Journal article

Efficient variance-reduced learning for fully decentralized on-device intelligence

This work develops a fully decentralized variance-reduced learning algorithm for on-device intelligence where nodes store and process the data locally and are only allowed to communicate with their immediate neighbors. In the proposed algorithm, there is no need for a central or master unit while the objective is to enable the dispersed nodes to learn the exact global model despite their limited localized interactions. The resulting algorithm is shown to have low memory requirement, guaranteed linear convergence, robustness to failure of links or nodes, scalability to the network size, and privacy-preserving properties. Moreover, the decentralized nature of the solution makes large-scale machine learning problems more tractable and also scalable since data is stored and processed locally at the nodes


  • There is no available fulltext. Please contact the lab or the authors.

Related material