Faster rates for compressed federated learning with client-variance reduction
Due to the communication bottleneck in distributed and federated learning applications, algorithms using communication compression have attracted significant attention and are widely used in practice. Moreover, the huge number, high heterogeneity, and limited availability of clients result in high c...
Saved in:
Main Authors: | ZHAO, Haoyu, BURLACHENKO, Konstantin, LI, Zhize, RICHTARIK, Peter |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2024
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9607 https://ink.library.smu.edu.sg/context/sis_research/article/10607/viewcontent/SIMODS24_cofig_av.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
MARINA: Faster non-convex distributed learning with compression
by: GORBUNOV, Eduard, et al.
Published: (2021) -
Escaping saddle points in heterogeneous federated learning via distributed SGD with communication compression
by: CHEN, Sijin, et al.
Published: (2024) -
Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference
by: LI, Zhize, et al.
Published: (2019) -
Reduction of microwave amplifier gain variance
by: Eccleston, K.W., et al.
Published: (2014) -
Joint client-and-sample selection for federated learning via bi-level optimization
by: Li, Anran, et al.
Published: (2024)