3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation

We propose and study a new class of gradient communication mechanisms for communication-efficient training -- three point compressors (3PC) -- as well as efficient distributed nonconvex optimization algorithms that can take advantage of them. Unlike most established approaches, which rely on a stati...

Full description

Saved in:
Bibliographic Details
Main Authors: RICHTARIK, Peter, SOKOLOV, Igor, FATKHULLIN, Ilyas, GASANOV, Elnur, LI, Zhize, GORBUNOV, Eduard
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8685
https://ink.library.smu.edu.sg/context/sis_research/article/9688/viewcontent/ICML22_full_3pc.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9688
record_format dspace
spelling sg-smu-ink.sis_research-96882024-03-28T09:02:49Z 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation RICHTARIK, Peter SOKOLOV, Igor FATKHULLIN, Ilyas GASANOV, Elnur LI, Zhize GORBUNOV, Eduard We propose and study a new class of gradient communication mechanisms for communication-efficient training -- three point compressors (3PC) -- as well as efficient distributed nonconvex optimization algorithms that can take advantage of them. Unlike most established approaches, which rely on a static compressor choice (e.g., Top-$K$), our class allows the compressors to {\em evolve} throughout the training process, with the aim of improving the theoretical communication complexity and practical efficiency of the underlying methods. We show that our general approach can recover the recently proposed state-of-the-art error feedback mechanism EF21 (Richt\'arik et al., 2021) and its theoretical properties as a special case, but also leads to a number of new efficient methods. Notably, our approach allows us to improve upon the state of the art in the algorithmic and theoretical foundations of the {\em lazy aggregation} literature (Chen et al., 2018). As a by-product that may be of independent interest, we provide a new and fundamental link between the lazy aggregation and error feedback literature. A special feature of our work is that we do not require the compressors to be unbiased. 2022-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8685 https://ink.library.smu.edu.sg/context/sis_research/article/9688/viewcontent/ICML22_full_3pc.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
spellingShingle Databases and Information Systems
RICHTARIK, Peter
SOKOLOV, Igor
FATKHULLIN, Ilyas
GASANOV, Elnur
LI, Zhize
GORBUNOV, Eduard
3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
description We propose and study a new class of gradient communication mechanisms for communication-efficient training -- three point compressors (3PC) -- as well as efficient distributed nonconvex optimization algorithms that can take advantage of them. Unlike most established approaches, which rely on a static compressor choice (e.g., Top-$K$), our class allows the compressors to {\em evolve} throughout the training process, with the aim of improving the theoretical communication complexity and practical efficiency of the underlying methods. We show that our general approach can recover the recently proposed state-of-the-art error feedback mechanism EF21 (Richt\'arik et al., 2021) and its theoretical properties as a special case, but also leads to a number of new efficient methods. Notably, our approach allows us to improve upon the state of the art in the algorithmic and theoretical foundations of the {\em lazy aggregation} literature (Chen et al., 2018). As a by-product that may be of independent interest, we provide a new and fundamental link between the lazy aggregation and error feedback literature. A special feature of our work is that we do not require the compressors to be unbiased.
format text
author RICHTARIK, Peter
SOKOLOV, Igor
FATKHULLIN, Ilyas
GASANOV, Elnur
LI, Zhize
GORBUNOV, Eduard
author_facet RICHTARIK, Peter
SOKOLOV, Igor
FATKHULLIN, Ilyas
GASANOV, Elnur
LI, Zhize
GORBUNOV, Eduard
author_sort RICHTARIK, Peter
title 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
title_short 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
title_full 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
title_fullStr 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
title_full_unstemmed 3PC: Three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
title_sort 3pc: three point compressors for communication-efficient distributed training and a better theory for lazy aggregation
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/8685
https://ink.library.smu.edu.sg/context/sis_research/article/9688/viewcontent/ICML22_full_3pc.pdf
_version_ 1795302172435415040