DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
Deep learning is a technique used in many domains to solve various problems. One of the popular example at the time of writing is ChatGPT, which uses GPT model made by OpenAI. A deep learning model consists of several layers, each with its own parameters. More complex models are developed by incr...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/74150 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
id |
id-itb.:74150 |
---|---|
spelling |
id-itb.:741502023-06-26T14:02:35ZDEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING Febryan Suryawan, Fransiskus Indonesia Final Project distributed deep learning, communication modification of Adam optimizer, weight compression, communication reduction INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/74150 Deep learning is a technique used in many domains to solve various problems. One of the popular example at the time of writing is ChatGPT, which uses GPT model made by OpenAI. A deep learning model consists of several layers, each with its own parameters. More complex models are developed by increasing the number of learned parameters. Increase of parameters in the model will cause communication overhead when learning in distributed architecture. Therefore, this final project will study about existing techniques to reduce communication overhead using reduction in commu- nication rounds and compression. This final project will review CADA (T. Chen dkk., 2021) and Efficient-Adam (C. Chen dkk., 2022). Then, this final project will incorporate ideas from both techniques, resulting in a technique which could reduce communications to as much as 0.97 times the communication rounds of CADA and using only 0.29 times the communication size of CADA. text |
institution |
Institut Teknologi Bandung |
building |
Institut Teknologi Bandung Library |
continent |
Asia |
country |
Indonesia Indonesia |
content_provider |
Institut Teknologi Bandung |
collection |
Digital ITB |
language |
Indonesia |
description |
Deep learning is a technique used in many domains to solve various problems. One
of the popular example at the time of writing is ChatGPT, which uses GPT model
made by OpenAI. A deep learning model consists of several layers, each with its
own parameters. More complex models are developed by increasing the number of
learned parameters.
Increase of parameters in the model will cause communication overhead when
learning in distributed architecture. Therefore, this final project will study about
existing techniques to reduce communication overhead using reduction in commu-
nication rounds and compression. This final project will review CADA (T. Chen
dkk., 2021) and Efficient-Adam (C. Chen dkk., 2022). Then, this final project will
incorporate ideas from both techniques, resulting in a technique which could reduce
communications to as much as 0.97 times the communication rounds of CADA and
using only 0.29 times the communication size of CADA. |
format |
Final Project |
author |
Febryan Suryawan, Fransiskus |
spellingShingle |
Febryan Suryawan, Fransiskus DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
author_facet |
Febryan Suryawan, Fransiskus |
author_sort |
Febryan Suryawan, Fransiskus |
title |
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
title_short |
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
title_full |
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
title_fullStr |
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
title_full_unstemmed |
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING |
title_sort |
development and implementation of communication modification on adam optimizer for distributed deep learning |
url |
https://digilib.itb.ac.id/gdl/view/74150 |
_version_ |
1822279794680659968 |