DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING

Deep learning is a technique used in many domains to solve various problems. One of the popular example at the time of writing is ChatGPT, which uses GPT model made by OpenAI. A deep learning model consists of several layers, each with its own parameters. More complex models are developed by incr...

Full description

Saved in:
Bibliographic Details
Main Author: Febryan Suryawan, Fransiskus
Format: Final Project
Language:Indonesia
Online Access:https://digilib.itb.ac.id/gdl/view/74150
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Institut Teknologi Bandung
Language: Indonesia
id id-itb.:74150
spelling id-itb.:741502023-06-26T14:02:35ZDEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING Febryan Suryawan, Fransiskus Indonesia Final Project distributed deep learning, communication modification of Adam optimizer, weight compression, communication reduction INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/74150 Deep learning is a technique used in many domains to solve various problems. One of the popular example at the time of writing is ChatGPT, which uses GPT model made by OpenAI. A deep learning model consists of several layers, each with its own parameters. More complex models are developed by increasing the number of learned parameters. Increase of parameters in the model will cause communication overhead when learning in distributed architecture. Therefore, this final project will study about existing techniques to reduce communication overhead using reduction in commu- nication rounds and compression. This final project will review CADA (T. Chen dkk., 2021) and Efficient-Adam (C. Chen dkk., 2022). Then, this final project will incorporate ideas from both techniques, resulting in a technique which could reduce communications to as much as 0.97 times the communication rounds of CADA and using only 0.29 times the communication size of CADA. text
institution Institut Teknologi Bandung
building Institut Teknologi Bandung Library
continent Asia
country Indonesia
Indonesia
content_provider Institut Teknologi Bandung
collection Digital ITB
language Indonesia
description Deep learning is a technique used in many domains to solve various problems. One of the popular example at the time of writing is ChatGPT, which uses GPT model made by OpenAI. A deep learning model consists of several layers, each with its own parameters. More complex models are developed by increasing the number of learned parameters. Increase of parameters in the model will cause communication overhead when learning in distributed architecture. Therefore, this final project will study about existing techniques to reduce communication overhead using reduction in commu- nication rounds and compression. This final project will review CADA (T. Chen dkk., 2021) and Efficient-Adam (C. Chen dkk., 2022). Then, this final project will incorporate ideas from both techniques, resulting in a technique which could reduce communications to as much as 0.97 times the communication rounds of CADA and using only 0.29 times the communication size of CADA.
format Final Project
author Febryan Suryawan, Fransiskus
spellingShingle Febryan Suryawan, Fransiskus
DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
author_facet Febryan Suryawan, Fransiskus
author_sort Febryan Suryawan, Fransiskus
title DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
title_short DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
title_full DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
title_fullStr DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
title_full_unstemmed DEVELOPMENT AND IMPLEMENTATION OF COMMUNICATION MODIFICATION ON ADAM OPTIMIZER FOR DISTRIBUTED DEEP LEARNING
title_sort development and implementation of communication modification on adam optimizer for distributed deep learning
url https://digilib.itb.ac.id/gdl/view/74150
_version_ 1822279794680659968