Optimization for efficient data communication in distributed machine training system
The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes....
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2017
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/72923 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-72923 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-729232023-03-03T20:32:42Z Optimization for efficient data communication in distributed machine training system Gan, Hsien Yan Wen Yonggang School of Computer Science and Engineering DRNTU::Engineering::Computer science and engineering The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes. Compared to interprocess communication, data exchange between nodes is relatively slow, high latency and high overhead cost. When the network link is shared among multiple nodes, limited bandwidth arises, which is a more undesirable property. This project is to minimize the data flow between nodes by adding a data filter and Snappy compression. The filter reduces the unnecessary data flow while the Snappy does data compression to reduce bandwidth consumption. This implementation successfully reduces the data flow to 8 percent and decrease training time to 76 percent. Due to the low required bandwidth, distributed system on different geographical area and hardware such as a mobile laptop is possible. Bachelor of Engineering (Computer Engineering) 2017-12-12T09:21:03Z 2017-12-12T09:21:03Z 2017 Final Year Project (FYP) http://hdl.handle.net/10356/72923 en Nanyang Technological University 47 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering |
spellingShingle |
DRNTU::Engineering::Computer science and engineering Gan, Hsien Yan Optimization for efficient data communication in distributed machine training system |
description |
The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes. Compared to interprocess communication, data exchange between nodes is relatively slow, high latency and high overhead cost. When the network link is shared among multiple nodes, limited bandwidth arises, which is a more undesirable property. This project is to minimize the data flow between nodes by adding a data filter and Snappy compression. The filter reduces the unnecessary data flow while the Snappy does data compression to reduce bandwidth consumption. This implementation successfully reduces the data flow to 8 percent and decrease training time to 76 percent. Due to the low required bandwidth, distributed system on different geographical area and hardware such as a mobile laptop is possible. |
author2 |
Wen Yonggang |
author_facet |
Wen Yonggang Gan, Hsien Yan |
format |
Final Year Project |
author |
Gan, Hsien Yan |
author_sort |
Gan, Hsien Yan |
title |
Optimization for efficient data communication in distributed machine training system |
title_short |
Optimization for efficient data communication in distributed machine training system |
title_full |
Optimization for efficient data communication in distributed machine training system |
title_fullStr |
Optimization for efficient data communication in distributed machine training system |
title_full_unstemmed |
Optimization for efficient data communication in distributed machine training system |
title_sort |
optimization for efficient data communication in distributed machine training system |
publishDate |
2017 |
url |
http://hdl.handle.net/10356/72923 |
_version_ |
1759853426017042432 |