Optimization for efficient data communication in distributed machine training system

The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes....

Full description

Saved in:
Bibliographic Details
Main Author: Gan, Hsien Yan
Other Authors: Wen Yonggang
Format: Final Year Project
Language:English
Published: 2017
Subjects:
Online Access:http://hdl.handle.net/10356/72923
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-72923
record_format dspace
spelling sg-ntu-dr.10356-729232023-03-03T20:32:42Z Optimization for efficient data communication in distributed machine training system Gan, Hsien Yan Wen Yonggang School of Computer Science and Engineering DRNTU::Engineering::Computer science and engineering The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes. Compared to interprocess communication, data exchange between nodes is relatively slow, high latency and high overhead cost. When the network link is shared among multiple nodes, limited bandwidth arises, which is a more undesirable property. This project is to minimize the data flow between nodes by adding a data filter and Snappy compression. The filter reduces the unnecessary data flow while the Snappy does data compression to reduce bandwidth consumption. This implementation successfully reduces the data flow to 8 percent and decrease training time to 76 percent. Due to the low required bandwidth, distributed system on different geographical area and hardware such as a mobile laptop is possible. Bachelor of Engineering (Computer Engineering) 2017-12-12T09:21:03Z 2017-12-12T09:21:03Z 2017 Final Year Project (FYP) http://hdl.handle.net/10356/72923 en Nanyang Technological University 47 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering
spellingShingle DRNTU::Engineering::Computer science and engineering
Gan, Hsien Yan
Optimization for efficient data communication in distributed machine training system
description The rising trend of deep learning causes the complexity and scale of machine learning to increase exponentially. But, the complexity is limited by hardware processing speed. To solve the issue, there are a few machine learning frameworks online, which support distributed training on multiple nodes. Compared to interprocess communication, data exchange between nodes is relatively slow, high latency and high overhead cost. When the network link is shared among multiple nodes, limited bandwidth arises, which is a more undesirable property. This project is to minimize the data flow between nodes by adding a data filter and Snappy compression. The filter reduces the unnecessary data flow while the Snappy does data compression to reduce bandwidth consumption. This implementation successfully reduces the data flow to 8 percent and decrease training time to 76 percent. Due to the low required bandwidth, distributed system on different geographical area and hardware such as a mobile laptop is possible.
author2 Wen Yonggang
author_facet Wen Yonggang
Gan, Hsien Yan
format Final Year Project
author Gan, Hsien Yan
author_sort Gan, Hsien Yan
title Optimization for efficient data communication in distributed machine training system
title_short Optimization for efficient data communication in distributed machine training system
title_full Optimization for efficient data communication in distributed machine training system
title_fullStr Optimization for efficient data communication in distributed machine training system
title_full_unstemmed Optimization for efficient data communication in distributed machine training system
title_sort optimization for efficient data communication in distributed machine training system
publishDate 2017
url http://hdl.handle.net/10356/72923
_version_ 1759853426017042432