Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, i...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2018
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/74869 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-74869 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-748692023-07-07T16:46:21Z Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures Kow, Li Ren Jiang Xudong School of Electrical and Electronic Engineering NVIDIA DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU) can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration. Bachelor of Engineering 2018-05-24T07:10:10Z 2018-05-24T07:10:10Z 2018 Final Year Project (FYP) http://hdl.handle.net/10356/74869 en Nanyang Technological University 64 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence |
spellingShingle |
DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Kow, Li Ren Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
description |
Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU)
can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration. |
author2 |
Jiang Xudong |
author_facet |
Jiang Xudong Kow, Li Ren |
format |
Final Year Project |
author |
Kow, Li Ren |
author_sort |
Kow, Li Ren |
title |
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
title_short |
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
title_full |
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
title_fullStr |
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
title_full_unstemmed |
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures |
title_sort |
benchmarking of the popular dl frameworks over multiple gpu cards on state-of-the-art cnn architectures |
publishDate |
2018 |
url |
http://hdl.handle.net/10356/74869 |
_version_ |
1772826171682062336 |