Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures

Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, i...

Full description

Saved in:
Bibliographic Details
Main Author: Kow, Li Ren
Other Authors: Jiang Xudong
Format: Final Year Project
Language:English
Published: 2018
Subjects:
Online Access:http://hdl.handle.net/10356/74869
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-74869
record_format dspace
spelling sg-ntu-dr.10356-748692023-07-07T16:46:21Z Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures Kow, Li Ren Jiang Xudong School of Electrical and Electronic Engineering NVIDIA DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU) can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration. Bachelor of Engineering 2018-05-24T07:10:10Z 2018-05-24T07:10:10Z 2018 Final Year Project (FYP) http://hdl.handle.net/10356/74869 en Nanyang Technological University 64 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
spellingShingle DRNTU::Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Kow, Li Ren
Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
description Neural networks get more difficult and longer time to train if the depth become deeper. As deep neural network going deeper, it has dominated mostly all the pattern recognition algorithm and application, especially on Natural Language Processing and computer vision. To train a deep neural network, it involves a lot of floating point matrix calculation and it will be time consuming training on a computer processing unit (CPU). Even graphic processing unit (GPU) can do better in floating point calculation but it still takes long time to complete the training if the dataset is large and models are deep. Hence, multiple GPU card could be used in parallel to accelerate the entire training process. It is important to understand how fast it can be with different kind of deep learning framework which include (Mxnet, Pytorch and Caffe2) and the key software and hardware factor in this parallel training process on a single node or multi node configuration.
author2 Jiang Xudong
author_facet Jiang Xudong
Kow, Li Ren
format Final Year Project
author Kow, Li Ren
author_sort Kow, Li Ren
title Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
title_short Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
title_full Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
title_fullStr Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
title_full_unstemmed Benchmarking of the popular DL Frameworks over multiple GPU cards on state-of-the-art CNN architectures
title_sort benchmarking of the popular dl frameworks over multiple gpu cards on state-of-the-art cnn architectures
publishDate 2018
url http://hdl.handle.net/10356/74869
_version_ 1772826171682062336