Decoupled neural network training with re-computation and weight prediction
To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a n...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/169712 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-169712 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1697122023-08-04T15:40:02Z Decoupled neural network training with re-computation and weight prediction Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Back Propagation Convolutional Neural Network To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. Agency for Science, Technology and Research (A*STAR) Nanyang Technological University Published version We wish to acknowledge the funding support for this project from Nanyang Technological University under the URECA Undergraduate Research Programme. This work was also supported in part by the Science and Engineering Research Council, Agency of Science, Technology and Research, Singapore, through the National Robotics Program under Grant 1922500054. 2023-08-01T01:47:12Z 2023-08-01T01:47:12Z 2023 Journal Article Peng, J., Xu, Y., Lin, Z., Weng, Z., Yang, Z. & Zhuang, H. (2023). Decoupled neural network training with re-computation and weight prediction. PloS One, 18(2), e0276427-. https://dx.doi.org/10.1371/journal.pone.0276427 1932-6203 https://hdl.handle.net/10356/169712 10.1371/journal.pone.0276427 36821537 2-s2.0-85148682326 2 18 e0276427 en 1922500054 URECA PloS one © 2023 Peng et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Electrical and electronic engineering Back Propagation Convolutional Neural Network |
spellingShingle |
Engineering::Electrical and electronic engineering Back Propagation Convolutional Neural Network Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping Decoupled neural network training with re-computation and weight prediction |
description |
To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping |
format |
Article |
author |
Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping |
author_sort |
Peng, Jiawei |
title |
Decoupled neural network training with re-computation and weight prediction |
title_short |
Decoupled neural network training with re-computation and weight prediction |
title_full |
Decoupled neural network training with re-computation and weight prediction |
title_fullStr |
Decoupled neural network training with re-computation and weight prediction |
title_full_unstemmed |
Decoupled neural network training with re-computation and weight prediction |
title_sort |
decoupled neural network training with re-computation and weight prediction |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/169712 |
_version_ |
1773551347369508864 |