Decoupled neural network training with re-computation and weight prediction

To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a n...

Full description

Saved in:
Bibliographic Details
Main Authors: Peng, Jiawei, Xu, Yicheng, Lin, Zhiping, Weng, Zhenyu, Yang, Zishuo, Zhuang, Huiping
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/169712
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-169712
record_format dspace
spelling sg-ntu-dr.10356-1697122023-08-04T15:40:02Z Decoupled neural network training with re-computation and weight prediction Peng, Jiawei Xu, Yicheng Lin, Zhiping Weng, Zhenyu Yang, Zishuo Zhuang, Huiping School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Back Propagation Convolutional Neural Network To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved. Agency for Science, Technology and Research (A*STAR) Nanyang Technological University Published version We wish to acknowledge the funding support for this project from Nanyang Technological University under the URECA Undergraduate Research Programme. This work was also supported in part by the Science and Engineering Research Council, Agency of Science, Technology and Research, Singapore, through the National Robotics Program under Grant 1922500054. 2023-08-01T01:47:12Z 2023-08-01T01:47:12Z 2023 Journal Article Peng, J., Xu, Y., Lin, Z., Weng, Z., Yang, Z. & Zhuang, H. (2023). Decoupled neural network training with re-computation and weight prediction. PloS One, 18(2), e0276427-. https://dx.doi.org/10.1371/journal.pone.0276427 1932-6203 https://hdl.handle.net/10356/169712 10.1371/journal.pone.0276427 36821537 2-s2.0-85148682326 2 18 e0276427 en 1922500054 URECA PloS one © 2023 Peng et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Back Propagation
Convolutional Neural Network
spellingShingle Engineering::Electrical and electronic engineering
Back Propagation
Convolutional Neural Network
Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
Decoupled neural network training with re-computation and weight prediction
description To break the three lockings during backpropagation (BP) process for neural network training, multiple decoupled learning methods have been investigated recently. These methods either lead to significant drop in accuracy performance or suffer from dramatic increase in memory usage. In this paper, a new form of decoupled learning, named decoupled neural network training scheme with re-computation and weight prediction (DTRP) is proposed. In DTRP, a re-computation scheme is adopted to solve the memory explosion problem, and a weight prediction scheme is proposed to deal with the weight delay caused by re-computation. Additionally, a batch compensation scheme is developed, allowing the proposed DTRP to run faster. Theoretical analysis shows that DTRP is guaranteed to converge to crical points under certain conditions. Experiments are conducted by training various convolutional neural networks on several classification datasets, showing comparable or better results than the state-of-the-art methods and BP. These experiments also reveal that adopting the proposed method, the memory explosion problem is effectively solved, and a significant acceleration is achieved.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
format Article
author Peng, Jiawei
Xu, Yicheng
Lin, Zhiping
Weng, Zhenyu
Yang, Zishuo
Zhuang, Huiping
author_sort Peng, Jiawei
title Decoupled neural network training with re-computation and weight prediction
title_short Decoupled neural network training with re-computation and weight prediction
title_full Decoupled neural network training with re-computation and weight prediction
title_fullStr Decoupled neural network training with re-computation and weight prediction
title_full_unstemmed Decoupled neural network training with re-computation and weight prediction
title_sort decoupled neural network training with re-computation and weight prediction
publishDate 2023
url https://hdl.handle.net/10356/169712
_version_ 1773551347369508864