Convolutional neural networks with dynamic regularization

Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement in the generalization performance. However, these methods lack a self-adaptive abilit...

Full description

Saved in:
Bibliographic Details
Main Authors: Wang, Yi, Bian, Zhen-Peng, Hou, Junhui, Chau, Lap-Pui
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2022
Subjects:
Online Access:https://hdl.handle.net/10356/159626
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-159626
record_format dspace
spelling sg-ntu-dr.10356-1596262022-06-28T08:09:44Z Convolutional neural networks with dynamic regularization Wang, Yi Bian, Zhen-Peng Hou, Junhui Chau, Lap-Pui School of Electrical and Electronic Engineering Engineering::Electrical and electronic engineering Convolutional Neural Network Generalization Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement in the generalization performance. However, these methods lack a self-adaptive ability throughout training. That is, the regularization strength is fixed to a predefined schedule, and manual adjustments are required to adapt to various network architectures. In this article, we propose a dynamic regularization method for CNNs. Specifically, we model the regularization strength as a function of the training loss. According to the change of the training loss, our method can dynamically adjust the regularization strength in the training procedure, thereby balancing the underfitting and overfitting of CNNs. With dynamic regularization, a large-scale model is automatically regularized by the strong perturbation, and vice versa. Experimental results show that the proposed method can improve the generalization capability on off-the-shelf network architectures and outperform state-of-the-art regularization methods. This work was supported in part by the Hong Kong Research Grants Council under Grant 9048123 (CityU 21211518) and Grant 9042820 (CityU 11219019). 2022-06-28T08:09:43Z 2022-06-28T08:09:43Z 2020 Journal Article Wang, Y., Bian, Z., Hou, J. & Chau, L. (2020). Convolutional neural networks with dynamic regularization. IEEE Transactions On Neural Networks and Learning Systems, 32(5), 2299-2304. https://dx.doi.org/10.1109/TNNLS.2020.2997044 2162-2388 https://hdl.handle.net/10356/159626 10.1109/TNNLS.2020.2997044 32511095 2-s2.0-85105605068 5 32 2299 2304 en IEEE Transactions on Neural Networks and Learning Systems © 2020 IEEE. All rights reserved.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Electrical and electronic engineering
Convolutional Neural Network
Generalization
spellingShingle Engineering::Electrical and electronic engineering
Convolutional Neural Network
Generalization
Wang, Yi
Bian, Zhen-Peng
Hou, Junhui
Chau, Lap-Pui
Convolutional neural networks with dynamic regularization
description Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement in the generalization performance. However, these methods lack a self-adaptive ability throughout training. That is, the regularization strength is fixed to a predefined schedule, and manual adjustments are required to adapt to various network architectures. In this article, we propose a dynamic regularization method for CNNs. Specifically, we model the regularization strength as a function of the training loss. According to the change of the training loss, our method can dynamically adjust the regularization strength in the training procedure, thereby balancing the underfitting and overfitting of CNNs. With dynamic regularization, a large-scale model is automatically regularized by the strong perturbation, and vice versa. Experimental results show that the proposed method can improve the generalization capability on off-the-shelf network architectures and outperform state-of-the-art regularization methods.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Wang, Yi
Bian, Zhen-Peng
Hou, Junhui
Chau, Lap-Pui
format Article
author Wang, Yi
Bian, Zhen-Peng
Hou, Junhui
Chau, Lap-Pui
author_sort Wang, Yi
title Convolutional neural networks with dynamic regularization
title_short Convolutional neural networks with dynamic regularization
title_full Convolutional neural networks with dynamic regularization
title_fullStr Convolutional neural networks with dynamic regularization
title_full_unstemmed Convolutional neural networks with dynamic regularization
title_sort convolutional neural networks with dynamic regularization
publishDate 2022
url https://hdl.handle.net/10356/159626
_version_ 1738844881128783872