Regularization of deep neural network using a multisample memory model
Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the...
Saved in:
Main Authors: | , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2025
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/182482 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-182482 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1824822025-02-04T05:30:32Z Regularization of deep neural network using a multisample memory model Tanveer, Muhammad Siyal, Mohammad Yakoob Rashid, Sheikh Faisal School of Electrical and Electronic Engineering Engineering Deeper architecture Overfitting Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin. 2025-02-04T05:30:31Z 2025-02-04T05:30:31Z 2024 Journal Article Tanveer, M., Siyal, M. Y. & Rashid, S. F. (2024). Regularization of deep neural network using a multisample memory model. Neural Computing and Applications, 36(36), 23295-23307. https://dx.doi.org/10.1007/s00521-024-10474-x 0941-0643 https://hdl.handle.net/10356/182482 10.1007/s00521-024-10474-x 2-s2.0-85206649904 36 36 23295 23307 en Neural Computing and Applications © 2024 The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Deeper architecture Overfitting |
spellingShingle |
Engineering Deeper architecture Overfitting Tanveer, Muhammad Siyal, Mohammad Yakoob Rashid, Sheikh Faisal Regularization of deep neural network using a multisample memory model |
description |
Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin. |
author2 |
School of Electrical and Electronic Engineering |
author_facet |
School of Electrical and Electronic Engineering Tanveer, Muhammad Siyal, Mohammad Yakoob Rashid, Sheikh Faisal |
format |
Article |
author |
Tanveer, Muhammad Siyal, Mohammad Yakoob Rashid, Sheikh Faisal |
author_sort |
Tanveer, Muhammad |
title |
Regularization of deep neural network using a multisample memory model |
title_short |
Regularization of deep neural network using a multisample memory model |
title_full |
Regularization of deep neural network using a multisample memory model |
title_fullStr |
Regularization of deep neural network using a multisample memory model |
title_full_unstemmed |
Regularization of deep neural network using a multisample memory model |
title_sort |
regularization of deep neural network using a multisample memory model |
publishDate |
2025 |
url |
https://hdl.handle.net/10356/182482 |
_version_ |
1823807384271716352 |