Regularization of deep neural network using a multisample memory model

Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the...

Full description

Saved in:
Bibliographic Details
Main Authors: Tanveer, Muhammad, Siyal, Mohammad Yakoob, Rashid, Sheikh Faisal
Other Authors: School of Electrical and Electronic Engineering
Format: Article
Language:English
Published: 2025
Subjects:
Online Access:https://hdl.handle.net/10356/182482
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-182482
record_format dspace
spelling sg-ntu-dr.10356-1824822025-02-04T05:30:32Z Regularization of deep neural network using a multisample memory model Tanveer, Muhammad Siyal, Mohammad Yakoob Rashid, Sheikh Faisal School of Electrical and Electronic Engineering Engineering Deeper architecture Overfitting Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin. 2025-02-04T05:30:31Z 2025-02-04T05:30:31Z 2024 Journal Article Tanveer, M., Siyal, M. Y. & Rashid, S. F. (2024). Regularization of deep neural network using a multisample memory model. Neural Computing and Applications, 36(36), 23295-23307. https://dx.doi.org/10.1007/s00521-024-10474-x 0941-0643 https://hdl.handle.net/10356/182482 10.1007/s00521-024-10474-x 2-s2.0-85206649904 36 36 23295 23307 en Neural Computing and Applications © 2024 The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature.
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Deeper architecture
Overfitting
spellingShingle Engineering
Deeper architecture
Overfitting
Tanveer, Muhammad
Siyal, Mohammad Yakoob
Rashid, Sheikh Faisal
Regularization of deep neural network using a multisample memory model
description Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin.
author2 School of Electrical and Electronic Engineering
author_facet School of Electrical and Electronic Engineering
Tanveer, Muhammad
Siyal, Mohammad Yakoob
Rashid, Sheikh Faisal
format Article
author Tanveer, Muhammad
Siyal, Mohammad Yakoob
Rashid, Sheikh Faisal
author_sort Tanveer, Muhammad
title Regularization of deep neural network using a multisample memory model
title_short Regularization of deep neural network using a multisample memory model
title_full Regularization of deep neural network using a multisample memory model
title_fullStr Regularization of deep neural network using a multisample memory model
title_full_unstemmed Regularization of deep neural network using a multisample memory model
title_sort regularization of deep neural network using a multisample memory model
publishDate 2025
url https://hdl.handle.net/10356/182482
_version_ 1823807384271716352