Robust test selection for deep neural networks

Deep Neural Networks (DNNs) have been widely used in various domains, such as computer vision and software engineering. Although many DNNs have been deployed to assist various tasks in the real world, similar to traditional software, they also suffer from defects that may lead to severe outcomes. DN...

Full description

Saved in:
Bibliographic Details
Main Authors: SUN, Weifeng, YAN, Meng, LIU, Zhongxin, LO, David
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2023
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/8374
https://ink.library.smu.edu.sg/context/sis_research/article/9377/viewcontent/RobustTextSelectionDNN_av.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-9377
record_format dspace
spelling sg-smu-ink.sis_research-93772024-01-12T02:45:34Z Robust test selection for deep neural networks SUN, Weifeng YAN, Meng LIU, Zhongxin LO, David Deep Neural Networks (DNNs) have been widely used in various domains, such as computer vision and software engineering. Although many DNNs have been deployed to assist various tasks in the real world, similar to traditional software, they also suffer from defects that may lead to severe outcomes. DNN testing is one of the most widely used methods to ensure the quality of DNNs. Such method needs rich test inputs with oracle information (expected output) to reveal the incorrect behaviors of a DNN model. However, manually labeling all the collected test inputs is a labor-intensive task, which delays the quality assurance process. tackles this problem by carefully selecting a small, more suspicious set of test inputs to label, enabling the failure detection of a DNN model with reduced effort. Researchers have proposed different test selection methods, including neuron-coverage-based and uncertainty-based methods, where the uncertainty-based method is arguably the most popular technique. Unfortunately, existing uncertainty-based selection methods meet the performance bottleneck due to one or several limitations: 1) they ignore noisy data in real scenarios; 2) they wrongly exclude many but rather include many (referring to those test inputs that are correctly predicted by the model); 3) they ignore the diversity of the selected test set. In this paper, we propose RTS, a obust est election method for deep neural networks to overcome the limitations mentioned above. First, RTS divides all unlabeled candidate test inputs into noise set, successful set, and suspicious set and assigns different selection prioritization to divided sets, which effectively alleviates the impact of noise and improves the ability to identify suspect test inputs. Subsequently, RTS leverages a probability-tier-matrix-based test metric for prioritizing the test inputs in each divided set (i.e., suspicious, successful, and noise set). As a result, RTS can select more suspicious test inputs within a limited selection size. We evaluate RTS by comparing it with 14 baseline methods under 5 widely-used DNN models and 6 widely-used datasets. The experimental results demonstrate that RTS can significantly outperform all test selection methods in failure detection capability and the test suites selected by RTS have the best model optimization capability. For example, when selecting 2.5% test input, RTS achieves an improvement of 9.37%-176.75% over baseline methods in terms of failure detection. 2023-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8374 info:doi/10.1109/TSE.2023.3330982 https://ink.library.smu.edu.sg/context/sis_research/article/9377/viewcontent/RobustTextSelectionDNN_av.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial neural networks Data models Deep learning testing deep neural networks Labeling Neurons Noise measurement Task analysis test selection Testing OS and Networks Software Engineering Theory and Algorithms
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial neural networks
Data models
Deep learning testing
deep neural networks
Labeling
Neurons
Noise measurement
Task analysis
test selection
Testing
OS and Networks
Software Engineering
Theory and Algorithms
spellingShingle Artificial neural networks
Data models
Deep learning testing
deep neural networks
Labeling
Neurons
Noise measurement
Task analysis
test selection
Testing
OS and Networks
Software Engineering
Theory and Algorithms
SUN, Weifeng
YAN, Meng
LIU, Zhongxin
LO, David
Robust test selection for deep neural networks
description Deep Neural Networks (DNNs) have been widely used in various domains, such as computer vision and software engineering. Although many DNNs have been deployed to assist various tasks in the real world, similar to traditional software, they also suffer from defects that may lead to severe outcomes. DNN testing is one of the most widely used methods to ensure the quality of DNNs. Such method needs rich test inputs with oracle information (expected output) to reveal the incorrect behaviors of a DNN model. However, manually labeling all the collected test inputs is a labor-intensive task, which delays the quality assurance process. tackles this problem by carefully selecting a small, more suspicious set of test inputs to label, enabling the failure detection of a DNN model with reduced effort. Researchers have proposed different test selection methods, including neuron-coverage-based and uncertainty-based methods, where the uncertainty-based method is arguably the most popular technique. Unfortunately, existing uncertainty-based selection methods meet the performance bottleneck due to one or several limitations: 1) they ignore noisy data in real scenarios; 2) they wrongly exclude many but rather include many (referring to those test inputs that are correctly predicted by the model); 3) they ignore the diversity of the selected test set. In this paper, we propose RTS, a obust est election method for deep neural networks to overcome the limitations mentioned above. First, RTS divides all unlabeled candidate test inputs into noise set, successful set, and suspicious set and assigns different selection prioritization to divided sets, which effectively alleviates the impact of noise and improves the ability to identify suspect test inputs. Subsequently, RTS leverages a probability-tier-matrix-based test metric for prioritizing the test inputs in each divided set (i.e., suspicious, successful, and noise set). As a result, RTS can select more suspicious test inputs within a limited selection size. We evaluate RTS by comparing it with 14 baseline methods under 5 widely-used DNN models and 6 widely-used datasets. The experimental results demonstrate that RTS can significantly outperform all test selection methods in failure detection capability and the test suites selected by RTS have the best model optimization capability. For example, when selecting 2.5% test input, RTS achieves an improvement of 9.37%-176.75% over baseline methods in terms of failure detection.
format text
author SUN, Weifeng
YAN, Meng
LIU, Zhongxin
LO, David
author_facet SUN, Weifeng
YAN, Meng
LIU, Zhongxin
LO, David
author_sort SUN, Weifeng
title Robust test selection for deep neural networks
title_short Robust test selection for deep neural networks
title_full Robust test selection for deep neural networks
title_fullStr Robust test selection for deep neural networks
title_full_unstemmed Robust test selection for deep neural networks
title_sort robust test selection for deep neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2023
url https://ink.library.smu.edu.sg/sis_research/8374
https://ink.library.smu.edu.sg/context/sis_research/article/9377/viewcontent/RobustTextSelectionDNN_av.pdf
_version_ 1789483242028531712