RobOT: Robustness-oriented testing for deep learning systems

Recently, there has been a significant growth of interest in applying software engineering techniques for the quality assurance of deep learning (DL) systems. One popular direction is deep learning testing, where adversarial examples (a.k.a. bugs) of DL systems are found either by fuzzing or guided...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Jingyi, CHEN, Jialuo, SUN, Youcheng, MA, Xingjun, WANG, Dongxia, SUN, Jun, CHENG, Peng
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/6056
https://ink.library.smu.edu.sg/context/sis_research/article/7059/viewcontent/sample_sigconf.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-7059
record_format dspace
spelling sg-smu-ink.sis_research-70592021-08-25T08:51:45Z RobOT: Robustness-oriented testing for deep learning systems WANG, Jingyi CHEN, Jialuo SUN, Youcheng MA, Xingjun WANG, Dongxia SUN, Jun CHENG, Peng Recently, there has been a significant growth of interest in applying software engineering techniques for the quality assurance of deep learning (DL) systems. One popular direction is deep learning testing, where adversarial examples (a.k.a. bugs) of DL systems are found either by fuzzing or guided search with the help of certain testing metrics. However, recent studies have revealed that the commonly used neuron coverage metrics by existing DL testing approaches are not correlated to model robustness. It is also not an effective measurement on the confidence of the model robustness after testing. In this work, we address this gap by proposing a novel testing framework called Robustness-Oriented Testing (RobOT). A key part of RobOT is a quantitative measurement on 1) the value of each test case in improving model robustness (often via retraining), and 2) the convergence quality of the model robustness improvement. RobOT utilizes the proposed metric to automatically generate test cases valuable for improving model robustness. The proposed metric is also a strong indicator on how well robustness improvement has converged through testing. Experiments on multiple benchmark datasets confirm the effectiveness and efficiency of RobOT in improving DL model robustness, with 67.02% increase on the adversarial robustness that is 50.65% higher than the state-of-the-art work DeepGini. 2021-05-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6056 info:doi/10.1109/ICSE43902.2021.00038 https://ink.library.smu.edu.sg/context/sis_research/article/7059/viewcontent/sample_sigconf.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Measurement Deep learning Benchmark testing Robustness Robots Software engineering Convergence Numerical Analysis and Scientific Computing Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Measurement
Deep learning
Benchmark testing
Robustness
Robots
Software engineering
Convergence
Numerical Analysis and Scientific Computing
Software Engineering
spellingShingle Measurement
Deep learning
Benchmark testing
Robustness
Robots
Software engineering
Convergence
Numerical Analysis and Scientific Computing
Software Engineering
WANG, Jingyi
CHEN, Jialuo
SUN, Youcheng
MA, Xingjun
WANG, Dongxia
SUN, Jun
CHENG, Peng
RobOT: Robustness-oriented testing for deep learning systems
description Recently, there has been a significant growth of interest in applying software engineering techniques for the quality assurance of deep learning (DL) systems. One popular direction is deep learning testing, where adversarial examples (a.k.a. bugs) of DL systems are found either by fuzzing or guided search with the help of certain testing metrics. However, recent studies have revealed that the commonly used neuron coverage metrics by existing DL testing approaches are not correlated to model robustness. It is also not an effective measurement on the confidence of the model robustness after testing. In this work, we address this gap by proposing a novel testing framework called Robustness-Oriented Testing (RobOT). A key part of RobOT is a quantitative measurement on 1) the value of each test case in improving model robustness (often via retraining), and 2) the convergence quality of the model robustness improvement. RobOT utilizes the proposed metric to automatically generate test cases valuable for improving model robustness. The proposed metric is also a strong indicator on how well robustness improvement has converged through testing. Experiments on multiple benchmark datasets confirm the effectiveness and efficiency of RobOT in improving DL model robustness, with 67.02% increase on the adversarial robustness that is 50.65% higher than the state-of-the-art work DeepGini.
format text
author WANG, Jingyi
CHEN, Jialuo
SUN, Youcheng
MA, Xingjun
WANG, Dongxia
SUN, Jun
CHENG, Peng
author_facet WANG, Jingyi
CHEN, Jialuo
SUN, Youcheng
MA, Xingjun
WANG, Dongxia
SUN, Jun
CHENG, Peng
author_sort WANG, Jingyi
title RobOT: Robustness-oriented testing for deep learning systems
title_short RobOT: Robustness-oriented testing for deep learning systems
title_full RobOT: Robustness-oriented testing for deep learning systems
title_fullStr RobOT: Robustness-oriented testing for deep learning systems
title_full_unstemmed RobOT: Robustness-oriented testing for deep learning systems
title_sort robot: robustness-oriented testing for deep learning systems
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/6056
https://ink.library.smu.edu.sg/context/sis_research/article/7059/viewcontent/sample_sigconf.pdf
_version_ 1770575776139182080