Cats are not fish: Deep learning testing calls for out-of-distribution awareness

As Deep Learning (DL) is continuously adopted in many industrial applications, its quality and reliability start to raise concerns. Similar to the traditional software development process, testing the DL software to uncover its defects at an early stage is an effective way to reduce risks after depl...

Full description

Saved in:
Bibliographic Details
Main Authors: BEREND, David, XIE, Xiaofei, MA, Lei, ZHOU, Lingjun, LIU, Yang, XU, Chi, ZHAO, Jianjun
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7076
https://ink.library.smu.edu.sg/context/sis_research/article/8079/viewcontent/3324884.3416609.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8079
record_format dspace
spelling sg-smu-ink.sis_research-80792022-04-07T08:08:43Z Cats are not fish: Deep learning testing calls for out-of-distribution awareness BEREND, David XIE, Xiaofei MA, Lei ZHOU, Lingjun LIU, Yang XU, Chi ZHAO, Jianjun As Deep Learning (DL) is continuously adopted in many industrial applications, its quality and reliability start to raise concerns. Similar to the traditional software development process, testing the DL software to uncover its defects at an early stage is an effective way to reduce risks after deployment. According to the fundamental assumption of deep learning, the DL software does not provide statistical guarantee and has limited capability in handling data that falls outside of its learned distribution, i.e., out-of-distribution (OOD) data. Although recent progress has been made in designing novel testing techniques for DL software, which can detect thousands of errors, the current state-of-the-art DL testing techniques usually do not take the distribution of generated test data into consideration. It is therefore hard to judge whether the "identified errors" are indeed meaningful errors to the DL application (i.e., due to quality issues of the model) or outliers that cannot be handled by the current model (i.e., due to the lack of training data). Tofill this gap, we take the first step and conduct a large scale empirical study, with a total of 451 experiment configurations, 42 deep neural networks (DNNs) and 1.2 million test data instances, to investigate and characterize the impact of OOD-awareness on DL testing. We further analyze the consequences when DL systems go into production by evaluating the effectiveness of adversarial retraining with distribution-aware errors. The results confirm that introducing data distribution awareness in both testing and enhancement phases outperforms distribution unaware retraining by up to 21.5%. 2020-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7076 info:doi/10.1145/3324884.3416609 https://ink.library.smu.edu.sg/context/sis_research/article/8079/viewcontent/3324884.3416609.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Deep learning testing quality assurance out of distribution OS and Networks Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Deep learning testing
quality assurance
out of distribution
OS and Networks
Software Engineering
spellingShingle Deep learning testing
quality assurance
out of distribution
OS and Networks
Software Engineering
BEREND, David
XIE, Xiaofei
MA, Lei
ZHOU, Lingjun
LIU, Yang
XU, Chi
ZHAO, Jianjun
Cats are not fish: Deep learning testing calls for out-of-distribution awareness
description As Deep Learning (DL) is continuously adopted in many industrial applications, its quality and reliability start to raise concerns. Similar to the traditional software development process, testing the DL software to uncover its defects at an early stage is an effective way to reduce risks after deployment. According to the fundamental assumption of deep learning, the DL software does not provide statistical guarantee and has limited capability in handling data that falls outside of its learned distribution, i.e., out-of-distribution (OOD) data. Although recent progress has been made in designing novel testing techniques for DL software, which can detect thousands of errors, the current state-of-the-art DL testing techniques usually do not take the distribution of generated test data into consideration. It is therefore hard to judge whether the "identified errors" are indeed meaningful errors to the DL application (i.e., due to quality issues of the model) or outliers that cannot be handled by the current model (i.e., due to the lack of training data). Tofill this gap, we take the first step and conduct a large scale empirical study, with a total of 451 experiment configurations, 42 deep neural networks (DNNs) and 1.2 million test data instances, to investigate and characterize the impact of OOD-awareness on DL testing. We further analyze the consequences when DL systems go into production by evaluating the effectiveness of adversarial retraining with distribution-aware errors. The results confirm that introducing data distribution awareness in both testing and enhancement phases outperforms distribution unaware retraining by up to 21.5%.
format text
author BEREND, David
XIE, Xiaofei
MA, Lei
ZHOU, Lingjun
LIU, Yang
XU, Chi
ZHAO, Jianjun
author_facet BEREND, David
XIE, Xiaofei
MA, Lei
ZHOU, Lingjun
LIU, Yang
XU, Chi
ZHAO, Jianjun
author_sort BEREND, David
title Cats are not fish: Deep learning testing calls for out-of-distribution awareness
title_short Cats are not fish: Deep learning testing calls for out-of-distribution awareness
title_full Cats are not fish: Deep learning testing calls for out-of-distribution awareness
title_fullStr Cats are not fish: Deep learning testing calls for out-of-distribution awareness
title_full_unstemmed Cats are not fish: Deep learning testing calls for out-of-distribution awareness
title_sort cats are not fish: deep learning testing calls for out-of-distribution awareness
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/7076
https://ink.library.smu.edu.sg/context/sis_research/article/8079/viewcontent/3324884.3416609.pdf
_version_ 1770576207306293248