An empirical study on data distribution-aware test selection for deep learning enhancement

Similar to traditional software that is constantly under evolution, deep neural networks need to evolve upon the rapid growth of test data for continuous enhancement (e.g., adapting to distribution shift in a new environment for deployment). However, it is labor intensive to manually label all of th...

Full description

Saved in:
Bibliographic Details
Main Authors: HU, Qiang, GUO, Yuejun, CORDY, Maxime, XIE, Xiaofei, MA, Lei, PAPADAKIS, Mike, LE TRAON, Yves
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7195
https://ink.library.smu.edu.sg/context/sis_research/article/8198/viewcontent/3511598.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8198
record_format dspace
spelling sg-smu-ink.sis_research-81982022-08-04T08:55:19Z An empirical study on data distribution-aware test selection for deep learning enhancement HU, Qiang GUO, Yuejun CORDY, Maxime XIE, Xiaofei MA, Lei PAPADAKIS, Mike LE TRAON, Yves Similar to traditional software that is constantly under evolution, deep neural networks need to evolve upon the rapid growth of test data for continuous enhancement (e.g., adapting to distribution shift in a new environment for deployment). However, it is labor intensive to manually label all of the collected test data. Test selection solves this problem by strategically choosing a small set to label. Via retraining with the selected set, deep neural networks will achieve competitive accuracy. Unfortunately, existing selection metrics involve three main limitations: (1) using different retraining processes, (2) ignoring data distribution shifts, and (3) being insufficiently evaluated. To fill this gap, we first conduct a systemically empirical study to reveal the impact of the retraining process and data distribution on model enhancement. Then based on our findings, we propose DAT, a novel distribution-aware test selection metric. Experimental results reveal that retraining using both the training and selected data outperforms using only the selected data. None of the selection metrics perform the best under various data distributions. By contrast, DAT effectively alleviates the impact of distribution shifts and outperforms the compared metrics by up to five times and 30.09% accuracy improvement for model enhancement on simulated and in-the-wild distribution shift scenarios, respectively. 2022-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7195 info:doi/10.1145/3511598 https://ink.library.smu.edu.sg/context/sis_research/article/8198/viewcontent/3511598.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Deep learning testing test selection data distribution Artificial Intelligence and Robotics Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Deep learning testing
test selection
data distribution
Artificial Intelligence and Robotics
Software Engineering
spellingShingle Deep learning testing
test selection
data distribution
Artificial Intelligence and Robotics
Software Engineering
HU, Qiang
GUO, Yuejun
CORDY, Maxime
XIE, Xiaofei
MA, Lei
PAPADAKIS, Mike
LE TRAON, Yves
An empirical study on data distribution-aware test selection for deep learning enhancement
description Similar to traditional software that is constantly under evolution, deep neural networks need to evolve upon the rapid growth of test data for continuous enhancement (e.g., adapting to distribution shift in a new environment for deployment). However, it is labor intensive to manually label all of the collected test data. Test selection solves this problem by strategically choosing a small set to label. Via retraining with the selected set, deep neural networks will achieve competitive accuracy. Unfortunately, existing selection metrics involve three main limitations: (1) using different retraining processes, (2) ignoring data distribution shifts, and (3) being insufficiently evaluated. To fill this gap, we first conduct a systemically empirical study to reveal the impact of the retraining process and data distribution on model enhancement. Then based on our findings, we propose DAT, a novel distribution-aware test selection metric. Experimental results reveal that retraining using both the training and selected data outperforms using only the selected data. None of the selection metrics perform the best under various data distributions. By contrast, DAT effectively alleviates the impact of distribution shifts and outperforms the compared metrics by up to five times and 30.09% accuracy improvement for model enhancement on simulated and in-the-wild distribution shift scenarios, respectively.
format text
author HU, Qiang
GUO, Yuejun
CORDY, Maxime
XIE, Xiaofei
MA, Lei
PAPADAKIS, Mike
LE TRAON, Yves
author_facet HU, Qiang
GUO, Yuejun
CORDY, Maxime
XIE, Xiaofei
MA, Lei
PAPADAKIS, Mike
LE TRAON, Yves
author_sort HU, Qiang
title An empirical study on data distribution-aware test selection for deep learning enhancement
title_short An empirical study on data distribution-aware test selection for deep learning enhancement
title_full An empirical study on data distribution-aware test selection for deep learning enhancement
title_fullStr An empirical study on data distribution-aware test selection for deep learning enhancement
title_full_unstemmed An empirical study on data distribution-aware test selection for deep learning enhancement
title_sort empirical study on data distribution-aware test selection for deep learning enhancement
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7195
https://ink.library.smu.edu.sg/context/sis_research/article/8198/viewcontent/3511598.pdf
_version_ 1770576267757748224