Revisiting neuron coverage metrics and quality of deep neural networks

Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have prop...

Full description

Saved in:
Bibliographic Details
Main Authors: YANG, Zhou, SHI, Jieke, ASYROFI, Muhammad Hilmi, LO, David
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7625
https://ink.library.smu.edu.sg/context/sis_research/article/8628/viewcontent/378600a408.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8628
record_format dspace
spelling sg-smu-ink.sis_research-86282023-01-10T04:01:57Z Revisiting neuron coverage metrics and quality of deep neural networks YANG, Zhou SHI, Jieke ASYROFI, Muhammad Hilmi LO, David Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have proposed neuron coverage metrics and coverage-driven methods to generate DNN test cases. However, Yan et al. doubt the usefulness of existing coverage criteria in DNN testing. They show that a coverage-driven method is less effective than a gradient-based method in terms of both uncovering defects and improving model robustness. In this paper, we conduct a replication study of the work by Yan et al. and extend the experiments for deeper analysis. A larger model and a dataset of higher resolution images are included to examine the generalizability of the results. We also extend the experiments with more test case generation techniques and adjust the process of improving model robustness to be closer to the practical life cycle of DNN development. Our experiment results confirm the conclusion from Yan et al. that coverage-driven methods are less effective than gradient-based methods. Yan et al. find that using gradient-based methods to retrain cannot repair defects uncovered by coverage-driven methods. They attribute this to the fact that the two types of methods use different perturbation strategies: gradient-based methods perform differentiable transformations while coverage-driven methods can perform additional non-differentiable transformations. We test several hypotheses and further show that even coverage-driven methods are constrained only to perform differentiable transformations, the uncovered defects still cannot be repaired by adversarial training with gradient-based methods. Thus, defensive strategies for coverage-driven methods should be further studied. 2022-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7625 info:doi/10.1109/SANER53432.2022.00056 https://ink.library.smu.edu.sg/context/sis_research/article/8628/viewcontent/378600a408.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Deep learning testing Coverage-driven testing Software quality Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Deep learning testing
Coverage-driven testing
Software quality
Software Engineering
spellingShingle Deep learning testing
Coverage-driven testing
Software quality
Software Engineering
YANG, Zhou
SHI, Jieke
ASYROFI, Muhammad Hilmi
LO, David
Revisiting neuron coverage metrics and quality of deep neural networks
description Deep neural networks (DNN) have been widely applied in modern life, including critical domains like autonomous driving, making it essential to ensure the reliability and robustness of DNN-powered systems. As an analogy to code coverage metrics for testing conventional software, researchers have proposed neuron coverage metrics and coverage-driven methods to generate DNN test cases. However, Yan et al. doubt the usefulness of existing coverage criteria in DNN testing. They show that a coverage-driven method is less effective than a gradient-based method in terms of both uncovering defects and improving model robustness. In this paper, we conduct a replication study of the work by Yan et al. and extend the experiments for deeper analysis. A larger model and a dataset of higher resolution images are included to examine the generalizability of the results. We also extend the experiments with more test case generation techniques and adjust the process of improving model robustness to be closer to the practical life cycle of DNN development. Our experiment results confirm the conclusion from Yan et al. that coverage-driven methods are less effective than gradient-based methods. Yan et al. find that using gradient-based methods to retrain cannot repair defects uncovered by coverage-driven methods. They attribute this to the fact that the two types of methods use different perturbation strategies: gradient-based methods perform differentiable transformations while coverage-driven methods can perform additional non-differentiable transformations. We test several hypotheses and further show that even coverage-driven methods are constrained only to perform differentiable transformations, the uncovered defects still cannot be repaired by adversarial training with gradient-based methods. Thus, defensive strategies for coverage-driven methods should be further studied.
format text
author YANG, Zhou
SHI, Jieke
ASYROFI, Muhammad Hilmi
LO, David
author_facet YANG, Zhou
SHI, Jieke
ASYROFI, Muhammad Hilmi
LO, David
author_sort YANG, Zhou
title Revisiting neuron coverage metrics and quality of deep neural networks
title_short Revisiting neuron coverage metrics and quality of deep neural networks
title_full Revisiting neuron coverage metrics and quality of deep neural networks
title_fullStr Revisiting neuron coverage metrics and quality of deep neural networks
title_full_unstemmed Revisiting neuron coverage metrics and quality of deep neural networks
title_sort revisiting neuron coverage metrics and quality of deep neural networks
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7625
https://ink.library.smu.edu.sg/context/sis_research/article/8628/viewcontent/378600a408.pdf
_version_ 1770576396861571072