Automatic fairness testing of neural classifiers through adversarial sampling

Although deep learning has demonstrated astonishing performance in many applications, there are still concerns about its dependability. One desirable property of deep learning applications with societal impact is fairness (i.e., non-discrimination). Unfortunately, discrimination might be intrinsical...

Full description

Saved in:
Bibliographic Details
Main Authors: ZHANG, Peixin, WANG, Jingyi, SUN, Jun, WANG, Xinyu, DONG, Guoliang, WANG, Xinggen, DAI, Ting, DONG, Jinsong
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2021
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/6224
https://ink.library.smu.edu.sg/context/sis_research/article/7227/viewcontent/09506918.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-7227
record_format dspace
spelling sg-smu-ink.sis_research-72272021-10-22T06:06:04Z Automatic fairness testing of neural classifiers through adversarial sampling ZHANG, Peixin WANG, Jingyi SUN, Jun WANG, Xinyu DONG, Guoliang WANG, Xinggen DAI, Ting DONG, Jinsong Although deep learning has demonstrated astonishing performance in many applications, there are still concerns about its dependability. One desirable property of deep learning applications with societal impact is fairness (i.e., non-discrimination). Unfortunately, discrimination might be intrinsically embedded into the models due to the discrimination in the training data. As a countermeasure, fairness testing systemically identifies discriminatory samples, which can be used to retrain the model and improve the model’s fairness. Existing fairness testing approaches however have two major limitations. Firstly, they only work well on traditional machine learning models and have poor performance (e.g., effectiveness and efficiency) on deep learning models. Secondly, they only work on simple structured (e.g., tabular) data and are not applicable for domains such as text. In this work, we bridge the gap by proposing a scalable and effective approach for systematically searching for discriminatory samples while extending existing fairness testing approaches to address a more challenging domain, i.e., text classification. Compared with state-of-the-art methods, our approach only employs lightweight procedures like gradient computation and clustering, which is significantly more scalable and effective. Experimental results show that on average, our approach explores the search space much more effectively (9.62 and 2.38 times more than the state-of-the-art methods respectively on tabular and text datasets) and generates much more discriminatory samples (24.95 and 2.68 times) within a same reasonable time. Moreover, the retrained models reduce discrimination by 57.2% and 60.2% respectively on average. 2021-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6224 info:doi/10.1109/TSE.2021.3101478 https://ink.library.smu.edu.sg/context/sis_research/article/7227/viewcontent/09506918.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Deep learning fairness testing individual discrimination gradient Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Deep learning
fairness testing
individual discrimination
gradient
Software Engineering
spellingShingle Deep learning
fairness testing
individual discrimination
gradient
Software Engineering
ZHANG, Peixin
WANG, Jingyi
SUN, Jun
WANG, Xinyu
DONG, Guoliang
WANG, Xinggen
DAI, Ting
DONG, Jinsong
Automatic fairness testing of neural classifiers through adversarial sampling
description Although deep learning has demonstrated astonishing performance in many applications, there are still concerns about its dependability. One desirable property of deep learning applications with societal impact is fairness (i.e., non-discrimination). Unfortunately, discrimination might be intrinsically embedded into the models due to the discrimination in the training data. As a countermeasure, fairness testing systemically identifies discriminatory samples, which can be used to retrain the model and improve the model’s fairness. Existing fairness testing approaches however have two major limitations. Firstly, they only work well on traditional machine learning models and have poor performance (e.g., effectiveness and efficiency) on deep learning models. Secondly, they only work on simple structured (e.g., tabular) data and are not applicable for domains such as text. In this work, we bridge the gap by proposing a scalable and effective approach for systematically searching for discriminatory samples while extending existing fairness testing approaches to address a more challenging domain, i.e., text classification. Compared with state-of-the-art methods, our approach only employs lightweight procedures like gradient computation and clustering, which is significantly more scalable and effective. Experimental results show that on average, our approach explores the search space much more effectively (9.62 and 2.38 times more than the state-of-the-art methods respectively on tabular and text datasets) and generates much more discriminatory samples (24.95 and 2.68 times) within a same reasonable time. Moreover, the retrained models reduce discrimination by 57.2% and 60.2% respectively on average.
format text
author ZHANG, Peixin
WANG, Jingyi
SUN, Jun
WANG, Xinyu
DONG, Guoliang
WANG, Xinggen
DAI, Ting
DONG, Jinsong
author_facet ZHANG, Peixin
WANG, Jingyi
SUN, Jun
WANG, Xinyu
DONG, Guoliang
WANG, Xinggen
DAI, Ting
DONG, Jinsong
author_sort ZHANG, Peixin
title Automatic fairness testing of neural classifiers through adversarial sampling
title_short Automatic fairness testing of neural classifiers through adversarial sampling
title_full Automatic fairness testing of neural classifiers through adversarial sampling
title_fullStr Automatic fairness testing of neural classifiers through adversarial sampling
title_full_unstemmed Automatic fairness testing of neural classifiers through adversarial sampling
title_sort automatic fairness testing of neural classifiers through adversarial sampling
publisher Institutional Knowledge at Singapore Management University
publishDate 2021
url https://ink.library.smu.edu.sg/sis_research/6224
https://ink.library.smu.edu.sg/context/sis_research/article/7227/viewcontent/09506918.pdf
_version_ 1770575894530752512