Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches
With the advancement of sentiment analysis (SA) models and their incorporation into our daily lives, fairness testing on these models is crucial, since unfair decisions can cause discrimination to a large population. Nevertheless, some challenges in fairness testing include the unknown oracle, the d...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2023
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/8514 https://ink.library.smu.edu.sg/context/sis_research/article/9517/viewcontent/Exploring_and_Repairing_Gender_Fairness_Violations_in_Word_Embedding_based_Sentiment_Analysis_Model_through_Adversarial_Patches.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-9517 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-95172024-01-22T15:09:18Z Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches KHOO, Lin Sze BAY, Jia Qi YAP, Ming Lee Kimberly LIM, Mei Kuan CHONG, Chun Yong YANG, Zhou LO, David With the advancement of sentiment analysis (SA) models and their incorporation into our daily lives, fairness testing on these models is crucial, since unfair decisions can cause discrimination to a large population. Nevertheless, some challenges in fairness testing include the unknown oracle, the difficulty in generating suitable test inputs, and the lack of a reliable way of fixing the issues. To fill in these gaps, BiasRV, a tool based on metamorphic testing (MT), was introduced and succeeded in uncovering fairness issues in a transformer-based model. However, the extent of unfairness in other SA models has not been thoroughly investigated. Our work conducts a more comprehensive empirical study to reveal the extent of fairness violations, specifically gender fairness, exhibited by other popular word embedding-based SA models. We define fairness violation as the behavior in which an SA model predicts variants created from a text, which merely differ in gender classes, to have different sentiments. Our inspection utilizing BiasRV uncovers at least 30 fairness violations (at BiasRV's default threshold) in all three SA models. Realizing the importance of addressing such significant violations, we introduce adversarial patches (AP) as a way of patch generation in an automated program repair (APR) system to fix them. We adopt adversarial fine-tuning in AP by retraining SA models using adversarial examples, which are bias-uncovering test cases dynamically generated by a tool named BiasFinder at runtime. Evaluation of the SA models shows that our proposed AP reduces fairness violations by at least 25%. 2023-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/8514 info:doi/10.1109/SANER56733.2023.00066 https://ink.library.smu.edu.sg/context/sis_research/article/9517/viewcontent/Exploring_and_Repairing_Gender_Fairness_Violations_in_Word_Embedding_based_Sentiment_Analysis_Model_through_Adversarial_Patches.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Fairness testing Automated program repair Sentiment analysis Software Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Fairness testing Automated program repair Sentiment analysis Software Engineering |
spellingShingle |
Fairness testing Automated program repair Sentiment analysis Software Engineering KHOO, Lin Sze BAY, Jia Qi YAP, Ming Lee Kimberly LIM, Mei Kuan CHONG, Chun Yong YANG, Zhou LO, David Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
description |
With the advancement of sentiment analysis (SA) models and their incorporation into our daily lives, fairness testing on these models is crucial, since unfair decisions can cause discrimination to a large population. Nevertheless, some challenges in fairness testing include the unknown oracle, the difficulty in generating suitable test inputs, and the lack of a reliable way of fixing the issues. To fill in these gaps, BiasRV, a tool based on metamorphic testing (MT), was introduced and succeeded in uncovering fairness issues in a transformer-based model. However, the extent of unfairness in other SA models has not been thoroughly investigated. Our work conducts a more comprehensive empirical study to reveal the extent of fairness violations, specifically gender fairness, exhibited by other popular word embedding-based SA models. We define fairness violation as the behavior in which an SA model predicts variants created from a text, which merely differ in gender classes, to have different sentiments. Our inspection utilizing BiasRV uncovers at least 30 fairness violations (at BiasRV's default threshold) in all three SA models. Realizing the importance of addressing such significant violations, we introduce adversarial patches (AP) as a way of patch generation in an automated program repair (APR) system to fix them. We adopt adversarial fine-tuning in AP by retraining SA models using adversarial examples, which are bias-uncovering test cases dynamically generated by a tool named BiasFinder at runtime. Evaluation of the SA models shows that our proposed AP reduces fairness violations by at least 25%. |
format |
text |
author |
KHOO, Lin Sze BAY, Jia Qi YAP, Ming Lee Kimberly LIM, Mei Kuan CHONG, Chun Yong YANG, Zhou LO, David |
author_facet |
KHOO, Lin Sze BAY, Jia Qi YAP, Ming Lee Kimberly LIM, Mei Kuan CHONG, Chun Yong YANG, Zhou LO, David |
author_sort |
KHOO, Lin Sze |
title |
Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
title_short |
Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
title_full |
Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
title_fullStr |
Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
title_full_unstemmed |
Exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
title_sort |
exploring and repairing gender fairness violations in word embedding-based sentiment analysis model through adversarial patches |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2023 |
url |
https://ink.library.smu.edu.sg/sis_research/8514 https://ink.library.smu.edu.sg/context/sis_research/article/9517/viewcontent/Exploring_and_Repairing_Gender_Fairness_Violations_in_Word_Embedding_based_Sentiment_Analysis_Model_through_Adversarial_Patches.pdf |
_version_ |
1789483256822890496 |