Debiasing NLU models via causal intervention and counterfactual reasoning

Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relying on annotation biases of the datasets as a shortcut, which goes against the underlying mechanisms of the task of interest. To reduce such biases, several recent works introduce debiasing methods to...

Full description

Saved in:
Bibliographic Details
Main Authors: TIAN, Bing, CAO, Yixin, ZHANG, Yong, XING, Chunxiao
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7454
https://ink.library.smu.edu.sg/context/sis_research/article/8457/viewcontent/21389_Article_Text_25402_1_2_20220628.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8457
record_format dspace
spelling sg-smu-ink.sis_research-84572022-10-20T07:22:55Z Debiasing NLU models via causal intervention and counterfactual reasoning TIAN, Bing CAO, Yixin ZHANG, Yong XING, Chunxiao Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relying on annotation biases of the datasets as a shortcut, which goes against the underlying mechanisms of the task of interest. To reduce such biases, several recent works introduce debiasing methods to regularize the training process of targeted NLU models. In this paper, we provide a new perspective with causal inference to fnd out the bias. On the one hand, we show that there is an unobserved confounder for the natural language utterances and their respective classes, leading to spurious correlations from training data. To remove such confounder, the backdoor adjustment with causal intervention is utilized to fnd the true causal effect, which makes the training process fundamentally different from the traditional likelihood estimation. On the other hand, in inference process, we formulate the bias as the direct causal effect and remove it by pursuing the indirect causal effect with counterfactual reasoning. We conduct experiments on large-scale natural language inference and fact verifcation benchmarks, evaluating on bias sensitive datasets that are specifcally designed to assess the robustness of models against known biases in the training data. Experimental results show that our proposed debiasing framework outperforms previous stateof-the-art debiasing methods while maintaining the original in-distribution performance. 2022-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7454 info:doi/10.1609/aaai.v36i10.21389 https://ink.library.smu.edu.sg/context/sis_research/article/8457/viewcontent/21389_Article_Text_25402_1_2_20220628.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Artificial Intelligence and Robotics Graphics and Human Computer Interfaces
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
spellingShingle Artificial Intelligence and Robotics
Graphics and Human Computer Interfaces
TIAN, Bing
CAO, Yixin
ZHANG, Yong
XING, Chunxiao
Debiasing NLU models via causal intervention and counterfactual reasoning
description Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relying on annotation biases of the datasets as a shortcut, which goes against the underlying mechanisms of the task of interest. To reduce such biases, several recent works introduce debiasing methods to regularize the training process of targeted NLU models. In this paper, we provide a new perspective with causal inference to fnd out the bias. On the one hand, we show that there is an unobserved confounder for the natural language utterances and their respective classes, leading to spurious correlations from training data. To remove such confounder, the backdoor adjustment with causal intervention is utilized to fnd the true causal effect, which makes the training process fundamentally different from the traditional likelihood estimation. On the other hand, in inference process, we formulate the bias as the direct causal effect and remove it by pursuing the indirect causal effect with counterfactual reasoning. We conduct experiments on large-scale natural language inference and fact verifcation benchmarks, evaluating on bias sensitive datasets that are specifcally designed to assess the robustness of models against known biases in the training data. Experimental results show that our proposed debiasing framework outperforms previous stateof-the-art debiasing methods while maintaining the original in-distribution performance.
format text
author TIAN, Bing
CAO, Yixin
ZHANG, Yong
XING, Chunxiao
author_facet TIAN, Bing
CAO, Yixin
ZHANG, Yong
XING, Chunxiao
author_sort TIAN, Bing
title Debiasing NLU models via causal intervention and counterfactual reasoning
title_short Debiasing NLU models via causal intervention and counterfactual reasoning
title_full Debiasing NLU models via causal intervention and counterfactual reasoning
title_fullStr Debiasing NLU models via causal intervention and counterfactual reasoning
title_full_unstemmed Debiasing NLU models via causal intervention and counterfactual reasoning
title_sort debiasing nlu models via causal intervention and counterfactual reasoning
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7454
https://ink.library.smu.edu.sg/context/sis_research/article/8457/viewcontent/21389_Article_Text_25402_1_2_20220628.pdf
_version_ 1770576351092277248