Learning to double-check model prediction from a causal perspective
The present machine learning schema typically uses a one-pass model inference (e.g., forward propagation) to make predictions in the testing phase. It is inherently different from human students who double-check the answer during examinations especially when the confidence is low. To bridge this gap...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/170568 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-170568 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1705682023-09-19T06:45:03Z Learning to double-check model prediction from a causal perspective Deng, Xun Feng, Fuli Wang, Xiang He, Xiangnan Zhang, Hanwang Chua, Tat-Seng School of Computer Science and Engineering Engineering::Computer science and engineering Causality Classification The present machine learning schema typically uses a one-pass model inference (e.g., forward propagation) to make predictions in the testing phase. It is inherently different from human students who double-check the answer during examinations especially when the confidence is low. To bridge this gap, we propose a learning to double-check (L2D) framework, which formulates double check as a learnable procedure with two core operations: recognizing unreliable predictions and revising predictions. To judge the correctness of a prediction, we resort to counterfactual faithfulness in causal theory and design a contrastive faithfulness measure. In particular, L2D generates counterfactual features by imagining: "what would the sample features be if its label was the predicted class" and judges the prediction by the faithfulness of the counterfactual features. Furthermore, we design a simple and effective revision module to revise the original model prediction according to the faithfulness. We apply the L2D framework to three classification models and conduct experiments on two public datasets for image classification, validating the effectiveness of L2D in prediction correctness judgment and revision. This work was supported in part by the National Key Research and Development Program of China under Grant 2020YFB1406703, in part by the National Natural Science Foundation of China under Grant 62272437 and Grant U21B2026, and in part by the CCCD Key Laboratory of the Ministry of Culture and Tourism. 2023-09-19T06:45:03Z 2023-09-19T06:45:03Z 2023 Journal Article Deng, X., Feng, F., Wang, X., He, X., Zhang, H. & Chua, T. (2023). Learning to double-check model prediction from a causal perspective. IEEE Transactions On Neural Networks and Learning Systems. https://dx.doi.org/10.1109/TNNLS.2023.3264712 2162-237X https://hdl.handle.net/10356/170568 10.1109/TNNLS.2023.3264712 37053061 2-s2.0-85153492532 en IEEE Transactions on Neural Networks and Learning Systems © 2023 IEEE. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Causality Classification |
spellingShingle |
Engineering::Computer science and engineering Causality Classification Deng, Xun Feng, Fuli Wang, Xiang He, Xiangnan Zhang, Hanwang Chua, Tat-Seng Learning to double-check model prediction from a causal perspective |
description |
The present machine learning schema typically uses a one-pass model inference (e.g., forward propagation) to make predictions in the testing phase. It is inherently different from human students who double-check the answer during examinations especially when the confidence is low. To bridge this gap, we propose a learning to double-check (L2D) framework, which formulates double check as a learnable procedure with two core operations: recognizing unreliable predictions and revising predictions. To judge the correctness of a prediction, we resort to counterfactual faithfulness in causal theory and design a contrastive faithfulness measure. In particular, L2D generates counterfactual features by imagining: "what would the sample features be if its label was the predicted class" and judges the prediction by the faithfulness of the counterfactual features. Furthermore, we design a simple and effective revision module to revise the original model prediction according to the faithfulness. We apply the L2D framework to three classification models and conduct experiments on two public datasets for image classification, validating the effectiveness of L2D in prediction correctness judgment and revision. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Deng, Xun Feng, Fuli Wang, Xiang He, Xiangnan Zhang, Hanwang Chua, Tat-Seng |
format |
Article |
author |
Deng, Xun Feng, Fuli Wang, Xiang He, Xiangnan Zhang, Hanwang Chua, Tat-Seng |
author_sort |
Deng, Xun |
title |
Learning to double-check model prediction from a causal perspective |
title_short |
Learning to double-check model prediction from a causal perspective |
title_full |
Learning to double-check model prediction from a causal perspective |
title_fullStr |
Learning to double-check model prediction from a causal perspective |
title_full_unstemmed |
Learning to double-check model prediction from a causal perspective |
title_sort |
learning to double-check model prediction from a causal perspective |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/170568 |
_version_ |
1779156735376752640 |