On the probability of necessity and sufficiency of explaining Graph Neural Networks: A lower bound optimization approach

The explainability of Graph Neural Networks (GNNs) is critical to various GNN applications, yet it remains a significant challenge. A convincing explanation should be both necessary and sufficient simultaneously. However, existing GNN explaining approaches focus on only one of the two aspects, neces...

全面介紹

Saved in:
書目詳細資料
Main Authors: CAI, Ruichu, ZHU, Yuxuan, CHEN, Xuexin, FANG, Yuan, WU, Min, QIAO, Jie, HAO, Zhifeng
格式: text
語言:English
出版: Institutional Knowledge at Singapore Management University 2025
主題:
在線閱讀:https://ink.library.smu.edu.sg/sis_research/10111
https://ink.library.smu.edu.sg/context/sis_research/article/11111/viewcontent/Prob_Necessity_GNN_sv.pdf
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
實物特徵
總結:The explainability of Graph Neural Networks (GNNs) is critical to various GNN applications, yet it remains a significant challenge. A convincing explanation should be both necessary and sufficient simultaneously. However, existing GNN explaining approaches focus on only one of the two aspects, necessity or sufficiency, or a heuristic trade-off between the two. Theoretically, the Probability of Necessity and Sufficiency (PNS) holds the potential to identify the most necessary and sufficient explanation since it can mathematically quantify the necessity and sufficiency of an explanation. Nevertheless, the difficulty of obtaining PNS due to non-monotonicity and the challenge of counterfactual estimation limit its wide use. To address the non-identifiability of PNS, we resort to a lower bound of PNS that can be optimized via counterfactual estimation, and propose a framework of Necessary and Sufficient Explanation for GNN (NSEG) via optimizing that lower bound. Specifically, we depict the GNN as a structural causal model (SCM), and estimate the probability of counterfactual via the intervention under the SCM. Additionally, we leverage continuous masks with a sampling strategy to optimize the lower bound to enhance the scalability. Empirical results demonstrate that NSEG outperforms state-of-the-art methods, consistently generating the most necessary and sufficient explanations. The implementation of our NSEG is available at https://github.com/EthanChu7/NSEG.