Neural logic vision language explainer

If we compare how humans reason and how deep models reason, humans reason in a symbolic manner with a formal language called logic, while most deep models reason in black-box. A natural question to ask is “Do the trained deep models reason similar as humans?” or “Can we explain the reasoning of...

Full description

Saved in:
Bibliographic Details
Main Authors: Yang, Xiaofeng, Liu, Fayao, Lin, Guosheng
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/172228
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-172228
record_format dspace
spelling sg-ntu-dr.10356-1722282023-12-08T15:35:51Z Neural logic vision language explainer Yang, Xiaofeng Liu, Fayao Lin, Guosheng School of Computer Science and Engineering Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Cognition Deep Learning If we compare how humans reason and how deep models reason, humans reason in a symbolic manner with a formal language called logic, while most deep models reason in black-box. A natural question to ask is “Do the trained deep models reason similar as humans?” or “Can we explain the reasoning of deep models in the language of logic?” . In this work, we present NeurLogX to explain the reasoning process of deep vision language models in the language of logic. Given a trained vision language model, our method starts by generating reasoning facts through augmenting the input data. We then develop a differentiable inductive logic programming framework to learn interpretable logic rules from the facts. We show our results on various popular vision language models. Interestingly, we observe that almost all of the tested models can reason logically. Ministry of Education (MOE) National Research Foundation (NRF) Submitted/Accepted version This research is supported by the National Research Foundation, Singapore under its AI Singapore Programme (AISG Award No: AISG-RP-2018-003), the MOE AcRF Tier-1 research grants: RG95/20, and the OPPO research grant. 2023-12-05T01:56:26Z 2023-12-05T01:56:26Z 2023 Journal Article Yang, X., Liu, F. & Lin, G. (2023). Neural logic vision language explainer. IEEE Transactions On Multimedia. https://dx.doi.org/10.1109/TMM.2023.3310277 1520-9210 https://hdl.handle.net/10356/172228 10.1109/TMM.2023.3310277 en AISG-RP-2018-003 RG95/20 IEEE Transactions on Multimedia © 2023 IEEE. All rights reserved. This article may be downloaded for personal use only. Any other use requires prior permission of the copyright holder. The Version of Record is available online at http://doi.org/10.1109/TMM.2023.3310277. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Cognition
Deep Learning
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Cognition
Deep Learning
Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
Neural logic vision language explainer
description If we compare how humans reason and how deep models reason, humans reason in a symbolic manner with a formal language called logic, while most deep models reason in black-box. A natural question to ask is “Do the trained deep models reason similar as humans?” or “Can we explain the reasoning of deep models in the language of logic?” . In this work, we present NeurLogX to explain the reasoning process of deep vision language models in the language of logic. Given a trained vision language model, our method starts by generating reasoning facts through augmenting the input data. We then develop a differentiable inductive logic programming framework to learn interpretable logic rules from the facts. We show our results on various popular vision language models. Interestingly, we observe that almost all of the tested models can reason logically.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
format Article
author Yang, Xiaofeng
Liu, Fayao
Lin, Guosheng
author_sort Yang, Xiaofeng
title Neural logic vision language explainer
title_short Neural logic vision language explainer
title_full Neural logic vision language explainer
title_fullStr Neural logic vision language explainer
title_full_unstemmed Neural logic vision language explainer
title_sort neural logic vision language explainer
publishDate 2023
url https://hdl.handle.net/10356/172228
_version_ 1784855584018792448