Towards interpreting recurrent neural networks through probabilistic abstraction

Neural networks are becoming a popular tool for solving many realworld problems such as object recognition and machine translation, thanks to its exceptional performance as an end-to-end solution. However, neural networks are complex black-box models, which hinders humans from interpreting and conse...

Full description

Saved in:
Bibliographic Details
Main Authors: DONG, Guoliang, WANG, Jingyi, SUN, Jun, ZHANG, Yang, WANG, Xinyu, DAI, Ting, DONG, Jin Song, WANG, Xingen
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/5947
https://ink.library.smu.edu.sg/context/sis_research/article/6950/viewcontent/1909.10023.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-6950
record_format dspace
spelling sg-smu-ink.sis_research-69502021-05-18T05:12:48Z Towards interpreting recurrent neural networks through probabilistic abstraction DONG, Guoliang WANG, Jingyi SUN, Jun ZHANG, Yang WANG, Xinyu DAI, Ting DONG, Jin Song WANG, Xingen Neural networks are becoming a popular tool for solving many realworld problems such as object recognition and machine translation, thanks to its exceptional performance as an end-to-end solution. However, neural networks are complex black-box models, which hinders humans from interpreting and consequently trusting them in making critical decisions. Towards interpreting neural networks, several approaches have been proposed to extract simple deterministic models from neural networks. The results are not encouraging (e.g., low accuracy and limited scalability), fundamentally due to the limited expressiveness of such simple models.In this work, we propose an approach to extract probabilistic automata for interpreting an important class of neural networks, i.e., recurrent neural networks. Our work distinguishes itself from existing approaches in two important ways. One is that probability is used to compensate for the loss of expressiveness. This is inspired by the observation that human reasoning is often 'probabilistic'. The other is that we adaptively identify the right level of abstraction so that a simple model is extracted in a request-specific way. We conduct experiments on several real-world datasets using state-of-the-art architectures including GRU and LSTM. The result shows that our approach significantly improves existing approaches in terms of accuracy or scalability. Lastly, we demonstrate the usefulness of the extracted models through detecting adversarial texts. 2020-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5947 info:doi/10.1145/3324884.3416592 https://ink.library.smu.edu.sg/context/sis_research/article/6950/viewcontent/1909.10023.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Abstraction Interpretation Probabilistic automata Recurrent neural networks Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Abstraction
Interpretation
Probabilistic automata
Recurrent neural networks
Software Engineering
spellingShingle Abstraction
Interpretation
Probabilistic automata
Recurrent neural networks
Software Engineering
DONG, Guoliang
WANG, Jingyi
SUN, Jun
ZHANG, Yang
WANG, Xinyu
DAI, Ting
DONG, Jin Song
WANG, Xingen
Towards interpreting recurrent neural networks through probabilistic abstraction
description Neural networks are becoming a popular tool for solving many realworld problems such as object recognition and machine translation, thanks to its exceptional performance as an end-to-end solution. However, neural networks are complex black-box models, which hinders humans from interpreting and consequently trusting them in making critical decisions. Towards interpreting neural networks, several approaches have been proposed to extract simple deterministic models from neural networks. The results are not encouraging (e.g., low accuracy and limited scalability), fundamentally due to the limited expressiveness of such simple models.In this work, we propose an approach to extract probabilistic automata for interpreting an important class of neural networks, i.e., recurrent neural networks. Our work distinguishes itself from existing approaches in two important ways. One is that probability is used to compensate for the loss of expressiveness. This is inspired by the observation that human reasoning is often 'probabilistic'. The other is that we adaptively identify the right level of abstraction so that a simple model is extracted in a request-specific way. We conduct experiments on several real-world datasets using state-of-the-art architectures including GRU and LSTM. The result shows that our approach significantly improves existing approaches in terms of accuracy or scalability. Lastly, we demonstrate the usefulness of the extracted models through detecting adversarial texts.
format text
author DONG, Guoliang
WANG, Jingyi
SUN, Jun
ZHANG, Yang
WANG, Xinyu
DAI, Ting
DONG, Jin Song
WANG, Xingen
author_facet DONG, Guoliang
WANG, Jingyi
SUN, Jun
ZHANG, Yang
WANG, Xinyu
DAI, Ting
DONG, Jin Song
WANG, Xingen
author_sort DONG, Guoliang
title Towards interpreting recurrent neural networks through probabilistic abstraction
title_short Towards interpreting recurrent neural networks through probabilistic abstraction
title_full Towards interpreting recurrent neural networks through probabilistic abstraction
title_fullStr Towards interpreting recurrent neural networks through probabilistic abstraction
title_full_unstemmed Towards interpreting recurrent neural networks through probabilistic abstraction
title_sort towards interpreting recurrent neural networks through probabilistic abstraction
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/5947
https://ink.library.smu.edu.sg/context/sis_research/article/6950/viewcontent/1909.10023.pdf
_version_ 1770575701306507264