Embedding WordNet knowledge for textual entailment
In this paper, we study how we can improve a deep learning approach to textual entailment by incorporating lexical entailment relations from WordNet. Our idea is to embed the lexical entailment knowledge contained in WordNet in specially-learned word vectors, which we call “entailment vectors.” We p...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2018
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/4280 https://ink.library.smu.edu.sg/context/sis_research/article/5283/viewcontent/Embedding_WordNet_2018_pv.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-5283 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-52832019-02-21T08:25:56Z Embedding WordNet knowledge for textual entailment LAN, Yunshi JIANG, Jing In this paper, we study how we can improve a deep learning approach to textual entailment by incorporating lexical entailment relations from WordNet. Our idea is to embed the lexical entailment knowledge contained in WordNet in specially-learned word vectors, which we call “entailment vectors.” We present a standard neural network model and a novel set-theoretic model to learn these entailment vectors from word pairs with known lexical entailment relations derived from WordNet. We further incorporate these entailment vectors into a decomposable attention model for textual entailment and evaluate the model on the SICK and the SNLI dataset. We find that using these special entailment word vectors, we can significantly improve the performance of textual entailment compared with a baseline that uses only standard word2vec vectors. The final performance of our model is close to or above the state of the art, but our method does not rely on any manually-crafted rules or extensive syntactic features. 2018-08-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4280 https://ink.library.smu.edu.sg/context/sis_research/article/5283/viewcontent/Embedding_WordNet_2018_pv.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Theory and Algorithms |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems Theory and Algorithms |
spellingShingle |
Databases and Information Systems Theory and Algorithms LAN, Yunshi JIANG, Jing Embedding WordNet knowledge for textual entailment |
description |
In this paper, we study how we can improve a deep learning approach to textual entailment by incorporating lexical entailment relations from WordNet. Our idea is to embed the lexical entailment knowledge contained in WordNet in specially-learned word vectors, which we call “entailment vectors.” We present a standard neural network model and a novel set-theoretic model to learn these entailment vectors from word pairs with known lexical entailment relations derived from WordNet. We further incorporate these entailment vectors into a decomposable attention model for textual entailment and evaluate the model on the SICK and the SNLI dataset. We find that using these special entailment word vectors, we can significantly improve the performance of textual entailment compared with a baseline that uses only standard word2vec vectors. The final performance of our model is close to or above the state of the art, but our method does not rely on any manually-crafted rules or extensive syntactic features. |
format |
text |
author |
LAN, Yunshi JIANG, Jing |
author_facet |
LAN, Yunshi JIANG, Jing |
author_sort |
LAN, Yunshi |
title |
Embedding WordNet knowledge for textual entailment |
title_short |
Embedding WordNet knowledge for textual entailment |
title_full |
Embedding WordNet knowledge for textual entailment |
title_fullStr |
Embedding WordNet knowledge for textual entailment |
title_full_unstemmed |
Embedding WordNet knowledge for textual entailment |
title_sort |
embedding wordnet knowledge for textual entailment |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2018 |
url |
https://ink.library.smu.edu.sg/sis_research/4280 https://ink.library.smu.edu.sg/context/sis_research/article/5283/viewcontent/Embedding_WordNet_2018_pv.pdf |
_version_ |
1770574598237061120 |