Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms
Understanding idioms is important in NLP. In this paper, we study to what extent pre-trained BERT model can encode the meaning of a potentially idiomatic expression (PIE) in a certain context. We make use of a few existing datasets and perform two probing tasks: PIE usage classification and idiom pa...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/6722 https://ink.library.smu.edu.sg/context/sis_research/article/7725/viewcontent/2021.ranlp_1.156.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-7725 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-77252022-01-27T11:13:06Z Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms TAN, Minghuan JIANG, Jing Understanding idioms is important in NLP. In this paper, we study to what extent pre-trained BERT model can encode the meaning of a potentially idiomatic expression (PIE) in a certain context. We make use of a few existing datasets and perform two probing tasks: PIE usage classification and idiom paraphrase identification. Our experiment results suggest that BERT indeed can separate the literal and idiomatic usages of a PIE with high accuracy. It is also able to encode the idiomatic meaning of a PIE to some extent. 2021-09-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/6722 info:doi/10.26615/978-954-452-072-4_156 https://ink.library.smu.edu.sg/context/sis_research/article/7725/viewcontent/2021.ranlp_1.156.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems Programming Languages and Compilers |
spellingShingle |
Databases and Information Systems Programming Languages and Compilers TAN, Minghuan JIANG, Jing Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
description |
Understanding idioms is important in NLP. In this paper, we study to what extent pre-trained BERT model can encode the meaning of a potentially idiomatic expression (PIE) in a certain context. We make use of a few existing datasets and perform two probing tasks: PIE usage classification and idiom paraphrase identification. Our experiment results suggest that BERT indeed can separate the literal and idiomatic usages of a PIE with high accuracy. It is also able to encode the idiomatic meaning of a PIE to some extent. |
format |
text |
author |
TAN, Minghuan JIANG, Jing |
author_facet |
TAN, Minghuan JIANG, Jing |
author_sort |
TAN, Minghuan |
title |
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
title_short |
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
title_full |
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
title_fullStr |
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
title_full_unstemmed |
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms |
title_sort |
does bert understand idioms? a probing-based empirical study of bert encodings of idioms |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2021 |
url |
https://ink.library.smu.edu.sg/sis_research/6722 https://ink.library.smu.edu.sg/context/sis_research/article/7725/viewcontent/2021.ranlp_1.156.pdf |
_version_ |
1770576054087319552 |