Cross-thought for sentence encoder pre-training
In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based seque...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/5602 https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-6605 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-66052021-01-07T13:54:35Z Cross-thought for sentence encoder pre-training WANG, Shuohang FANG, Yuwei SUN, Siqi GAN, Zhe CHENG, Yu LIU, Jingjing JIANG, Jing In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based sequence encoder over a large set of short sequences, which allows the model to automatically select the most useful information for predicting masked words. Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders trained with continuous sentence signals as well as traditional masked language modeling baselines. Our proposed approach also achieves new state of the art on HotpotQA (full-wiki setting) by improving intermediate information retrieval performance. 2020-11-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5602 info:doi/10.18653/v1/2020.emnlp-main.30 https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems Programming Languages and Compilers |
spellingShingle |
Databases and Information Systems Programming Languages and Compilers WANG, Shuohang FANG, Yuwei SUN, Siqi GAN, Zhe CHENG, Yu LIU, Jingjing JIANG, Jing Cross-thought for sentence encoder pre-training |
description |
In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based sequence encoder over a large set of short sequences, which allows the model to automatically select the most useful information for predicting masked words. Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders trained with continuous sentence signals as well as traditional masked language modeling baselines. Our proposed approach also achieves new state of the art on HotpotQA (full-wiki setting) by improving intermediate information retrieval performance. |
format |
text |
author |
WANG, Shuohang FANG, Yuwei SUN, Siqi GAN, Zhe CHENG, Yu LIU, Jingjing JIANG, Jing |
author_facet |
WANG, Shuohang FANG, Yuwei SUN, Siqi GAN, Zhe CHENG, Yu LIU, Jingjing JIANG, Jing |
author_sort |
WANG, Shuohang |
title |
Cross-thought for sentence encoder pre-training |
title_short |
Cross-thought for sentence encoder pre-training |
title_full |
Cross-thought for sentence encoder pre-training |
title_fullStr |
Cross-thought for sentence encoder pre-training |
title_full_unstemmed |
Cross-thought for sentence encoder pre-training |
title_sort |
cross-thought for sentence encoder pre-training |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2020 |
url |
https://ink.library.smu.edu.sg/sis_research/5602 https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf |
_version_ |
1770575526390398976 |