Cross-thought for sentence encoder pre-training

In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based seque...

Full description

Saved in:
Bibliographic Details
Main Authors: WANG, Shuohang, FANG, Yuwei, SUN, Siqi, GAN, Zhe, CHENG, Yu, LIU, Jingjing, JIANG, Jing
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/5602
https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-6605
record_format dspace
spelling sg-smu-ink.sis_research-66052021-01-07T13:54:35Z Cross-thought for sentence encoder pre-training WANG, Shuohang FANG, Yuwei SUN, Siqi GAN, Zhe CHENG, Yu LIU, Jingjing JIANG, Jing In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based sequence encoder over a large set of short sequences, which allows the model to automatically select the most useful information for predicting masked words. Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders trained with continuous sentence signals as well as traditional masked language modeling baselines. Our proposed approach also achieves new state of the art on HotpotQA (full-wiki setting) by improving intermediate information retrieval performance. 2020-11-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5602 info:doi/10.18653/v1/2020.emnlp-main.30 https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Databases and Information Systems
Programming Languages and Compilers
spellingShingle Databases and Information Systems
Programming Languages and Compilers
WANG, Shuohang
FANG, Yuwei
SUN, Siqi
GAN, Zhe
CHENG, Yu
LIU, Jingjing
JIANG, Jing
Cross-thought for sentence encoder pre-training
description In this paper, we propose Cross-Thought, a novel approach to pre-training sequence encoder, which is instrumental in building reusable sequence embeddings for large-scale NLP tasks such as question answering. Instead of using the original signals of full sentences, we train a Transformer-based sequence encoder over a large set of short sequences, which allows the model to automatically select the most useful information for predicting masked words. Experiments on question answering and textual entailment tasks demonstrate that our pre-trained encoder can outperform state-of-the-art encoders trained with continuous sentence signals as well as traditional masked language modeling baselines. Our proposed approach also achieves new state of the art on HotpotQA (full-wiki setting) by improving intermediate information retrieval performance.
format text
author WANG, Shuohang
FANG, Yuwei
SUN, Siqi
GAN, Zhe
CHENG, Yu
LIU, Jingjing
JIANG, Jing
author_facet WANG, Shuohang
FANG, Yuwei
SUN, Siqi
GAN, Zhe
CHENG, Yu
LIU, Jingjing
JIANG, Jing
author_sort WANG, Shuohang
title Cross-thought for sentence encoder pre-training
title_short Cross-thought for sentence encoder pre-training
title_full Cross-thought for sentence encoder pre-training
title_fullStr Cross-thought for sentence encoder pre-training
title_full_unstemmed Cross-thought for sentence encoder pre-training
title_sort cross-thought for sentence encoder pre-training
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/5602
https://ink.library.smu.edu.sg/context/sis_research/article/6605/viewcontent/EMNLP_2020a.pdf
_version_ 1770575526390398976