Zero-shot text classification via self-supervised tuning
Existing solutions to zero-shot text classification either conduct prompting with pre-trained language models, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised...
Saved in:
Main Authors: | , , , , , , |
---|---|
Other Authors: | |
Format: | Conference or Workshop Item |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/168505 https://2023.aclweb.org/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-168505 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1685052023-06-11T15:38:44Z Zero-shot text classification via self-supervised tuning Liu, Chaoqun Zhang, Wenxuan Chen, Guizhen Wu, Xiaobao Luu, Anh Tuan Chang, Chip Hong Bing, Lidong Interdisciplinary Graduate School (IGS) 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023) Alibaba Group Alibaba-NTU Singapore Joint Research Institute Computer Science - Computation and Language Computer Science - Artificial Intelligence Engineering::Computer science and engineering::Computing methodologies::Document and text processing Zero-Shot Text Classification Self-Supervised Tuning Existing solutions to zero-shot text classification either conduct prompting with pre-trained language models, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised learning to solve zero-shot text classification tasks by tuning the language models with unlabeled data, called self-supervised tuning. By exploring the inherent structure of free texts, we propose a new learning objective called first sentence prediction to bridge the gap between unlabeled data and text classification tasks. After tuning the model to learn to predict the first sentence in a paragraph based on the rest, the model is able to conduct zero-shot inference on unseen tasks such as topic classification and sentiment analysis. Experimental results show that our model outperforms the state-of-the-art baselines on 7 out of 10 tasks. Moreover, the analysis reveals that our model is less sensitive to the prompt design. Our code and pre-trained models are publicly available at https://github.com/DAMO-NLP-SG/SSTuning . Ministry of Education (MOE) Submitted/Accepted version This research is supported, in part, by Alibaba Group through Alibaba Innovative Research (AIR) Program and Alibaba-NTU Singapore Joint Research Institute (JRI), Nanyang Technological University, Singapore. Chaoqun Liu and Guizhen Chen extend their gratitude to Interdisciplinary Graduate Programme and School of Computer Science and Engineering, Nanyang Technological University, Singapore, for their support. This research is also supported by the Ministry of Education Tier 1 grant (MOE Tier 1 RS21/20). 2023-06-06T04:43:56Z 2023-06-06T04:43:56Z 2023 Conference Paper Liu, C., Zhang, W., Chen, G., Wu, X., Luu, A. T., Chang, C. H. & Bing, L. (2023). Zero-shot text classification via self-supervised tuning. 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023). https://hdl.handle.net/10356/168505 arXiv: 2305.11442 https://2023.aclweb.org/ en Alibaba-NTU-AIR2021B6 MOE-T1-RS21/20 © 2023 Association for Computational Linguistics. All rights reserved. This paper was published in the Proceedings of 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023) and is made available with permission of Association for Computational Linguistics. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer Science - Computation and Language Computer Science - Artificial Intelligence Engineering::Computer science and engineering::Computing methodologies::Document and text processing Zero-Shot Text Classification Self-Supervised Tuning |
spellingShingle |
Computer Science - Computation and Language Computer Science - Artificial Intelligence Engineering::Computer science and engineering::Computing methodologies::Document and text processing Zero-Shot Text Classification Self-Supervised Tuning Liu, Chaoqun Zhang, Wenxuan Chen, Guizhen Wu, Xiaobao Luu, Anh Tuan Chang, Chip Hong Bing, Lidong Zero-shot text classification via self-supervised tuning |
description |
Existing solutions to zero-shot text classification either conduct prompting
with pre-trained language models, which is sensitive to the choices of
templates, or rely on large-scale annotated data of relevant tasks for
meta-tuning. In this work, we propose a new paradigm based on self-supervised
learning to solve zero-shot text classification tasks by tuning the language
models with unlabeled data, called self-supervised tuning. By exploring the
inherent structure of free texts, we propose a new learning objective called
first sentence prediction to bridge the gap between unlabeled data and text
classification tasks. After tuning the model to learn to predict the first
sentence in a paragraph based on the rest, the model is able to conduct
zero-shot inference on unseen tasks such as topic classification and sentiment
analysis. Experimental results show that our model outperforms the
state-of-the-art baselines on 7 out of 10 tasks. Moreover, the analysis reveals
that our model is less sensitive to the prompt design. Our code and pre-trained
models are publicly available at https://github.com/DAMO-NLP-SG/SSTuning . |
author2 |
Interdisciplinary Graduate School (IGS) |
author_facet |
Interdisciplinary Graduate School (IGS) Liu, Chaoqun Zhang, Wenxuan Chen, Guizhen Wu, Xiaobao Luu, Anh Tuan Chang, Chip Hong Bing, Lidong |
format |
Conference or Workshop Item |
author |
Liu, Chaoqun Zhang, Wenxuan Chen, Guizhen Wu, Xiaobao Luu, Anh Tuan Chang, Chip Hong Bing, Lidong |
author_sort |
Liu, Chaoqun |
title |
Zero-shot text classification via self-supervised tuning |
title_short |
Zero-shot text classification via self-supervised tuning |
title_full |
Zero-shot text classification via self-supervised tuning |
title_fullStr |
Zero-shot text classification via self-supervised tuning |
title_full_unstemmed |
Zero-shot text classification via self-supervised tuning |
title_sort |
zero-shot text classification via self-supervised tuning |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/168505 https://2023.aclweb.org/ |
_version_ |
1772825864383234048 |