Zero-shot text classification via self-supervised tuning

Existing solutions to zero-shot text classification either conduct prompting with pre-trained language models, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised...

Full description

Saved in:
Bibliographic Details
Main Authors: Liu, Chaoqun, Zhang, Wenxuan, Chen, Guizhen, Wu, Xiaobao, Luu, Anh Tuan, Chang, Chip Hong, Bing, Lidong
Other Authors: Interdisciplinary Graduate School (IGS)
Format: Conference or Workshop Item
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/168505
https://2023.aclweb.org/
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-168505
record_format dspace
spelling sg-ntu-dr.10356-1685052023-06-11T15:38:44Z Zero-shot text classification via self-supervised tuning Liu, Chaoqun Zhang, Wenxuan Chen, Guizhen Wu, Xiaobao Luu, Anh Tuan Chang, Chip Hong Bing, Lidong Interdisciplinary Graduate School (IGS) 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023) Alibaba Group Alibaba-NTU Singapore Joint Research Institute Computer Science - Computation and Language Computer Science - Artificial Intelligence Engineering::Computer science and engineering::Computing methodologies::Document and text processing Zero-Shot Text Classification Self-Supervised Tuning Existing solutions to zero-shot text classification either conduct prompting with pre-trained language models, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised learning to solve zero-shot text classification tasks by tuning the language models with unlabeled data, called self-supervised tuning. By exploring the inherent structure of free texts, we propose a new learning objective called first sentence prediction to bridge the gap between unlabeled data and text classification tasks. After tuning the model to learn to predict the first sentence in a paragraph based on the rest, the model is able to conduct zero-shot inference on unseen tasks such as topic classification and sentiment analysis. Experimental results show that our model outperforms the state-of-the-art baselines on 7 out of 10 tasks. Moreover, the analysis reveals that our model is less sensitive to the prompt design. Our code and pre-trained models are publicly available at https://github.com/DAMO-NLP-SG/SSTuning . Ministry of Education (MOE) Submitted/Accepted version This research is supported, in part, by Alibaba Group through Alibaba Innovative Research (AIR) Program and Alibaba-NTU Singapore Joint Research Institute (JRI), Nanyang Technological University, Singapore. Chaoqun Liu and Guizhen Chen extend their gratitude to Interdisciplinary Graduate Programme and School of Computer Science and Engineering, Nanyang Technological University, Singapore, for their support. This research is also supported by the Ministry of Education Tier 1 grant (MOE Tier 1 RS21/20). 2023-06-06T04:43:56Z 2023-06-06T04:43:56Z 2023 Conference Paper Liu, C., Zhang, W., Chen, G., Wu, X., Luu, A. T., Chang, C. H. & Bing, L. (2023). Zero-shot text classification via self-supervised tuning. 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023). https://hdl.handle.net/10356/168505 arXiv: 2305.11442 https://2023.aclweb.org/ en Alibaba-NTU-AIR2021B6 MOE-T1-RS21/20 © 2023 Association for Computational Linguistics. All rights reserved. This paper was published in the Proceedings of 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023) and is made available with permission of Association for Computational Linguistics. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer Science - Computation and Language
Computer Science - Artificial Intelligence
Engineering::Computer science and engineering::Computing methodologies::Document and text processing
Zero-Shot Text Classification
Self-Supervised Tuning
spellingShingle Computer Science - Computation and Language
Computer Science - Artificial Intelligence
Engineering::Computer science and engineering::Computing methodologies::Document and text processing
Zero-Shot Text Classification
Self-Supervised Tuning
Liu, Chaoqun
Zhang, Wenxuan
Chen, Guizhen
Wu, Xiaobao
Luu, Anh Tuan
Chang, Chip Hong
Bing, Lidong
Zero-shot text classification via self-supervised tuning
description Existing solutions to zero-shot text classification either conduct prompting with pre-trained language models, which is sensitive to the choices of templates, or rely on large-scale annotated data of relevant tasks for meta-tuning. In this work, we propose a new paradigm based on self-supervised learning to solve zero-shot text classification tasks by tuning the language models with unlabeled data, called self-supervised tuning. By exploring the inherent structure of free texts, we propose a new learning objective called first sentence prediction to bridge the gap between unlabeled data and text classification tasks. After tuning the model to learn to predict the first sentence in a paragraph based on the rest, the model is able to conduct zero-shot inference on unseen tasks such as topic classification and sentiment analysis. Experimental results show that our model outperforms the state-of-the-art baselines on 7 out of 10 tasks. Moreover, the analysis reveals that our model is less sensitive to the prompt design. Our code and pre-trained models are publicly available at https://github.com/DAMO-NLP-SG/SSTuning .
author2 Interdisciplinary Graduate School (IGS)
author_facet Interdisciplinary Graduate School (IGS)
Liu, Chaoqun
Zhang, Wenxuan
Chen, Guizhen
Wu, Xiaobao
Luu, Anh Tuan
Chang, Chip Hong
Bing, Lidong
format Conference or Workshop Item
author Liu, Chaoqun
Zhang, Wenxuan
Chen, Guizhen
Wu, Xiaobao
Luu, Anh Tuan
Chang, Chip Hong
Bing, Lidong
author_sort Liu, Chaoqun
title Zero-shot text classification via self-supervised tuning
title_short Zero-shot text classification via self-supervised tuning
title_full Zero-shot text classification via self-supervised tuning
title_fullStr Zero-shot text classification via self-supervised tuning
title_full_unstemmed Zero-shot text classification via self-supervised tuning
title_sort zero-shot text classification via self-supervised tuning
publishDate 2023
url https://hdl.handle.net/10356/168505
https://2023.aclweb.org/
_version_ 1772825864383234048