Virtual prompt pre-training for prototype-based few-shot relation extraction
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding performance by reducing the gap between pre-training tasks and various downstream applications, which requires additional labor efforts in label word mappings and prompt template engineering. However, in a label intensive...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/170494 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-170494 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1704942023-09-15T06:39:46Z Virtual prompt pre-training for prototype-based few-shot relation extraction He, Kai Huang, Yucheng Mao, Rui Gong, Tieliang Li, Chen Cambria, Erik School of Computer Science and Engineering Engineering::Computer science and engineering Few-shot Learning Information Extraction Prompt tuning with pre-trained language models (PLM) has exhibited outstanding performance by reducing the gap between pre-training tasks and various downstream applications, which requires additional labor efforts in label word mappings and prompt template engineering. However, in a label intensive research domain, e.g., few-shot relation extraction (RE), manually defining label word mappings is particularly challenging, because the number of utilized relation label classes with complex relation names can be extremely large. Besides, the manual prompt development in natural language is subjective to individuals. To tackle these issues, we propose a virtual prompt pre-training method, projecting the virtual prompt to latent space, then fusing with PLM parameters. The pre-training is entity-relation-aware for RE, including the tasks of mask entity prediction, entity typing, distant supervised RE, and contrastive prompt pre-training. The proposed pre-training method can provide robust initialization for prompt encoding, while maintaining the interaction with the PLM. Furthermore, the virtual prompt can effectively avoid the labor efforts and the subjectivity issue in label word mapping and prompt template engineering. Our proposed prompt-based prototype network delivers a novel learning paradigm to model entities and relations via the probability distribution and Euclidean distance of the predictions of query instances and prototypes. The results indicate that our model yields an averaged accuracy gain of 4.21% on two few-shot datasets over strong RE baselines. Based on our proposed framework, our pre-trained model outperforms the strongest RE-related PLM by 6.52%. Agency for Science, Technology and Research (A*STAR) This research is supported by the Agency for Science, Technology and Research (A*STAR) under its AME Programmatic Funding Scheme (Project #A18A2b0046). This work is also supported by the Key Research and Development Program of Ningxia Hui Nationality Autonomous Region (2022BEG02025); The Key Research and Development Program of Shaanxi Province (2021GXLH-Z-095); The Innovative Research Group of the National Natural Science Foundation of China (61721002); The innovation team from the Ministry of Education (IRT_17R86). 2023-09-15T06:39:45Z 2023-09-15T06:39:45Z 2023 Journal Article He, K., Huang, Y., Mao, R., Gong, T., Li, C. & Cambria, E. (2023). Virtual prompt pre-training for prototype-based few-shot relation extraction. Expert Systems With Applications, 213(Part A), 118927-. https://dx.doi.org/10.1016/j.eswa.2022.118927 0957-4174 https://hdl.handle.net/10356/170494 10.1016/j.eswa.2022.118927 2-s2.0-85139597157 Part A 213 118927 en A18A2b0046 Expert Systems with Applications © 2022 Elsevier Ltd. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Few-shot Learning Information Extraction |
spellingShingle |
Engineering::Computer science and engineering Few-shot Learning Information Extraction He, Kai Huang, Yucheng Mao, Rui Gong, Tieliang Li, Chen Cambria, Erik Virtual prompt pre-training for prototype-based few-shot relation extraction |
description |
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding performance by reducing the gap between pre-training tasks and various downstream applications, which requires additional labor efforts in label word mappings and prompt template engineering. However, in a label intensive research domain, e.g., few-shot relation extraction (RE), manually defining label word mappings is particularly challenging, because the number of utilized relation label classes with complex relation names can be extremely large. Besides, the manual prompt development in natural language is subjective to individuals. To tackle these issues, we propose a virtual prompt pre-training method, projecting the virtual prompt to latent space, then fusing with PLM parameters. The pre-training is entity-relation-aware for RE, including the tasks of mask entity prediction, entity typing, distant supervised RE, and contrastive prompt pre-training. The proposed pre-training method can provide robust initialization for prompt encoding, while maintaining the interaction with the PLM. Furthermore, the virtual prompt can effectively avoid the labor efforts and the subjectivity issue in label word mapping and prompt template engineering. Our proposed prompt-based prototype network delivers a novel learning paradigm to model entities and relations via the probability distribution and Euclidean distance of the predictions of query instances and prototypes. The results indicate that our model yields an averaged accuracy gain of 4.21% on two few-shot datasets over strong RE baselines. Based on our proposed framework, our pre-trained model outperforms the strongest RE-related PLM by 6.52%. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering He, Kai Huang, Yucheng Mao, Rui Gong, Tieliang Li, Chen Cambria, Erik |
format |
Article |
author |
He, Kai Huang, Yucheng Mao, Rui Gong, Tieliang Li, Chen Cambria, Erik |
author_sort |
He, Kai |
title |
Virtual prompt pre-training for prototype-based few-shot relation extraction |
title_short |
Virtual prompt pre-training for prototype-based few-shot relation extraction |
title_full |
Virtual prompt pre-training for prototype-based few-shot relation extraction |
title_fullStr |
Virtual prompt pre-training for prototype-based few-shot relation extraction |
title_full_unstemmed |
Virtual prompt pre-training for prototype-based few-shot relation extraction |
title_sort |
virtual prompt pre-training for prototype-based few-shot relation extraction |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/170494 |
_version_ |
1779156807652999168 |