Injecting descriptive meta-information into pre-trained language models with hypernetworks
Pre-trained language models have been widely adopted as backbones in various natural language processing tasks. However, existing pre-trained language models ignore the descriptive meta-information in the text such as the distinction between the title and the mainbody, leading to over-weighted atten...
Saved in:
Main Authors: | DUAN, Wenying, HE, Xiaoxi, ZHOU, Zimu, RAO, Hong, THIELE, Lothar |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/6237 https://ink.library.smu.edu.sg/context/sis_research/article/7240/viewcontent/Injecting_Descriptive_Meta_Information_into_Pre_Trained_Language_Models_with_Hypernetworks.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Generative AI art - hypernetworks
by: Chee, Mei Qi
Published: (2024) -
Using pre-trained models for vision-language understanding tasks
by: CAO, Rui
Published: (2024) -
On the transferability of pre-trained language models for low-resource programming languages
by: CHEN, Fuxiang, et al.
Published: (2022) -
p-Meta: Towards on-device deep model adaptation
by: QU, Zhongnan, et al.
Published: (2022) -
Pre-training graph transformer with multimodal side information for recommendation
by: Liu, Yong, et al.
Published: (2022)