Wav-BERT: Cooperative acoustic and linguistic representation learning for low-resource speech recognition
Unifying acoustic and linguistic representation learning has become increasingly crucial to transfer the knowledge learned on the abundance of high-resource language data for low-resource speech recognition. Existing approaches simply cascade pre-trained acoustic and language models to learn the tra...
Saved in:
Main Authors: | ZHENG, Guolin, XIAO, Yubei, GONG, Ke, ZHOU, Pan, LIANG, Xiaodan, LIN, Liang |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2021
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/9000 https://ink.library.smu.edu.sg/context/sis_research/article/10003/viewcontent/2021_EMNLP_Wav_BERT.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
Similar Items
-
Adversarial meta sampling for multilingual low-resource speech recognition
by: XIAO, Yubei, et al.
Published: (2021) -
Let’s think outside the box: Exploring leap-of-thought in large language models with multimodal humor generation
by: ZHONG, Shanshan, et al.
Published: (2024) -
Position-guided text prompt for vision-language pre-training
by: WANG, Alex Jinpeng, et al.
Published: (2023) -
Does BERT understand idioms? A probing-based empirical study of BERT encodings of idioms
by: TAN, Minghuan, et al.
Published: (2021) -
Aligning images in the wild
by: LIN, Wen-yan, et al.
Published: (2012)