Exploring and adapting Chinese GPT to pinyin input method
While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyi...
Saved in:
Main Authors: | , , , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2022
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7474 https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-8477 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-84772023-04-04T02:51:53Z Exploring and adapting Chinese GPT to pinyin input method TAN, Minghuan DAI, Yong TANG, Duyu FENG, Zhangyin HUANG, Guoping JIANG, Jing LI, Jiwei SHI, Shuming While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyin. However, the performance drops dramatically when the input includes abbreviated pinyin. A reason is that an abbreviated pinyin can be mapped to many perfect pinyin, which links to even larger number of Chinese characters. We mitigate this issue with two strategies, including enriching the context with pinyin and optimizing the training process to help distinguish homophones. To further facilitate the evaluation of pinyin input method, we create a dataset consisting of 270K instances from fifteen domains. Results show that our approach improves the performance on abbreviated pinyin across all domains. Model analysis demonstrates that both strategies contribute to the performance boost. 2022-05-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7474 info:doi/10.18653/v1/2022.acl-long.133 https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf http://creativecommons.org/licenses/by/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Chinese characters Input methods ITS applications Modeling analyzes training process Artificial Intelligence and Robotics Databases and Information Systems Programming Languages and Compilers |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Chinese characters Input methods ITS applications Modeling analyzes training process Artificial Intelligence and Robotics Databases and Information Systems Programming Languages and Compilers |
spellingShingle |
Chinese characters Input methods ITS applications Modeling analyzes training process Artificial Intelligence and Robotics Databases and Information Systems Programming Languages and Compilers TAN, Minghuan DAI, Yong TANG, Duyu FENG, Zhangyin HUANG, Guoping JIANG, Jing LI, Jiwei SHI, Shuming Exploring and adapting Chinese GPT to pinyin input method |
description |
While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyin. However, the performance drops dramatically when the input includes abbreviated pinyin. A reason is that an abbreviated pinyin can be mapped to many perfect pinyin, which links to even larger number of Chinese characters. We mitigate this issue with two strategies, including enriching the context with pinyin and optimizing the training process to help distinguish homophones. To further facilitate the evaluation of pinyin input method, we create a dataset consisting of 270K instances from fifteen domains. Results show that our approach improves the performance on abbreviated pinyin across all domains. Model analysis demonstrates that both strategies contribute to the performance boost. |
format |
text |
author |
TAN, Minghuan DAI, Yong TANG, Duyu FENG, Zhangyin HUANG, Guoping JIANG, Jing LI, Jiwei SHI, Shuming |
author_facet |
TAN, Minghuan DAI, Yong TANG, Duyu FENG, Zhangyin HUANG, Guoping JIANG, Jing LI, Jiwei SHI, Shuming |
author_sort |
TAN, Minghuan |
title |
Exploring and adapting Chinese GPT to pinyin input method |
title_short |
Exploring and adapting Chinese GPT to pinyin input method |
title_full |
Exploring and adapting Chinese GPT to pinyin input method |
title_fullStr |
Exploring and adapting Chinese GPT to pinyin input method |
title_full_unstemmed |
Exploring and adapting Chinese GPT to pinyin input method |
title_sort |
exploring and adapting chinese gpt to pinyin input method |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2022 |
url |
https://ink.library.smu.edu.sg/sis_research/7474 https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf |
_version_ |
1770576352946159616 |