Exploring and adapting Chinese GPT to pinyin input method

While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyi...

Full description

Saved in:
Bibliographic Details
Main Authors: TAN, Minghuan, DAI, Yong, TANG, Duyu, FENG, Zhangyin, HUANG, Guoping, JIANG, Jing, LI, Jiwei, SHI, Shuming
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7474
https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8477
record_format dspace
spelling sg-smu-ink.sis_research-84772023-04-04T02:51:53Z Exploring and adapting Chinese GPT to pinyin input method TAN, Minghuan DAI, Yong TANG, Duyu FENG, Zhangyin HUANG, Guoping JIANG, Jing LI, Jiwei SHI, Shuming While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyin. However, the performance drops dramatically when the input includes abbreviated pinyin. A reason is that an abbreviated pinyin can be mapped to many perfect pinyin, which links to even larger number of Chinese characters. We mitigate this issue with two strategies, including enriching the context with pinyin and optimizing the training process to help distinguish homophones. To further facilitate the evaluation of pinyin input method, we create a dataset consisting of 270K instances from fifteen domains. Results show that our approach improves the performance on abbreviated pinyin across all domains. Model analysis demonstrates that both strategies contribute to the performance boost. 2022-05-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7474 info:doi/10.18653/v1/2022.acl-long.133 https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf http://creativecommons.org/licenses/by/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Chinese characters Input methods ITS applications Modeling analyzes training process Artificial Intelligence and Robotics Databases and Information Systems Programming Languages and Compilers
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Chinese characters
Input methods
ITS applications
Modeling analyzes
training process
Artificial Intelligence and Robotics
Databases and Information Systems
Programming Languages and Compilers
spellingShingle Chinese characters
Input methods
ITS applications
Modeling analyzes
training process
Artificial Intelligence and Robotics
Databases and Information Systems
Programming Languages and Compilers
TAN, Minghuan
DAI, Yong
TANG, Duyu
FENG, Zhangyin
HUANG, Guoping
JIANG, Jing
LI, Jiwei
SHI, Shuming
Exploring and adapting Chinese GPT to pinyin input method
description While GPT has become the de-facto method for text generation tasks, its application to pinyin input method remains unexplored. In this work, we make the first exploration to leverage Chinese GPT for pinyin input method. We find that a frozen GPT achieves state-of-the-art performance on perfect pinyin. However, the performance drops dramatically when the input includes abbreviated pinyin. A reason is that an abbreviated pinyin can be mapped to many perfect pinyin, which links to even larger number of Chinese characters. We mitigate this issue with two strategies, including enriching the context with pinyin and optimizing the training process to help distinguish homophones. To further facilitate the evaluation of pinyin input method, we create a dataset consisting of 270K instances from fifteen domains. Results show that our approach improves the performance on abbreviated pinyin across all domains. Model analysis demonstrates that both strategies contribute to the performance boost.
format text
author TAN, Minghuan
DAI, Yong
TANG, Duyu
FENG, Zhangyin
HUANG, Guoping
JIANG, Jing
LI, Jiwei
SHI, Shuming
author_facet TAN, Minghuan
DAI, Yong
TANG, Duyu
FENG, Zhangyin
HUANG, Guoping
JIANG, Jing
LI, Jiwei
SHI, Shuming
author_sort TAN, Minghuan
title Exploring and adapting Chinese GPT to pinyin input method
title_short Exploring and adapting Chinese GPT to pinyin input method
title_full Exploring and adapting Chinese GPT to pinyin input method
title_fullStr Exploring and adapting Chinese GPT to pinyin input method
title_full_unstemmed Exploring and adapting Chinese GPT to pinyin input method
title_sort exploring and adapting chinese gpt to pinyin input method
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7474
https://ink.library.smu.edu.sg/context/sis_research/article/8477/viewcontent/2022.acl_long.133.pdf
_version_ 1770576352946159616