Diffusion models for natural language processing

The autoregressive method has long been fundamental in natural language processing (NLP), where models predict the next token based on previously generated context. While effective, this approach suffers from inherent limitations, including slow generation speed and sequential dependency, hindering...

Full description

Saved in:
Bibliographic Details
Main Author: Hoang, Minh Nhat
Other Authors: Luu Anh Tuan
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/174904
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-174904
record_format dspace
spelling sg-ntu-dr.10356-1749042024-04-19T15:44:40Z Diffusion models for natural language processing Hoang, Minh Nhat Luu Anh Tuan School of Computer Science and Engineering anhtuan.luu@ntu.edu.sg Computer and Information Science Natural language processing The autoregressive method has long been fundamental in natural language processing (NLP), where models predict the next token based on previously generated context. While effective, this approach suffers from inherent limitations, including slow generation speed and sequential dependency, hindering its scalability and efficiency in large-scale applications. Non-autoregressive methods offer a promising solution by enabling parallel decoding and mitigating the sequential bottleneck. In recent years, diffusion models have proven its superior performance in continuous domain such as image and video, and have also recently emerged as a non-autoregressive alternative in NLP. This paper explores the application of diffusion models across various NLP tasks, ranging from language modeling to sequence generation. Through extensive experimentation, we demonstrate that diffusion models not only rival but surpass predominated autoregressive architecture such as the fine-tuned GPT-2 Large models. Moreover, our analysis reveals that diffusion models exhibit a remarkable ability to generate diverse outputs under identical settings, indicating their potential to revolutionize NLP tasks. In summary, this paper presents a comprehensive investigation into the application of diffusion models in NLP, shedding light on their efficacy and potential to address longstanding challenges associated with autoregressive methods. These findings underscore the significance of diffusion models as a viable alternative for enhancing efficiency and scalability in natural language processing tasks. Bachelor's degree 2024-04-16T01:47:47Z 2024-04-16T01:47:47Z 2024 Final Year Project (FYP) Hoang, M. N. (2024). Diffusion models for natural language processing. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/174904 https://hdl.handle.net/10356/174904 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Computer and Information Science
Natural language processing
spellingShingle Computer and Information Science
Natural language processing
Hoang, Minh Nhat
Diffusion models for natural language processing
description The autoregressive method has long been fundamental in natural language processing (NLP), where models predict the next token based on previously generated context. While effective, this approach suffers from inherent limitations, including slow generation speed and sequential dependency, hindering its scalability and efficiency in large-scale applications. Non-autoregressive methods offer a promising solution by enabling parallel decoding and mitigating the sequential bottleneck. In recent years, diffusion models have proven its superior performance in continuous domain such as image and video, and have also recently emerged as a non-autoregressive alternative in NLP. This paper explores the application of diffusion models across various NLP tasks, ranging from language modeling to sequence generation. Through extensive experimentation, we demonstrate that diffusion models not only rival but surpass predominated autoregressive architecture such as the fine-tuned GPT-2 Large models. Moreover, our analysis reveals that diffusion models exhibit a remarkable ability to generate diverse outputs under identical settings, indicating their potential to revolutionize NLP tasks. In summary, this paper presents a comprehensive investigation into the application of diffusion models in NLP, shedding light on their efficacy and potential to address longstanding challenges associated with autoregressive methods. These findings underscore the significance of diffusion models as a viable alternative for enhancing efficiency and scalability in natural language processing tasks.
author2 Luu Anh Tuan
author_facet Luu Anh Tuan
Hoang, Minh Nhat
format Final Year Project
author Hoang, Minh Nhat
author_sort Hoang, Minh Nhat
title Diffusion models for natural language processing
title_short Diffusion models for natural language processing
title_full Diffusion models for natural language processing
title_fullStr Diffusion models for natural language processing
title_full_unstemmed Diffusion models for natural language processing
title_sort diffusion models for natural language processing
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/174904
_version_ 1800916241339121664