Composition distillation for semantic sentence embeddings
The increasing demand for Natural Language Processing (NLP) solutions is driven by an exponential growth in digital content, communication platforms, and the undeniable need for sophisticated language understanding. This surge in demand also reflects the critical role of NLP in enabling machines to...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177524 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
Summary: | The increasing demand for Natural Language Processing (NLP) solutions is driven by an exponential growth in digital content, communication platforms, and the undeniable need for sophisticated language understanding. This surge in demand also reflects the critical role of NLP in enabling machines to comprehend, interpret, and generate human-like text, which makes it a crucial technology in modern AI applications.
Semantics, the study of meaning in languages, plays a pivotal role in NLP, encompassing the understanding of context, relationships, and nuances within multiple textual data. In recent years, there has been remarkable progress in utilizing pre-trained language models like BERT (Bidirectional Encoder Representations from Transformers) and GPT-3/4 (Generative Pre-trained Transformer) for semantic embeddings in NLP tasks. This project identifies and addresses a critical challenge within NLP that is commonly overlooked. The intricate composition of semantics within sentences often gets lost during model training, resulting in a lack of depth and precision in understanding the input language, leading to potential misinterpretations of textual data.
This gap is hence addressed by enhancing already existing methods to distil semantic information from texts into smaller and more efficient models. By building upon the foundation laid by previous models, this project aims to improve the performance and accuracy of NLP systems by enhancing the quality and depth of semantic embeddings. |
---|