Composition distillation for semantic sentence embeddings

The increasing demand for Natural Language Processing (NLP) solutions is driven by an exponential growth in digital content, communication platforms, and the undeniable need for sophisticated language understanding. This surge in demand also reflects the critical role of NLP in enabling machines to...

وصف كامل

محفوظ في:
التفاصيل البيبلوغرافية
المؤلف الرئيسي: Vaanavan, Sezhiyan
مؤلفون آخرون: Lihui Chen
التنسيق: Final Year Project
اللغة:English
منشور في: Nanyang Technological University 2024
الموضوعات:
NLP
LLM
الوصول للمادة أونلاين:https://hdl.handle.net/10356/177524
الوسوم: إضافة وسم
لا توجد وسوم, كن أول من يضع وسما على هذه التسجيلة!
الوصف
الملخص:The increasing demand for Natural Language Processing (NLP) solutions is driven by an exponential growth in digital content, communication platforms, and the undeniable need for sophisticated language understanding. This surge in demand also reflects the critical role of NLP in enabling machines to comprehend, interpret, and generate human-like text, which makes it a crucial technology in modern AI applications. Semantics, the study of meaning in languages, plays a pivotal role in NLP, encompassing the understanding of context, relationships, and nuances within multiple textual data. In recent years, there has been remarkable progress in utilizing pre-trained language models like BERT (Bidirectional Encoder Representations from Transformers) and GPT-3/4 (Generative Pre-trained Transformer) for semantic embeddings in NLP tasks. This project identifies and addresses a critical challenge within NLP that is commonly overlooked. The intricate composition of semantics within sentences often gets lost during model training, resulting in a lack of depth and precision in understanding the input language, leading to potential misinterpretations of textual data. This gap is hence addressed by enhancing already existing methods to distil semantic information from texts into smaller and more efficient models. By building upon the foundation laid by previous models, this project aims to improve the performance and accuracy of NLP systems by enhancing the quality and depth of semantic embeddings.