Disentangling syntactics, semantics, and pragmatics in natural language processing
In the era of deep learning, the natural language processing (NLP) community has become increasingly reliant on large language models (LLM), which are essentially probabilistic black-boxes. Hence, when applying them to downstream applications, we can only entrust that the LLMs has learned the variou...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Doctor of Philosophy |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177426 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-177426 |
---|---|
record_format |
dspace |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Computer and Information Science |
spellingShingle |
Computer and Information Science Zhang, Xulang Disentangling syntactics, semantics, and pragmatics in natural language processing |
description |
In the era of deep learning, the natural language processing (NLP) community has become increasingly reliant on large language models (LLM), which are essentially probabilistic black-boxes. Hence, when applying them to downstream applications, we can only entrust that the LLMs has learned the various linguistic phenomena in human language implicitly through extensive pretraining. Although LLMs are versatile and competent for many NLP tasks, they have yet to achieve human-like natural language understanding due to their stochastic nature. Additionally, LLMs offer limited interpretability and trustworthiness, making it difficult for humans to be involved in their learning process. To advance NLP studies, it is worth investigating whether a human-centered, neurosymbolic approach can address these issues.
One way to combat the limitations of LLMs is task decomposition, i.e., deconstructing a downstream task into its constituent subtasks. Such approach allows the symbolic representations of different linguistic layers, namely syntactics, semantics, and pragmatics, to be incorporated into the subsymbolic neural model. Specifically, syntactic layer extracts meanings from sentence constituents, and establish sentence structure.Semantic layer deconstructs texts into concepts, and resolves references.
Pragmatic layer is then able to extract meaning from both sentence structure and semantics obtained from prior layers.As such, in this thesis, we disentangle the syntactics, semantics, and pragmatics in NLP by studying tasks within each linguistic layer.
First, we investigate how syntactic tasks of different granularity might complement each other in a multitask learning setting. The tasks of study include sentence boundary disambiguation, text chunking, and part-of-speech tagging. We propose a soft-parameter-sharing multitask learning model that share local and global dependency information between tasks, and a curriculum learning mechanism to address the non-parallel label data problem. We evaluate our methods on the benchmark datasets, and provide findings on the pair-wise complementarity of the involved tasks.
Second, we examine the effective application of semantic processing in downstream task, i.e., sentiment analysis. Word sense disambiguation is a well-studied semantic task, whose application to downstream tasks has been limited by nature of its task setup. By reframing the disambiguation of word sense as a lexical substitution problem that replaces ambiguous words with more definitive ones, we propose a neurosymbolic sentiment analysis method that enables an explicit, dynamic, and explainable incorporation of word sense into the sentiment analysis task. We conduct experiments to prove that our framework is indeed effective in improving the performance of sentiment analysis.
Third, using sentiment analysis as a case study, we propose a sentiment knowledge base as a source of pragmatic symbolic representations. Our knowledge base differs from existing sentiment-annotated knowledge bases because it is human-centric and domain-adaptable. It accommodates varying sentiment perceptions of a word in different domains through the use of multi-dimensional representation, addressing the neglect of human subjectivity and disagreement in the prevalent single-score annotation paradigm.
Lastly, we present a shallow neurosymbolic model that alleviates the heaviness and rigidness of deep neural network and LLMs by coupling pragmatic and syntactic representations. The former is the sentiment-pragmatic representations from our proposed sentiment knowledge base, and the latter is the part-of-speech sequence of the input, provided by off-the-shelf tagger. Our model stands out for its lightweight nature, robustness across domains and languages, efficient few-shot training, and rapid convergence. We conduct extensive experiments to both show the effectiveness and robustness of our lightweight neurosymbolic framework, as well as the validity of our SenticVec knowledge base.
In summary, this thesis deconstructed NLP into syntactic, semantic, and pragmatic processing, and proposed novel solutions for tasks in each categories with an outlook towards human-centric neurosymbolic NLP. |
author2 |
Erik Cambria |
author_facet |
Erik Cambria Zhang, Xulang |
format |
Thesis-Doctor of Philosophy |
author |
Zhang, Xulang |
author_sort |
Zhang, Xulang |
title |
Disentangling syntactics, semantics, and pragmatics in natural language processing |
title_short |
Disentangling syntactics, semantics, and pragmatics in natural language processing |
title_full |
Disentangling syntactics, semantics, and pragmatics in natural language processing |
title_fullStr |
Disentangling syntactics, semantics, and pragmatics in natural language processing |
title_full_unstemmed |
Disentangling syntactics, semantics, and pragmatics in natural language processing |
title_sort |
disentangling syntactics, semantics, and pragmatics in natural language processing |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/177426 |
_version_ |
1814047098805420032 |
spelling |
sg-ntu-dr.10356-1774262024-06-03T06:51:19Z Disentangling syntactics, semantics, and pragmatics in natural language processing Zhang, Xulang Erik Cambria School of Computer Science and Engineering cambria@ntu.edu.sg Computer and Information Science In the era of deep learning, the natural language processing (NLP) community has become increasingly reliant on large language models (LLM), which are essentially probabilistic black-boxes. Hence, when applying them to downstream applications, we can only entrust that the LLMs has learned the various linguistic phenomena in human language implicitly through extensive pretraining. Although LLMs are versatile and competent for many NLP tasks, they have yet to achieve human-like natural language understanding due to their stochastic nature. Additionally, LLMs offer limited interpretability and trustworthiness, making it difficult for humans to be involved in their learning process. To advance NLP studies, it is worth investigating whether a human-centered, neurosymbolic approach can address these issues. One way to combat the limitations of LLMs is task decomposition, i.e., deconstructing a downstream task into its constituent subtasks. Such approach allows the symbolic representations of different linguistic layers, namely syntactics, semantics, and pragmatics, to be incorporated into the subsymbolic neural model. Specifically, syntactic layer extracts meanings from sentence constituents, and establish sentence structure.Semantic layer deconstructs texts into concepts, and resolves references. Pragmatic layer is then able to extract meaning from both sentence structure and semantics obtained from prior layers.As such, in this thesis, we disentangle the syntactics, semantics, and pragmatics in NLP by studying tasks within each linguistic layer. First, we investigate how syntactic tasks of different granularity might complement each other in a multitask learning setting. The tasks of study include sentence boundary disambiguation, text chunking, and part-of-speech tagging. We propose a soft-parameter-sharing multitask learning model that share local and global dependency information between tasks, and a curriculum learning mechanism to address the non-parallel label data problem. We evaluate our methods on the benchmark datasets, and provide findings on the pair-wise complementarity of the involved tasks. Second, we examine the effective application of semantic processing in downstream task, i.e., sentiment analysis. Word sense disambiguation is a well-studied semantic task, whose application to downstream tasks has been limited by nature of its task setup. By reframing the disambiguation of word sense as a lexical substitution problem that replaces ambiguous words with more definitive ones, we propose a neurosymbolic sentiment analysis method that enables an explicit, dynamic, and explainable incorporation of word sense into the sentiment analysis task. We conduct experiments to prove that our framework is indeed effective in improving the performance of sentiment analysis. Third, using sentiment analysis as a case study, we propose a sentiment knowledge base as a source of pragmatic symbolic representations. Our knowledge base differs from existing sentiment-annotated knowledge bases because it is human-centric and domain-adaptable. It accommodates varying sentiment perceptions of a word in different domains through the use of multi-dimensional representation, addressing the neglect of human subjectivity and disagreement in the prevalent single-score annotation paradigm. Lastly, we present a shallow neurosymbolic model that alleviates the heaviness and rigidness of deep neural network and LLMs by coupling pragmatic and syntactic representations. The former is the sentiment-pragmatic representations from our proposed sentiment knowledge base, and the latter is the part-of-speech sequence of the input, provided by off-the-shelf tagger. Our model stands out for its lightweight nature, robustness across domains and languages, efficient few-shot training, and rapid convergence. We conduct extensive experiments to both show the effectiveness and robustness of our lightweight neurosymbolic framework, as well as the validity of our SenticVec knowledge base. In summary, this thesis deconstructed NLP into syntactic, semantic, and pragmatic processing, and proposed novel solutions for tasks in each categories with an outlook towards human-centric neurosymbolic NLP. Doctor of Philosophy 2024-05-24T13:22:16Z 2024-05-24T13:22:16Z 2024 Thesis-Doctor of Philosophy Zhang, X. (2024). Disentangling syntactics, semantics, and pragmatics in natural language processing. Doctoral thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177426 https://hdl.handle.net/10356/177426 10.32657/10356/177426 en This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (CC BY-NC 4.0). application/pdf Nanyang Technological University |