STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM.
Sentiment Analysis or SA is one of the most widely used fields in Natural Language Processing or NLP today. SA itself also has a lot of branches such as aspect-based, end2end, targeted, and more. Because of this many branches, it becomes difficult to track the development of SA as whole. To overc...
Saved in:
Main Author: | |
---|---|
Format: | Final Project |
Language: | Indonesia |
Online Access: | https://digilib.itb.ac.id/gdl/view/77876 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Institut Teknologi Bandung |
Language: | Indonesia |
id |
id-itb.:77876 |
---|---|
spelling |
id-itb.:778762023-09-15T08:34:32ZSTRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. Putra Tjandra, Andrianata Indonesia Final Project Structured Sentiment Analysis; BART; bi-LSTM; AI; NLP INSTITUT TEKNOLOGI BANDUNG https://digilib.itb.ac.id/gdl/view/77876 Sentiment Analysis or SA is one of the most widely used fields in Natural Language Processing or NLP today. SA itself also has a lot of branches such as aspect-based, end2end, targeted, and more. Because of this many branches, it becomes difficult to track the development of SA as whole. To overcome this, a solution is given in the form of Structured Sentiment Analysis which aims to predict the holder, target, expression, and polarity of sentences collectively in a tuple. One of the methods used is the BART-based Encoder-Decoder, but this method has a weakness in handling long sentences and sentences without opinions. Therefore, a solution is proposed to overcome these weaknesses by combining BART with an RNN or Recurrent Neural Network. From a number of alternative RNNs that have been studied, a bi-LSTM was finally selected for use. From the results obtained, the use of bi-LSTM can improve the performance of BART-based Encoder-Decoder. text |
institution |
Institut Teknologi Bandung |
building |
Institut Teknologi Bandung Library |
continent |
Asia |
country |
Indonesia Indonesia |
content_provider |
Institut Teknologi Bandung |
collection |
Digital ITB |
language |
Indonesia |
description |
Sentiment Analysis or SA is one of the most widely used fields in Natural Language Processing
or NLP today. SA itself also has a lot of branches such as aspect-based, end2end, targeted, and
more. Because of this many branches, it becomes difficult to track the development of SA as
whole. To overcome this, a solution is given in the form of Structured Sentiment Analysis which
aims to predict the holder, target, expression, and polarity of sentences collectively in a tuple.
One of the methods used is the BART-based Encoder-Decoder, but this method has a weakness
in handling long sentences and sentences without opinions. Therefore, a solution is proposed to
overcome these weaknesses by combining BART with an RNN or Recurrent Neural Network.
From a number of alternative RNNs that have been studied, a bi-LSTM was finally selected for
use. From the results obtained, the use of bi-LSTM can improve the performance of BART-based
Encoder-Decoder. |
format |
Final Project |
author |
Putra Tjandra, Andrianata |
spellingShingle |
Putra Tjandra, Andrianata STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
author_facet |
Putra Tjandra, Andrianata |
author_sort |
Putra Tjandra, Andrianata |
title |
STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
title_short |
STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
title_full |
STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
title_fullStr |
STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
title_full_unstemmed |
STRUCTURED SENTIMENT ANALYSIS MODELING WITH BART BASED ENCODER-DECODER AND BI-LSTM. |
title_sort |
structured sentiment analysis modeling with bart based encoder-decoder and bi-lstm. |
url |
https://digilib.itb.ac.id/gdl/view/77876 |
_version_ |
1822280867000614912 |