Sentiment analysis for software engineering: How far can pre-trained transformer models go?

Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions i...

Full description

Saved in:
Bibliographic Details
Main Authors: ZHANG, Ting, XU, Bowen, Ferdian, Thung, AGUS HARYONO, Stefanus, LO, David, JIANG, Lingxiao
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2020
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/5535
https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-6538
record_format dspace
spelling sg-smu-ink.sis_research-65382021-05-18T01:59:42Z Sentiment analysis for software engineering: How far can pre-trained transformer models go? ZHANG, Ting XU, Bowen Ferdian, Thung AGUS HARYONO, Stefanus LO, David JIANG, Lingxiao Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions in Stack Overflow and code review comments). Even so, there is still much room for improvement. Recently, pre-trained Transformer-based models (e.g., BERT, XLNet) have brought considerable breakthroughs in the field of natural language processing (NLP). In this work, we conducted a systematic evaluation of five existing SA4SE tools and variants of four state-of-the-art pre-trained Transformer-based models on six SE datasets. Our work is the first to fine-tune pre-trained Transformer-based models for the SA4SE task. Empirically, across all six datasets, our fine-tuned pre-trained Transformer-based models outperform the existing SA4SE tools by 6.5-35.6% in terms of macro/micro-averaged F1 scores. 2020-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5535 info:doi/10.1109/ICSME46990.2020.00017 https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Natural Language Processing Pre-trained Models Sentiment Analysis Software Mining Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Natural Language Processing
Pre-trained Models
Sentiment Analysis
Software Mining
Software Engineering
spellingShingle Natural Language Processing
Pre-trained Models
Sentiment Analysis
Software Mining
Software Engineering
ZHANG, Ting
XU, Bowen
Ferdian, Thung
AGUS HARYONO, Stefanus
LO, David
JIANG, Lingxiao
Sentiment analysis for software engineering: How far can pre-trained transformer models go?
description Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions in Stack Overflow and code review comments). Even so, there is still much room for improvement. Recently, pre-trained Transformer-based models (e.g., BERT, XLNet) have brought considerable breakthroughs in the field of natural language processing (NLP). In this work, we conducted a systematic evaluation of five existing SA4SE tools and variants of four state-of-the-art pre-trained Transformer-based models on six SE datasets. Our work is the first to fine-tune pre-trained Transformer-based models for the SA4SE task. Empirically, across all six datasets, our fine-tuned pre-trained Transformer-based models outperform the existing SA4SE tools by 6.5-35.6% in terms of macro/micro-averaged F1 scores.
format text
author ZHANG, Ting
XU, Bowen
Ferdian, Thung
AGUS HARYONO, Stefanus
LO, David
JIANG, Lingxiao
author_facet ZHANG, Ting
XU, Bowen
Ferdian, Thung
AGUS HARYONO, Stefanus
LO, David
JIANG, Lingxiao
author_sort ZHANG, Ting
title Sentiment analysis for software engineering: How far can pre-trained transformer models go?
title_short Sentiment analysis for software engineering: How far can pre-trained transformer models go?
title_full Sentiment analysis for software engineering: How far can pre-trained transformer models go?
title_fullStr Sentiment analysis for software engineering: How far can pre-trained transformer models go?
title_full_unstemmed Sentiment analysis for software engineering: How far can pre-trained transformer models go?
title_sort sentiment analysis for software engineering: how far can pre-trained transformer models go?
publisher Institutional Knowledge at Singapore Management University
publishDate 2020
url https://ink.library.smu.edu.sg/sis_research/5535
https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf
_version_ 1770575502726135808