Sentiment analysis for software engineering: How far can pre-trained transformer models go?
Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions i...
Saved in:
Main Authors: | , , , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2020
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/5535 https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-6538 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-65382021-05-18T01:59:42Z Sentiment analysis for software engineering: How far can pre-trained transformer models go? ZHANG, Ting XU, Bowen Ferdian, Thung AGUS HARYONO, Stefanus LO, David JIANG, Lingxiao Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions in Stack Overflow and code review comments). Even so, there is still much room for improvement. Recently, pre-trained Transformer-based models (e.g., BERT, XLNet) have brought considerable breakthroughs in the field of natural language processing (NLP). In this work, we conducted a systematic evaluation of five existing SA4SE tools and variants of four state-of-the-art pre-trained Transformer-based models on six SE datasets. Our work is the first to fine-tune pre-trained Transformer-based models for the SA4SE task. Empirically, across all six datasets, our fine-tuned pre-trained Transformer-based models outperform the existing SA4SE tools by 6.5-35.6% in terms of macro/micro-averaged F1 scores. 2020-10-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/5535 info:doi/10.1109/ICSME46990.2020.00017 https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Natural Language Processing Pre-trained Models Sentiment Analysis Software Mining Software Engineering |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Natural Language Processing Pre-trained Models Sentiment Analysis Software Mining Software Engineering |
spellingShingle |
Natural Language Processing Pre-trained Models Sentiment Analysis Software Mining Software Engineering ZHANG, Ting XU, Bowen Ferdian, Thung AGUS HARYONO, Stefanus LO, David JIANG, Lingxiao Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
description |
Extensive research has been conducted on sentiment analysis for software engineering (SA4SE). Researchers have invested much effort in developing customized tools (e.g., SentiStrength-SE, SentiCR) to classify the sentiment polarity for Software Engineering (SE) specific contents (e.g., discussions in Stack Overflow and code review comments). Even so, there is still much room for improvement. Recently, pre-trained Transformer-based models (e.g., BERT, XLNet) have brought considerable breakthroughs in the field of natural language processing (NLP). In this work, we conducted a systematic evaluation of five existing SA4SE tools and variants of four state-of-the-art pre-trained Transformer-based models on six SE datasets. Our work is the first to fine-tune pre-trained Transformer-based models for the SA4SE task. Empirically, across all six datasets, our fine-tuned pre-trained Transformer-based models outperform the existing SA4SE tools by 6.5-35.6% in terms of macro/micro-averaged F1 scores. |
format |
text |
author |
ZHANG, Ting XU, Bowen Ferdian, Thung AGUS HARYONO, Stefanus LO, David JIANG, Lingxiao |
author_facet |
ZHANG, Ting XU, Bowen Ferdian, Thung AGUS HARYONO, Stefanus LO, David JIANG, Lingxiao |
author_sort |
ZHANG, Ting |
title |
Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
title_short |
Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
title_full |
Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
title_fullStr |
Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
title_full_unstemmed |
Sentiment analysis for software engineering: How far can pre-trained transformer models go? |
title_sort |
sentiment analysis for software engineering: how far can pre-trained transformer models go? |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2020 |
url |
https://ink.library.smu.edu.sg/sis_research/5535 https://ink.library.smu.edu.sg/context/sis_research/article/6538/viewcontent/icsme20SA4SE.pdf |
_version_ |
1770575502726135808 |