PerfLearner: learning from bug reports to understand and generate performance test frames

Software performance is important for ensuring the quality of software products. Performance bugs, defined as programming errors that cause significant performance degradation, can lead to slow systems and poor user experience. While there has been some research on automated performance testing such...

Full description

Saved in:
Bibliographic Details
Main Authors: HAN, Xue, YU, Tingting, LO, David
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2018
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/4298
https://ink.library.smu.edu.sg/context/sis_research/article/5301/viewcontent/ase18main_p354_p.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-5301
record_format dspace
spelling sg-smu-ink.sis_research-53012019-06-06T07:14:35Z PerfLearner: learning from bug reports to understand and generate performance test frames HAN, Xue YU, Tingting LO, David Software performance is important for ensuring the quality of software products. Performance bugs, defined as programming errors that cause significant performance degradation, can lead to slow systems and poor user experience. While there has been some research on automated performance testing such as test case generation, the main idea is to select workload values to increase the program execution times. These techniques often assume the initial test cases have the right combination of input parameters and focus on evolving values of certain input parameters. However, such an assumption may not hold for highly configurable real-word applications, in which the combinations of input parameters can be very large. In this paper, we manually analyze 300 bug reports from three large open source projects - Apache HTTP Server, MySQL, and Mozilla Firefox. We found that 1) exposing performance bugs often requires combinations of multiple input parameters, and 2) certain input parameters are frequently involved in exposing performance bugs. Guided by these findings, we designed and evaluated an automated approach, PerfLearner, to extract execution commands and input parameters from descriptions of performance bug reports and use them to generate test frames for guiding actual performance test case generation. 2018-12-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4298 info:doi/10.1145/3238147.3238204 https://ink.library.smu.edu.sg/context/sis_research/article/5301/viewcontent/ase18main_p354_p.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Performance bugs Software mining Software testing Programming Languages and Compilers Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Performance bugs
Software mining
Software testing
Programming Languages and Compilers
Software Engineering
spellingShingle Performance bugs
Software mining
Software testing
Programming Languages and Compilers
Software Engineering
HAN, Xue
YU, Tingting
LO, David
PerfLearner: learning from bug reports to understand and generate performance test frames
description Software performance is important for ensuring the quality of software products. Performance bugs, defined as programming errors that cause significant performance degradation, can lead to slow systems and poor user experience. While there has been some research on automated performance testing such as test case generation, the main idea is to select workload values to increase the program execution times. These techniques often assume the initial test cases have the right combination of input parameters and focus on evolving values of certain input parameters. However, such an assumption may not hold for highly configurable real-word applications, in which the combinations of input parameters can be very large. In this paper, we manually analyze 300 bug reports from three large open source projects - Apache HTTP Server, MySQL, and Mozilla Firefox. We found that 1) exposing performance bugs often requires combinations of multiple input parameters, and 2) certain input parameters are frequently involved in exposing performance bugs. Guided by these findings, we designed and evaluated an automated approach, PerfLearner, to extract execution commands and input parameters from descriptions of performance bug reports and use them to generate test frames for guiding actual performance test case generation.
format text
author HAN, Xue
YU, Tingting
LO, David
author_facet HAN, Xue
YU, Tingting
LO, David
author_sort HAN, Xue
title PerfLearner: learning from bug reports to understand and generate performance test frames
title_short PerfLearner: learning from bug reports to understand and generate performance test frames
title_full PerfLearner: learning from bug reports to understand and generate performance test frames
title_fullStr PerfLearner: learning from bug reports to understand and generate performance test frames
title_full_unstemmed PerfLearner: learning from bug reports to understand and generate performance test frames
title_sort perflearner: learning from bug reports to understand and generate performance test frames
publisher Institutional Knowledge at Singapore Management University
publishDate 2018
url https://ink.library.smu.edu.sg/sis_research/4298
https://ink.library.smu.edu.sg/context/sis_research/article/5301/viewcontent/ase18main_p354_p.pdf
_version_ 1770574603461066752