Data and Documentation
Open Data Policy
FAQ
EN
DE
FR
Suchbegriff
Advanced search
Publication
Back to overview
Scented since the beginning: On the diffuseness of test smells in automatically generated test code
Type of publication
Peer-reviewed
Publikationsform
Original article (peer-reviewed)
Author
Grano Giovanni, Palomba Fabio, Nucci Dario Di, Lucia Andrea De, Gall Harald C.,
Project
SURF-MobileAppsData
Show all
Original article (peer-reviewed)
Journal
Journal on Systems and Software
Volume (Issue)
156
Page(s)
312 - 327
Title of proceedings
Journal on Systems and Software
DOI
10.1016/j.jss.2019.07.016
Open Access
URL
https://www.zora.uzh.ch/id/eprint/172208/
Type of Open Access
Repository (Green Open Access)
Abstract
Software testing represents a key software engineering practice to ensure source code quality and reliability. To support developers in this activity and reduce testing effort, several automated unit test generation tools have been proposed. Most of these approaches have the main goal of covering as more branches as possible. While these approaches have good performance, little is still known on the maintainability of the test code they produce, i.e.,whether the generated tests have a good code quality and if they do not possibly introduce issues threatening their effectiveness. To bridge this gap, in this paper we study to what extent existing automated test case generation tools produce potentially problematic test code. We consider seven test smells, i.e.,suboptimal design choices applied by programmers during the development of test cases, as measure of code quality of the generated tests, and evaluate their diffuseness in the unit test classes automatically generated by three state-of-the-art tools such as Randoop, JTExpert, and Evosuite. Moreover, we investigate whether there are characteristics of test and production code influencing the generation of smelly tests. Our study shows that all the considered tools tend to generate a high quantity of two specific test smell types, i.e.,Assertion Roulette and Eager Test, which are those that previous studies showed to negatively impact the reliability of production code. We also discover that test size is correlated with the generation of smelly tests. Based on our findings, we argue that more effective automated generation algorithms that explicitly take into account test code quality should be further investigated and devised.
-