Computing Reviews
Today's Issue Hot Topics Search Browse Recommended My Account Log In
Review Help
Search
Empirically evaluating the quality of automatically generated and manually written test suites
Kracht J., Petrovic J., Walcott-Justice K.  QSIC 2014 (Proceedings of the 2014 14th International Conference on Quality Software, Dallas, TX, Oct 2-3, 2014)256-265.2014.Type:Proceedings
Date Reviewed: Mar 25 2015

Are automatically generated test suites better than manually written test suites? To answer this question, ten Java applications with existing manually written test suites were tested using the EVOSUITE and CodePro automated test generation tools. Branch coverage and mutation scores were used to assess the quality of test suites. The Jacoco and MAJOR tools were used to calculate these measures.

EVOSUITE covered 31.86 percent of branches on average and had an average mutation score of 39.89 percent. For the manually written test suites, the figures were, respectively, 31.5 percent and 42.14 percent. The authors conclude that their results should encourage use of a tool such as EVOSUITE for test production. CodePro’s test quality was found to be much lower and this was attributed to absent or weaker oracles.

Also investigated was the relationship between branch coverage and mutation score. By inspection, Figures 7 and 8 do indeed suggest correlations are present for EVOSUITE and the manually written test suites. It is unclear, however, what the actual correlation scores are. The investigators imply they calculated non-linear fits, but in Figures 7 and 8, linear lines are drawn.

The analysis presented has two major weaknesses. First, there is no discussion or treatment of equivalent mutants. There can be sizable changes in mutation scores when equivalent mutants are factored out. Second, there is no discussion or treatment of the degree to which branches actually tested overlapped with branches actually containing mutations.

Despite the shortcomings identified, this paper is strongly recommended to those working in software testing.

Reviewer:  Andy Brooks Review #: CR143277 (1506-0487)
Bookmark and Share
  Reviewer Selected
Featured Reviewer
 
 
Testing Tools (D.2.5 ... )
 
 
Process Metrics (D.2.8 ... )
 
Would you recommend this review?
yes
no
Other reviews under "Testing Tools": Date
Automatic generation of random self-checking test cases
Bird D., Munoz C. IBM Systems Journal 22(3): 229-245, 1983. Type: Article
Aug 1 1985
Program testing by specification mutation
Budd T., Gopal A. Information Systems 10(1): 63-73, 1985. Type: Article
Feb 1 1986
SEES--a software testing environment support system
Roussopoulos N., Yeh R. (ed) IEEE Transactions on Software Engineering SE-11(4): 355-366, 1985. Type: Article
Apr 1 1986
more...

E-Mail This Printer-Friendly
Send Your Comments
Contact Us
Reproduction in whole or in part without permission is prohibited.   Copyright 1999-2024 ThinkLoud®
Terms of Use
| Privacy Policy