An important goal of automatic testing techniques, including random testing is to achieve high code coverage with a minimum set of test cases. To meet this goal, random testing researchers have proposed many techniques to generate test inputs and method call sequences that yield higher code coverage. However, most proposed random testing techniques are only suitable for toy systems, and they achieve low code coverage rates while generating too many unnecessary test cases on large-scale software systems. We propose GENRED, a tool that utilizes three approaches: input on demand creation and coverage-based method selection techniques that enhance Randoop, a state-of-the-art feedback-directed random testing technique, and finally, a sequence-based reduction technique that removes redundant test cases without executing them. We evaluate GENRED as a tool to test four open-source systems. The results show that these techniques achieve branch coverage
Hojun Jaygarl, Kai-Shin Lu, Carl K. Chang