SBFT Tool Competition 2024 - Python Test Case Generation Track
Creators
Description
Test case generation (TCG) for Python poses distinctive challenges due to the language’s dynamic nature and the absence of strict type information. Previous research has successfully explored automated unit TCG for Python, with solutions outperforming random test generation methods. Nevertheless, fundamental issues persist, hindering the practical adoption of existing test case generators. To address these challenges, we report on the organization, challenges, and results of the first edition of the Python Testing Competition. Four tools, i.e., UTBotPython, Klara, Hypothesis Ghostwriter, and Pynguin were executed on a benchmark with 35 Python source files sampled from 7 open-source Python projects, for a time budget of 400 seconds. We considered one configuration for each test subject and evaluated the tools’ effectiveness in terms of code and mutation coverage. This paper describes our methodology, the analysis of the results together with the competing tools, and the challenges faced while running the competition experiments.
Files
python-tool-competition-2024-hypothesis-ghostwriter-main.zip
Files
(351.5 kB)
Name | Size | Download all |
---|---|---|
md5:1ef0a000689dba5e855b2f9cab2e113b
|
46.6 kB | Preview Download |
md5:c3315293d84bfe09dee8c6c3a05c7664
|
46.6 kB | Preview Download |
md5:8d6eb547edebe81fd8d9044d4ed31ab2
|
193.8 kB | Preview Download |
md5:47eb7d105c4782fee704762a2f9cccb1
|
59.4 kB | Preview Download |
md5:a44f85684a7394dc9a98a783f3ac25bc
|
5.1 kB | Preview Download |