Large Language Models for Software Testing: A Research Roadmap
By: Cristian Augusto , Antonia Bertolino , Guglielmo De Angelis and more
Potential Business Impact:
Helps computers find mistakes in programs.
Large Language Models (LLMs) are starting to be profiled as one of the most significant disruptions in the Software Testing field. Specifically, they have been successfully applied in software testing tasks such as generating test code, or summarizing documentation. This potential has attracted hundreds of researchers, resulting in dozens of new contributions every month, hardening researchers to stay at the forefront of the wave. Still, to the best of our knowledge, no prior work has provided a structured vision of the progress and most relevant research trends in LLM-based testing. In this article, we aim to provide a roadmap that illustrates its current state, grouping the contributions into different categories, and also sketching the most promising and active research directions for the field. To achieve this objective, we have conducted a semi-systematic literature review, collecting articles and mapping them into the most prominent categories, reviewing the current and ongoing status, and analyzing the open challenges of LLM-based software testing. Lastly, we have outlined several expected long-term impacts of LLMs over the whole software testing field.
Similar Papers
Large Language Models for Unit Test Generation: Achievements, Challenges, and the Road Ahead
Software Engineering
Helps computers write better code tests automatically.
Large Language Models for Unit Testing: A Systematic Literature Review
Software Engineering
Helps computers write and check their own code.
Software Testing with Large Language Models: An Interview Study with Practitioners
Software Engineering
Helps testers use smart computer programs better.