A Study on the Improvement of Code Generation Quality Using Large Language Models Leveraging Product Documentation
By: Takuro Morimoto, Harumi Haraguchi
Potential Business Impact:
Makes apps work right by testing them automatically.
Research on using Large Language Models (LLMs) in system development is expanding, especially in automated code and test generation. While E2E testing is vital for ensuring application quality, most test generation research has focused on unit tests, with limited work on E2E test code. This study proposes a method for automatically generating E2E test code from product documentation such as manuals, FAQs, and tutorials using LLMs with tailored prompts. The two step process interprets documentation intent and produces executable test code. Experiments on a web app with six key features (e.g., authentication, profile, discussion) showed that tests generated from product documentation had high compilation success and functional coverage, outperforming those based on requirement specs and user stories. These findings highlight the potential of product documentation to improve E2E test quality and, by extension, software quality.
Similar Papers
Large Language Models for Unit Test Generation: Achievements, Challenges, and the Road Ahead
Software Engineering
Helps computers write better code tests automatically.
Software Testing with Large Language Models: An Interview Study with Practitioners
Software Engineering
Helps testers use smart computer programs better.
Large Language Models for Unit Testing: A Systematic Literature Review
Software Engineering
Helps computers write and check their own code.