Technique to Baseline QE Artefact Generation Aligned to Quality Metrics
By: Eitan Farchi , Kiran Nayak , Papia Ghosh Majumdar and more
Potential Business Impact:
Checks computer-made plans for quality.
Large Language Models (LLMs) are transforming Quality Engineering (QE) by automating the generation of artefacts such as requirements, test cases, and Behavior Driven Development (BDD) scenarios. However, ensuring the quality of these outputs remains a challenge. This paper presents a systematic technique to baseline and evaluate QE artefacts using quantifiable metrics. The approach combines LLM-driven generation, reverse generation , and iterative refinement guided by rubrics technique for clarity, completeness, consistency, and testability. Experimental results across 12 projects show that reverse-generated artefacts can outperform low-quality inputs and maintain high standards when inputs are strong. The framework enables scalable, reliable QE artefact validation, bridging automation with accountability.
Similar Papers
Advancing Software Quality: A Standards-Focused Review of LLM-Based Assurance Techniques
Software Engineering
AI helps make computer programs better and safer.
Large Language Models for Unit Test Generation: Achievements, Challenges, and the Road Ahead
Software Engineering
Helps computers write better code tests automatically.
A Study on the Improvement of Code Generation Quality Using Large Language Models Leveraging Product Documentation
Software Engineering
Makes apps work right by testing them automatically.