Evaluating Classical Software Process Models as Coordination Mechanisms for LLM-Based Software Generation
By: Duc Minh Ha , Phu Trac Kien , Tho Quan and more
Potential Business Impact:
Guides AI teams to build better software faster.
[Background] Large Language Model (LLM)-based multi-agent systems (MAS) are transforming software development by enabling autonomous collaboration. Classical software processes such asWaterfall, V-Model, and Agile offer structured coordination patterns that can be repurposed to guide these agent interactions. [Aims] This study explores how traditional software development processes can be adapted as coordination scaffolds for LLM based MAS and examines their impact on code quality, cost, and productivity. [Method] We executed 11 diverse software projects under three process models and four GPT variants, totaling 132 runs. Each output was evaluated using standardized metrics for size (files, LOC), cost (execution time, token usage), and quality (code smells, AI- and human detected bugs). [Results] Both process model and LLM choice significantly affected system performance. Waterfall was most efficient, V-Model produced the most verbose code, and Agile achieved the highest code quality, albeit at higher computational cost. [Conclusions] Classical software processes can be effectively instantiated in LLM-based MAS, but each entails trade-offs across quality, cost, and adaptability. Process selection should reflect project goals, whether prioritizing efficiency, robustness, or structured validation.
Similar Papers
Evaluating Software Process Models for Multi-Agent Class-Level Code Generation
Software Engineering
Helps AI write better, cleaner computer code.
Evaluation of LLMs for Process Model Analysis and Optimization
Artificial Intelligence
Helps computers find mistakes in business plans.
Evaluating the Process Modeling Abilities of Large Language Models -- Preliminary Foundations and Results
Computation and Language
Helps computers create better step-by-step plans.