Debate-Feedback: A Multi-Agent Framework for Efficient Legal Judgment Prediction
By: Xi Chen , Mao Mao , Shuo Li and more
Potential Business Impact:
AI helps predict court cases faster.
The use of AI in legal analysis and prediction (LegalAI) has gained widespread attention, with past research focusing on retrieval-based methods and fine-tuning large models. However, these approaches often require large datasets and underutilize the capabilities of modern large language models (LLMs). In this paper, inspired by the debate phase of real courtroom trials, we propose a novel legal judgment prediction model based on the Debate-Feedback architecture, which integrates LLM multi-agent debate and reliability evaluation models. Unlike traditional methods, our model achieves significant improvements in efficiency by minimizing the need for large historical datasets, thus offering a lightweight yet robust solution. Comparative experiments show that it outperforms several general-purpose and domain-specific legal models, offering a dynamic reasoning process and a promising direction for future LegalAI research.
Similar Papers
Multi-Agent Debate for LLM Judges with Adaptive Stability Detection
Artificial Intelligence
Debating computers make better judgments than voting ones.
Multi-Agent LLM Judge: automatic personalized LLM judge design for evaluating natural language generation applications
Computation and Language
Helps computers judge writing better than people.
Efficient LLM Safety Evaluation through Multi-Agent Debate
Artificial Intelligence
Makes AI safer and cheaper to test.