ReGal: A First Look at PPO-based Legal AI for Judgment Prediction and Summarization in India
By: Shubham Kumar Nigam , Tanuj Tyagi , Siddharth Shukla and more
This paper presents an early exploration of reinforcement learning methodologies for legal AI in the Indian context. We introduce Reinforcement Learning-based Legal Reasoning (ReGal), a framework that integrates Multi-Task Instruction Tuning with Reinforcement Learning from AI Feedback (RLAIF) using Proximal Policy Optimization (PPO). Our approach is evaluated across two critical legal tasks: (i) Court Judgment Prediction and Explanation (CJPE), and (ii) Legal Document Summarization. Although the framework underperforms on standard evaluation metrics compared to supervised and proprietary models, it provides valuable insights into the challenges of applying RL to legal texts. These challenges include reward model alignment, legal language complexity, and domain-specific adaptation. Through empirical and qualitative analysis, we demonstrate how RL can be repurposed for high-stakes, long-document tasks in law. Our findings establish a foundation for future work on optimizing legal reasoning pipelines using reinforcement learning, with broader implications for building interpretable and adaptive legal AI systems.
Similar Papers
Part I: Tricks or Traps? A Deep Dive into RL for LLM Reasoning
Machine Learning (CS)
Makes smart computer programs learn better and faster.
ReST-RL: Achieving Accurate Code Reasoning of LLMs with Optimized Self-Training and Decoding
Artificial Intelligence
Makes computers better at solving coding problems.
A Technical Survey of Reinforcement Learning Techniques for Large Language Models
Artificial Intelligence
Teaches computers to follow instructions better.