Multi-Objective Reinforcement Learning for Efficient Tactical Decision Making for Trucks in Highway Traffic
By: Deepthi Pathare, Leo Laine, Morteza Haghir Chehreghani
Potential Business Impact:
Trucks learn to drive safely, fast, and cheap.
Balancing safety, efficiency, and operational costs in highway driving poses a challenging decision-making problem for heavy-duty vehicles. A central difficulty is that conventional scalar reward formulations, obtained by aggregating these competing objectives, often obscure the structure of their trade-offs. We present a Proximal Policy Optimization based multi-objective reinforcement learning framework that learns a continuous set of policies explicitly representing these trade-offs and evaluates it on a scalable simulation platform for tactical decision making in trucks. The proposed approach learns a continuous set of Pareto-optimal policies that capture the trade-offs among three conflicting objectives: safety, quantified in terms of collisions and successful completion; energy efficiency and time efficiency, quantified using energy cost and driver cost, respectively. The resulting Pareto frontier is smooth and interpretable, enabling flexibility in choosing driving behavior along different conflicting objectives. This framework allows seamless transitions between different driving policies without retraining, yielding a robust and adaptive decision-making strategy for autonomous trucking applications.
Similar Papers
Multi-Objective Reinforcement Learning for Large-Scale Mixed Traffic Control
Multiagent Systems
Makes traffic lights fairer and safer for all cars.
End-to-end Deep Reinforcement Learning for Stochastic Multi-objective Optimization in C-VRPTW
Machine Learning (CS)
Helps delivery trucks plan faster, safer routes.
Balancing Progress and Safety: A Novel Risk-Aware Objective for RL in Autonomous Driving
Robotics
Makes self-driving cars safer by learning better driving rules.