CrashAgent: Crash Scenario Generation via Multi-modal Reasoning
By: Miao Li , Wenhao Ding , Haohong Lin and more
Potential Business Impact:
Makes self-driving cars learn from accidents.
Training and evaluating autonomous driving algorithms requires a diverse range of scenarios. However, most available datasets predominantly consist of normal driving behaviors demonstrated by human drivers, resulting in a limited number of safety-critical cases. This imbalance, often referred to as a long-tail distribution, restricts the ability of driving algorithms to learn from crucial scenarios involving risk or failure, scenarios that are essential for humans to develop driving skills efficiently. To generate such scenarios, we utilize Multi-modal Large Language Models to convert crash reports of accidents into a structured scenario format, which can be directly executed within simulations. Specifically, we introduce CrashAgent, a multi-agent framework designed to interpret multi-modal real-world traffic crash reports for the generation of both road layouts and the behaviors of the ego vehicle and surrounding traffic participants. We comprehensively evaluate the generated crash scenarios from multiple perspectives, including the accuracy of layout reconstruction, collision rate, and diversity. The resulting high-quality and large-scale crash dataset will be publicly available to support the development of safe driving algorithms in handling safety-critical situations.
Similar Papers
From Accidents to Insights: Leveraging Multimodal Data for Scenario-Driven ADS Testing
Software Engineering
Finds car crashes to make self-driving cars safer.
AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework
Robotics
Makes self-driving cars safer by creating tricky test situations.
Research on Driving Scenario Technology Based on Multimodal Large Lauguage Model Optimization
CV and Pattern Recognition
Helps self-driving cars see and react better.