Robustness of LLM-enabled vehicle trajectory prediction under data security threats
By: Feilong Wang, Fuqiang Liu
Potential Business Impact:
Makes self-driving cars easily tricked by bad data.
The integration of large language models (LLMs) into automated driving systems has opened new possibilities for reasoning and decision-making by transforming complex driving contexts into language-understandable representations. Recent studies demonstrate that fine-tuned LLMs can accurately predict vehicle trajectories and lane-change intentions by gathering and transforming data from surrounding vehicles. However, the robustness of such LLM-based prediction models for safety-critical driving systems remains unexplored, despite the increasing concerns about the trustworthiness of LLMs. This study addresses this gap by conducting a systematic vulnerability analysis of LLM-enabled vehicle trajectory prediction. We propose a one-feature differential evolution attack that perturbs a single kinematic feature of surrounding vehicles within the LLM's input prompts under a black-box setting. Experiments on the highD dataset reveal that even minor, physically plausible perturbations can significantly disrupt model outputs, underscoring the susceptibility of LLM-based predictors to adversarial manipulation. Further analyses reveal a trade-off between accuracy and robustness, examine the failure mechanism, and explore potential mitigation solutions. The findings provide the very first insights into adversarial vulnerabilities of LLM-driven automated vehicle models in the context of vehicular interactions and highlight the need for robustness-oriented design in future LLM-based intelligent transportation systems.
Similar Papers
Evaluation of Large Language Models for Anomaly Detection in Autonomous Vehicles
Robotics
Helps self-driving cars spot tricky road problems.
Large Language Models and Their Applications in Roadway Safety and Mobility Enhancement: A Comprehensive Review
Artificial Intelligence
Helps cars understand traffic better for safer roads.
POLAR: Automating Cyber Threat Prioritization through LLM-Powered Assessment
Cryptography and Security
Makes computers better at finding online dangers.