Score: 0

Deep Reinforcement Learning with anticipatory reward in LSTM for Collision Avoidance of Mobile Robots

Published: August 11, 2025 | arXiv ID: 2508.07941v1

By: Olivier Poulet, Frédéric Guinand, François Guérin

Potential Business Impact:

Robots predict crashes, then avoid them safely.

This article proposes a collision risk anticipation method based on short-term prediction of the agents position. A Long Short-Term Memory (LSTM) model, trained on past trajectories, is used to estimate the next position of each robot. This prediction allows us to define an anticipated collision risk by dynamically modulating the reward of a Deep Q-Learning Network (DQN) agent. The approach is tested in a constrained environment, where two robots move without communication or identifiers. Despite a limited sampling frequency (1 Hz), the results show a significant decrease of the collisions number and a stability improvement. The proposed method, which is computationally inexpensive, appears particularly attractive for implementation on embedded systems.

Country of Origin
🇫🇷 France

Page Count
6 pages

Category
Computer Science:
Artificial Intelligence