Score: 0

Real-Time Object Tracking with On-Device Deep Learning for Adaptive Beamforming in Dynamic Acoustic Environments

Published: November 24, 2025 | arXiv ID: 2511.19396v1

By: Jorge Ortigoso-Narro , Jose A. Belloch , Adrian Amor-Martin and more

Potential Business Impact:

Listens to and follows specific sounds in noisy places.

Business Areas:
Image Recognition Data and Analytics, Software

Advances in object tracking and acoustic beamforming are driving new capabilities in surveillance, human-computer interaction, and robotics. This work presents an embedded system that integrates deep learning-based tracking with beamforming to achieve precise sound source localization and directional audio capture in dynamic environments. The approach combines single-camera depth estimation and stereo vision to enable accurate 3D localization of moving objects. A planar concentric circular microphone array constructed with MEMS microphones provides a compact, energy-efficient platform supporting 2D beam steering across azimuth and elevation. Real-time tracking outputs continuously adapt the array's focus, synchronizing the acoustic response with the target's position. By uniting learned spatial awareness with dynamic steering, the system maintains robust performance in the presence of multiple or moving sources. Experimental evaluation demonstrates significant gains in signal-to-interference ratio, making the design well-suited for teleconferencing, smart home devices, and assistive technologies.

Country of Origin
🇪🇸 Spain

Page Count
11 pages

Category
Computer Science:
Sound