Real-Time Object Tracking with On-Device Deep Learning for Adaptive Beamforming in Dynamic Acoustic Environments
By: Jorge Ortigoso-Narro , Jose A. Belloch , Adrian Amor-Martin and more
Potential Business Impact:
Listens to and follows specific sounds in noisy places.
Advances in object tracking and acoustic beamforming are driving new capabilities in surveillance, human-computer interaction, and robotics. This work presents an embedded system that integrates deep learning-based tracking with beamforming to achieve precise sound source localization and directional audio capture in dynamic environments. The approach combines single-camera depth estimation and stereo vision to enable accurate 3D localization of moving objects. A planar concentric circular microphone array constructed with MEMS microphones provides a compact, energy-efficient platform supporting 2D beam steering across azimuth and elevation. Real-time tracking outputs continuously adapt the array's focus, synchronizing the acoustic response with the target's position. By uniting learned spatial awareness with dynamic steering, the system maintains robust performance in the presence of multiple or moving sources. Experimental evaluation demonstrates significant gains in signal-to-interference ratio, making the design well-suited for teleconferencing, smart home devices, and assistive technologies.
Similar Papers
Beamformed 360° Sound Maps: U-Net-Driven Acoustic Source Segmentation and Localization
Audio and Speech Processing
Pinpoints sounds from all directions accurately
Multimodal Real-Time Anomaly Detection and Industrial Applications
Sound
Watches and listens to spot problems.
Frequency-Invariant Beamforming in Elevation and Azimuth via Autograd and Concentric Circular Microphone Arrays
Sound
Focuses sound better in all directions.