Multimodal Remote Inference
By: Keyuan Zhang, Yin Sun, Bo Ji
Potential Business Impact:
Smartly picks sensor data to improve AI guesses.
We consider a remote inference system with multiple modalities, where a multimodal machine learning (ML) model performs real-time inference using features collected from remote sensors. As sensor observations may change dynamically over time, fresh features are critical for inference tasks. However, timely delivering features from all modalities is often infeasible due to limited network resources. To this end, we study a two-modality scheduling problem to minimize the ML model's inference error, which is expressed as a penalty function of AoI for both modalities. We develop an index-based threshold policy and prove its optimality. Specifically, the scheduler switches modalities when the current modality's index function exceeds a threshold. We show that the two modalities share the same threshold, and both the index functions and the threshold can be computed efficiently. The optimality of our policy holds for (i) general AoI functions that are \emph{non-monotonic} and \emph{non-additive} and (ii) \emph{heterogeneous} transmission times. Numerical results show that our policy reduces inference error by up to 55% compared to round-robin and uniform random policies, which are oblivious to the AoI-based inference error function. Our results shed light on how to improve remote inference accuracy by optimizing task-oriented AoI functions.
Similar Papers
Multimodal Remote Inference
Machine Learning (CS)
Smartly picks sensor data to improve predictions.
Task-oriented Age of Information for Remote Inference with Hybrid Language Models
Information Theory
Smart AI chooses fast or smart models for speed.
Real-Time Inference for Distributed Multimodal Systems under Communication Delay Uncertainty
Machine Learning (CS)
Lets computers understand events with changing delays.