Joint Optimization of Offloading, Batching and DVFS for Multiuser Co-Inference
By: Yaodan Xu, Sheng Zhou, Zhisheng Niu
Potential Business Impact:
Saves phone battery by sharing tasks with a server.
With the growing integration of artificial intelligence in mobile applications, a substantial number of deep neural network (DNN) inference requests are generated daily by mobile devices. Serving these requests presents significant challenges due to limited device resources and strict latency requirements. Therefore, edge-device co-inference has emerged as an effective paradigm to address these issues. In this study, we focus on a scenario where multiple mobile devices offload inference tasks to an edge server equipped with a graphics processing unit (GPU). For finer control over offloading and scheduling, inference tasks are partitioned into smaller sub-tasks. Additionally, GPU batch processing is employed to boost throughput and improve energy efficiency. This work investigates the problem of minimizing total energy consumption while meeting hard latency constraints. We propose a low-complexity Joint DVFS, Offloading, and Batching strategy (J-DOB) to solve this problem. The effectiveness of the proposed algorithm is validated through extensive experiments across varying user numbers and deadline constraints. Results show that J-DOB can reduce energy consumption by up to 51.30% and 45.27% under identical and different deadlines, respectively, compared to local computing.
Similar Papers
Batching-Aware Joint Model Onloading and Offloading for Hierarchical Multi-Task Inference
Machine Learning (CS)
Lets phones do many smart jobs at once.
Joint Memory Frequency and Computing Frequency Scaling for Energy-efficient DNN Inference
Machine Learning (CS)
Saves phone power by adjusting chip speeds.
E4: Energy-Efficient DNN Inference for Edge Video Analytics Via Early-Exit and DVFS
CV and Pattern Recognition
Saves phone battery by making smart video analysis faster.