Efficient Multi-Model Orchestration for Self-Hosted Large Language Models
By: Bhanu Prakash Vangala, Tanu Malik
Potential Business Impact:
Makes big AI programs run cheaper and faster.
Self-hosting large language models (LLMs) is increasingly appealing for organizations seeking privacy, cost control, and customization. Yet deploying and maintaining in-house models poses challenges in GPU utilization, workload routing, and reliability. We introduce Pick and Spin, a practical framework that makes self-hosted LLM orchestration scalable and economical. Built on Kubernetes, it integrates a unified Helm-based deployment system, adaptive scale-to-zero automation, and a hybrid routing module that balances cost, latency, and accuracy using both keyword heuristics and a lightweight DistilBERT classifier. We evaluate four models, Llama-3 (90B), Gemma-3 (27B), Qwen-3 (235B), and DeepSeek-R1 (685B) across eight public benchmark datasets, with five inference strategies, and two routing variants encompassing 31,019 prompts and 163,720 inference runs. Pick and Spin achieves up to 21.6% higher success rates, 30% lower latency, and 33% lower GPU cost per query compared with static deployments of the same models.
Similar Papers
Scalable Engine and the Performance of Different LLM Models in a SLURM based HPC architecture
Distributed, Parallel, and Cluster Computing
Makes smart computer programs run much faster.
Small Language Models as Compiler Experts: Auto-Parallelization for Heterogeneous Systems
Machine Learning (CS)
Makes computers run programs much faster.
Evaluating Open-Source Large Language Models for Technical Telecom Question Answering
Networking and Internet Architecture
Tests AI for phone network questions.