T-SAR: A Full-Stack Co-design for CPU-Only Ternary LLM Inference via In-Place SIMD ALU Reorganization
By: Hyunwoo Oh , KyungIn Nam , Rajat Bhattacharjya and more
Potential Business Impact:
Makes smart AI run faster on small devices.
Recent advances in LLMs have outpaced the computational and memory capacities of edge platforms that primarily employ CPUs, thereby challenging efficient and scalable deployment. While ternary quantization enables significant resource savings, existing CPU solutions rely heavily on memory-based lookup tables (LUTs) which limit scalability, and FPGA or GPU accelerators remain impractical for edge use. This paper presents T-SAR, the first framework to achieve scalable ternary LLM inference on CPUs by repurposing the SIMD register file for dynamic, in-register LUT generation with minimal hardware modifications. T-SAR eliminates memory bottlenecks and maximizes data-level parallelism, delivering 5.6-24.5x and 1.1-86.2x improvements in GEMM latency and GEMV throughput, respectively, with only 3.2% power and 1.4% area overheads in SIMD units. T-SAR achieves up to 2.5-4.9x the energy efficiency of an NVIDIA Jetson AGX Orin, establishing a practical approach for efficient LLM inference on edge platforms.
Similar Papers
TENET: An Efficient Sparsity-Aware LUT-Centric Architecture for Ternary LLM Inference On Edge
Hardware Architecture
Makes smart computer programs run much faster.
Fast and Compact Tsetlin Machine Inference on CPUs Using Instruction-Level Optimization
Machine Learning (CS)
Makes computers think faster using clever tricks.
SAIL: SRAM-Accelerated LLM Inference System with Lookup-Table-based GEMV
Hardware Architecture
Makes AI smarter on regular computers.