Computational Economics in Large Language Models: Exploring Model Behavior and Incentive Design under Resource Constraints
By: Sandeep Reddy , Kabir Khan , Rohit Patil and more
Potential Business Impact:
Makes smart computer programs use less power.
Large language models (LLMs) are limited by substantial computational cost. We introduce a "computational economics" framework that treats an LLM as an internal economy of resource-constrained agents (attention heads and neuron blocks) that must allocate scarce computation to maximize task utility. First, we show empirically that when computation is scarce, standard LLMs reallocate attention toward high-value tokens while preserving accuracy. Building on this observation, we propose an incentive-driven training paradigm that augments the task loss with a differentiable computation cost term, encouraging sparse and efficient activations. On GLUE (MNLI, STS-B, CoLA) and WikiText-103, the method yields a family of models that trace a Pareto frontier and consistently dominate post-hoc pruning; for a similar accuracy we obtain roughly a forty percent reduction in FLOPS and lower latency, together with more interpretable attention patterns. These results indicate that economic principles offer a principled route to designing efficient, adaptive, and more transparent LLMs under strict resource constraints.
Similar Papers
Efficient Strategy for Improving Large Language Model (LLM) Capabilities
Computation and Language
Makes smart computer programs run faster with less power.
Evaluating Large Language Models for Workload Mapping and Scheduling in Heterogeneous HPC Systems
Distributed, Parallel, and Cluster Computing
Lets computers solve hard scheduling puzzles from words.
Constraint-Compliant Network Optimization through Large Language Models
Networking and Internet Architecture
Makes computer networks follow rules perfectly.