An Explorative Study on Distributed Computing Techniques in Training and Inference of Large Language Models
By: Sheikh Azizul Hakim, Saem Hasan
Potential Business Impact:
Lets big AI run on normal computers.
Large language models (LLM) are advanced AI systems trained on extensive textual data, leveraging deep learning techniques to understand and generate human-like language. Today's LLMs with billions of parameters are so huge that hardly any single computing node can train, fine-tune, or infer from them. Therefore, several distributed computing techniques are being introduced in the literature to properly utilize LLMs. We have explored the application of distributed computing techniques in LLMs from two angles. \begin{itemize} \item We study the techniques that democratize the LLM, that is, how large models can be run on consumer-grade computers. Here, we also implement a novel metaheuristics-based modification to an existing system. \item We perform a comparative study on three state-of-the-art LLM serving techniques. \end{itemize}
Similar Papers
Scaling Performance of Large Language Model Pretraining
Distributed, Parallel, and Cluster Computing
Teaches computers to learn faster with less power.
Evaluating Large Language Models for Workload Mapping and Scheduling in Heterogeneous HPC Systems
Distributed, Parallel, and Cluster Computing
Lets computers solve hard scheduling puzzles from words.
Collaborative Inference and Learning between Edge SLMs and Cloud LLMs: A Survey of Algorithms, Execution, and Open Challenges
Distributed, Parallel, and Cluster Computing
Smart computers work together for faster, private AI.