Cost-Performance Analysis: A Comparative Study of CPU-Based Serverless and GPU-Based Training Architectures
By: Amine Barrak, Fabio Petrillo, Fehmi Jaafar
Potential Business Impact:
Trains computer brains faster and cheaper.
The field of distributed machine learning (ML) faces increasing demands for scalable and cost-effective training solutions, particularly in the context of large, complex models. Serverless computing has emerged as a promising paradigm to address these challenges by offering dynamic scalability and resource-efficient execution. Building upon our previous work, which introduced the Serverless Peer Integrated for Robust Training (SPIRT) architecture, this paper presents a comparative analysis of several serverless distributed ML architectures. We examine SPIRT alongside established architectures like ScatterReduce, AllReduce, and MLLess, focusing on key metrics such as training time efficiency, cost-effectiveness, communication overhead, and fault tolerance capabilities. Our findings reveal that SPIRT provides significant improvements in reducing training times and communication overhead through strategies such as parallel batch processing and in-database operations facilitated by RedisAI. However, traditional architectures exhibit scalability challenges and varying degrees of vulnerability to faults and adversarial attacks. The cost analysis underscores the long-term economic benefits of SPIRT despite its higher initial setup costs. This study not only highlights the strengths and limitations of current serverless ML architectures but also sets the stage for future research aimed at developing new models that combine the most effective features of existing systems.
Similar Papers
Serverless GPU Architecture for Enterprise HR Analytics: A Production-Scale BDaaS Implementation
Distributed, Parallel, and Cluster Computing
Makes computer analysis faster, cheaper, and trustworthy.
Towards Energy-Efficient Serverless Computing with Hardware Isolation
Distributed, Parallel, and Cluster Computing
Saves energy by giving each task its own tiny computer.
High-Dimensional Data Processing: Benchmarking Machine Learning and Deep Learning Architectures in Local and Distributed Environments
Distributed, Parallel, and Cluster Computing
Teaches computers to learn from lots of information.