Combining Serverless and High-Performance Computing Paradigms to support ML Data-Intensive Applications
By: Mills Staylor , Arup Kumar Sarker , Gregor von Laszewski and more
Potential Business Impact:
Lets computers process big data faster without big machines.
Data is found everywhere, from health and human infrastructure to the surge of sensors and the proliferation of internet-connected devices. To meet this challenge, the data engineering field has expanded significantly in recent years in both research and industry. Traditionally, data engineering, Machine Learning, and AI workloads have been run on large clusters within data center environments, requiring substantial investment in hardware and maintenance. With the rise of the public cloud, it is now possible to run large applications across nodes without owning or maintaining hardware. Serverless functions such as AWS Lambda provide horizontal scaling and precise billing without the hassle of managing traditional cloud infrastructure. However, when processing large datasets, users often rely on external storage options that are significantly slower than direct communication typical of HPC clusters. We introduce Cylon, a high-performance distributed data frame solution that has shown promising results for data processing using Python. We describe how we took inspiration from the FMI library and designed a serverless communicator to tackle communication and performance issues associated with serverless functions. With our design, we demonstrate that the performance of AWS Lambda falls below one percent of strong scaling experiments compared to serverful AWS (EC2) and HPCs based on implementing direct communication via NAT Traversal TCP Hole Punching.
Similar Papers
High-Dimensional Data Processing: Benchmarking Machine Learning and Deep Learning Architectures in Local and Distributed Environments
Distributed, Parallel, and Cluster Computing
Teaches computers to learn from lots of information.
Towards Energy-Efficient Serverless Computing with Hardware Isolation
Distributed, Parallel, and Cluster Computing
Saves energy by giving each task its own tiny computer.
Automated Dynamic AI Inference Scaling on HPC-Infrastructure: Integrating Kubernetes, Slurm and vLLM
Distributed, Parallel, and Cluster Computing
Makes supercomputers run AI faster for many people.