Score: 2

Performant Unified GPU Kernels for Portable Singular Value Computation Across Hardware and Precision

Published: August 8, 2025 | arXiv ID: 2508.06339v1

By: Evelyne Ringoot , Rabab Alomairy , Valentin Churavy and more

BigTech Affiliations: Massachusetts Institute of Technology

Potential Business Impact:

Makes computers learn faster with better math.

This paper presents a portable, GPU-accelerated implementation of a QR-based singular value computation algorithm in Julia. The singular value ecomposition (SVD) is a fundamental numerical tool in scientific computing and machine learning, providing optimal low-rank matrix approximations. Its importance has increased even more in large-scale machine learning pipelines, including large language models (LLMs), where it enables low-rank adaptation (LoRA). The implemented algorithm is based on the classic two-stage QR reduction, consisting of successive matrix reduction to band form and bidiagonal form. Our implementation leverages Julia's multiple dispatch and metaprogramming capabilities, integrating with the GPUArrays and KernelAbstractions frameworks to provide a unified type and hardware-agnostic function. It supports diverse GPU architectures and data types, and is, to our knowledge, the first GPU-accelerated singular value implementation to support Apple Metal GPUs and half precision. Performance results on multiple GPU backends and data types demonstrate that portability does not require sacrificing performance: the unified function outperforms most linear algebra libraries (MAGMA, SLATE, rocSOLVER, oneMKL) for matrix sizes larger than 1024x1024, and achieves 80%-90% of the performance of cuSOLVER for large matrices.

Country of Origin
πŸ‡©πŸ‡ͺ πŸ‡ΊπŸ‡Έ United States, Germany

Page Count
12 pages

Category
Computer Science:
Distributed, Parallel, and Cluster Computing