ParaCodex: A Profiling-Guided Autonomous Coding Agent for Reliable Parallel Code Generation and Translation
By: Erel Kaplan , Tomer Bitan , Lian Ghrayeb and more
Potential Business Impact:
Makes supercomputers run much faster automatically.
Parallel programming is central to HPC and AI, but producing code that is correct and fast remains challenging, especially for OpenMP GPU offload, where data movement and tuning dominate. Autonomous coding agents can compile, test, and profile on target hardware, but outputs are brittle without domain scaffolding. We present ParaCodex, an HPC-engineer workflow that turns a Codex-based agent into an autonomous OpenMP GPU offload system using staged hotspot analysis, explicit data planning, correctness gating, and profiling-guided refinement. We evaluate translation from serial CPU kernels to OpenMP GPU offload kernels on HeCBench, Rodinia, and NAS. After excluding five kernels, ParaCodex succeeded on all 31 valid kernels. The generated kernels improved GPU time over reference OpenMP implementations in 25/31 cases, achieving geometric-mean speedups of 3x on HeCBench and 5x on Rodinia, and outperforming a zero-shot Codex baseline on all suites. We also evaluate CUDA to OpenMP offload translation on ParEval, where ParaCodex maintains high compilation and validation rates in code-only and end-to-end settings.
Similar Papers
VibeCodeHPC: An Agent-Based Iterative Prompting Auto-Tuner for HPC Code Generation Using LLMs
Software Engineering
Makes computer programs run much faster on new hardware.
UniPar: A Unified LLM-Based Framework for Parallel and Accelerated Code Translation in HPC
Distributed, Parallel, and Cluster Computing
Helps computers rewrite code for faster processing.
Theoretical Foundations of GPU-Native Compilation for Rapid Code Iteration
Distributed, Parallel, and Cluster Computing
Makes AI code run much, much faster.