ResBench: Benchmarking LLM-Generated FPGA Designs with Resource Awareness
By: Ce Guo, Tong Zhao
Potential Business Impact:
Helps computers design better chips faster.
Field-Programmable Gate Arrays (FPGAs) are widely used in modern hardware design, yet writing Hardware Description Language (HDL) code for FPGA implementation remains a complex and time-consuming task. Large Language Models (LLMs) have emerged as a promising tool for HDL generation, but existing benchmarks for LLM-based code generation primarily focus on functional correctness while overlooking hardware resource usage. Furthermore, current benchmarks offer limited diversity and do not fully represent the wide range of real-world FPGA applications. To address these shortcomings, we introduce ResBench, the first resource-focused benchmark explicitly designed to distinguish between resource-optimized and inefficient LLM-generated HDL code. ResBench consists of 56 problems across 12 categories, covering applications from finite state machines to financial computing. Our open-source evaluation framework automatically tests LLMs by generating Verilog code, verifying correctness, and measuring resource usage. The experiments, which primarily analyze Lookup Table (LUT) usage, reveal significant differences among LLMs, demonstrating ResBench's capability to identify models that generate more resource-optimized FPGA designs.
Similar Papers
ForgeBench: A Machine Learning Benchmark Suite and Auto-Generation Framework for Next-Generation HLS Tools
Hardware Architecture
Makes computer chips for AI design faster.
ProtocolLLM: RTL Benchmark for SystemVerilog Generation of Communication Protocols
Hardware Architecture
Helps computers design computer chips faster.
QuanBench: Benchmarking Quantum Code Generation with Large Language Models
Software Engineering
Tests how well computers write quantum computer code.