Score: 0

Quantum State Preparation via Large-Language-Model-Driven Evolution

Published: May 9, 2025 | arXiv ID: 2505.06347v1

By: Qing-Hong Cao , Zong-Yue Hou , Ying-Ying Li and more

Potential Business Impact:

Finds better ways to build quantum computers.

Business Areas:
Quantum Computing Science and Engineering

We propose an automated framework for quantum circuit design by integrating large-language models (LLMs) with evolutionary optimization to overcome the rigidity, scalability limitations, and expert dependence of traditional ones in variational quantum algorithms. Our approach (FunSearch) autonomously discovers hardware-efficient ans\"atze with new features of scalability and system-size-independent number of variational parameters entirely from scratch. Demonstrations on the Ising and XY spin chains with n = 9 qubits yield circuits containing 4 parameters, achieving near-exact energy extrapolation across system sizes. Implementations on quantum hardware (Zuchongzhi chip) validate practicality, where two-qubit quantum gate noises can be effectively mitigated via zero-noise extrapolations for a spin chain system as large as 20 sites. This framework bridges algorithmic design and experimental constraints, complementing contemporary quantum architecture search frameworks to advance scalable quantum simulations.

Page Count
10 pages

Category
Physics:
Quantum Physics