Score: 1

Deep Reinforcement Learning Xiangqi Player with Monte Carlo Tree Search

Published: June 18, 2025 | arXiv ID: 2506.15880v1

By: Berk Yilmaz, Junyu Hu, Jinsong Liu

Potential Business Impact:

Makes computers play Chinese chess better.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

This paper presents a Deep Reinforcement Learning (DRL) system for Xiangqi (Chinese Chess) that integrates neural networks with Monte Carlo Tree Search (MCTS) to enable strategic self-play and self-improvement. Addressing the underexplored complexity of Xiangqi, including its unique board layout, piece movement constraints, and victory conditions, our approach combines policy-value networks with MCTS to simulate move consequences and refine decision-making. By overcoming challenges such as Xiangqi's high branching factor and asymmetrical piece dynamics, our work advances AI capabilities in culturally significant strategy games while providing insights for adapting DRL-MCTS frameworks to domain-specific rule systems.

Country of Origin
🇺🇸 United States

Repos / Data Links

Page Count
8 pages

Category
Computer Science:
Artificial Intelligence