Score: 0

ORPP: Self-Optimizing Role-playing Prompts to Enhance Language Model Capabilities

Published: June 3, 2025 | arXiv ID: 2506.02480v1

By: Yifan Duan , Yihong Tang , Kehai Chen and more

Potential Business Impact:

Makes AI better at tasks by giving it roles.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

High-quality prompts are crucial for eliciting outstanding performance from large language models (LLMs) on complex tasks. Existing research has explored model-driven strategies for prompt optimization. However, these methods often suffer from high computational overhead or require strong optimization capabilities from the model itself, which limits their broad applicability.To address these challenges, we propose ORPP (Optimized Role-Playing Prompt),a framework that enhances model performance by optimizing and generating role-playing prompts. The core idea of ORPP is to confine the prompt search space to role-playing scenarios, thereby fully activating the model's intrinsic capabilities through carefully crafted, high-quality role-playing prompts. Specifically, ORPP first performs iterative optimization on a small subset of training samples to generate high-quality role-playing prompts. Then, leveraging the model's few-shot learning capability, it transfers the optimization experience to efficiently generate suitable prompts for the remaining samples.Our experimental results show that ORPP not only matches but in most cases surpasses existing mainstream prompt optimization methods in terms of performance. Notably, ORPP demonstrates superior "plug-and-play" capability. In most cases, it can be integrated with various other prompt methods and further enhance their effectiveness.

Country of Origin
🇨🇳 China

Page Count
15 pages

Category
Computer Science:
Computation and Language