Score: 2

Training-Free Adaptation of New-Generation LLMs using Legacy Clinical Models

Published: January 6, 2026 | arXiv ID: 2601.03423v1

By: Sasha Ronaghi , Chloe Stanwyck , Asad Aali and more

BigTech Affiliations: Stanford University

Potential Business Impact:

Makes old medical AI work with new AI.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Adapting language models to the clinical domain through continued pretraining and fine-tuning requires costly retraining for each new model generation. We propose Cross-Architecture Proxy Tuning (CAPT), a model-ensembling approach that enables training-free adaptation of state-of-the-art general-domain models using existing clinical models. CAPT supports models with disjoint vocabularies, leveraging contrastive decoding to selectively inject clinically relevant signals while preserving the general-domain model's reasoning and fluency. On six clinical classification and text-generation tasks, CAPT with a new-generation general-domain model and an older-generation clinical model consistently outperforms both models individually and state-of-the-art ensembling approaches (average +17.6% over UniTE, +41.4% over proxy tuning across tasks). Through token-level analysis and physician case studies, we demonstrate that CAPT amplifies clinically actionable language, reduces context errors, and increases clinical specificity.

Country of Origin
🇺🇸 United States

Page Count
29 pages

Category
Computer Science:
Computation and Language