Training-Free Adaptation of New-Generation LLMs using Legacy Clinical Models
By: Sasha Ronaghi , Chloe Stanwyck , Asad Aali and more
Potential Business Impact:
Makes old medical AI work with new AI.
Adapting language models to the clinical domain through continued pretraining and fine-tuning requires costly retraining for each new model generation. We propose Cross-Architecture Proxy Tuning (CAPT), a model-ensembling approach that enables training-free adaptation of state-of-the-art general-domain models using existing clinical models. CAPT supports models with disjoint vocabularies, leveraging contrastive decoding to selectively inject clinically relevant signals while preserving the general-domain model's reasoning and fluency. On six clinical classification and text-generation tasks, CAPT with a new-generation general-domain model and an older-generation clinical model consistently outperforms both models individually and state-of-the-art ensembling approaches (average +17.6% over UniTE, +41.4% over proxy tuning across tasks). Through token-level analysis and physician case studies, we demonstrate that CAPT amplifies clinically actionable language, reduces context errors, and increases clinical specificity.
Similar Papers
Ultra-Light Test-Time Adaptation for Vision--Language Models
CV and Pattern Recognition
Makes AI better at seeing new things.
Mitigating Spurious Correlations in LLMs via Causality-Aware Post-Training
Machine Learning (CS)
Teaches computers to think better, not just copy.
Generalization of Medical Large Language Models through Cross-Domain Weak Supervision
Computation and Language
Makes AI understand and answer medical questions better.