How Instruction-Tuning Imparts Length Control: A Cross-Lingual Mechanistic Analysis
By: Elisabetta Rocchetti, Alfio Ferrara
Potential Business Impact:
Makes AI write exactly how long you want.
Adhering to explicit length constraints, such as generating text with a precise word count, remains a significant challenge for Large Language Models (LLMs). This study aims at investigating the differences between foundation models and their instruction-tuned counterparts, on length-controlled text generation in English and Italian. We analyze both performance and internal component contributions using Cumulative Weighted Attribution, a metric derived from Direct Logit Attribution. Our findings reveal that instruction-tuning substantially improves length control, primarily by specializing components in deeper model layers. Specifically, attention heads in later layers of IT models show increasingly positive contributions, particularly in English. In Italian, while attention contributions are more attenuated, final-layer MLPs exhibit a stronger positive role, suggesting a compensatory mechanism. These results indicate that instruction-tuning reconfigures later layers for task adherence, with component-level strategies potentially adapting to linguistic context.
Similar Papers
Towards Alignment-Centric Paradigm: A Survey of Instruction Tuning in Large Language Models
Computation and Language
Teaches AI to follow instructions better.
Plan-and-Write: Structure-Guided Length Control for LLMs without Model Retraining
Computation and Language
Makes AI write exactly the right amount of words.
In-context Learning vs. Instruction Tuning: The Case of Small and Multilingual Language Models
Computation and Language
Teaches computers to follow instructions better.