From Teacher to Student: Tracking Memorization Through Model Distillation
By: Simardeep Singh
Potential Business Impact:
Makes AI models safer by reducing memorization.
Large language models (LLMs) are known to memorize parts of their training data, raising important concerns around privacy and security. While previous research has focused on studying memorization in pre-trained models, much less is known about how knowledge distillation (KD) affects memorization.In this study, we explore how different KD methods influence the memorization of fine-tuned task data when a large teacher model is distilled into smaller student variants.This study demonstrates that distilling a larger teacher model, fine-tuned on a dataset, into a smaller variant not only lowers computational costs and model size but also significantly reduces the memorization risks compared to standard fine-tuning approaches.
Similar Papers
Membership and Memorization in LLM Knowledge Distillation
Machine Learning (CS)
Keeps private info safe when small AI learns from big AI.
Knowledge Distillation of Domain-adapted LLMs for Question-Answering in Telecom
Computation and Language
Makes big AI models smaller and smarter.
Feature Alignment and Representation Transfer in Knowledge Distillation for Large Language Models
Computation and Language
Makes smart computer programs smaller and faster.