Mo' Memory, Mo' Problems: Stream-Native Machine Unlearning
By: Kennon Stewart
Potential Business Impact:
Lets computers forget old info faster and cheaper.
Machine unlearning work assumes a static, i.i.d training environment that doesn't truly exist. Modern ML pipelines need to learn, unlearn, and predict continuously on production streams of data. We translate the notion of the batch unlearning scenario to the online setting using notions of regret, sample complexity, and deletion capacity. We further tighten regret bounds to a logarithmic $\mathcal{O}(\ln{T})$, a first for a machine unlearning algorithm. And we swap out an expensive Hessian inversion with online variant of L-BFGS optimization, removing a memory footprint that scales linearly with time. Such changes extend the lifespan of an ML model before expensive retraining, making for a more efficient unlearning process.
Similar Papers
Mo' Memory, Mo' Problems: Stream-Native Machine Unlearning
Machine Learning (Stat)
Lets computers forget old info without retraining.
Machine Unlearning for Streaming Forgetting
Machine Learning (CS)
Removes data from AI without retraining it.
When to Forget? Complexity Trade-offs in Machine Unlearning
Machine Learning (Stat)
Removes old data from AI without full rebuild.