Shift is Good: Mismatched Data Mixing Improves Test Performance
By: Marko Medvedev , Kaifeng Lyu , Zhiyuan Li and more
Potential Business Impact:
Makes AI learn better with mixed-up practice.
We consider training and testing on mixture distributions with different training and test proportions. We show that in many settings, and in some sense generically, distribution shift can be beneficial, and test performance can improve due to mismatched training proportions, even if the components are unrelated and with no transfer between components. In a variety of scenarios, we identify the optimal training proportions and the extent to which such distribution shift can be beneficial. We show how the same analysis applies also to a compositional setting with differing distribution of component "skills'' at training and test.
Similar Papers
Handling Out-of-Distribution Data: A Survey
Machine Learning (CS)
Helps computers learn from changing information.
When the Past Misleads: Rethinking Training Data Expansion Under Temporal Distribution Shifts
Computers and Society
More data doesn't always make computer predictions better.
When the Past Misleads: Rethinking Training Data Expansion Under Temporal Distribution Shifts
Computers and Society
More old data can hurt computer predictions.