Test-Time Training with Self-Supervision for Generalization under Distribution Shifts
Abstract
Test-Time Training enhances model predictions by adjusting parameters using self-supervised learning on individual test samples, improving robustness to distribution shifts in image classification tasks.
In this paper, we propose Test-Time Training, a general approach for improving the performance of predictive models when training and test data come from different distributions. We turn a single unlabeled test sample into a self-supervised learning problem, on which we update the model parameters before making a prediction. This also extends naturally to data in an online stream. Our simple approach leads to improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper