Test-time adaptation (TTA) methods, which generally rely on the model's predictions (e.g., entropy minimization) to adapt the source pretrained model to the unlabeled target domain, suffer from noisy signals originating from 1) incorrect or 2) open-set predictions. Long-term stable adaptation is hampered by such noisy signals, so training models without such error accumulation is crucial for practical TTA. To address these issues, including open-set TTA, we propose a simple yet effective sample selection method inspired by the following crucial empirical finding. While entropy minimization compels the model to increase the probability of its predicted label (i.e., confidence values), we found that noisy samples rather show decreased confide...
Experiencing domain shifts during test-time is nearly inevitable in practice and likely results in a...
We introduce a framework for calibrating machine learning models so that their predictions satisfy e...
We consider various versions of adaptive Gibbs and Metropolis- within-Gibbs samplers, which update ...
Test-time adaptation (TTA) seeks to tackle potential distribution shifts between training and testin...
Fully Test-Time Adaptation (TTA), which aims at adapting models to data drifts, has recently attract...
While the design of blind image quality assessment (IQA) algorithms has improved significantly, the ...
Test-time adaptation harnesses test inputs to improve the accuracy of a model trained on source data...
Models should have the ability to adapt to unseen data during test-time to avoid performance drop ca...
The model selection procedure is usually a single-criterion decision making in which we select the m...
In survey methodology, inverse probability weighted (Horvitz-Thompson) estimation has become an indi...
Test-Time Adaptation aims to adapt source domain model to testing data at inference stage with succe...
We highlight the utility of a certain property of model training: instead of drawing training data f...
We study the prevalent problem when a test distribution differs from the training distribution. We c...
Model overconfidence and poor calibration are common in machine learning and difficult to account fo...
Current supervised visual detectors, though impressive within their training distribution, often fai...
Experiencing domain shifts during test-time is nearly inevitable in practice and likely results in a...
We introduce a framework for calibrating machine learning models so that their predictions satisfy e...
We consider various versions of adaptive Gibbs and Metropolis- within-Gibbs samplers, which update ...
Test-time adaptation (TTA) seeks to tackle potential distribution shifts between training and testin...
Fully Test-Time Adaptation (TTA), which aims at adapting models to data drifts, has recently attract...
While the design of blind image quality assessment (IQA) algorithms has improved significantly, the ...
Test-time adaptation harnesses test inputs to improve the accuracy of a model trained on source data...
Models should have the ability to adapt to unseen data during test-time to avoid performance drop ca...
The model selection procedure is usually a single-criterion decision making in which we select the m...
In survey methodology, inverse probability weighted (Horvitz-Thompson) estimation has become an indi...
Test-Time Adaptation aims to adapt source domain model to testing data at inference stage with succe...
We highlight the utility of a certain property of model training: instead of drawing training data f...
We study the prevalent problem when a test distribution differs from the training distribution. We c...
Model overconfidence and poor calibration are common in machine learning and difficult to account fo...
Current supervised visual detectors, though impressive within their training distribution, often fai...
Experiencing domain shifts during test-time is nearly inevitable in practice and likely results in a...
We introduce a framework for calibrating machine learning models so that their predictions satisfy e...
We consider various versions of adaptive Gibbs and Metropolis- within-Gibbs samplers, which update ...