scispace - formally typeset
Proceedings ArticleDOI

Continual Test-Time Domain Adaptation

Qin Wang, +3 more
- pp 7191-7201
Reads0
Chats0
TLDR
The proposed CoTTA proposes to stochastically restore a small part of the neurons to the source pre-trained weights during each iteration to help preserve source knowledge in the longterm and demonstrates the effectiveness of the approach on four classification tasks and a segmentation task for continual test-time adaptation.
Abstract
Test-time domain adaptation aims to adapt a source pre-trained model to a target domain without using any source data. Existing works mainly consider the case where the target domain is static. However, real-world machine perception systems are running in non-stationary and continually changing environments where the target domain distribution can change over time. Existing methods, which are mostly based on self-training and entropy regularization, can suffer from these non-stationary environments. Due to the distribution shift over time in the target domain, pseudo-labels become unreliable. The noisy pseudo-labels can further lead to error accumulation and catastrophic forgetting. To tackle these issues, we propose a continual test-time adaptation approach (CoTTA) which comprises two parts. Firstly, we propose to reduce the error accumulation by using weight-averaged and augmentation-averaged predictions which are often more accurate. On the other hand, to avoid catastrophic forgetting, we propose to stochastically restore a small part of the neurons to the source pre-trained weights during each iteration to help preserve source knowledge in the longterm. The proposed method enables the longterm adaptation for all parameters in the network. CoTTA is easy to implement and can be readily incorporated in off-the-shelf pre-trained models. We demonstrate the effectiveness of our approach on four classification tasks and a segmentation task for continual test-time adaptation, on which we outperform existing methods. Our code is available at https://gin.ee/cotta.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Efficient Test-Time Model Adaptation without Forgetting

TL;DR: An active sample selection criterion is proposed to identify reliable and non-redundant samples, on which the model is updated to minimize the entropy loss for test-time adaptation, and a Fisher regularizer is introduced to constrain important model parameters from drastic changes.
Proceedings ArticleDOI

HRDA: Context-Aware High-Resolution Domain-Adaptive Semantic Segmentation

TL;DR: HRDA is proposed, a multi-resolution training approach for UDA that combines the strengths of small high-resolution crops to preserve fine segmentation details and large low- resolution crops to capture long-range context dependencies with a learned scale attention, while maintaining a manageable GPU memory footprint.
Proceedings ArticleDOI

Test-Time Adaptation via Conjugate Pseudo-labels

TL;DR: This paper analyzes test-time adaptation through the lens of the training losses’s convex conjugate function, and shows that under natural conditions, this (unsupervised) conjugates can be viewed as a good local approximation to the original supervised loss and indeed, it recovers the “best” losses found by meta-learning.
Proceedings ArticleDOI

Towards Stable Test-Time Adaptation in Dynamic Wild World

TL;DR: This paper proposed a sharpness-aware and reliable entropy minimization method, called SAR, for further stabilizing TTA from two aspects: 1) remove partial noisy samples with large gradients, and 2) encourage model weights to go to a flat minimum so that the model is robust to the remaining noisy samples.
Proceedings Article

NOTE: Robust Continual Test-time Adaptation Against Temporal Correlation

TL;DR: This work presents a new test-time adaptation scheme that is robust against non-i.i.d. test data streams, and demonstrates that the proposed robust TTA not only outperforms state-of-the-art TTA algorithms in the non- i.i-i-d.
References
More filters
Journal ArticleDOI

ImageNet classification with deep convolutional neural networks

TL;DR: A large, deep convolutional neural network was trained to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes and employed a recently developed regularization method called "dropout" that proved to be very effective.
Journal ArticleDOI

A survey on Image Data Augmentation for Deep Learning

TL;DR: This survey will present existing methods for Data Augmentation, promising developments, and meta-level decisions for implementing DataAugmentation, a data-space solution to the problem of limited data.
Posted Content

The Cityscapes Dataset for Semantic Urban Scene Understanding

TL;DR: Cityscapes as discussed by the authors is a large-scale dataset for semantic urban scene understanding, consisting of 5000 images with high quality pixel-level annotations and 200,000 additional images with coarse annotations.
Posted Content

Learning Transferable Features with Deep Adaptation Networks

TL;DR: A new Deep Adaptation Network (DAN) architecture is proposed, which generalizes deep convolutional neural network to the domain adaptation scenario and can learn transferable features with statistical guarantees, and can scale linearly by unbiased estimate of kernel embedding.
Posted Content

Unsupervised Domain Adaptation by Backpropagation

TL;DR: In this paper, a gradient reversal layer is proposed to promote the emergence of deep features that are discriminative for the main learning task on the source domain and invariant with respect to the shift between the domains.