AdaMatch: A Unified Approach to Semi-Supervised Learning and Domain Adaptation
@article{Berthelot2021AdaMatchAU, title={AdaMatch: A Unified Approach to Semi-Supervised Learning and Domain Adaptation}, author={David Berthelot and Rebecca Roelofs and Kihyuk Sohn and Nicholas Carlini and Alexey Kurakin}, journal={ArXiv}, year={2021}, volume={abs/2106.04732} }
We extend semi-supervised learning to the problem of domain adaptation to learn significantly higher-accuracy models that train on one data distribution and test on a different one. With the goal of generality, we introduce AdaMatch, a method that unifies the tasks of unsupervised domain adaptation (UDA), semi-supervised learning (SSL), and semi-supervised domain adaptation (SSDA). In an extensive experimental study, we compare its behavior with respective state-of-the-art techniques from SSL…
Figures and Tables from this paper
table 1 figure 1 table 2 figure 2 table 3 figure 3 table 4 figure 4 figure 5 table 5 table 6 table 7 table 8 table 9 table 10 table 11 table 12 table 13 table 14 table 15 table 16 table 17 table 18 table 19 table 20 table 21 table 22 table 23 table 24 table 25 table 26 table 27 table 28 table 29 table 30 table 31 table 32 table 33 table 34 table 35 table 36 table 37 table 38 table 39 table 40 table 41 table 42 table 43 table 44 table 45 table 46 table 47 table 48 table 49 table 50 table 51 table 52 table 53 table 54 table 55 table 56 table 57 table 58 table 59 table 60 table 61 table 62 table 63 table 64 table 65 table 66 table 67 table 68 table 69 table 70 table 71 table 72 table 73 table 74 table 75 table 76 table 77 table 78 table 79 table 80 table 81 table 82 table 83 table 84 table 85 table 86 table 87 table 89 table 90 table 92 table 93 table 94 table 97 table 98 table 99 table 100 table 101 table 102 table 106 table 110 table 111 table 112 table 113 table 114 table 117 table 118 table 120 table 121 table 122 table 123 table 124 table 126 table 127 table 128 table 129 table 130 table 131 table 132 table 133 table 134 table 135 table 136 table 137 table 138 table 139 table 140 table 141
17 Citations
Pick up the PACE: Fast and Simple Domain Adaptation via Ensemble Pseudo-Labeling
- Computer ScienceArXiv
- 2022
A fast and simple DA method consisting of three stages: domain alignment by covariance matching, pseudo-labeling, and ensembling that exceeds previous state-of-the-art methods on most benchmark adaptation tasks without training a neural network.
How Does Contrastive Pre-training Connect Disparate Domains?
- Computer Science
- 2021
A new measure of connectivity —the relative connection strengths between same and different classes across domains— is proposed that governs the success of contrastive pre-training for domain adaptation in a simple example and strongly correlates with the results on benchmark datasets.
Not All Labels Are Equal: Rationalizing The Labeling Costs for Training Object Detection
- Computer Science
- 2021
This work proposes a unified framework for active learning, that considers both the uncertainty and the robustness of the detector, ensuring that the network performs well in all classes, and leverages auto-labeling to suppress a potential distribution drift while boosting the performance of the model.
Holistic Semi-Supervised Approaches for EEG Representation Learning
- Computer ScienceICASSP
- 2022
Three state-of-the-art holistic semi-supervised approaches, namely MixMatch, Fix-Match, and AdaMatch are adapted, as well as five classical semi- supervised methods for EEG learning, which show strong results even when only 1 labeled sample is used per class.
Extending the WILDS Benchmark for Unsupervised Adaptation
- Computer ScienceArXiv
- 2021
The Wilds 2.0 update is presented, which extends 8 of the 10 datasets in the Wilds benchmark of distribution shifts to include curated unlabeled data that would be realistically obtainable in deployment, and systematically benchmark state-of-the-art methods that leverage unlabeling data, including domain-invariant, self-training, and self-supervised methods.
FreeMatch: Self-adaptive Thresholding for Semi-supervised Learning
- Computer ScienceArXiv
- 2022
This work proposes FreeMatch to be used and introduces a self-adaptive class fairness regularization penalty that encourages the model to produce diverse predictions during the early stages of training and indicates the superiority of FreeMatch especially when the labeled data are extremely rare.
Self-Supervised Contrastive Pre-Training For Time Series via Time-Frequency Consistency
- Computer ScienceArXiv
- 2022
A decomposable pre-training model, where the self-supervised signal is provided by the distance between time and frequency components, each individually trained by contrastive estimation is proposed, motivated by TF-C.
Sample Efficiency of Data Augmentation Consistency Regularization
- Computer ScienceArXiv
- 2022
A simple and novel analysis for linear regression with label invariant augmentations is presented, demonstrating that data augmentation consistency (DAC) is intrinsically more efficient than empirical risk minimization on augmented data (DA-ERM).
Ensembles and Cocktails: Robust Finetuning for Natural Language Generation
- Computer Science
- 2021
This work presents methods to combine the benefits of full and lightweight finetuning, achieving strong performance both ID and OOD, and provides some explanatory theory in a multiclass logistic regression setting with a large number of classes.
ProxyMix: Proxy-based Mixup Training with Label Refinery for Source-Free Domain Adaptation
- Computer ScienceArXiv
- 2022
This work proposes an effective method named Proxy-based Mixup training with label refinery (ProxyMix), which defines the weights of the classifier as the class prototypes and constructs a class-balanced proxy source domain by the nearest neighbors of the prototypes to bridge the unseen source domain and the target domain.
References
SHOWING 1-10 OF 41 REFERENCES
Temporal Ensembling for Semi-Supervised Learning
- Computer ScienceICLR
- 2017
Self-ensembling is introduced, where it is shown that this ensemble prediction can be expected to be a better predictor for the unknown labels than the output of the network at the most recent training epoch, and can thus be used as a target for training.
Semi-supervised Domain Adaptation with Subspace Learning for visual recognition
- Computer Science2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
- 2015
A novel domain adaptation framework, named Semi-supervised Domain Adaptation with Subspace Learning (SDASL), which jointly explores invariant low-dimensional structures across domains to correct data distribution mismatch and leverages available unlabeled target examples to exploit the underlying intrinsic information in the target domain.
Unsupervised Data Augmentation for Consistency Training
- Computer ScienceNeurIPS
- 2020
A new perspective on how to effectively noise unlabeled examples is presented and it is argued that the quality of noising, specifically those produced by advanced data augmentation methods, plays a crucial role in semi-supervised learning.
Fast Generalized Distillation for Semi-Supervised Domain Adaptation
- Computer ScienceAAAI
- 2017
It is shown that without accessing the source data, GDSDA can effectively utilize the unlabeled data to transfer the knowledge from the source models to efficiently solve the SDA problem.
Frustratingly Easy Semi-Supervised Domain Adaptation
- Computer Science
- 2010
This work builds on the notion of augmented space and harnesses unlabeled data in target domain to ameliorate the transfer of information from source to target and can be applied as a pre-processing step to any supervised learner.
Semi-supervised Domain Adaptation with Instance Constraints
- Computer Science2013 IEEE Conference on Computer Vision and Pattern Recognition
- 2013
It is shown that imposing smoothness constraints on the classifier scores over the unlabeled data can lead to improved adaptation results, and this work proposes techniques that build on existing domain adaptation methods by explicitly modeling these relationships, and demonstrates empirically that they improve recognition accuracy.
Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks
- Computer Science2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
- 2017
This generative adversarial network (GAN)-based method adapts source-domain images to appear as if drawn from the target domain, and outperforms the state-of-the-art on a number of unsupervised domain adaptation scenarios by large margins.
Self-ensembling for visual domain adaptation
- Computer ScienceICLR
- 2018
In small image benchmarks, the use of self-ensembling for visual domain adaptation problems not only outperforms prior art, but can also achieve accuracy that is close to that of a classifier trained in a supervised fashion.
Gotta Adapt 'Em All: Joint Pixel and Feature-Level Domain Adaptation for Recognition in the Wild
- Computer Science2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
- 2019
This work uses 3D geometry and image synthesis based on a generalized appearance flow to preserve identity across pose transformations, while using an attribute-conditioned CycleGAN to translate a single source into multiple target images that differ in lower-level properties such as lighting.
Generalizing to unseen domains via distribution matching
- Computer Science
- 2019
This work focuses on domain generalization: a formalization where the data generating process at test time may yield samples from never-before-seen domains (distributions), and relies on a simple lemma to derive a generalization bound for this setting.