Class Prior Estimation under Covariate Shift - no Problem?
@article{Tasche2022ClassPE, title={Class Prior Estimation under Covariate Shift - no Problem?}, author={Dirk Tasche}, journal={ArXiv}, year={2022}, volume={abs/2206.02449} }
. We show that in the context of classification the property of source and target distributions to be related by covariate shift may break down when the information content captured in the covariates is reduced, for instance by discretization of the covariates, dropping some of them, or by any transformation of the covariates even if it is domain-invariant. The consequences of this observation for class prior estimation under covariate shift are discussed. A probing algorithm as alternative…
One Citation
Factorizable Joint Shift in Multinomial Classification
- Computer Science, MathematicsMach. Learn. Knowl. Extr.
- 2022
For the multinomial (multiclass) classification setting, a representation of factorizable joint shift is derived in terms of the source (training) distribution, the target (test) prior class probabilities and the target marginal distribution of the features.
References
SHOWING 1-10 OF 31 REFERENCES
Probability theory - a comprehensive course
- MathematicsUniversitext
- 2008
Convergence Theorems are applied to the interpretation of Brownian Motion and the law of the Iterated Logarithm as well as to Martingales and Exchangeability.
Support and Invertibility in Domain-Invariant Representations
- Computer ScienceAISTATS
- 2019
This work gives generalization bounds for unsupervised domain adaptation that hold for any representation function by acknowledging the cost of non-invertibility and proposes a bound based on measuring the extent to which the support of the source domain covers the target domain.
When Training and Test Sets Are Different: Characterizing Learning Transfer
- Computer Science
- 2009
This chapter contains sections titled: Introduction, Conditional and Generative Models, real- life Reasons for Dataset Shift, Real-Life Reasons for dataset shift, Simple Covariate Shift, Prior Probability Shift, Sample Selection Bias, Imbalanced Data, Domain Shift, Source Component Shift and Gaussian Process Methods.
A Generalized Neyman-Pearson Criterion for Optimal Domain Adaptation
- Computer ScienceALT
- 2019
This work studies a class of domain adaptation problems that generalizes both the covariate shift assumption and a model for feature-dependent label noise, and establishes optimal classification on the target domain despite not having access to labelled data from this domain.
The Importance of Calibration for Estimating Proportions from Annotations
- Computer ScienceNAACL
- 2018
This paper identifies and differentiate between two relevant data generating scenarios (intrinsic vs. extrinsic labels), introduces a simple but novel method which emphasizes the importance of calibration, and analyzes and experimentally validate the appropriateness of various methods for each of the two scenarios.
Minimising quantifier variance under prior probability shift
- MathematicsCIKM Workshops
- 2021
It is found that the asymptotic variance of the maximum likelihood estimator is a function of the Brier score for the regression of the class label against the features under the test data set distribution, which suggests that optimising the accuracy of a base classifier on the training data set helps to reduce thevariance of the related quantifiers on the testData set.
Exact Fit of Simple Finite Mixture Models
- Computer Science
- 2014
It is pointed out that the maximum-likelihood (ML) approach to fitting the mixture distribution not only gives an optimum but even an exact fit if the authors allow the mixture components to vary but keep their density ratio fixed.
Domain Adaptation with Conditional Transferable Components
- MathematicsICML
- 2016
This paper aims to extract conditional transferable components whose conditional distribution is invariant after proper location-scale (LS) transformations, and identifies how P(Y) changes between domains simultaneously.
Calibrating sufficiently
- Computer ScienceStatistics
- 2021
The probing reduction approach is revisited and it is found that it produces an estimator of probabilistic classifiers that reduces grouping loss and identifies comonotonicity as a useful criterion for sufficiency.
Domain Adaptation with Factorizable Joint Shift
- Computer ScienceArXiv
- 2022
This paper proposes a new assumption, Factorizable Joint Shift (FJS), to handle the co-existence of sampling bias in covariates and labels, and provides theoretical and empirical understandings about when FJS degenerates to prior assumptions and when it is necessary.