# Null-sampling for Interpretable and Fair Representations

@inproceedings{Kehrenberg2020NullsamplingFI, title={Null-sampling for Interpretable and Fair Representations}, author={Thomas Kehrenberg and Myles Bartlett and Oliver Thomas and Novi Quadrianto}, booktitle={ECCV}, year={2020} }

We propose to learn invariant representations, in the data domain, to achieve interpretability in algorithmic fairness. Invariance implies a selectivity for high level, relevant correlations w.r.t. class label annotations, and a robustness to irrelevant correlations with protected characteristics such as race or gender. We introduce a non-trivial setup in which the training set exhibits a strong bias such that class label annotations are irrelevant and spurious correlations cannot be… Expand

#### Figures, Tables, and Topics from this paper

#### 3 Citations

Fair Normalizing Flows

- Computer Science
- ArXiv
- 2021

This work presents Fair Normalizing Flows (FNF), a new approach offering more rigorous fairness guarantees for learned representations and experiments demonstrate the effectiveness of FNF in enforcing various group fairness notions, as well as other attractive properties such as interpretability and transfer learning. Expand

Personalizing Pre-trained Models

- Computer Science
- ArXiv
- 2021

This work developed a technique, called Multi-label Weight Imprinting (MWI), for multi-label, continual, and few-shot learning, and CLIPPER (CLIP PERsonalized) uses image representations from CLIP, a large-scale image representation learning model trained using weak natural language supervision. Expand

Fair Representation: Guaranteeing Approximate Multiple Group Fairness for Unknown Tasks

- Computer Science
- ArXiv
- 2021

It is proved that, although fair representation might not guarantee fairness for all prediction tasks, it does guarantee fairnessFor an important subset of tasks—the tasks for which the representation is discriminative. Expand

#### References

SHOWING 1-10 OF 48 REFERENCES

Discovering Fair Representations in the Data Domain

- Computer Science
- 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
- 2019

This work proposes to cast the problem ofpretability and fairness in computer vision and machine learning applications as data-to-data translation, i.e. learning a mapping from an input domain to a fair target domain, where a fairness definition is being enforced. Expand

Flexibly Fair Representation Learning by Disentanglement

- Computer Science, Mathematics
- ICML
- 2019

This work proposes an algorithm for learning compact representations of datasets that are useful for reconstruction and prediction, but are also flexible fair, meaning they can be easily modified at test time to achieve subgroup demographic parity. Expand

Learning Fair Representations via an Adversarial Framework

- Computer Science, Mathematics
- ArXiv
- 2019

A minimax adversarial framework with a generator to capture the data distribution and generate latent representations, and a critic to ensure that the distributions across different protected groups are similar provides a theoretical guarantee with respect to statistical parity and individual fairness. Expand

Unsupervised Adversarial Invariance

- Computer Science, Mathematics
- NeurIPS
- 2018

This work presents a novel unsupervised invariance induction framework for neural networks that learns a split representation of data through competitive training between the prediction task and a reconstruction task coupled with disentanglement, without needing any labeled information about nuisance factors or domain knowledge. Expand

Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations

- Computer Science
- ArXiv
- 2017

An adversarial training procedure is used to remove information about the sensitive attribute from the latent representation learned by a neural network, and the data distribution empirically drives the adversary's notion of fairness. Expand

Discovering Interpretable Representations for Both Deep Generative and Discriminative Models

- Computer Science
- ICML
- 2018

This work provides an interpretable lens for an existing model, and proposes two interpretability frameworks which rely on joint optimization for a representation which is both maximally informative about the side information and maximally compressive about the non-interpretable data factors. Expand

The Variational Fair Autoencoder

- Mathematics, Computer Science
- ICLR
- 2016

This model is based on a variational autoencoding architecture with priors that encourage independence between sensitive and latent factors of variation that is more effective than previous work in removing unwanted sources of variation while maintaining informative latent representations. Expand

beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework

- Computer Science
- ICLR
- 2017

Learning an interpretable factorised representation of the independent data generative factors of the world without supervision is an important precursor for the development of artificial… Expand

Censoring Representations with an Adversary

- Computer Science, Mathematics
- ICLR
- 2016

This work forms the adversarial model as a minimax problem, and optimize that minimax objective using a stochastic gradient alternate min-max optimizer, and demonstrates the ability to provide discriminant free representations for standard test problems, and compares with previous state of the art methods for fairness. Expand

Learning Fair Representations

- Mathematics, Computer Science
- ICML
- 2013

We propose a learning algorithm for fair classification that achieves both group fairness (the proportion of members in a protected group receiving positive classification is identical to the… Expand