Disaggregated Interventions to Reduce Inequality

  title={Disaggregated Interventions to Reduce Inequality},
  author={Lucius E.J. Bynum and Joshua R. Loftus and Julia Stoyanovich},
  journal={Equity and Access in Algorithms, Mechanisms, and Optimization},
A significant body of research in the data sciences considers unfair discrimination against social categories such as race or gender that could occur or be amplified as a result of algorithmic decisions. Simultaneously, real-world disparities continue to exist, even before algorithmic decisions are made. In this work, we draw on insights from the social sciences brought into the realm of causal modeling and constrained optimization, and develop a novel algorithmic framework for tackling pre… 

Figures and Tables from this paper

The Many Facets of Data Equity

This paper attempts to address data equity broadly, identify different ways in which it is manifest in data-driven systems and proposes a research agenda.



Causal Multi-level Fairness

The problem of multi-level fairness is formalized using tools from causal inference in a manner that allows one to assess and account for effects of sensitive attributes at multiple levels, and importance of the problem is illustrated by illustrating residual unfairness if macro-level sensitive attributes are not accounted for, or included without accounting for their multi- level nature.

Race as a Bundle of Sticks: Designs that Estimate Effects of Seemingly Immutable Characteristics

Although understanding the role of race, ethnicity, and identity is central to political science, methodological debates persist about whether it is possible to estimate the effect of something

Learning Optimal Fair Policies

This paper uses methods from causal inference and constrained optimization to learn optimal policies in a way that addresses multiple potential biases which afflict data analysis in sensitive contexts, extending the approach of Nabi & Shpitser (2018).

Algorithmic Fairness and the Social Welfare Function

It is argued that it would be beneficial to model fairness and algorithmic bias more holistically, including both a generative model of the underlying social phenomena and a description of a global welfare function.

Subsidy Allocations in the Presence of Income Shocks

A model of welfare is introduced that incorporates income, wealth, and income shocks and it is shown that it can vary, at times substantially, from measures of welfare that only use income or wealth.

Towards a critical race methodology in algorithmic fairness

It is argued that algorithmic fairness researchers need to take into account the multidimensionality of race, take seriously the processes of conceptualizing and operationalizing race, focus on social processes which produce racial inequality, and consider perspectives of those most affected by sociotechnical systems.

Counterfactual Fairness

This paper develops a framework for modeling fairness using tools from causal inference and demonstrates the framework on a real-world problem of fair prediction of success in law school.

Fairness, Equality, and Power in Algorithmic Decision-Making

This work argues that leading notions of fairness suffer from three key limitations: they legitimize inequalities justified by "merit;" they are narrowly bracketed, considering only differences of treatment within the algorithm; and they consider between-group and not within-group differences.

Measurement and Fairness

It is argued that many of the harms discussed in the literature on fairness in computational systems are direct results of such mismatches, and it is shown how some of these harms could have been anticipated and mitigated if viewed through the lens of measurement modeling.

What Do Randomized Studies of Housing Mobility Demonstrate?

A frame-work for causal inference when interference is present is developed and a number of causal estimands of interest are defined, which are of great importance for a researcher who fails to recognize that a treatment is beneficial when in fact it is universally harmful.