by Oliver Thomas and Thomas Kehrenberg
...using statistical techniques to give computer systems the ability to "learn" (e.g., progressively improve performance on a specific task) from data, without being explicitly programmed.
But there are problems:
Independence based fairness (i.e. Statistical Parity)
$$ \hat{Y} \perp S $$Separation based fairness (i.e. Equalised Odds/Opportunity)
$$ \hat{Y} \perp S | Y $$For both to hold, then either $S \perp Y$, our data is fair, or $\hat{Y} \perp Y$, we have a random predictor.
Similarly, Sufficiency cannot hold with either notion of fairness.
Consider a university, and we are in charge of administration!
We can only accept 50% of all applicants.
10,000 applicants are female and 10,000 of applicants are male.
We have been tasked with being fair with regard to gender.
We have an acceptance criteria that is highly predictive of success.
80% of those who meet the acceptance criteria will successfully graduate.
Only 10% of those who don't meet the acceptance criteria will successfully graduate.
As we're a good university we have a lot of applications from people who don't meet the acceptance criteria.
60% of female applicants meet the acceptance criteria.
40% of male applicants meet the acceptance criteria.
Remember, we can only accept 50% of all applicants
Accepted | Not | |
---|---|---|
Actually Graduate | ||
Don't Graduate |
Accepted | Not | |
---|---|---|
Actually Graduate | ||
Don't Graduate |
How would we solve this problem being fair using Statistical Parity as our measure?
Accepted | Not | |
---|---|---|
Actually Graduate | 4000 | 1200 |
Don't Graduate | 1000 | 3800 |
Accepted | Not | |
---|---|---|
Actually Graduate | 3300 | 500 |
Don't Graduate | 1700 | 4500 |
How would we solve this problem being fair using Equal Opportunity as our measure?
Accepted | Not | |
---|---|---|
Actually Graduate | 4440 | 760 |
Don't Graduate | 1110 | 3690 |
Accepted | Not | |
---|---|---|
Actually Graduate | 3245 | 555 |
Don't Graduate | 1205 | 4995 |
How would we solve this problem being fair using Calibration by Group as our measure?
Accepted | Not | |
---|---|---|
Actually Graduate | 4800 | 400 |
Don't Graduate | 1200 | 3600 |
Accepted | Not | |
---|---|---|
Actually Graduate | 3200 | 600 |
Don't Graduate | 800 | 5400 |
There's no right answer, all the above are "fair". It's important to consult domain experts to find which is the best fit for each problem. There is no one-size fits all.
The learned representation is uninterpretable by default. Recently Quadrianto et al constrained the representation to be in the same same as the input so that we could look at what changed
Referred to as "fair pipelines". Work has only just begun exploring these. Current research shows that these don't work (at the moment!)
Given we have a loss function, $\mathcal{L}(\theta)$.
In an unconstrained classifier, we would expect to see
$$ \min{\mathcal{L}(\theta)} $$To reduce Disparate Impact, Zafar adds a constraint to the loss function.
$$ \begin{aligned} \text{min } & \mathcal{L}(\theta) \\ \text{subject to } & P(\hat{y} \neq y|s = 0) − P(\hat{y} \neq y|s = 1) \leq \epsilon \\ \text{subject to } & P(\hat{y} \neq y|s = 0) − P(\hat{y} \neq y|s = 1) \geq -\epsilon \end{aligned} $$Calders and Verwer (2010) train two separate models: one for all datapoints with $s=0$ and another one for $s=1$
The thresholds of the model are then tweaked until they produce the same positive rate ($P(\hat{y}=1|s=0)=P(\hat{y}=1|s=1)$)
Disadvantage: $s$ has to be known for making predictions in order to choose the correct model.
https://tinyurl.com/ethicml
https://developers.google.com/machine-learning/crash-course/fairness
http://course18.fast.ai/lessons/lesson13.html
In the real world there are implications.
An individual doesn't just cease to exist after we've made our loan or bail decision.
The decision we make has consequences.
Area | Description |
---|---|
Active Harm | Expected change in credit score of an individual is negative |
Stagnation | Expected change in credit score of an individual is 0 |
Improvement | Expected change in credit score of an individual is positive |
Area | Description |
---|---|
Relative Harm | Expected change in credit score of an individual is less than if the selection policy had been to maximize profit |
Relative Improvement | Expected change in credit score of an individual is better than if the selection policy had been to maximize profit |
For those interested in more of an explanation of the equations, see Appendix
"Correlation doesn't imply causation"
But what is causation?
If we can understand what causes unfair behavior, then we can take steps to mitigate it.
Basic idea: if a sensitive attribute has no causal influence on prediction, don't use it.
But how do we model causation?
Solution: build causal graphs of your problem
Problem: causality cannot be inferred from observational data
Observational data can only show correlations
For causal information we have to do experiments. (But that is often not ethical.)
Example: Law school success
Task: given GPA score and LSAT (law school entry exam), predict grades after one year in law school: FYA (first year average)
Additionally two sensitive attributes: race and gender
Two possible graphs
$U$: set of all unobserved background variables
$P(\hat{y}_{s=i}(U) = 1|x, s=i)=P(\hat{y}_{s=j}(U) = 1|x, s=i)$
$i, j \in \{0, 1\}$
$\hat{y}_{s=k}$: prediction in the counterfactual world where $s=k$
practical consequence: $\hat{y}$ is counterfactually fair if it does not causally depend (as defined by the causal model) on $s$ or any descendants of $s$.
Alternative idea: