Computer Vision MT25, Ethics and privacy


Flashcards

Why is it not sufficient to omit sensitive features (e.g. race) from an ML model used to make decisions?


Other features may correlate with the sensitive features.

Suppose:

  • $Y$ is a target variable (e.g. recidivism)
  • $R$ is the output of a classifier
  • $A$ is a sensitive attribute

@Define the fairness-related “independence” condition in this context.


The classifier response is independent from the sensitive attribute, i.e.

\[\mathbb P(R \mid A) = \mathbb P(R)\]

Suppose:

  • $Y$ is a target variable (e.g. recidivism)
  • $R$ is the output of a classifier
  • $A$ is a sensitive attribute

@Define the fairness-related “separation” condition in this context.


The classifier response is conditionally independent from the sensitive attribute given the target, i.e.

\[\mathbb P(R, A \mid Y) = \mathbb P(R \mid Y) \mathbb P(A \mid Y)\]

Suppose:

  • $Y$ is a target variable (e.g. recidivism)
  • $R$ is the output of a classifier
  • $A$ is a sensitive attribute

Then the fairness-related “separation” condition in this context is that the classifier response is conditionally independent from the sensitive attribute given the target, i.e.

\[\mathbb P(R, A \mid Y) = \mathbb P(R \mid Y) \mathbb P(A \mid Y)\]

What does this entail in terms of the error rates of the classifier?


All groups (delineated by $A$) experience the same false negative and false positive rate.

Suppose:

  • $Y$ is a target variable (e.g. recidivism)
  • $R$ is the output of a classifier
  • $A$ is a sensitive attribute

Then we have the fairness conditions:

  • Independence: $\mathbb P(R, A) = \mathbb P(R) \mathbb P(A)$
  • Separation: $\mathbb P(R, A \mid Y) = \mathbb P(R \mid Y) \mathbb P(A \mid Y)$

@State an unfortunate theorem in this context.


Suppose further:

  • $Y$ is binary
  • $A$ is not independent of $Y$
  • $R$ is not independent of $Y$

Then:

  • Both independence and separation cannot hold simultaneously.

@Define allocative harms in the context of ML-decision making.


Harms causes by a system allocating resources unfairly.

Can you classify these representational harms into the (potentially simultaneous) categories of:

  • denigration
  • stereotype
  • recognition
  • under-representation
  • ex-nomination

@Define the gender bias of a category $C$.


\[\frac{\text{total number of percieved male instances of }C}{\text{total number of instances of }C}\]



Related posts