Antonella Basso (she/her/ella)

Testing Bias

I. Analyzing a Real-World Scenario for Sources of Bias

Task:

Using the figure below, we identify potential sources of bias in the following scenario and list them with justification, distinguishing whether these are likely to appear in the data collection/representation stage, the training/model building phase, or in the inference/action phase of the data processing pipeline.

Categories of AI Bias

Scenario:

Health care systems use automated tools to determine whether a patient should be inducted into an intensive “care management system” to cure what ails them. The idea is to determine which patients can benefit the most from being admitted to the system because it is expensive, and not everyone can be enrolled. It is hard to know who will benefit the most, so the systems instead try to predict a proxy: who will incur the most health care costs. The idea is that if your health care costs are predicted to be large, you probably need intensive treatments to be cured, and so you’re likely to benefit. The algorithm that makes this prediction uses the following training data set: input features for an individual are a list of insurance claim data from the previous year. Each such claim consists of

  1. Demographic information (age, sex, but not race)
  2. Type of insurance (private, medicare, medicaid, hmo, etc.)
  3. Disease diagnosis (usually expressed in a standard code)
  4. Procedures performed (X-rays, MRIs, surgeries, etc.—each expressed with a specific code)
  5. Medications
  6. Costs

The goal is to predict the cost of care this year. Formally, this is a regression problem with a mix of numerical and categorical variables, and a numeric output.

Response:

This is my favorite example of racial discrimination in automated decision making. I often use it to explain (to others) how an algorithm may still favor a particular social group(s) in prediction, despite being “blind” to group membership in the data. Namely, this case exemplifies the ways in which algorithmic neutrality not only fails to prevent biased decisions, but leads individuals to believe that their outcomes are fair and accept injustice willingly. It’s no surprise, given our nation’s historical legacies of oppression and racism towards Black and Brown individuals, that communities of color continue to experience systemic injustices on disproportionate levels. Among these are both the degree and quality of healthcare provision, as reflected in insurance claims data, which proves that healthcare expenditures in the U.S. are disparately low for marginalized groups. Thus, any model that utilizes this data to predict future healthcare costs is bound to replicate the systemic biases embedded within it to generate an adverse feedback loop of discriminatory outcomes—that is, unless used in tandem with a processing strategy explicitly designed to remove them. In this way, healthcare expenditure becomes a proxy for race, making its use as a basis for allocating those in more urgent need of care an implicit method for discriminating against vulnerable populations and keeping marginalized individuals on the margins.

Selection/Sampling Bias:

Processing/Validation Bias:

Use & Interpretation Bias:

II. Analyzing a Particular (Mathematical) Notion of Fairness

Task:

Against the backdrop of fair clustering, we assess the degree to which the fairness measure described below eliminates any form of bias that one might be concerned with in the subsequent scenarios provided.

In various exploratory data mining tasks the goal is to cluster a collection of objects into groups and determine whether the groups have some meaningful structure. Formally, we are given a set of \(n\) points, where each point is represented by a \(d\)-dimensional feature vector \(x \in \mathcal{R}^d\). Defining the distance between two points as the Euclidean distance:

\[ d(x,y) = \sqrt{\sum_{i=1}^d (x_i - y_i)^2}, \]

we can then outline the \(k\)-means problem as follows. Partition the points into \(k\) clusters \(C_1, \dots, C_k\) such that the sum of squared distance from each point to its cluster center is minimized, where the center of a cluster is defined as the centroid of the cluster:

\[ \mu(C) = \sum_{x \in C} \frac{x}{\mid C \mid}, \]

where \(\mid C \mid\) is the total number of points in the cluster. More precisely, the goal is to find \(C_1, \dots, C_k\) such that the following expression is minimized.

\[ \sum_{j=1}^k \sum_{x \in C_j} d^2 (x, \mu(C_j))) \]

In the exploration of fairness measures (How Do Design Choices Affect Fairness?) we’ve considered how to evaluate the fairness of a binary classifier that produces a 0-1 output. However, we now consider a “fair” equivalent of the \(k\)-means problem.

Suppose each point \(x\) also has a color feature \(g(x)\), such that we can determine the fraction of points with a particular color within each cluster. Then the goal then becomes to make sure these cluster-specific or “local” fractional values match the overall “global” proportions of colors reflected in the data. For example, if we have 15 points of which 5 are red and 10 are blue, and we want to cluster them into 5 clusters, then each cluster should contain 1 red and 2 blue points.

Let us consider two scenarios in which one might wish to cluster points.

Scenarios:

  1. Each point encodes different kinds of qualifications. The clusters represent people with similar qualifications. The “color” of a point is a binary gender encoding. The goal of the clustering is to group people into categories to target them with different kinds of job ads.

  2. Each point represents the location of a voter in a state. The clusters represent voting districts for a state assembly. The “color” of a point is the person’s registered political affiliation, assumming that there are only two parties.

Response:

Scenario 1:

Metric Evaluation:

Scenario 2:

Metric Evaluation: