Cornell Law School Logo - white on transparent background

Vol. 107, Issue 7

Article

Discredited Data

Ngozi Okidegbe, Associate Professor of Law & Assistant Professor of Computing and Data Science, Boston University

1 Nov 2022

Jurisdictions are increasingly employing pretrial algorithms as a solution to the racial and socioeconomic inequities in the bail system. But in practice, pretrial algorithms have reproduced the very inequities they were intended to correct. Scholars have diagnosed this problem as the biased data problem: pretrial algorithms generate racially and socioeconomically biased predictions because they are constructed and trained with biased data.

This Article contends that biased data is not the sole cause of algorithmic discrimination. Another reason pretrial algorithms produce biased results is that they are exclusively built and trained with data from carceral knowledge sources—the police, pretrial services agencies, and the court system. Redressing this problem will require a paradigmatic shift away from carceral knowledge sources toward non-carceral knowledge sources. This Article explores knowledge produced by communities most impacted by the criminal legal system (“community knowledge sources”) as one category of non-carceral knowledge sources worth utilizing. Though data derived from community knowledge sources have traditionally been discredited and excluded in the construction of pretrial algorithms, tapping into them offers a path toward developing algorithms that have the potential to produce racially and socioeconomically just outcomes.

To read this Article, please click here: Discredited Data.