San Francisco District Attorney George Gascón says the city is using artificial intelligence to remove racial bias from the process of deciding whom to charge with crimes. A new AI tool scans police reports and automatically redacts any race information.
It’s part of an effort to remove implicit bias caused by social conditioning and learned associations, Gascón’s office said in a press release Wednesday.
“Lady Justice is depicted wearing a blindfold to signify impartiality of the law, but it is blindingly clear that the criminal justice system remains biased when it comes to race,” Gascón said. “This technology will reduce the threat that implicit bias poses to the purity of decisions.”
Stage one of the tool, bias mitigation review, removes details that could be connected to race, such as officer, witness and suspect names; officer star numbers; specific neighborhoods and districts; and hair and eye color.
Once investigators record a preliminary charge, they’ll gain access to the unredacted incident report and body camera footage. This second stage is called full review, and prosecutors will be required to record why the unredacted information led to any changes in their charges.
This information will be used to refine the AI tool, which is set to be fully implemented by the SFDA’s general felonies teams from July 1, 2019.
The tool, reported on earlier by the San Francisco Chronicle, was created by the Stanford Computational Policy Lab at no cost to Gascón’s office.
The tool’s lead developer, Stanford Assistant Professor Sharad Goel, said it will “reduce unnecessary incarceration.”
back in May, citing a breach of citizens’ civil liberties.