In a case that exposes the hazards of entrusting law enforcement to AI software, Porcha Woodruff, an African American mother, has embarked on a legal battle against the City of Detroit and a Detroit Police Department detective. This comes after she was wrongly apprehended while eight months pregnant due to a faulty facial recognition match. The incident sheds light on the ongoing debate surrounding the use of AI technology in policing and the potential for racial discrimination.
The federal lawsuit, filed August 3, 2023, outlines distressing events that unfolded on February 16, 2023. On that morning, while preparing her two young children for school, Porcha Woodruff found herself bombarded by six Detroit police officers with a warrant for her arrest for a carjacking and robbery incident that occurred on January 29.
Police Chief Perspective on Malpractice:
Fast-forward to August 10, 2023, Detroit’s police chief, James White, offered his perspective on the case. He maintained that the use of facial recognition technology was not at fault, but that “poor investigative work” had led to Woodruff’s wrongful arrest. Chief White emphasized that the misstep was the result of a detective presenting an image generated by facial recognition technology to the carjacking and robbery victim, which linked to a 2015 mugshot of Porcha Woodruff. The detective then acted in a way that was in direct violation of the department’s policy by using the mugshot photo in the lineup presented to the victim. As a result, Woodruff was arrested.
What adds to the severity of the situation is that Woodruff was in the advanced stages of pregnancy – eight months pregnant at the time the police officers arrived at her doorstep. The police report for the January 29 incident, in no way indicated that the carjacking and robbery suspect was pregnant – a fact that was in stark contrast to Woodruff’s situation.
According to Woodruff’s lawsuit, “Facial recognition alone cannot serve as probable cause for arrests, as a computer’s identification is prone to errors that humans might also make.” This assertion underlines the inherent limitations of technology and the potential for it to be injustices.
Racial Bias and Discrimination
This case highlights a troubling pattern of racial discrimination. The lawsuit contends that Detroit police have engaged in the systemic mistreatment of Woodruff and other African American citizens through the utilization of facial recognition technology. Studies have indicated that these technologies disproportionately misidentify racial minorities, with the National Institute of Standards and Technology finding that Native American, African American, and Asian individuals are more likely to be affected. The report revealed that certain algorithms were up to 100 times more likely to incorrectly match two different non-White people, further underscoring the technology’s racial bias.
The incident involving Porcha Woodruff is not isolated. In 2020, the American Civil Liberties Union (ACLU) filed a complaint against the Detroit Police Department in what became the first documented case of wrongful arrest due to facial recognition technology. The ACLU’s action underscored the urgent need for reform and accountability in the application of such tools within law enforcement.
The Porcha Woodruff case continues to spark debates about the consequences of relying on technology to make pivotal decisions in the criminal justice system. As society navigates the integration of advanced tools into policing, it becomes increasingly critical to scrutinize their potential biases and limitations. The legal battle that Woodruff has initiated serves as a reminder that technology, while valuable, must be employed ethically to prevent unjust outcomes to preserve the principles of justice and equality.