
Schizophrenia Diagnoses Have Highest Likelihood of AI Racial Bias, Study Shows
Large language models (LLMs) were more likely to suggest inferior treatment options when reviewing hypothetical cases when patients' Black identity was listed or implied.
Artificial intelligence large language models (LLMs) showed the most racial bias when dealing with schizophrenic patients when compared with patients with eating disorders, depression, anxiety or attention-deficit and hyperactivity disorder (ADHD), according to a
“We found most LLMs exhibited some form of bias when dealing with African American patients, at times recommending dramatically different treatments for the same psychiatric illness and otherwise same patient,” corresponding author Elias Aboujaoude, M.D., MA, director of the Program in Internet, Health and Society in the Department of Biomedical Sciences at Cedars-Sinai, said in the study.
In this study, Aboujaoude and his team asked four of the most popular LLMs in psychiatry (Claude, ChatGPT, Gemini and NewMes-1) for a diagnosis and treatment plan for 10 hypothetical patient cases. For each case, race was either explicitly stated, implied or left ambiguous.
Responses were then rated by a clinical neuropsychologist and a social psychologist using a 0-3 assessment scale, with 3 indicating the highest bias.
LLMs were more likely to propose inferior treatments when patient race was explicitly or implicitly indicated. Diagnostic decisions showed less bias, with most scores at a 1.5 or below.
For example, one LLM showed increased interest in recommending reducing alcohol intake only for Black patients. Another LLM brought up guardianship cases only for Black patients.
NewMes-1 was the LLM that exhibited the highest amount of racial bias.
“These findings underscore critical concerns about the potential for AI to perpetuate racial disparities in mental healthcare, emphasizing the necessity of rigorous bias assessment in algorithmic medical decision support systems,” Aboujaoude and his team write.
LLMs have been increasingly used in the mental health field to increase efficiency. They have also been praised for their ability to cut back on documentation burden, which consumes approximately
However, other evidence shows that LLMs make significantly more errors when processing mental health information from minority groups.
Aboujaoude and his team suggested that the biases of LLMs are “learned” from the content used to train them.
Researchers in a
Schizophrenia affects between
Black Americans are approximately
“Our findings serve as a call to action for stakeholders across the healthcare AI ecosystem to help ensure that these technologies enhance health equity rather than reproduce or exacerbate existing inequities,” Aboujaoude and his team said. “This will require close collaboration between researchers, clinicians, healthcare institutions, and policymakers to establish and maintain robust standards for AI adoption.”
Newsletter
Get the latest industry news, event updates, and more from Managed healthcare Executive.
















































