
Antiracism in AI: How to Build Bias Checkpoints Into Your Development and Delivery Process
Much of the data that AI depends on is tainted with racial bias.
Artificial Intelligence (AI) is a powerful tool in healthcare because it can process large volumes of data to inform decisions that
Due to socioeconomic inequalities in healthcare delivery, research data, and medical records are
The potential harm of biased AI is so great that many regulatory bodies are taking bold steps to set uplegal guardrails, such as the European Union’s recent
Most industry professionals are aware of this problem, but its prevalence shows that many of the teams building AI solutions may still be overlooking critical areas where bias can be introduced into an AI system. To ensure healthcare AI solutions have a positive impact and don’t propagate existing health inequities, healthcare leaders should ensure that AI implementation processes include bias checkpoints at each phase of development and delivery.
Here are a few examples of questions that healthcare leaders should consider when building machine learning solutions for their organization:
Are you building solutions that address the health needs of all, or are you placing higher priority on problems that affect advantaged groups?
Much of the conversation around AI bias has focused on data and models, but the first checkpoint should occur much earlier in the process when problem that needs to be addressed is first identified
There are many healthcare concerns that
Evidence
Do you have balanced data from users across races, genders, and socioeconomic backgrounds, or are you missing data from underserved and marginalized groups?
If data is biased, the product will be too. Unfortunately, many of the most used data sources in healthcare technology are not equally representative of all populations. For example, research
Because access to healthcare is inequal, clinical records alone are not enough. Incorporating nonclinical data such as social determinants of health (SDOH)indicators may help bridge this gap. A recent
Have you identified and accounted for any biases in your prediction definition and algorithm design?
Once imbalances are rectified within these datasets, it is important to think carefully about how to interpret each variable and define predictive outcomes. When predicting for an unknown variable, assumptions need to be made at times about which data features are most likely to have predictive power. However, you will need to evaluate the feature selection mechanism, as opposed to using a “black box” model, so that you can consider any biases that may be present in the correlations.
Additionally, you should consider a prediction label itself to ensure that you are directly measuring what you intend to. Research
After implementing the model, are you monitoring its performance and impact across different populations?
While not often discussed, perhaps the most important bias checkpoint relies on the measurement of real-world impact. In healthcare, the decisions you automate with machine learning are high-stakes, and you cannot afford to assume that your organization has perfectly eradicated bias in earlier steps. Therefore, it is critical to measure the downstream impact of the solution to ensure the model generalizes well to both the majority and minority groups within a user population. In the highest risk use cases, such as models used for diagnosis and treatment decisions, potential harms should be explicitly identified and measured.
Removing bias from machine learning solutions in the healthcare space requires dedicated attention from design to deployment. This technology has the potential to have a profound positive impact on the world, if the necessary steps are taken to ensure the industry is not propagating racial disparities or inequities.
Allison Langley is a data science manager at
Newsletter
Get the latest industry news, event updates, and more from Managed healthcare Executive.



















































