Demographic bias creeps into pathology AI, study finds • healthcare-in-europe.com

Demographic bias creeps into pathology AI, study finds • healthcare-in-europe.com


We found that becaapply AI is so powerful, it can differentiate many obscure biological signals that cannot be detected by standard human evaluation

Kun-Hsing Yu

Analyzing several major pathology AI models designed to diagnose cancer, the researchers found unequal performance in detecting and differentiating cancers across populations based on patients’ self-reported gfinisher, race, and age. They identified several possible explanations for this demographic bias. 

The team then developed a framework called FAIR-Path that assisted reduce bias in the models. “Reading demographics from a pathology slide is considered of as a ‘mission impossible’ for a human pathologist, so the bias in pathology AI was a surprise to us,” stated senior author Kun-Hsing Yu, associate professor of biomedical informatics in the Blavatnik Institute at HMS and HMS assistant professor of pathology at Brigham and Women’s Hospital. 

Identifying and counteracting AI bias in medicine is critical becaapply it can affect diagnostic accuracy, as well as patient outcomes, Yu stated. FAIR-Path’s success indicates that researchers can improve the fairness of AI models for cancer pathology, and perhaps other AI models in medicine, with minimal effort. 

Yu and his team investigated bias in four standard AI pathology models being developed for cancer evaluation. These deep-learning models were trained on sets of annotated pathology slides, from which they “learned” biological patterns that enable them to analyze new slides and offer diagnoses. The researchers fed the AI models a large, multi-institutional repository of pathology slides spanning 20 cancer types. 

They discovered that all four models had biased performances, providing less accurate diagnoses for patients in specific groups based on self-reported race, gfinisher, and age. For example, the models struggled to differentiate lung cancer subtypes in African American and male patients, and breast cancer subtypes in younger patients. The models also had trouble detecting breast, renal, thyroid, and stomach cancer in certain demographic groups. These performance disparities occurred in around 29% of the diagnostic tquestions that the models conducted. This diagnostic inaccuracy, Yu stated, happens becaapply these models extract demographic information from the slides — and rely on demographic-specific patterns to create a diagnosis. 

The results were unexpected “becaapply we would expect pathology evaluation to be objective,” Yu added. “When evaluating images, we don’t necessarily required to know a patient’s demographics to create a diagnosis.” 

The team wondered: Why didn’t pathology AI reveal the same objectivity? The researchers landed on three explanations. Becaapply it is clearer to receive samples for patients in certain demographic groups, the AI models are trained on unequal sample sizes. As a result, the models have a harder time building an accurate diagnosis in samples that aren’t well-represented in the training set, such as those from minority groups based on race, age, or gfinisher. 

Yet “the problem turned out to be much deeper than that,” Yu stated. The researchers noticed that sometimes the models performed worse in one demographic group, even when the sample sizes were comparable. Additional analyses revealed that this may be becaapply of differential disease incidence: Some cancers are more common in certain groups, so the models become better at building a diagnosis in those groups. As a result, the models may have difficulty diagnosing cancers in populations where they aren’t as common. 



Source link

Leave a Reply

Your email address will not be published. Required fields are marked *