Why healthcare researchers must be wary of misusing AI

Image source: Wikimedia Commons/Doctoroftcm (x-ray); Unsplash/Michael Dziedzic (circuitry); mashup: HiE/Behrends

News • Devil in the details

Why healthcare researchers must be wary of misusing AI

An international team of researchers advises that strong care needs to be taken not to misuse or overuse machine learning (ML) in healthcare research.

The commentary was published in Nature Medicine

“I absolutely believe in the power of ML but it has to be a relevant addition,” said neurosurgeon-in-training and statistics editor Dr Victor Volovici, first author of the comment, from Erasmus MC University Medical Center, The Netherlands. “Sometimes ML algorithms do not perform better than traditional statistical methods, leading to the publication of papers that lack clinical or scientific value.”

portrait of nan liu
Associate Professor Nan Liu

Image source: Duke-NUS

Real world examples have shown that the misuse of algorithms in healthcare could perpetuate human prejudices or inadvertently cause harm when the machines are trained on biased datasets. “Many believe ML will revolutionise healthcare because machines make choices more objectively than humans. But without proper oversight, ML models may do more harm than good,” said Associate Professor Nan Liu, senior author of the comment, from the Centre for Quantitative Medicine and Health Services & Systems Research Programme at Duke-NUS Medical School, Singapore. “If, through ML, we uncover patterns that we otherwise would not see—like in radiology and pathology images—we should be able to explain how the algorithms got there, to allow for checks and balances.” 

Together with a group of scientists from the UK and Singapore, the researchers highlight that although guidelines have been formulated to regulate the use of ML in clinical research, these guidelines are only applicable once a decision to use ML has been made and do not ask whether or when its use is appropriate in the first place. 

For example, companies have successfully trained ML algorithms to recognise faces and road objects using billions of images and videos. But when it comes to their use in healthcare settings, they are often trained on data in the tens, hundreds or thousands. “This underscores the relative poverty of big data in healthcare and the importance of working towards achieving sample sizes that have been attained in other industries, as well as the importance of a concerted, international big data sharing effort for health data,” the researchers write. 

lung x-ray of a healthy person
Information like the "R" marking of the right lung (shown here in a non-related x-ray of a healthy patient) can lead to confusion in machine learning algorithms, a prior example showed

Another issue is that most ML and deep learning algorithms (that do not receive explicit instructions regarding the outcome) are often still regarded as a ‘black box’. For example, at the start of the Covid-19 pandemic, scientists published an algorithm that could predict coronavirus infections from lung photos. Afterwards, it turned out that the algorithm had drawn conclusions based on the imprint of the letter ‘R’ (for ‘Right Lung’) in the photos, which was always found in a slightly different spot on the scans. “We have to get rid of the idea that ML can discover patterns in data that we cannot understand,” said Dr Volovici about the incident. “ML can very well discover patterns that we cannot see directly, but then you have to be able to explain how you came to that conclusion. In order to do that, the algorithm has to be able to show what steps it took, and that requires innovation.” 

The researchers advise that ML algorithms should be evaluated against traditional statistical approaches (when applicable) before they are used in clinical research. And when deemed appropriate, they should complement clinician decision-making, rather than replace it. “ML researchers should recognise the limits of their algorithms and models in order to prevent their overuse and misuse, which could otherwise sow distrust and cause patient harm,” the researchers write. 

The team is working on organising an international effort to provide guidance on the use of ML and traditional statistics, and also to set up a large database of anonymised clinical data that can harness the power of ML algorithms. 


Source: Duke-NUS Medical School

15.09.2022

Read all latest stories

Related articles

Photo

News • Award-winning research publication

Evaluating brain tumours with AI

Can AI help better evaluate images of brain tumours? A publication from German researchers on this topic presented at this years' ICIS conference won the Best Paper Award.

Photo

News • Algorithm-assisted diagnostics

AI in imaging: not as reliable as you'd think

Machine learning and AI are highly unstable in medical image reconstruction, and may lead to false positives and false negatives, a new study suggests. A team of researchers, led by the University of…

Photo

Video • Automation in radiology

Machine learning techniques generate clinical labels of medical scans

Researchers used machine learning techniques, including natural language processing algorithms, to identify clinical concepts in radiologist reports for CT scans, according to a study conducted at…

Related products

Subscribe to Newsletter