Skin Cancer Diagnoses Using AI Are as Reliable as those Made by Medical Experts

Artificial intelligence (AI) is already widely used in medical diagnostics. An Austrian-Australian research team led by dermatologist Harald Kittler from MedUni Vienna investigated the extent to which diagnosis and therapy of pigmented skin lesions benefit from it in a realistic clinical scenario. In a study published by The Lancet Digital Health, the team compared the accuracy in diagnosis and therapy recommendation of two different algorithms in smartphone applications with that of doctors. The results show that the AI application generally performs well in diagnosis. However, doctors were clearly superior when it came to treatment decisions.

The research team tested the AI application under realistic clinical conditions in two skin cancer centres, the University Department of Dermatology at MedUni Vienna and the Sydney Melanoma Diagnostic Centre in Australia. The prospective study consisted of two scenarios, with AI being used in scenario A for changes suspicious of skin cancer and in scenario B for patients with many moles. The AI-assisted application was compared in both cases with both medical experts and less experienced physicians.

In scenario A, 172 suspicious pigmented lesions (of which 84 were malignant) were examined in 124 patients; in scenario B, the research team analysed 5,696 pigmented lesions (of which 18 were malignant) in 66 patients. Two different AI-based smartphone applications were used: a novel 7-class AI algorithm and an ISIC algorithm already used in retrospective preliminary studies. In scenario A, the 7-class AI algorithm showed equivalent diagnostic accuracy compared to the experts while it was significantly superior to the less experienced physicians. The ISIC algorithm, on the other hand, performed significantly worse compared to experts, but better than the inexperienced users.

A critical view of AI decisions

In terms of treatment decisions, the 7-class algorithm was significantly inferior to the experts but superior to the inexperienced users. The results suggest that an AI-assisted smartphone application for skin cancer diagnosis makes similarly good diagnostic decisions as experts in a real clinical scenario. When it came to treatment decisions, however, the experts were superior to the AI. Kittler: "The AI application tends to remove more benign lesions in the treatment recommendation than experts would. If you take this into account, the AI application can certainly be used. It should also be borne in mind that if it is used uncritically, too many false-positive findings would have to be clarified."

Menzies SW, Sinz C, Menzies M, Lo SN, Yolland W, Lingohr J, Razmara M, Tschandl P, Guitera P, Scolyer RA, Boltz F, Borik-Heil L, Herbert Chan H, Chromy D, Coker DJ, Collgros H, Eghtedari M, Corral Forteza M, Forward E, Gallo B, Geisler S, Gibson M, Hampel A, Ho G, Junez L, Kienzl P, Martin A, Moloney FJ, Regio Pereira A, Ressler JM, Richter S, Silic K, Silly T, Skoll M, Tittes J, Weber P, Weninger W, Weiss D, Woo-Sampson P, Zilberg C, Kittler H.
Comparison of humans versus mobile phone-powered artificial intelligence for the diagnosis and management of pigmented skin cancer in secondary care: a multicentre, prospective, diagnostic, clinical trial.
Lancet Digit Health. 2023 Oct;5(10):e679-e691. doi: 10.1016/S2589-7500(23)00130-9

Most Popular Now

ChatGPT can Produce Medical Record Notes…

The AI model ChatGPT can write administrative medical notes up to ten times faster than doctors without compromising quality. This is according to a new study conducted by researchers at...

Can Language Models Read the Genome? Thi…

The same class of artificial intelligence that made headlines coding software and passing the bar exam has learned to read a different kind of text - the genetic code. That code...

Study Shows Human Medical Professionals …

When looking for medical information, people can use web search engines or large language models (LLMs) like ChatGPT-4 or Google Bard. However, these artificial intelligence (AI) tools have their limitations...

Bayer and Google Cloud to Accelerate Dev…

Bayer and Google Cloud announced a collaboration on the development of artificial intelligence (AI) solutions to support radiologists and ultimately better serve patients. As part of the collaboration, Bayer will...

Shared Digital NHS Prescribing Record co…

Implementing a single shared digital prescribing record across the NHS in England could avoid nearly 1 million drug errors every year, stopping up to 16,000 fewer patients from being harmed...

Ask Chat GPT about Your Radiation Oncolo…

Cancer patients about to undergo radiation oncology treatment have lots of questions. Could ChatGPT be the best way to get answers? A new Northwestern Medicine study tested a specially designed ChatGPT...

North West Anglia Works with Clinisys to…

North West Anglia NHS Foundation Trust has replaced two, legacy laboratory information systems with a single instance of Clinisys WinPath. The trust, which serves a catchment of 800,000 patients in North...

Can AI Techniques Help Clinicians Assess…

Investigators have applied artificial intelligence (AI) techniques to gait analyses and medical records data to provide insights about individuals with leg fractures and aspects of their recovery. The study, published in...

AI Makes Retinal Imaging 100 Times Faste…

Researchers at the National Institutes of Health applied artificial intelligence (AI) to a technique that produces high-resolution images of cells in the eye. They report that with AI, imaging is...

SPARK TSL Acquires Sentean Group

SPARK TSL is acquiring Sentean Group, a Dutch company with a complementary background in hospital entertainment and communication, and bringing its Fusion Bedside platform for clinical and patient apps to...

Standing Up for Health Tech and SMEs: Sh…

AS the new chair of the health and social care council at techUK, Shane Tickell talked to Highland Marketing about his determination to support small and innovative companies, by having...

GPT-4 Matches Radiologists in Detecting …

Large language model GPT-4 matched the performance of radiologists in detecting errors in radiology reports, according to research published in Radiology, a journal of the Radiological Society of North America...