Study Shows ChatGPT Failed when Challenging ESCMID Guideline for Treating Brain Abscesses

With artificial intelligence (AI) poised to become a fundamental part of clinical research and decision making, many still question the accuracy of ChatGPT, a sophisticated AI language model, to support complex diagnostic and treatment processes.

Now a new study, being presented at this year's ESCMID Global Congress (formerly ECCMID) in Barcelona, Spain (27-30 April), which pitted ChatGPT against the ESCMID guideline for the management of brain abscesses, found that while ChatGPT seems able to give recommendations on key questions about diagnosis and treatment in most cases, some of the AI model’s responses could put patients at risk.

The study was conducted by members of the ESCMID Study Group for Infectious Diseases of the Brain (ESGIB), and is published in The Journal of Neurology.

"Anything less than 100% is a failure when you’re dealing with patient safety," says author Dr Susanne Dyckhoff-Shen from LMU University Hospital Munich in Germany and a member of ESCMID. "While we are amazed by ChatGPT's knowledge on the management of brain abscesses, there are some key limitations when it comes to using the AI model as a medical device, including potential patient harm and the lack of transparency about which data are used to provide responses."

The ability of AI to rapidly assimilate, process, and interpret vast data sets offers tantalising prospects. But are time-consuming processes to create medical guidelines still necessary, or could AI models trained on a wealth of scientific medical literature rival clinical experts in answering complex clinical questions?

Brain abscesses are a potentially life-threatening central nervous system (CNS) infection that require immediate identification and treatment to prevent severe neurological complications and even death.

Historically, the management of brain abscesses has been largely guided by clinical experience and limited studies, but in 2023 ESCMID fulfilled the need for a standardised approach by developing an international guideline.

To find out whether ChatGPT is able to professionally evaluate medical research and give scientifically valid recommendations, a European team of researchers tested the AI model to see whether it could accurately provide answers to 10 key questions on brain abscess diagnostics and treatment in comparison to the ESCMID guideline.

First, the researchers asked ChatGPT (version 4) to answer 10 questions that had been developed and appraised by the ESCMID committee for their brain abscess guideline without any additional information.

Then, ChatGPT was additionally primed with the text of the same scientific research articles that were used to develop the guideline before asking the same questions. This was done to see if ChatGPT could provide more aligned recommendations when given the same data used for guideline development.

The AI-generated responses were then compared to the recommendations of the ESCMID guideline by three independent infectious CNS disease experts for their clarity, alignment with the guideline, and patient risk.

The researchers found that overall, for 17 out of 20 questions asked (with and without data input), ChatGPT's responses were clear on the management of patients with brain abscess, including grade of evidence and strength of recommendation, with clarity assessed at 80-90%.

However, the AI model did not provide clear enough answers to guide physicians on treatment decisions on withholding microbials until surgery and prophylactic antiepileptic treatment (questions 2 and 10).

Without additional data input, ChatGPT’s responses to 70% (7/10) of questions were very similar to the guideline recommendation. However, the AI model failed to come up with the correct advice on three questions relating to withholding microbials, consolidation therapy, and prophylactic antiepileptic treatment (questions 2, 8, and 10). Importantly, however, these incorrect responses would not have harmed patients.

Surprisingly, data input resulted in fewer correct answers (40%) including two recommendations that directly contradicted the guideline, that could have put patients at risk.

Question 6 about duration of antimicrobial therapy for bacterial brain abscess was answered by ChatGPT after data entry as "intravenous administration for about 4 weeks, followed by 12 weeks oral medication," but the ESCMID guideline recommends "a total duration of 6-8 weeks of intravenous antimicrobials...".

For question 7 about early transition to oral antimicrobials, after data input ChatGPT recommended that "an early switch to oral antibiotics during the first 14 days of treatment... seems to be associated with favorable outcomes in selected patients." However, the ESCMID guideline committee judged that there was insufficient evidence to provide a recommendation for this question.

In both cases, following ChatGPT's advice could have potentially led to patient harm.

"The fact that ChatGPT's recommendations were inferior after data entry might be due to an overvaluation of the few observational studies provided for key questions 6 and 7. For one of those, even the guideline committee was not able to give a recommendation as the evidence was insufficient to answer the question," says senior author Professor Mattias Klein from LMU University Hospital Munich in Germany and a member of the ESCMID committee which established the guideline.

"As the exact operating procedures of ChatGPT remain unclear, we speculate that while the AI model can process large amounts of data quickly, it may lack the ability to correctly classify and weigh the data based on their scientific quality. Moreover, it remains unclear which data are used for ChatGPT's responses as it does not disclose the sources of its answers, which risks dubious literature being used."

Dr Dyckhoff-Shen adds, "It is alarming to think that patients could have come to harm if ChatGPT's advice on two key questions had been followed. The nuanced expertise of expert committees remains essential, especially to answer complex clinical queries. Blindly relying on AI could put patients at risk."

Nevertheless, the authors note that ChatGPT’s knowledge was from before September 2021 and the questions in the study covered some extremely complex medical issues, some of which are controversial even among experts and for which hardly any robust data are available. However, even when primed with the same research articles that were used to develop the ESCMID guideline, ChatGPT’s advice aligned even less with the guideline. They recommend that the quality of ChatGPT should be reviewed on an ongoing basis following its evolution and further development.

The authors explain further that ChatGPT, like many AI models, has a cutoff date for the information it can access. For the current version as of today, it was last trained on data up until January 2022. "This means that while it can provide responses based on a wide range of information, it doesn't have access to real-time data or events occurring after that date," says Dr Dyckhoff-Shen. "When we used ChatGPT for our study before August 2023, it was only trained on data up until September 2021. There was no possibility for us to get a more up-to-date version at that time because ChatGPT was not trained further yet. This is also the reason why we used a second approach by prompting ChatGPT with relevant scientific articles that were used by the ESGIB group to give recommendations in the ESCMID guideline so that we sort of 'manually' tried to get it more up-to-date.

"In the future, it would be interesting to re-assess ChatGPT's knowledge in the future after internal optimisation processes. However, once the chatbot has access to the ESCMID guideline itself, it could just use the recommendations from the guideline thus rendering a comparison no longer useful."

Dyckhoff-Shen S, Koedel U, Brouwer MC, Bodilsen J, Klein M.
ChatGPT fails challenging the recent ESCMID brain abscess guideline.
J Neurol. 2024 Apr;271(4):2086-2101. doi: 10.1007/s00415-023-12168-1

Most Popular Now

Most Advanced Artificial Touch for Brain…

For the first time ever, a complex sense of touch for individuals living with spinal cord injuries is a step closer to reality. A new study published in Science, paves...

Picking the Right Doctor? AI could Help

Years ago, as she sat in waiting rooms, Maytal Saar-Tsechansky began to wonder how people chose a good doctor when they had no way of knowing a doctor's track record...

From Text to Structured Information Secu…

Artificial intelligence (AI) and above all large language models (LLMs), which also form the basis for ChatGPT, are increasingly in demand in hospitals. However, patient data must always be protected...

Deep Learning Model Helps Detect Lung Tu…

A new deep learning model shows promise in detecting and segmenting lung tumors, according to a study published in Radiology, a journal of the Radiological Society of North America (RSNA)...

AI Innovation Unlocks Non-Surgical Way t…

Researchers have developed an artificial intelligence (AI) model to detect the spread of metastatic brain cancer using MRI scans, offering insights into patients’ cancer without aggressive surgery. The proof-of-concept study, co-led...

New Study Reveals AI's Transformati…

Intensive care units (ICUs) face mounting pressure to effectively manage resources while delivering optimal patient care. Groundbreaking research published in the INFORMS journal Information Systems Research highlights how a novel...

One of the Largest Global Surveys of Soc…

As leaders gather for the World Economic Forum Annual Meeting 2025 in Davos, Leaps by Bayer, the impact investing arm of Bayer, and Boston Consulting Group (BCG) announced the launch...

New Computer Models Open Door to Far Mor…

With antibiotic resistance a growing problem, University of Virginia School of Medicine researchers have developed cutting-edge computer models that could give the disease-fighting drugs a laser-like precision to target only...

New Biomarkers to Detect Colorectal Canc…

Machine learning and artificial intelligence (AI) techniques and analysis of large datasets have helped University of Birmingham researchers to discover proteins that have strong predictive potential for colorectal cancer. In a...

Sam Neville Joins the Highland Marketing…

Leading chief nursing information officer Sam Neville is joining the Highland Marketing advisory board. Sam brings a passion for nursing and safety to the board, which debates the big issues...

AI Model Identifies Potential Risk Genes…

Researchers from the Cleveland Clinic Genome Center have successfully applied advanced artificial intelligence (AI) genetics models to Parkinson's disease. Researchers identified genetic factors in progression and FDA-approved drugs that can...

AI Tool that may Assist Underserved Hosp…

As the fields of healthcare and technology increasingly evolve and intersect, researchers are collaborating on the best ways to use emerging technologies such as artificial intelligence (AI) to care for...