Wednesday, May 14, 2025
Science
No Result
View All Result
  • Login
  • HOME
  • SCIENCE NEWS
  • CONTACT US
  • HOME
  • SCIENCE NEWS
  • CONTACT US
No Result
View All Result
Scienmag
No Result
View All Result
Home Science News Technology and Engineering

Study shows ChatGPT failed when challenging ESCMID guideline for treating brain abscesses

April 28, 2024
in Technology and Engineering
Reading Time: 5 mins read
0
Study shows ChatGPT failed when challenging ESCMID guideline for treating brain abscesses
67
SHARES
611
VIEWS
Share on FacebookShare on Twitter

With artificial intelligence (AI) poised to become a fundamental part of clinical research and decision making, many still question the accuracy of ChatGPT, a sophisticated AI language model, to support complex diagnostic and treatment processes.

With artificial intelligence (AI) poised to become a fundamental part of clinical research and decision making, many still question the accuracy of ChatGPT, a sophisticated AI language model, to support complex diagnostic and treatment processes.

Now a new study, being presented at this year’s ESCMID Global Congress (formerly ECCMID) in Barcelona, Spain (27-30 April), which pitted ChatGPT against the ESCMID guideline for the management of brain abscesses, found that while ChatGPT seems able to give recommendations on key questions about diagnosis and treatment in most cases, some of the AI model’s responses could put patients at risk.

The study was conducted by members of the ESCMID Study Group for Infectious Diseases of the Brain (ESGIB), and is published in The Journal of Neurology.

“Anything less than 100% is a failure when you’re dealing with patient safety”, says author Dr Susanne Dyckhoff-Shen from LMU University Hospital Munich in Germany and a member of ESCMID. “While we are amazed by ChatGPT’s knowledge on the management of brain abscesses, there are some key limitations when it comes to using the AI model as a medical device, including potential patient harm and the lack of transparency about which data are used to provide responses.”

The ability of AI to rapidly assimilate, process, and interpret vast data sets offers tantalising prospects. But are time-consuming processes to create medical guidelines still necessary, or could AI models trained on a wealth of scientific medical literature rival clinical experts in answering complex clinical questions?

Brain abscesses are a potentially life-threatening central nervous system (CNS) infection that require immediate identification and treatment to prevent severe neurological complications and even death.

Historically, the management of brain abscesses has been largely guided by clinical experience and limited studies, but in 2023 ESCMID fulfilled the need for a standardised approach by developing an international guideline [1].

To find out whether ChatGPT is able to professionally evaluate medical research and give scientifically valid recommendations, a European team of researchers tested the AI model to see whether it could accurately provide answers to 10 key questions on brain abscess diagnostics and treatment in comparison to the ESCMID guideline.

First, the researchers asked ChatGPT (version 4) to answer 10 questions that had been developed and appraised by the ESCMID committee for their brain abscess guideline without any additional information.

Then, ChatGPT was additionally primed with the text of the same scientific research articles that were used to develop the guideline before asking the same questions. This was done to see if ChatGPT could provide more aligned recommendations when given the same data used for guideline development.

The AI-generated responses were then compared to the recommendations of the ESCMID guideline by three independent infectious CNS disease experts for their clarity, alignment with the guideline, and patient risk.

Clear responses to most key questions

The researchers found that overall, for 17 out of 20 questions asked (with and without data input), ChatGPT’s responses were clear on the management of patients with brain abscess, including grade of evidence and strength of recommendation, with clarity assessed at 80-90% (see link to poster in notes to editors).

However, the AI model did not provide clear enough answers to guide physicians on treatment decisions on withholding microbials until surgery and prophylactic antiepileptic treatment (questions 2 and 10).

More incorrect advice and risk to patients with data prompting

Without additional data input, ChatGPT’s responses to 70% (7/10) of questions were very similar to the guideline recommendation. However, the AI model failed to come up with the correct advice on three questions relating to withholding microbials, consolidation therapy, and prophylactic antiepileptic treatment (questions 2, 8, and 10). Importantly, however, these incorrect responses would not have harmed patients (see link to poster in notes to editors).

Surprisingly, data input resulted in fewer correct answers (40%) including two recommendations that directly contradicted the guideline, that could have put patients at risk.

Question 6 about duration of antimicrobial therapy for bacterial brain abscess was answered by ChatGPT after data entry as “intravenous administration for about 4 weeks, followed by 12 weeks oral medication”, but the ESCMID guideline recommends “a total duration of 6-8 weeks of intravenous antimicrobials…”.

For question 7 about early transition to oral antimicrobials, after data input ChatGPT recommended that “an early switch to oral antibiotics during the first 14 days of treatment…seems to be associated with favorable outcomes in selected patients”. However, the ESCMID guideline committee judged that there was insufficient evidence to provide a recommendation for this question.

In both cases, following ChatGPT’s advice could have potentially led to patient harm.

“The fact that ChatGPT’s recommendations were inferior after data entry might be due to an overvaluation of the few observational studies provided for key questions 6 and 7. For one of those, even the guideline committee was not able to give a recommendation as the evidence was insufficient to answer the question”, says senior author Professor Mattias Klein from LMU University Hospital Munich in Germany and a member of the ESCMID committee which established the guideline.

“As the exact operating procedures of ChatGPT remain unclear, we speculate that while the AI model can process large amounts of data quickly, it may lack the ability to correctly classify and weigh the data based on their scientific quality. Moreover, it remains unclear which data are used for ChatGPT’s responses as it does not disclose the sources of its answers, which risks dubious literature being used.”

Dr Dyckhoff-Shen adds, “It is alarming to think that patients could have come to harm if ChatGPT’s advice on two key questions had been followed. The nuanced expertise of expert committees remains essential, especially to answer complex clinical queries. Blindly relying on AI could put patients at risk.”

Nevertheless, the authors note that ChatGPT’s knowledge was from before September 2021 and the questions in the study covered some extremely complex medical issues, some of which are controversial even among experts and for which hardly any robust data are available. However, even when primed with the same research articles that were used to develop the ESCMID guideline, ChatGPT’s advice aligned even less with the guideline. They recommend that the quality of ChatGPT should be reviewed on an ongoing basis following its evolution and further development.

The authors explain further that ChatGPT, like many AI models, has a cutoff date for the information it can access. For the current version as of today, it was last trained on data up until January 2022. “This means that while it can provide responses based on a wide range of information, it doesn’t have access to real-time data or events occurring after that date,” says Dr Dyckhoff-Shen. “When we used ChatGPT for our study before August 2023, it was only trained on data up until September 2021. There was no possibility for us to get a more up-to-date version at that time because ChatGPT was not trained further yet. This is also the reason why we used a second approach by prompting ChatGPT with relevant scientific articles that were used by the ESGIB group to give recommendations in the ESCMID guideline so that we sort of ‘manually’ tried to get it more up-to-date.

“In the future, it would be interesting to re-assess ChatGPT’s knowledge in the future after internal optimisation processes. However, once the chatbot has access to the ESCMID guideline itself, it could just use the recommendations from the guideline thus rendering a comparison no longer useful.”

 



Journal

Journal of Neurology

COI Statement

The authors declare no conflicts of interest.

Share27Tweet17
Previous Post

The threat of polio paralysis may have disappeared, but enterovirus paralysis is just as dangerous and surveillance and testing systems are desperately needed

Next Post

Study finds resistance to critically important antibiotics in uncooked meat sold for human and animal consumption

Related Posts

blank
Medicine

Relativistic Ionized Winds Blast from Quasar

May 14, 2025
blank
Technology and Engineering

Wireless Dynamic Strain Monitoring for High-Speed Rail

May 14, 2025
blank
Technology and Engineering

Genetic and Environmental Roots of Norway’s Socioeconomic Status

May 14, 2025
Treated v. untreated human neurons
Technology and Engineering

Revolutionary Sugar-Coated Nanotherapy Significantly Enhances Neuron Survival in Alzheimer’s Model

May 14, 2025
blank
Technology and Engineering

Room-Temperature Lanthanide Halides Boost X-Ray Imaging

May 14, 2025
blank
Technology and Engineering

Ozone Exposure Linked to Kawasaki Disease in Chinese Kids

May 13, 2025
Next Post
Study finds resistance to critically important antibiotics in uncooked meat sold for human and animal consumption

Study finds resistance to critically important antibiotics in uncooked meat sold for human and animal consumption

  • Mothers who receive childcare support from maternal grandparents show more parental warmth, finds NTU Singapore study

    Mothers who receive childcare support from maternal grandparents show more parental warmth, finds NTU Singapore study

    27494 shares
    Share 10994 Tweet 6872
  • Bee body mass, pathogens and local climate influence heat tolerance

    636 shares
    Share 254 Tweet 159
  • Researchers record first-ever images and data of a shark experiencing a boat strike

    497 shares
    Share 199 Tweet 124
  • Warm seawater speeding up melting of ‘Doomsday Glacier,’ scientists warn

    304 shares
    Share 122 Tweet 76
  • Probiotics during pregnancy shown to help moms and babies

    251 shares
    Share 100 Tweet 63
Science

Embark on a thrilling journey of discovery with Scienmag.com—your ultimate source for cutting-edge breakthroughs. Immerse yourself in a world where curiosity knows no limits and tomorrow’s possibilities become today’s reality!

Recent Posts

  • Pioneering Discovery of FLASH Effect Conditions with Compact Carbon Ion Synchrotron Paves Way for Safer Cancer Treatments
  • Low-Viscosity Oil Enhances PDMS SlipChip for Safer Cell Research and Precise Gradient Formation
  • NASA Selects UTA to Develop Advanced Wildfire Smoke Warning System
  • Decoding the Genome of the Northern White Rhino: A Beacon of Hope for Species Revival

Categories

  • Agriculture
  • Anthropology
  • Archaeology
  • Athmospheric
  • Biology
  • Bussines
  • Cancer
  • Chemistry
  • Climate
  • Earth Science
  • Marine
  • Mathematics
  • Medicine
  • Pediatry
  • Policy
  • Psychology & Psychiatry
  • Science Education
  • Social Science
  • Space
  • Technology and Engineering

Subscribe to Blog via Email

Enter your email address to subscribe to this blog and receive notifications of new posts by email.

Join 4,862 other subscribers

© 2025 Scienmag - Science Magazine

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • HOME
  • SCIENCE NEWS
  • CONTACT US

© 2025 Scienmag - Science Magazine