Stanford Medicine Study Suggests Physician's Medical Decisions Benefit from Chatbot

Artificial intelligence-powered chatbots are getting pretty good at diagnosing some diseases, even when they are complex. But how do chatbots do when guiding treatment and care after the diagnosis? For example, how long before surgery should a patient stop taking prescribed blood thinners? Should a patient's treatment protocol change if they've had adverse reactions to similar drugs in the past? These sorts of questions don't have a textbook right or wrong answer - it's up to physicians to use their judgment.

Jonathan H. Chen, MD, PhD, assistant professor of medicine, and a team of researchers are exploring whether chatbots, a type of large language model, or LLM, can effectively answer such nuanced questions, and whether physicians supported by chatbots perform better.

The answers, it turns out, are yes and yes. The research team tested how a chatbot performed when faced with a variety of clinical crossroads. A chatbot on its own outperformed doctors who could access only an internet search and medical references, but armed with their own LLM, the doctors, from multiple regions and institutions across the United States, kept up with the chatbots.

"For years I've said that, when combined, human plus computer is going to do better than either one by itself," Chen said. "I think this study challenges us to think about that more critically and ask ourselves, 'What is a computer good at? What is a human good at?' We may need to rethink where we use and combine those skills and for which tasks we recruit AI."

A study detailing these results published in Nature Medicine on Feb. 5. Chen and Adam Rodman, MD, assistant professor at Harvard University, are co-senior authors. Postdoctoral scholars Ethan Goh, MD, and Robert Gallo, MD, are co-lead author.

In October 2024, Chen and Goh led a team that ran a study, published in JAMA Network Open, that tested how the chatbot performed when diagnosing diseases and that found its accuracy was higher than that of doctors, even if they were using a chatbot. The current paper digs into the squishier side of medicine, evaluating chatbot and physician performance on questions that fall into a category called "clinical management reasoning."

Goh explains the difference like this: Imagine you’re using a map app on your phone to guide you to a certain destination. Using an LLM to diagnose a disease is sort of like using the map to pinpoint the correct location. How you get there is the management reasoning part - do you take backroads because there’s traffic? Stay the course, bumper to bumper? Or wait and hope the roads clear up?

In a medical context, these decisions can get tricky. Say a doctor incidentally discovers a hospitalized patient has a sizeable mass in the upper part of the lung. What would the next steps be? The doctor (or chatbot) should recognize that a large nodule in the upper lobe of the lung statistically has a high chance of spreading throughout the body. The doctor could immediately take a biopsy of the mass, schedule the procedure for a later date or order imaging to try to learn more.

Determining which approach is best suited for the patient comes down to a host of details, starting with the patient’s known preferences. Are they reticent to undergo an invasive procedure? Does the patient’s history show a lack of following up on appointments? Is the hospital’s health system reliable when organizing follow-up appointments? What about referrals? These types of contextual factors are crucial to consider, Chen said.

The team designed a trial to study clinical management reasoning performance in three groups: the chatbot alone, 46 doctors with chatbot support, and 46 doctors with access only to internet search and medical references. They selected five de-identified patient cases and gave them to the chatbot and to the doctors, all of whom provided a written response that detailed what they would do in each case, why and what they considered when making the decision.

In addition, the researchers tapped a group of board-certified doctors to create a rubric that would qualify a medical judgment or decision as appropriately assessed. The decisions were then scored against the rubric.

To the team's surprise, the chatbot outperformed the doctors who had access only to the internet and medical references, ticking more items on the rubric than the doctors did. But the doctors who were paired with a chatbot performed as well as the chatbot alone.

Exactly what gave the physician-chatbot collaboration a boost is up for debate. Does using the LLM force doctors to be more thoughtful about the case? Or is the LLM providing guidance that the doctors wouldn't have thought of on their own? It's a future direction of exploration, Chen said.

The positive outcomes for chatbots and physicians paired with chatbots beg an ever-popular question: Are AI doctors on their way?

"Perhaps it's a point in AI’s favor," Chen said. But rather than replacing physicians, the results suggest that doctors might want to welcome a chatbot assist. "This doesn't mean patients should skip the doctor and go straight to chatbots. Don't do that," he said. "There's a lot of good information out there, but there's also bad information. The skill we all have to develop is discerning what's credible and what's not right. That's more important now than ever."

Researchers from VA Palo Alto Health Care System, Beth Israel Deaconess Medical Center, Harvard University, University of Minnesota, University of Virginia, Microsoft and Kaiser contributed to this work.

The study was funded by the Gordon and Betty Moore Foundation, the Stanford Clinical Excellence Research Center and the VA Advanced Fellowship in Medical Informatics.

Stanford's Department of Medicine also supported the work.

Goh E, Gallo RJ, Strong E, Weng Y, Kerman H, Freed JA, Cool JA, Kanjee Z, Lane KP, Parsons AS, Ahuja N, Horvitz E, Yang D, Milstein A, Olson APJ, Hom J, Chen JH, Rodman A.
GPT-4 assistance for improvement of physician performance on patient care tasks: a randomized controlled trial.
Nat Med. 2025 Feb 5. doi: 10.1038/s41591-024-03456-y

Most Popular Now

Unlocking the 10 Year Health Plan

The government's plan for the NHS is a huge document. Jane Stephenson, chief executive of SPARK TSL, argues the key to unlocking its digital ambitions is to consider what it...

Alcidion Grows Top Talent in the UK, wit…

Alcidion has today announced the addition of three new appointments to their UK-based team, with one internal promotion and two external recruits. Dr Paul Deffley has been announced as the...

AI can Find Cancer Pathologists Miss

Men assessed as healthy after a pathologist analyses their tissue sample may still have an early form of prostate cancer. Using AI, researchers at Uppsala University have been able to...

AI, Full Automation could Expand Artific…

Automated insulin delivery (AID) systems such as the UVA Health-developed artificial pancreas could help more type 1 diabetes patients if the devices become fully automated, according to a new review...

How AI could Speed the Development of RN…

Using artificial intelligence (AI), MIT researchers have come up with a new way to design nanoparticles that can more efficiently deliver RNA vaccines and other types of RNA therapies. After training...

MIT Researchers Use Generative AI to Des…

With help from artificial intelligence, MIT researchers have designed novel antibiotics that can combat two hard-to-treat infections: drug-resistant Neisseria gonorrhoeae and multi-drug-resistant Staphylococcus aureus (MRSA). Using generative AI algorithms, the research...

New Training Year Starts at Siemens Heal…

In September, 197 school graduates will start their vocational training or dual studies in Germany at Siemens Healthineers. 117 apprentices and 80 dual students will begin their careers at Siemens...

AI Hybrid Strategy Improves Mammogram In…

A hybrid reading strategy for screening mammography, developed by Dutch researchers and deployed retrospectively to more than 40,000 exams, reduced radiologist workload by 38% without changing recall or cancer detection...

Penn Developed AI Tools and Datasets Hel…

Doctors treating kidney disease have long depended on trial-and-error to find the best therapies for individual patients. Now, new artificial intelligence (AI) tools developed by researchers in the Perelman School...

Are You Eligible for a Clinical Trial? C…

A new study in the academic journal Machine Learning: Health discovers that ChatGPT can accelerate patient screening for clinical trials, showing promise in reducing delays and improving trial success rates. Researchers...

New AI Tool Addresses Accuracy and Fairn…

A team of researchers at the Icahn School of Medicine at Mount Sinai has developed a new method to identify and reduce biases in datasets used to train machine-learning algorithms...

Global Study Reveals How Patients View M…

How physicians feel about artificial intelligence (AI) in medicine has been studied many times. But what do patients think? A team led by researchers at the Technical University of Munich...