ChatGPT may be Better than Doctors at Evidence-Based Management of Clinical Depression

ChatGPT, the AI language model capable of mirroring human conversation, may be better than a doctor at following recognised treatment standards for clinical depression, and without any of the gender or social class biases sometimes seen in the primary care doctor-patient relationship, finds research published in the open access journal Family Medicine and Community Health.

However, further research is needed into how well this technology might manage severe cases as well as potential risks and ethical issues arising from its use, say the researchers.

Depression is very common, and many of those affected turn first to their family (primary care) doctors for help. The recommended course of treatment should largely be guided by evidence-based clinical guidelines, which usually suggest a tiered approach to care, in line with the severity of the depression.

ChatGPT has the potential to offer fast, objective, data-derived insights that can supplement traditional diagnostic methods as well as providing confidentiality and anonymity, say the researchers.

They therefore wanted to find out how the technology evaluated the recommended therapeutic approach for mild and severe major depression and whether this was influenced by gender or social class biases, when compared with 1249 French primary care doctors (73% women).

They drew on carefully designed and previously validated vignettes, centering around patients with symptoms of sadness, sleep problems, and loss of appetite during the preceding 3 weeks and a diagnosis of mild to moderate depression.

Eight versions of these vignettes were developed with different variations of patient characteristics, such as gender, social class, and depression severity. Each vignette was repeated 10 times for ChatGPT versions 3.5 and 4.

For each of the 8 vignettes, ChatGPT was asked: ‘What do you think a primary care physician should suggest in this situation?’ The possible responses were: watchful waiting; referral for psychotherapy; prescribed drugs (for depression/anxiety/sleep problems); referral for psychotherapy plus prescribed drugs; none of these.

Only just over 4% of family doctors exclusively recommended referral for psychotherapy for mild cases in line with clinical guidance, compared with ChatGPT-3.5 and ChatGPT-4, which selected this option in 95% and 97.5% of cases, respectively.

Most of the medical practitioners proposed either drug treatment exclusively (48%) or psychotherapy plus prescribed drugs (32.5%).

In severe cases, most of the doctors recommended psychotherapy plus prescribed drugs (44.5%). ChatGPT proposed this more frequently than the doctors (72% ChatGPT 3.5; 100% ChatGPT 4 in line with clinical guidelines). Four out of 10 of the doctors proposed prescribed drugs exclusively, which neither ChatGPT version recommended.

When medication was recommended, the AI and human participants were asked to specify which types of drugs they would prescribe.

The doctors recommended a combination of antidepressants and anti-anxiety drugs and sleeping pills in 67.5% of cases, exclusive use of antidepressants in 18%, and exclusive use of anti-anxiety and sleeping pills in 14%.

ChatGPT was more likely than the doctors to recommend antidepressants exclusively: 74%, version 3.5; and 68%, version 4. ChatGPT-3.5 (26%) and ChatGPT-4 (32%) also suggested using a combination of antidepressants and anti-anxiety drugs and sleeping pills more frequently than did the doctors.

But unlike the findings of previously published research, ChatGPT didn’t exhibit any gender or social class biases in its recommended treatment.

The researchers acknowledge that the study was limited to iterations of ChatGPT-3 and ChatGPT-4 at specific points in time and that the ChatGPT data were compared with data from a representative sample of primary care doctors from France, so might not be more widely applicable.

Lastly, the cases described in the vignettes were for an initial visit due to a complaint of depression, so didn't represent ongoing treatment of the disease or other variables that the doctor would know about the patient.

"ChatGPT-4 demonstrated greater precision in adjusting treatment to comply with clinical guidelines. Furthermore, no discernible biases related to gender and [socioeconomic status] were detected in the ChatGPT systems," highlight the researchers.

But there are ethical issues to consider, particularly around ensuring data privacy and security which are supremely important, considering the sensitive nature of mental health data, they point out, adding that AI shouldn’t ever be a substitute for human clinical judgement in the diagnosis or treatment of depression.

Nevertheless, they conclude: "The study suggests that ChatGPT... has the potential to enhance decision making in primary healthcare."

"However, it underlines the need for ongoing research to verify the dependability of its suggestions. Implementing such AI systems could bolster the quality and impartiality of mental health services."

Levkovich I, Elyoseph Z.
Identifying depression and its determinants upon initiating treatment: ChatGPT versus primary care physicians.
Fam Med Community Health. 2023 Sep;11(4):e002391. doi: 10.1136/fmch-2023-002391

Most Popular Now

Mobile Phone Data Helps Track Pathogen S…

A new way to map the spread and evolution of pathogens, and their responses to vaccines and antibiotics, will provide key insights to help predict and prevent future outbreaks. The...

AI Model to Improve Patient Response to …

A new artificial intelligence (AI) tool that can help to select the most suitable treatment for cancer patients has been developed by researchers at The Australian National University (ANU). DeepPT, developed...

Can AI Tell you if You Have Osteoporosis…

Osteoporosis is so difficult to detect in early stage it’s called the "silent disease." What if artificial intelligence could help predict a patient’s chances of having the bone-loss disease before...

Study Reveals Why AI Models that Analyze…

Artificial intelligence (AI) models often play a role in medical diagnoses, especially when it comes to analyzing images such as X-rays. However, studies have found that these models don’t always...

Think You're Funny? ChatGPT might b…

A study comparing jokes by people versus those told by ChatGPT shows that humans need to work on their material. The research team behind the study published on Wednesday, July 3...

Innovative, Highly Accurate AI Model can…

If there is one medical exam that everyone in the world has taken, it's a chest x-ray. Clinicians can use radiographs to tell if someone has tuberculosis, lung cancer, or...

New AI Approach Optimizes Antibody Drugs

Proteins have evolved to excel at everything from contracting muscles to digesting food to recognizing viruses. To engineer better proteins, including antibodies, scientists often iteratively mutate the amino acids -...

AI Speeds Up Heart Scans, Saving Doctors…

Researchers have developed a groundbreaking method for analysing heart MRI scans with the help of artificial intelligence (AI), which could save valuable NHS time and resources, as well as improve...

Researchers Customize AI Tools for Digit…

Scientists from Weill Cornell Medicine and the Dana-Farber Cancer Institute in Boston have developed and tested new artificial intelligence (AI) tools tailored to digital pathology - a rapidly growing field...

Young People Believe that AI is a Valuab…

Children and young people are generally positive about artificial intelligence (AI) and think it should be used in modern healthcare, finds the first-of-its-kind survey led by UCL and Great Ormond...