ChatGPT: Revolutionizing Clinical Decision Making with Artificial Intelligence

A recent study conducted by Mass General Brigham has demonstrated the impressive accuracy of ChatGPT, an artificial intelligence (AI) chatbot, in clinical decision making. The study revealed that ChatGPT achieved an overall accuracy rate of approximately 72% in various aspects of clinical care, ranging from initial diagnoses to final care management decisions. The findings, published in the Journal of Medical Internet Research, highlight the potential of large-language model (LLM) AI systems in supporting medical professionals’ decision-making processes.

Led by corresponding author Marc Succi, MD, the research team conducted an extensive assessment of ChatGPT’s capabilities. They found that the AI chatbot performed equally well in primary care and emergency settings across all medical specialties. In fact, the study estimated that ChatGPT’s performance was akin to that of a medical school graduate, such as an intern or resident. This suggests that LLMs have the potential to significantly enhance the practice of medicine by providing augmenting tools for clinical decision making.

Artificial intelligence has rapidly transformed various sectors, and healthcare is no exception. However, the extent to which LLMs can assist in clinical care has not been extensively studied until now. In this comprehensive study, the team explored how ChatGPT could support clinical advisement and decision making throughout the entire patient encounter. The researchers hypothesized that ChatGPT could recommend diagnostic workups, determine management courses, and ultimately reach a final diagnosis.

To test this, the team inputted standardized clinical vignettes into ChatGPT, allowing the tool to generate possible diagnoses, make management decisions, and provide a final diagnosis. The AI chatbot achieved an overall accuracy of 72%, with its highest accuracy observed in making final diagnoses (77%). However, it performed less accurately in differential diagnoses (60%) and clinical management decisions (68%).

One notable finding was that ChatGPT’s responses did not demonstrate gender bias, and its performance remained consistent in both primary and emergency care scenarios. While the study has showcased the potential of AI tools like ChatGPT, the authors emphasize the need for further research and regulatory guidance before integrating such tools into clinical care.


Q: What is ChatGPT?
A: ChatGPT is an artificial intelligence chatbot that shows impressive accuracy in clinical decision making.

Q: What was the accuracy rate of ChatGPT in the study?
A: The study found that ChatGPT achieved an overall accuracy rate of approximately 72% in clinical decision making.

Q: How well did ChatGPT perform in different aspects of clinical care?
A: ChatGPT performed best in making final diagnoses (77% accuracy), while its accuracy was lower in differential diagnoses (60%) and clinical management decisions (68%).

Q: Did ChatGPT show any gender bias in its responses?
A: No, the study revealed that ChatGPT’s responses did not demonstrate any gender bias.

Q: Can AI tools like ChatGPT be integrated into clinical care?
A: Before integration, further benchmark research and regulatory guidance are necessary to ensure their safe and effective use in clinical settings.