Welcome to PsychiatryAI.com: [PubMed] - Psychiatry AI Latest

ChatGPT as a Tool for Medical Education and Clinical Decision-Making on the Wards: Case Study

Evidence

JMIR Form Res. 2024 May 8;8:e51346. doi: 10.2196/51346.

ABSTRACT

BACKGROUND: Large language models (LLMs) are computational artificial intelligence systems with advanced natural language processing capabilities that have recently been popularized among health care students and educators due to their ability to provide real-time access to a vast amount of medical knowledge. The adoption of LLM technology into medical education and training has varied, and little empirical evidence exists to support its use in clinical teaching environments.

OBJECTIVE: The aim of the study is to identify and qualitatively evaluate potential use cases and limitations of LLM technology for real-time ward-based educational contexts.

METHODS: A brief, single-site exploratory evaluation of the publicly available ChatGPT-3.5 (OpenAI) was conducted by implementing the tool into the daily attending rounds of a general internal medicine inpatient service at a large urban academic medical center. ChatGPT was integrated into rounds via both structured and organic use, using the web-based “chatbot” style interface to interact with the LLM through conversational free-text and discrete queries. A qualitative approach using phenomenological inquiry was used to identify key insights related to the use of ChatGPT through analysis of ChatGPT conversation logs and associated shorthand notes from the clinical sessions.

RESULTS: Identified use cases for ChatGPT integration included addressing medical knowledge gaps through discrete medical knowledge inquiries, building differential diagnoses and engaging dual-process thinking, challenging medical axioms, using cognitive aids to support acute care decision-making, and improving complex care management by facilitating conversations with subspecialties. Potential additional uses included engaging in difficult conversations with patients, exploring ethical challenges and general medical ethics teaching, personal continuing medical education resources, developing ward-based teaching tools, supporting and automating clinical documentation, and supporting productivity and task management. LLM biases, misinformation, ethics, and health equity were identified as areas of concern and potential limitations to clinical and training use. A code of conduct on ethical and appropriate use was also developed to guide team usage on the wards.

CONCLUSIONS: Overall, ChatGPT offers a novel tool to enhance ward-based learning through rapid information querying, second-order content exploration, and engaged team discussion regarding generated responses. More research is needed to fully understand contexts for educational use, particularly regarding the risks and limitations of the tool in clinical settings and its impacts on trainee development.

PMID:38717811 | DOI:10.2196/51346

Document this CPD Copy URL Button

Google

Google Keep Add to Google Keep

LinkedIn Share Share on Linkedin Share on Linkedin

Estimated reading time: 7 minute(s)

Latest: Psychiatryai.com #RAISR4D

Cool Evidence: Engaging Young People and Students in Real-World Evidence ☀️

Real-Time Evidence Search [Psychiatry]

AI Research [Andisearch.com]

ChatGPT as a Tool for Medical Education and Clinical Decision-Making on the Wards: Case Study

Copy WordPress Title

🌐 90 Days

Evidence Blueprint

ChatGPT as a Tool for Medical Education and Clinical Decision-Making on the Wards: Case Study

QR Code

☊ AI-Driven Related Evidence Nodes

(recent articles with at least 5 words in title)

More Evidence

ChatGPT as a Tool for Medical Education and Clinical Decision-Making on the Wards: Case Study

🌐 365 Days

Floating Tab
close chatgpt icon
ChatGPT

Enter your request.