Search Articles

View query in Help articles search

Search Results (1 to 10 of 333 Results)

Download search results: CSV END BibTex RIS


Evaluation of ChatGPT-4 as an Online Outpatient Assistant in Puerperal Mastitis Management: Content Analysis of an Observational Study

Evaluation of ChatGPT-4 as an Online Outpatient Assistant in Puerperal Mastitis Management: Content Analysis of an Observational Study

Chat GPT-4 was selected as the AI model for evaluation due to its widespread adoption in recent health-related studies and its extensive documentation in the current literature. Using Chat GPT-4 allowed for comparability with prior research, ensuring consistency and alignment with similar investigations examining large language models’ (LLMs’) performance in medical information delivery.

Fatih Dolu, Oğuzhan Fatih Ay, Aydın Hakan Kupeli, Enes Karademir, Muhammed Huseyin Büyükavcı

JMIR Med Inform 2025;13:e68980

Improving Large Language Models’ Summarization Accuracy by Adding Highlights to Discharge Notes: Comparative Evaluation

Improving Large Language Models’ Summarization Accuracy by Adding Highlights to Discharge Notes: Comparative Evaluation

In the study by Ma et al [39], the researchers developed a method called Impression GPT to summarize the “impression” section of radiology reports using Chat GPT. They used a dynamic prompt generation and iterative optimization approach to improve the performance of Chat GPT in this task.

Mahshad Koohi Habibi Dehkordi, Yehoshua Perl, Fadi P Deek, Zhe He, Vipina K Keloth, Hao Liu, Gai Elhanan, Andrew J Einstein

JMIR Med Inform 2025;13:e66476

Clinical Performance and Communication Skills of ChatGPT Versus Physicians in Emergency Medicine: Simulated Patient Study

Clinical Performance and Communication Skills of ChatGPT Versus Physicians in Emergency Medicine: Simulated Patient Study

Chat GPT (version 4.0) was used in this study, with all examinations conducted between September and October 2023. For Chat GPT, a prompt designed to emulate the role of a medical professional was provided (Table S2 in Multimedia Appendix 2). Chat GPT was instructed to perform history taking and, when a physical examination was required, to obtain relevant information through targeted questions.

ChulHyoung Park, Min Ho An, Gyubeom Hwang, Rae Woong Park, Juho An

JMIR Med Inform 2025;13:e68409

Evaluating the Readability of Pediatric Neurocutaneous Syndromes–Related Patient Education Material Created by a Custom GPT With Retrieval Augmentation

Evaluating the Readability of Pediatric Neurocutaneous Syndromes–Related Patient Education Material Created by a Custom GPT With Retrieval Augmentation

This study aims to evaluate a GPT assistant’s ability to provide readable patient information on pediatric neurocutaneous syndromes in comparison to Chat GPT-4. A GPT assistant was developed by using Python and Open AI’s application program interface (API; Figure 1). It was not programmed to answer questions at a specific reading level.

Nneka Ede, Robyn Okereke

JMIR Dermatol 2025;8:e59054

Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

Several studies have shown that Chat GPT provides appropriate, accurate, and reliable knowledge across a wide range of cardiac and noncardiac medical conditions, including heart failure [11-16]. In addition to accuracy, Chat GPT has been found to deliver more empathetic responses to real-world patient questions than physicians in online forums [17]. As prior data regarding accuracy have been promising, an emerging focus has been on investigating the readability of the model’s output.

Ryan C King, Jamil S Samaan, Joseph Haquang, Vishnu Bharani, Samuel Margolis, Nitin Srinivasan, Yuxin Peng, Yee Hui Yeo, Roxana Ghashghaei

JMIR Cardio 2025;9:e68817

Assessing the Accuracy of ChatGPT in Appropriately Triaging Common Postoperative Concerns Regarding Mohs Micrographic Surgery

Assessing the Accuracy of ChatGPT in Appropriately Triaging Common Postoperative Concerns Regarding Mohs Micrographic Surgery

As of April 2024, a pilot program in Louisiana incorporated Chat GPT-4.0 into electronic health record (EHR) messaging to generate preliminary responses that clinicians subsequently reviewed for validity [3]. Despite Chat GPT-4.0’s advances, the study demonstrated that human oversight in AI-generated communication remains essential [3]. Such initiatives demonstrate AI’s potential to reduce administrative workload, but they also underscore its role in improving patient education.

Chloe Fernandez, Victoria Dukharan, Nathaniel A Marroquin, Rebecca Bolen, Adam Leavitt, Nicole C Cabbad

JMIR Dermatol 2025;8:e72706

Co-Design of a Health Screening Program Fact Sheet by People Experiencing Homelessness and ChatGPT: Focus Group Study

Co-Design of a Health Screening Program Fact Sheet by People Experiencing Homelessness and ChatGPT: Focus Group Study

In this research project, we aimed to co-design an awareness-raising fact sheet for an oral cancer screening program with people experiencing homelessness as experts by experience and Chat GPT. The latter was used to present textual alternatives for this health information piece, so we could also test the usability of Chat GPT in designing adequate information materials serving the needs of people experiencing homelessness.

Nóra Radó, Orsolya Németh, Sándor Békási

JMIR Form Res 2025;9:e68316

Evaluating a Large Language Model’s Ability to Synthesize a Health Science Master’s Thesis: Case Study

Evaluating a Large Language Model’s Ability to Synthesize a Health Science Master’s Thesis: Case Study

In an exploratory case study, we asked Chat GPT (Chat GPT-4o and Chat GPT-o1) to synthesize two datasets: one qualitative and one quantitative, inspired by existing real-world datasets the authors had previously analyzed. The generation of synthetic datasets and subsequent data analysis took place in October and November of 2024.

Pål Joranger, Sara Rivenes Lafontan, Asgeir Brevik

JMIR Form Res 2025;9:e73248

Exploring Inflammatory Bowel Disease Discourse on Reddit Throughout the COVID-19 Pandemic Using OpenAI’s GPT-3.5 Turbo Model: Classification Model Validation and Case Study

Exploring Inflammatory Bowel Disease Discourse on Reddit Throughout the COVID-19 Pandemic Using OpenAI’s GPT-3.5 Turbo Model: Classification Model Validation and Case Study

Reference 13: ChatGPT outperforms humans in emotional awareness evaluations Reference 15: Evaluation of ChatGPT for NLP-based mental health applications(https://arxiv.org/abs/2303.15727 Reference 16: Bias in emotion recognition with ChatGPT(https://arxiv.org/abs/2310.11753) Reference 19: Human vs. machine: a comparative analysis of qualitative coding by humans and ChatGPT-4chatgptGenerative Language Models Including ChatGPT

Tyler Babinski, Sara Karley, Marita Cooper, Salma Shaik, Y Ken Wang

J Med Internet Res 2025;27:e53332

Evaluating the Quality of Psychotherapy Conversational Agents: Framework Development and Cross-Sectional Study

Evaluating the Quality of Psychotherapy Conversational Agents: Framework Development and Cross-Sectional Study

Custom GPTs are derivations of the baseline Chat GPT model (at the time of evaluation: GPT-4o) developed by Open AI that have been modified by members of the public with customized instructions and behavior for specific applications (eg, psychotherapy chatbot). In May 2024, we indexed both sites using the search feature to emulate what an end-user may experience with the following search terms: “therapy,” “anxiety,” “depression,” “mental health,” “therapist,” and “psychologist.”

Kunmi Sobowale, Daniel Kevin Humphrey

JMIR Form Res 2025;9:e65605