Search Articles

View query in Help articles search

Search Results (1 to 10 of 287 Results)

Download search results: CSV END BibTex RIS


Large Language Models for Thematic Summarization in Qualitative Health Care Research: Comparative Analysis of Model and Human Performance

Large Language Models for Thematic Summarization in Qualitative Health Care Research: Comparative Analysis of Model and Human Performance

A novel 7-point scale was developed following a pilot test conducted by two of the authors to address the complexities of comparing codes generated by humans and Chat GPT. This scale, presented in the first column of Table 2, focuses on exploring the complementary and divergent insights between human-generated and Chat GPT-generated codes.

Arturo Castellanos, Haoqiang Jiang, Paulo Gomes, Debra Vander Meer, Alfred Castillo

JMIR AI 2025;4:e64447

Using a Hybrid of AI and Template-Based Method in Automatic Item Generation to Create Multiple-Choice Questions in Medical Education: Hybrid AIG

Using a Hybrid of AI and Template-Based Method in Automatic Item Generation to Create Multiple-Choice Questions in Medical Education: Hybrid AIG

Unlike the template-based method, this method uses the ability of artificial intelligence (AI) to generate content dynamically, for example, using Chat GPT, which is an AI-based chatbot developed by Open AI, for creating items based on specific topics or learning outcomes provided by users [17-21]. This approach allows for the generation of diverse and complex questions in seconds, offering flexibility and efficiency in item development.

Yavuz Selim Kıyak, Andrzej A Kononowicz

JMIR Form Res 2025;9:e65726

Large Language Models in Summarizing Radiology Report Impressions for Lung Cancer in Chinese: Evaluation Study

Large Language Models in Summarizing Radiology Report Impressions for Lung Cancer in Chinese: Evaluation Study

Recently, large language models (LLMs) such as Chat GPT [12] and GPT-4 [13] have captured worldwide attention due to their astonishing text-generation capabilities. Through pretraining on vast amounts of data, LLMs demonstrate remarkable performance on unseen downstream tasks using zero-shot, one-shot, or few-shot prompts without parameter updates [14].

Danqing Hu, Shanyuan Zhang, Qing Liu, Xiaofeng Zhu, Bing Liu

J Med Internet Res 2025;27:e65547

Authors’ Reply: Citation Accuracy Challenges Posed by Large Language Models

Authors’ Reply: Citation Accuracy Challenges Posed by Large Language Models

We appreciate the thoughtful critique of our manuscript “Perceptions and earliest experiences of medical students and faculty with Chat GPT in medical education: qualitative study” [1] by Zhao and Zhang [2]. Concerns over the generation of hallucinated citations by large language models (LLMs), such as Open AI’s Chat GPT, Google’s Gemini, and Hangzhou’s Deep Seek, warrant exploring advanced and novel methodologies to ensure citation accuracy and overall output integrity [3].

Mohamad-Hani Temsah, Ayman Al-Eyadhy, Amr Jamal, Khalid Alhasan, Khalid H Malki

JMIR Med Educ 2025;11:e73698

Citation Accuracy Challenges Posed by Large Language Models

Citation Accuracy Challenges Posed by Large Language Models

Large language models (LLMs) such as Deep Seek, Chat GPT, and Chat GLM have significant limitations in generating citations, raising concerns about the quality and reliability of academic research. These models tend to produce citations that are correctly formatted but fictional in content, misleading users and undermining academic rigor.

Manlin Zhang, Tianyu Zhao

JMIR Med Educ 2025;11:e72998

Authors’ Reply: The Importance of Comparing New Technologies (AI) to Existing Tools for Patient Education on Common Dermatologic Conditions: A Commentary

Authors’ Reply: The Importance of Comparing New Technologies (AI) to Existing Tools for Patient Education on Common Dermatologic Conditions: A Commentary

Juels Parker commented on our study comparing the sufficiency of Chat GPT, Google Bard, and Bing artificial intelligence (AI) in generating patient-facing responses to questions about five dermatological diagnoses [1,2]. He highlights an important need to compare AI to existing patient education tools, such as handouts, peer-reviewed articles, and patient-centered websites. We agree that AI is not a benign entity, and many resources exist for patients to learn about their conditions, aside from AI [3,4].

Courtney Chau, Hao Feng, Gabriela Cobos, Joyce Park

JMIR Dermatol 2025;8:e72540

The Importance of Comparing New Technologies (AI) to Existing Tools for Patient Education on Common Dermatologic Conditions: A Commentary

The Importance of Comparing New Technologies (AI) to Existing Tools for Patient Education on Common Dermatologic Conditions: A Commentary

Reference 1: The comparative sufficiency of ChatGPT, Google Bard, and Bing AI in answering diagnosis,chatgptTheme Issue (2023): AI and ChatGPT in Dermatology

Parker Juels

JMIR Dermatol 2025;8:e71768

Evaluating the Diagnostic Accuracy of ChatGPT-4 Omni and ChatGPT-4 Turbo in Identifying Melanoma: Comparative Study

Evaluating the Diagnostic Accuracy of ChatGPT-4 Omni and ChatGPT-4 Turbo in Identifying Melanoma: Comparative Study

There has a been rapid popularization of the LLM, Chat GPT for home-based medical inquiries [3]. Minimal research exists on Chat GPT’s accuracy in detecting melanoma. Given that patients are increasingly presenting internet-derived diagnostics during cancer consultations, it is imperative to understand the capabilities of commonly used AI engines, such as Chat GPT [4].

Samantha S. Sattler, Nitin Chetla, Matthew Chen, Tamer Rajai Hage, Joseph Chang, William Young Guo, Jeremy Hugh

JMIR Dermatol 2025;8:e67551

The Impact of ChatGPT Exposure on User Interactions With a Motivational Interviewing Chatbot: Quasi-Experimental Study

The Impact of ChatGPT Exposure on User Interactions With a Motivational Interviewing Chatbot: Quasi-Experimental Study

To determine the extent of exposure to Chat GPT, for each participant in MIBot (version 5.2 A), we included an additional short survey in the 1-week-later survey referred to as the Chat GPT survey. It contained 8 new questions designed to evaluate the participant’s knowledge and use of Chat GPT prior to engaging in MIBot (version 5.2 A). The full Chat GPT survey can be found in Multimedia Appendix 1.

Jiading Zhu, Alec Dong, Cindy Wang, Scott Veldhuizen, Mohamed Abdelwahab, Andrew Brown, Peter Selby, Jonathan Rose

JMIR Form Res 2025;9:e56973