Search Articles

View query in Help articles search

Search Results (1 to 10 of 329 Results)

Download search results: CSV END BibTex RIS


Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

Several studies have shown that Chat GPT provides appropriate, accurate, and reliable knowledge across a wide range of cardiac and noncardiac medical conditions, including heart failure [11-16]. In addition to accuracy, Chat GPT has been found to deliver more empathetic responses to real-world patient questions than physicians in online forums [17]. As prior data regarding accuracy have been promising, an emerging focus has been on investigating the readability of the model’s output.

Ryan C King, Jamil S Samaan, Joseph Haquang, Vishnu Bharani, Samuel Margolis, Nitin Srinivasan, Yuxin Peng, Yee Hui Yeo, Roxana Ghashghaei

JMIR Cardio 2025;9:e68817

Assessing the Accuracy of ChatGPT in Appropriately Triaging Common Postoperative Concerns Regarding Mohs Micrographic Surgery

Assessing the Accuracy of ChatGPT in Appropriately Triaging Common Postoperative Concerns Regarding Mohs Micrographic Surgery

As of April 2024, a pilot program in Louisiana incorporated Chat GPT-4.0 into electronic health record (EHR) messaging to generate preliminary responses that clinicians subsequently reviewed for validity [3]. Despite Chat GPT-4.0’s advances, the study demonstrated that human oversight in AI-generated communication remains essential [3]. Such initiatives demonstrate AI’s potential to reduce administrative workload, but they also underscore its role in improving patient education.

Chloe Fernandez, Victoria Dukharan, Nathaniel A Marroquin, Rebecca Bolen, Adam Leavitt, Nicole C Cabbad

JMIR Dermatol 2025;8:e72706

Co-Design of a Health Screening Program Fact Sheet by People Experiencing Homelessness and ChatGPT: Focus Group Study

Co-Design of a Health Screening Program Fact Sheet by People Experiencing Homelessness and ChatGPT: Focus Group Study

In this research project, we aimed to co-design an awareness-raising fact sheet for an oral cancer screening program with people experiencing homelessness as experts by experience and Chat GPT. The latter was used to present textual alternatives for this health information piece, so we could also test the usability of Chat GPT in designing adequate information materials serving the needs of people experiencing homelessness.

Nóra Radó, Orsolya Németh, Sándor Békási

JMIR Form Res 2025;9:e68316

Evaluating a Large Language Model’s Ability to Synthesize a Health Science Master’s Thesis: Case Study

Evaluating a Large Language Model’s Ability to Synthesize a Health Science Master’s Thesis: Case Study

In an exploratory case study, we asked Chat GPT (Chat GPT-4o and Chat GPT-o1) to synthesize two datasets: one qualitative and one quantitative, inspired by existing real-world datasets the authors had previously analyzed. The generation of synthetic datasets and subsequent data analysis took place in October and November of 2024.

Pål Joranger, Sara Rivenes Lafontan, Asgeir Brevik

JMIR Form Res 2025;9:e73248

Exploring Inflammatory Bowel Disease Discourse on Reddit Throughout the COVID-19 Pandemic Using OpenAI’s GPT-3.5 Turbo Model: Classification Model Validation and Case Study

Exploring Inflammatory Bowel Disease Discourse on Reddit Throughout the COVID-19 Pandemic Using OpenAI’s GPT-3.5 Turbo Model: Classification Model Validation and Case Study

Reference 13: ChatGPT outperforms humans in emotional awareness evaluations Reference 15: Evaluation of ChatGPT for NLP-based mental health applications(https://arxiv.org/abs/2303.15727 Reference 16: Bias in emotion recognition with ChatGPT(https://arxiv.org/abs/2310.11753) Reference 19: Human vs. machine: a comparative analysis of qualitative coding by humans and ChatGPT-4chatgptGenerative Language Models Including ChatGPT

Tyler Babinski, Sara Karley, Marita Cooper, Salma Shaik, Y Ken Wang

J Med Internet Res 2025;27:e53332

Evaluating the Quality of Psychotherapy Conversational Agents: Framework Development and Cross-Sectional Study

Evaluating the Quality of Psychotherapy Conversational Agents: Framework Development and Cross-Sectional Study

Custom GPTs are derivations of the baseline Chat GPT model (at the time of evaluation: GPT-4o) developed by Open AI that have been modified by members of the public with customized instructions and behavior for specific applications (eg, psychotherapy chatbot). In May 2024, we indexed both sites using the search feature to emulate what an end-user may experience with the following search terms: “therapy,” “anxiety,” “depression,” “mental health,” “therapist,” and “psychologist.”

Kunmi Sobowale, Daniel Kevin Humphrey

JMIR Form Res 2025;9:e65605

Enhancing Magnetic Resonance Imaging (MRI) Report Comprehension in Spinal Trauma: Readability Analysis of AI-Generated Explanations for Thoracolumbar Fractures

Enhancing Magnetic Resonance Imaging (MRI) Report Comprehension in Spinal Trauma: Readability Analysis of AI-Generated Explanations for Thoracolumbar Fractures

Recent analyses of Chat GPT-4o (Open AI) have demonstrated its effectiveness and accuracy in summarizing diagnostic radiology reports, with the ability to translate medical terminology to an 8th-grade reading level [3]. When prompted to explain medical imaging reports to a child using simplified and basic language, Chat GPT-4o generated 15 different reports, which were evaluated by 15 radiologists.

David C Sing, Kishan S Shah, Michael Pompliano, Paul H Yi, Calogero Velluto, Ali Bagheri, Robert K Eastlack, Stephen R Stephan, Gregory M Mundis Jr

JMIR AI 2025;4:e69654

Performance of Large Language Models in the Non-English Context: Qualitative Study of Models Trained on Different Languages in Chinese Medical Examinations

Performance of Large Language Models in the Non-English Context: Qualitative Study of Models Trained on Different Languages in Chinese Medical Examinations

Research has shown that Chat GPT, primarily trained on English-language corpora, is able to pass the United States Medical Licensing Examination [5], and can effectively addressing various clinical medical inquiries [6]. However, in some non-English medical examinations, Chat GPT fails to even reach a passing grade [7-9].

Zhong Yao, Liantan Duan, Shuo Xu, Lingyi Chi, Dongfang Sheng

JMIR Med Inform 2025;13:e69485

Sentiment Analysis Using a Large Language Model–Based Approach to Detect Opioids Mixed With Other Substances Via Social Media: Method Development and Validation

Sentiment Analysis Using a Large Language Model–Based Approach to Detect Opioids Mixed With Other Substances Via Social Media: Method Development and Validation

For transformer-based models and LLMs, we used pretrained embeddings from models like BERT and Chat GPT. These models capture deep contextual meanings by analyzing entire sentences rather than individual words. Unlike traditional methods, transformers dynamically understand context, improving sentiment analysis accuracy by recognizing complex language patterns. The inverse document frequency of a term reflects the inverse proportion of documents that contain that term.

Muhammad Ahmad, Ildar Batyrshin, Grigori Sidorov

JMIR Infodemiology 2025;5:e70525

Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives

Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives

The last years brought an unmatched rise of large language models (LLMs) such as Open AI’s Chat GPT [1,2], Google’s Bidirectional Encoder Representations from Transformers (BERT) [3], and Meta’s Llama [4]. LLMs have dramatically extended the abilities of natural language processing through generating text by repeatedly adding the most likely following words [5]. Most LLMs are based on large amounts of general-purpose text data on which the models were trained [6-8].

Florian Leiser, Richard Guse, Ali Sunyaev

J Med Internet Res 2025;27:e70315