Search Articles

View query in Help articles search

Search Results (1 to 10 of 2371 Results)

Download search results: CSV END BibTex RIS


Performance of Large Language Models in Numerical Versus Semantic Medical Knowledge: Cross-Sectional Benchmarking Study on Evidence-Based Questions and Answers

Performance of Large Language Models in Numerical Versus Semantic Medical Knowledge: Cross-Sectional Benchmarking Study on Evidence-Based Questions and Answers

For example, Katz et al [28] reported that GPT-4’s accuracy rates ranged from 17.42% (n=21) to 74.7% (n=90) across various medical disciplines. In contrast, our study found GPT-4’s accuracy rates ranged more narrowly, from 53.5% (n=704) to 60.35% (n=1076). This discrepancy could be partially attributed to the differing medical disciplines emphasized in each study, as well as variations in question structure.

Eden Avnat, Michal Levy, Daniel Herstain, Elia Yanko, Daniel Ben Joya, Michal Tzuchman Katz, Dafna Eshel, Sahar Laros, Yael Dagan, Shahar Barami, Joseph Mermelstein, Shahar Ovadia, Noam Shomron, Varda Shalev, Raja-Elie E Abdulnour

J Med Internet Res 2025;27:e64452

Networked Behaviors Associated With a Large-Scale Secure Messaging Network: Cross-Sectional Secondary Data Analysis

Networked Behaviors Associated With a Large-Scale Secure Messaging Network: Cross-Sectional Secondary Data Analysis

This is also reflected in cluster 1’s increased message volume and messaging behavior, most likely highlighting their role in clinical decision-making. However, such a centralized communication structure may also increase physicians’ workload and cognitive burden arising from an increased messaging volume [57]. Similarly, there was one cluster (cluster 4) of nurses and medical assistants who had fewer connections and were not as central within the network.

Laura Rosa Baratta, Linlin Xia, Daphne Lew, Elise Eiden, Y Jasmine Wu, Noshir Contractor, Bruce L Lambert, Sunny S Lou, Thomas Kannampallil

JMIR Med Inform 2025;13:e66544

Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

Improving the Readability of Institutional Heart Failure–Related Patient Education Materials Using GPT-4: Observational Study

We aim to expand on the previous literature by assessing the readability of heart failure–related online PEMs from renowned cardiology institutions, assessing GPT-4’s ability to improve the readability of these PEMs, and comparing the accuracy and comprehensiveness between institutional PEMs and GPT-4’s revised PEMs.

Ryan C King, Jamil S Samaan, Joseph Haquang, Vishnu Bharani, Samuel Margolis, Nitin Srinivasan, Yuxin Peng, Yee Hui Yeo, Roxana Ghashghaei

JMIR Cardio 2025;9:e68817