<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Cancer</journal-id><journal-id journal-id-type="publisher-id">cancer</journal-id><journal-id journal-id-type="index">21</journal-id><journal-title>JMIR Cancer</journal-title><abbrev-journal-title>JMIR Cancer</abbrev-journal-title><issn pub-type="epub">2369-1999</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v11i1e65984</article-id><article-id pub-id-type="doi">10.2196/65984</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>Large Language Model Applications for Health Information Extraction in Oncology: Scoping Review</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Chen</surname><given-names>David</given-names></name><degrees>BMSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Alnassar</surname><given-names>Saif Addeen</given-names></name><degrees>BASc</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Avison</surname><given-names>Kate Elizabeth</given-names></name><degrees>BASc</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Huang</surname><given-names>Ryan S</given-names></name><degrees>MSc</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Raman</surname><given-names>Srinivas</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib></contrib-group><aff id="aff1"><institution>Temerty Faculty of Medicine, University of Toronto</institution><addr-line>Toronto</addr-line><addr-line>ON</addr-line><country>Canada</country></aff><aff id="aff2"><institution>Department of Systems Design Engineering, University of Waterloo</institution><addr-line>Waterloo</addr-line><addr-line>ON</addr-line><country>Canada</country></aff><aff id="aff3"><institution>Department of Radiation Oncology, BC Cancer Vancouver</institution><addr-line>600 W 10th Ave</addr-line><addr-line>Vancouver</addr-line><addr-line>BC</addr-line><country>Canada</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Cahill</surname><given-names>Naomi</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Hu</surname><given-names>Danqing</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>You</surname><given-names>Kisung</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Patel</surname><given-names>Krishnan</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Srinivas Raman, MD, Department of Radiation Oncology, BC Cancer Vancouver, 600 W 10th Ave, Vancouver, BC, V5Z 4E6, Canada, 1 416-946-4501; <email>srinivas.raman@bccancer.bc.ca</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>28</day><month>3</month><year>2025</year></pub-date><volume>11</volume><elocation-id>e65984</elocation-id><history><date date-type="received"><day>30</day><month>08</month><year>2024</year></date><date date-type="rev-recd"><day>23</day><month>01</month><year>2025</year></date><date date-type="accepted"><day>27</day><month>01</month><year>2025</year></date></history><copyright-statement>&#x00A9; David Chen, Saif Addeen Alnassar, Kate Elizabeth Avison, Ryan S Huang, Srinivas Raman. Originally published in JMIR Cancer (<ext-link ext-link-type="uri" xlink:href="https://cancer.jmir.org">https://cancer.jmir.org</ext-link>), 28.3.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Cancer, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://cancer.jmir.org/">https://cancer.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://cancer.jmir.org/2025/1/e65984"/><abstract><sec><title>Background</title><p>Natural language processing systems for data extraction from unstructured clinical text require expert-driven input for labeled annotations and model training. The natural language processing competency of large language models (LLM) can enable automated data extraction of important patient characteristics from electronic health records, which is useful for accelerating cancer clinical research and informing oncology care.</p></sec><sec><title>Objective</title><p>This scoping review aims to map the current landscape, including definitions, frameworks, and future directions of LLMs applied to data extraction from clinical text in oncology.</p></sec><sec sec-type="methods"><title>Methods</title><p>We queried Ovid MEDLINE for primary, peer-reviewed research studies published since 2000 on June 2, 2024, using oncology- and LLM-related keywords. This scoping review included studies that evaluated the performance of an LLM applied to data extraction from clinical text in oncology contexts. Study attributes and main outcomes were extracted to outline key trends of research in LLM-based data extraction.</p></sec><sec sec-type="results"><title>Results</title><p>The literature search yielded 24 studies for inclusion. The majority of studies assessed original and fine-tuned variants of the BERT LLM (n=18, 75%) followed by the Chat-GPT conversational LLM (n=6, 25%). LLMs for data extraction were commonly applied in pan-cancer clinical settings (n=11, 46%), followed by breast (n=4, 17%), and lung (n=4, 17%) cancer contexts, and were evaluated using multi-institution datasets (n=18, 75%). Comparing the studies published in 2022&#x2010;2024 versus 2019&#x2010;2021, both the total number of studies (18 vs 6) and the proportion of studies using prompt engineering increased (5/18, 28% vs 0/6, 0%), while the proportion using fine-tuning decreased (8/18, 44.4% vs 6/6, 100%). Advantages of LLMs included positive data extraction performance and reduced manual workload.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>LLMs applied to data extraction in oncology can serve as useful automated tools to reduce the administrative burden of reviewing patient health records and increase time for patient-facing care. Recent advances in prompt-engineering and fine-tuning methods, and multimodal data extraction present promising directions for future research. Further studies are needed to evaluate the performance of LLM-enabled data extraction in clinical domains beyond the training dataset and to assess the scope and integration of LLMs into real-world clinical environments.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>chatbot</kwd><kwd>data extraction</kwd><kwd>AI</kwd><kwd>conversational agent</kwd><kwd>health information</kwd><kwd>oncology</kwd><kwd>scoping review</kwd><kwd>natural language processing</kwd><kwd>NLP</kwd><kwd>large language model</kwd><kwd>LLM</kwd><kwd>digital health</kwd><kwd>health technology</kwd><kwd>electronic health record</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The advent of electronic health records (EHR) has allowed clinicians to leverage their access to vast amounts of longitudinal, patient-level clinical text data that inform patient diagnoses, prognoses, and management [<xref ref-type="bibr" rid="ref1">1</xref>]. However, the majority of useful clinical data are stored as unstructured free text that requires manual extraction into meaningful clinical features; therefore, clinicians spend more time on administrative work reviewing EHRs instead of practising patient-facing medicine [<xref ref-type="bibr" rid="ref1">1</xref>]. To address this task of extracting key attributes from unstructured clinical text, natural language processing (NLP) methods have classically applied rule-based and machine-learning methods to identify important entities in text and categorize them based on categories of interest [<xref ref-type="bibr" rid="ref2">2</xref>]. For instance, the extraction of cancer staging information from clinical text requires an NLP algorithm to recognize references to cancer staging in clinical texts and categorize these references according to defined cancer staging nomenclature, such as the TNM classification of malignant tumors system.</p><p>Rule-based classification relies on domain expert-designed rules, heuristics, ontologies, and pattern-matching techniques to extract information from text. In contrast, machine learning-based approaches use statistical models trained on large-scale labeled text data to automatically learn patterns and generalize these learned competencies in data extraction to unlabeled testing data. The emergence of deep learning models, a subfield of machine learning that focuses on artificial neural network models with multiple processing layers, has been particularly effective at modeling the hierarchical structure of natural language and demonstrated superior performance across diverse NLP tasks, including but not limited to data extraction [<xref ref-type="bibr" rid="ref3">3</xref>].</p><p>One particularly promising deep learning architecture, known as the transformer model, has gained worldwide attention for its generative language competency and strong performance in question answering, sentence completion, and sentence classification tasks compared to other deep learning models [<xref ref-type="bibr" rid="ref4">4</xref>]. Deep learning&#x2013;based transformer models may require less time and fewer resources needed to manually annotate training datasets compared to classical machine learning models and can better address nuanced edge cases in data extraction that may not be explicitly accounted for in rule-based data approaches [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>]. However, these models are often limited by their need for large-scale computational resources and training data [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>Modern LLMs are commonly built using adaptations of the transformer architecture and trained on large corpora of text to enable human-like natural language competency. Due to their extensive training dataset, LLMs such as BERT and GPT may have zero-shot capabilities, meaning they can perform tasks without prior task-specific training [<xref ref-type="bibr" rid="ref9">9</xref>]. Emerging research on fine-tuning LLMs with custom datasets and prompt engineering for conversational LLMs has yielded promising performance improvements for specialized NLP tasks compared to baseline LLMs.</p><p>Given the longitudinal nature of cancer care, the vast amount of clinical text associated with cancer patient EHRs necessitates the development of automated methods for data extraction from these clinical records into structured data, which is useful for review by oncologists. The broad natural language competency of LLMs encourages the design of specialized LLM applications for data extraction from unstructured clinical text, reducing the oncologists&#x2019; time and effort spent in manually reviewing patient EHRs to extract key information to inform their clinical decision-making.</p><p>The emergence of several recent pilot studies of LLM-enabled data extraction prompts the need for a scoping review to map the current landscape, including definitions, frameworks, and future directions for this novel tool in clinical data extraction. This review seeks to address this gap in the literature by characterizing primary research articles that evaluated an LLM tool applied to data extraction from unstructured clinical text into structured data.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><p>We queried OVID Medline on June 2, 2024, using oncology (&#x201C;neoplasms,&#x201D; &#x201C;cancer,&#x201D; &#x201C;onco,&#x201D; &#x201C;tumor&#x201D;) and generative LLM (&#x201C;natural language processing,&#x201D; &#x201C;artificial intelligence,&#x201D; &#x201C;generative,&#x201D; &#x201C;large language model&#x201D;) keywords in consultation with a librarian. Non-English articles, nonprimary research articles, articles published before 2000, and articles published in nonpeer-reviewed settings were excluded. The full search strategy is detailed in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. Following the deduplication of articles (n=10) using the Covidence review management tool, the literature search yielded 817 articles for manual screening.</p><p>We conducted abstract screening followed by full-text screening of articles in duplicate (KA and SA), including primary research articles that tested a large language model, were applied in oncology contexts, and evaluated the performance of data extraction from text. The articles that evaluated an NLP-based algorithm that did not assess an LLM, were secondary research articles, applied in only nononcology settings, and did not evaluate or report the performance of data extraction from the clinical text were excluded. Screening conflicts were resolved through consensus discussion with a third reviewer (DC).</p><p>We extracted key study attributes from the included full-text papers in duplicate (KA and SA), including clinical domain, LLM attributes (eg, model, use of fine-tuning, use of prompt engineering), the dataset used for training and testing, primary study outcomes, model training methodology, and model evaluation processes. The LLMs were coded as baseline if they were applied &#x201C;out of the box&#x201D; without additional fine-tuning. LLMs were coded as (1) fine-tuned LLMs: the study described training the baseline LLM on a custom dataset intended to yield improved data extraction performance compared to the baseline LLM alone; (2) zero-shot LLMs: they were applied &#x201C;out-of-the-box&#x201D; without additional prompt engineering, (3) prompt engineered LLMs: the study described adaptations to prompting procedures, such as one-shot or few-shot prompting, designed to yield improved data extraction performance compared to the baseline LLM alone. Data extraction conflicts were resolved through consensus discussion with a third reviewer (DC).</p><p>The synthesis of extracted data involved grouping studies based on similarities in the evaluated specific model, clinical domain applied, and shared themes of strengths and limitations, based on outcomes reported by the studies. The appraisal process involved the completion of a standardized data extraction form to systematically code in duplicate (KA and SA) which articles commented on which themes of strengths and limitations, and the discrepancies were resolved through discussion (DC and SR). The risk of bias was assessed using ROBINS-I (Version 2) in duplicate (KA and SA), with conflicts resolved through consensus discussion with a third reviewer (DC). Cohen <italic>&#x03BA;</italic> score was used to assess inter-rater concordance. This scoping review followed the PRISMA-ScR reporting guideline.</p></sec><sec id="s3" sec-type="results"><title>Results</title><p>The literature search yielded 817 papers, of which 24 papers met the inclusion criteria (<xref ref-type="fig" rid="figure1">Figure 1</xref>). Most included papers exhibited moderate (n=15, 62.5%) risk or low (n=9, 37.5%) risk of bias (<xref ref-type="fig" rid="figure2">Figure 2</xref>). The most common domains for moderate risk of bias included bias due to confounding (n=21, 87.5%) and bias in the selection of the reported result (n=21, 87.5%). No papers scored a high risk of bias in any domain. ROBINS-I risk of bias assessment exhibited moderate inter-rater concordance based on an <italic>&#x03BA;</italic> score of 0.43.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Search and filtering strategy used to select large language model studies evaluating data extraction performance for inclusion in this review. LLM: large language model.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="cancer_v11i1e65984_fig01.png"/></fig><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Risk of bias assessment using the ROBINS-I tool displayed as a traffic light plot for each included study [<xref ref-type="bibr" rid="ref10">1</xref>,<xref ref-type="bibr" rid="ref11">3</xref>,<xref ref-type="bibr" rid="ref3">5</xref>,<xref ref-type="bibr" rid="ref1">6</xref>,<xref ref-type="bibr" rid="ref5">7</xref>,<xref ref-type="bibr" rid="ref12">8</xref>-<xref ref-type="bibr" rid="ref16">12</xref>,<xref ref-type="bibr" rid="ref7">13</xref>,<xref ref-type="bibr" rid="ref17">14</xref>-<xref ref-type="bibr" rid="ref19">16</xref>,<xref ref-type="bibr" rid="ref6">17</xref>,<xref ref-type="bibr" rid="ref9">18</xref>,<xref ref-type="bibr" rid="ref8">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>-<xref ref-type="bibr" rid="ref26">26</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="cancer_v11i1e65984_fig02.png"/></fig><p>Characteristics of the studies included in the study and published between 2019&#x2010;2024 are shown in <xref ref-type="table" rid="table1">Table 1</xref>. The most common LLMs reported in these studies included BERT and its variants, as well as ChatGPT. Additional details related to methodology are reported in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Characteristics of studies included in the review.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Study ID</td><td align="left" valign="bottom">Clinical domain</td><td align="left" valign="bottom">Baseline model</td><td align="left" valign="bottom">Baseline or fine-tuned LLM<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td><td align="left" valign="bottom">Zero-shot or prompt -engineered LLM</td><td align="left" valign="bottom">LLM main outcomes</td></tr></thead><tbody><tr><td align="left" valign="top">Solarte-Pabon 2023[<xref ref-type="bibr" rid="ref10">10</xref>]</td><td align="left" valign="top">Breast</td><td align="left" valign="top">BERT; RoBERTa</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">F-scores: BETA: 0.9371; Multilingual BERT: 0.9463; RoBERTa Biomedical: 0.9501; RoBERTa BNE: 0.9454</td></tr><tr><td align="left" valign="top">Choi 2023 [<xref ref-type="bibr" rid="ref11">11</xref>]</td><td align="left" valign="top">Breast</td><td align="left" valign="top">ChatGPT-3.5</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Prompt-engineered</td><td align="left" valign="top">Accuracy: 87.7%</td></tr><tr><td align="left" valign="top">Zhou 2022 [<xref ref-type="bibr" rid="ref3">3</xref>]</td><td align="left" valign="top">Breast</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">F1-score: 0.866 and 0.904 for exact and permissive matches respectively</td></tr><tr><td align="left" valign="top">Zhang 2019 [<xref ref-type="bibr" rid="ref1">1</xref>]</td><td align="left" valign="top">Breast</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">NER:<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> 93.53%; Relation extraction: 96.73% (best model, BERT+ Bi-LSTM-CRF)</td></tr><tr><td align="left" valign="top">Seong 2023 [<xref ref-type="bibr" rid="ref5">5</xref>]</td><td align="left" valign="top">Colorectal</td><td align="left" valign="top">Bi-LSTM with a CRF layer; BioBERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Bi-LSTM-CRF:<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup> Precision: 0.9844; F1-score:0.9848; Pre trained word embedding performed better than the one hot encoding pre-processing</td></tr><tr><td align="left" valign="top">Laios 2023 [<xref ref-type="bibr" rid="ref12">12</xref>]</td><td align="left" valign="top">Gynecology</td><td align="left" valign="top">RoBERTa</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">AUROC:<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup> 0.86; AUPRC:<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup> 0.87; F1: 0.77; Accuracy: 0.81</td></tr><tr><td align="left" valign="top">Liu 2021 [<xref ref-type="bibr" rid="ref13">13</xref>]</td><td align="left" valign="top">Liver</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">APHE<sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup>: 98.40%; PDPH<sup><xref ref-type="table-fn" rid="table1fn7">g</xref></sup>: 90.67%</td></tr><tr><td align="left" valign="top">Fink 2023 [<xref ref-type="bibr" rid="ref14">14</xref>]</td><td align="left" valign="top">Lung</td><td align="left" valign="top">ChatGPT-3.5; ChatGPT-4.0</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Prompt-engineered</td><td align="left" valign="top">Overall accuracy: GPT-4: 98.6%; GPT-3.5: 84%<break/>Metastatic ID accuracy: GPT-4: 98.1%; GPT-3.5: 90.3%<break/>Oncologic progression accuracy: GPT-4 F1: 0.96; GPT-3.5: 0.91<break/>Oncologic reasoning correctness: GPT-4: 4.3; GPT-3.5: 3.9<break/>accuracy: GPT-4: 4.4; GPT-3.5: 3.3</td></tr><tr><td align="left" valign="top">Chen 2023 [<xref ref-type="bibr" rid="ref15">15</xref>]</td><td align="left" valign="top">Lung</td><td align="left" valign="top">&#x2003;BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Macro F1-score: Task 1:0.92; Task 2: 0.82; Task 3: 0.74</td></tr><tr><td align="left" valign="top">Lyu 2023 [<xref ref-type="bibr" rid="ref16">16</xref>]</td><td align="left" valign="top">Lung</td><td align="left" valign="top">ChatGPT-4.0</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Translate: 4.27/5; Provided specific suggestions based on findings in 37% of all cases</td></tr><tr><td align="left" valign="top">Yu 2021 [<xref ref-type="bibr" rid="ref7">7</xref>]</td><td align="left" valign="top">Lung</td><td align="left" valign="top">BERT; RoBERTa</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">BERT Lenient: 0.8999<break/>BERT Strict: 0.8791</td></tr><tr><td align="left" valign="top">Martin-Noguerol 2024 [<xref ref-type="bibr" rid="ref17">17</xref>]</td><td align="left" valign="top">Neurology</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-Shot</td><td align="left" valign="top">HGG: Precision: 79.17; Sensitivity: 76; F1:77.55; Metastasis: Precision: 73.91; Sensitivity: 77.27; F1: 75.56; AUC: 76.64</td></tr><tr><td align="left" valign="top">Fang 2022_1 [<xref ref-type="bibr" rid="ref18">18</xref>]</td><td align="left" valign="top">Endocrine</td><td align="left" valign="top">BERT-BiLSTM-CRF</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Strict F1-score: 91.27%; Relaxed F1-score: 95.57%</td></tr><tr><td align="left" valign="top">Huang 2024 [<xref ref-type="bibr" rid="ref19">19</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">ChatGPT-3.5</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Prompt-engineered</td><td align="left" valign="top">Accuracy 0.89; F1 0.88; Kappa 0.80; Recall 0.89; Precision 0.89, Coverage 0.95</td></tr><tr><td align="left" valign="top">Arya 2024 [<xref ref-type="bibr" rid="ref6">6</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Predict imaging scan site: Precision:99.4%; Recall:99.4%; F1-score: 99.3%; AUROC:99.4%; Accuracy:99.9%; Predict cancer presence: Precision:88.8%; Recall:89.2%; F1:88.8%; AUROC:97.6%; Accuracy:93.4%; Predict cancer status: Precision:85.6%; Recall:85.5%; F1-score: 85.5%; AUROC:97%; Accuracy:93.1%</td></tr><tr><td align="left" valign="top">Truhn 2024 [<xref ref-type="bibr" rid="ref9">9</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">ChatGPT-4.0</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Experiment 1: Correct T-stage: 99%; Correct N-stage: 95; Correct M stage: 94; Lymph nodes; 99%<break/>Experiment 3: 100% accuracy</td></tr><tr><td align="left" valign="top">Hu 2024 [<xref ref-type="bibr" rid="ref8">8</xref>]</td><td align="left" valign="top">Lung</td><td align="left" valign="top">ChatGPT-4.0</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Prompt-engineered</td><td align="left" valign="top">Prompt Base: Accuracy: 0.937; Precision: 0.860; Recall: 0.917; F1-score:0.882; Prior medical knowledge: Accuracy: 0.940; Precision:0.900; Recall: 0.864; F1:0.867; PMK-EN<sup><xref ref-type="table-fn" rid="table1fn8">h</xref></sup>: Accuracy: 0.896; Precision:0.871: Recall:0.776; F1: 0.786</td></tr><tr><td align="left" valign="top">Elmarakeby 2023 [<xref ref-type="bibr" rid="ref20">20</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">AUC: ClinicalBERT: 0.93; DFCI-ImagingBERT: 0.95<break/>F1: ClinicalBERT: 0.72; DFCI-ImagingBERT: 0.78</td></tr><tr><td align="left" valign="top">Tan 2023 [<xref ref-type="bibr" rid="ref21">21</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">GatorTron; BERT; PubMedGPT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Prompt-engineered</td><td align="left" valign="top">Accuracy: GatorTron: 0.8916; BioMegatron:0.8861; BioBERT:0.8861; RoBERTa:0.8813; PubMedGPT:0.8762; DeBERTa:0.8746; BioClinicalBERT: 0.8746; BERT: 0.8708</td></tr><tr><td align="left" valign="top">Fang 2022_2 [<xref ref-type="bibr" rid="ref22">22</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Baseline</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">ROC:<sup><xref ref-type="table-fn" rid="table1fn9">i</xref></sup> 0.94</td></tr><tr><td align="left" valign="top">Mitchell 2022 [<xref ref-type="bibr" rid="ref23">23</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Group level site accuracy: 93.53%; Histology codes: 97.6%</td></tr><tr><td align="left" valign="top">Lu 2021 [<xref ref-type="bibr" rid="ref24">24</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">Symptom domains: 0.931; problems with cognitive and social attributes on pain interference: 0.916; problems on fatigue: 0.929</td></tr><tr><td align="left" valign="top">Percha 2021 [<xref ref-type="bibr" rid="ref25">25</xref>]</td><td align="left" valign="top">Breast</td><td align="left" valign="top">ALBERT; BART; ELECTRA; RoBERTa; XLNet</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">ALBERT was the best-performing model in 22 out of the 43 fields</td></tr><tr><td align="left" valign="top">Gao 2021 [<xref ref-type="bibr" rid="ref26">26</xref>]</td><td align="left" valign="top">Pan-cancer</td><td align="left" valign="top">BlueBERT</td><td align="left" valign="top">Fine-tuned</td><td align="left" valign="top">Zero-shot</td><td align="left" valign="top">BERT does not outperform baseline models&#x2013;quantifiable measures not available</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>LLM: large language model.</p></fn><fn id="table1fn2"><p><sup>b</sup>NER: named entity recognition.</p></fn><fn id="table1fn3"><p><sup>c</sup>Bi-LSTM-CRF: bidirectional-long short term memory-conditional random field.</p></fn><fn id="table1fn4"><p><sup>d</sup>AUROC: area under the receiver operating characteristic.</p></fn><fn id="table1fn5"><p><sup>e</sup>AUPRC: area under the precision-recall curve.</p></fn><fn id="table1fn6"><p><sup>f</sup>APHE: hyperintense enhancement in the arterial phase.</p></fn><fn id="table1fn7"><p><sup>g</sup>PDPH: hypointense in the portal and delayed phases.</p></fn><fn id="table1fn8"><p><sup>h</sup>PMK-EN: Prior Medical Knowledge-English Prompt</p></fn><fn id="table1fn9"><p><sup>i</sup>ROC: receiver operating characteristic.</p></fn></table-wrap-foot></table-wrap><p>Most studies evaluated either the original or fine-tuned variants of the BERT LLM (n=18, 75%) in studies published between 2019&#x2010;2024, followed by the Chat-GPT conversational LLM (n=6, 25%), upon application to data extraction from clinical texts in oncology, in studies published between 2023&#x2010;2024. The LLMs for data extraction were commonly applied in pan-cancer clinical settings (n=11, 46%), followed by breast (n=4, 17%), lung (n=4, 17%), neurological (n=2, 8%), colorectal (n=1, 4%), gynecological (n=1, 4%), and liver (n=1, 4%) cancer contexts. The author teams of these studies belonged to institutions in the United States (n=11, 46%), China (n=4, 17%), Korea (n=3, 12%), Germany (n=2, 8%), Spain (n=2, 8%), the United Kingdom (n=1, 4%), and Singapore (n=1, 4%). Most studies were evaluated on datasets sourced from multiple institutional centers (n=18, 75%) compared to a single institutional center (n=6, 25%). Regarding the year of study publication, we observed a higher number of studies published between 2022&#x2010;2024 (n=18, 75%) compared to 2019&#x2010;2021 (n=6, 25%) (<xref ref-type="fig" rid="figure3">Figure 3</xref>). Notably, upon a comparison of studies published between 2022&#x2010;2024 with studies between 2019&#x2010;2021, the proportion of studies that reported the use of the fine-tuning method was lower (10/18, 55.6% vs 6/6, 100%) (<xref ref-type="fig" rid="figure3">Figure 3A</xref>), whereas the proportion of studies that reported the use of prompt engineering was higher (5/18, 28% vs 0/6, 0%) (<xref ref-type="fig" rid="figure3">Figure 3B</xref>).</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Number of studies that evaluated (A) fine-tuning and (B) prompt engineering methodologies to optimize large language model data extraction performance.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="cancer_v11i1e65984_fig03.png"/></fig></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>Our scoping review of 24 studies highlights significant research interest in designing, evaluating, and deploying LLMs for data extraction from clinical text in oncology. The most commonly used LLMs for data extraction from clinical text in oncology include BERT and Chat-GPT, two of the most well-known LLMs in NLP research. These LLMs were most frequently applied in pan-cancer clinical contexts, reflecting their generalized natural language competency, regardless of clinical domain and context-specific terminologies and nomenclature. We observed a notable trend toward increasing utilization and refinement of LLM techniques over time, particularly in the areas of fine-tuning and prompt engineering. Given the common application of fine-tuning [<xref ref-type="bibr" rid="ref26">26</xref>-<xref ref-type="bibr" rid="ref28">28</xref>] and prompt-engineering [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref30">30</xref>] techniques in the design of deep learning- and LLM-based models in oncology, respectively, the emergence of optimized LLMs using these techniques represents a promising future direction for enhancing their data-processing capabilities. Despite these advancements, mixed reports of data extraction performance underscore the imperative for further assessment of these models across specific topics and use cases before their deployment as tools in cancer research and clinical care. Compared to historical statistical NLP and machine learning-based methods for data extraction in oncology, LLMs have been broadly evaluated for comparable applications, such as extracting tumor and cancer characteristics and patient-related demographic data [<xref ref-type="bibr" rid="ref31">31</xref>].</p><p>The data processing competency of LLMs makes them a useful tool for automating repetitive, rule-based tasks, such as data extraction from clinical text on EHRs, to generate medical evidence about specific patients and patient populations that can inform patient care and population health guidelines respectively. Notably, LLMs have already shown competency in pilot studies of automated data extraction in biology [<xref ref-type="bibr" rid="ref32">32</xref>], materials science materials science [<xref ref-type="bibr" rid="ref33">33</xref>], and pharmacology [<xref ref-type="bibr" rid="ref33">33</xref>], suggesting their generalized ability to extract relevant named entities from the clinical text that may be useful to synthesize medical knowledge. Across the included studies in this review, we found that LLMs offer several benefits for data extraction in clinical oncology, though further benchmarking against representative datasets and classical machine learning or statistical NLP approaches is required to determine their superior performance. In general, LLMs exhibited positive performance metrics compared to baseline human or statistical NLP approaches or were deemed feasible and acceptable in cross-sectional studies. These LLMs harbor the potential to balance accuracy and efficiency when processing large-scale, complex, unstructured text datasets found in EHRs [<xref ref-type="bibr" rid="ref19">19</xref>]. Using LLM approaches for clinical data extraction as a supportive tool along with a human reviewer may reduce the potential for errors associated with human-led manual data extraction alone, thereby enhancing the reliability of clinical data analyses and interpretations [<xref ref-type="bibr" rid="ref34">34</xref>].</p><p>Moreover, LLMs may curtail the resources required for data extraction, which is traditionally a labor- and time-intensive process [<xref ref-type="bibr" rid="ref35">35</xref>]. For instance, our review highlighted the generalized performance of LLM-enabled data extraction across various text types in oncology, including histological and pathological classification [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref36">36</xref>], imaging report classification [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref14">14</xref>], and data extraction from postoperative surgery reports [<xref ref-type="bibr" rid="ref5">5</xref>]. By automating the extraction and preliminary analysis of clinical text data, these models may free up valuable time for health care professionals, allowing them to focus more on patient-facing care and synthesis of medical knowledge from LLM-extracted information rather than the burden of administrative data management [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref37">37</xref>]. This shift not only improves clinical efficiency and cost-effectiveness but also reduces the serious risks of burnout among clinical staff by mitigating some of the repetitive administrative tasks associated with data handling [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref38">38</xref>].</p><p>Additionally, the versatility of LLMs across different clinical text contexts is notable. Whether dealing with structured data formats or the myriad forms of unstructured data present in EHRs, such as physician&#x2019;s notes and diagnostic reports, the general human-like natural language competencies of LLMs enable these &#x201C;out-of-the-box&#x201D; solutions to automatically adapt to and extract relevant information from varied data sources. This adaptability is crucial in precision oncology, where data from multiple data formats&#x2014;such as imaging reports, next-generation sequencing results, and laboratory results&#x2014;must be integrated and analyzed to generate personalized patient profiles and treatment strategies [<xref ref-type="bibr" rid="ref39">39</xref>]. Our review highlighted that current state-of-the-art evaluations of LLMs for data extraction in oncology have primarily focused on clinical text as input. However, we also highlight the recent emergence of multimodal LLMs capable of processing both image- and text-based inputs, serving as a new frontier for clinical decision support [<xref ref-type="bibr" rid="ref40">40</xref>]. Taken together, future research to optimize data extraction for specific text formats in oncology&#x2014;each with their own nuances&#x2014;may improve extraction accuracy, enhance reliability, and produce results that can be trusted by clinicians and readily inform clinical decision-making [<xref ref-type="bibr" rid="ref41">41</xref>].</p><p>The distribution of studies included in our scoping review reflects a predominant application of LLMs in pan-cancer clinical domains, accounting for nearly half of all research studies. This suggests that researchers leverage the versatility of LLMs to address broad oncological challenges across multiple cancer types, likely due to the generalizable nature of these models for various cancer data [<xref ref-type="bibr" rid="ref42">42</xref>]. Breast and lung cancer also constituted a large portion of the studies, which can likely be attributed to their high prevalence and extensive clinical data availability, providing a rich dataset for deploying and testing the efficacy of LLMs [<xref ref-type="bibr" rid="ref43">43</xref>]. The focus on these specific cancers indicates a targeted approach, where models are fine-tuned to address unique data extraction challenges, such as cancer type-specific nomenclature and lexicons. This underscores the potential of LLMs to be customized for specialized medical fields while also highlighting their broad &#x201C;out-of-the-box&#x201D; utility in general oncology. For instance, Gao et al [<xref ref-type="bibr" rid="ref44">44</xref>] reported that BlueBERT did not outperform baseline nonLLM models in pan-cancer contexts, while Fang et al [<xref ref-type="bibr" rid="ref22">22</xref>] and Mitchell et al (2022) [<xref ref-type="bibr" rid="ref23">23</xref>] reported that the data extraction performance of BERT exceeded 90% accuracy in pan-cancer contexts. The mixed performance reported by different pilot studies of data extraction performance within the same clinical domain may be confounded by study-specific factors, including the prompting methodology, benchmark dataset, and definitions of performance metrics. These findings align with similar reports of mixed performance across different tasks and clinical text datasets within cancer type-specific domains [<xref ref-type="bibr" rid="ref45">45</xref>-<xref ref-type="bibr" rid="ref47">47</xref>], highlighting the need for systematic benchmarks to assess LLM data extraction reliability and domain-specific limitations. Standardizing performance metrics and defining critical thresholds for acceptable performance of data extraction accuracy remain open research questions to be addressed.</p><p>Our analysis reveals an increasing trend in the use of fine-tuning and prompt-engineering techniques in studies on LLMs, with 16 (67%) studies incorporating fine-tuning and 5 (21%) using prompt engineering. This progression suggests a maturation in the application of LLMs in clinical settings, where research has transitioned from developing baseline models for simple data extraction to the optimization of existing models using novel model adaptations and prompting methodologies tailored to the intricacies of medical data extraction. Fine-tuning allows models to adapt to the unique linguistic and contextual challenges presented by medical texts, potentially improving the accuracy and relevance of extracted information [<xref ref-type="bibr" rid="ref29">29</xref>]. In comparison, prompt engineering enables the creation of more effective queries that align closely with the specific information needs of specialty fields such as oncology, steering LLMs toward more precise data retrieval [<xref ref-type="bibr" rid="ref48">48</xref>]. For instance, Huang et al [<xref ref-type="bibr" rid="ref19">19</xref>] demonstrated that providing LLMs with example outputs for few-shot learning and chain-of-thought reasoning methods for prompting yielded higher classification performance compared to baseline zero-shot applications of LLMs for data extraction. The careful design of prompting methodologies personalized to specific tasks and clinical domains within oncology may yield more accurate and efficient data extraction performance [<xref ref-type="bibr" rid="ref49">49</xref>].</p><p>Despite the promising applications of LLMs in clinical oncology, our review also highlights notable disadvantages, particularly in cases of poor data extraction accuracy and performance [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. Among the 24 reviewed studies, 9 (38%) cited accuracy as a limitation of LLMs for data extraction. These shortcomings underscore the critical need for cautious integration of LLMs into clinical workflows. The variability in performance can be attributed to the complex and diverse nature of clinical data, which may include nuanced medical terminologies and varied presentation styles across different documents [<xref ref-type="bibr" rid="ref50">50</xref>]. These challenges emphasize the necessity for ongoing refinement and testing of these models under real-world conditions. Another minor disadvantage is the token limit of many LLMs, including both BERT and ChatGPT [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref42">42</xref>,<xref ref-type="bibr" rid="ref44">44</xref>]. This limitation may complicate the extraction process, requiring models to be adapted to longer texts and resulting in reduced performance of these models [<xref ref-type="bibr" rid="ref51">51</xref>]. Future research directions, as indicated by the reviewed studies, should involve performance benchmarks against existing statistical and machine learning&#x2013;based methods and the extension of LLM tool validation to external, hold-out cohorts from additional clinical domains beyond those used in initial training datasets [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref24">24</xref>]. This would help ensure that the models are robust and reliable across various medical specialties and global oncology patient populations. While LLMs hold significant potential to revolutionize data management in oncology, their integration into clinical practice must be approached with careful planning and systematic evaluation to truly harness their capabilities without compromising patient care quality and privacy. The interpretation of both advantages and disadvantages of LLMs requires individualized consideration of each study, on a case-by-case basis given the heterogeneity in benchmark datasets, study designs, and reported outcomes.</p></sec><sec id="s4-2"><title>Limitations</title><p>We acknowledge the limitations inherent in our scoping review. First, the rapid evolution of LLM technologies means that newer advancements may not have been fully represented in the reviewed studies due to the delays in publication cycles, leading to the omission of recent models. Second, the heterogeneity in study designs, datasets, and methodologies across included articles may affect the generalizability of findings in external contexts not evaluated in the same conditions as the original studies. Third, the majority of included studies originated from high-resource settings, primarily the United States, which may limit the applicability of results to lower-resource or structurally different health care systems. Fourth, while the risk of publication bias was not formally evaluated in our review, the tendency to publish studies with positive results may overrepresent the strengths of these LLMs without an understanding and consideration of their limitations and nonpublished, negative results. Fifth, more recent journals that publish artificial intelligence research may not be indexed in the search databases yet, limiting the completeness of the search results in this scoping review. Sixth, this scoping review searched only one literature database, which may have resulted in the omission of relevant studies from other sources and limited the comprehensiveness of the findings.</p></sec><sec id="s4-3"><title>Conclusion</title><p>In conclusion, the application of LLMs in oncology represents a forward leap in the digital transformation of health care data management. The potential to enhance data extraction processes and improve clinical decision-making is significant yet tempered by the current technological and methodological limitations. Ongoing research and development will be vital in harnessing the full potential of these models, ultimately leading to their more widespread adoption in clinical practice.</p></sec></sec></body><back><fn-group><fn fn-type="con"><p>Conceptualization: DC, SR</p><p>Data curation: KA, RH, SA</p><p>Formal analysis: DC, KA, RH, SA</p><p>Funding acquisition: SR</p><p>Investigation: DC, KA, RH, SA</p><p>Methodology: DC, SR</p><p>Project administration: DC, SR</p><p>Visualization: DC</p><p>Supervision: SR</p><p>Writing &#x2013; original draft: DC, KA, RH, SA</p><p>Writing &#x2013; review &#x0026; editing: SR</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">EHR</term><def><p>electronic health record</p></def></def-item><def-item><term id="abb2">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb3">NLP</term><def><p>natural language processing</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Q</given-names> </name><etal/></person-group><article-title>Extracting comprehensive clinical information for breast cancer using deep learning methods</article-title><source>Int J Med Inform</source><year>2019</year><month>12</month><volume>132</volume><fpage>103985</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2019.103985</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Rastegar-Mojarad</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Clinical information extraction applications: a literature review</article-title><source>J Biomed Inform</source><year>2018</year><month>01</month><volume>77</volume><fpage>34</fpage><lpage>49</lpage><pub-id pub-id-type="doi">10.1016/j.jbi.2017.11.011</pub-id><pub-id pub-id-type="medline">29162496</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>N</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>R</given-names> </name></person-group><article-title>CancerBERT: a cancer domain-specific language model for extracting breast cancer phenotypes from electronic health records</article-title><source>J Am Med Inform Assoc</source><year>2022</year><month>06</month><day>14</day><volume>29</volume><issue>7</issue><fpage>1208</fpage><lpage>1216</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocac040</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Devlin</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>MW</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>K</given-names> </name><name name-style="western"><surname>Toutanova</surname><given-names>K</given-names> </name></person-group><article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title><source>arXiv</source><comment>Preprint posted online on  May 24, 2019</comment><comment><ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/1810.04805">http://arxiv.org/abs/1810.04805</ext-link></comment></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Seong</surname><given-names>D</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Shin</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>BK</given-names> </name></person-group><article-title>Deep learning approach to detection of colonoscopic information from unstructured reports</article-title><source>BMC Med Inform Decis Mak</source><year>2023</year><month>02</month><day>7</day><volume>23</volume><issue>1</issue><fpage>28</fpage><pub-id pub-id-type="doi">10.1186/s12911-023-02121-7</pub-id><pub-id pub-id-type="medline">36750932</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arya</surname><given-names>A</given-names> </name><name name-style="western"><surname>Niederhausern</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bahadur</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Artificial intelligence-assisted cancer status detection in radiology reports</article-title><source>Cancer Res Commun</source><year>2024</year><month>04</month><day>9</day><volume>4</volume><issue>4</issue><fpage>1041</fpage><lpage>1049</lpage><pub-id pub-id-type="doi">10.1158/2767-9764.CRC-24-0064</pub-id><pub-id pub-id-type="medline">38592452</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Dang</surname><given-names>C</given-names> </name><etal/></person-group><article-title>A study of social and behavioral determinants of health in lung cancer patients using transformers-based natural language processing models</article-title><source>AMIA Annu Symp Proc</source><year>2021</year><volume>2021</volume><fpage>1225</fpage><lpage>1233</lpage><pub-id pub-id-type="medline">35309014</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>B</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Lu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>N</given-names> </name></person-group><article-title>Zero-shot information extraction from radiological reports using ChatGPT</article-title><source>Int J Med Inform</source><year>2024</year><month>03</month><volume>183</volume><fpage>105321</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2023.105321</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Truhn</surname><given-names>D</given-names> </name><name name-style="western"><surname>Loeffler</surname><given-names>CM</given-names> </name><name name-style="western"><surname>M&#x00FC;ller-Franzes</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Extracting structured information from unstructured histopathology reports using generative pre-trained transformer 4 (GPT-4)</article-title><source>J Pathol</source><year>2024</year><month>03</month><volume>262</volume><issue>3</issue><fpage>310</fpage><lpage>319</lpage><pub-id pub-id-type="doi">10.1002/path.6232</pub-id><pub-id pub-id-type="medline">38098169</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Solarte-Pab&#x00F3;n</surname><given-names>O</given-names> </name><name name-style="western"><surname>Montenegro</surname><given-names>O</given-names> </name><name name-style="western"><surname>Garc&#x00ED;a-Barrag&#x00E1;n</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Transformers for extracting breast cancer information from Spanish clinical narratives</article-title><source>Artif Intell Med</source><year>2023</year><month>09</month><volume>143</volume><fpage>102625</fpage><pub-id pub-id-type="doi">10.1016/j.artmed.2023.102625</pub-id><pub-id pub-id-type="medline">37673566</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>HS</given-names> </name><name name-style="western"><surname>Song</surname><given-names>JY</given-names> </name><name name-style="western"><surname>Shin</surname><given-names>KH</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Jang</surname><given-names>BS</given-names> </name></person-group><article-title>Developing prompts from large language model for extracting clinical information from pathology and ultrasound reports in breast cancer</article-title><source>Radiat Oncol J</source><year>2023</year><month>09</month><volume>41</volume><issue>3</issue><fpage>209</fpage><lpage>216</lpage><pub-id pub-id-type="doi">10.3857/roj.2023.00633</pub-id><pub-id pub-id-type="medline">37793630</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Laios</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kalampokis</surname><given-names>E</given-names> </name><name name-style="western"><surname>Mamalis</surname><given-names>ME</given-names> </name><etal/></person-group><article-title>RoBERTa-assisted outcome prediction in ovarian cancer cytoreductive surgery using operative notes</article-title><source>Cancer Control</source><year>2023</year><volume>30</volume><fpage>10732748231209892</fpage><pub-id pub-id-type="doi">10.1177/10732748231209892</pub-id><pub-id pub-id-type="medline">37915208</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Use of BERT (Bidirectional Encoder Representations from Transformers)-based deep learning method for extracting evidences in Chinese radiology reports: development of a computer-aided liver cancer diagnosis framework</article-title><source>J Med Internet Res</source><year>2021</year><month>01</month><day>12</day><volume>23</volume><issue>1</issue><fpage>e19689</fpage><pub-id pub-id-type="doi">10.2196/19689</pub-id><pub-id pub-id-type="medline">33433395</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fink</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Bischoff</surname><given-names>A</given-names> </name><name name-style="western"><surname>Fink</surname><given-names>CA</given-names> </name><etal/></person-group><article-title>Potential of ChatGPT and GPT-4 for data mining of free-text CT reports on lung cancer</article-title><source>Radiology</source><year>2023</year><month>09</month><volume>308</volume><issue>3</issue><fpage>e231362</fpage><pub-id pub-id-type="doi">10.1148/radiol.231362</pub-id><pub-id pub-id-type="medline">37724963</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Guevara</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ramirez</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Natural language processing to automatically extract the presence and severity of esophagitis in notes of patients undergoing radiotherapy</article-title><source>JCO Clin Cancer Inform</source><year>2023</year><month>07</month><volume>7</volume><issue>7</issue><fpage>e2300048</fpage><pub-id pub-id-type="doi">10.1200/CCI.23.00048</pub-id><pub-id pub-id-type="medline">37506330</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lyu</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zapadka</surname><given-names>ME</given-names> </name><etal/></person-group><article-title>Translating radiology reports into plain language using ChatGPT and GPT-4 with prompt learning: results, limitations, and potential</article-title><source>Vis Comput Ind Biomed Art</source><year>2023</year><month>05</month><day>18</day><volume>6</volume><issue>1</issue><fpage>9</fpage><pub-id pub-id-type="doi">10.1186/s42492-023-00136-5</pub-id><pub-id pub-id-type="medline">37198498</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mart&#x00ED;n-Noguerol</surname><given-names>T</given-names> </name><name name-style="western"><surname>L&#x00F3;pez-&#x00DA;beda</surname><given-names>P</given-names> </name><name name-style="western"><surname>Pons-Escoda</surname><given-names>A</given-names> </name><name name-style="western"><surname>Luna</surname><given-names>A</given-names> </name></person-group><article-title>Natural language processing deep learning models for the differential between high-grade gliomas and metastasis: what if the key is how we report them?</article-title><source>Eur Radiol</source><year>2024</year><month>03</month><volume>34</volume><issue>3</issue><fpage>2113</fpage><lpage>2120</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10202-4</pub-id><pub-id pub-id-type="medline">37665389</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fang</surname><given-names>A</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>W</given-names> </name><etal/></person-group><article-title>Extracting clinical named entity for pituitary adenomas from Chinese electronic medical records</article-title><source>BMC Med Inform Decis Mak</source><year>2022</year><month>03</month><day>23</day><volume>22</volume><issue>1</issue><fpage>72</fpage><pub-id pub-id-type="doi">10.1186/s12911-022-01810-z</pub-id><pub-id pub-id-type="medline">35321705</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Rong</surname><given-names>R</given-names> </name><etal/></person-group><article-title>A critical assessment of using ChatGPT for extracting structured data from clinical notes</article-title><source>NPJ Digit Med</source><year>2024</year><month>05</month><day>1</day><volume>7</volume><issue>1</issue><fpage>106</fpage><pub-id pub-id-type="doi">10.1038/s41746-024-01079-8</pub-id><pub-id pub-id-type="medline">38693429</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Elmarakeby</surname><given-names>HA</given-names> </name><name name-style="western"><surname>Trukhanov</surname><given-names>PS</given-names> </name><name name-style="western"><surname>Arroyo</surname><given-names>VM</given-names> </name><etal/></person-group><article-title>Empirical evaluation of language modeling to ascertain cancer outcomes from clinical text reports</article-title><source>BMC Bioinformatics</source><year>2023</year><month>09</month><day>2</day><volume>24</volume><issue>1</issue><fpage>328</fpage><pub-id pub-id-type="doi">10.1186/s12859-023-05439-1</pub-id><pub-id pub-id-type="medline">37658330</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tan</surname><given-names>RSYC</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Low</surname><given-names>GH</given-names> </name><etal/></person-group><article-title>Inferring cancer disease response from radiology reports using large language models with data augmentation and prompting</article-title><source>J Am Med Inform Assoc</source><year>2023</year><month>09</month><day>25</day><volume>30</volume><issue>10</issue><fpage>1657</fpage><lpage>1664</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocad133</pub-id><pub-id pub-id-type="medline">37451682</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Markuzon</surname><given-names>N</given-names> </name><name name-style="western"><surname>Patel</surname><given-names>N</given-names> </name><name name-style="western"><surname>Rueda</surname><given-names>JD</given-names> </name></person-group><article-title>Natural language processing for automated classification of qualitative data from interviews of patients with cancer</article-title><source>Value Health</source><year>2022</year><month>12</month><volume>25</volume><issue>12</issue><fpage>1995</fpage><lpage>2002</lpage><pub-id pub-id-type="doi">10.1016/j.jval.2022.06.004</pub-id><pub-id pub-id-type="medline">35840523</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mitchell</surname><given-names>JR</given-names> </name><name name-style="western"><surname>Szepietowski</surname><given-names>P</given-names> </name><name name-style="western"><surname>Howard</surname><given-names>R</given-names> </name><etal/></person-group><article-title>A question-and-answer system to extract data from free-text oncological pathology reports (CancerBERT Network): development study</article-title><source>J Med Internet Res</source><year>2022</year><month>03</month><day>23</day><volume>24</volume><issue>3</issue><fpage>e27210</fpage><pub-id pub-id-type="doi">10.2196/27210</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Sim</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>JX</given-names> </name><etal/></person-group><article-title>Natural language processing and machine learning methods to characterize unstructured patient-reported outcomes: validation study</article-title><source>J Med Internet Res</source><year>2021</year><month>11</month><day>3</day><volume>23</volume><issue>11</issue><fpage>e26777</fpage><pub-id pub-id-type="doi">10.2196/26777</pub-id><pub-id pub-id-type="medline">34730546</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Percha</surname><given-names>B</given-names> </name><name name-style="western"><surname>Pisapati</surname><given-names>K</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Schmidt</surname><given-names>H</given-names> </name></person-group><article-title>Natural language inference for curation of structured clinical registries from unstructured text</article-title><source>J Am Med Inform Assoc</source><year>2021</year><month>12</month><day>28</day><volume>29</volume><issue>1</issue><fpage>97</fpage><lpage>108</lpage><pub-id pub-id-type="doi">10.1093/jamia/ocab243</pub-id><pub-id pub-id-type="medline">34791282</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Roslidar</surname><given-names>R</given-names> </name><name name-style="western"><surname>Saddami</surname><given-names>K</given-names> </name><name name-style="western"><surname>Arnia</surname><given-names>F</given-names> </name><name name-style="western"><surname>Syukri</surname><given-names>M</given-names> </name><name name-style="western"><surname>Munadi</surname><given-names>K</given-names> </name></person-group><article-title>A study of fine-tuning CNN models based on thermal imaging for breast cancer classification</article-title><conf-name>2019 IEEE International Conference on Cybernetics and Computational Intelligence (CyberneticsCom)</conf-name><conf-date>Aug 22-24, 2019</conf-date><conf-loc>Banda Aceh, Indonesia</conf-loc><fpage>77</fpage><lpage>81</lpage><pub-id pub-id-type="doi">10.1109/CYBERNETICSCOM.2019.8875661</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chougrad</surname><given-names>H</given-names> </name><name name-style="western"><surname>Zouaki</surname><given-names>H</given-names> </name><name name-style="western"><surname>Alheyane</surname><given-names>O</given-names> </name></person-group><article-title>Deep convolutional neural networks for breast cancer screening</article-title><source>Comput Methods Programs Biomed</source><year>2018</year><month>04</month><volume>157</volume><fpage>19</fpage><lpage>30</lpage><pub-id pub-id-type="doi">10.1016/j.cmpb.2018.01.011</pub-id><pub-id pub-id-type="medline">29477427</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nasir</surname><given-names>MU</given-names> </name><name name-style="western"><surname>Ghazal</surname><given-names>TM</given-names> </name><name name-style="western"><surname>Khan</surname><given-names>MA</given-names> </name><etal/></person-group><article-title>Breast cancer prediction empowered with fine-tuning</article-title><source>Comput Intell Neurosci</source><year>2022</year><volume>2022</volume><fpage>5918686</fpage><pub-id pub-id-type="doi">10.1155/2022/5918686</pub-id><pub-id pub-id-type="medline">35720929</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nguyen</surname><given-names>D</given-names> </name><name name-style="western"><surname>Swanson</surname><given-names>D</given-names> </name><name name-style="western"><surname>Newbury</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>YH</given-names> </name></person-group><article-title>Evaluation of ChatGPT and Google Bard using prompt engineering in cancer screening algorithms</article-title><source>Acad Radiol</source><year>2024</year><month>05</month><volume>31</volume><issue>5</issue><fpage>1799</fpage><lpage>1804</lpage><pub-id pub-id-type="doi">10.1016/j.acra.2023.11.002</pub-id><pub-id pub-id-type="medline">38103973</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Khanmohammadi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ghanem</surname><given-names>AI</given-names> </name><name name-style="western"><surname>Verdecchia</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Iterative prompt refinement for radiation oncology symptom extraction using teacher-student large language models</article-title><source>arXiv</source><comment>Preprint posted online on 2024</comment><comment><ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2402.04075">https://arxiv.org/abs/2402.04075</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Savova</surname><given-names>GK</given-names> </name><name name-style="western"><surname>Danciu</surname><given-names>I</given-names> </name><name name-style="western"><surname>Alamudun</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Use of natural language processing to extract clinical cancer phenotypes from electronic medical records</article-title><source>Cancer Res</source><year>2019</year><month>11</month><day>1</day><volume>79</volume><issue>21</issue><fpage>5463</fpage><lpage>5470</lpage><pub-id pub-id-type="doi">10.1158/0008-5472.CAN-19-0579</pub-id><pub-id pub-id-type="medline">31395609</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Jung</surname><given-names>SJ</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>H</given-names> </name><name name-style="western"><surname>Jang</surname><given-names>KS</given-names> </name></person-group><article-title>LLM based biological named entity recognition from scientific literature</article-title><conf-name>2024 IEEE International Conference on Big Data and Smart Computing (BigComp)</conf-name><conf-date>Feb 18-21, 2024</conf-date><conf-loc>Bangkok, Thailand</conf-loc><fpage>433</fpage><lpage>435</lpage><pub-id pub-id-type="doi">10.1109/BigComp60711.2024.00095</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Schilling-Wilhelmi</surname><given-names>M</given-names> </name><name name-style="western"><surname>R&#x00ED;os-Garc&#x00ED;a</surname><given-names>M</given-names> </name><name name-style="western"><surname>Shabih</surname><given-names>S</given-names> </name><etal/></person-group><article-title>From text to insight: large language models for materials science data extraction</article-title><source>arXiv</source><comment>Preprint posted online on  Dec 2, 2024</comment><comment><ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2407.16867">http://arxiv.org/abs/2407.16867</ext-link></comment></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hong</surname><given-names>MKH</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>HHI</given-names> </name><name name-style="western"><surname>Pedersen</surname><given-names>JS</given-names> </name><etal/></person-group><article-title>Error rates in a clinical data repository: lessons from the transition to electronic data transfer--a descriptive study</article-title><source>BMJ Open</source><year>2013</year><month>05</month><day>28</day><volume>3</volume><issue>5</issue><fpage>e002406</fpage><pub-id pub-id-type="doi">10.1136/bmjopen-2012-002406</pub-id><pub-id pub-id-type="medline">23793682</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Meddeb</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ebert</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bressem</surname><given-names>KK</given-names> </name><etal/></person-group><article-title>Evaluating local open-source large language models for data extraction from unstructured reports on mechanical thrombectomy in patients with ischemic stroke</article-title><source>J NeuroIntervent Surg</source><year>2024</year><month>08</month><day>2</day><fpage>jnis-2024-022078</fpage><pub-id pub-id-type="doi">10.1136/jnis-2024-022078</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>H</given-names> </name><name name-style="western"><surname>Lim</surname><given-names>FXY</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>GT</given-names> </name><etal/></person-group><article-title>Natural language processing in urology: Automated extraction of clinical information from histopathology reports of uro-oncology procedures</article-title><source>Heliyon</source><year>2023</year><month>04</month><volume>9</volume><issue>4</issue><fpage>e14793</fpage><pub-id pub-id-type="doi">10.1016/j.heliyon.2023.e14793</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Davenport</surname><given-names>T</given-names> </name><name name-style="western"><surname>Kalakota</surname><given-names>R</given-names> </name></person-group><article-title>The potential for artificial intelligence in healthcare</article-title><source>Future Healthc J</source><year>2019</year><month>06</month><volume>6</volume><issue>2</issue><fpage>94</fpage><lpage>98</lpage><pub-id pub-id-type="doi">10.7861/futurehosp.6-2-94</pub-id><pub-id pub-id-type="medline">31363513</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>De Hert</surname><given-names>S</given-names> </name></person-group><article-title>Burnout in healthcare workers: prevalence, impact and preventative strategies</article-title><source>Local Reg Anesth</source><year>2020</year><volume>13</volume><fpage>171</fpage><lpage>183</lpage><pub-id pub-id-type="doi">10.2147/LRA.S240564</pub-id><pub-id pub-id-type="medline">33149664</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Boehm</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Khosravi</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vanguri</surname><given-names>R</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shah</surname><given-names>SP</given-names> </name></person-group><article-title>Harnessing multimodal data integration to advance precision oncology</article-title><source>Nat Rev Cancer</source><year>2022</year><month>02</month><volume>22</volume><issue>2</issue><fpage>114</fpage><lpage>126</lpage><pub-id pub-id-type="doi">10.1038/s41568-021-00408-3</pub-id><pub-id pub-id-type="medline">34663944</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>D</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>RS</given-names> </name><name name-style="western"><surname>Jomy</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Performance of multimodal artificial intelligence chatbots evaluated on clinical oncology cases</article-title><source>JAMA Netw Open</source><year>2024</year><month>10</month><day>1</day><volume>7</volume><issue>10</issue><fpage>e2437711</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2024.37711</pub-id><pub-id pub-id-type="medline">39441598</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Belyaeva</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cosentino</surname><given-names>J</given-names> </name><name name-style="western"><surname>Hormozdiari</surname><given-names>F</given-names> </name><name name-style="western"><surname>Eswaran</surname><given-names>K</given-names> </name><name name-style="western"><surname>Shetty</surname><given-names>S</given-names> </name><name name-style="western"><surname>Corrado</surname><given-names>G</given-names> </name><etal/></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Maier</surname><given-names>AK</given-names> </name><name name-style="western"><surname>Schnabel</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Tiwari</surname><given-names>P</given-names> </name><name name-style="western"><surname>Stegle</surname><given-names>O</given-names> </name></person-group><article-title>Multimodal llms for health grounded in individual-specific data</article-title><source>Machine Learning for Multimodal Healthcare Data</source><year>2024</year><volume>14315</volume><publisher-name>Springer Nature Switzerland</publisher-name><fpage>86</fpage><lpage>102</lpage><series>Lecture Notes in Computer Science</series><comment><ext-link ext-link-type="uri" xlink:href="https://link.springer.com/10.1007/978-3-031-47679-2_7">https://link.springer.com/10.1007/978-3-031-47679-2_7</ext-link></comment><pub-id pub-id-type="doi">10.1007/978-3-031-47679-2_7</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Truhn</surname><given-names>D</given-names> </name><name name-style="western"><surname>Eckardt</surname><given-names>JN</given-names> </name><name name-style="western"><surname>Ferber</surname><given-names>D</given-names> </name><name name-style="western"><surname>Kather</surname><given-names>JN</given-names> </name></person-group><article-title>Large language models and multimodal foundation models for precision oncology</article-title><source>NPJ Precis Oncol</source><year>2024</year><month>03</month><day>22</day><volume>8</volume><issue>1</issue><fpage>72</fpage><pub-id pub-id-type="doi">10.1038/s41698-024-00573-2</pub-id><pub-id pub-id-type="medline">38519519</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hwang</surname><given-names>KT</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Noh</surname><given-names>DY</given-names> </name><name name-style="western"><surname>Han</surname><given-names>W</given-names> </name><name name-style="western"><surname>Toi</surname><given-names>M</given-names> </name></person-group><article-title>Clinical databases for breast cancer research</article-title><source>Translational Research in Breast Cancer</source><year>2021</year><publisher-name>Springer Singapore</publisher-name><fpage>493</fpage><lpage>509</lpage><series>Advances in Experimental Medicine and Biology</series><pub-id pub-id-type="doi">10.1007/978-981-32-9620-6_26</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gao</surname><given-names>S</given-names> </name><name name-style="western"><surname>Alawad</surname><given-names>M</given-names> </name><name name-style="western"><surname>Young</surname><given-names>MT</given-names> </name><etal/></person-group><article-title>Limitations of transformers on clinical text classification</article-title><source>IEEE J Biomed Health Inform</source><year>2021</year><month>09</month><volume>25</volume><issue>9</issue><fpage>3596</fpage><lpage>3607</lpage><pub-id pub-id-type="doi">10.1109/JBHI.2021.3062322</pub-id><pub-id pub-id-type="medline">33635801</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>D</given-names> </name><name name-style="western"><surname>Parsa</surname><given-names>R</given-names> </name><name name-style="western"><surname>Hope</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Physician and artificial intelligence chatbot responses to cancer questions from social media</article-title><source>JAMA Oncol</source><year>2024</year><month>07</month><day>1</day><volume>10</volume><issue>7</issue><fpage>956</fpage><lpage>960</lpage><pub-id pub-id-type="doi">10.1001/jamaoncol.2024.0836</pub-id><pub-id pub-id-type="medline">38753317</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Samaan</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Yeo</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Rajeev</surname><given-names>N</given-names> </name><etal/></person-group><article-title>Assessing the accuracy of responses by the language model ChatGPT to questions regarding bariatric surgery</article-title><source>Obes Surg</source><year>2023</year><month>06</month><volume>33</volume><issue>6</issue><fpage>1790</fpage><lpage>1796</lpage><pub-id pub-id-type="doi">10.1007/s11695-023-06603-5</pub-id><pub-id pub-id-type="medline">37106269</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Haver</surname><given-names>HL</given-names> </name><name name-style="western"><surname>Ambinder</surname><given-names>EB</given-names> </name><name name-style="western"><surname>Bahl</surname><given-names>M</given-names> </name><name name-style="western"><surname>Oluyemi</surname><given-names>ET</given-names> </name><name name-style="western"><surname>Jeudy</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name></person-group><article-title>Appropriateness of breast cancer prevention and screening recommendations provided by ChatGPT</article-title><source>Radiology</source><year>2023</year><month>05</month><volume>307</volume><issue>4</issue><fpage>e230424</fpage><pub-id pub-id-type="doi">10.1148/radiol.230424</pub-id><pub-id pub-id-type="medline">37014239</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Marvin</surname><given-names>G</given-names> </name><name name-style="western"><surname>Hellen</surname><given-names>N</given-names> </name><name name-style="western"><surname>Jjingo</surname><given-names>D</given-names> </name><name name-style="western"><surname>Nakatumba-Nabende</surname><given-names>J</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Jacob</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Piramuthu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Falkowski-Gilski</surname><given-names>P</given-names> </name></person-group><article-title>Prompt engineering in large language models</article-title><source>Data Intelligence and Cognitive Informatics</source><year>2024</year><publisher-name>Springer Nature Singapore</publisher-name><fpage>387</fpage><lpage>402</lpage><series>Algorithms for Intelligent Systems</series><comment><ext-link ext-link-type="uri" xlink:href="https://link.springer.com/10.1007/978-981-99-7962-2_30">https://link.springer.com/10.1007/978-981-99-7962-2_30</ext-link></comment></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Shi</surname><given-names>E</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Ma</surname><given-names>C</given-names> </name><name name-style="western"><surname>Dai</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Prompt engineering for healthcare: methodologies and applications</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 23, 2024</comment><comment><ext-link ext-link-type="uri" xlink:href="http://arxiv.org/abs/2304.14670">http://arxiv.org/abs/2304.14670</ext-link></comment></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ullah</surname><given-names>E</given-names> </name><name name-style="western"><surname>Parwani</surname><given-names>A</given-names> </name><name name-style="western"><surname>Baig</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>R</given-names> </name></person-group><article-title>Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review</article-title><source>Diagn Pathol</source><year>2024</year><month>02</month><day>27</day><volume>19</volume><issue>1</issue><fpage>43</fpage><pub-id pub-id-type="doi">10.1186/s13000-024-01464-7</pub-id><pub-id pub-id-type="medline">38414074</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>NF</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>K</given-names> </name><name name-style="western"><surname>Hewitt</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Lost in the middle: how language models use long contexts</article-title><source>Trans Assoc Comput Linguist</source><year>2024</year><month>02</month><day>23</day><volume>12</volume><fpage>157</fpage><lpage>173</lpage><pub-id pub-id-type="doi">10.1162/tacl_a_00638</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Scoping review full search strategy for MEDLINE.</p><media xlink:href="cancer_v11i1e65984_app1.docx" xlink:title="DOCX File, 7 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Methodology characteristics of included studies.</p><media xlink:href="cancer_v11i1e65984_app2.xlsx" xlink:title="XLSX File, 12 KB"/></supplementary-material><supplementary-material id="app3"><label>Checklist 1</label><p>PRISMA-ScR reporting guideline.</p><media xlink:href="cancer_v11i1e65984_app3.pdf" xlink:title="PDF File, 677 KB"/></supplementary-material></app-group></back></article>