<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="review-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id><journal-title-group>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title></journal-title-group>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2025.1732759</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Mini Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Large language models for neurology: a mini review</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes"><name><surname>Wunsch III</surname><given-names>Donald C.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2122031/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Hier</surname><given-names>Daniel B.</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/1230804/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Saint Louis University School of Medicine</institution>, <city>St. Louis</city>, <state>MO</state>, <country country="us">United States</country></aff>
<aff id="aff2"><label>2</label><institution>Center for Artificial Intelligence and Autonomous Systems, Kummer Institute, Missouri University of Science and Technology</institution>, <city>Rolla</city>, <state>MO</state>, <country country="us">United States</country></aff>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Donald C. Wunsch III <email xlink:href="mailto:donaldcwunsch@gmail.com">donaldcwunsch@gmail.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-06"><day>06</day><month>01</month><year>2026</year></pub-date>
<pub-date publication-format="electronic" date-type="collection"><year>2025</year></pub-date>
<volume>7</volume><elocation-id>1732759</elocation-id>
<history>
<date date-type="received"><day>26</day><month>10</month><year>2025</year></date>
<date date-type="rev-recd"><day>03</day><month>12</month><year>2025</year></date>
<date date-type="accepted"><day>11</day><month>12</month><year>2025</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 Wunsch III and Hier.</copyright-statement>
<copyright-year>2026</copyright-year><copyright-holder>Wunsch III and Hier</copyright-holder><license><ali:license_ref start_date="2026-01-06">https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p></license>
</permissions>
<abstract>
<p>Large language models have the potential to transform neurology by augmenting diagnostic reasoning, streamlining documentation, and improving workflow efficiency. This Mini Review surveys emerging applications of large language models in Alzheimer&#x2019;s disease, Parkinson&#x2019;s disease, multiple sclerosis, and epilepsy, with emphasis on ambient documentation, multimodal data integration, and clinical decision support. Key barriers to adoption include bias, privacy, reliability, and regulatory alignment. Looking ahead, neurology-focused language models may develop greater fluency in biomedical ontologies and FHIR standards, improving data interoperability and supporting more seamless collaboration between clinicians and AI systems. Two future developments have the potential to be particularly impactful: (1) the integration of multi-omic and neuroimaging data with digital-twin simulations to advance precision neurology, and (2) broader adoption of ambient documentation and other language-model&#x2013;based efficiencies that could reduce administrative and cognitive burden. Ultimately, the clinical success of large language models will depend on continued progress in model robustness, ethical governance, and careful implementation.</p>
</abstract>
<kwd-group>
<kwd>ambient documentation</kwd>
<kwd>digital twins</kwd>
<kwd>documentation burden</kwd>
<kwd>ethical AI</kwd>
<kwd>large language models</kwd>
<kwd>multimodal AI</kwd>
<kwd>neurology</kwd>
<kwd>precision neurology</kwd>
</kwd-group><funding-group><funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement></funding-group><counts>
<fig-count count="1"/>
<table-count count="1"/><equation-count count="4"/><ref-count count="76"/><page-count count="9"/><word-count count="68548"/></counts><custom-meta-group><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Informatics</meta-value></custom-meta></custom-meta-group>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>The ascent of large language models marks a tipping point in clinical medicine, integrating clinical informatics, artificial intelligence (AI), data analytics, and precision medicine into a unified paradigm [<xref ref-type="bibr" rid="B1">1</xref>]. These models now leverage deep learning architectures to process vast datasets, interpret complex clinical relationships, and generate human-like text [<xref ref-type="bibr" rid="B2">2</xref>]. As their scale and contextual depth have expanded, large language models&#x2014;originally designed as next-token predictors&#x2014;have acquired emergent capabilities [<xref ref-type="bibr" rid="B3">3</xref>]: functioning as repositories of medical knowledge, summarizers of clinical text, and reasoning engines that exhibit physician-like competence in diagnosis, prognosis, and therapeutics. Neurology, characterized by complex diagnostic reasoning, detailed clinico-anatomic correlation, extensive unstructured documentation, and integration of multimodal data streams including radiologic and electrophysiologic inputs, is uniquely positioned to benefit from their entry into clinical medicine [<xref ref-type="bibr" rid="B4">4</xref>].</p>
<p>The American Academy of Neurology (AAN) has released a position statement on the use of large language models in neurology [<xref ref-type="bibr" rid="B1">1</xref>]. Their stance is hopeful yet disciplined: these models are framed as a potential breakthrough in neurological efficiency and quality&#x2014;provided they are adopted deliberately, validated rigorously, and governed ethically. The emphasis is on <italic>responsible innovation</italic>, embracing the transformative potential of the technology while acknowledging its immaturity and risks in real-world practice. The AAN statement expresses enthusiasm for the promise of large language models to improve documentation, reduce administrative burden, and enable quality measurement, yet tempers this optimism with concerns about safety, bias, reliability, and governance. Avoiding both technological pessimism and uncritical endorsement, the statement characterizes these systems as powerful but immature tools requiring robust oversight, transparent validation, and ethical governance.</p>
<p>We performed a targeted literature search using Google Scholar (<ext-link ext-link-type="uri" xlink:href="https://scholar.google.com/">https://scholar.google.com/</ext-link>) and Consensus GPT (<ext-link ext-link-type="uri" xlink:href="https://consensus.app/">https://consensus.app/</ext-link>) to identify key primary publications on large language models in neurology, using the search terms <monospace>large language models AND neurology</monospace>. We also conducted forward citation tracking to identify subsequent articles citing these foundational studies, resulting in a core list of 31 relevant publications (<xref ref-type="table" rid="T1">Table&#x00A0;1</xref>). The primary search covered papers published from January 2023 through November 2025. Papers were classified by topic area (e.g., Diagnosis, Documentation, Management) and by type (Review, Research, Opinion). Additional articles were included as needed to ensure completeness of the Mini Review. Papers were selected based on relevance to neurology, contribution to understanding LLM capabilities, and overall quality as independently judged by both co-authors. This process yielded a focused&#x2014;though not exhaustive&#x2014;survey appropriate for a Mini Review.</p>
<table-wrap id="T1" position="float"><label>Table&#x00A0;1</label>
<caption><p>Core references utilized in mini review.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Author</th>
<th valign="top" align="center">Year</th>
<th valign="top" align="center">Topic</th>
<th valign="top" align="center">Type</th>
<th valign="top" align="center">Comment</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Jones</td>
<td valign="top" align="center">2022</td>
<td valign="top" align="left">AI in neurology</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">Forecast for 2035</td>
</tr>
<tr>
<td valign="top" align="left">Moura</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">AI in neurology</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">AAN position statement</td>
</tr>
<tr>
<td valign="top" align="left">Rizzo</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">AI in neurology</td>
<td valign="top" align="left">Review</td>
<td valign="top" align="left">Comprehensive overview</td>
</tr>
<tr>
<td valign="top" align="left">Romano</td>
<td valign="top" align="center">2023</td>
<td valign="top" align="left">AI in neurology</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">Ethical challenges</td>
</tr>
<tr>
<td valign="top" align="left">Westover</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">AI in neurology</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">Risks vs. benefits</td>
</tr>
<tr>
<td valign="top" align="left">Barrit</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLM outperforms neurologists</td>
</tr>
<tr>
<td valign="top" align="left">Cano-Besquet</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLMs comparable to neurologists</td>
</tr>
<tr>
<td valign="top" align="left">Ford</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Limited accuracy on seizures</td>
</tr>
<tr>
<td valign="top" align="left">Habibi</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Limited accuracy</td>
</tr>
<tr>
<td valign="top" align="left">Joseph</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Limited accuracy for MS</td>
</tr>
<tr>
<td valign="top" align="left">Maiorana</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Neurologists outperform LLMs</td>
</tr>
<tr>
<td valign="top" align="left">Qadri</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Review</td>
<td valign="top" align="left">Dementia diagnosis</td>
</tr>
<tr>
<td valign="top" align="left">Song</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Stroke diagnosis from notes</td>
</tr>
<tr>
<td valign="top" align="left">Sorka</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLM outperformed neurologists</td>
</tr>
<tr>
<td valign="top" align="left">Twala</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Multimodal ML for parkinsonism</td>
</tr>
<tr>
<td valign="top" align="left">Yang</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLM identifies seizure locus</td>
</tr>
<tr>
<td valign="top" align="left">Zamai</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Diagnosis</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Dementia MRI interpretation</td>
</tr>
<tr>
<td valign="top" align="left">Chadehumbe</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Documentation</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Workflow with ambient AI</td>
</tr>
<tr>
<td valign="top" align="left">Ge</td>
<td valign="top" align="center">2023</td>
<td valign="top" align="left">Documentation</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">Benefits vs. risks</td>
</tr>
<tr>
<td valign="top" align="left">Chiang</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Localization</td>
<td valign="top" align="left">Opinion</td>
<td valign="top" align="left">Cautious optimism</td>
</tr>
<tr>
<td valign="top" align="left">Dani</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Localization</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Localizes seizure zones</td>
</tr>
<tr>
<td valign="top" align="left">Lee</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Localization</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">High accuracy on stroke cases</td>
</tr>
<tr>
<td valign="top" align="left">Amin</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Disease management</td>
<td valign="top" align="left">Review</td>
<td valign="top" align="left">MS management</td>
</tr>
<tr>
<td valign="top" align="left">Harrison</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Disease management</td>
<td valign="top" align="left">Review</td>
<td valign="top" align="left">Dementia management</td>
</tr>
<tr>
<td valign="top" align="left">Naji</td>
<td valign="top" align="center">2023</td>
<td valign="top" align="left">Disease management</td>
<td valign="top" align="left">Review</td>
<td valign="top" align="left">MS management</td>
</tr>
<tr>
<td valign="top" align="left">Mavrych</td>
<td valign="top" align="center">2025</td>
<td valign="top" align="left">Question answering</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Advanced LLMs did better</td>
</tr>
<tr>
<td valign="top" align="left">Ros-Arlanzon</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Question answering</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">GPT-4 outperformed GPT-3.5</td>
</tr>
<tr>
<td valign="top" align="left">Schubert</td>
<td valign="top" align="center">2023</td>
<td valign="top" align="left">Question answering</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLMs outperformed humans</td>
</tr>
<tr>
<td valign="top" align="left">Shojaee-Mend</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Question answering</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">Limitations noted</td>
</tr>
<tr>
<td valign="top" align="left">Shu</td>
<td valign="top" align="center">2024</td>
<td valign="top" align="left">Question answering</td>
<td valign="top" align="left">Research</td>
<td valign="top" align="left">LLM outperformed humans</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="TF1"><p>LLM, Large Language Model; MS, Multiple Sclerosis; MRI, Magnetic Resonance Imaging; ML, Machine Learning; AAN, American Academy of Neurology.</p></fn>
<fn id="TF2"><p>All works listed in this table are cited in the main text and appear in the reference list.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>In this Mini Review, we examine how large language models may improve neurological care and reduce documentation and administrative burdens. We review assessments of neurological knowledge, including foundational neuroscience understanding, lesion localization, and diagnostic reasoning. We then discuss their role in workflow efficiency, disease management, and the challenges that accompany their implementation. Finally, we present our predictions for the state of neurology-focused large language models in 2035 [<xref ref-type="bibr" rid="B5">5</xref>].</p>
</sec>
<sec id="s2"><label>2</label><title>Assessing neurological competency in large language models</title>
<p>Large language models have demonstrated emerging capabilities across several domains of neurology&#x2014;including board&#x2014;style question answering, neurological diagnosis, and lesion localization. Proficiency in these domains is essential for establishing the credibility and clinical utility of a large language model.</p>
<sec id="s2a"><label>2.1</label><title>Neurological knowledge</title>
<p>Large language models have demonstrated substantial progress in mastering the foundational knowledge base of neurology, as reflected in their performance on standardized board-style examinations. GPT-4 correctly answered 82&#x0025; of U.S. neurology board-style questions (<italic>n</italic> = 1,956), surpassing GPT-3.5 (66&#x0025;) and approaching the lower range of specialist performance [<xref ref-type="bibr" rid="B6">6</xref>]. Three large language models (Bard, Claude 2, and GPT-3.5) were evaluated on 20 essay-style advanced neurophysiology questions, which were scored by physiologists on a 0&#x2013;5 scale, yielding a mean score of 3.9/5 across models [<xref ref-type="bibr" rid="B7">7</xref>]. On the Spanish Neurology Specialist Examination (77 multiple-choice questions), GPT-4 scored 81.8&#x0025; correct&#x2014;ranking seventeenth among 120 neurologists who took the examination [<xref ref-type="bibr" rid="B8">8</xref>]. On 200 multiple-choice questions comparable to the neuroscience section of the United States Medical Licensing Examination (USMLE), Claude (88.0&#x0025;) and GPT-4 (81.7&#x0025;) outperformed the student average (74.6&#x0025;) [<xref ref-type="bibr" rid="B9">9</xref>]. On the NeuroReady&#x00AE; board preparation question bank (<italic>n</italic> = 400), GPT-4 scored 75&#x0025;, exceeding the passing threshold (70&#x0025;) and the average test-taker score (69&#x0025;) [<xref ref-type="bibr" rid="B10">10</xref>]. Collectively, these findings suggest that advanced large language models such as GPT-4 have reached near-human competence in factual neurology knowledge [<xref ref-type="bibr" rid="B6">6</xref>&#x2013;<xref ref-type="bibr" rid="B12">12</xref>].</p>
</sec>
<sec id="s2b"><label>2.2</label><title>Neurological lesion localization</title>
<p>Accurate lesion localization based on signs and symptoms remains a defining cognitive skill of neurologists [<xref ref-type="bibr" rid="B13">13</xref>&#x2013;<xref ref-type="bibr" rid="B15">15</xref>]. In a structured evaluation of 46 acute stroke vignettes derived from published cases, GPT-4 localized brain lesions with an F1 score of 0.85 for brain region and 0.74 for lesion side. Although no direct human comparison was performed, performance varied by region, with best results for cerebral and spinal lesions and poorest for cerebellar lesions [<xref ref-type="bibr" rid="B16">16</xref>]. Most localization errors arose from incomplete input data or reasoning gaps rather than factual hallucination, supporting GPT-4&#x2019;s potential for structured neuroanatomical reasoning.</p>
</sec>
<sec id="s2c"><label>2.3</label><title>Neurological diagnosis</title>
<p>The diagnostic accuracy of large language models in neurology varies by setting and model. Current evidence suggests that while general-purpose large language models lag behind human neurologists in real-world practice, domain-specialized models can approach or exceed human performance in selected scenarios.</p>
<p>In a comparative study using 28 real-world anonymized patients, neurologists achieved substantially higher diagnostic accuracy (75&#x0025;) than general-purpose large language models such as GPT-3.5 (54&#x0025;) and Gemini (46&#x0025;) [<xref ref-type="bibr" rid="B17">17</xref>]. These models struggled with complex clinical reasoning, contextual integration, and subtle diagnostic differentiation.</p>
<p>Similarly, OpenBioLLM&#x2014;a domain-specialized model&#x2014;achieved only 38&#x0025; diagnostic accuracy on 25 cases from the textbook <italic>Clinical Cases in Neurology</italic>. Although the model frequently identified relevant symptoms and correctly localized the lesion, it often failed to arrive at an accurate etiologic diagnosis [<xref ref-type="bibr" rid="B18">18</xref>].</p>
<p>In contrast, the neurology-specialized model <italic>Neura</italic> demonstrated substantially higher diagnostic capability [<xref ref-type="bibr" rid="B19">19</xref>]. In a blinded comparison with 13 neurologists using five difficult <italic>Clinical Reasoning</italic> cases from the journal <bold>Neurology</bold>, <italic>Neura</italic> achieved scores of 86&#x0025; overall, 85&#x0025; for differential diagnosis, and 88&#x0025; for final diagnosis. Neurologists scored 55&#x0025;, 46&#x0025;, and 71&#x0025;, respectively. <italic>Neura</italic> produced rapid, citation-backed responses with minimal hallucination.</p>
<p>General-purpose large language models have also shown strong performance in specific tasks such as seizure localization. When presented with 1,269 clinical epilepsy narratives, Mistral-8x7B (F1 = 51.7) and GPT-4 (F1 = 52.3) localized epileptogenic zones across seven brain regions as accurately as expert neurologists (F1 = 48.8) [<xref ref-type="bibr" rid="B20">20</xref>].</p>
</sec>
</sec>
<sec id="s3"><label>3</label><title>Neurologist workflow and efficiency</title>
<p>Large language models have the potential to improve the workflow and efficiency of neurologists. They can generate patient-specific educational materials in real-time by drawing directly from the electronic health record (EHR), reducing the need for manual customization [<xref ref-type="bibr" rid="B21">21</xref>]. In the outpatient setting, ambient AI systems can capture and document the patient encounter&#x2014;recording the history, examination findings, assessment, and plan&#x2014;thereby reducing documentation burden on the neurologist [<xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B23">23</xref>]. In a pilot study, 10 of 13 neurology providers reported improved efficiency by implementing ambient AI for documentation of outpatient neurology visits [<xref ref-type="bibr" rid="B23">23</xref>]. Large language model&#x2014;based summarization tools can expedite pre-visit preparation by synthesizing entire EHRs into concise, clinically relevant overviews. Additional efficiency gains are possible through faster creation of discharge summaries, accelerated drug authorizations, improved prior approval workflows, streamlined scheduling for diagnostic testing, and automated chart coding [<xref ref-type="bibr" rid="B24">24</xref>]. Together, these capabilities of large language models could meaningfully ease the administrative and documentation burden faced by neurologists.</p>
</sec>
<sec id="s4"><label>4</label><title>Disease-specific applications</title>
<sec id="s4a"><label>4.1</label><title>Stroke</title>
<p>Acute stroke management is time-critical and data-intensive, making it an ideal domain for integration with large language models. Modern stroke care produces large volumes of unstructured text&#x2014;from triage notes, radiology findings, and procedural reports&#x2014;that must be interpreted rapidly for treatment decisions [<xref ref-type="bibr" rid="B25">25</xref>]. Large language models optimized for stroke care can integrate clinical narratives with non-contrast CT findings to diagnose stroke, assess eligibility for intravenous thrombolysis, and detect large vessel occlusions. Song et al. [<xref ref-type="bibr" rid="B26">26</xref>] fine-tuned ChatGLM-6B based on 1,885 patients with and without stroke. Patients were divided into training and validation sets. The model distinguished between hemorrhage and infarction with 100&#x0025; accuracy, identified large vessel occlusions with 80&#x0025; accuracy, and screened patients for intravenous thrombolysis with 89.4&#x0025; accuracy. The model input was a non-contrast CT scan and the clinical notes [<xref ref-type="bibr" rid="B26">26</xref>]. These findings highlight the potential of large language models to streamline time-sensitive stroke workflows, speed diagnosis, and enhance decision support.</p>
</sec>
<sec id="s4b"><label>4.2</label><title>Alzheimer&#x2019;s disease and dementia</title>
<p>Large language models are increasingly applied to the diagnosis and management of neurodegenerative disorders, particularly Alzheimer&#x2019;s disease. They show promise for early detection by identifying subtle, complex symptom patterns within unstructured clinical text that may escape human recognition. Harrison et al. [<xref ref-type="bibr" rid="B27">27</xref>] provide a comprehensive review of emerging applications of large language models in improving Alzheimer&#x2019;s disease diagnosis. Qadri et al. [<xref ref-type="bibr" rid="B28">28</xref>] review the utility of large language models in the diagnosis of neurodegenerative disorders such as Alzheimer&#x2019;s disease and Parkinson&#x2019;s disease. In an innovative approach, Zamai et al. [<xref ref-type="bibr" rid="B29">29</xref>] used fine-tuned large language models to classify 615 MRI images into four diagnostic categories (normal, Alzheimer&#x2019;s disease, frontotemporal dementia, and primary progressive aphasia). The authors first converted each MRI into a synthetic radiology-style text report (an image-to-text approach) and then fine-tuned a language model to interpret these reports. With fine-tuning, Qwen-3.1-8B achieved a balanced accuracy of 68.4&#x0025;, outperforming GPT-4o (balanced accuracy 55.5&#x0025;) on the same classification task [<xref ref-type="bibr" rid="B29">29</xref>].</p>
</sec>
<sec id="s4c"><label>4.3</label><title>Parkinson&#x2019;s disease</title>
<p>The diagnosis of Parkinson&#x2019;s disease relies heavily on clinical observation of motor symptoms, which can lead to delayed or uncertain diagnosis in early stages. Advanced AI frameworks that integrate deep learning with natural language processing have been developed to analyze voice, gait, and motor patterns, enabling detection of subtle features of parkinsonism that may escape routine clinical examination. Twala [<xref ref-type="bibr" rid="B30">30</xref>] evaluated a multimodal model that combined gait analysis, voice analysis, and visual motor assessments to classify 847 synthetic patient profiles as having Parkinson&#x2019;s disease or not. Using this synthetic dataset, the system achieved a diagnostic accuracy of 94.2&#x0025;. Although these results are preliminary and limited by reliance on synthetic data, they highlight the potential of multimodal AI systems to enhance early detection of Parkinson&#x2019;s disease.</p>
</sec>
<sec id="s4d"><label>4.4</label><title>Multiple sclerosis</title>
<p>Large language models such as GPT-4 are being evaluated for classifying multiple sclerosis (MS) status based on clinical notes. When aligned with diagnostic frameworks such as the 2017 McDonald criteria, they achieve classification accuracies of up to 74&#x0025; [<xref ref-type="bibr" rid="B31">31</xref>]. Venkatesh et al. [<xref ref-type="bibr" rid="B31">31</xref>] used GPT-4 to reclassify 125 patients (105 with MS, 10 with related disorders, and 10 healthy controls) based on their clinical notes that included laboratory findings. GPT-4 correctly classified 74&#x0025; (93/125) patients, including 70&#x0025; of patients with MS, 100&#x0025; of related disorders, and 90&#x0025; of healthy controls. Large language model&#x2013;based systems are increasingly applied to prognosis, integrating MRI data and large clinical registries to identify key predictors of disease progression and treatment response in multiple sclerosis [<xref ref-type="bibr" rid="B32">32</xref>].</p>
</sec>
<sec id="s4e"><label>4.5</label><title>Epilepsy and seizures</title>
<p>In epilepsy, large language models have been evaluated for challenging diagnostic distinctions, such as differentiating epileptic seizures from functional or dissociative seizures. In a study using patient-generated symptom descriptions, Ford et al. [<xref ref-type="bibr" rid="B33">33</xref>] tested GPT-4 on 41 cases (16 epilepsy, 25 functional/dissociative seizures). In the zero-shot condition, GPT-4 achieved a balanced accuracy of 57&#x0025; (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM1"><mml:mi>&#x03BA;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.15</mml:mn></mml:math></inline-formula>), which improved to 64&#x0025; (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM2"><mml:mi>&#x03BA;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.27</mml:mn></mml:math></inline-formula>) after a single example (one-shot prompting). Additional examples (two- and three-shot) did not further improve performance. In contrast, three experienced neurologists achieved a mean balanced accuracy of 71&#x0025; (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM3"><mml:mi>&#x03BA;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.42</mml:mn></mml:math></inline-formula>). Notably, in the subset of 18 cases correctly diagnosed by all three neurologists, GPT-4 achieved a balanced accuracy of 81&#x0025; (<inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM4"><mml:mi>&#x03BA;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.66</mml:mn></mml:math></inline-formula>), suggesting that performance improves substantially when clinical descriptions are clear and internally consistent. Large language models have also been explored for presurgical planning. Fine-tuned systems can localize seizure origins to epileptogenic zones using clinical narratives, streamlining early stages of the preoperative workflow [<xref ref-type="bibr" rid="B34">34</xref>]. In addition, language models are increasingly used to assist with generating structured reports for electroencephalograms (EEG), electromyograms (EMG), and nerve conduction studies [<xref ref-type="bibr" rid="B25">25</xref>].</p>
<p>Taken together, these disease-specific applications should be viewed as exploratory rather than definitive. Most studies used small or moderate sample sizes, were conducted at single centers, or relied on synthetic data, vignettes, or retrospectively assembled datasets. External validation was limited, and few evaluations tested performance in real-time clinical workflows. As a result, the reported accuracies are best interpreted as suggestive of what large language models may be able to do under controlled conditions, rather than as evidence that they are ready for routine clinical deployment.</p>
</sec>
</sec>
<sec id="s5"><label>5</label><title>Challenges and controversies</title>
<p>The widespread adoption of large language models in neurology faces significant challenges and controversies that require debate and resolution.</p>
<sec id="s5a"><label>5.1</label><title>Bias</title>
<p>Bias arises from imbalanced training data. The overrepresentation of specific populations, institutions, or languages produce systematic inequities in diagnosis, classification, and treatment recommendations [<xref ref-type="bibr" rid="B25">25</xref>, <xref ref-type="bibr" rid="B35">35</xref>&#x2013;<xref ref-type="bibr" rid="B37">37</xref>].</p>
</sec>
<sec id="s5b"><label>5.2</label><title>Privacy</title>
<p>Large language models risk re-identifying protected health information through memorization or inadvertent data exposure. Breaches, leaks, and technical vulnerabilities must be managed through robust encryption, audit trails, and data-minimization protocols [<xref ref-type="bibr" rid="B38">38</xref>].</p>
</sec>
<sec id="s5c"><label>5.3</label><title>Trust</title>
<p>Trust depends on reliability, validity, and explainability. Reliability ensures consistent results; validity aligns with clinical ground truth; and explainability enables oversight and clinician confidence [<xref ref-type="bibr" rid="B39">39</xref>&#x2013;<xref ref-type="bibr" rid="B43">43</xref>].</p>
</sec>
<sec id="s5d"><label>5.4</label><title>Accurate model inputs</title>
<p>Neurological reasoning depends heavily on subtle bedside findings that must be correctly observed and documented by the clinician [<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B44">44</xref>]. Large language models can only reason over what is recorded. This reflects the classical Garbage In, Garbage Out (GIGO) principle: incomplete or imprecise clinical inputs propagate into incomplete or erroneous model outputs. No current AI system can substitute for the neurologist&#x2019;s direct examination; the fidelity of LLM- or LMM-generated reasoning is ultimately constrained by the quality of the clinician&#x2019;s initial observations.</p>
</sec>
<sec id="s5e"><label>5.5</label><title>Regulation</title>
<p>Some large language models may qualify as medical devices, requiring validation, traceability, and postmarket surveillance. Current regulatory frameworks remain incomplete, demanding coordinated oversight among developers, clinicians, and regulators [<xref ref-type="bibr" rid="B45">45</xref>&#x2013;<xref ref-type="bibr" rid="B47">47</xref>].</p>
</sec>
<sec id="s5f"><label>5.6</label><title>Multimodal integration</title>
<p>Neurological diagnosis requires synthesizing text, imaging, and physiologic signals. While contemporary large language models (LLMs) are primarily trained on text and therefore depend on narrative descriptions of MRI, EEG, and EMG findings, emerging <italic>large multimodal models</italic> (LMMs) can natively integrate information from multiple data streams [<xref ref-type="bibr" rid="B48">48</xref>]. These multimodal architectures show promise in harmonizing radiologic, electrophysiologic, and textual inputs, potentially reducing the need for intermediate text-based summaries in the future [<xref ref-type="bibr" rid="B49">49</xref>&#x2013;<xref ref-type="bibr" rid="B53">53</xref>]. Although the terminology is still evolving, and the boundary between LLMs and LMMs is not yet fixed, neurology is likely to benefit disproportionately from models that can directly process high-dimensional clinical signals.</p>
</sec>
<sec id="s5g"><label>5.7</label><title>Keeping neurology large language models current</title>
<p>Retraining and manual updates are slow and prone to catastrophic forgetting. Retrieval-augmented generation (RAG) architectures provide a more scalable solution, coupling dynamic knowledge sources with stable reasoning engines [<xref ref-type="bibr" rid="B54">54</xref>, <xref ref-type="bibr" rid="B55">55</xref>].</p>
</sec>
<sec id="s5h"><label>5.8</label><title>Specialized vs. foundation models</title>
<p>Foundation models offer scalability and multimodal reasoning, but domain-specific models achieve higher accuracy and lower hallucination rates for clinical tasks [<xref ref-type="bibr" rid="B19">19</xref>].</p>
</sec>
<sec id="s5i"><label>5.9</label><title>From textbook knowledge to real-world usability</title>
<p>Large language models perform well on structured examinations, yet show variable accuracy in real clinical contexts. Real-world validation remains the decisive test of their readiness for clinical use [<xref ref-type="bibr" rid="B56">56</xref>].</p>
</sec>
<sec id="s5j"><label>5.10</label><title>Evolving skills that support neurologist workflow</title>
<p>Neurology large language models should target reduction of documentation burden, automation of text summarization, and seamless integration with the EHR. EHR burden remains one of the greatest challenges facing neurologists [<xref ref-type="bibr" rid="B57">57</xref>, <xref ref-type="bibr" rid="B58">58</xref>]. The design of large language models for neurologists must reflect clinician priorities&#x2014;not replace&#x2014;neurologist expertise [<xref ref-type="bibr" rid="B59">59</xref>, <xref ref-type="bibr" rid="B60">60</xref>].</p>
</sec>
</sec>
<sec id="s6"><label>6</label><title>Neurology large language models in 2035</title>
<p>We agree with Jones and Kerber [<xref ref-type="bibr" rid="B5">5</xref>] that, by 2035, neurology-focused large language models may evolve from static repositories of neurologic knowledge into more dynamic, ontology-aware computational tools that support neurological reasoning and synchronize with emerging biomedical information. Several developments could unfold along three major dimensions (<xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>).</p>
<fig id="F1" position="float"><label>Figure&#x00A0;1</label>
<caption><p>Conceptual overview of how neurology-focused large language models may transform neurology by 2035 along three major dimensions: (1) neurologist&#x2013;AI collaboration, (2) precision and personalization through digital twins and multi-omic integration, and (3) ontology awareness and FHIR-native interoperability enabling a computable EHR.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-07-1732759-g001.tif"><alt-text content-type="machine-generated">Flowchart illustrating the role of neurology-focused large language models in 2035. The central box points to three areas: \"Neurologist&#x2013;AI collaboration\" (enhancing diagnosis and reducing EHR burden), \"Precision Neurology\" (using multi-omics and digital twins), and \"Data Interoperability\" (with standards like SNOMED and FHIR).</alt-text>
</graphic>
</fig>
<sec id="s6a"><label>6.1</label><title>The evolving neurologist&#x2013;AI collaboration</title>
<p>Large language models may function as active cognitive partners&#x2014;suggesting differential diagnoses, proposing tests, or challenging initial hypotheses&#x2014;while neurologists retain executive authority. The clinician&#x2019;s role could shift subtly from synthesizing raw data toward validating and contextualizing AI recommendations, particularly in ambiguous or ethically complex cases [<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B25">25</xref>]. It is possible that advances in language-model&#x2013;driven documentation will reduce the long-standing burdens associated with electronic health records, although the extent of such improvement will depend on technical reliability and clinical integration.</p>
</sec>
<sec id="s6b"><label>6.2</label><title>The emergence of precision and personalized neurology</title>
<p>Integration of digital twins and multi-omic data may enable more personalized disease trajectory forecasting and treatment selection [<xref ref-type="bibr" rid="B61">61</xref>&#x2013;<xref ref-type="bibr" rid="B64">64</xref>]. Future multimodal frameworks could cross-validate predictions to reduce bias and achieve greater fluency across text, imaging, and physiological signal data [<xref ref-type="bibr" rid="B8">8</xref>]. AI-generated digital twins [<xref ref-type="bibr" rid="B65">65</xref>&#x2013;<xref ref-type="bibr" rid="B67">67</xref>] might one day support <italic>in silico</italic> comparisons of therapeutic strategies&#x2014;analogous to virtual &#x201C;clinical trials of one&#x201D;&#x2014;for disorders such as multiple sclerosis or Parkinson&#x2019;s disease. Whether such simulations will become routine clinical tools remains uncertain and will depend on regulatory oversight, validation, and clinician acceptance.</p>
</sec>
<sec id="s6c"><label>6.3</label><title>Infrastructure and interoperability for neurological data</title>
<p>A persistent barrier to real-world deployment of neurology-focused AI is the transformation of free-text clinical documentation into structured, computable, and interoperable data. Neurology-specific models may eventually achieve deeper fluency with biomedical ontologies&#x2014;including SNOMED CT, HPO, GO, RxNorm, ICD-CM and with FHIR resource standards, thereby improving the consistency and computability of clinical data. Continuous alignment with curated databases and knowledge graphs may further reduce the need for frequent retraining, although such capabilities remain aspirational at present [<xref ref-type="bibr" rid="B68">68</xref>, <xref ref-type="bibr" rid="B69">69</xref>]. Real progress will require: (1) consistent use of standardized vocabularies to encode diagnoses, medications, laboratory values, and neurological findings; (2) broad adoption of FHIR resources and profiles to represent encounters, observations, imaging studies, and procedures; and (3) robust NLP and mapping pipelines that convert narrative notes into coded concepts without losing clinically relevant nuance. Our preliminary work suggests that large language models can preprocess physician-written notes to improve downstream extraction of standard ontology terms and facilitate FHIR resource generation [<xref ref-type="bibr" rid="B70">70</xref>, <xref ref-type="bibr" rid="B71">71</xref>]. Given that 60&#x0025;&#x2013;80&#x0025; of clinically relevant information in U.S. electronic health records remains buried in free text [<xref ref-type="bibr" rid="B72">72</xref>, <xref ref-type="bibr" rid="B73">73</xref>], expanding computability could meaningfully enhance both clinical care and research. However, this transition will require secure data platforms, governance frameworks, and auditable interfaces that allow AI systems to query and write back to the EHR safely. Without such infrastructure, even highly capable neurology models will remain confined to pilot settings rather than routine clinical practice.</p>
</sec>
</sec>
<sec id="s7" sec-type="discussion"><label>7</label><title>Discussion</title>
<p>Before transformer-based large language models [<xref ref-type="bibr" rid="B74">74</xref>], artificial intelligence in neurology remained largely confined to narrow research prototypes. With the advent of large language models, diagnostic reasoning, documentation, and workflow assistance have moved from aspiration to implementation&#x2014;a transition from theory to practice. Neurology has always been intellectually demanding, yet neurologists rarely cite cognitive challenge as their source of fatigue. Rather, burnout stems from administrative overload and documentation burden [<xref ref-type="bibr" rid="B57">57</xref>, <xref ref-type="bibr" rid="B58">58</xref>]. Large language models now offer a pragmatic remedy: relieving the clerical weight that adds to cognitive work while amplifying the neurologist&#x2019;s capacity for insight and care.</p>
<p>Beyond efficiency, neurology-focused large language models herald a new integrative intelligence. As these systems evolve into large multimodal models (LMMs) capable of processing diverse data modalities, they can correlate multi-omic data (radiomic, proteomic, genomic, phenomic), fuse multimodal streams (text, imaging, waveforms), interpret digital-twin simulations, retrieve and summarize biomedical literature, and distill electronic health records into structured, comprehensible narratives [<xref ref-type="bibr" rid="B49">49</xref>, <xref ref-type="bibr" rid="B51">51</xref>]. Increasingly, these models are coupled with retrieval-augmented generation (RAG) systems, allowing them to access curated, continuously updated corpora of guidelines, trial results, and institutional protocols rather than relying solely on static parametric memory. For each of these data streams, the model acts as both purveyor and interpreter&#x2014;a never-tiring colleague who assists in recollection and reasoning within a setting of great complexity. An underappreciated capability of these emerging architectures is their capacity to explain complex machine-learning outputs and multimodal data flows. As clinical AI systems incorporate increasingly sophisticated pipelines&#x2014;spanning imaging models, temporal predictors, and graph-based representations&#x2014;multimodal LMMs with RAG can serve as interpreters of this complexity, translating opaque analytical steps into clinically meaningful explanations that support oversight, safety, and trust. As these capabilities mature, such systems will not replace the neurologist&#x2019;s judgment; rather, they will reinforce it by clarifying what is known, suggesting what is possible, and documenting what has been done.</p>
<p>The promise of neurology-focused large language models is tempered by significant technical, ethical, and practical barriers [<xref ref-type="bibr" rid="B75">75</xref>]. The use of large language models in neurology will require continual updating of their factual foundations as well as expansion of their cognitive and procedural skill sets. Bias must be measured and mitigated, privacy protected, and trust earned through transparency and validation. Regulation must evolve to keep pace with increasingly capable systems. Multimodal data fusion&#x2014;essential for integrating textual, imaging, and physiologic data&#x2014;remains an unfinished scientific project [<xref ref-type="bibr" rid="B75">75</xref>, <xref ref-type="bibr" rid="B76">76</xref>]. The alignment of structured medical knowledge with the realities of clinical practice also lags behind expectations. Ultimate success will depend on increasing technological sophistication that is coupled with sustained engagement of neurologists, data scientists, implementers, and regulators to ensure that these systems amplify, rather than erode, clinical judgment [<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B75">75</xref>].</p>
</sec>
</body>
<back>
<sec id="s8" sec-type="author-contributions"><title>Author contributions</title>
<p>DW: Conceptualization, Formal analysis, Investigation, Methodology, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. DH: Conceptualization, Methodology, Project administration, Supervision, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec id="s10" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s11" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declared that generative AI was used in the creation of this manuscript. Portions of this manuscript-including literature searching, organization, and language refinement-were supported by AI-assisted tools (Consensus GPT and ChatGPT). All scientific judgments, interpretations, and final editing decisions were made by the authors. No text, data, or references were accepted without verification, and all content was reviewed for accuracy, originality, and compliance with journal guidelines.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="s12" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moura</surname> <given-names>L</given-names></name> <name><surname>Jones</surname> <given-names>DT</given-names></name> <name><surname>Sheikh</surname> <given-names>IS</given-names></name> <name><surname>Murphy</surname> <given-names>S</given-names></name> <name><surname>Kalfin</surname> <given-names>M</given-names></name> <name><surname>Kummer</surname> <given-names>BR</given-names></name><etal/></person-group> <article-title>Implications of large language models for quality and efficiency of neurologic care: emerging issues in neurology</article-title>. <source>Neurology</source>. (<year>2024</year>) <volume>102</volume>(<issue>11</issue>):<fpage>e209497</fpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000209497</pub-id><pub-id pub-id-type="pmid">38759131</pub-id></mixed-citation></ref>
<ref id="B2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rhee</surname> <given-names>J</given-names></name> <name><surname>Sounack</surname> <given-names>T</given-names></name> <name><surname>Tentor</surname> <given-names>Z</given-names></name> <name><surname>Davis</surname> <given-names>J</given-names></name> <name><surname>Durieux</surname> <given-names>B</given-names></name> <name><surname>Miller</surname> <given-names>P</given-names></name><etal/></person-group> <article-title>Validating patient symptoms in the electronic health record with large language models for scalable tracking of symptoms in neuro-oncology (p3-11.004)</article-title>. <source>Neurology</source>. (<year>2025</year>) <volume>104</volume>(<issue>7&#x005F;Supplement&#x005F;1</issue>):<fpage>2978</fpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000210712</pub-id></mixed-citation></ref>
<ref id="B3"><label>3.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Berti</surname> <given-names>L</given-names></name> <name><surname>Giorgi</surname> <given-names>F</given-names></name> <name><surname>Kasneci</surname> <given-names>G</given-names></name></person-group>. <article-title>Emergent abilities in large language models: a survey</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2503.05788</italic> (2025)</comment>.</mixed-citation></ref>
<ref id="B4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ge</surname> <given-names>W</given-names></name> <name><surname>Rice</surname> <given-names>HJ</given-names></name> <name><surname>Sheikh</surname> <given-names>IS</given-names></name> <name><surname>Westover</surname> <given-names>MB</given-names></name> <name><surname>Weathers</surname> <given-names>AL</given-names></name> <name><surname>Jones</surname> <given-names>LK</given-names></name><etal/></person-group> <article-title>Improving neurology clinical care with natural language processing tools</article-title>. <source>Neurology</source>. (<year>2023</year>) <volume>101</volume>(<issue>22</issue>):<fpage>1010</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000207853</pub-id><pub-id pub-id-type="pmid">37816638</pub-id></mixed-citation></ref>
<ref id="B5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jones</surname> <given-names>DT</given-names></name> <name><surname>Kerber</surname> <given-names>KA</given-names></name></person-group>. <article-title>Artificial intelligence and the practice of neurology in 2035: the neurology future forecasting series</article-title>. <source>Neurology</source>. (<year>2022</year>) <volume>98</volume>(<issue>6</issue>):<fpage>238</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000013200</pub-id><pub-id pub-id-type="pmid">35131918</pub-id></mixed-citation></ref>
<ref id="B6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schubert</surname> <given-names>MC</given-names></name> <name><surname>Wick</surname> <given-names>W</given-names></name> <name><surname>Venkataramani</surname> <given-names>V</given-names></name></person-group>. <article-title>Performance of large language models on a neurology board&#x2013;style examination</article-title>. <source>JAMA Netw Open</source>. (<year>2023</year>) <volume>6</volume>(<issue>12</issue>):<fpage>e2346721</fpage>. <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2023.46721</pub-id><pub-id pub-id-type="pmid">38060223</pub-id></mixed-citation></ref>
<ref id="B7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shojaee-Mend</surname> <given-names>H</given-names></name> <name><surname>Mohebbati</surname> <given-names>R</given-names></name> <name><surname>Amiri</surname> <given-names>M</given-names></name> <name><surname>Atarodi</surname> <given-names>A</given-names></name></person-group>. <article-title>Evaluating the strengths and weaknesses of large language models in answering neurophysiology questions</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>10785</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-60405-y</pub-id><pub-id pub-id-type="pmid">38734712</pub-id></mixed-citation></ref>
<ref id="B8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ros-Arlanz&#x00F3;n</surname> <given-names>P</given-names></name> <name><surname>Perez-Sempere</surname> <given-names>A</given-names></name></person-group>. <article-title>Evaluating AI competence in specialized medicine: comparative analysis of chatgpt and neurologists in a neurology specialist examination in spain</article-title>. <source>JMIR Med Educ</source>. (<year>2024</year>) <volume>10</volume>:<fpage>e56762</fpage>. <pub-id pub-id-type="doi">10.2196/56762</pub-id></mixed-citation></ref>
<ref id="B9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mavrych</surname> <given-names>V</given-names></name> <name><surname>Yaqinuddin</surname> <given-names>A</given-names></name> <name><surname>Bolgova</surname> <given-names>O</given-names></name></person-group>. <article-title>Claude, ChatGPT, Copilot, and gemini performance versus students in different topics of neuroscience</article-title>. <source>Adv Physiol Educ</source>. (<year>2025</year>) <volume>49</volume>(<issue>2</issue>):<fpage>430</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1152/advan.00093.2024</pub-id><pub-id pub-id-type="pmid">39824512</pub-id></mixed-citation></ref>
<ref id="B10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shu</surname> <given-names>L</given-names></name> <name><surname>Mandel</surname> <given-names>D</given-names></name> <name><surname>Tang</surname> <given-names>OY</given-names></name> <name><surname>Jiang</surname> <given-names>Z</given-names></name> <name><surname>Goldstein</surname> <given-names>E</given-names></name> <name><surname>Mahta</surname> <given-names>A</given-names></name></person-group>. <article-title>Large language model performance in neurology board questions (s33. 001)</article-title>. <source>Neurology</source>. (<year>2024</year>) <volume>102</volume>(<issue>7&#x005F;Supplement&#x005F;1</issue>):<fpage>2826</fpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000204763</pub-id></mixed-citation></ref>
<ref id="B11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Erdogan</surname> <given-names>M</given-names></name></person-group>. <article-title>Evaluation of responses of the large language model GPT to the neurology question of the week</article-title>. <source>Neurol Sci</source>. (<year>2024</year>) <volume>45</volume>(<issue>9</issue>):<fpage>4605</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1007/s10072-024-07580-y</pub-id><pub-id pub-id-type="pmid">38717580</pub-id></mixed-citation></ref>
<ref id="B12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fitzgerald</surname> <given-names>S</given-names></name></person-group>. <article-title>Chatgpt will take your neurology boards now: large-language models vs. humans on exams</article-title>. <source>Neurol Today</source>. (<year>2024</year>) <volume>24</volume>(<issue>3</issue>):<fpage>10</fpage>&#x2013;<lpage>1</lpage>. <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://neurologytoday.aan.com/doi/10.1097/01.NT.0001007276.74435.dc">https://neurologytoday.aan.com/doi/10.1097/01.NT.0001007276.74435.dc</ext-link></mixed-citation></ref>
<ref id="B13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chimowitz</surname> <given-names>MI</given-names></name> <name><surname>Logigian</surname> <given-names>EL</given-names></name> <name><surname>Caplan</surname> <given-names>LR</given-names></name></person-group>. <article-title>The accuracy of bedside neurological diagnoses</article-title>. <source>Ann Neurol</source>. (<year>1990</year>) <volume>28</volume>(<issue>1</issue>):<fpage>78</fpage>&#x2013;<lpage>85</lpage>. <pub-id pub-id-type="doi">10.1002/ana.410280114</pub-id><pub-id pub-id-type="pmid">2375637</pub-id></mixed-citation></ref>
<ref id="B14"><label>14.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Alpert</surname> <given-names>JN</given-names></name></person-group>, <source>The Neurologic Diagnosis: A Practical Bedside Approach</source>. <edition>Second Edition</edition>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2019</year>). <pub-id pub-id-type="doi">10.1007/978-3-319-95951-1</pub-id></mixed-citation></ref>
<ref id="B15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Berger</surname> <given-names>JR</given-names></name></person-group>. <article-title>Neurologists: the last bedside physician-scientists</article-title>. <source>JAMA Neurol</source>. (<year>2013</year>) <volume>70</volume>(<issue>8</issue>):<fpage>965</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1001/jamaneurol.2013.2977</pub-id><pub-id pub-id-type="pmid">23732945</pub-id></mixed-citation></ref>
<ref id="B16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>J-H</given-names></name> <name><surname>Choi</surname> <given-names>E</given-names></name> <name><surname>McDougal</surname> <given-names>R</given-names></name> <name><surname>Lytton</surname> <given-names>WW</given-names></name></person-group>. <article-title>GPT-4 performance for neurologic localization</article-title>. <source>Neurol Clin Pract</source>. (<year>2024</year>) <volume>14</volume>(<issue>3</issue>):<fpage>e200293</fpage>. <pub-id pub-id-type="doi">10.1212/CPJ.0000000000200293</pub-id><pub-id pub-id-type="pmid">38596779</pub-id></mixed-citation></ref>
<ref id="B17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Maiorana</surname> <given-names>NV</given-names></name> <name><surname>Marceglia</surname> <given-names>S</given-names></name> <name><surname>Treddenti</surname> <given-names>M</given-names></name> <name><surname>Tosi</surname> <given-names>M</given-names></name> <name><surname>Guidetti</surname> <given-names>M</given-names></name> <name><surname>Creta</surname> <given-names>MF</given-names></name><etal/></person-group> <article-title>Large language models in neurological practice: real-world study</article-title>. <source>J Med Internet Res</source>. (<year>2025</year>) <volume>27</volume>:<fpage>e73212</fpage>. <pub-id pub-id-type="doi">10.2196/73212</pub-id><pub-id pub-id-type="pmid">40982758</pub-id></mixed-citation></ref>
<ref id="B18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Habibi</surname> <given-names>G</given-names></name> <name><surname>Rajai Firouzabadi</surname> <given-names>S</given-names></name> <name><surname>Mohammadi</surname> <given-names>I</given-names></name> <name><surname>Gargari</surname> <given-names>OK</given-names></name></person-group>. <article-title>Evaluating the diagnostic performance of openbioA in neurology: a case-based assessment of a medical large language model</article-title>. <source>PLoS One</source>. (<year>2025</year>) <volume>20</volume>(<issue>9</issue>):<fpage>e0332196</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0332196</pub-id><pub-id pub-id-type="pmid">40997070</pub-id></mixed-citation></ref>
<ref id="B19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Barrit</surname> <given-names>S</given-names></name> <name><surname>Torcida</surname> <given-names>N</given-names></name> <name><surname>Mazeraud</surname> <given-names>A</given-names></name> <name><surname>Boulogne</surname> <given-names>S</given-names></name> <name><surname>Benoit</surname> <given-names>J</given-names></name> <name><surname>Carette</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Specialized large language model outperforms neurologists at complex diagnosis in blinded case-based evaluation</article-title>. <source>Brain Sci</source>. (<year>2025</year>) <volume>15</volume>(<issue>4</issue>):<fpage>347</fpage>. <pub-id pub-id-type="doi">10.3390/brainsci15040347</pub-id><pub-id pub-id-type="pmid">40309809</pub-id></mixed-citation></ref>
<ref id="B20"><label>20.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Dani</surname> <given-names>M</given-names></name> <name><surname>Prakash</surname> <given-names>MJ</given-names></name> <name><surname>Akata</surname> <given-names>Z</given-names></name> <name><surname>Liebe</surname> <given-names>S</given-names></name></person-group>. <article-title>SemioA: evaluating large language models for diagnostic reasoning from unstructured clinical narratives in epilepsy</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2407.03004</italic> (2024)</comment>.</mixed-citation></ref>
<ref id="B21"><label>21.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Haq</surname> <given-names>M</given-names></name> <name><surname>Ur Rehman</surname> <given-names>MM</given-names></name> <name><surname>Derhab</surname> <given-names>M</given-names></name> <name><surname>Saeed</surname> <given-names>R</given-names></name> <name><surname>Kalia</surname> <given-names>J</given-names></name></person-group>. <article-title>Bridging language gaps in neurology patient education through large language models: a comparative analysis of chatgpt, gemini, and claude</article-title>. <comment><italic>medRxiv</italic> [Preprint]. (2024)</comment>.</mixed-citation></ref>
<ref id="B22"><label>22.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>J-H</given-names></name> <name><surname>Choi</surname> <given-names>E</given-names></name> <name><surname>Angulo Castro</surname> <given-names>S</given-names></name> <name><surname>McDougal</surname> <given-names>R</given-names></name> <name><surname>Lytton</surname> <given-names>W</given-names></name></person-group>. <article-title>Neurological history both twinned and queried by generative artificial intelligence (p1-2.006)</article-title>. <source>Neurology</source>. (<year>2025</year>) <volume>104</volume>(<issue>7&#x005F;Supplement&#x005F;1</issue>):<fpage>179</fpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000208430</pub-id></mixed-citation></ref>
<ref id="B23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chadehumbe</surname> <given-names>T</given-names></name> <name><surname>Mintz</surname> <given-names>J</given-names></name></person-group>. <article-title>Enhancing physician efficiency with ambient artificial intelligence</article-title>. <source>Neurology</source>. (<year>2025</year>) <volume>104</volume>:<fpage>1703</fpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000208693</pub-id></mixed-citation></ref>
<ref id="B24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Johnsen</surname> <given-names>M</given-names></name> <name><surname>Meng</surname> <given-names>X</given-names></name> <name><surname>Zhang</surname> <given-names>S</given-names></name> <name><surname>Lee</surname> <given-names>D</given-names></name></person-group>. <article-title>Large language models in healthcare and medical applications: a comprehensive review</article-title>. <source>Front Digit Health</source>. (<year>2025</year>) <volume>3</volume>:<fpage>12189880</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering12060631</pub-id></mixed-citation></ref>
<ref id="B25"><label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Romano</surname> <given-names>M</given-names></name> <name><surname>Shih</surname> <given-names>L</given-names></name> <name><surname>Paschalidis</surname> <given-names>I</given-names></name> <name><surname>Au</surname> <given-names>R</given-names></name> <name><surname>Kolachalama</surname> <given-names>V</given-names></name></person-group>. <article-title>Large language models in neurology research and future practice</article-title>. <source>Neurology</source>. (<year>2023</year>) <volume>101</volume>:<fpage>1058</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000207967</pub-id><pub-id pub-id-type="pmid">37816646</pub-id></mixed-citation></ref>
<ref id="B26"><label>26.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>J</given-names></name> <name><surname>He</surname> <given-names>F</given-names></name> <name><surname>Yin</surname> <given-names>W</given-names></name> <name><surname>Ma</surname> <given-names>W</given-names></name> <name><surname>Wu</surname> <given-names>J</given-names></name></person-group>. <article-title>Stroke diagnosis and prediction tool using ChatGLM: development and validation study</article-title>. <source>J Med Internet Res</source>. (<year>2025</year>) <volume>27</volume>:<fpage>e67010</fpage>. <pub-id pub-id-type="doi">10.2196/67010</pub-id><pub-id pub-id-type="pmid">40009850</pub-id></mixed-citation></ref>
<ref id="B27"><label>27.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Harrison</surname> <given-names>JR</given-names></name> <name><surname>Tang</surname> <given-names>SL</given-names></name> <name><surname>Burston</surname> <given-names>B</given-names></name> <name><surname>Robertson</surname> <given-names>A</given-names></name> <name><surname>Liang</surname> <given-names>H</given-names></name> <name><surname>Taylor</surname> <given-names>JP</given-names></name></person-group>. <article-title>Large language models: a paradigm shift for dementia diagnosis and care</article-title>. <source>Br J Hosp Med</source>. (<year>2025</year>) <volume>86</volume>:<fpage>1</fpage>&#x2013;<lpage>19</lpage>. <pub-id pub-id-type="doi">10.12968/hmed.2024.0666</pub-id></mixed-citation></ref>
<ref id="B28"><label>28.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Qadri</surname> <given-names>YA</given-names></name> <name><surname>Ahmad</surname> <given-names>K</given-names></name> <name><surname>Kim</surname> <given-names>SW</given-names></name></person-group>. <article-title>Artificial general intelligence for the detection of neurodegenerative disorders</article-title>. <source>Sensors</source>. (<year>2024</year>) <volume>24</volume>(<issue>20</issue>):<fpage>6658</fpage>. <pub-id pub-id-type="doi">10.3390/s24206658</pub-id><pub-id pub-id-type="pmid">39460138</pub-id></mixed-citation></ref>
<ref id="B29"><label>29.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zamai</surname> <given-names>A</given-names></name> <name><surname>Fijalkow</surname> <given-names>N</given-names></name> <name><surname>Mansencal</surname> <given-names>B</given-names></name> <name><surname>Simon</surname> <given-names>L</given-names></name> <name><surname>Navet</surname> <given-names>E</given-names></name> <name><surname>Coupe</surname> <given-names>P</given-names></name></person-group>. <article-title>An explainable diagnostic framework for neurodegenerative dementias via reinforcement-optimized a reasoning</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2505.19954</italic> (2025)</comment>.</mixed-citation></ref>
<ref id="B30"><label>30.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Twala</surname> <given-names>B</given-names></name></person-group>. <article-title>AI-driven precision diagnosis and treatment in Parkinson&#x2019;s disease: a comprehensive review and experimental analysis</article-title>. <source>Front Aging Neurosci</source>. (<year>2025</year>) <volume>17</volume>:<fpage>1638340</fpage>. <pub-id pub-id-type="doi">10.3389/fnagi.2025.1638340</pub-id><pub-id pub-id-type="pmid">40791245</pub-id></mixed-citation></ref>
<ref id="B31"><label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Venkatesh</surname> <given-names>S</given-names></name> <name><surname>DelSignore</surname> <given-names>M</given-names></name> <name><surname>Wu</surname> <given-names>X</given-names></name> <name><surname>Morris</surname> <given-names>M</given-names></name> <name><surname>Kerr</surname> <given-names>W</given-names></name> <name><surname>Visweswaran</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Deconstructing complex diagnostic criteria and leveraging generative artificial intelligence to facilitate multiple sclerosis diagnosis</article-title>. <source>Mult Scler J</source>. (<year>2025</year>) <volume>31</volume>(<issue>2</issue>):<fpage>244</fpage>. <pub-id pub-id-type="doi">10.1177/13524585251333228</pub-id></mixed-citation></ref>
<ref id="B32"><label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Naji</surname> <given-names>Y</given-names></name> <name><surname>Mahdaoui</surname> <given-names>M</given-names></name> <name><surname>Klevor</surname> <given-names>R</given-names></name> <name><surname>Kissani</surname> <given-names>N</given-names></name> <name><surname>Raymond</surname> <given-names>K</given-names></name></person-group>. <article-title>Artificial intelligence and multiple sclerosis: up-to-date review</article-title>. <source>Cureus</source>. (<year>2023</year>) <volume>15</volume>(<issue>9</issue>):<fpage>e45412</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.45412</pub-id><pub-id pub-id-type="pmid">37854769</pub-id></mixed-citation></ref>
<ref id="B33"><label>33.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ford</surname> <given-names>J</given-names></name> <name><surname>Pevy</surname> <given-names>N</given-names></name> <name><surname>Grunewald</surname> <given-names>R</given-names></name> <name><surname>Howell</surname> <given-names>S</given-names></name> <name><surname>Reuber</surname> <given-names>M</given-names></name></person-group>. <article-title>Can artificial intelligence diagnose seizures based on patients&#x2019; descriptions? a study of GPT-4</article-title>. <source>Epilepsia</source>. (<year>2025</year>) <volume>66</volume>:<fpage>1959</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1111/epi.18322</pub-id><pub-id pub-id-type="pmid">40013839</pub-id></mixed-citation></ref>
<ref id="B34"><label>34.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>S</given-names></name> <name><surname>Luo</surname> <given-names>Y</given-names></name> <name><surname>Fotedar</surname> <given-names>N</given-names></name> <name><surname>Jiao</surname> <given-names>M</given-names></name> <name><surname>Rao</surname> <given-names>VR</given-names></name> <name><surname>Ju</surname> <given-names>X</given-names></name><etal/></person-group> <article-title>EpiSemoLLM: a fine-tuned large language model for epileptogenic zone localization based on seizure semiology with a performance comparable to epileptologists</article-title>. <comment><italic>MedRxiv</italic> [Preprint]. (2024)</comment>.</mixed-citation></ref>
<ref id="B35"><label>35.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Suenghataiphorn</surname> <given-names>T</given-names></name> <name><surname>Tribuddharat</surname> <given-names>N</given-names></name> <name><surname>Danpanichkul</surname> <given-names>P</given-names></name> <name><surname>Kulthamrongsri</surname> <given-names>N</given-names></name></person-group>. <article-title>Bias in large language models across clinical applications: a systematic review</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2504.02917</italic> (2025)</comment>.</mixed-citation></ref>
<ref id="B36"><label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mahajan</surname> <given-names>A</given-names></name> <name><surname>Obermeyer</surname> <given-names>Z</given-names></name> <name><surname>Daneshjou</surname> <given-names>R</given-names></name> <name><surname>Lester</surname> <given-names>J</given-names></name> <name><surname>Powell</surname> <given-names>D</given-names></name></person-group>. <article-title>Cognitive bias in clinical large language models</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>1</issue>):<fpage>428</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-01790-0</pub-id><pub-id pub-id-type="pmid">40640549</pub-id></mixed-citation></ref>
<ref id="B37"><label>37.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abujaber</surname> <given-names>AA</given-names></name> <name><surname>Nashwan</surname> <given-names>AJ</given-names></name></person-group>. <article-title>Ethical framework for artificial intelligence in healthcare research: a path to integrity</article-title>. <source>World J Methodol</source>. (<year>2024</year>) <volume>14</volume>(<issue>3</issue>):<fpage>94071</fpage>. <pub-id pub-id-type="doi">10.5662/wjm.v14.i3.94071</pub-id><pub-id pub-id-type="pmid">39310239</pub-id></mixed-citation></ref>
<ref id="B38"><label>38.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Das</surname> <given-names>BC</given-names></name> <name><surname>Amini</surname> <given-names>MH</given-names></name> <name><surname>Wu</surname> <given-names>Y</given-names></name></person-group>. <article-title>Security and privacy challenges of large language models: a survey</article-title>. <source>ACM Comput Surv</source>. (<year>2025</year>) <volume>57</volume>(<issue>6</issue>):<fpage>1</fpage>&#x2013;<lpage>39</lpage>. <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://doi-org.proxy.cc.uic.edu/10.1145/3712001">https://doi-org.proxy.cc.uic.edu/10.1145/3712001</ext-link></mixed-citation></ref>
<ref id="B39"><label>39.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chiang</surname> <given-names>C-C</given-names></name> <name><surname>Fries</surname> <given-names>JA</given-names></name></person-group>. <article-title>Exploring the potential of large language models in neurology, using neurologic localization as an example</article-title>. <source>Neurol Clin Pract</source>. (<year>2024</year>) <volume>14</volume>(<issue>3</issue>):<fpage>e200311</fpage>. <pub-id pub-id-type="doi">10.1212/CPJ.0000000000200311</pub-id><pub-id pub-id-type="pmid">38601858</pub-id></mixed-citation></ref>
<ref id="B40"><label>40.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jung</surname> <given-names>K-H</given-names></name></person-group>. <article-title>Large language models in medicine: clinical applications, technical challenges, and ethical considerations</article-title>. <source>Healthc Inform Res</source>. (<year>2025</year>) <volume>31</volume>(<issue>2</issue>):<fpage>114</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.4258/hir.2025.31.2.114</pub-id><pub-id pub-id-type="pmid">40384063</pub-id></mixed-citation></ref>
<ref id="B41"><label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Busch</surname> <given-names>F</given-names></name> <name><surname>Hoffmann</surname> <given-names>L</given-names></name> <name><surname>Rueger</surname> <given-names>C</given-names></name> <name><surname>van Dijk</surname> <given-names>EHC</given-names></name> <name><surname>Kader</surname> <given-names>R</given-names></name> <name><surname>Ortiz-Prado</surname> <given-names>E</given-names></name><etal/></person-group> <article-title>Current applications and challenges in large language models for patient care: a systematic review</article-title>. <source>Commun Med</source>. (<year>2025</year>) <volume>5</volume>(<issue>1</issue>):<fpage>26</fpage>. <pub-id pub-id-type="doi">10.1038/s43856-024-00717-2</pub-id><pub-id pub-id-type="pmid">39838160</pub-id></mixed-citation></ref>
<ref id="B42"><label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McCoy</surname> <given-names>LG</given-names></name> <name><surname>Swamy</surname> <given-names>R</given-names></name> <name><surname>Sagar</surname> <given-names>N</given-names></name> <name><surname>Wang</surname> <given-names>M</given-names></name> <name><surname>Bacchi</surname> <given-names>S</given-names></name> <name><surname>Fong</surname> <given-names>JMN</given-names></name><etal/></person-group> <article-title>Assessment of large language models in clinical reasoning: a novel benchmarking study</article-title>. <source>NEJM AI</source>. (<year>2025</year>) <volume>2</volume>(<issue>10</issue>):<fpage>AIdbp2500120</fpage>. <pub-id pub-id-type="doi">10.1056/AIdbp2500120</pub-id></mixed-citation></ref>
<ref id="B43"><label>43.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mehandru</surname> <given-names>N</given-names></name> <name><surname>Miao</surname> <given-names>BY</given-names></name> <name><surname>Almaraz</surname> <given-names>ER</given-names></name> <name><surname>Sushil</surname> <given-names>M</given-names></name> <name><surname>Butte</surname> <given-names>AJ</given-names></name> <name><surname>Alaa</surname> <given-names>A</given-names></name></person-group>. <article-title>Evaluating large language models as agents in the clinic</article-title>. <source>NPJ Digit Med</source>. (<year>2024</year>) <volume>7</volume>(<issue>1</issue>):<fpage>84</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-024-01083-y</pub-id><pub-id pub-id-type="pmid">38570554</pub-id></mixed-citation></ref>
<ref id="B44"><label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shinar</surname> <given-names>D</given-names></name> <name><surname>Gross</surname> <given-names>CR</given-names></name> <name><surname>Mohr</surname> <given-names>JP</given-names></name> <name><surname>Caplan</surname> <given-names>LR</given-names></name> <name><surname>Price</surname> <given-names>TR</given-names></name> <name><surname>Wolf</surname> <given-names>PA</given-names></name><etal/></person-group> <article-title>Interobserver variability in the assessment of neurologic history and examination in the stroke data bank</article-title>. <source>Arch Neurol</source>. (<year>1985</year>) <volume>42</volume>(<issue>6</issue>):<fpage>557</fpage>&#x2013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.1001/archneur.1985.04060060059010</pub-id><pub-id pub-id-type="pmid">4004598</pub-id></mixed-citation></ref>
<ref id="B45"><label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Busis</surname> <given-names>NA</given-names></name> <name><surname>Marolia</surname> <given-names>D</given-names></name> <name><surname>Montgomery</surname> <given-names>R</given-names></name> <name><surname>Balcer</surname> <given-names>LJ</given-names></name> <name><surname>Galetta</surname> <given-names>SL</given-names></name> <name><surname>Grossman</surname> <given-names>SN</given-names></name></person-group>. <article-title>Navigating the us regulatory landscape for neurologic digital health technologies</article-title>. <source>NPJ Digit Med</source>. (<year>2024</year>) <volume>7</volume>(<issue>1</issue>):<fpage>94</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-024-01098-5</pub-id><pub-id pub-id-type="pmid">38609447</pub-id></mixed-citation></ref>
<ref id="B46"><label>46.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Weissman</surname> <given-names>GE</given-names></name> <name><surname>Mankowitz</surname> <given-names>T</given-names></name> <name><surname>Kanter</surname> <given-names>GP</given-names></name></person-group>. <article-title>Unregulated large language models produce medical device&#x2013;like output</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>3</issue>):<fpage>215</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-01544-y</pub-id><pub-id pub-id-type="pmid">40254642</pub-id></mixed-citation></ref>
<ref id="B47"><label>47.</label><mixed-citation publication-type="other"><collab>U.S. Food and Drug Administration</collab>. <article-title>Artificial intelligence and machine learning in software as a medical device (SaMD)</article-title>. <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-software-medical-device">https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-software-medical-device</ext-link> (Accessed March 2025). FDA guidance outlining oversight principles for AI- and LLM-enabled medical software under the Software as a Medical Device framework</comment>.</mixed-citation></ref>
<ref id="B48"><label>48.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Peng</surname> <given-names>C</given-names></name> <name><surname>Zhang</surname> <given-names>K</given-names></name> <name><surname>Lyu</surname> <given-names>M</given-names></name> <name><surname>Liu</surname> <given-names>H</given-names></name> <name><surname>Sun</surname> <given-names>L</given-names></name> <name><surname>Wu</surname> <given-names>Y</given-names></name></person-group>. <article-title>Scaling up biomedical vision-language models: fine-tuning, instruction tuning, and multi-modal learning</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2505.17436</italic> (2025)</comment>.</mixed-citation></ref>
<ref id="B49"><label>49.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Ye</surname> <given-names>H</given-names></name> <name><surname>Li</surname> <given-names>S</given-names></name></person-group>. <article-title>LLMs help alleviate the cross-subject variability in brain signal and language alignment</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2501.02621</italic> (2025)</comment>.</mixed-citation></ref>
<ref id="B50"><label>50.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chiang</surname> <given-names>S</given-names></name> <name><surname>Picard</surname> <given-names>RW</given-names></name> <name><surname>Chiong</surname> <given-names>W</given-names></name> <name><surname>Moss</surname> <given-names>R</given-names></name> <name><surname>Worrell</surname> <given-names>GA</given-names></name> <name><surname>Rao</surname> <given-names>VR</given-names></name><etal/></person-group> <article-title>Guidelines for conducting ethical artificial intelligence research in neurology: a systematic approach for clinicians and researchers</article-title>. <source>Neurology</source>. (<year>2021</year>) <volume>97</volume>(<issue>13</issue>):<fpage>632</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000012570</pub-id><pub-id pub-id-type="pmid">34315785</pub-id></mixed-citation></ref>
<ref id="B51"><label>51.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>E</given-names></name> <name><surname>Chu</surname> <given-names>X</given-names></name> <name><surname>Zhang</surname> <given-names>W</given-names></name> <name><surname>Meng</surname> <given-names>X</given-names></name> <name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Ji</surname> <given-names>X</given-names></name><etal/></person-group> <article-title>Large language models in medicine: applications, challenges, and future directions</article-title>. <source>Int J Med Sci</source>. (<year>2025</year>) <volume>22</volume>(<issue>11</issue>):<fpage>2792</fpage>. <pub-id pub-id-type="doi">10.7150/ijms.111780</pub-id><pub-id pub-id-type="pmid">40520893</pub-id></mixed-citation></ref>
<ref id="B52"><label>52.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rizzo</surname> <given-names>M</given-names></name> <name><surname>Dawson</surname> <given-names>JD</given-names></name></person-group>. <article-title>AI in neurology: everything, everywhere, all at once part 1: principles and practice</article-title>. <source>Ann Neurol</source>. (<year>2025</year>) <volume>98</volume>:<fpage>211</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1002/ana.27225</pub-id><pub-id pub-id-type="pmid">40536356</pub-id></mixed-citation></ref>
<ref id="B53"><label>53.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Maiorana</surname> <given-names>NV</given-names></name> <name><surname>Marceglia</surname> <given-names>S</given-names></name> <name><surname>Treddenti</surname> <given-names>M</given-names></name> <name><surname>Tosi</surname> <given-names>M</given-names></name> <name><surname>Guidetti</surname> <given-names>M</given-names></name> <name><surname>Creta</surname> <given-names>MF</given-names></name><etal/></person-group> <article-title>Large language models in neurological practice: real-world study</article-title>. <source>J Med Internet Res</source>. (<year>2025</year>) <volume>27</volume>:<fpage>e73212</fpage>. <pub-id pub-id-type="doi">10.2196/73212</pub-id><pub-id pub-id-type="pmid">40982758</pub-id></mixed-citation></ref>
<ref id="B54"><label>54.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Gupta</surname> <given-names>A</given-names></name> <name><surname>Patel</surname> <given-names>S</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name></person-group>. <article-title>Retrieval-augmented generation for large language models in healthcare: a systematic review</article-title>. <source>PLoS Digit Health</source>. (<year>2025</year>) <volume>4</volume>:<fpage>e0000877</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000877</pub-id><pub-id pub-id-type="pmid">40498738</pub-id></mixed-citation></ref>
<ref id="B55"><label>55.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Ovadia</surname> <given-names>O</given-names></name> <name><surname>Brief</surname> <given-names>M</given-names></name> <name><surname>Mishaeli</surname> <given-names>M</given-names></name> <name><surname>Elisha</surname> <given-names>O</given-names></name></person-group>. <article-title>Fine-tuning or retrieval? Comparing knowledge injection in LLMs</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2312.05934</italic> (2023)</comment>.</mixed-citation></ref>
<ref id="B56"><label>56.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Agrawal</surname> <given-names>M</given-names></name> <name><surname>Chen</surname> <given-names>IY</given-names></name> <name><surname>Gulamali</surname> <given-names>F</given-names></name> <name><surname>Joshi</surname> <given-names>S</given-names></name></person-group>. <article-title>The evaluation illusion of large language models in medicine</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>1</issue>):<fpage>600</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-01963-x</pub-id><pub-id pub-id-type="pmid">41057566</pub-id></mixed-citation></ref>
<ref id="B57"><label>57.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Busis</surname> <given-names>NA</given-names></name></person-group>. <article-title>Study suggests 60&#x0025; of U.S. neurologists experiencing burnout</article-title>. <comment><italic>American Academy of Neurology Press Release</italic> (January 2017). AAN workforce survey reporting that burnout among neurologists is primarily associated with administrative workload, clerical duties, and electronic documentation requirements</comment>.</mixed-citation></ref>
<ref id="B58"><label>58.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Busis</surname> <given-names>NA</given-names></name> <name><surname>Shanafelt</surname> <given-names>TD</given-names></name> <name><surname>Keran</surname> <given-names>CM</given-names></name> <name><surname>Cascino</surname> <given-names>GD</given-names></name> <name><surname>Vidic</surname> <given-names>TR</given-names></name> <name><surname>Miyasaki</surname> <given-names>JM</given-names></name></person-group>. <article-title>Burnout, career satisfaction, and well-being among U.S. neurologists in 2016</article-title>. <source>Neurology</source>. (<year>2017</year>) <volume>88</volume>(<issue>8</issue>):<fpage>797</fpage>&#x2013;<lpage>808</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000003640</pub-id><pub-id pub-id-type="pmid">28122905</pub-id></mixed-citation></ref>
<ref id="B59"><label>59.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miyasaki</surname> <given-names>JM</given-names></name> <name><surname>Rhea</surname> <given-names>D</given-names></name> <name><surname>Marras</surname> <given-names>C</given-names></name> <name><surname>Schoenberg</surname> <given-names>M</given-names></name> <name><surname>Tanner</surname> <given-names>CM</given-names></name> <name><surname>Weiner</surname> <given-names>WJ</given-names></name><etal/></person-group> <article-title>Burnout in neurologists: a survey of us neurologists in 2016</article-title>. <source>Neurology</source>. (<year>2017</year>) <volume>89</volume>(<issue>16</issue>):<fpage>1730</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000004526</pub-id><pub-id pub-id-type="pmid">28931640</pub-id></mixed-citation></ref>
<ref id="B60"><label>60.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rodr&#x00ED;guez-Fern&#x00E1;ndez</surname> <given-names>JM</given-names></name> <name><surname>Loeb</surname> <given-names>JA</given-names></name> <name><surname>Hier</surname> <given-names>DB</given-names></name></person-group>. <article-title>It&#x2019;s time to change our documentation philosophy: writing better neurology notes without the burnout</article-title>. <source>Front Digit Health</source>. (<year>2022</year>) <volume>4</volume>:<fpage>1063141</fpage>. <pub-id pub-id-type="doi">10.3389/fdgth.2022.1063141</pub-id></mixed-citation></ref>
<ref id="B61"><label>61.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sad&#x00E9;e</surname> <given-names>C</given-names></name> <name><surname>Testa</surname> <given-names>S</given-names></name> <name><surname>Barba</surname> <given-names>T</given-names></name> <name><surname>Hartmann</surname> <given-names>K</given-names></name> <name><surname>Schuessler</surname> <given-names>M</given-names></name> <name><surname>Thieme</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Medical digital twins: enabling precision medicine and proactive care</article-title>. <source>Lancet Digit Health</source>. (<year>2025</year>) <volume>7</volume>(<issue>3</issue>):<fpage>e215</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.1016/j.landig.2025.02.004</pub-id></mixed-citation></ref>
<ref id="B62"><label>62.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vall&#x00E9;e</surname> <given-names>A</given-names></name></person-group>. <article-title>Digital twins for personalized medicine require epidemiological data and mathematical modeling</article-title>. <source>J Med Internet Res</source>. (<year>2025</year>) <volume>27</volume>:<fpage>e72411</fpage>. <pub-id pub-id-type="doi">10.2196/72411</pub-id></mixed-citation></ref>
<ref id="B63"><label>63.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Iqbal</surname> <given-names>JD</given-names></name> <name><surname>Krauthammer</surname> <given-names>M</given-names></name> <name><surname>Witt</surname> <given-names>CM</given-names></name> <name><surname>Biller-Andorno</surname> <given-names>N</given-names></name> <name><surname>Christen</surname> <given-names>M</given-names></name></person-group>. <article-title>A consensus statement on the use of digital twins in medicine</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>1</issue>):<fpage>484</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-01897-4</pub-id><pub-id pub-id-type="pmid">40721854</pub-id></mixed-citation></ref>
<ref id="B64"><label>64.</label><mixed-citation publication-type="other"><collab>Unlearn.AI Research Team</collab>. <article-title>Digital twin generators for disease modeling</article-title>. <comment><italic>arXiv</italic> [e-prints]. (2024)</comment>.</mixed-citation></ref>
<ref id="B65"><label>65.</label><mixed-citation publication-type="other"><collab>Stanford Medicine</collab>. <article-title>Medical digital twins: a new frontier in personalized healthcare</article-title>. (<year>2024</year>). <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://med.stanford.edu/medicine/news/current-news/standard-news/medical-digital-twins.html">https://med.stanford.edu/medicine/news/current-news/standard-news/medical-digital-twins.html</ext-link> (Accessed October 25, 2025)</comment>.</mixed-citation></ref>
<ref id="B66"><label>66.</label><mixed-citation publication-type="other"><collab>Palos Publishing</collab>. <article-title>Combining large language models with digital twin simulations. (2024)</article-title>. <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://palospublishing.com/combining-llms-with-digital-twin-simulations/">https://palospublishing.com/combining-llms-with-digital-twin-simulations/</ext-link> (Accessed October 25, 2025)</comment>.</mixed-citation></ref>
<ref id="B67"><label>67.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Fu</surname> <given-names>T</given-names></name> <name><surname>Xu</surname> <given-names>Y</given-names></name> <name><surname>Ma</surname> <given-names>Z</given-names></name> <name><surname>Xu</surname> <given-names>H</given-names></name> <name><surname>Du</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>TWIN-GPT: digital twins for clinical trials via large language model</article-title>. <source>ACM Trans Multimed Comput Commun Appl</source>. (<year>2024</year>). p. <fpage>1</fpage>&#x2013;<lpage>19</lpage>. <comment>Available online at</comment>: <ext-link ext-link-type="uri" xlink:href="https://doi-org.proxy.cc.uic.edu/10.1145/3674838">https://doi-org.proxy.cc.uic.edu/10.1145/3674838</ext-link></mixed-citation></ref>
<ref id="B68"><label>68.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Giglou</surname> <given-names>HB</given-names></name> <name><surname>D&#x2019;Souza</surname> <given-names>J</given-names></name> <name><surname>Mihindukulasooriya</surname> <given-names>N</given-names></name> <name><surname>Auer</surname> <given-names>S</given-names></name></person-group>. <article-title>As4ol 2025 overview: the second large language models for ontology learning challenge</article-title>. <comment>In: <italic>Open Conference Proceedings</italic>. (2025). Vol. 6</comment>.</mixed-citation></ref>
<ref id="B69"><label>69.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Engelke</surname> <given-names>M</given-names></name> <name><surname>Baldini</surname> <given-names>G</given-names></name> <name><surname>Kleesiek</surname> <given-names>J</given-names></name> <name><surname>Nensa</surname> <given-names>F</given-names></name> <name><surname>Dada</surname> <given-names>A</given-names></name></person-group>. <article-title>FHIR-former: enhancing clinical predictions through fast healthcare interoperability resources and large language models</article-title>. <source>J Am Med Inform Assoc</source>. (<year>2025</year>) <volume>32</volume>:<fpage>1793</fpage>&#x2013;<lpage>801</lpage>. <pub-id pub-id-type="doi">10.1093/jamia/ocaf165</pub-id><pub-id pub-id-type="pmid">41082356</pub-id></mixed-citation></ref>
<ref id="B70"><label>70.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hier</surname> <given-names>DB</given-names></name> <name><surname>Carrithers</surname> <given-names>MA</given-names></name> <name><surname>Platt</surname> <given-names>SK</given-names></name> <name><surname>Nguyen</surname> <given-names>A</given-names></name> <name><surname>Giannopoulos</surname> <given-names>I</given-names></name> <name><surname>Obafemi-Ajayi</surname> <given-names>T</given-names></name></person-group>. <article-title>Preprocessing of physician notes by LLMs improves clinical concept extraction without information loss</article-title>. <source>Information</source>. (<year>2025</year>) <volume>16</volume>(<issue>6</issue>):<fpage>446</fpage>. <pub-id pub-id-type="doi">10.3390/info16060446</pub-id></mixed-citation></ref>
<ref id="B71"><label>71.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Hier</surname> <given-names>DB</given-names></name> <name><surname>Carrithers</surname> <given-names>MD</given-names></name> <name><surname>Do</surname> <given-names>TS</given-names></name> <name><surname>Obafemi-Ajayi</surname> <given-names>T</given-names></name></person-group>. <article-title>Remote: a framework to create fast healthcare interoperability resources (fHIR) from unstructured clinical data</article-title>. <comment>In: <italic>2025 47th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</italic> (2025). p. 1&#x2013;6</comment>.</mixed-citation></ref>
<ref id="B72"><label>72.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Murdoch</surname> <given-names>TB</given-names></name> <name><surname>Detsky</surname> <given-names>AS</given-names></name></person-group>. <article-title>The inevitable application of big data to health care</article-title>. <source>JAMA</source>. (<year>2013</year>) <volume>309</volume>(<issue>13</issue>):<fpage>1351</fpage>&#x2013;<lpage>2</lpage>. <pub-id pub-id-type="doi">10.1001/jama.2013.393</pub-id><pub-id pub-id-type="pmid">23549579</pub-id></mixed-citation></ref>
<ref id="B73"><label>73.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Rastegar-Mojarad</surname> <given-names>M</given-names></name> <name><surname>Moon</surname> <given-names>S</given-names></name> <name><surname>Shen</surname> <given-names>F</given-names></name> <name><surname>Afzal</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>Clinical information extraction applications: a literature review</article-title>. <source>J Biomed Inform</source>. (<year>2018</year>) <volume>77</volume>:<fpage>34</fpage>&#x2013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2017.11.011</pub-id><pub-id pub-id-type="pmid">29162496</pub-id></mixed-citation></ref>
<ref id="B74"><label>74.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nerella</surname> <given-names>S</given-names></name> <name><surname>Bandyopadhyay</surname> <given-names>S</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Contreras</surname> <given-names>M</given-names></name> <name><surname>Siegel</surname> <given-names>S</given-names></name> <name><surname>Bumin</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Transformers and large language models in healthcare: a review</article-title>. <source>Artif Intell Med</source>. (<year>2024</year>) <volume>154</volume>:<fpage>102900</fpage>. <pub-id pub-id-type="doi">10.1016/j.artmed.2024.102900</pub-id><pub-id pub-id-type="pmid">38878555</pub-id></mixed-citation></ref>
<ref id="B75"><label>75.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Westover</surname> <given-names>MB</given-names></name> <name><surname>Westover</surname> <given-names>AM</given-names></name></person-group>. <article-title>General AI may revolutionize neurology&#x2014;or it might be bad</article-title>. <source>JAMA Neurol</source>. (<year>2025</year>) <volume>82</volume>:<fpage>977</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1001/jamaneurol.2025.0905</pub-id><pub-id pub-id-type="pmid">40323622</pub-id></mixed-citation></ref>
<ref id="B76"><label>76.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z</given-names></name> <name><surname>Zhao</surname> <given-names>S</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Huang</surname> <given-names>H</given-names></name> <name><surname>Xie</surname> <given-names>S</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Re-task: revisiting LLM tasks from capability, skill, and knowledge perspectives</article-title>. <comment><italic>arXiv</italic> [Preprint]. <italic>arXiv:2408.06904</italic> (2024)</comment>.</mixed-citation></ref></ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/896495/overview">Kezhi Li</ext-link>, University College London, United Kingdom</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/31886/overview">Romain Carron</ext-link>, H&#x00F4;pital de la Timone, France</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3276722/overview">Zhijun Guo</ext-link>, University College London, United Kingdom</p></fn>
</fn-group>
</back>
</article>