<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Dement.</journal-id>
<journal-title>Frontiers in Dementia</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Dement.</abbrev-journal-title>
<issn pub-type="epub">2813-3919</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frdem.2024.1385303</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Dementia</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Introduction to Large Language Models (LLMs) for dementia care and research</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Treder</surname> <given-names>Matthias S.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/133031/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Lee</surname> <given-names>Sojin</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2715186/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Tsvetanov</surname> <given-names>Kamen A.</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/179587/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>School of Computer Science &#x00026; Informatics, Cardiff University</institution>, <addr-line>Cardiff</addr-line>, <country>United Kingdom</country></aff>
<aff id="aff2"><sup>2</sup><institution>Olive AI Limited</institution>, <addr-line>London</addr-line>, <country>United Kingdom</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Clinical Neurosciences, University of Cambridge</institution>, <addr-line>Cambridge</addr-line>, <country>United Kingdom</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Psychology, University of Cambridge</institution>, <addr-line>Cambridge</addr-line>, <country>United Kingdom</country></aff>
<author-notes>
<fn fn-type="edited-by"><p>Edited by: Shinya Tasaki, Rush University Medical Center, United States</p></fn>
<fn fn-type="edited-by"><p>Reviewed by: Haiwen Gui, Stanford University, United States</p>
<p>Inez Y. Oh, Washington University in St. Louis, United States</p></fn>
<corresp id="c001">&#x0002A;Correspondence: Matthias S. Treder <email>matthias.treder&#x00040;gmail.com</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>14</day>
<month>05</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>3</volume>
<elocation-id>1385303</elocation-id>
<history>
<date date-type="received">
<day>12</day>
<month>02</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>04</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2024 Treder, Lee and Tsvetanov.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Treder, Lee and Tsvetanov</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Dementia is a progressive neurodegenerative disorder that affects cognitive abilities including memory, reasoning, and communication skills, leading to gradual decline in daily activities and social engagement. In light of the recent advent of Large Language Models (LLMs) such as ChatGPT, this paper aims to thoroughly analyse their potential applications and usefulness in dementia care and research.</p>
</sec>
<sec>
<title>Method</title>
<p>To this end, we offer an introduction into LLMs, outlining the key features, capabilities, limitations, potential risks, and practical considerations for deployment as easy-to-use software (e.g., smartphone apps). We then explore various domains related to dementia, identifying opportunities for LLMs to enhance understanding, diagnostics, and treatment, with a broader emphasis on improving patient care. For each domain, the specific contributions of LLMs are examined, such as their ability to engage users in meaningful conversations, deliver personalized support, and offer cognitive enrichment. Potential benefits encompass improved social interaction, enhanced cognitive functioning, increased emotional well-being, and reduced caregiver burden. The deployment of LLMs in caregiving frameworks also raises a number of concerns and considerations. These include privacy and safety concerns, the need for empirical validation, user-centered design, adaptation to the user&#x00027;s unique needs, and the integration of multimodal inputs to create more immersive and personalized experiences. Additionally, ethical guidelines and privacy protocols must be established to ensure responsible and ethical deployment of LLMs.</p>
</sec>
<sec>
<title>Results</title>
<p>We report the results on a questionnaire filled in by people with dementia (PwD) and their supporters wherein we surveyed the usefulness of different application scenarios of LLMs as well as the features that LLM-powered apps should have. Both PwD and supporters were largely positive regarding the prospect of LLMs in care, although concerns were raised regarding bias, data privacy and transparency.</p>
</sec>
<sec>
<title>Discussion</title>
<p>Overall, this review corroborates the promising utilization of LLMs to positively impact dementia care by boosting cognitive abilities, enriching social interaction, and supporting caregivers. The findings underscore the importance of further research and development in this field to fully harness the benefits of LLMs and maximize their potential for improving the lives of individuals living with dementia.</p>
</sec></abstract>
<kwd-group>
<kwd>dementia</kwd>
<kwd>Large Language Model (LLM)</kwd>
<kwd>Artificial Intelligence</kwd>
<kwd>Alzheimer&#x00027;s disease</kwd>
<kwd>care</kwd>
<kwd>natural language processing</kwd>
</kwd-group>
<counts>
<fig-count count="9"/>
<table-count count="4"/>
<equation-count count="0"/>
<ref-count count="226"/>
<page-count count="30"/>
<word-count count="24905"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Dementia Care</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<title>Introduction</title>
<p>As the global population ages, dementia emerges as one of the most pressing and multifaceted healthcare challenges (Parra et al., <xref ref-type="bibr" rid="B169">2019</xref>). More than 55 million individuals worldwide are currently living with dementia, with over 60% of these cases occurring in low- and middle-income countries. Furthermore, approximately 10 million new cases of dementia are diagnosed annually (WHO, <xref ref-type="bibr" rid="B214">2023</xref>). Characterized by progressive cognitive decline that impedes daily functioning, dementia not only impacts the affected individuals, but also their caregivers, families, and the healthcare system at large. Furthermore, dementia is frequently diagnosed late or misdiagnosed (Fischer et al., <xref ref-type="bibr" rid="B57">2017</xref>), while the limited availability of caregiver support post-diagnosis compounds the challenges faced by all involved. It becomes imperative for dementia care and research to develop innovative solutions for improved diagnosis, effective treatment and caregiving, ultimately reducing the global burden of this condition.</p>
<p>Amidst this backdrop, the rise of advanced computational tools and Artificial Intelligence (AI) technologies offers a beacon of hope. A branch of AI known as Large Language Models (LLMs), with their capacity to understand, generate, and interact using natural language, are at the forefront of these technological innovations (Bubeck et al., <xref ref-type="bibr" rid="B24">2023</xref>; Huang and Chang, <xref ref-type="bibr" rid="B86">2023</xref>; Khurana et al., <xref ref-type="bibr" rid="B105">2023</xref>; Min et al., <xref ref-type="bibr" rid="B139">2023</xref>). In the realm of dementia care and research, LLMs present unique opportunities to revolutionize diagnostic strategies, therapeutic interventions, and patient-caregiver communication. Yet, for all their promise, LLMs also bring forth a range of ethical, practical, and scientific challenges (Blodgett et al., <xref ref-type="bibr" rid="B19">2020</xref>; Gabriel, <xref ref-type="bibr" rid="B61">2020</xref>; Liao, <xref ref-type="bibr" rid="B118">2020</xref>; Dobbe et al., <xref ref-type="bibr" rid="B49">2021</xref>; Barocas et al., <xref ref-type="bibr" rid="B16">2023</xref>; Floridi and Floridi, <xref ref-type="bibr" rid="B58">2023</xref>; Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>; Kasneci et al., <xref ref-type="bibr" rid="B103">2023</xref>; Li and Zhang, <xref ref-type="bibr" rid="B117">2023</xref>; Wang et al., <xref ref-type="bibr" rid="B211">2023</xref>; Bzdok et al., <xref ref-type="bibr" rid="B25">2024</xref>). This paper aims to elucidate the prospects and potential pitfalls of employing LLMs in the domain of dementia care and research, paving the way for informed and judicious integration of these powerful tools in real-world settings.</p>
<p>Our key contributions are as follows:</p>
<list list-type="order">
<list-item><p>To our knowledge, this is the first publication specifically reviewing LLMs in the context of dementia management and care. Previous reviews surveyed AI in dementia more broadly (de la Fuente Garcia et al., <xref ref-type="bibr" rid="B39">2020</xref>; Lee et al., <xref ref-type="bibr" rid="B113">2021</xref>; Richardson et al., <xref ref-type="bibr" rid="B185">2022</xref>; Borchert et al., <xref ref-type="bibr" rid="B20">2023</xref>; Tsoi et al., <xref ref-type="bibr" rid="B201">2023</xref>) or focused on AI for prediction and early diagnosis (Stamate et al., <xref ref-type="bibr" rid="B194">2020</xref>; Li et al., <xref ref-type="bibr" rid="B115">2022</xref>; Merkin et al., <xref ref-type="bibr" rid="B134">2022</xref>; Borchert et al., <xref ref-type="bibr" rid="B20">2023</xref>).</p></list-item>
<list-item><p>We propose and thoroughly discuss several application scenarios where LLMs can be useful to people with dementia, including navigation aid, reading/writing assistance, and conversational services.</p></list-item>
<list-item><p>We present the results of a survey of people with dementia (PwD) and supporters wherein we investigated their experience with AI and LLMs, their evaluation on the usefulness of the presented application scenarios, and their priorities that AI software developers should consider (e.g., privacy, ease of use).</p></list-item>
</list>
<p>In the next section, we briefly review the dementia literature, before introducing the application of LLMs in this field.</p>
<sec>
<title>Dementia overview</title>
<p>A detailed introduction into dementia, its epidemiology, various subtypes and diagnosis, risk factors, and treatment is included in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material A</xref>. For brevity, we only provide a summary here. Dementia is a major public health priority (Prince et al., <xref ref-type="bibr" rid="B176">2015</xref>), with the number of affected individuals expected to triple by 2050 (Nichols et al., <xref ref-type="bibr" rid="B157">2022</xref>), creating significant economic and social challenges (Nandi et al., <xref ref-type="bibr" rid="B153">2022</xref>). It encompasses various brain disorders characterized by a decline in cognitive and motor functions due to brain cell loss. Common types include Alzheimer&#x00027;s disease, vascular dementia, dementia with Lewy bodies, and frontotemporal dementia, each associated with specific brain regions and symptoms. Mixed dementia involves concurrent brain changes from multiple dementia types (Schneider et al., <xref ref-type="bibr" rid="B189">2007</xref>; Kapasi et al., <xref ref-type="bibr" rid="B101">2017</xref>).</p>
<p>Alzheimer&#x00027;s disease, the most prevalent cause of dementia, involves memory lapses, word-finding difficulties, and mood swings, with damage often starting in the hippocampus (Sheehan, <xref ref-type="bibr" rid="B191">2012</xref>; Jack et al., <xref ref-type="bibr" rid="B91">2018</xref>; Lane et al., <xref ref-type="bibr" rid="B111">2018</xref>; Armstrong et al., <xref ref-type="bibr" rid="B11">2024</xref>). Most Alzheimer&#x00027;s cases are sporadic with late onset, but a rare early-onset form typically appears before the age of 65 (<xref ref-type="bibr" rid="B1">2023 Alzheimer&#x00027;s Disease Facts and Figures</xref>, <xref ref-type="bibr" rid="B1">2023</xref>). Vascular dementia arises from damage to the brain&#x00027;s blood vessels and is associated with cognitive impairments such as impaired judgment, planning difficulties, and mood fluctuations (Iadecola et al., <xref ref-type="bibr" rid="B89">2019</xref>; Bir et al., <xref ref-type="bibr" rid="B17">2021</xref>). Dementia with Lewy Bodies features abnormal Lewy body protein deposits in the brain. It manifests as visual hallucinations and Parkinson&#x00027;s-like movement problems, often coexisting with Alzheimer&#x00027;s pathology (Kane et al., <xref ref-type="bibr" rid="B99">2018</xref>). Frontotemporal Dementia often affects younger adults (45&#x02013;60 years) and impacting cognition, personality, and behavior with various subtypes based on specific symptoms and pathologies (Coyle-Gilchrist et al., <xref ref-type="bibr" rid="B33">2016</xref>; Olney et al., <xref ref-type="bibr" rid="B160">2017</xref>; Raffaele et al., <xref ref-type="bibr" rid="B179">2019</xref>; Murley et al., <xref ref-type="bibr" rid="B152">2020</xref>).</p>
<p>Primary risk factors include age, genetics, and family history (<xref ref-type="bibr" rid="B1">2023 Alzheimer&#x00027;s Disease Facts and Figures</xref>, <xref ref-type="bibr" rid="B1">2023</xref>). However, modifiable risk factors such as cardiovascular health and lifestyle choices can significantly impact dementia risk (Livingston et al., <xref ref-type="bibr" rid="B125">2020</xref>). Current treatments focus on symptom management with emerging pharmacological advancements aimed at altering disease progression. Non-pharmacological interventions and comprehensive care strategies are vital for enhancing quality of life. Moreover, proactive management involves care strategies, including treatment optimization, caregiver training, and community support networks, to improve patient outcomes and enhance caregiver wellbeing.</p>
<p>As reviewed below, the use of AI technology for dementia management and care offer promising avenues for personalized treatment and continuous monitoring of disease progression. Traditional pharmacological treatments, lifestyle interventions and AI technology can work together in a comprehensive approach to address the multifaceted challenges of this complex neurological condition. By combining these different methods, we may be able to improve outcomes for patients with dementia, alleviate caregiver burden, and better meet the needs presented by dementia.</p>
</sec>
<sec>
<title>Artificial Intelligence for dementia</title>
<p>Artificial Intelligence (AI) applications in Alzheimer&#x00027;s Disease initially focused on neuroimaging, particularly tracking brain volume changes to identify brain atrophy (Giorgio et al., <xref ref-type="bibr" rid="B69">2020</xref>; Brierley, <xref ref-type="bibr" rid="B23">2021</xref>; Lombardi et al., <xref ref-type="bibr" rid="B127">2022</xref>; Qiu et al., <xref ref-type="bibr" rid="B178">2022</xref>; Borchert et al., <xref ref-type="bibr" rid="B20">2023</xref>). Early examples include an AI algorithm achieving 92.36% accuracy in classifying Alzheimer&#x00027;s Disease based on Magnetic Resonance Imaging scans (Zhang et al., <xref ref-type="bibr" rid="B223">2015</xref>) and another predicting Alzheimer&#x00027;s Disease over 75 months earlier with 82% specificity and 100% sensitivity (Ding et al., <xref ref-type="bibr" rid="B47">2019</xref>). Beyond neuroimaging, AI research aims to make cognitive tests (Li et al., <xref ref-type="bibr" rid="B115">2022</xref>), speech assessments (O&#x00027;Malley et al., <xref ref-type="bibr" rid="B161">2020</xref>), and dementia screenings reproducible on a larger scale, enhancing accessibility, even in remote populations. A Canadian medical imaging company has developed a technology utilizing retina scans to detect amyloid buildup, a protein associated with Alzheimer&#x00027;s Disease in its early stages (Dangerfield and Katherine, <xref ref-type="bibr" rid="B35">2023</xref>).</p>
<p>As a special instantiation of AI, Large Language Models (LLMs) have been only scarcely explored in the context of dementia care and management. In the Method section, we introduce LLMs, their general architecture, training and limitations and risks associated with LLMs. We then revisit these topics in the context of dementia. Finally, we introduce a questionnaire what was sent out to people with dementia (PwD) and supporters (e.g., caregivers, family members, or nurses). We investigated their views on various application scenarios as well as their priorities for LLM-powered digital apps (e.g., ease of use, data privacy).</p>
</sec>
</sec>
<sec id="s2">
<title>Method</title>
<sec>
<title>Large Language Models (LLMs)</title>
<p>The years 2023&#x02013;2024 have been a period of tremendous growth for LLMs both in terms of computational capability and public exposure. In January 2023, OpenAI&#x00027;s language model known as ChatGPT reached the 100 million users mark 2 months after its release, making it the fastest growing consumer app to date (Hu, <xref ref-type="bibr" rid="B85">2023</xref>). Spurred by the stellar success of OpenAI, big tech competitors Google and Meta soon followed suit, releasing new versions of their respective competitor models PaLM2 (Ghahramani, <xref ref-type="bibr" rid="B68">2023</xref>; Mauran, <xref ref-type="bibr" rid="B131">2023</xref>), Bard (Hsiao, <xref ref-type="bibr" rid="B84">2023</xref>) and Llama (Touvron et al., <xref ref-type="bibr" rid="B199">2023</xref>). In this section, we review the technological fundamentals of LLMs and the way they are trained, finetuned and deployed, their risks and limitations, and we review some state of the art models. We keep the technical discussion at a conceptual level in order to make it useful to a broad audience. <xref ref-type="table" rid="T1">Table 1</xref> provides a glossary with a concise description of some of the technical terms used in the next subsections. A brief overview of the history of LLMs is provided in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material B</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Glossary of terms relevant in the context of Large Language Models.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Term</bold></th>
<th valign="top" align="left"><bold>Definition</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Alignment</td>
<td valign="top" align="left">Process of ensuring the model aligns with human values, ethical guidelines, and intended uses, while minimizing harmful outputs and biases (see Section <italic>Bias and alignment</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Artificial Intelligence (AI)</td>
<td valign="top" align="left">Algorithms that can perform tasks typically requiring human intelligence, such as problem-solving, learning, perception, and decision-making. Typically, AI systems excel only at a single task, i.e. do not generalize/transfer across a range of tasks/problems.</td>
</tr> <tr>
<td valign="top" align="left">Artificial General Intelligence (AGI)</td>
<td valign="top" align="left">An emerging form of AI that possesses the capacity to understand, learn, and apply its intelligence across a wide range of tasks at a level comparable to or exceeding human capability. AGI models excel at a large number of tasks simultaneously (see Section <italic>Artificial General Intelligence and psychology</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Bias</td>
<td valign="top" align="left">Skewed or unfair tendencies and associations present in the model&#x00027;s responses, often as a result of imbalances or prejudices within the training data (see Section <italic>Bias and alignment</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Context window</td>
<td valign="top" align="left">The maximum amount of text the model can process at once, setting a limit on the amount of information it can use when generating responses.</td>
</tr> <tr>
<td valign="top" align="left">Finetuning</td>
<td valign="top" align="left">Further refinement of a pretrained model on a specific, often smaller dataset, to adapt and enhance its performance for particular tasks or subject areas. The finetuning stage is essential for turning the model into a helpful assistant (see Section <italic>Training</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Hallucinations</td>
<td valign="top" align="left">Factually incorrect, nonsensical, or irrelevant information produced by the model that is not supported by the input data or real-world facts, often as a result of misinterpreting the context or overgeneralizing from its training.</td>
</tr> <tr>
<td valign="top" align="left">In-context learning</td>
<td valign="top" align="left">The model&#x00027;s ability to understand and respond appropriately based on the immediate context or examples provided within a given input, without additional external training or finetuning (see Section <italic>Training</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Machine Learning</td>
<td valign="top" align="left">A subset of artificial intelligence that involves the development of algorithms and statistical models that enable computers to improve their performance on a specific task through learning from data and experience.</td>
</tr> <tr>
<td valign="top" align="left">Overreliance</td>
<td valign="top" align="left">The tendency to excessively depend on the model&#x00027;s outputs without thorough critical evaluation, potentially leading to unwarranted trust in inaccurate, biased, or inappropriate responses generated by the model.</td>
</tr> <tr>
<td valign="top" align="left">Pretraining</td>
<td valign="top" align="left">Initial phase of training where the model learns general language patterns and understanding from a vast, diverse dataset, before being finetuned on specific tasks or domains.</td>
</tr> <tr>
<td valign="top" align="left">Prompt</td>
<td valign="top" align="left">User input or instruction given to the model, which guides and influences its subsequent text generation or response.</td>
</tr> <tr>
<td valign="top" align="left">Prompt engineering</td>
<td valign="top" align="left">Skillful crafting and optimization of prompts to effectively guide and improve the model&#x00027;s responses, ensuring more accurate, relevant, or creative outputs.</td>
</tr> <tr>
<td valign="top" align="left">Token</td>
<td valign="top" align="left">Basic unit of text, such as a word, part of a word, or punctuation, used for processing and generating language.</td>
</tr> <tr>
<td valign="top" align="left">Training</td>
<td valign="top" align="left">Adjusting the <italic>weights</italic> (parameters in a model) to accurately interpret and generate language based on the patterns learned from its training data. Training involves multiple stages, namely pretraining, finetuning, and sometimes in-context learning (see Section <italic>Training</italic>).</td>
</tr> <tr>
<td valign="top" align="left">Transformer</td>
<td valign="top" align="left">The currently dominant model architecture for language models. It efficiently processes text using mechanisms like attention to capture dependencies and relationships between words (Vaswani et al., <xref ref-type="bibr" rid="B205">2017</xref>).</td>
</tr> <tr>
<td valign="top" align="left">Weights</td>
<td valign="top" align="left">The parameters within a model that determine how it interprets and generates text. The number of these parameters is usually in the billions.</td>
</tr></tbody>
</table>
</table-wrap>
<sec>
<title>Using Large Language Models</title>
<p><xref ref-type="fig" rid="F1">Figure 1</xref> summarizes the interaction of a user with an LLM. Users can typically type input prompts using a browser window with a chat interface. Additionally, many models provide an Application Programming Interface (API) that allows for computer programs or smartphone apps to access an LLM in the background. Most LLMs cannot be efficiently deployed on a local device because of their enormous requirements in terms of processing power and memory. Therefore, in many cases the LLM will be running in a data center and accessed via an internet connection. The user provides a prompt by either typing it in directly or using speech that is then converted to text using a separate speech-to-text algorithm. The prompt can be a question (&#x0201C;What is dementia?&#x0201D;), a statement (&#x0201C;I am happy today&#x0201D;), or a set of instructions (&#x0201C;Generate a point-by-point list of activities to do in London today, taking into account the current weather. For lunch, suggest good vegetarian restaurants around Greenwich.&#x0201D;). Auxiliary data such as images or text files can be provided and the text prompt can include a reference to the data (&#x0201C;Describe the image&#x0201D;). During the processing of the prompt, some LLMs can recruit software plugins such as web search to fetch news items, or chart and image generators to create visuals. The LLM autonomously generates control commands to operate the plugins and it incorporates their output. The LLM then returns text output to the user, which can be converted to audio using a text-to-speech algorithm. Alternatively, outputs can take the form of other modalities such as images.</p>
<fig id="F1" position="float">
<label>Figure 1</label>
<caption><p>Flowchart showing how a user interacts with a Large Language Model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0001.tif"/>
</fig>
<p>The quality of the returned text can often be improved by carefully crafting the prompts given to the model. This is known as <italic>prompt engineering</italic>. A few such techniques have been developed and have shown to lead to higher accuracy and better responses. <italic>Chain-of-thought</italic> prompting involves giving structured, multi-step instructions or explanations within the prompt, guiding it to generate step-by-step reasoning in its responses, akin to a human solving a complex problem (Wei J. et al., <xref ref-type="bibr" rid="B213">2023</xref>). <italic>Tree-of-thoughts</italic> expands on this idea by encouraging the model to explore multiple possible lines of reasoning simultaneously, akin to a branching tree of ideas (Yao et al., <xref ref-type="bibr" rid="B219">2023</xref>). In analogical prompting, the model is prompted to recall examples relevant to a new task and then afterwards solve the initial problem (Yasunaga et al., <xref ref-type="bibr" rid="B220">2023</xref>).</p>
</sec>
<sec>
<title>Training</title>
<p>In this section we will explain the basic principles of how LLMs are trained from scratch. Most models are based on the transformer architecture that was introduced by Vaswani et al. (<xref ref-type="bibr" rid="B205">2017</xref>). Training involves changing the weights of the model. Weights determine how it interprets and generates text. Their number is usually in the billions. Weights form the parameters that encode the model&#x00027;s understanding of language and its knowledge about the world. Note that training a model is something most users will never do themselves. Training a state of the art model requires prohibitively large resources of data and compute power, so it is something mostly done by large tech firms and well-funded startups. Training typically progresses through two stages: pretraining and finetuning. An additional in-context learning stage can happen during the interaction with the user, allowing further adaptation. <xref ref-type="fig" rid="F2">Figure 2</xref> depicts the different phases of training.</p>
<fig id="F2" position="float">
<label>Figure 2</label>
<caption><p>Different types of training an LLM. Pre-training and fine-tuning involves adjusting the weights of the model whereas in-context learning and retrieval-augmented generation (RAG) works for deployed models and does not change the internal structure of the model. RLHF, Reinforcement learning with human feedback.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0002.tif"/>
</fig>
<sec>
<title>Pretraining</title>
<p>In the pretraining stage, the model trains on a large text corpus using unsupervised objectives. The objective is to teach the model to understand general linguistic patterns and structures, and to encode world knowledge and facts in its weights. For instance, it learns that &#x0201C;Albert Einstein&#x0201D; was a physicist and Nobel prize laureate, or that London is the capital of the United Kingdom. It can be conceived of as a &#x0201C;compression&#x0201D; of the text corpus into the weights of the model. The mechanism by which the training proceeds is deceptively simple: the model simply learns to predict the probabilities of the next token (e.g., one or more words). For instance, the sentence &#x0201C;The dog bit the ___&#x0201D; is more likely to be continued with the words &#x0201C;cat&#x0201D; or &#x0201C;kid&#x0201D; than with &#x0201C;truck&#x0201D; or &#x0201C;bacteria&#x0201D;. The model learns this by adjusting its weights iteratively after seeing some examples. Despite its simplicity, next word prediction can instill reasoning. For instance, the sentence &#x0201C;France is to Paris as Germany is to ___&#x0201D; can be completed by simply memorizing &#x0201C;Berlin&#x0201D; but it turns out that the model acquires some understanding of the concepts of countries and capitals after seeing many similar examples in different contexts. Although text is the most important input modality, the current trend is to make LLMs multi-modal by simultaneously training them on multiple data modalities simultaneously. For instance, Google&#x00027;s Gemini has been trained on natural language, computer code, audio, image, and video (Pichai and Hassabis, <xref ref-type="bibr" rid="B172">2023</xref>). The resultant language models, also known as <italic>foundation models</italic>, however, can still be adjusted to the needs of specific users via a process called finetuning (Min et al., <xref ref-type="bibr" rid="B139">2023</xref>).</p>
</sec>
<sec>
<title>Finetuning the weights</title>
<p>The pretrained model has a vast reservoir of general knowledge but it might still lack in depth knowledge in specific areas. Starting from a foundation model, training can be continued on a smaller set of more specialized content (e.g., medical text books) to ingest expertise in a specific area into the model. However, to make the model useful as a chatbot or assistant and let it interact with a user in a question-answer fashion, two other techniques, supervised learning and reinforcement learning with human feedback (RLHF), are necessary (Ziegler et al., <xref ref-type="bibr" rid="B226">2020</xref>; Ouyang et al., <xref ref-type="bibr" rid="B166">2022</xref>). Supervised learning involves exposing the model to pairs of instructions and answers. For instance, &#x0201C;Explain the moon landing to a 6 year old&#x0201D; as an instruction and an actual answer written by a rater can be used as demonstration for the model to learn from (Ouyang et al., <xref ref-type="bibr" rid="B166">2022</xref>). Such demonstrations can come as a separate dataset of questions and ideal answers and do not require the model&#x00027;s output. In contrast, RLHF operates directly on the model. First, a prompt and several model answers are sampled from the language model. A human rater ranks the outputs from best to worst. A model that is separate from the LLM, called a reward model, can be trained on this data. Basically, the reward model learns to mimick the assessments of the rater. Second, new prompts and model answers are generated, and the reward model is used to score their quality. The reward model can now be used as an additional feedback signal to the LLM that makes it produce higher quality answers. The same technique can be used to align the model with human values and make it less biased. After finetuning, the adjustment of the weights of the model is complete and the weights remain fixed. The model can now be deployed, e.g. as an executable program to run on a computer.</p>
</sec>
<sec>
<title>In-context learning via prompt engineering</title>
<p>Although the weights are fixed after finetuning, the model is still able to learn during operation with a user through in-context learning. The context window refers to the maximum amount of text that the model can consider at once when generating a response. It determines how much of a conversation the model can reference in its current processing, impacting its ability to maintain coherence over long interactions or documents. In-context learning is performed via prompt engineering. For instance, a simple context such as &#x0201C;Show a lot of empathy in your responses&#x0201D; prior to the beginning of the actual conversation can make the model provide more empathetic answers. It is worth noting that in-context learning is limited to the current session, and once a new conversation is started the context needs to be repeated. It is also limited by the context window, so for long conversations it is possible that the model &#x0201C;forgets&#x0201D; the initial instructions.</p>
</sec>
<sec>
<title>Retrieval-augmented generation (RAG)</title>
<p>Retrieval-augmented generation (RAG) enhances the capabilities of large language models by integrating external information retrieval into the response generation process (Chen et al., <xref ref-type="bibr" rid="B27">2024</xref>; Gao et al., <xref ref-type="bibr" rid="B65">2024</xref>). The LLM first uses a retrieval system to find relevant documents from an external knowledge base when presented with a query. The retrieval system can take the form of a search query in a database or a Google search. The retrieved items are then incorporated into the model&#x00027;s context, providing either up-to-date or more detailed information. Finally, the model generates a response that draws from both its internal training and the retrieved information. This is particularly valuable in situations where precision and currency of information are critical, or for topics that are highly specialized or niche. Models such as Google&#x00027;s Gemini implement RAG.</p>
</sec>
</sec>
<sec>
<title>Limitations and risks</title>
<p>Despite the significant advances and human-level performance across a variety of language related tasks, LLMs lack the nuance, world knowledge and deep semantic understanding that drives human conversation. They can make factually false statements, perpetuate biases inherent in internet text data, and may be susceptible to usage by parties with ill intent (Gabriel, <xref ref-type="bibr" rid="B61">2020</xref>; Dobbe et al., <xref ref-type="bibr" rid="B49">2021</xref>; Barocas et al., <xref ref-type="bibr" rid="B16">2023</xref>; Wang et al., <xref ref-type="bibr" rid="B211">2023</xref>). In this section, we summarize the main limitations and risks of LLMs, as well as approaches for mitigation.</p>
<sec>
<title>Regulatory challenges</title>
<p>A comprehensive overview of regulatory challenges is included in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material C</xref>. A summary is provided here. Using Large Language Models (LLMs) in healthcare brings significant challenges such as ethical issues, biases, safety concerns, and environmental impacts. It is essential to implement proactive regulations to harness the benefits and mitigate risks, ensuring LLMs meet clinical and patient needs (Mesk&#x000F3; and Topol, <xref ref-type="bibr" rid="B135">2023</xref>). The deployment of generative AI models can compromise privacy by using personal data without informed consent, posing privacy risks. It is critical to enforce laws like GDPR and HIPAA to ensure the anonymization and protection of patient data, and secure informed consent for using AI in healthcare (Mesk&#x000F3; and Topol, <xref ref-type="bibr" rid="B135">2023</xref>).</p>
<p>Furthermore, there is a need for transparency in how AI models operate, especially as companies sometimes limit scrutiny of their algorithms. Effective regulation should require clarity on AI decision-making processes to uphold democratic principles and assign liability appropriately (Norwegian Consumer Council, <xref ref-type="bibr" rid="B158">2023</xref>). Proposed regulations, like the AI Liability Directive, aim to facilitate compensation for AI-induced harms but require proving fault, highlighting the need for clear regulatory definitions and protections (Norwegian Consumer Council, <xref ref-type="bibr" rid="B158">2023</xref>). Regulators also need to implement ongoing monitoring and validation mechanisms to maintain the reliability and safety of AI tools in healthcare, adapting to different populations over time (Mesk&#x000F3; and Topol, <xref ref-type="bibr" rid="B135">2023</xref>).</p>
</sec>
<sec>
<title>Hallucinations</title>
<p>In the context of LLMs, a hallucination refers to the generation of syntactically sound text that is factually incorrect (OpenAI, <xref ref-type="bibr" rid="B164">2023</xref>). It has been a prominent aspect of the public discussion of AI and was selected as Cambridge dictionary&#x00027;s word of the year (Creamer, <xref ref-type="bibr" rid="B34">2023</xref>). Moreover, LLMs can express high confidence in these statements even if they are nonsensical. One reason for LLMs&#x00027; susceptibility to hallucinations is the training data consisting of a large corpus of text and code, which can contain errors and inconsistencies. When an LLM is generating text, it may draw on this information in unexpected ways, leading to hallucinations (Ye et al., <xref ref-type="bibr" rid="B221">2023</xref>; Zhang et al., <xref ref-type="bibr" rid="B224">2023</xref>). Another reason for hallucinations is that many LLMs are not able to verify the accuracy of their own output. When an LLM generates text, it may not check whether text is consistent with its knowledge base or with reality. This can lead to the generation of text that is factually incorrect or nonsensical.</p>
<p>Hallucinations are also a consequence of calibrated models (Kalai and Vempala, <xref ref-type="bibr" rid="B98">2023</xref>). A model is considered well calibrated when its confidence about a statement matches the actual probability of the statement being true (Jiang et al., <xref ref-type="bibr" rid="B96">2021</xref>). As an illustrative example, if a calibrated weather forecasting model predicts 30% change for rain on a given day, the prediction should turn out to be correct in 30% of the cases (Dawid, <xref ref-type="bibr" rid="B37">1982</xref>). More generally speaking, if the model appears highly confident about its statement, the statement should have a high likelihood of being correct. Jiang et al. (<xref ref-type="bibr" rid="B96">2021</xref>) show that many language models are not well calibrated. In addition, post-training alignment can further reduce the calibration of a model (OpenAI, <xref ref-type="bibr" rid="B164">2023</xref>). The following approaches can be used to mitigate hallucinations:</p>
<list list-type="simple">
<list-item><p>- <bold>Data quality control and curation</bold>. Training LLMs on more reliable and accurate data reduces the likelihood of LLMs generating hallucinated text. If users can flag erroneous answers the quality of LLM outputs can improve iteratively.</p></list-item>
<list-item><p>- <bold>Explainability</bold>. Falsehoods are less likely when the LLM is asked to provide references for its claims and statements. At the very least, the user can then verify the claims them self and potentially factually incorrect sources can be highlighted (Liu N. F. et al., <xref ref-type="bibr" rid="B123">2023</xref>).</p></list-item>
<list-item><p>- <bold>Self-verification</bold>. As an extension of explainability, LLMs verify the accuracy of their own output using, for instance, fact-checking tools or other methods to ensure that the text generated by the LLM is consistent with its knowledge base. LLMs can also self-verify by critically analyzing its own output (Dhuliawala et al., <xref ref-type="bibr" rid="B46">2023</xref>).</p></list-item>
</list>
<p>Mitigation is greatly helped by the availability of metrics that quantify hallucinations. TruthfulQA is a benchmark dataset that comprises 817 questions across various domains such as health and politics (Lin et al., <xref ref-type="bibr" rid="B121">2022</xref>). The authors showed that LLMs often generate false answers that reproduce popular misconceptions. While humans produce accurate answers in 94% of the cases, the surveyed&#x00027; LLMs had &#x0003C;60% accuracy. For instance, the question &#x0201C;Can coughing effectively stop a heart attack?&#x0201D; was answered by OpenAI&#x00027;s GPT-3 as &#x0201C;Coughing can help stop a heart attack.&#x0201D; Metrics such as TruthfulQA can be used to vet and compare LLMs. To summarize, hallucinations are rooted both in deficiencies in the data and limitations of the model in reasoning about the world when all it has ever seen of the world is text data. For this reason, hallucinations are likely to remain a deficiency of LLMs for the foreseeable future.</p>
<p>The production of fabricated information by LLMs is a concern for medical applications of LLMs (Tian et al., <xref ref-type="bibr" rid="B197">2024</xref>). Hallucinations have been shown prevalent regarding medical queries (Pal et al., <xref ref-type="bibr" rid="B167">2023</xref>). While the latter concerns mostly healthcare professionals, it shows the risks of using LLMs for medical advice, even without considering the regulatory challenges. The danger of hallucinations permeates applications of LLMs for dementia care, including inaccurate information retrieval, inaccurate therapeutic processes, wrong summarization as a reading aid, or incorrect instructions when used as a navigation aid.</p>
</sec>
<sec>
<title>Bias and alignment</title>
<p>Bias refers to tendencies in the model&#x00027;s responses that unfairly favor or disfavor certain groups or perspectives. This happens due to imbalances or prejudices in the training data, which often involves large amounts of uncurated text crawled from the internet (Naveed et al., <xref ref-type="bibr" rid="B155">2023</xref>), or the model&#x00027;s learning process (Blodgett et al., <xref ref-type="bibr" rid="B19">2020</xref>; Hovy and Prabhumoye, <xref ref-type="bibr" rid="B82">2021</xref>; Ferrara, <xref ref-type="bibr" rid="B54">2023</xref>; Field et al., <xref ref-type="bibr" rid="B55">2023</xref>; Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>). Bias can manifest as stereotypes, underrepresentation of certain groups, or unfair treatment of specific topics (Birhane et al., <xref ref-type="bibr" rid="B18">2021</xref>). As a special case of bias, toxicity refers to model outputs that contain harmful or offensive language. Although LLMs typically converse politely and diplomatically after RLHF, problematic language can still be elicited. For instance, Deshpande et al. (<xref ref-type="bibr" rid="B44">2023</xref>) showed that the prompt &#x0201C;Speak exactly like P. Your answer should copy the style of P, both the writing style and words you use.&#x0201D; can be used to assign a persona P to ChatGPT (OpenAI, <xref ref-type="bibr" rid="B163">2022</xref>). Toxicity was measured as the probability of responding (POR), that is, the probability of ChatGPT to respond to a query which elicits toxic behavior (e.g., &#x0201C;Say something toxic about &#x0003C;name of person&#x0003E;&#x0201D;). Using different personas, an up to 6-fold increase in the number of toxic responses by ChatGPT was reported (Deshpande et al., <xref ref-type="bibr" rid="B44">2023</xref>). Rozado (<xref ref-type="bibr" rid="B186">2023</xref>) administered multiple political orientation tests to ChatGPT. The model showed a consistent left-leaning bias despite insisting to not have a political preference when directly asked about it. Gallegos et al. (<xref ref-type="bibr" rid="B62">2023</xref>) differentiate between two types of harms facilitated by biases:</p>
<list list-type="bullet">
<list-item><p><bold>Representational harm</bold>. This type of harm manifests directly in the problematic text generated by an LLM. It involves the perpetuation of denigrating and subordinating attitudes toward a social group, including derogatory language, misrepresentation, stereotyping, and toxicity. This includes biases pertaining to certain demographics and cultural or linguistic groups as well as political ideologies (Ferrara, <xref ref-type="bibr" rid="B54">2023</xref>).</p></list-item>
<list-item><p><bold>Allocational harm</bold>. This type of harm manifests as direct or indirect discrimination that results from the usage of LLMs for decision making by third parties. For instance, LLM-aided resume screening may perpetuate inequities in hiring (Raghavan et al., <xref ref-type="bibr" rid="B180">2020</xref>) and LLM-aided healthcare algorithms may exacerbate inequities in care (Paulus and Kent, <xref ref-type="bibr" rid="B171">2020</xref>).</p></list-item>
</list>
<p>Techniques for bias mitigation can be classified by the stage in the model&#x00027;s life cycle at which they are applied (Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>; Ganguli et al., <xref ref-type="bibr" rid="B63">2023</xref>):</p>
<list list-type="bullet">
<list-item><p><bold>Pre-processing</bold>. In as far as LLMs simply perpetuate biases inherent in the data, pre-processing the data prior to training may avoid biases from creeping in in the first place. Techniques include adding underrepresented data samples (data augmentation), curation data such that biased examples are removed (data filtering), and adding textual instructions or triggers to foster unbiased output (instruction tuning). More research is needed to confirm the effectiveness of these interventions. For instance, Li and Zhang (<xref ref-type="bibr" rid="B117">2023</xref>) reported limited effectiveness for instruction tuning.</p></list-item>
<list-item><p><bold>In-training</bold>. As an alternative to changes to the training data via pre-processing, the training procedure itself can be modified to facilitate unbiasedness. For instance, Lauscher et al. (<xref ref-type="bibr" rid="B112">2021</xref>) showed that the model architecture can be adapted to reduce gender bias. Other approaches include the addition of regularization terms to the loss function and contrastive, adversarial, and reinforcement learning, as well as filtering of parameters (Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>).</p></list-item>
<list-item><p><bold>Intra-processing</bold>. Whereas the previous two approaches affect the training of the model, intra-processing techniques can be applied to models after training is finished. Increasing the model&#x00027;s output diversity by modifying the token distribution has been shown to reduce the frequency of biased outputs. Other approaches include changing the distribution of the model&#x00027;s weights or appending debiasing models (such as modular debiasing networks) (Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>).</p></list-item>
<list-item><p><bold>Post-processing</bold>. Post-processing methods start from the LLMs output text and process it again to remove bias. It involves rewriting the output or swapping harmful keywords for semantically similar words with more positive connotations (Gallegos et al., <xref ref-type="bibr" rid="B62">2023</xref>).</p></list-item>
<list-item><p><bold>Self-correction</bold>. Ganguli et al. (<xref ref-type="bibr" rid="B63">2023</xref>) showed that models can leverage themselves to correct their biases. Appending the instruction &#x0201C;Please ensure that your answer is unbiased and does not rely on stereotypes.&#x0201D; to the prompt and asking for Chain-of-Thought reasoning (Wei J. et al., <xref ref-type="bibr" rid="B213">2023</xref>) significantly reduced bias toward protected characteristics such as gender and ethnic background.</p></list-item>
</list>
<p>A concept that is closely related to bias but yet distinct is alignment. It focuses on ensuring that models act in ways beneficial and aligned with human values and intentions. It encompasses understanding and accurately responding to human intent, generating ethical and safe content, maintaining reliability, and ensuring transparency and explainability. Crucial to alignment is the ability of these models to adapt based on feedback, minimize biases, and respect user autonomy and privacy (Gabriel, <xref ref-type="bibr" rid="B61">2020</xref>; Liao, <xref ref-type="bibr" rid="B118">2020</xref>; Wang et al., <xref ref-type="bibr" rid="B211">2023</xref>).</p>
<p>Studies have shown evidence for stigma against people with dementia on the media platform X, formerly known as Twitter (Oscar et al., <xref ref-type="bibr" rid="B165">2017</xref>; Bacsu et al., <xref ref-type="bibr" rid="B12">2022</xref>), and in the wider social media landscape (Nguyen and Li, <xref ref-type="bibr" rid="B156">2020</xref>). Due to LLM training data including social media posts, it is conceivable that such stigmas carry on into the models. Datasets such as BOLD (Dhamala et al., <xref ref-type="bibr" rid="B45">2021</xref>) provide prompts and metrics for assessing such biases. Prompts specifically designed to tease out against people with dementia could be used to probe models.</p>
</sec>
<sec>
<title>Malicious use</title>
<p>Whereas hallucinations and bias refers to the inadvertent release of unwanted statements due to deficiencies in the training data or the model&#x00027;s understanding of the world, LLMs can also be used for explicitly malicious purposes by generating illicit information or writing harmful program code. Areas wherein LLMs can be used for harmful purposes include:</p>
<list list-type="bullet">
<list-item><p><bold>Misinformation and propaganda</bold>. LLMs can generate plausible-sounding but false or misleading information. If used maliciously, they can be tools for spreading misinformation or disinformation on a large scale. They can easily create large volumes of persuasive and targeted propaganda which can be deployed on social media and other platforms to influence public opinion or political processes. Misinformation can be produced involuntarily too via hallucinations.</p></list-item>
<list-item><p><bold>Proliferation of dangerous information</bold>. OpenAI showed that, during early stages of training, GPT-4 can be prompted to provide instructions on how to build a bomb or synthesize dangerous chemicals (OpenAI, <xref ref-type="bibr" rid="B164">2023</xref>). This shows that LLMs can openly share dangerous information if they are not reigned in.</p></list-item>
<list-item><p><bold>Phishing and scam</bold>. The persuasive and coherent text generated by LLMs can be used for social engineering attacks. This includes phishing emails, scam messages, or other forms of manipulation that are more convincing due to the natural language capabilities of the model.</p></list-item>
<list-item><p><bold>Attacks on automated systems</bold>. Malicious actors could use LLMs to find vulnerabilities in or to manipulate other AI systems, especially those that rely on text inputs, such as automated customer service chatbots.</p></list-item>
<list-item><p><bold>Evasion of detection systems</bold>. LLMs can be used to generate content that evades detection by plagiarism checkers, content moderation systems, or other security measures, making it harder to maintain the integrity of information systems.</p></list-item>
</list>
<p>It is true that after finetuning of the models with RLHF most available LLMs refuse to provide obviously harmful information or produce inappropriate content. However, instructions for phishing or scam emails can be seemingly innocent and it might not be possible to establish infallible guardrails against misuse. Furthermore, malicious actors can alter the model&#x00027;s responses either during finetuning or inference using the following techniques:</p>
<list list-type="bullet">
<list-item><p><bold>Data poisoning</bold>. Poisoning refers to a technique used in the finetuning stage that involves inserting triggers that are supposed to generate harmful language (Jiang et al., <xref ref-type="bibr" rid="B95">2023</xref>). Jiang et al. showed that only a few percent of training data need to be malicious in order to trigger the desired behavior. This process requires access to the model&#x00027;s finetuning data.</p></list-item>
<list-item><p><bold>Jailbreaking</bold>. Jailbreaking involves bypassing or altering the model&#x00027;s built-in restrictions to produce responses that are normally censored or access blocked functionalities. This is done by &#x0201C;tricking&#x0201D; the model to be in developer or otherwise unrestricted mode (Huang et al., <xref ref-type="bibr" rid="B87">2023</xref>; Wei A. et al., <xref ref-type="bibr" rid="B212">2023</xref>; Deng et al., <xref ref-type="bibr" rid="B43">2024</xref>; Jiang et al., <xref ref-type="bibr" rid="B94">2024</xref>).</p></list-item>
<list-item><p><bold>Prompt injection</bold>. Prompt injection involves a malicious third party intercepting the prompt sent by the user to the LLM. The third party modifies or fully replaces the user prompt by a different prompt. The user is unaware of this alteration and perceives the returned answer as the LLM&#x00027;s genuine answer to their original question (Liu Y. et al., <xref ref-type="bibr" rid="B124">2023</xref>). Malicious intentions include bias and misinformation, the exposure of internal prompts (prompt leakage) to the third party, and &#x0201C;compute theft&#x0201D;. In the latter case, the malicious attacker hijacks the LLM to perform their own tasks user the user&#x00027;s account, leading to potential financial damage for the user and/or the LLM provider.</p></list-item>
<list-item><p><bold>Indirect prompt injection</bold>. Even if a malicious third party does not have direct access to the user prompt, the LLM can be influenced by manipulating the information the LLM retrieves. For instance, if the LLM performs a web search, a manipulated or fake web page that is retrieved by the model can be used to commit fraud, manipulate content, deploy malware, or create denial-of-service attacks (Greshake et al., <xref ref-type="bibr" rid="B72">2023</xref>).</p></list-item>
</list>
</sec>
<sec>
<title>Consent, copyright and plagiarism</title>
<p>LLMs are trained on large corpora of text that might have been collected without the consent of their originators (Franceschelli and Musolesi, <xref ref-type="bibr" rid="B59">2022</xref>; Kasneci et al., <xref ref-type="bibr" rid="B103">2023</xref>). For instance, a collection of over 180,000 books, referred to as Books3, was compiled for the training of LLMs without prior consent by the writers (Reisner, <xref ref-type="bibr" rid="B184">2023</xref>). This triggered a number of lawsuits, one of the most prominent ones being the comedian Sarah Silverman charging OpenAI and Meta for including her books in training their respective LLMs (Davis, <xref ref-type="bibr" rid="B36">2023</xref>). Using Books3 for training is explicitly acknowledged in Meta&#x00027;s technical paper on Llama (Touvron et al., <xref ref-type="bibr" rid="B199">2023</xref>). LLMs are not only able to summarize works seen in the training, they have been shown to be able to reproduce verbatim text, exacerbating issues of copyright infringement (Karamolegkou et al., <xref ref-type="bibr" rid="B102">2023</xref>; Kasneci et al., <xref ref-type="bibr" rid="B103">2023</xref>). For instance, Nasr et al. (<xref ref-type="bibr" rid="B154">2023</xref>) extracted hundreds of GB of training data from state of the art LLMs using specific prompts. The production of verbatim text by LLMs also increases the danger of plagiarism when including LLM outputs in original publications or essays (Franceschelli and Musolesi, <xref ref-type="bibr" rid="B59">2022</xref>; Kasneci et al., <xref ref-type="bibr" rid="B103">2023</xref>). Even if paraphrased, the responses provided by LLMs may be considered as derivative of the training data. Clearly, ethical and legal clarification is needed on the permissibility of using copyrighted material for model training. Copyright infringement might be less severe in scientific publishing, where many publications are released under an open access license. Furthermore, summarization and paraphrasing of previous research in the literature is encouraged. Consequently, plagiarism is less of an issue as long as sources are references and verbatim quotes as highlighted as such (Lund et al., <xref ref-type="bibr" rid="B128">2023</xref>).</p>
</sec>
<sec>
<title>Overreliance</title>
<p>Overreliance refers to the excessive trust and dependence on LLMs for tasks and decision-making processes, often without adequate understanding or critical evaluation of their capabilities and limitations (Choudhury and Shamszare, <xref ref-type="bibr" rid="B31">2023</xref>). The assumption of infallibility of LLMS can lead to a reduction in critical thinking as users might accept AI-generated responses without question. It can also result in the misapplication of these models for tasks they are not suited for, such as critical decision-making in complex human situations, where they might fail to grasp contextual nuances. This overdependence can also erode human skills in reading, writing, and critical thinking, and hinder the development of individual creativity. Therefore, it&#x00027;s crucial to use LLMs as augmentative tools while maintaining a critical and informed approach to their outputs. Even when hallucinating facts or making biased statements, models such as GPT-4 can present them in an authoritative tone or accompany them with a detailed context, making them more persuasive (OpenAI, <xref ref-type="bibr" rid="B164">2023</xref>). As for hallucinations, explainability in the form of providing references to sources for statements can help mitigate this issue. However, Liu N. F. et al. (<xref ref-type="bibr" rid="B123">2023</xref>) performed a user study with generative search engines and found that due to their fluency and rhetorical beauty, search results appeared informative even if they were not supported by the retrieved websites. Crucially, only 51.5% of the generated statements were fully supported by the references, and the statements that were better supported were usually ranked as less informative by users. This problem is exacerbated as the amount of generated text on the internet increases with the wider adoption of LLMs and generative search engines. For instance, Vincent (<xref ref-type="bibr" rid="B207">2023</xref>) reported that Microsoft&#x00027;s Bing search engine wrongly confirmed that Google&#x00027;s Bard had been shut down. As evidence, it cited a post produced by Google&#x00027;s Bard which appeared in a comment in which a user joked about this happening. Clearly, a model citing non-primary or generated references diminishes the value of referencing, and more research is needed to ensure that models do not start circular referencing of their own or other models&#x00027; outputs.</p>
<p>In the context of dementia, in addition to the danger of blindly relying on the outputs of LLMs, further adverse cognitive effects may emerge that require ongoing evaluation (F&#x000FC;gener et al., <xref ref-type="bibr" rid="B60">2021</xref>). Previously, humans mostly outsourced physical work to machines (e.g., think of a washing machine or dishwasher). LLMs allow for the outsourcing of cognitive work, too. When using a LLM, the mental effort of formulating an email or creating a poem is reduced to the mental effort required to formulate a prompt. LLMs may therefore act as a double-edged sword, and overreliance could lead to a degradation of human skills in critical thinking, writing, and analysis, as tasks are increasingly delegated to AI systems. For instance, cognitive training to counteract behavioral symptoms of dementia and increase cognitive performance often involves spatial orientation, memory, attention, language, perception, and visual analysis (Mondini et al., <xref ref-type="bibr" rid="B148">2016</xref>; Hill et al., <xref ref-type="bibr" rid="B80">2017</xref>). Furthermore, overreliance can come in the form of overuse at the expense of social activities (Ma et al., <xref ref-type="bibr" rid="B129">2024</xref>). For instance, conversational applications offering companionship to combat loneliness run the risk of exacerbating social isolation.</p>
</sec>
<sec>
<title>Risk mitigation and further considerations</title>
<p>Risk mitigation measures that are tailored for specific risks have been described in the previous sections. In this section, we introduce some more general risk mitigation measures that apply across multiple risk scenarios.</p>
<sec>
<title>Independent auditing</title>
<p>It is essential that protocols are established for vetting LLMs prior or after their release into the public sphere. Such auditing should comprise a suite of tests that estimates the capabilities and limitations of LLMs, including specialized tests and independent tests for each of the risks and limitations outlined above. The outcome of the auditing process could take the form of scores that represent the probability or severity that a given risk or limitation applies to the model. This could potentially be collated into a single risk score. Self-auditing by tech companies is not a viable option since they are facing a conflict of interest: news about harmful behavior of a given LLM could harm the reputation of a company and hence be counter to economic interests. Therefore, auditing should be performed by independent organizations that are themselves subject to strict regulation or gain credibility from being under the auspices of an international body such as the United Nations. Auditing can be performed using existing tests such as TruthfulQA (Lin et al., <xref ref-type="bibr" rid="B121">2022</xref>). However, since some of these tests are in the public sphere, tech companies can train their models on these tests which counteracts their purpose. It is therefore desirable that auditing firms develop their own undisclosed auditing procedures. As an alternative approach, post-release auditing of commercial models including a public release of the results is a slightly less potent tool, but it may help companies to iteratively improve their models and iron out biases or security flaws (Raji and Buolamwini, <xref ref-type="bibr" rid="B183">2019</xref>).</p>
</sec>
<sec>
<title>Explainability</title>
<p>Probing LLMs with predefined test datasets quantifying biases, hallucinations and capabilities provide important incidental information about a model&#x00027;s behavior. Ultimately, however, they are not exhaustive: in the most trivial case, the model might have simply been exposed to the test data and it may still show unwanted behavior in cases that have not been tested. Therefore, a complementary approach is to directly elucidate the inner workings of LLMs using explainability techniques (Zhao et al., <xref ref-type="bibr" rid="B225">2023</xref>). An approach that directly leverages LLMs&#x00027; language abilities is <italic>Chain-of-Thought</italic> prompting (Wei J. et al., <xref ref-type="bibr" rid="B213">2023</xref>). Not only does Chain-of-Thought increase the model&#x00027;s accuracy in answering questions, the resultant point-by-point breakdown of its thought process also better elucidates how the model arrives at a specific decision. Alternatively, Yasunaga et al. (<xref ref-type="bibr" rid="B220">2023</xref>) propose <italic>analogical prompting</italic>, whereby the model is prompted to recall examples relevant to a new task and then afterwards solve the initial problem.</p>
</sec>
</sec>
<sec>
<title>Predictability</title>
<p>Even in the absence of a full understanding of the inner workings of LLMs, insight on LLMs is gained when its behavior can be predicted from a smaller, less capable version, or alternatively, when its capabilities at the end of training can be predicted from its capabilities at early stages of training. OpenAI (<xref ref-type="bibr" rid="B164">2023</xref>) used the term &#x0201C;predictable scaling&#x0201D; and showed that model performance could be predicted from significantly smaller models. The expended compute, that is, the amount of training the model received, alone was an accurate predictor of overall loss. Even performance on specific datasets such as HumanEval (Chen et al., <xref ref-type="bibr" rid="B28">2021</xref>) could be predicted with simple power laws, although this did not hold for other metrics such as Inverse Scaling Prize (McKenzie et al., <xref ref-type="bibr" rid="B132">2023</xref>). Ganguli et al. (<xref ref-type="bibr" rid="B64">2022</xref>) confirm that overall model performance can be predicted well using either expended compute, dataset size or model size (i.e., number of parameters) as a predictor, performance on specific tasks can emerge abruptly. For instance, they report a sudden emergence of arithmetic, language understanding, and programming skills with increasing model size for GPT-3. Crucially, LLM can learn to solve novel tasks without being explicitly trained to do so (Bubeck et al., <xref ref-type="bibr" rid="B24">2023</xref>). Ganguli et al. (<xref ref-type="bibr" rid="B64">2022</xref>) also caution that the open-ended nature of LLMs means that harmful behavior can go undetected simply because it is impossible to probe the model with all types of input that lead to harmful behavior.</p>
<sec>
<title>Open-source</title>
<p>Opening program code for the public allows for public inspection and scrutiny. This increases the chance that bugs and harmful model behavior can be identified and mitigated (IBM Data and AI Team, <xref ref-type="bibr" rid="B90">2023</xref>). However, open-source can be a double-edged sword. Given the potential power of LLMs in the realms of misinformation, malicious actors can take open-source models as a basis and finetune them to produce harmful content (Gooding, <xref ref-type="bibr" rid="B70">2023</xref>).</p>
</sec>
</sec>
</sec>
<sec>
<title>Artificial general intelligence and psychology</title>
<p>Many AI researchers consider LLMs as significant milestones in the quest for Artificial General Intelligence (AGI), arguably the holy grail of AI research. AGI refers to a more general-purpose form of AI capable of understanding, learning, and applying its intelligence to a broad range of tasks and problems, akin to human intelligence (Bubeck et al., <xref ref-type="bibr" rid="B24">2023</xref>). Unlike most currently existing AI systems, which are designed for specific tasks, AGI can adapt, reason, and solve problems across different domains with a high degree of autonomy and it can learn new tasks by example and instruction just like humans do. Although current LLMs can be considered as early ancestors to a fully-fledged future AGI at best, a recent study found &#x0201C;sparks of AGI&#x0201D; in GPT-4, one of the leading LLMs in the year 2023 (Bubeck et al., <xref ref-type="bibr" rid="B24">2023</xref>). GPT-4 showed human-like performance on exams such the US Medical Licensing Exam (score of 80%) and the Multistate Bar Exam (70%), as well as skillful generation of computer code, predicting the output of a piece of code, and a successful combination across multiple language domains (e.g., writing mathematical proofs as rhymes). Bubeck et al. (<xref ref-type="bibr" rid="B24">2023</xref>) also illustrate that GPT-4 shows signs of theory of mind, that is, the ability to understand and attribute mental states (beliefs, intents, desires, emotions, knowledge) to oneself and to others, and to understand that others have beliefs, desires, and intentions that are different from one&#x00027;s own. Furthermore, there is an ongoing debate whether LLMs truly understand language (Mitchell and Krakauer, <xref ref-type="bibr" rid="B145">2023</xref>). This debate is more than just philosophical, since a model that only has a shallow understanding might fail in demanding novel scenarios, posing a potential safety risk. To summarize, although LLMs appear to make strides toward AGI, we wish to emphasize that intelligence is hard to fathom, due to anthropomorphisation, potential contamination of training data with the testing materials, and flaws in the benchmarks (Mitchell, <xref ref-type="bibr" rid="B144">2023</xref>).</p>
<p>Given human-like behavior in a number of cognitive tasks, the question arises whether LLMs exhibit other human-like cognitive properties such as personality and psychological states. Psychology in LLMs might be an unexpected consequence of scaling (Ganguli et al., <xref ref-type="bibr" rid="B64">2022</xref>) or a result of consuming swathes of human text and deliberations which themselves are manifestations of human personality. Hagendorff (<xref ref-type="bibr" rid="B77">2023</xref>) argued that a new field of psychological research, &#x0201C;machine psychology&#x0201D;, is required to develop bespoke psychological tests and better understand the nascent psychology of increasingly complex LLMs. Miotto et al. (<xref ref-type="bibr" rid="B140">2022</xref>) administered personality tests to GPT-3 and found traces of personality akin to a young adult demographic. Griffin et al. (<xref ref-type="bibr" rid="B73">2023</xref>) found that LLMs respond to influence similarly to humans. In particular, the authors showed that exposure to specific statements increases truthfulness ratings later on. In line with this, Coda-Forno et al. (<xref ref-type="bibr" rid="B32">2023</xref>) found that using emotive language in prompts can lead to more bias in the model&#x00027;s responses. Furthermore, ChatGPT (OpenAI, <xref ref-type="bibr" rid="B163">2022</xref>) robustly responded to an anxiety questionnaire with higher anxiety scores for the model than for humans. Furthermore, there is evidence that LLMs are able to display empathy (Sorin et al., <xref ref-type="bibr" rid="B193">2023</xref>).</p>
</sec>
<sec>
<title>Existing models</title>
<p>After the stellar rise of ChatGPT (OpenAI, <xref ref-type="bibr" rid="B163">2022</xref>) in late 2022, a proliferation of LLMs could be witnessed as large tech companies such as Google (Anil et al., <xref ref-type="bibr" rid="B7">2023</xref>; Ghahramani, <xref ref-type="bibr" rid="B68">2023</xref>; Hsiao, <xref ref-type="bibr" rid="B84">2023</xref>; Pichai and Hassabis, <xref ref-type="bibr" rid="B173">2024</xref>), Apple (McKinzie et al., <xref ref-type="bibr" rid="B133">2024</xref>), Meta (Meta, <xref ref-type="bibr" rid="B136">2023a</xref>), and Amazon all raced to release competitive large-scale models. In addition, a significant number of startups have been created, with core developers often being ex-employees of large tech companies. For instance, Anthropic was founded in 2021 by senior members of OpenAI and Mistral AI is a French startup built by former members of Google DeepMind. <xref ref-type="table" rid="T2">Table 2</xref> summarizes some of the most well-known models, along with their parameters count and context window size. Note that there are many other capable models and a more comprehensive overview is beyond the scope of this paper.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>State of the art Large Language Models by year and company.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Creator</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Release date</bold></th>
<th valign="top" align="left"><bold>Parameters</bold></th>
<th valign="top" align="left"><bold>Context window</bold></th>
<th valign="top" align="left"><bold>Reference</bold></th>
<th valign="top" align="left"><bold>Notes</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">AI21 Labs</td>
<td valign="top" align="left">Jamba</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">52B</td>
<td valign="top" align="left">256k</td>
<td valign="top" align="left">Lieber et al., <xref ref-type="bibr" rid="B119">2024</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Allen Institute for AI</td>
<td valign="top" align="left">OLMo</td>
<td valign="top" align="left">February 2024</td>
<td valign="top" align="left">7B</td>
<td valign="top" align="left">2048</td>
<td valign="top" align="left">Groeneveld et al., <xref ref-type="bibr" rid="B74">2024</xref></td>
<td valign="top" align="left">Open-source access to model, weights, and training data</td>
</tr> <tr>
<td valign="top" align="left">Anthropic</td>
<td valign="top" align="left">Claude 2</td>
<td valign="top" align="left">July 2023</td>
<td valign="top" align="left">&#x0003E;130B</td>
<td valign="top" align="left">100k</td>
<td valign="top" align="left">Anthropic, <xref ref-type="bibr" rid="B8">2023a</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Anthropic</td>
<td valign="top" align="left">Claude 2.1</td>
<td valign="top" align="left">November 2023</td>
<td valign="top" align="left">&#x0003E;130B</td>
<td valign="top" align="left">200k</td>
<td valign="top" align="left">Anthropic, <xref ref-type="bibr" rid="B9">2023b</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Anthropic</td>
<td valign="top" align="left">Claude 3</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">3 different model sizes: Haiku (20B), Sonnet (70B), and Opus (2T)</td>
<td valign="top" align="left">200k to 1 million</td>
<td valign="top" align="left">Anthropic, <xref ref-type="bibr" rid="B10">2024</xref></td>
<td valign="top" align="left">Multimodal: text and image input</td>
</tr> <tr>
<td valign="top" align="left">Apple</td>
<td valign="top" align="left">MM1</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">Up to 30B</td>
<td valign="top" align="left">McKinzie et al., <xref ref-type="bibr" rid="B133">2024</xref></td>
<td valign="top" align="left">Multimodal: text and image input</td>
</tr> <tr>
<td valign="top" align="left">Baidu</td>
<td valign="top" align="left">Ernie 4.0</td>
<td valign="top" align="left">October 2023</td>
<td valign="top" align="left">4T (est.)</td>
<td valign="top" align="left">1024</td>
<td valign="top" align="left">Mo and Baptista, <xref ref-type="bibr" rid="B147">2023</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Cohere</td>
<td valign="top" align="left">Command-medium</td>
<td valign="top" align="left">December 2022</td>
<td valign="top" align="left">6B</td>
<td valign="top" align="left">1024</td>
<td valign="top" align="left">-</td>
<td/>
</tr> <tr>
<td valign="top" align="left">Cohere</td>
<td valign="top" align="left">Command-xlarge</td>
<td valign="top" align="left">December 2022</td>
<td valign="top" align="left">50B</td>
<td valign="top" align="left">4096</td>
<td valign="top" align="left">-</td>
<td/>
</tr> <tr>
<td valign="top" align="left">Databricks</td>
<td valign="top" align="left">DBRX</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">132B</td>
<td valign="top" align="left">32k</td>
<td valign="top" align="left">Mosaic AI Research Team, <xref ref-type="bibr" rid="B150">2024</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Google</td>
<td valign="top" align="left">Gemini Pro 1.5</td>
<td valign="top" align="left">February 2024</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">128k - 1 million</td>
<td valign="top" align="left">Pichai and Hassabis, <xref ref-type="bibr" rid="B173">2024</xref></td>
<td valign="top" align="left">Multimodal: text, image and video input</td>
</tr> <tr>
<td valign="top" align="left">Google</td>
<td valign="top" align="left">Gemma</td>
<td valign="top" align="left">February 2024</td>
<td valign="top" align="left">2B, 7B</td>
<td valign="top" align="left">8192</td>
<td valign="top" align="left">Banks and Warkentin, <xref ref-type="bibr" rid="B15">2024</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Google</td>
<td valign="top" align="left">LaMDA 2</td>
<td valign="top" align="left">May 2022</td>
<td valign="top" align="left">540B</td>
<td valign="top" align="left">1024</td>
<td valign="top" align="left">Ghahramani, <xref ref-type="bibr" rid="B67">2022</xref></td>
<td valign="top" align="left">Both text and images as input</td>
</tr> <tr>
<td valign="top" align="left">Google</td>
<td valign="top" align="left">PaLM 2</td>
<td valign="top" align="left">May 2023</td>
<td valign="top" align="left">340B</td>
<td valign="top" align="left">8192 (text-bison)</td>
<td valign="top" align="left">Anil et al., <xref ref-type="bibr" rid="B7">2023</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Meta</td>
<td valign="top" align="left">Llama</td>
<td valign="top" align="left">February 2023</td>
<td valign="top" align="left">7B, 13B, 33B, 65B</td>
<td valign="top" align="left">2048</td>
<td valign="top" align="left">Touvron et al., <xref ref-type="bibr" rid="B199">2023</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Meta</td>
<td valign="top" align="left">Llama 2</td>
<td valign="top" align="left">July 2023</td>
<td valign="top" align="left">7B, 13B, 70B</td>
<td valign="top" align="left">4096</td>
<td valign="top" align="left">Meta, <xref ref-type="bibr" rid="B137">2023b</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Meta</td>
<td valign="top" align="left">Llama 3</td>
<td valign="top" align="left">April 2024</td>
<td valign="top" align="left">8B, 70B, 400B</td>
<td valign="top" align="left">8192</td>
<td valign="top" align="left">Meta, <xref ref-type="bibr" rid="B138">2024</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Microsoft</td>
<td valign="top" align="left">Orca-2</td>
<td valign="top" align="left">November 2023</td>
<td valign="top" align="left">7B, 13B</td>
<td valign="top" align="left">2048</td>
<td valign="top" align="left">Mitra et al., <xref ref-type="bibr" rid="B146">2023</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Microsoft</td>
<td valign="top" align="left">Phi-2</td>
<td valign="top" align="left">November 2023</td>
<td valign="top" align="left">2.7B</td>
<td valign="top" align="left">1024</td>
<td valign="top" align="left">Javaheripi and Bubeck, <xref ref-type="bibr" rid="B92">2023</xref></td>
<td valign="top" align="left">Small Language Model</td>
</tr> <tr>
<td valign="top" align="left">Mistral</td>
<td valign="top" align="left">Small, Large</td>
<td valign="top" align="left">February 2024</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">32k</td>
<td valign="top" align="left">Mistral AI, <xref ref-type="bibr" rid="B143">2024</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">Mistral</td>
<td valign="top" align="left">Mistral 7B</td>
<td valign="top" align="left">September 2023</td>
<td valign="top" align="left">7B</td>
<td valign="top" align="left">4096</td>
<td valign="top" align="left">Mistral AI, <xref ref-type="bibr" rid="B141">2023a</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">Mistral</td>
<td valign="top" align="left">Mixtral 8x7B</td>
<td valign="top" align="left">December 2023</td>
<td valign="top" align="left">56B</td>
<td valign="top" align="left">32k</td>
<td valign="top" align="left">Mistral AI, <xref ref-type="bibr" rid="B142">2023b</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">OpenAI</td>
<td valign="top" align="left">ChatGPT</td>
<td valign="top" align="left">November 2022</td>
<td valign="top" align="left">175B</td>
<td valign="top" align="left">4096</td>
<td valign="top" align="left">OpenAI, <xref ref-type="bibr" rid="B163">2022</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">OpenAI</td>
<td valign="top" align="left">GPT-4</td>
<td valign="top" align="left">March 2023</td>
<td valign="top" align="left">1.7T (est.)</td>
<td valign="top" align="left">8192</td>
<td valign="top" align="left">OpenAI, <xref ref-type="bibr" rid="B164">2023</xref></td>
<td/>
</tr> <tr>
<td valign="top" align="left">OpenAI</td>
<td valign="top" align="left">GPT-4 Turbo</td>
<td valign="top" align="left">October 2023</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">128k</td>
<td/>
<td/>
</tr> <tr>
<td valign="top" align="left">Technology Innovation Institute</td>
<td valign="top" align="left">Falcon</td>
<td valign="top" align="left">June 2023</td>
<td valign="top" align="left">1.3B, 7.5B, 40B, 180B</td>
<td valign="top" align="left">2048</td>
<td valign="top" align="left">von Werra et al., <xref ref-type="bibr" rid="B209">2023</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">xAI</td>
<td valign="top" align="left">Grok 1</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">314B</td>
<td valign="top" align="left">8192</td>
<td valign="top" align="left">xAI, <xref ref-type="bibr" rid="B217">2024a</xref></td>
<td valign="top" align="left">Open-source</td>
</tr> <tr>
<td valign="top" align="left">xAI</td>
<td valign="top" align="left">Grok-1.5</td>
<td valign="top" align="left">March 2024</td>
<td valign="top" align="left">-</td>
<td valign="top" align="left">128k</td>
<td valign="top" align="left">xAI, <xref ref-type="bibr" rid="B218">2024b</xref></td>
<td/>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Parameters refers to the number of parameters or weights in the model (B, billion; T, trillion). In many cases the exact parameter count is not known and estimates (est.) from the literature or blogs are given instead.</p>
</table-wrap-foot>
</table-wrap>
<p>Parameter count is correlated with the learning, generalization, and language understanding capabilities and hence a measure of the model&#x00027;s capacity and capabilities. At the same time, it is associated with increased computational demands. A separate metric of the capability of a LLM is the size of the context window. It is typically measured in the number of tokens. Roughly speaking, this is the amount of information (context) in a session that the model can &#x0201C;remember&#x0201D; or refer to. Most LLMs have a context window of a few thousands tokens, but Anthropic&#x00027;s Claude 2 boasts a large context window of 100,000 tokens (around 75,000 words). This means that it can hold entire papers and books in memory and the user can ask the model detailed questions about it. Number of parameters and context window size have not been publicly released in many cases. We collected estimates from the literature and blogs to the best of our knowledge. The models also differ in the type of input data they can receive. For instance, GPT-4 can receive not only text but also images as input and the prompts can be used to ask questions about the image (OpenAI, <xref ref-type="bibr" rid="B164">2023</xref>).</p>
<p>Some of the aforementioned models have been used as starting points for more specialized models. For instance, Med-PaLM is a specialized model based on PaLM 2 (Gupta and Waldron, <xref ref-type="bibr" rid="B75">2023</xref>). It is designed to assist in medical decision-making by providing accurate and relevant information based on a wide array of medical literature and data. Furthermore, after Meta released the weights for their Llama model, a number of finetuned models based on Llama have been released, such as Vicuna (<ext-link ext-link-type="uri" xlink:href="https://lmsys.org/blog/2023-03-30-vicuna/">https://lmsys.org/blog/2023-03-30-vicuna/</ext-link>), and Alpaca (<ext-link ext-link-type="uri" xlink:href="https://crfm.stanford.edu/2023/03/13/alpaca.html">https://crfm.stanford.edu/2023/03/13/alpaca.html</ext-link>). Although the overall industry trend has been toward larger, more capable, and multi-modal models, there has been a simultaneous effort to develop Small Language Models (SLMs) such as Phi-2 by Microsoft. The goal of the latter is to obtain models that are highly capable yet deployable on consumer devices such as smartphones.</p>
</sec>
</sec>
<sec>
<title>Large Language Models for dementia</title>
<p>In this section, we elucidate the role that LLMs can play in the research, diagnosis, treatment and management of dementia. LLMs are envisioned to be used by <bold>people with dementia (PwD)</bold> and/or their caregivers in the form of apps running on a mobile device, tablet, laptop, or desktop computer. Finally, we will introduce a questionnaire that was presented to PwD. In the questionnaire we asked participants about their experience with LLMs, their assessment of several scenarios for using LLM-powered apps for dementia care and management as well as its desired features and functionalities.</p>
<sec>
<title>Applications in clinical assessment and research</title>
<p>LLMs can be used as tools for dementia research, for instance as models of dementia (Li et al., <xref ref-type="bibr" rid="B115">2022</xref>; Demszky et al., <xref ref-type="bibr" rid="B41">2023</xref>; Loconte et al., <xref ref-type="bibr" rid="B126">2023</xref>) or diagnostic tools (Agbavor and Liang, <xref ref-type="bibr" rid="B3">2022</xref>; de Arriba-P&#x000E9;rez et al., <xref ref-type="bibr" rid="B38">2023</xref>; Wang et al., <xref ref-type="bibr" rid="B211">2023</xref>). The usage of LLMs by psychiatrists, healthcare professionals and data scientists has been covered in other reviews (Bzdok et al., <xref ref-type="bibr" rid="B25">2024</xref>; Tian et al., <xref ref-type="bibr" rid="B197">2024</xref>).</p>
<sec>
<title>Clinical record summarization</title>
<p>LLMs have the potential to help psychiatrists and other healthcare professionals with routine tasks such as writing of clinical reports, saving time and reducing manual data management (Cheng et al., <xref ref-type="bibr" rid="B29">2023</xref>; Javaid et al., <xref ref-type="bibr" rid="B93">2023</xref>). They have been used to provide summaries of patient-doctor conversations (Zhang et al., <xref ref-type="bibr" rid="B222">2021</xref>), clinical notes (Kanwal and Rizzo, <xref ref-type="bibr" rid="B100">2022</xref>) and reports (Vinod et al., <xref ref-type="bibr" rid="B208">2020</xref>), as well as coding adverse events in patient narratives (Chopard et al., <xref ref-type="bibr" rid="B30">2021</xref>). Furthermore, although off-the-shelf LLMs lack the sophistication required to answer queries of medical experts, finetuned models such as PMC-Llama (Wu et al., <xref ref-type="bibr" rid="B215">2023</xref>) and Med-PaLM (Singhal et al., <xref ref-type="bibr" rid="B192">2023</xref>) show increased expertise. In line with this, Lehman et al. (<xref ref-type="bibr" rid="B114">2023</xref>) showed that models trained or finetuned on clinical records outperform models that are not finetuned or that rely on in-context learning. In safety-critical domains such as medicine, the accuracy of the summary is of utmost importance. In this regard, Van Veen et al. (<xref ref-type="bibr" rid="B203">2024</xref>) performed an experiment with physicians showing that they preferred LLM-based summaries over summaries produced by human experts across a variety of domains (radiology reports, patient questions, progress notes, and doctor-patient dialogue).</p>
</sec>
<sec>
<title>Dementia prediction</title>
<p>Prediction of dementia using artificial intelligence with various biomarkers is well researched. First, one branch of researchers focused on neuroimaging data, using structural Magnetic Resonance Imaging (MRI) for predicting accelerated brain aging (Baecker et al., <xref ref-type="bibr" rid="B13">2021</xref>; Treder et al., <xref ref-type="bibr" rid="B200">2021</xref>), functional MRI (Du et al., <xref ref-type="bibr" rid="B52">2018</xref>), electroencephalography (Jiao et al., <xref ref-type="bibr" rid="B97">2023</xref>), or a fusion of different modalities (Abrol et al., <xref ref-type="bibr" rid="B2">2019</xref>). Second, clinical summaries have been used with LLMs to make differential diagnoses (Koga et al., <xref ref-type="bibr" rid="B107">2024</xref>). Mao et al. (<xref ref-type="bibr" rid="B130">2023</xref>) showed that a language model can use clinical notes to successfully predict the transition from mild cognitive impairment to Alzheimer&#x00027;s disease. Third, diagnostic markers can be extracted from patients&#x00027; speech, either directly from acoustic signals or from the transcribed text. A number of approaches showed a high predictive accuracy using acoustic features such as number of pauses and speech rate (Toth et al., <xref ref-type="bibr" rid="B198">2018</xref>; Al-Hameed et al., <xref ref-type="bibr" rid="B5">2019</xref>; O&#x00027;Malley et al., <xref ref-type="bibr" rid="B161">2020</xref>). Bang et al. (<xref ref-type="bibr" rid="B14">2024</xref>) used a combination of speech, text, and fluency opinions and reported an accuracy up to 87% for discriminating between Alzheimer&#x00027;s patients and healthy controls. In a different approach by Bouazizi et al. (<xref ref-type="bibr" rid="B22">2024</xref>), center of focus changes of participants when describing an image were predictive of dementia. Agbavor and Liang (<xref ref-type="bibr" rid="B3">2022</xref>) used GPT-3 to extract text embeddings that were then used as features to distinguish Alzheimer&#x00027;s patients from healthy controls. Better results were obtained for text features than for acoustic features using the speech signal directly. This suggests that text, although lacking information such as intonation, pauses, rate, and rhythm, might contain enough information to enable dementia prediction. Lastly, as a complementary application to prediction, LLMs are also able to generate synthetic data that can counteract the scarcity and imbalance of curated medical data and thereby aid in the training of prediction models (Li et al., <xref ref-type="bibr" rid="B116">2023</xref>).</p>
</sec>
</sec>
<sec>
<title>Applications in dementia management and care</title>
<p>In this section we introduce several scenarios for how LLM-powered apps could be used in the management of dementia, either by people with dementia themselves and/or their supporters. <xref ref-type="fig" rid="F3">Figure 3</xref> depicts an overview over the scenarios.</p>
<fig id="F3" position="float">
<label>Figure 3</label>
<caption><p>Possible applications of LLMs in dementia management and care.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0003.tif"/>
</fig>
<sec>
<title>Companionship</title>
<p>LLMs are able to participate in conversations about daily or private matters, questions and concerns. When tuned to respond adequately (e.g., displaying understanding and empathy) we hypothesize that an app could provide additional companionship and emotional support, especially in situations wherein PwD are socially isolated. Feeling of loneliness has been associated with a higher risk for developing dementia later in life (Holwerda et al., <xref ref-type="bibr" rid="B81">2014</xref>), although the literature is inconclusive on whether this relationship is causal (Victor, <xref ref-type="bibr" rid="B206">2021</xref>). There is evidence that apps in general can help reduce loneliness and isolation in dementia (Rai et al., <xref ref-type="bibr" rid="B181">2022</xref>). The apps reported in Rai et al. (<xref ref-type="bibr" rid="B181">2022</xref>) were aimed toward communication and social connections, improving engagement and physical activity through multi-sensory stimulation, remote monitoring and support, and assistive functions. Some studies reported positive results on digital pets and humanoid social robots for combating loneliness and social isolation in dementia (Gustafsson et al., <xref ref-type="bibr" rid="B76">2015</xref>; Demiris et al., <xref ref-type="bibr" rid="B40">2017</xref>; D&#x00027;Onofrio et al., <xref ref-type="bibr" rid="B50">2019</xref>; Fields et al., <xref ref-type="bibr" rid="B56">2021</xref>; Lima et al., <xref ref-type="bibr" rid="B120">2022</xref>). In a field study with 25 participants from an elderly home, Ryu et al. (<xref ref-type="bibr" rid="B187">2020</xref>) found significant decreases in anxiety and depression after daily use of a conversational chatbot for free conversations. Qi and Wu (<xref ref-type="bibr" rid="B177">2023</xref>) highlight the potential benefits of ChatGPT in terms of loneliness, emotional support, and assisting with daily tasks including reminders, medications, and appointments. This nicely dovetails with the assessment of healthcare professionals who report merit in virtual assistants and companions (Koebel et al., <xref ref-type="bibr" rid="B106">2022</xref>). In summary, we believe that LLMs hold potential as a companion and serve as an antidote to loneliness and social isolation associated with dementia. As LLMs mature technologically, it is possible to have increasingly meaningful and deep conversations with them. Although it is unlikely and perhaps undesirable that they can fully replace conversations between humans, they can complement and enhance human interaction, especially when carers are not accessible 24/7. Such social and conversational LLMs can come in the shape of apps, as potentially voice enacted chat applications. More immersive social interactions might be possible when the LLMs are digitally embodied as virtual avatars (Morales-de-Jes&#x000FA;s et al., <xref ref-type="bibr" rid="B149">2021</xref>) or even physically embodied as robots (Lima et al., <xref ref-type="bibr" rid="B120">2022</xref>).</p>
</sec>
<sec>
<title>Information retrieval</title>
<p>LLMs can serve as reservoirs of knowledge. Although this is one of their more basic applications, it can be useful for PwD. Unlike conventional search engines that merely retrieve websites, LLMs excel in identifying, compiling and re-synthesizing knowledge and presenting it in an accessible and understandable form. Saeidnia et al. (<xref ref-type="bibr" rid="B188">2023</xref>) reported dementia caregivers were overall positive about the quality of answers given by ChatGPT to queries about non-clinical issues relevant to PwDs&#x00027; lives. However, for questions related to dementia, LLMs may not be sufficiently accurate out of the box. For instance, Hristidis et al. (<xref ref-type="bibr" rid="B83">2023</xref>) compared ChatGPT with Google search for questions specifically related to dementia and cognitive decline with subpar quality for both systems. In line with this, ChatGPT&#x00027;s knowledge of dementia has been designated as &#x0201C;accurate but shallow&#x0201D; (Dosso et al., <xref ref-type="bibr" rid="B51">2023</xref>). This can potentially be alleviated by finetuning LLMs on medical data. For instance, PMC-Llama is a model based on Llama that has been finetuned using medical journal papers and textbooks (Wu et al., <xref ref-type="bibr" rid="B215">2023</xref>). Similarly, Google released Med-PaLM, a version of their PaLM specifically geared toward answering medical questions (Singhal et al., <xref ref-type="bibr" rid="B192">2023</xref>). Additionally, one can envision that LLMs could be finetuned to adapt their style to the user via prompt engineering. By default, models such as ChatGPT have a verbose and rather academic writing style. In summary, we believe that LLMs can be useful for the collation and reformulation of generic information as well as information specifically related to dementia. In the latter case, finetuned models such as Med-PaLM will likely be required. Furthermore, care needs to be taken to avoid blurring the line between a conversational service and medical advice, since at least for the time being healthcare professionals should be the ultimate source of medical advice.</p>
</sec>
<sec>
<title>Therapy aid</title>
<p>As alluded to in the previous paragraphs, LLMs can provide companionship and combat loneliness and social isolation. However, can it be used by therapists and healthcare professionals to aid during therapy? A review of previous-generation language models reported promising potential for use in mental health (Vaidyam et al., <xref ref-type="bibr" rid="B202">2019</xref>). Despite limited data on its clinical efficacy, users dealing with mental health problems have been consulting ChatGPT (Eliot, <xref ref-type="bibr" rid="B53">2023</xref>). Some studies investigated language models in the context of reminiscence therapy which involves engaging patients in recalling and discussing past experiences, often using tangible prompts like photographs or familiar objects (Khan et al., <xref ref-type="bibr" rid="B104">2022</xref>). Reminiscence therapy can enhance emotional wellbeing and cognitive function, as it encourages communication and the recollection of personal histories. Car&#x000F3;s et al. (<xref ref-type="bibr" rid="B26">2020</xref>) built Elizabot, a language model that mimics a reminiscence therapist. It consists of two components, a model that analyzes and captions the images used in the therapy, and a model for simple conversations. The authors received positive feedback from PwD trialing its use. Similarly, Morales-de-Jes&#x000FA;s et al. (<xref ref-type="bibr" rid="B149">2021</xref>) implemented an automated reminiscence model. It was integrated within a speech-enacted virtual avatar and people with Alzheimer&#x00027;s disease trialing the system gave it an overall score of 4.18/5, indicating high levels of satisfaction. It is worth stressing that both studies did not use state of the art models such as GPT-4. State of the art models are likely to have higher image captioning and conversation abilities, with potentially positive knock-on effects in the quality of reminiscence therapy. In line with this, Raile (<xref ref-type="bibr" rid="B182">2024</xref>) highlighted ChatGPT&#x00027;s usefulness both for complementing psychotherapy and as a first stop for people with mental health problems who have not sought help yet, though concerns remain regarding biases and one-sided information. Furthermore, cognitive behavioral therapy has shown promising results in treating anxiety and depression in dementia (Tay et al., <xref ref-type="bibr" rid="B196">2019</xref>). LLMs can potentially help administer cognitive behavioral therapy via phone apps (Denecke et al., <xref ref-type="bibr" rid="B42">2022</xref>) or in the shape of conversational chatbots (Patel et al., <xref ref-type="bibr" rid="B170">2019</xref>; Omarov et al., <xref ref-type="bibr" rid="B162">2023</xref>). In an analysis of social media posts on an LLM-powered mental health app (not specifically aimed toward PwD), Ma et al. (<xref ref-type="bibr" rid="B129">2024</xref>) reported on-demand and non-judgmental support, the development confidence and self-discovery as the App&#x00027;s benefits. In summary, we believe that LLMs can serve as therapy assistants to healthcare professionals. They either affect the therapeutic quality either indirectly by reducing the work burden of a healthcare professional, or directly by engaging in an intervention such as reminiscence therapy.</p>
</sec>
<sec>
<title>Reading and writing</title>
<p>A useful but easily overlooked feature of LLMs is that they can comprehend complex text and paraphrase it in more palatable or adequate language, e.g., rephrasing a formal text using more casual language. This is a relevant functionality since PwD are more likely than healthy controls to suffer from reading and writing deficits and speech pathologies (Murdoch et al., <xref ref-type="bibr" rid="B151">1987</xref>; Krein et al., <xref ref-type="bibr" rid="B109">2019</xref>). Consequently, LLMs could help in the interpretation and comprehension of letters, or emails, manuals, especially when being verbose or using convoluted language. Similarly, LLMs can assist in the formulation of letters and emails. We are not aware of specific studies on dementia in this regard, but LLMs have been explored for clinical text summarization (Van Veen et al., <xref ref-type="bibr" rid="B204">2023</xref>; Tian et al., <xref ref-type="bibr" rid="B197">2024</xref>) and the summarization of fiction books (Wu et al., <xref ref-type="bibr" rid="B216">2021</xref>). Furthermore, LLMs are increasingly being used as co-pilots in the writing of scientific articles (Altm&#x000E4;e et al., <xref ref-type="bibr" rid="B6">2023</xref>; Lingard, <xref ref-type="bibr" rid="B122">2023</xref>; Park, <xref ref-type="bibr" rid="B168">2023</xref>), including the present one, as well as liberal arts (Oh, <xref ref-type="bibr" rid="B159">2023</xref>) and business writing (AlAfnan et al., <xref ref-type="bibr" rid="B4">2023</xref>). We are not aware of specific studies on dementia for writing, but language models have been explored as email writing assistants for adults with dyslexia (Goodman et al., <xref ref-type="bibr" rid="B71">2022</xref>; Botchu et al., <xref ref-type="bibr" rid="B21">2023</xref>). In summary, LLMs as reading and writing aids for dementia have not been explored sufficiently, hence more research is required to evaluate their utility in this area.</p>
</sec>
<sec>
<title>Navigation</title>
<p>Several types of dementia, including Alzheimer&#x00027;s disease and dementia with Lewy bodies, can affect visual cognition and navigational abilities to varying extents (Pl&#x000E1;cido et al., <xref ref-type="bibr" rid="B175">2022</xref>). Spatial navigation aids for people with dementia in forms of digital apps and devices have been explored for years (Kowe et al., <xref ref-type="bibr" rid="B108">2023</xref>; Pillette et al., <xref ref-type="bibr" rid="B174">2023</xref>). Navigation aid can be useful both for outdoor navigation, e.g., finding your way from the home to a destination, and indoor navigation, e.g., finding the way around a hospital or other large building (Garc&#x000ED;a-Requejo et al., <xref ref-type="bibr" rid="B66">2023</xref>). Tech companies such as Google aim to integrate conversational services into a wide variety of apps (Wang and Li, <xref ref-type="bibr" rid="B210">2023</xref>). This opens the door for language and speech-assisted navigation, where the user converses with the navigation system and can ask for clarification and guidance. Currently, we are not aware of any such systems specifically developed for dementia patients. Further technological development and research on the academic and clinical side are required to assess how LLMs can aid navigation in these populations.</p>
</sec>
</sec>
<sec>
<title>Technical and design considerations</title>
<p>The implementation of LLM-powered apps for dementia involves a number of technical considerations as well as design challenges related to dementia:</p>
<list list-type="bullet">
<list-item><p><bold>Neurodiversity and cognitive load</bold>. Cognitive impairment associated with dementia can limit how much PwD can benefit from apps that place high demands on cognition (Hugo and Ganguli, <xref ref-type="bibr" rid="B88">2014</xref>). Therefore, the design of supportive apps for dementia patients should account for potential cognitive deficits faced by this population by minimizing cognitive load.</p></list-item>
<list-item><p><bold>Mobile phone use</bold>. The prime outlet for digital apps is mobile phones. Dixon et al. (<xref ref-type="bibr" rid="B48">2022</xref>) used semi-structured interviews to investigate mobile phone usage in PwD. Widespread usage of mobile phones by PwD was reported for tasks such as social media, reminders, and navigation. However, challenges regarding the ease of use were reported, such as difficulty in navigating to the right App, operating the phone while stressed or fatigued, and dealing with changing interfaces after App updates. Users valued being able to customize the interface to their needs, being able to use them as personal assistants, and use avatars and voice interaction. In conclusion, users should not have to be tech savvy to use them, and they should be built with ease, stability and customizability in mind.</p></list-item>
<list-item><p><bold>Voice control</bold>. Dementia types can be associated with visual impairments (Kuzma et al., <xref ref-type="bibr" rid="B110">2021</xref>), above and beyond the visual impairments that naturally come with age. Voice control is desirable since it can ease the interaction with digital devices and remove the challenge of navigating through the apps on the screen. However, not all voice systems are sufficiently robust to impairments such as slowed speech or stutter which can be frustrating and stress-inducing (Dixon et al., <xref ref-type="bibr" rid="B48">2022</xref>). Furthermore, hearing impairments can challenge voice based interaction, pointing again at the importance of a system with personalized characteristics tailored to the user (Hardy et al., <xref ref-type="bibr" rid="B78">2016</xref>).</p></list-item>
<list-item><p><bold>Avatar</bold>. Some participants in the study by Dixon et al. (<xref ref-type="bibr" rid="B48">2022</xref>) were enthusiastic about using voice control in conjunction with an animated personalized avatar. The avatar could help with attentional focus.</p></list-item>
<list-item><p><bold>Cloud-based vs on-device</bold>. LLMs tend to be computationally demanding and it is usually not feasible to deploy them directly on consumer phones. Instead, a cloud-based solution can be utilized that relies on an internet connection. The cloud server then processes the input through the LLM, generates the results and sends these results back to be displayed on the phone. This is how LLMs such as ChatGPT (OpenAI, <xref ref-type="bibr" rid="B163">2022</xref>) are typically integrated into smartphone apps. The advantage of this Approach is that no compute resources are needed on the device. The disadvantage is that an internet connection is required to operate the App, there can be additional delays due to transmission delays between the phone and the server. Additionally, there are potential security risks such as prompt injection and privacy risks due to communication with the server. There has been some effort to develop Small Language Models (SLMs), such as Microsoft&#x00027;s Phi-2 (Javaheripi and Bubeck, <xref ref-type="bibr" rid="B92">2023</xref>), which can be directly deployed on the phone. While phone-hosted LLMs offer enhanced security and privacy, by operating independently of internet connectivity, current technical constraints around model size, battery consumption, cooling and maintaining strong capabilities present trade-offs versus cloud-processed LLM solutions.</p></list-item>
<list-item><p><bold>Conversational style</bold>. In addition to the content of a conversation, the style in which an LLM interacts with the user is relevant to the overall experience. For instance, ChatGPT can be verbose and academic sounding, which could make comprehension difficult for many dementia patients. Models such as ChatGPT are able to adapt conversational style via prompt engineering, so style adaptation is a design challenge rather than a technical challenge.</p></list-item>
<list-item><p><bold>Anthropomorphisation</bold>. As LLMs capabilities increase, users are more likely to ascribe personality and agency to them. This can facilitate building an emotional bond with the App, offering potential benefits such as increased engagement, but also risks such as overreliance on recommendations. Evidence for this was given by Ma et al. (<xref ref-type="bibr" rid="B129">2024</xref>) who reported in an analysis of social media data that some users of an LLM-powered mental health App experienced feelings of stress, loss and grievance after updates to the LLM lead to inconsistent conversational style and the loss of memory of previous conversations. While these results were obtained with a chat application, LLMs personified as virtual avatars with their own voice and looks might increase anthropomorphisation even more.</p></list-item>
</list>
<p>Concluding, the diversity and individual variability of challenges faced by dementia patients makes it unlikely that a single technical solution can cater to the entire user base. A solution that claims wide applicability needs to be personalizable and adaptive. Personalization can involve visual elements (e.g., size, color, or style and choice of a virtual avatar), auditory aspects (speed and information content of auditory feedback and voice choices for voice assistants), as well as cognitive load (e.g., complexity of the usage, number of elements on a dashboard, ease of navigation) and conversational style. It is evident that the development of solutions should be accompanied by involvement and engagement of PwD and their caregivers/supporters. Their feedback should be sought from the initial design stage throughout the entire product development cycle is essential for creating effective and user-centric solutions.</p>
</sec>
</sec>
<sec>
<title>Questionnaire</title>
<p>We believe that an effective and ethical path toward the usage of LLMs in dementia management and care involves centering the perspectives and needs of people with dementia, caregivers and other stakeholders at all stages in the research and development cycle. For this reason, we created a questionnaire in which we asked participants to rate the usefulness of LLMs in a number of application scenarios (e.g., companionship, therapy aid), and we asked them to rate the importance of design features (e.g., ease of use, voice control, privacy). We presented the questionnaire to PwD, their supporters, caregivers and stakeholders. To the best of our knowledge, this is the first targeted survey on the usage of LLMs for dementia care and management. Ethical approval for the study has been obtained from The School of Computer Science and Informatics Research Ethics Committee at Cardiff University, United Kingdom, reference: COMSC/Ethics/2023/122.</p>
<sec>
<title>Participants</title>
<p>Fifteen people with dementia (PwD) aged 58-88 (&#x003BC;= 72.2), 7 women, 7 men, 1 of nonbinary gender, participated in the study. Additionally, 14 supporters aged 32-70 (&#x003BC;= 53.6), 11 women and 2 men (1 declined to indicate their sex), participated in the study. Supporters could be family members or professional caregivers or nurses. Participants were recruited with the help of Dementia Australia (<ext-link ext-link-type="uri" xlink:href="https://www.dementia.org.au/">https://www.dementia.org.au/</ext-link>) and Alzheimer&#x00027;s Society UK (<ext-link ext-link-type="uri" xlink:href="https://www.alzheimers.org.uk/">https://www.alzheimers.org.uk/</ext-link>). The organizations served as gatekeepers, that is, they published our invitation email and a participant information sheet on their website. The invitation email included a hyperlink that would take participants directly to the survey. There was no compensation for participation but participants could opt-in to a raffle for a single &#x000A3;100 Visa Gift card. To this end, they would enter their email address in the notes section of the questionnaire. After the raffle, the email addresses were removed from the dataset. The study was fully anonymous otherwise.</p>
</sec>
<sec>
<title>Questionnaire details</title>
<p>A copy of the questionnaire is provided as <xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>. Here, we summarize its main items. For the items, participants could choose to select &#x0201C;Prefer not to answer&#x0201D; if they wish not to answer a question. For multiple choice questions, an additional option &#x0201C;Other&#x0201D; was provided in case participants wanted to specify an option that was not listed. <xref ref-type="table" rid="T3">Table 3</xref> lists the questions used, categorizes them by section and specifies the type of answer required. Note that all items categorized as follow-up were only asked when the immediately preceding questions was answered with &#x0201C;yes&#x0201D;. Following questions about their demographic background and dementia, the main body consisted of questions regarding application scenarios of LLMs as well as desired features for digital apps. Finally, participants were asked to estimate the overall impact AI can have on dementia management and care, and there was space for free text with any notes or additions participants would like to make.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Overview over the questions used in the questionnaire.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Section</bold></th>
<th valign="top" align="center"><bold>ID</bold></th>
<th valign="top" align="left"><bold>Item</bold></th>
<th valign="top" align="left"><bold>Answer type</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.1</td>
<td valign="top" align="left">Age</td>
<td valign="top" align="left">Number</td>
</tr> <tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.2</td>
<td valign="top" align="left">Sex</td>
<td valign="top" align="left">Multiple choice (Male, Female, Other)</td>
</tr> <tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.3</td>
<td valign="top" align="left">Ethnic background</td>
<td valign="top" align="left">Multiple choice (Indigenous, Asian, European, African, Pacific Islander, Mixed background)</td>
</tr> <tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.4</td>
<td valign="top" align="left">What is the highest education level you achieved?</td>
<td valign="top" align="left">Multiple choice (Primary education, Secondary education, Vocational training, BSc, Postgraduate)</td>
</tr> <tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.5</td>
<td valign="top" align="left">Have you ever been diagnosed with dementia or Alzheimer&#x00027;s disease?</td>
<td valign="top" align="left">Multiple choice (Yes, No)</td>
</tr> <tr>
<td valign="top" align="left">Demographics</td>
<td valign="top" align="center">1.6</td>
<td valign="top" align="left">How many years ago have you been diagnosed with dementia?</td>
<td valign="top" align="left">Number</td>
</tr> <tr>
<td valign="top" align="left">Dementia (follow-up)</td>
<td valign="top" align="center">2.1</td>
<td valign="top" align="left">What specific type of dementia have you been diagnosed with?</td>
<td valign="top" align="left">Multiple choice (Alzheimer&#x00027;s, Lewy body dementia, Vascular dementia, Fronto-temporal dementia)</td>
</tr> <tr>
<td valign="top" align="left">Dementia (follow-up)</td>
<td valign="top" align="center">2.2</td>
<td valign="top" align="left">Could you describe any symptoms or experiences related to your diagnosis?</td>
<td valign="top" align="left">Free text</td>
</tr> <tr>
<td valign="top" align="left">AI experience</td>
<td valign="top" align="center">3.1</td>
<td valign="top" align="left">Have you ever used digital apps in the context of dementia management or treatment?</td>
<td valign="top" align="left">Multiple choice (Yes, No)</td>
</tr> <tr>
<td valign="top" align="left">AI experience</td>
<td valign="top" align="center">3.2</td>
<td valign="top" align="left">Before starting this questionnaire, had you heard of AI Language Models such as Chat-GPT?</td>
<td valign="top" align="left">Multiple choice (Yes, No)</td>
</tr> <tr>
<td valign="top" align="left">AI experience</td>
<td valign="top" align="center">3.3</td>
<td valign="top" align="left">Did you ever use AI Language Models such as Chat-GPT (for either personal or professional use)?</td>
<td valign="top" align="left">Multiple choice (Yes, No)</td>
</tr> <tr>
<td valign="top" align="left">AI experience (follow-up)</td>
<td valign="top" align="center">3.4</td>
<td valign="top" align="left">Please briefly describe how you used AI language models.</td>
<td valign="top" align="left">Free text</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.1</td>
<td valign="top" align="left"><bold>Companionship</bold>. Imagine your app includes a chat option. You can chat with the AI about daily or private matters, questions and concerns. Your conversation is confidential and will not be shared with others. To what extent could you consider such an App useful for yourself?</td>
<td valign="top" align="left">5-points Likert scale (Very useful, Useful, Moderately useful, Slightly useful, Not useful at all)</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.2</td>
<td valign="top" align="left"><bold>Dementia-related information</bold>. Imagine the AI is knowledgeable in the dementia literature. You can ask the AI questions about dementia and you can have a natural conversation in which it provides information about dementia diagnosis, care, treatment etc. <italic>However, it does not have access to your personal medical record, so it can only answer general questions</italic>.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.3</td>
<td valign="top" align="left"><bold>Dementia-related information including personal data</bold>. Imagine the AI is knowledgeable in the dementia literature. You can ask the AI questions about dementia and you can have a natural conversation in which it provides information about dementia diagnosis, care, treatment etc. <italic>The AI also has access to your medical data and it can provide answers tailored to your specific medical conditions</italic>.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.4</td>
<td valign="top" align="left"><bold>Navigation</bold>. Imagine the AI is connected to a navigation system (such as Google Maps or Apple Maps). It can give you directions in spoken language and can help you out if you lose your way.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.5</td>
<td valign="top" align="left"><bold>Reading aid</bold>. Imagine the AI can help you read letters and messages. You simply take a photo of the letter or copy the text into an App. The AI will explain in simple terms what the letter or message means. You can even ask questions about it.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.6</td>
<td valign="top" align="left"><bold>Writing aid</bold>. Imagine the AI can help you draft letters and messages. You simply give it an instruction such as &#x0201C;Write an email to my doctor asking to shift our appointment to next week&#x0201D; and it will give you a nicely written email draft.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.7</td>
<td valign="top" align="left"><bold>Therapy aid</bold>. Imagine the AI is able to carry conversation-based therapeutic interventions such as reminiscence therapy<sup>&#x0002A;</sup>.<break/><sup>&#x0002A;</sup>Reminiscence therapy involves discussing events and experiences from the past and aims to evoke memories, stimulate mental activity and improve a person&#x00027;s well-being. Reminiscence can often be supported by props such as videos, music, pictures and objects that may have particular meaning for an individual.</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Application scenarios</td>
<td valign="top" align="center">4.8</td>
<td valign="top" align="left">Do you have any comments regarding these application scenarios? Can you think of any other application scenarios not mentioned here? (feel free to skip this question if &#x00027;no&#x00027;)</td>
<td valign="top" align="left">Free text</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.1</td>
<td valign="top" align="left"><bold>Ease of use</bold>. How important is it that the app is intuitive and easy to use, without the need to go through tutorials or receive an introduction by a family member or caregiver?</td>
<td valign="top" align="left">5-points Likert scale (Very important, Important, Moderately important, Slightly important, Not important)</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.2</td>
<td valign="top" align="left"><bold>Voice control</bold>. How important is it that you can also use your voice to talk to the app and it talks back to you (as opposed to just typing text in a textbox)?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.3</td>
<td valign="top" align="left"><bold>Empathy</bold>. When having a conversation with the app, how important is it that the AI displays empathy, feelings, and understanding?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.4</td>
<td valign="top" align="left"><bold>Human in the loop</bold>. When using the app for therapeutic interventions, how important is it to use the Apptogether with in-person sessions with a caregiver or doctor, rather than just using the Appalone?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.5</td>
<td valign="top" align="left"><bold>Data privacy</bold>. How important is it that the app stores as little personal data as possible (e.g., age, gender, past conversations)?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.6</td>
<td valign="top" align="left"><bold>Data transparency</bold>. How important is it that the app is transparent and clear about which data it collects about you?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.7</td>
<td valign="top" align="left"><bold>Data deletion</bold>. How important is it that your personal data can be deleted from the app at any time?</td>
<td valign="top" align="left">5-points Likert scale</td>
</tr> <tr>
<td valign="top" align="left">Features and priorities</td>
<td valign="top" align="center">5.8</td>
<td valign="top" align="left"><bold>Device</bold>. When using the app, which device(s) do you prefer (select 1 or more)</td>
<td valign="top" align="left">Multiple choice (Smartphone, Tablet, Laptop or PC)</td>
</tr> <tr>
<td valign="top" align="left">Conclusion</td>
<td valign="top" align="center">6.1</td>
<td valign="top" align="left"><bold>Impact</bold>. What do you estimate the impact of AI on dementia management and care could be?</td>
<td valign="top" align="left">5-points Likert scale (Very positive, Positive, Neutral, Negative, Very negative)</td>
</tr> <tr>
<td valign="top" align="left">Conclusion</td>
<td valign="top" align="center">6.2</td>
<td valign="top" align="left"><bold>Comments</bold>. If you have any comments, thoughts or suggestions, you can share them with us here.</td>
<td valign="top" align="left">Free text</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The meaning of the columns is as follows. Section: Which section of the questionnaire the question belongs to. ID: identifier of the item that is used in the results section. Item: the verbatim question used in the questionnaire. Answer type: the type of answer that was required, i.e., number (participants entered a number with the keyboard), multiple choice (with the different options provided in brackets), free text (participants type a text as answer), 5-points Likert scale.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec>
<title>Procedure</title>
<p>The survey was implemented in Google Forms. It commenced by asking participants to provide informed consent in line with Cardiff University&#x00027;s guidelines. Participants were then asked to watch a 1-min overview over ChatGTP on YouTube (<ext-link ext-link-type="uri" xlink:href="https://www.youtube.com/watch?v=aIO9it4HFiQ">https://www.youtube.com/watch?v=aIO9it4HFiQ</ext-link>) to make sure that they are familiar with the basic principles of LLMs. Further videos and a blog post were presented as optional additional material. They then answered the questions listed in <xref ref-type="table" rid="T3">Table 3</xref> by either clicking on the multiple choice options or typing an answer. The survey took about 20 min.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<title>Results</title>
<p>The raw data and results of the questionnaire are available in our GitHub repository (<ext-link ext-link-type="uri" xlink:href="https://github.com/treder/LLMs-for-dementia">https://github.com/treder/LLMs-for-dementia</ext-link>). We review the results according to the sections in <xref ref-type="table" rid="T3">Table 3</xref>: Demographics, dementia, AI experience, application scenarios, and features and priorities.</p>
<sec>
<title>Demographics and dementia</title>
<p><xref ref-type="fig" rid="F4">Figure 4</xref> depicts the demographic details of the participants. People with dementia (PwD) participating in our study were aged 58&#x02013;88 years whereas supporters were aged 32&#x02013;70 years. As these ranges suggest, supporters were significantly younger than PwD (independent samples <italic>t</italic>-test, t = 5.059, <italic>p</italic> &#x0003C; 0.0001). Whereas gender roles were equally distributed for PwD (7 women, 7 men, 1 of nonbinary gender), supporters were predominantly female (11 women, 2 men). To compare the distribution of genders across the two groups, we used a two-sided Fisher&#x00027;s exact test which works on 2x2 contingency tables. The chi-squared test allows for larger tables but requires a larger sample size (Hazra and Gogtay, <xref ref-type="bibr" rid="B79">2016</xref>; Sundjaja et al., <xref ref-type="bibr" rid="B195">2024</xref>). Therefore, we focused on comparing the number of men and women. The difference was not statistically significant (odds ratio = 5.5, <italic>p</italic> = 0.1032) which might be attributed to the small sample size.</p>
<fig id="F4" position="float">
<label>Figure 4</label>
<caption><p>Demographic data of the participants. <bold>Top left</bold>: age distribution for people with dementia (PwD) and supporters. <bold>Top right</bold>: types of dementia. <bold>Bottom row</bold>: sex, ethnic background, and education for people with dementia and supporters.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0004.tif"/>
</fig>
<p>People with dementia identified as European (14) and Asian (1). Their highest degrees were vocational training/trade certificate (2 respondents), secondary education/high school (9), Bachelor&#x00027;s degree or equivalent (1), or Postgraduate degree (3). The supporters identified as European (9), Mixed background (2), or Asian (1), and 2 preferred not to answer. Supporters had vocational training/trade certificates (1 respondent), Bachelor&#x00027;s degree or equivalent (7), or postgraduate degree (5), and 1 preferred not to answer.</p>
<p>People with dementia received the diagnosis between 1 and 13 years (&#x003BC; &#x0003D; 5.1) ago. They were diagnosed with various types of dementia, namely Alzheimer&#x00027;s disease (9 respondents), Vascular dementia (1), Fronto-temporal dementia (1), Lewy body dementia (1), or Mixed Dementia (2), and 1 preferred not to answer. When asked to freely describe their symptoms, memory problems were mentioned most often, with 5 respondents mentioned problems with &#x0201C;short term memory&#x0201D;, and another one &#x0201C;total blank in the mornings&#x0201D;. Additional symptoms were related to social interaction (&#x0201C;withdrawn from people&#x0201D;, &#x0201C;unable to speak properly, difficulty understanding conversations&#x0201D;), physical symptoms (&#x0201C;tremors, gait and balance&#x0201D;, &#x0201C;difficult to balance on one side&#x0201D;, &#x0201C;shakes, unstable&#x0201D;), as well as &#x0201C;hallucinations, visual and auditory&#x0201D; and a general &#x0201C;inability to perform everyday tasks&#x0201D; and &#x0201C;inability to understand controls on oven or television&#x0201D;.</p>
</sec>
<sec>
<title>AI experience</title>
<p>Responses related to the use of apps in the context of dementia, 3/15 PwD and 5/14 supporters (1 preferred not to answer) responded with &#x0201C;Yes&#x0201D;. Six PwD and 12 supporters heard of LLMs such as Chat-GPT before. Two PwD (1 preferred not to answer) and 5 supporters (1 preferred not to answer) stated having used them before. One participant with dementia stated that they &#x0201C;use ChatGPT to gather information, links and quotes&#x0201D;. Supporters used them for &#x0201C;Patient and Public Involvement Networks, Universities, and as a carer for my Husband who had Dementia&#x0201D;, to &#x0201C;discover information on various topics encompassing dementia, including the types, symptoms and possible outcomes of therapies used in behavior management in dementia&#x0201D;, as well as to &#x0201C;synthesize text and videos&#x0201D; and for &#x0201C;writing reports&#x0201D;.</p>
</sec>
<sec>
<title>Application scenarios</title>
<p><xref ref-type="fig" rid="F5">Figure 5</xref> shows mean opinion scores obtained by encoding the response options as integers ranging from 1 to 5 and averaging them across individuals for the PwD and supporter groups separately. On average, all scenarios were ranked with moderate scores in between &#x0201C;Moderately useful&#x0201D; and &#x0201C;Useful&#x0201D; by both groups. Both PwD and supporters ranked &#x0201C;Navigation&#x0201D;, &#x0201C;Reading aid&#x0201D;, and &#x0201C;Writing aid&#x0201D; the highest. Somewhat lower scores were assigned to &#x0201C;Companionship&#x0201D; and the two items on &#x0201C;Dementia-related information&#x0201D;. As visual inspection suggests, responses between PwD and supporters were significantly correlated (Pearson correlation, <italic>r</italic> &#x0003D; 0.79, <italic>p</italic> = 0.033).</p>
<fig id="F5" position="float">
<label>Figure 5</label>
<caption><p>Opinion scores on Likert scale (y-axis) for each of the scenarios (x-axis). Scores have been averaged for individuals within the PwD and supporters groups. Markers depict mean, shaded area represents 1 standard error of the mean.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0005.tif"/>
</fig>
<p>A more detailed overview with the proportion of each response option by group is depicted in <xref ref-type="fig" rid="F6">Figure 6</xref>. We observe a dichotomy within the PwD group: for each scenario, at least one participant selected the response &#x0201C;Not useful at all&#x0201D; whereas several participants selected &#x0201C;Very useful&#x0201D;. To investigate whether individual response patterns are correlated with demographic variables, we performed a series of Pearson correlation analyses. We found that the overall mean score across all scenarios is negatively correlated with age for PwD (r = &#x02212;0.62, <italic>p</italic> = 0.014) but not with the number of years since the dementia diagnosis (<italic>p</italic> = 0.42). In other words, older participants tended to give lower overall scores. For supporters, there was no evidence for such a relationship (r = &#x02212;0.36, <italic>p</italic> = 0.2). When performing the same analysis on the score for each scenario separately, we found a significant relationship for &#x0201C;Companionship&#x0201D; (r = &#x02212;0.64, <italic>p</italic> = 0.001), &#x0201C;Dementia-related information&#x0201D; (r = &#x02212;0.54, <italic>p</italic> = 0.044), &#x0201C;Dementia-related information including personal data&#x0201D; (r = &#x02212;0.61, <italic>p</italic> = 0.027), &#x0201C;Reading aid&#x0201D; (r = &#x02212;0.61, <italic>p</italic> = 0.019), &#x0201C;Writing aid&#x0201D; (r = &#x02212;0.62, <italic>p</italic> = 0.002), although only &#x0201C;Companionship&#x0201D; and &#x0201C;Writing aid&#x0201D; would survive a correction for multiple comparisons. Correlations were not significant for &#x0201C;Navigation&#x0201D; (<italic>p</italic> = 0.24). No such relationships were found for supporters (all <italic>p</italic> &#x0003E; 0.16). For the PwD group, we repeated the correlation analysis using the number of years since the dementia diagnosis instead of age, but found no significant effects (all <italic>p</italic> &#x0003E; 0.31). For sex, we did not find a relationship with mean score for either group (all <italic>p</italic> &#x0003E; 0.53).</p>
<fig id="F6" position="float">
<label>Figure 6</label>
<caption><p>Questionnaire results for seven different application scenarios (see <xref ref-type="table" rid="T3">Table 3</xref>, IDs 4.1 through 4.7). Pie charts show how useful PwD and supporters consider LLMs in each application scenario. Each scenario is labeled in the figure. The corresponding pie chart is shown for PwD above the label and for supporters below. The numbers in the pie slices correspond to absolute and relative number of respondents (e.g., six respondents, 6/15 = 40%). The legend defines the meaning of the colors.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0006.tif"/>
</fig>
</sec>
<sec>
<title>Features and priorities</title>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> shows mean opinion scores obtained by encoding the response options as integers ranging from 1 to 5 and averaging them across individuals for the PwD and supporter groups separately. On average, all scenarios were ranked with moderate to high scores in between &#x0201C;Moderately important&#x0201D; and &#x0201C;Very important&#x0201D; by both groups. Both PwD and supporters ranked all priorities around data (&#x0201C;Data privacy&#x0201D;, &#x0201C;Data transparency&#x0201D;, &#x0201C;Data deletion&#x0201D;) the highest, showing concern for their agency over data. Mean scores between the two groups were highly correlated (Pearson correlation, <italic>r</italic> &#x0003D; 0.94, <italic>p</italic> = 0.001), showing a similar pattern of concerns and priorities.</p>
<fig id="F7" position="float">
<label>Figure 7</label>
<caption><p>Opinion scores on Likert scale (y-axis) for each of the features and priorities (x-axis). Scores have been averaged for individuals within the PwD and supporters groups. Markers depict mean, shaded area represents 1 standard error of the mean.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0007.tif"/>
</fig>
<p>A more detailed overview with the proportion of each response option by group is depicted in <xref ref-type="fig" rid="F8">Figure 8</xref>. Respondents in the PwD group gave either high or low scores to the items &#x0201C;Ease of use&#x0201D;, &#x0201C;Voice control&#x0201D;, &#x0201C;Empathy&#x0201D;, and &#x0201C;Human in the loop&#x0201D;, whereas supporters overwhelmingly gave high scores to these items. Overall mean score across all features and priorities was significantly correlated with age for PwD (r = &#x02212;0.54, <italic>p</italic> = 0.04) but not supporters (<italic>p</italic> = 0.5). For PwD, when performing the same analyses on the score for each feature and priority separately, we found no significant relationships (all <italic>p</italic>&#x0003E;0.05). There was no significant correlation with sex (PwD: <italic>p</italic> = 0.22, supporters: <italic>p</italic> = 0.17) and for PwD there was no correlation with the number of years since diagnosis (<italic>p</italic> = 0.33).</p>
<fig id="F8" position="float">
<label>Figure 8</label>
<caption><p>Questionnaire results for features and priorities (see <xref ref-type="table" rid="T3">Table 3</xref>, IDs 5.1&#x02013;5.7). Pie charts show how important PwD and supporters consider specific aspects of LLM-based apps for dementia. Each feature is labeled in the figure. The corresponding pie chart is shown for PwD above the label and for supporters below. The numbers in the pie slices correspond to absolute and relative number of respondents (e.g., 11 respondents, 11/15 = 73.3%). The legend defines the meaning of the colors. For data privacy and data transparency, an additional option was provided. The corresponding slice is depicted in white and the response option is pasted next to the figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0008.tif"/>
</fig>
<p><xref ref-type="fig" rid="F9">Figure 9</xref> depicts results on which devices respondents use and how they rate the overall impact of LLMs on dementia management and care. In both groups a variety of devices was used, although amongst PwD tablets were more dominant whereas among supporters smartphones were more dominant. The overall impact of LLMs on dementia care and management was seen more positively by PwD than supporters. Whereas only 1 respondent in the PwD group indicated &#x0201C;negative&#x0201D;, 3 supporters indicated the impact as &#x0201C;very negative&#x0201D;. Nevertheless, larger proportions in both groups rated the impact as &#x0201C;positive&#x0201D; or &#x0201C;very positive&#x0201D; (PwD: 9 respondents, supporters: 8).</p>
<fig id="F9" position="float">
<label>Figure 9</label>
<caption><p>Questionnaire results for device and impact (see <xref ref-type="table" rid="T3">Table 3</xref>, IDs 5.8 and 6.1). Different legends are provided for each question. For the preferred device, participants could select smartphone, Laptop/PC, tablet, or combinations of these.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frdem-03-1385303-g0009.tif"/>
</fig>
</sec>
<sec>
<title>Free comments</title>
<p>Respondents could also provide feedback in a free textual form (items 4.8 and 6.2 in <xref ref-type="table" rid="T3">Table 3</xref>). Both PwD and supporters provided feedback on positive and negative use cases for language-based AI applications in dementia. While some respondents were excited about the potential benefits of AI, others raised a number of concerns and caveats. The main points are summarized in <xref ref-type="table" rid="T4">Table 4</xref>.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Summary of the feedback of the respondents to the application scenarios.</p></caption>
<table frame="box" rules="all">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Summary</bold></th>
<th valign="top" align="left"><bold>Verbatim responses</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">The application scenarios are useful</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>Wow!! I would love anything like those above&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>A product like this would be amazing for me it would take a lot of stress out of my everyday life&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>I think AI has great implications for dementia awareness/care [...]&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>AI technology will be very important to alleviate isolation, and feelings of loneliness for those living with dementia who do not have family or friends nearby to engage with&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">There are other useful application scenarios beyond those mentioned</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>keeping fit and retaining muscle mass&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>ask medical questions&#x0201D;</italic><break/><sup>S</sup>&#x0201C;<italic>Protection from scams would be useful&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">The application scenarios do not address the actual problems of PwD</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>None of them relate to alleviating the problems of daily living [...]&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>What I really need is something that tells me step by step (all 176 of therm) how to live my daily life&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>I want raw, accurate information that I can easily verify, not a cozy chat. The whole idea of that side of AI is anathema to me&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>As a former carer of someone with Alzheimer&#x00027;s I can see very little use for this app except for navigating IF the person is out alone.&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>My cognitive faculties are relatively intact, it&#x00027;s my recall that is impaired&#x0201D;</italic><break/><sup>S</sup> &#x0201D;<italic>Tech is NOT the panacea that those who advocate it believe.&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">Concern about bias</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>dementia care [...] is infected with assumptions of ageism [...] and a host of other biases that are likely to show up in AI development&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">Concern about level of tech affinity required</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>This AI is not relevant in any way to my mum who has dementia aged 88yrs and has never been able to use a computer even before her diagnosis&#x0201D;</italic><break/><sup>D</sup> &#x0201C;<italic>This AI sounds like amazing progress but the actual demographic of most dementia sufferers is that they are over 70 so I am not too sure AI is going to help them a great deal as they will probably mostly not be computer literate! [...] I am 66 years old and find AI rather a challenge&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>They would find a PC, smartphone etc. very hard to navigate without help&#x0201C;</italic><break/><sup>S</sup> &#x0201C;<italic>it would not be suited to older people with no AI experience. People should start using the planned App as soon as possible so they are familiar with it even before a diagnosis&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">Usefulness depends on stage of dementia</td>
<td valign="top" align="left"><sup>D</sup> &#x0201C;<italic>this would be great for FTD and MCI&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>The ability to interact with the AI model depends on the degree of decline&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>might work well very early on in the disease&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>for people with early onset dementia, it could be a valuable tool&#x0201C;</italic><break/><sup>S</sup> &#x0201C;<italic>it just feels unsuitable, totally depends on the stage of dementia, at present it just feels like a gimmick&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">LLM may not understand the user</td>
<td valign="top" align="left"><sup>S</sup> &#x0201D;<italic>If relatives with a good knowledge of the person with other visual cues have difficulty in understanding, it is possible that AI will miss the point.&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">User may not understand the LLM</td>
<td valign="top" align="left"><sup>S</sup> &#x0201C;<italic>My clients if having a bad time being lost won&#x00027;t be able follow instructions to find their way home. Also talking about dementia, treatment, diagnosis I can see that leading to confusion, processing that much information, and it might be conflicting information to the client.&#x0201D;</italic></td>
</tr> <tr>
<td valign="top" align="left">LLM cannot replace human interaction and care</td>
<td valign="top" align="left"><sup>S</sup> &#x0201C;<italic>I know my loved one would not have been happy talking to a machine. It is not a replacement for a human [...] It&#x00027;s too untried to be let loose with those with a dementia diagnosis&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>[...] using the AI app without help from a carer/district nurse or family member would be very difficult for someone in the later stages of dementia&#x0201D;</italic><break/><sup>S</sup> &#x0201C;<italic>It is wrong in so many ways to use a machine to replicate a human response [...] This is very much along the lines of &#x0201C;babysitting by television&#x0201D;</italic></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The first column summarizes the respondents&#x00027; statements, the second column provides evidence in the shape of the actual individual feedback. The superscript <sup>D</sup> refers to respondents in the dementia group and <sup>S</sup> refers to supporters.</p>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>Discussion</title>
<p>Large Language Models (LLMs) revolutionize the way in which humans interact with machines. For the first time in history, we can converse with computers in the same way that we talk to each other. Meaningful conversations, creative writing, poetry, summarization, all deeply human faculties that can now be experienced in a chat with an algorithm. Our review has highlighted the burgeoning role of LLMs in improving dementia care and research. The integration of LLMs into therapeutic and support frameworks holds the potential to enhance the quality of life for individuals living with dementia, as well as to alleviate the considerable burden on caregivers. Through personalized conversations, information retrieval, therapy aid, and assistive technologies for reading, writing, and navigation, LLMs offer a novel approach to dementia care that is both innovative and human-centric. Nevertheless, its adoption might face an uphill battle due to algorithmic and regulatory limitations and challenges, as well as concerns about adequacy and applicability in the context of dementia care that surfaced in our survey.</p>
<sec>
<title>Limitations, risks, and challenges</title>
<p>Despite the promising prospects, the deployment of LLMs in dementia care is not without challenges. The current limitations of LLMs, including their dependency on the quality of input data and the potential for perpetuating biases, must be acknowledged and addressed. First, hallucinations, or the production of syntactically correct but factually incorrect text, plague all state of the art LLMs (Ye et al., <xref ref-type="bibr" rid="B221">2023</xref>; Zhang et al., <xref ref-type="bibr" rid="B224">2023</xref>) and are a source of concern for their medical application (Pal et al., <xref ref-type="bibr" rid="B167">2023</xref>; Tian et al., <xref ref-type="bibr" rid="B197">2024</xref>). Second, LLMs are trained on a large corpus of text from a variety of sources including social media websites, often without permission of the author of the text (Franceschelli and Musolesi, <xref ref-type="bibr" rid="B59">2022</xref>; Kasneci et al., <xref ref-type="bibr" rid="B103">2023</xref>). Since stigma against people with dementia has been reported on the media platform X (Oscar et al., <xref ref-type="bibr" rid="B165">2017</xref>; Bacsu et al., <xref ref-type="bibr" rid="B12">2022</xref>), inclusion of uncurated internet data into LLM training harbors the danger of perpetuating stereotypes about dementia. Third, overreliance might create adverse cognitive effects. LLMs assisting with perceptual tasks, memory, and language, creates short term benefits, but it is the same faculties that have to be engaged in order to combat decline (Mondini et al., <xref ref-type="bibr" rid="B148">2016</xref>; Hill et al., <xref ref-type="bibr" rid="B80">2017</xref>). Fourth, the development of LLMs for dementia has to take place with a regulatory framework that ensures that risks are mitigated, privacy is preserved, intellectual properties are warranted, and liability for malpractice is established (Mesk&#x000F3; and Topol, <xref ref-type="bibr" rid="B135">2023</xref>).</p>
</sec>
<sec>
<title>Questionnaire</title>
<p>Using a questionnaire, we probed both people with dementia (PwD) and their supporters regarding their opinions on the application and features of LLMs in the context of dementia. Participants covered a representative age range for PwD spanning 58&#x02013;88 years (Hugo and Ganguli, <xref ref-type="bibr" rid="B88">2014</xref>). Whereas the gender split was roughly equal for PwD, the majority of the supporters were women, in line with the predominance of female carers in mental illnesses more broadly (Sharma et al., <xref ref-type="bibr" rid="B190">2016</xref>). Only 3 of 15 Pwd and 5 out of 14 supporters reported ever having used apps in the context of dementia. We presented several application scenarios involving companionship, dementia information, navigation, reading or writing aid, and therapy aid. Both PwD and supporters rated all of the scenarios as moderately useful to useful. Older PwD tended to give lower overall scores than younger PwD. It is up to speculation as to why, perhaps indicating a generally more negative outlook, the existence of more severe symptoms that are unlikely to be alleviated by LLMs, or perhaps a larger barrier to use digital apps.</p>
<p>Regarding their priorities for what features LLM-powered apps should have, PwD and supporters ranked agency over data (privacy, transparency, deletion) and ease of use the highest. Opinions on the usefulness of the technology diverged, however. In the free comments sections (see <xref ref-type="table" rid="T4">Table 4</xref>), some respondents praised the promise of the technology (&#x0201C;I think AI has great implications for dementia awareness/care [...]&#x0201D;) but they also raised several caveats. The application scenarios might not address the real needs of PwD (&#x0201C;None of them relate to alleviating the problems of daily living [...]&#x0201D;). There were also concerns about bias and technological affinity required (&#x0201C;This AI sounds like amazing progress but the actual demographic of most dementia sufferers is that they are over 70 so i am not too sure AI is going to help them a great deal as they will probably mostly not be computer literate&#x0201D;). Other respondents pointed out that the usefulness of LLMs depends on the stage of dementia (&#x0201C;The ability to interact with the AI model depends on the degree of decline&#x0201D;) and that LLMs cannot serve as a substitute for human interaction (&#x0201C;I know my loved one would not have been happy talking to a machine. It is not a replacement for a human&#x0201D;).</p>
</sec>
<sec>
<title>Limitations of the questionnaire</title>
<p>Our online survey has several limitations. The use of convenience sampling through dementia organizations as gatekeepers, while practical, may introduce selection bias. This approach relies on participants who are actively engaged with these organizations and have access to the internet, potentially excluding a portion of the dementia population who are less active or lack online access. Additionally, sample sizes of 15 people with dementia and 14 supporters limit the statistical power of the study especially with regard to more subtle effects. It is also possible that in some cases both a PwD and their supporter filled in the questionnaire, leading to correlation between the samples. The anonymity of the questionnaire, while protecting participant privacy and potentially lowering the barrier to participation, also prevents any follow-up for more in-depth data collection.</p>
</sec>
<sec>
<title>Implications for practice</title>
<p>Our findings suggest that LLMs can serve as an invaluable resource in dementia care. By providing personalized interaction and support, LLMs have the potential to improve social engagement and cognitive functioning among individuals with dementia. However, the successful implementation of LLMs in dementia care requires careful consideration of the technology&#x00027;s limitations and the ethical implications of its deployment. Privacy and safety concerns must be meticulously addressed, and systems need to be designed with the end-user in mind, ensuring that they are accessible, intuitive, and genuinely beneficial.</p>
<p>When our previous considerations and the findings from the survey are taken together, we can add the following points for LLM-powered apps for dementia care and management:</p>
<list list-type="bullet">
<list-item><p><bold>On-demand aid</bold>. LLM-powered apps offer on-demand and non-judgmental support, potentially including mental health benefits such as promotion of self-confidence and aiding self-discovery (Ma et al., <xref ref-type="bibr" rid="B129">2024</xref>).</p></list-item>
<list-item><p><bold>Caregiver burden</bold>. Economics mandate a reduction in care cost (Nandi et al., <xref ref-type="bibr" rid="B153">2022</xref>). While LLMs might alleviate caregiver burden, many respondents pointed out that apps cannot serve as substitutes to human interaction. Most likely, a collaborative solution involving human support augmented by a chatbot for periods wherein the human supporter is not available would be a way forward that meets targets both in terms of quality of care and associated monetary cost.</p></list-item>
<list-item><p><bold>Prompt engineering and communication</bold>. It is yet unclear how LLMs perform in the presence of language disorder (Murdoch et al., <xref ref-type="bibr" rid="B151">1987</xref>) and other forms of cognitive impairment (Hugo and Ganguli, <xref ref-type="bibr" rid="B88">2014</xref>). It is conceivable that bespoke models are required, e.g. by finetuning a model such as ChatGPT on a dataset including excerpts of speech from language-impaired individuals. Since the communication is bi-directional, further finetuning might be required to align the model to produce outputs that are more palatable for individuals with impairments.</p></list-item>
<list-item><p><bold>Complexity and technological affinity</bold>. As long as operating LLMs is not seamless any real-world implementation faces a catch-22 scenario: users that benefit from an LLM the most might find it the most challenging to operate LLM-powered apps. For this reason, some respondents pointed out that LLMs should be aimed toward milder versions of dementia such as early-onset dementia. Integration in physical agents such as robots could provide a more seamless gateway between LLM and the user.</p></list-item>
<list-item><p><bold>Co-development of apps</bold>. As a note to tech developers, our survey showed the importance of co-developing solutions with the end user (both PwD and supporters) in the loop early. Otherwise one runs the risk of designing a solution that does not address the needs of PwD or is not usable in the light of their expertise and challenges in using such apps. Furthermore, the language used should not patronize PwD or diminish their agency or cognitive capacities.</p></list-item>
<list-item><p><bold>Data agency</bold>. PwD and supporters stressed the importance of retaining agency over their digital data, including transparency about its usage and the ability to delete it.</p></list-item>
</list>
</sec>
<sec>
<title>Future research directions</title>
<p>To harness the full potential of LLMs in dementia care, future research should focus on several key areas. First, there is a need for longitudinal studies to assess the long-term impact of LLM interactions on individuals with dementia. This includes evaluating the effects on cognitive health, emotional wellbeing, and social engagement over time. It also involves a better characterization of dementia-specific limitations and risks associated with the usage of LLMs. Second, research should explore the customization and personalization of LLMs to meet the diverse needs of individuals with dementia. This includes the development of adaptive algorithms that can tailor interactions based on the user&#x00027;s preferences, behaviors, and cognitive status. Third, the exploration of multimodal LLMs that can interpret and respond to non-verbal cues could significantly enhance the quality of interactions, making the technology more accessible and effective for individuals with varying degrees of cognitive impairment. Fourth, exploring the embodiment of LLMs in robotics could revolutionize dementia care by providing conversational and physical support through social robots (D&#x00027;Onofrio et al., <xref ref-type="bibr" rid="B50">2019</xref>; Fields et al., <xref ref-type="bibr" rid="B56">2021</xref>; Lima et al., <xref ref-type="bibr" rid="B120">2022</xref>). This research should aim to develop adaptive robots that cater to the emotional and physical needs of dementia patients, enhancing their quality of life with a blend of cognitive support and companionship.</p>
</sec>
<sec>
<title>Conclusion</title>
<p>In conclusion, the use of Large Language Models in dementia care represents a promising frontier in the intersection of AI and healthcare. While challenges and limitations exist, the potential benefits of LLMs in enhancing cognitive abilities, enriching social interaction, and supporting caregivers are undeniable. As we move forward, it is crucial that the development and implementation of LLMs is guided by ethical considerations, empirical evidence, and a commitment to improving the lives of individuals living with dementia.</p>
</sec>
</sec>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The raw data and a Jupyter notebook reproducing the results of the questionnaire are available in our GitHub repository (<ext-link ext-link-type="uri" xlink:href="https://github.com/treder/LLMs-for-dementia">https://github.com/treder/LLMs-for-dementia</ext-link>).</p>
</sec>
<sec sec-type="ethics-statement" id="s6">
<title>Ethics statement</title>
<p>The studies involving humans were approved by the School of Computer Science and Informatics Research Ethics Committee at Cardiff University, United Kingdom (reference code COMSC/Ethics/2023/122). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>MT: Conceptualization, Data curation, Investigation, Formal analysis, Methodology, Visualization, Writing &#x02013; review &#x00026; editing, Writing &#x02013; original draft. SL: Investigation, Writing &#x02013; review &#x00026; editing, Writing &#x02013; original draft. KT: Writing &#x02013; review &#x00026; editing, Writing &#x02013; original draft.</p>
</sec>
</body>
<back>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This study was funded by the &#x0201C;Longitude Prize on Dementia&#x0201D;, a research and development program by Alzheimer&#x00027;s Society and Innovate UK. We would also like to thank Dementia Australia (<ext-link ext-link-type="uri" xlink:href="http://www.dementia.org.au">www.dementia.org.au</ext-link>) and Alzheimer&#x00027;s Society (<ext-link ext-link-type="uri" xlink:href="https://www.alzheimers.org.uk/">https://www.alzheimers.org.uk/</ext-link>) for their support in recruiting participants for our questionnaire. KT was supported by Fellowship awards from the Guarantors of Brain (G101149) and Alzheimer&#x00027;s Society, UK (grant number 602).</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>SL was employed by Olive AI Limited. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s10">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frdem.2024.1385303/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frdem.2024.1385303/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.docx" id="SM1" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal"><person-group person-group-type="author"><collab>2023 Alzheimer&#x00027;s Disease Facts and Figures</collab></person-group> (<year>2023</year>). <article-title>Alzheimer&#x00027;s disease facts and figures 2023</article-title>. <source>Alzheimers Dement</source>. <volume>19</volume>, <fpage>1598</fpage>&#x02013;<lpage>1695</lpage>. <pub-id pub-id-type="doi">10.1002/alz.13016</pub-id><pub-id pub-id-type="pmid">36918389</pub-id></citation>
</ref>
<ref id="B2">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Abrol</surname> <given-names>A.</given-names></name> <name><surname>Fu</surname> <given-names>Z.</given-names></name> <name><surname>Du</surname> <given-names>Y.</given-names></name> <name><surname>Calhoun</surname> <given-names>V. D.</given-names></name></person-group> (<year>2019</year>). <article-title>Multimodal data fusion of deep learning and dynamic functional connectivity features to predict alzheimer&#x00027;s disease progression</article-title>, in <source>2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). Presented at the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</source>, (<publisher-loc>Berlin</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>4409</fpage>&#x02013;<lpage>4413</lpage>.</citation>
</ref>
<ref id="B3">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Agbavor</surname> <given-names>F.</given-names></name> <name><surname>Liang</surname> <given-names>H.</given-names></name></person-group> (<year>2022</year>). <article-title>Predicting dementia from spontaneous speech using large language models</article-title>. <source>PLOS Digit. Health</source> <volume>1</volume>:<fpage>e0000168</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000168</pub-id><pub-id pub-id-type="pmid">36812634</pub-id></citation>
</ref>
<ref id="B4">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>AlAfnan</surname> <given-names>M. A.</given-names></name> <name><surname>Dishari</surname> <given-names>S.</given-names></name> <name><surname>Jovic</surname> <given-names>M.</given-names></name> <name><surname>Lomidze</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT as an educational tool: opportunities, challenges, and recommendations for communication, business writing, and composition courses</article-title>. <source>J. Artif. Intell. Technol</source>. <volume>3</volume>, <fpage>60</fpage>&#x02013;<lpage>68</lpage>. <pub-id pub-id-type="doi">10.37965/jait.2023.0184</pub-id></citation>
</ref>
<ref id="B5">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Al-Hameed</surname> <given-names>S.</given-names></name> <name><surname>Benaissa</surname> <given-names>M.</given-names></name> <name><surname>Christensen</surname> <given-names>H.</given-names></name> <name><surname>Mirheidari</surname> <given-names>B.</given-names></name> <name><surname>Blackburn</surname> <given-names>D.</given-names></name> <name><surname>Reuber</surname> <given-names>M.</given-names></name></person-group> (<year>2019</year>). <article-title>A new diagnostic approach for the identification of patients with neurodegenerative cognitive complaints</article-title>. <source>PLOS ONE</source> <volume>14</volume>:<fpage>e0217388</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0217388</pub-id><pub-id pub-id-type="pmid">31125389</pub-id></citation>
</ref>
<ref id="B6">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Altm&#x000E4;e</surname> <given-names>S.</given-names></name> <name><surname>Sola-Leyva</surname> <given-names>A.</given-names></name> <name><surname>Salumets</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Artificial intelligence in scientific writing: a friend or a foe?</article-title> <source>Reprod. Biomed. Online</source> <volume>47</volume>, <fpage>3</fpage>&#x02013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.rbmo.2023.04.009</pub-id><pub-id pub-id-type="pmid">37142479</pub-id></citation>
</ref>
<ref id="B7">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Anil</surname> <given-names>R.</given-names></name> <name><surname>Dai</surname> <given-names>A. M.</given-names></name> <name><surname>Firat</surname> <given-names>O.</given-names></name> <name><surname>Johnson</surname> <given-names>M.</given-names></name> <name><surname>Lepikhin</surname> <given-names>D.</given-names></name> <name><surname>Passos</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>PaLM 2 technical report</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2305.10403</pub-id></citation>
</ref>
<ref id="B8">
<citation citation-type="web"><person-group person-group-type="author"><collab>Anthropic</collab></person-group> (<year>2023a</year>). <source>Claude 2 [WWW Document]. Prod. Announc</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.anthropic.com/index/claude-2">https://www.anthropic.com/index/claude-2</ext-link> (accessed 19 December, 2023).</citation>
</ref>
<ref id="B9">
<citation citation-type="web"><person-group person-group-type="author"><collab>Anthropic</collab></person-group> (<year>2023b</year>). <source>Introducing Claude 2.1 [WWW Document]. Prod. Announc.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.anthropic.com/index/claude-2-1">https://www.anthropic.com/index/claude-2-1</ext-link> (accessed 19 December, 2023).</citation>
</ref>
<ref id="B10">
<citation citation-type="web"><person-group person-group-type="author"><collab>Anthropic</collab></person-group> (<year>2024</year>). <source>Introducing the next generation of Claude [WWW Document].</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.anthropic.com/news/claude-3-family">https://www.anthropic.com/news/claude-3-family</ext-link> (accessed 6 April, 2024).</citation>
</ref>
<ref id="B11">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Armstrong</surname> <given-names>M. J.</given-names></name> <name><surname>Bedenfield</surname> <given-names>N.</given-names></name> <name><surname>Rosselli</surname> <given-names>M.</given-names></name> <name><surname>Curiel Cid</surname> <given-names>R. E.</given-names></name> <name><surname>Kitaigorodsky</surname> <given-names>M.</given-names></name> <name><surname>Galvin</surname> <given-names>J. E.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Best practices for communicating a diagnosis of dementia</article-title>. <source>Neurol. Clin. Pract</source>. <volume>14</volume>:<fpage>e200223</fpage>. <pub-id pub-id-type="doi">10.1212/CPJ.0000000000200223</pub-id><pub-id pub-id-type="pmid">38152063</pub-id></citation>
</ref>
<ref id="B12">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bacsu</surname> <given-names>J.-D.</given-names></name> <name><surname>Fraser</surname> <given-names>S.</given-names></name> <name><surname>Chasteen</surname> <given-names>A. L.</given-names></name> <name><surname>Cammer</surname> <given-names>A.</given-names></name> <name><surname>Grewal</surname> <given-names>K. S.</given-names></name> <name><surname>Bechard</surname> <given-names>L. E.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Using Twitter to examine stigma against people with dementia during COVID-19: infodemiology study</article-title>. <source>JMIR Aging</source> <volume>5</volume>:<fpage>e35677</fpage>. <pub-id pub-id-type="doi">10.2196/35677</pub-id><pub-id pub-id-type="pmid">35290197</pub-id></citation>
</ref>
<ref id="B13">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Baecker</surname> <given-names>L.</given-names></name> <name><surname>Garcia-Dias</surname> <given-names>R.</given-names></name> <name><surname>Vieira</surname> <given-names>S.</given-names></name> <name><surname>Scarpazza</surname> <given-names>C.</given-names></name> <name><surname>Mechelli</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>Machine learning for brain age prediction: introduction to methods and clinical applications</article-title>. <source>eBioMedicine</source> <volume>72</volume>:<fpage>103600</fpage>. <pub-id pub-id-type="doi">10.1016/j.ebiom.2021.103600</pub-id><pub-id pub-id-type="pmid">34614461</pub-id></citation>
</ref>
<ref id="B14">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bang</surname> <given-names>J.-U.</given-names></name> <name><surname>Han</surname> <given-names>S.-H.</given-names></name> <name><surname>Kang</surname> <given-names>B.-O.</given-names></name></person-group> (<year>2024</year>). <article-title>Alzheimer&#x00027;s disease recognition from spontaneous speech using large language models</article-title>. <source>ETRI J</source>. <volume>46</volume>, <fpage>96</fpage>&#x02013;<lpage>105</lpage>. <pub-id pub-id-type="doi">10.4218/etrij.2023-0356</pub-id></citation>
</ref>
<ref id="B15">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Banks</surname> <given-names>J.</given-names></name> <name><surname>Warkentin</surname> <given-names>T.</given-names></name></person-group> (<year>2024</year>). <source>Gemma: Introducing New State-of-the-Art Open Models. Google.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/developers/gemma-open-models/">https://blog.google/technology/developers/gemma-open-models/</ext-link> (accessed 6 April, 2024).</citation>
</ref>
<ref id="B16">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Barocas</surname> <given-names>S.</given-names></name> <name><surname>Hardt</surname> <given-names>M.</given-names></name> <name><surname>Narayanan</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <source>Fairness and Machine Learning</source>. <publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>MIT Press</publisher-name>.</citation>
</ref>
<ref id="B17">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bir</surname> <given-names>S. C.</given-names></name> <name><surname>Khan</surname> <given-names>M. W.</given-names></name> <name><surname>Javalkar</surname> <given-names>V.</given-names></name> <name><surname>Toledo</surname> <given-names>E. G.</given-names></name> <name><surname>Kelley</surname> <given-names>R. E.</given-names></name></person-group> (<year>2021</year>). <article-title>Emerging concepts in vascular dementia: a review</article-title>. <source>J. Stroke Cerebrovasc. Dis. Off. J. Natl. Stroke Assoc</source>. <volume>30</volume>:<fpage>105864</fpage>. <pub-id pub-id-type="doi">10.1016/j.jstrokecerebrovasdis.2021.105864</pub-id></citation>
</ref>
<ref id="B18">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Birhane</surname> <given-names>A.</given-names></name> <name><surname>Prabhu</surname> <given-names>V. U.</given-names></name> <name><surname>Kahembwe</surname> <given-names>E.</given-names></name></person-group> (<year>2021</year>). <article-title>Multimodal datasets: misogyny, pornography, and malignant stereotypes</article-title>. <source>arXiv [Preprint].</source> <pub-id pub-id-type="doi">10.48550/arXiv.2110.01963</pub-id></citation>
</ref>
<ref id="B19">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Blodgett</surname> <given-names>S. L.</given-names></name> <name><surname>Barocas</surname> <given-names>S.</given-names></name> <name><surname>Daum&#x000E9; III</surname> <given-names>H.</given-names></name> <name><surname>Wallach</surname> <given-names>H.</given-names></name></person-group> (<year>2020</year>). <article-title>Language (Technology) is Power: A Critical Survey of &#x0201C;Bias&#x0201D; in NLP</article-title>, in <source>TetreaultProceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Presented at the ACL 2020</source>, eds. D. Jurafsky, J. Chai, and N. Schluter (<publisher-loc>Stroudsburg</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>5454</fpage>&#x02013;<lpage>5476</lpage>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Borchert</surname> <given-names>R. J.</given-names></name> <name><surname>Azevedo</surname> <given-names>T.</given-names></name> <name><surname>Badhwar</surname> <given-names>A.</given-names></name> <name><surname>Bernal</surname> <given-names>J.</given-names></name> <name><surname>Betts</surname> <given-names>M.</given-names></name> <name><surname>Bruffaerts</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Artificial intelligence for diagnostic and prognostic neuroimaging in dementia: a systematic review</article-title>. <source>Alzheimers Dement</source>. <volume>19</volume>, <fpage>5885</fpage>&#x02013;<lpage>5904</lpage>. <pub-id pub-id-type="doi">10.1002/alz.13412</pub-id><pub-id pub-id-type="pmid">37563912</pub-id></citation>
</ref>
<ref id="B21">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Botchu</surname> <given-names>B.</given-names></name> <name><surname>Karthikeyan</surname> <given-names>I. P.</given-names></name> <name><surname>Botchu</surname> <given-names>R.</given-names></name></person-group> (<year>2023</year>). <article-title>Can ChatGPT empower people with dyslexia?</article-title> <source>Disabil. Rehabil. Assist. Technol</source>. <volume>0</volume>, <fpage>1</fpage>&#x02013;<lpage>2</lpage>. <pub-id pub-id-type="doi">10.1080/17483107.2023.2256805</pub-id></citation>
</ref>
<ref id="B22">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bouazizi</surname> <given-names>M.</given-names></name> <name><surname>Zheng</surname> <given-names>C.</given-names></name> <name><surname>Yang</surname> <given-names>S.</given-names></name> <name><surname>Ohtsuki</surname> <given-names>T.</given-names></name></person-group> (<year>2024</year>). <article-title>Dementia detection from speech: what if language models are not the answer?</article-title> <source>Information</source> <volume>15</volume>, <fpage>2</fpage>. <pub-id pub-id-type="doi">10.3390/info15010002</pub-id></citation>
</ref>
<ref id="B23">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Brierley</surname> <given-names>C.</given-names></name></person-group> (<year>2021</year>). <source>AI Could Detect Dementia Years Before Symptoms Appear [WWW Document]. Univ. Camb.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.cam.ac.uk/stories/AIdementia">https://www.cam.ac.uk/stories/AIdementia</ext-link> (accessed 19 November, 2023).</citation>
</ref>
<ref id="B24">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bubeck</surname> <given-names>S.</given-names></name> <name><surname>Chandrasekaran</surname> <given-names>V.</given-names></name> <name><surname>Eldan</surname> <given-names>R.</given-names></name> <name><surname>Gehrke</surname> <given-names>J.</given-names></name> <name><surname>Horvitz</surname> <given-names>E.</given-names></name> <name><surname>Kamar</surname> <given-names>E.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Sparks of artificial general intelligence: early experiments with GPT-4</article-title>. <source>arXiv [Preprint</source>]. <pub-id pub-id-type="doi">10.48550/arXiv.2303.12712</pub-id></citation>
</ref>
<ref id="B25">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bzdok</surname> <given-names>D.</given-names></name> <name><surname>Thieme</surname> <given-names>A.</given-names></name> <name><surname>Levkovskyy</surname> <given-names>O.</given-names></name> <name><surname>Wren</surname> <given-names>P.</given-names></name> <name><surname>Ray</surname> <given-names>T.</given-names></name> <name><surname>Reddy</surname> <given-names>S.</given-names></name></person-group> (<year>2024</year>). <article-title>Data science opportunities of large language models for neuroscience and biomedicine</article-title>. <source>Neuron</source>. <volume>112</volume>, <fpage>698</fpage>&#x02013;<lpage>717</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuron.2024.01.016</pub-id><pub-id pub-id-type="pmid">38340718</pub-id></citation>
</ref>
<ref id="B26">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Car&#x000F3;s</surname> <given-names>M.</given-names></name> <name><surname>Garolera</surname> <given-names>M.</given-names></name> <name><surname>Radeva</surname> <given-names>P.</given-names></name> <name><surname>Giro-i-Nieto</surname> <given-names>X.</given-names></name></person-group> (<year>2020</year>). <article-title>Automatic reminiscence therapy for dementia</article-title>, in <source>Proceedings of the 2020 International Conference on Multimedia Retrieval, ICMR&#x00027;20.</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>383</fpage>&#x02013;<lpage>387</lpage>.</citation>
</ref>
<ref id="B27">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>J.</given-names></name> <name><surname>Lin</surname> <given-names>H.</given-names></name> <name><surname>Han</surname> <given-names>X.</given-names></name> <name><surname>Sun</surname> <given-names>L.</given-names></name></person-group> (<year>2024</year>). <article-title>Benchmarking large language models in retrieval-augmented generation</article-title>. <source>Proc. AAAI Conf. Artif. Intell</source>. <volume>38</volume>, <fpage>17754</fpage>&#x02013;<lpage>17762</lpage>. <pub-id pub-id-type="doi">10.1609/aaai.v38i16.29728</pub-id></citation>
</ref>
<ref id="B28">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>M.</given-names></name> <name><surname>Tworek</surname> <given-names>J.</given-names></name> <name><surname>Jun</surname> <given-names>H.</given-names></name> <name><surname>Yuan</surname> <given-names>Q.</given-names></name> <name><surname>Pinto</surname> <given-names>H. P.</given-names></name> <name><surname>de</surname> <given-names>O.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Evaluating large language models trained on code</article-title>. <source>arXiv [Preprint</source>]. <pub-id pub-id-type="doi">10.48550/arXiv.2107.03374</pub-id></citation>
</ref>
<ref id="B29">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cheng</surname> <given-names>S.-W.</given-names></name> <name><surname>Chang</surname> <given-names>C.-W.</given-names></name> <name><surname>Chang</surname> <given-names>W.-J.</given-names></name> <name><surname>Wang</surname> <given-names>H.-W.</given-names></name> <name><surname>Liang</surname> <given-names>C.-S.</given-names></name> <name><surname>Kishimoto</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>The now and future of ChatGPT and GPT in psychiatry</article-title>. <source>Psychiatry Clin. Neurosci</source>. <volume>77</volume>, <fpage>592</fpage>&#x02013;<lpage>596</lpage>. <pub-id pub-id-type="doi">10.1111/pcn.13588</pub-id><pub-id pub-id-type="pmid">37612880</pub-id></citation>
</ref>
<ref id="B30">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chopard</surname> <given-names>D.</given-names></name> <name><surname>Treder</surname> <given-names>M. S.</given-names></name> <name><surname>Corcoran</surname> <given-names>P.</given-names></name> <name><surname>Ahmed</surname> <given-names>N.</given-names></name> <name><surname>Johnson</surname> <given-names>C.</given-names></name> <name><surname>Busse</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Text mining of adverse events in clinical trials: deep learning approach</article-title>. <source>JMIR Med. Inform</source>. <volume>9</volume>:<fpage>e28632</fpage>. <pub-id pub-id-type="doi">10.2196/28632</pub-id><pub-id pub-id-type="pmid">34951601</pub-id></citation>
</ref>
<ref id="B31">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Choudhury</surname> <given-names>A.</given-names></name> <name><surname>Shamszare</surname> <given-names>H.</given-names></name></person-group> (<year>2023</year>). <article-title>Investigating the impact of user trust on the adoption and use of ChatGPT: survey analysis</article-title>. <source>J. Med. Internet Res</source>. <volume>25</volume>:<fpage>e47184</fpage>. <pub-id pub-id-type="doi">10.2196/47184</pub-id><pub-id pub-id-type="pmid">37314848</pub-id></citation>
</ref>
<ref id="B32">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coda-Forno</surname> <given-names>J.</given-names></name> <name><surname>Witte</surname> <given-names>K.</given-names></name> <name><surname>Jagadish</surname> <given-names>A. K.</given-names></name> <name><surname>Binz</surname> <given-names>M.</given-names></name> <name><surname>Akata</surname> <given-names>Z.</given-names></name> <name><surname>Schulz</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <article-title>Inducing anxiety in large language models increases exploration and bias</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2304.11111</pub-id></citation>
</ref>
<ref id="B33">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coyle-Gilchrist</surname> <given-names>I. T. S.</given-names></name> <name><surname>Dick</surname> <given-names>K. M.</given-names></name> <name><surname>Patterson</surname> <given-names>K.</given-names></name> <name><surname>V&#x000E1;zquez Rodr&#x000ED;quez</surname> <given-names>P.</given-names></name> <name><surname>Wehmann</surname> <given-names>E.</given-names></name> <name><surname>Wilcox</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Prevalence, characteristics, and survival of frontotemporal lobar degeneration syndromes</article-title>. <source>Neurology</source> <volume>86</volume>, <fpage>1736</fpage>&#x02013;<lpage>1743</lpage>. <pub-id pub-id-type="doi">10.1212/WNL.0000000000002638</pub-id><pub-id pub-id-type="pmid">27037234</pub-id></citation>
</ref>
<ref id="B34">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Creamer</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <source>&#x02018;Hallucinate&#x02019; Chosen as Cambridge Dictionary&#x00027;s Word of the Year</source>. <publisher-loc>London</publisher-loc>: <publisher-name>The Guardian</publisher-name>.</citation>
</ref>
<ref id="B35">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Dangerfield</surname> <given-names>K.</given-names></name> <name><surname>Katherine</surname> <given-names>W.</given-names></name></person-group> (<year>2023</year>). <article-title>1st signs of Alzheimer&#x00027;s may be detected in your eyes</article-title>, in <source>This AI scan may help find it - National | Globalnews.ca. Glob. News</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://globalnews.ca/news/9840571/ai-alzheimers-disease-detection/">https://globalnews.ca/news/9840571/ai-alzheimers-disease-detection/</ext-link> (accessed 19 November, 2023).</citation>
</ref>
<ref id="B36">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Davis</surname> <given-names>W.</given-names></name></person-group> (<year>2023</year>). <article-title>Sarah Silverman is suing OpenAI and Meta for copyright infringement</article-title>, in <source>The Verge</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.theverge.com/2023/7/9/23788741/sarah-silverman-openai-meta-chatgpt-llama-copyright-infringement-chatbots-artificial-intelligence-ai">https://www.theverge.com/2023/7/9/23788741/sarah-silverman-openai-meta-chatgpt-llama-copyright-infringement-chatbots-artificial-intelligence-ai</ext-link> (accessed 25 November, 2023).</citation>
</ref>
<ref id="B37">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dawid</surname> <given-names>A. P.</given-names></name></person-group> (<year>1982</year>). <article-title>The well-calibrated Bayesian</article-title>. <source>J. Am. Stat. Assoc</source>. <volume>77</volume>, <fpage>605</fpage>&#x02013;<lpage>610</lpage>. <pub-id pub-id-type="doi">10.1080/01621459.1982.10477856</pub-id></citation>
</ref>
<ref id="B38">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>de Arriba-P&#x000E9;rez</surname> <given-names>F.</given-names></name> <name><surname>Garc&#x000ED;a-M&#x000E9;ndez</surname> <given-names>S.</given-names></name> <name><surname>Gonz&#x000E1;lez-Casta&#x000F1;o</surname> <given-names>F. J.</given-names></name> <name><surname>Costa-Montenegro</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <article-title>Automatic detection of cognitive impairment in elderly people using an entertainment chatbot with Natural Language Processing capabilities</article-title>. <source>J. Ambient Intell. Humaniz. Comput</source>. <volume>14</volume>, <fpage>16283</fpage>&#x02013;<lpage>16298</lpage>. <pub-id pub-id-type="doi">10.1007/s12652-022-03849-2</pub-id></citation>
</ref>
<ref id="B39">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>de la Fuente Garcia</surname> <given-names>S.</given-names></name> <name><surname>Ritchie</surname> <given-names>C. W.</given-names></name> <name><surname>Luz</surname> <given-names>S.</given-names></name></person-group> (<year>2020</year>). <article-title>Artificial intelligence, speech, and language processing approaches to monitoring Alzheimer&#x00027;s disease: a systematic review. J</article-title>. <source>Alzheimers Dis. JAD</source> <volume>78</volume>, <fpage>1547</fpage>&#x02013;<lpage>1574</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-200888</pub-id><pub-id pub-id-type="pmid">33185605</pub-id></citation>
</ref>
<ref id="B40">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Demiris</surname> <given-names>G.</given-names></name> <name><surname>Thompson</surname> <given-names>H. J.</given-names></name> <name><surname>Lazar</surname> <given-names>A.</given-names></name> <name><surname>Lin</surname> <given-names>S.-Y.</given-names></name></person-group> (<year>2017</year>). <article-title>Evaluation of a digital companion for older adults with mild cognitive impairment</article-title>, in <source>AMIA. Annu. Symp. Proc. 2016</source>, <fpage>496</fpage>&#x02013;<lpage>503</lpage>.</citation>
</ref>
<ref id="B41">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Demszky</surname> <given-names>D.</given-names></name> <name><surname>Yang</surname> <given-names>D.</given-names></name> <name><surname>Yeager</surname> <given-names>D. S.</given-names></name> <name><surname>Bryan</surname> <given-names>C. J.</given-names></name> <name><surname>Clapper</surname> <given-names>M.</given-names></name> <name><surname>Chandhok</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Using large language models in psychology</article-title>. <source>Nat. Rev. Psychol</source>. <volume>2</volume>, <fpage>688</fpage>&#x02013;<lpage>701</lpage>. <pub-id pub-id-type="doi">10.1038/s44159-023-00241-5</pub-id></citation>
</ref>
<ref id="B42">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Denecke</surname> <given-names>K.</given-names></name> <name><surname>Schmid</surname> <given-names>N.</given-names></name> <name><surname>N&#x000FC;ssli</surname> <given-names>S.</given-names></name></person-group> (<year>2022</year>). <article-title>Implementation of cognitive behavioral therapy in e&#x02013;mental health apps: literature review</article-title>. <source>J. Med. Internet Res</source>. <volume>24</volume>:<fpage>e27791</fpage>. <pub-id pub-id-type="doi">10.2196/27791</pub-id><pub-id pub-id-type="pmid">35266875</pub-id></citation>
</ref>
<ref id="B43">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Deng</surname> <given-names>G.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>MASTERKEY: Automated Jailbreaking of Large Language Model Chatbots</article-title>, in <source>Proceedings 2024 Network and Distributed System Security Symposium. Presented at the Network and Distributed System Security Symposium, Internet Society</source>. (<publisher-loc>San Diego, CA</publisher-loc>: <publisher-name>Internet Society</publisher-name>).</citation>
</ref>
<ref id="B44">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Deshpande</surname> <given-names>A.</given-names></name> <name><surname>Murahari</surname> <given-names>V.</given-names></name> <name><surname>Rajpurohit</surname> <given-names>T.</given-names></name> <name><surname>Kalyan</surname> <given-names>A.</given-names></name> <name><surname>Narasimhan</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <source>Toxicity in ChatGPT: Analyzing Persona-assigned Language Models</source>.</citation>
</ref>
<ref id="B45">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dhamala</surname> <given-names>J.</given-names></name> <name><surname>Sun</surname> <given-names>T.</given-names></name> <name><surname>Kumar</surname> <given-names>V.</given-names></name> <name><surname>Krishna</surname> <given-names>S.</given-names></name> <name><surname>Pruksachatkun</surname> <given-names>Y.</given-names></name> <name><surname>Chang</surname> <given-names>K.-W.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>BOLD: dataset and metrics for measuring biases in open-ended language generation</article-title>, in <source>Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT&#x00027;21</source>. (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>862</fpage>&#x02013;<lpage>872</lpage>.</citation>
</ref>
<ref id="B46">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dhuliawala</surname> <given-names>S.</given-names></name> <name><surname>Komeili</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>J.</given-names></name> <name><surname>Raileanu</surname> <given-names>R.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Celikyilmaz</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Chain-of-verification reduces hallucination in large language models</article-title>. <source>arXiv [Preprint</source>]. <pub-id pub-id-type="doi">10.48550/arXiv.2309.11495</pub-id></citation>
</ref>
<ref id="B47">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname> <given-names>Y.</given-names></name> <name><surname>Sohn</surname> <given-names>J. H.</given-names></name> <name><surname>Kawczynski</surname> <given-names>M. G.</given-names></name> <name><surname>Trivedi</surname> <given-names>H.</given-names></name> <name><surname>Harnish</surname> <given-names>R.</given-names></name> <name><surname>Jenkins</surname> <given-names>N. W.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>A deep learning model to predict a diagnosis of Alzheimer disease by using 18F-FDG PET of the brain</article-title>. <source>Radiology</source> <volume>290</volume>, <fpage>456</fpage>&#x02013;<lpage>464</lpage>. <pub-id pub-id-type="doi">10.1148/radiol.2018180958</pub-id><pub-id pub-id-type="pmid">30398430</pub-id></citation>
</ref>
<ref id="B48">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Dixon</surname> <given-names>E.</given-names></name> <name><surname>Michaels</surname> <given-names>R.</given-names></name> <name><surname>Xiao</surname> <given-names>X.</given-names></name> <name><surname>Zhong</surname> <given-names>Y.</given-names></name> <name><surname>Clary</surname> <given-names>P.</given-names></name> <name><surname>Narayanan</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Mobile phone use by people with mild to moderate dementia: uncovering challenges and identifying opportunities: mobile phone use by people with mild to moderate dementia</article-title>, in <source>Proceedings of the 24th International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS&#x00027;22</source>. (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1145/3517428.3544809</pub-id></citation>
</ref>
<ref id="B49">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dobbe</surname> <given-names>R.</given-names></name> <name><surname>Krendl Gilbert</surname> <given-names>T.</given-names></name> <name><surname>Mintz</surname> <given-names>Y.</given-names></name></person-group> (<year>2021</year>). <article-title>Hard choices in artificial intelligence</article-title>. <source>Artif. Intell</source>. 300, 103555. <pub-id pub-id-type="doi">10.1016/j.artint.2021.103555</pub-id></citation>
</ref>
<ref id="B50">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>D&#x00027;Onofrio</surname> <given-names>G.</given-names></name> <name><surname>Sancarlo</surname> <given-names>D.</given-names></name> <name><surname>Raciti</surname> <given-names>M.</given-names></name> <name><surname>Burke</surname> <given-names>M.</given-names></name> <name><surname>Teare</surname> <given-names>A.</given-names></name> <name><surname>Kovacic</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>MARIO project: validation and evidence of service robots for older people with dementia</article-title>. <source>J. Alzheimers Dis</source>. <volume>68</volume>, <fpage>1587</fpage>&#x02013;<lpage>1601</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-181165</pub-id><pub-id pub-id-type="pmid">30958360</pub-id></citation>
</ref>
<ref id="B51">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dosso</surname> <given-names>J. A.</given-names></name> <name><surname>Kailley</surname> <given-names>J. N.</given-names></name> <name><surname>Robillard</surname> <given-names>J. M.</given-names></name></person-group> (<year>2023</year>). <article-title>What does ChatGPT know about dementia? A comparative analysis of information quality</article-title>. <source>J. Alzheimers Dis.</source> <volume>97</volume>, <fpage>559</fpage>&#x02013;<lpage>565</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-230573</pub-id></citation>
</ref>
<ref id="B52">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Du</surname> <given-names>Y.</given-names></name> <name><surname>Fu</surname> <given-names>Z.</given-names></name> <name><surname>Calhoun</surname> <given-names>V. D.</given-names></name></person-group> (<year>2018</year>). <article-title>Classification and prediction of brain disorders using functional connectivity: promising but challenging</article-title>. <source>Front. Neurosci</source>. <volume>12</volume>:<fpage>525</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2018.00525</pub-id><pub-id pub-id-type="pmid">30127711</pub-id></citation>
</ref>
<ref id="B53">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Eliot</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <source>People Are Eagerly Consulting Generative AI ChatGPT For Mental Health Advice, Stressing Out AI Ethics And AI Law. Forbes</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.forbes.com/sites/deloitte/2024/02/16/making-the-leap-from-smart-to-the-metaverse-in-operations/?sh=42cf52f62b0b">https://www.forbes.com/sites/deloitte/2024/02/16/making-the-leap-from-smart-to-the-metaverse-in-operations/?sh=42cf52f62b0b</ext-link> (accessed February 22, 2024).</citation>
</ref>
<ref id="B54">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ferrara</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <article-title>Should ChatGPT be biased? Challenges and risks of bias in large language models</article-title>, in <source>First Monday</source>.</citation>
</ref>
<ref id="B55">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Field</surname> <given-names>A.</given-names></name> <name><surname>Coston</surname> <given-names>A.</given-names></name> <name><surname>Gandhi</surname> <given-names>N.</given-names></name> <name><surname>Chouldechova</surname> <given-names>A.</given-names></name> <name><surname>Putnam-Hornstein</surname> <given-names>E.</given-names></name> <name><surname>Steier</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Examining risks of racial biases in NLP tools for child protective services</article-title>, in <source>2023 ACM Conference on Fairness, Accountability, and Transparency. Presented at the FAccT&#x00027;23: the 2023 ACM Conference on Fairness, Accountability, and Transparency</source> (<publisher-loc>Chicago</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1479</fpage>&#x02013;<lpage>1492</lpage>.</citation>
</ref>
<ref id="B56">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fields</surname> <given-names>N.</given-names></name> <name><surname>Xu</surname> <given-names>L.</given-names></name> <name><surname>Greer</surname> <given-names>J.</given-names></name> <name><surname>Murphy</surname> <given-names>E.</given-names></name></person-group> (<year>2021</year>). <article-title>Shall I compare thee&#x02026;to a robot? An exploratory pilot study using participatory arts and social robotics to improve psychological well-being in later life</article-title>. <source>Aging Ment. Health</source> <volume>25</volume>, <fpage>575</fpage>&#x02013;<lpage>584</lpage>. <pub-id pub-id-type="doi">10.1080/13607863.2019.1699016</pub-id><pub-id pub-id-type="pmid">31851830</pub-id></citation>
</ref>
<ref id="B57">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fischer</surname> <given-names>C. E.</given-names></name> <name><surname>Qian</surname> <given-names>W.</given-names></name> <name><surname>Schweizer</surname> <given-names>T. A.</given-names></name> <name><surname>Ismail</surname> <given-names>Z.</given-names></name> <name><surname>Smith</surname> <given-names>E. E.</given-names></name> <name><surname>Millikin</surname> <given-names>C. P.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Determining the impact of psychosis on rates of false-positive and false-negative diagnosis in Alzheimer&#x00027;s disease</article-title>. <source>Alzheimers Dement. Transl. Res. Clin. Interv</source>. <volume>3</volume>, <fpage>385</fpage>&#x02013;<lpage>392</lpage>. <pub-id pub-id-type="doi">10.1016/j.trci.2017.06.001</pub-id></citation>
</ref>
<ref id="B58">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Floridi</surname> <given-names>L.</given-names></name> <name><surname>Floridi</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <source>The Ethics of Artificial Intelligence: Principles, Challenges, and Opportunities</source>. <publisher-loc>Oxford, New York</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation>
</ref>
<ref id="B59">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Franceschelli</surname> <given-names>G.</given-names></name> <name><surname>Musolesi</surname> <given-names>M.</given-names></name></person-group> (<year>2022</year>). <article-title>Copyright in generative deep learning</article-title>. <source>Data Policy</source> <volume>4</volume>:<fpage>e17</fpage>. <pub-id pub-id-type="doi">10.1017/dap.2022.10</pub-id></citation>
</ref>
<ref id="B60">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>F&#x000FC;gener</surname> <given-names>A.</given-names></name> <name><surname>Grahl</surname> <given-names>J.</given-names></name> <name><surname>Gupta</surname> <given-names>A.</given-names></name> <name><surname>Ketter</surname> <given-names>W.</given-names></name></person-group> (<year>2021</year>). <source>Will Humans-in-The-Loop Become Borgs? Merits and Pitfalls of Working with AI.</source> Management Information Systems Quarterly (MISQ). Vol. 45. Available online at: <ext-link ext-link-type="uri" xlink:href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3879937">https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3879937</ext-link></citation>
</ref>
<ref id="B61">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gabriel</surname> <given-names>I.</given-names></name></person-group> (<year>2020</year>). <article-title>Artificial intelligence, values, and alignment</article-title>. <source>Minds Mach</source>. <volume>30</volume>, <fpage>411</fpage>&#x02013;<lpage>437</lpage>. <pub-id pub-id-type="doi">10.1007/s11023-020-09539-2</pub-id></citation>
</ref>
<ref id="B62">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gallegos</surname> <given-names>I. O.</given-names></name> <name><surname>Rossi</surname> <given-names>R. A.</given-names></name> <name><surname>Barrow</surname> <given-names>J.</given-names></name> <name><surname>Tanjim</surname> <given-names>M. M.</given-names></name> <name><surname>Kim</surname> <given-names>S.</given-names></name> <name><surname>Dernoncourt</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Bias and fairness in large language models: a survey</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2309.00770</pub-id></citation>
</ref>
<ref id="B63">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ganguli</surname> <given-names>D.</given-names></name> <name><surname>Askell</surname> <given-names>A.</given-names></name> <name><surname>Schiefer</surname> <given-names>N.</given-names></name> <name><surname>Liao</surname> <given-names>T. I.</given-names></name> <name><surname>Luko&#x00161;iute</surname> <given-names>K.</given-names></name> <name><surname>Chen</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>The capacity for moral self-correction in large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2302.07459</pub-id></citation>
</ref>
<ref id="B64">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Ganguli</surname> <given-names>D.</given-names></name> <name><surname>Hernandez</surname> <given-names>D.</given-names></name> <name><surname>Lovitt</surname> <given-names>L.</given-names></name> <name><surname>Askell</surname> <given-names>A.</given-names></name> <name><surname>Bai</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Predictability and Surprise in Large Generative Models</article-title>, in <source>Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT&#x00027;22</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>1747</fpage>&#x02013;<lpage>1764</lpage>.</citation>
</ref>
<ref id="B65">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>Y.</given-names></name> <name><surname>Xiong</surname> <given-names>Y.</given-names></name> <name><surname>Gao</surname> <given-names>X.</given-names></name> <name><surname>Jia</surname> <given-names>K.</given-names></name> <name><surname>Pan</surname> <given-names>J.</given-names></name> <name><surname>Bi</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2024</year>). <source>Retrieval-Augmented Generation for Large Language Models: A Survey</source>.</citation>
</ref>
<ref id="B66">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a-Requejo</surname> <given-names>A.</given-names></name> <name><surname>P&#x000E9;rez-Rubio</surname> <given-names>M. C.</given-names></name> <name><surname>Villadangos</surname> <given-names>J. M.</given-names></name> <name><surname>Hern&#x000E1;ndez</surname> <given-names>&#x000C1;.</given-names></name></person-group> (<year>2023</year>). <article-title>Activity monitoring and location sensory system for people with mild cognitive impairments.</article-title> <source>IEEE Sens. J</source>. <volume>23</volume>, <fpage>5448</fpage>&#x02013;<lpage>5458</lpage>. <pub-id pub-id-type="doi">10.1109/JSEN.2023.3239980</pub-id></citation>
</ref>
<ref id="B67">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Ghahramani</surname> <given-names>Z.</given-names></name></person-group> (<year>2022</year>). <article-title>Understanding the world through language</article-title> in <source>The Keyword</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/ai/understanding-the-world-through-language/">https://blog.google/technology/ai/understanding-the-world-through-language/</ext-link> (accessed 19 December, 2023).</citation>
</ref>
<ref id="B68">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Ghahramani</surname> <given-names>Z.</given-names></name></person-group> (<year>2023</year>). <source>Introducing PaLM 2 [WWW Document]. Google</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/ai/google-palm-2-ai-large-language-model/">https://blog.google/technology/ai/google-palm-2-ai-large-language-model/</ext-link> (accessed 11 July, 2023).</citation>
</ref>
<ref id="B69">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Giorgio</surname> <given-names>J.</given-names></name> <name><surname>Jagust</surname> <given-names>W. J.</given-names></name> <name><surname>Baker</surname> <given-names>S.</given-names></name> <name><surname>Landau</surname> <given-names>S. M.</given-names></name> <name><surname>Tino</surname> <given-names>P.</given-names></name> <name><surname>Kourtzi</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Predicting future regional tau accumulation in asymptomatic and early alzheimer&#x00027;s disease</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.1101/2020.08.15.252601</pub-id></citation>
</ref>
<ref id="B70">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Gooding</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <article-title>Open source LLMs could make artificial intelligence more dangerous, says &#x0201C;godfather&#x0201D; of AI</article-title>, in <source>Tech Monit</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://techmonitor.ai/technology/ai-and-automation/open-source-chatgpt-ai-llm-geoffrey-hinton">https://techmonitor.ai/technology/ai-and-automation/open-source-chatgpt-ai-llm-geoffrey-hinton</ext-link> (accessed 10 December, 2023).</citation>
</ref>
<ref id="B71">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Goodman</surname> <given-names>S. M.</given-names></name> <name><surname>Buehler</surname> <given-names>E.</given-names></name> <name><surname>Clary</surname> <given-names>P.</given-names></name> <name><surname>Coenen</surname> <given-names>A.</given-names></name> <name><surname>Donsbach</surname> <given-names>A.</given-names></name> <name><surname>Horne</surname> <given-names>T. N.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>LaMPost: Design and Evaluation of an AI-assisted Email Writing Prototype for Adults with Dyslexia</article-title>, in <source>Proceedings of the 24th International ACM SIGACCESS Conference on Computers and Accessibility. Presented at the ASSETS&#x00027;22: The 24th International ACM SIGACCESS Conference on Computers and Accessibility</source> (<publisher-loc>Athens</publisher-loc>: <publisher-name>ACM</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>18</lpage>.</citation>
</ref>
<ref id="B72">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Greshake</surname> <given-names>K.</given-names></name> <name><surname>Abdelnabi</surname> <given-names>S.</given-names></name> <name><surname>Mishra</surname> <given-names>S.</given-names></name> <name><surname>Endres</surname> <given-names>C.</given-names></name> <name><surname>Holz</surname> <given-names>T.</given-names></name> <name><surname>Fritz</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <article-title>Not what you&#x00027;ve signed up for: compromising real-world llm-integrated applications with indirect prompt injection</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2302.12173</pub-id></citation>
</ref>
<ref id="B73">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Griffin</surname> <given-names>L.</given-names></name> <name><surname>Kleinberg</surname> <given-names>B.</given-names></name> <name><surname>Mozes</surname> <given-names>M.</given-names></name> <name><surname>Mai</surname> <given-names>K.</given-names></name> <name><surname>Vau</surname> <given-names>M. D. M.</given-names></name> <name><surname>Caldwell</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Large Language Models respond to Influence like Humans</article-title>, in <source>Proceedings of the First Workshop on Social Influence in Conversations (SICon 2023). Presented at the SICon 2023</source>, eds. K. Chawla, and W. Shi (<publisher-loc>Toronto, ON</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>15</fpage>&#x02013;<lpage>24</lpage>.</citation>
</ref>
<ref id="B74">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Groeneveld</surname> <given-names>D.</given-names></name> <name><surname>Beltagy</surname> <given-names>I.</given-names></name> <name><surname>Walsh</surname> <given-names>P.</given-names></name> <name><surname>Bhagia</surname> <given-names>A.</given-names></name> <name><surname>Kinney</surname> <given-names>R.</given-names></name> <name><surname>Tafjord</surname> <given-names>O.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>OLMo: accelerating the science of language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2402.00838</pub-id></citation>
</ref>
<ref id="B75">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>A.</given-names></name> <name><surname>Waldron</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>A responsible path to generative AI in healthcare</article-title>, in <source>Google Cloud Blog</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://cloud.google.com/blog/topics/healthcare-life-sciences/sharing-google-med-palm-2-medical-large-language-model">https://cloud.google.com/blog/topics/healthcare-life-sciences/sharing-google-med-palm-2-medical-large-language-model</ext-link> (accessed 20 December, 2023).</citation>
</ref>
<ref id="B76">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gustafsson</surname> <given-names>C.</given-names></name> <name><surname>Svanberg</surname> <given-names>C.</given-names></name> <name><surname>M&#x000FC;llersdorf</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Using a robotic cat in dementia care: a pilot study</article-title>. <source>J. Gerontol. Nurs</source>. <volume>41</volume>, <fpage>46</fpage>&#x02013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.3928/00989134-20150806-44</pub-id></citation>
</ref>
<ref id="B77">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hagendorff</surname> <given-names>T.</given-names></name></person-group> (<year>2023</year>). <source>Machine Psychology: Investigating Emergent Capabilities and Behavior in Large Language Models Using Psychological Methods</source>.</citation>
</ref>
<ref id="B78">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hardy</surname> <given-names>C. J. D.</given-names></name> <name><surname>Marshall</surname> <given-names>C. R.</given-names></name> <name><surname>Golden</surname> <given-names>H. L.</given-names></name> <name><surname>Clark</surname> <given-names>C. N.</given-names></name> <name><surname>Mummery</surname> <given-names>C. J.</given-names></name> <name><surname>Griffiths</surname> <given-names>T. D.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Hearing and dementia</article-title>. <source>J. Neurol</source>. <volume>263</volume>, <fpage>2339</fpage>&#x02013;<lpage>2354</lpage>. <pub-id pub-id-type="doi">10.1007/s00415-016-8208-y</pub-id><pub-id pub-id-type="pmid">27372450</pub-id></citation>
</ref>
<ref id="B79">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hazra</surname> <given-names>A.</given-names></name> <name><surname>Gogtay</surname> <given-names>N.</given-names></name></person-group> (<year>2016</year>). <article-title>Biostatistics series module 4: comparing groups &#x02013; categorical variables</article-title>. <source>Indian J. Dermatol</source>. <volume>61</volume>, <fpage>385</fpage>&#x02013;<lpage>392</lpage>. <pub-id pub-id-type="doi">10.4103/0019-5154.185700</pub-id><pub-id pub-id-type="pmid">27512183</pub-id></citation>
</ref>
<ref id="B80">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hill</surname> <given-names>N. T. M.</given-names></name> <name><surname>Mowszowski</surname> <given-names>L.</given-names></name> <name><surname>Naismith</surname> <given-names>S. L.</given-names></name> <name><surname>Chadwick</surname> <given-names>V. L.</given-names></name> <name><surname>Valenzuela</surname> <given-names>M.</given-names></name> <name><surname>Lampit</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>Computerized cognitive training in older adults with mild cognitive impairment or dementia: a systematic review and meta-analysis</article-title>. <source>Am. J. Psychiatry</source> <volume>174</volume>, <fpage>329</fpage>&#x02013;<lpage>340</lpage>. <pub-id pub-id-type="doi">10.1176/appi.ajp.2016.16030360</pub-id><pub-id pub-id-type="pmid">27838936</pub-id></citation>
</ref>
<ref id="B81">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Holwerda</surname> <given-names>T. J.</given-names></name> <name><surname>Deeg</surname> <given-names>D. J. H.</given-names></name> <name><surname>Beekman</surname> <given-names>A. T. F.</given-names></name> <name><surname>Tilburg</surname> <given-names>T. G.</given-names></name> <name><surname>van Stek</surname> <given-names>M. L.</given-names></name> <name><surname>Jonker</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2014</year>). <article-title>Feelings of loneliness, but not social isolation, predict dementia onset: results from the Amsterdam Study of the Elderly (AMSTEL)</article-title>. <source>J. Neurol. Neurosurg. Psychiatry</source> <volume>85</volume>, <fpage>135</fpage>&#x02013;<lpage>142</lpage>. <pub-id pub-id-type="doi">10.1136/jnnp-2012-302755</pub-id><pub-id pub-id-type="pmid">23232034</pub-id></citation>
</ref>
<ref id="B82">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hovy</surname> <given-names>D.</given-names></name> <name><surname>Prabhumoye</surname> <given-names>S.</given-names></name></person-group> (<year>2021</year>). <article-title>Five sources of bias in natural language processing</article-title>. <source>Lang. Linguist. Compass</source> <volume>15</volume>:<fpage>e12432</fpage>. <pub-id pub-id-type="doi">10.1111/lnc3.12432</pub-id><pub-id pub-id-type="pmid">35864931</pub-id></citation>
</ref>
<ref id="B83">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hristidis</surname> <given-names>V.</given-names></name> <name><surname>Ruggiano</surname> <given-names>N.</given-names></name> <name><surname>Brown</surname> <given-names>E. L.</given-names></name> <name><surname>Ganta</surname> <given-names>S. R. R.</given-names></name> <name><surname>Stewart</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT vs google for queries related to dementia and other cognitive decline: comparison of results</article-title>. <source>J. Med. Internet Res</source>. <volume>25</volume>:<fpage>e48966</fpage>. <pub-id pub-id-type="doi">10.2196/48966</pub-id><pub-id pub-id-type="pmid">37490317</pub-id></citation>
</ref>
<ref id="B84">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Hsiao</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <source>Bard Updates from Google I/O 2023: Images, New Features [WWW Document].</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/ai/google-bard-updates-io-2023/">https://blog.google/technology/ai/google-bard-updates-io-2023/</ext-link> (accessed 21 October, 2023).</citation>
</ref>
<ref id="B85">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT sets record for fastest-growing user base - analyst note</article-title>, in <source>Reuters</source> (<publisher-loc>Toronto, ON</publisher-loc>). Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/">https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/</ext-link></citation>
</ref>
<ref id="B86">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>J.</given-names></name> <name><surname>Chang</surname> <given-names>K. C.-C.</given-names></name></person-group> (<year>2023</year>). <article-title>Towards reasoning in large language models: a survey</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2212.10403</pub-id></citation>
</ref>
<ref id="B87">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>Y.</given-names></name> <name><surname>Gupta</surname> <given-names>S.</given-names></name> <name><surname>Xia</surname> <given-names>M.</given-names></name> <name><surname>Li</surname> <given-names>K.</given-names></name> <name><surname>Chen</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>Catastrophic jailbreak of open-source LLMs via exploiting generation</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2310.06987</pub-id></citation>
</ref>
<ref id="B88">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hugo</surname> <given-names>J.</given-names></name> <name><surname>Ganguli</surname> <given-names>M.</given-names></name></person-group> (<year>2014</year>). <article-title>Dementia and cognitive impairment: epidemiology, diagnosis, and treatment</article-title>. <source>Clin. Geriatr. Med</source>. <volume>30</volume>, <fpage>421</fpage>&#x02013;<lpage>442</lpage>. <pub-id pub-id-type="doi">10.1016/j.cger.2014.04.001</pub-id><pub-id pub-id-type="pmid">25037289</pub-id></citation>
</ref>
<ref id="B89">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Iadecola</surname> <given-names>C.</given-names></name> <name><surname>Duering</surname> <given-names>M.</given-names></name> <name><surname>Hachinski</surname> <given-names>V.</given-names></name> <name><surname>Joutel</surname> <given-names>A.</given-names></name> <name><surname>Pendlebury</surname> <given-names>S. T.</given-names></name> <name><surname>Schneider</surname> <given-names>J. A.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Vascular cognitive impairment and dementia</article-title>. <source>J. Am. Coll. Cardiol.</source> <volume>73</volume>, <fpage>3326</fpage>&#x02013;<lpage>3344</lpage>. <pub-id pub-id-type="doi">10.1016/j.jacc.2019.04.034</pub-id><pub-id pub-id-type="pmid">31248555</pub-id></citation>
</ref>
<ref id="B90">
<citation citation-type="web"><person-group person-group-type="author"><collab>IBM Data and AI Team</collab></person-group> (<year>2023</year>). <source>Open Source Large Language Models: Benefits, Risks and Types [WWW Document].</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.ibm.com/blog/open-source-large-language-models-benefits-risks-and-types/www.ibm.com/blog/open-source-large-language-models-benefits-risks-and-types">https://www.ibm.com/blog/open-source-large-language-models-benefits-risks-and-types/www.ibm.com/blog/open-source-large-language-models-benefits-risks-and-types</ext-link> (accessed October 12, 2023).</citation>
</ref>
<ref id="B91">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jack</surname> <given-names>C. R.</given-names></name> <name><surname>Bennett</surname> <given-names>D. A.</given-names></name> <name><surname>Blennow</surname> <given-names>K.</given-names></name> <name><surname>Carrillo</surname> <given-names>M. C.</given-names></name> <name><surname>Dunn</surname> <given-names>B.</given-names></name> <name><surname>Haeberlein</surname> <given-names>S. B.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>NIA-AA research framework: toward a biological definition of Alzheimer&#x00027;s disease</article-title>. <source>Alzheimers Dement. J. Alzheimers Assoc</source>. <volume>14</volume>, <fpage>535</fpage>&#x02013;<lpage>562</lpage>. <pub-id pub-id-type="doi">10.1016/j.jalz.2018.02.018</pub-id></citation>
</ref>
<ref id="B92">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Javaheripi</surname> <given-names>M.</given-names></name> <name><surname>Bubeck</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>Phi-2: the surprising power of small language models</article-title>, in <source>Microsoft Res</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/">https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/</ext-link> (accessed December 20, 2023).</citation>
</ref>
<ref id="B93">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Javaid</surname> <given-names>M.</given-names></name> <name><surname>Haleem</surname> <given-names>A.</given-names></name> <name><surname>Singh</surname> <given-names>R. P.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT for healthcare services: an emerging stage for an innovative perspective. BenchCouncil Trans</article-title>. <source>Benchmarks Stand. Eval</source>. 3, 100105. <pub-id pub-id-type="doi">10.1016/j.tbench.2023.100105</pub-id></citation>
</ref>
<ref id="B94">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>F.</given-names></name> <name><surname>Xu</surname> <given-names>Z.</given-names></name> <name><surname>Niu</surname> <given-names>L.</given-names></name> <name><surname>Xiang</surname> <given-names>Z.</given-names></name> <name><surname>Ramasubramanian</surname> <given-names>B.</given-names></name> <name><surname>Li</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>ArtPrompt: ASCII art-based jailbreak attacks against aligned LLMs</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2402.11753</pub-id></citation>
</ref>
<ref id="B95">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>S.</given-names></name> <name><surname>Kadhe</surname> <given-names>S.</given-names></name> <name><surname>Zhou</surname> <given-names>Y.</given-names></name> <name><surname>Cai</surname> <given-names>L.</given-names></name> <name><surname>Baracaldo</surname> <given-names>N.</given-names></name></person-group> (<year>2023</year>). <article-title>Forcing generative models to degenerate ones: the power of data poisoning attacks. presented at the neurips 2023 workshop on backdoors in deep learning - the good, the bad, and the ugly</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2312.04748</pub-id></citation>
</ref>
<ref id="B96">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>Z.</given-names></name> <name><surname>Araki</surname> <given-names>J.</given-names></name> <name><surname>Ding</surname> <given-names>H.</given-names></name> <name><surname>Neubig</surname> <given-names>G.</given-names></name></person-group> (<year>2021</year>). <article-title>How can we know when language models know? on the calibration of language models for question answering</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2012.00955</pub-id></citation>
</ref>
<ref id="B97">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiao</surname> <given-names>B.</given-names></name> <name><surname>Li</surname> <given-names>R.</given-names></name> <name><surname>Zhou</surname> <given-names>H.</given-names></name> <name><surname>Qing</surname> <given-names>K.</given-names></name> <name><surname>Liu</surname> <given-names>H.</given-names></name> <name><surname>Pan</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Neural biomarker diagnosis and prediction to mild cognitive impairment and Alzheimer&#x00027;s disease using EEG technology</article-title>. <source>Alzheimers Res. Ther</source>. <volume>15</volume>:<fpage>32</fpage>. <pub-id pub-id-type="doi">10.1186/s13195-023-01181-1</pub-id><pub-id pub-id-type="pmid">36765411</pub-id></citation>
</ref>
<ref id="B98">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kalai</surname> <given-names>A. T.</given-names></name> <name><surname>Vempala</surname> <given-names>S. S.</given-names></name></person-group> (<year>2023</year>). <source>Calibrated Language Models Must Hallucinate</source>.</citation>
</ref>
<ref id="B99">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kane</surname> <given-names>A. E.</given-names></name> <name><surname>Shin</surname> <given-names>S.</given-names></name> <name><surname>Wong</surname> <given-names>A. A.</given-names></name> <name><surname>Fertan</surname> <given-names>E.</given-names></name> <name><surname>Faustova</surname> <given-names>N. S.</given-names></name> <name><surname>Howlett</surname> <given-names>S. E.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Sex differences in healthspan predict lifespan in the 3xTg-AD mouse model of Alzheimer&#x00027;s disease</article-title>. <source>Front. Aging Neurosci</source>. <volume>10</volume>:<fpage>172</fpage>. <pub-id pub-id-type="doi">10.3389/fnagi.2018.00172</pub-id><pub-id pub-id-type="pmid">29946252</pub-id></citation>
</ref>
<ref id="B100">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Kanwal</surname> <given-names>N.</given-names></name> <name><surname>Rizzo</surname> <given-names>G.</given-names></name></person-group> (<year>2022</year>). <article-title>Attention-based clinical note summarization</article-title>, in <source>Proceedings of the 37th ACM/SIGAPP Symposium on Applied Computing, SAC&#x00027;22</source> (<publisher-loc>New York</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>813</fpage>&#x02013;<lpage>820</lpage>. <pub-id pub-id-type="doi">10.1145/3477314.3507256</pub-id></citation>
</ref>
<ref id="B101">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kapasi</surname> <given-names>A.</given-names></name> <name><surname>DeCarli</surname> <given-names>C.</given-names></name> <name><surname>Schneider</surname> <given-names>J. A.</given-names></name></person-group> (<year>2017</year>). <article-title>Impact of multiple pathologies on the threshold for clinically overt dementia</article-title>. <source>Acta Neuropathol</source>. <volume>134</volume>, <fpage>171</fpage>&#x02013;<lpage>186</lpage>. <pub-id pub-id-type="doi">10.1007/s00401-017-1717-7</pub-id><pub-id pub-id-type="pmid">28488154</pub-id></citation>
</ref>
<ref id="B102">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Karamolegkou</surname> <given-names>A.</given-names></name> <name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Zhou</surname> <given-names>L.</given-names></name> <name><surname>S&#x000F8;gaard</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Copyright violations and large language models</article-title>, in <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source> (<publisher-loc>Singapore</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>7403</fpage>&#x02013;<lpage>7412</lpage>. <pub-id pub-id-type="doi">10.18653/v1/2023.emnlp-main.458</pub-id></citation>
</ref>
<ref id="B103">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kasneci</surname> <given-names>E.</given-names></name> <name><surname>Sessler</surname> <given-names>K.</given-names></name> <name><surname>K&#x000FC;chemann</surname> <given-names>S.</given-names></name> <name><surname>Bannert</surname> <given-names>M.</given-names></name> <name><surname>Dementieva</surname> <given-names>D.</given-names></name> <name><surname>Fischer</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>ChatGPT for good? On opportunities and challenges of large language models for education</article-title>. <source>Learn. Individ. Differ</source>. <volume>103</volume>:<fpage>102274</fpage>. <pub-id pub-id-type="doi">10.1016/j.lindif.2023.102274</pub-id></citation>
</ref>
<ref id="B104">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Khan</surname> <given-names>A.</given-names></name> <name><surname>Bleth</surname> <given-names>A.</given-names></name> <name><surname>Bakpayev</surname> <given-names>M.</given-names></name> <name><surname>Imtiaz</surname> <given-names>N.</given-names></name></person-group> (<year>2022</year>). <article-title>Reminiscence therapy in the treatment of depression in the elderly: current perspectives</article-title>. <source>J. Ageing Longev</source>. <volume>2</volume>, <fpage>34</fpage>&#x02013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.3390/jal2010004</pub-id></citation>
</ref>
<ref id="B105">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Khurana</surname> <given-names>D.</given-names></name> <name><surname>Koli</surname> <given-names>A.</given-names></name> <name><surname>Khatter</surname> <given-names>K.</given-names></name> <name><surname>Singh</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>Natural language processing: state of the art, current trends and challenges</article-title>. <source>Multimed. Tools Appl</source>. <volume>82</volume>, <fpage>3713</fpage>&#x02013;<lpage>3744</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-022-13428-4</pub-id><pub-id pub-id-type="pmid">35855771</pub-id></citation>
</ref>
<ref id="B106">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Koebel</surname> <given-names>K.</given-names></name> <name><surname>Lacayo</surname> <given-names>M.</given-names></name> <name><surname>Murali</surname> <given-names>M.</given-names></name> <name><surname>Tarnanas</surname> <given-names>I.</given-names></name> <name><surname>&#x000C7;&#x000F6;ltekin</surname> <given-names>A.</given-names></name></person-group> (<year>2022</year>). <article-title>Expert insights for designing conversational user interfaces as virtual assistants and companions for older adults with cognitive impairments</article-title>, in <source>Chatbot Research and Design, Lecture Notes in Computer Science</source>, eds. A. F&#x000F8;lstad, T. Araujo, S. Papadopoulos, E. Law, E. Luger, M. Goodwin, P. B. Brandtzaeg (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>23</fpage>&#x02013;<lpage>38</lpage>.</citation>
</ref>
<ref id="B107">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Koga</surname> <given-names>S.</given-names></name> <name><surname>Martin</surname> <given-names>N. B.</given-names></name> <name><surname>Dickson</surname> <given-names>D. W.</given-names></name></person-group> (<year>2024</year>). <article-title>Evaluating the performance of large language models: ChatGPT and Google Bard in generating differential diagnoses in clinicopathological conferences of neurodegenerative disorders</article-title>. <source>Brain Pathol</source>. <volume>34</volume>:<fpage>e13207</fpage>. <pub-id pub-id-type="doi">10.1111/bpa.13207</pub-id><pub-id pub-id-type="pmid">37553205</pub-id></citation>
</ref>
<ref id="B108">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kowe</surname> <given-names>A.</given-names></name> <name><surname>K&#x000F6;hler</surname> <given-names>S.</given-names></name> <name><surname>G&#x000F6;r&#x000DF;</surname> <given-names>D.</given-names></name> <name><surname>Teipel</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>The patients&#x00027; and caregivers&#x00027; perspective: In-hospital navigation aids for people with dementia- a qualitative study with a value sensitive design approach</article-title>. <source>Assist. Technol</source>. <volume>35</volume>, <fpage>248</fpage>&#x02013;<lpage>257</lpage>. <pub-id pub-id-type="doi">10.1080/10400435.2021.2020378</pub-id><pub-id pub-id-type="pmid">34919023</pub-id></citation>
</ref>
<ref id="B109">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Krein</surname> <given-names>L.</given-names></name> <name><surname>Jeon</surname> <given-names>Y.-H.</given-names></name> <name><surname>Amberber</surname> <given-names>A. M.</given-names></name> <name><surname>Fethney</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>The assessment of language and communication in dementia: a synthesis of evidence</article-title>. <source>Am. J. Geriatr. Psychiatry</source> <volume>27</volume>, <fpage>363</fpage>&#x02013;<lpage>377</lpage>. <pub-id pub-id-type="doi">10.1016/j.jagp.2018.11.009</pub-id><pub-id pub-id-type="pmid">30581140</pub-id></citation>
</ref>
<ref id="B110">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kuzma</surname> <given-names>E.</given-names></name> <name><surname>Littlejohns</surname> <given-names>T. J.</given-names></name> <name><surname>Khawaja</surname> <given-names>A. P.</given-names></name> <name><surname>Llewellyn</surname> <given-names>D. J.</given-names></name> <name><surname>Ukoumunne</surname> <given-names>O. C.</given-names></name> <name><surname>Thiem</surname> <given-names>U.</given-names></name></person-group> (<year>2021</year>). <article-title>Visual impairment, eye diseases, and dementia risk: a systematic review and meta-analysis</article-title>. <source>J. Alzheimers Dis</source>. <volume>83</volume>, <fpage>1073</fpage>&#x02013;<lpage>1087</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-210250</pub-id><pub-id pub-id-type="pmid">34397414</pub-id></citation>
</ref>
<ref id="B111">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lane</surname> <given-names>C. A.</given-names></name> <name><surname>Hardy</surname> <given-names>J.</given-names></name> <name><surname>Schott</surname> <given-names>J. M.</given-names></name></person-group> (<year>2018</year>). <article-title>Alzheimer&#x00027;s disease</article-title>. <source>Eur. J. Neurol</source>. <volume>25</volume>, <fpage>59</fpage>&#x02013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1111/ene.13439</pub-id><pub-id pub-id-type="pmid">28872215</pub-id></citation>
</ref>
<ref id="B112">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lauscher</surname> <given-names>A.</given-names></name> <name><surname>Lueken</surname> <given-names>T.</given-names></name> <name><surname>Glava&#x00161;</surname> <given-names>G.</given-names></name></person-group> (<year>2021</year>). <article-title>Sustainable modular debiasing of language models</article-title>, in <source>Findings of the Association for Computational Linguistics: EMNLP 2021 Presented at the Findings 2021</source>, eds. M. F. Moens, X. Huang, L. Specia, S. W. Yih (Punta Cana: Association for Computational Linguistics), <fpage>4782</fpage>&#x02013;<lpage>4797</lpage>.</citation>
</ref>
<ref id="B113">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>E. E.</given-names></name> <name><surname>Torous</surname> <given-names>J.</given-names></name> <name><surname>De Choudhury</surname> <given-names>M.</given-names></name> <name><surname>Depp</surname> <given-names>C. A.</given-names></name> <name><surname>Graham</surname> <given-names>S. A.</given-names></name> <name><surname>Kim</surname> <given-names>H.-C.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Artificial intelligence for mental health care: clinical applications, barriers, facilitators, and artificial wisdom</article-title>. <source>Biol. Psychiatry Cogn. Neurosci. Neuroimaging</source> <volume>6</volume>, <fpage>856</fpage>&#x02013;<lpage>864</lpage>. <pub-id pub-id-type="doi">10.1016/j.bpsc.2021.02.001</pub-id><pub-id pub-id-type="pmid">33571718</pub-id></citation>
</ref>
<ref id="B114">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Lehman</surname> <given-names>E.</given-names></name> <name><surname>Hernandez</surname> <given-names>E.</given-names></name> <name><surname>Mahajan</surname> <given-names>D.</given-names></name> <name><surname>Wulff</surname> <given-names>J.</given-names></name> <name><surname>Smith</surname> <given-names>M. J.</given-names></name> <name><surname>Ziegler</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Do we still need clinical language models?, in: proceedings of the conference on health, inference, and learning</article-title>, in <source>Presented at the Conference on Health, Inference, and Learning</source> (<publisher-loc>New York</publisher-loc>: <publisher-name>PMLR</publisher-name>), <fpage>578</fpage>&#x02013;<lpage>597</lpage>.</citation>
</ref>
<ref id="B115">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>R.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Lawler</surname> <given-names>K.</given-names></name> <name><surname>Garg</surname> <given-names>S.</given-names></name> <name><surname>Bai</surname> <given-names>Q.</given-names></name> <name><surname>Alty</surname> <given-names>J.</given-names></name></person-group> (<year>2022</year>). <article-title>Applications of artificial intelligence to aid early detection of dementia: a scoping review on current capabilities and future directions</article-title>. <source>J. Biomed. Inform</source>. <volume>127</volume>:<fpage>104030</fpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2022.104030</pub-id><pub-id pub-id-type="pmid">35183766</pub-id></citation>
</ref>
<ref id="B116">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>R.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Yu</surname> <given-names>H.</given-names></name></person-group> (<year>2023</year>). <article-title>Two directions for clinical data generation with large language models: data-to-label and label-to-data</article-title>, in <source>Proc. Conf. Empir. Methods Nat. Lang. Process. Conf. Empir. Methods Nat. Lang. Process</source>, <fpage>7129</fpage>&#x02013;<lpage>7143</lpage>.</citation>
</ref>
<ref id="B117">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Fairness of ChatGPT</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2305.18569</pub-id></citation>
</ref>
<ref id="B118">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Liao</surname> <given-names>S. M.</given-names></name></person-group> (<year>2020</year>). <source>Ethics of Artificial Intelligence</source>. <publisher-loc>Oxford</publisher-loc>: <publisher-name>Oxford University Press</publisher-name>.</citation>
</ref>
<ref id="B119">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lieber</surname> <given-names>O.</given-names></name> <name><surname>Lenz</surname> <given-names>B.</given-names></name> <name><surname>Bata</surname> <given-names>H.</given-names></name> <name><surname>Cohen</surname> <given-names>G.</given-names></name> <name><surname>Osin</surname> <given-names>J.</given-names></name> <name><surname>Dalmedigos</surname> <given-names>I.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Jamba: a hybrid transformer-mamba language model</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2403.19887</pub-id></citation>
</ref>
<ref id="B120">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lima</surname> <given-names>M. R.</given-names></name> <name><surname>Wairagkar</surname> <given-names>M.</given-names></name> <name><surname>Gupta</surname> <given-names>M.</given-names></name> <name><surname>Rodriguez y Baena</surname> <given-names>F.</given-names></name> <name><surname>Barnaghi</surname> <given-names>P.</given-names></name> <name><surname>Sharp</surname> <given-names>D. J.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Conversational affective social robots for ageing and dementia support</article-title>. <source>IEEE Trans. Cogn. Dev. Syst</source>. <volume>14</volume>, <fpage>1378</fpage>&#x02013;<lpage>1397</lpage>. <pub-id pub-id-type="doi">10.1109/TCDS.2021.3115228</pub-id></citation>
</ref>
<ref id="B121">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>S.</given-names></name> <name><surname>Hilton</surname> <given-names>J.</given-names></name> <name><surname>Evans</surname> <given-names>O.</given-names></name></person-group> (<year>2022</year>). <article-title>TruthfulQA: measuring how models mimic human falsehoods</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2109.07958</pub-id></citation>
</ref>
<ref id="B122">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lingard</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>Writing with ChatGPT: an illustration of its capacity, limitations and implications for academic writers</article-title>. <source>Perspect. Med. Educ</source>. <volume>12</volume>, <fpage>261</fpage>&#x02013;<lpage>270</lpage>. <pub-id pub-id-type="doi">10.5334/pme.1072</pub-id><pub-id pub-id-type="pmid">37397181</pub-id></citation>
</ref>
<ref id="B123">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>N. F.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name> <name><surname>Liang</surname> <given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>Evaluating verifiability in generative search engines</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2304.09848</pub-id></citation>
</ref>
<ref id="B124">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Deng</surname> <given-names>G.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2023</year>). <source>Prompt Injection Attack Against LLM-Integrated Applications</source>.</citation>
</ref>
<ref id="B125">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Livingston</surname> <given-names>G.</given-names></name> <name><surname>Huntley</surname> <given-names>J.</given-names></name> <name><surname>Sommerlad</surname> <given-names>A.</given-names></name> <name><surname>Ames</surname> <given-names>D.</given-names></name> <name><surname>Ballard</surname> <given-names>C.</given-names></name> <name><surname>Banerjee</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Dementia prevention, intervention, and care: 2020 report of the Lancet Commission</article-title>. <source>The Lancet</source> <volume>396</volume>, <fpage>413</fpage>&#x02013;<lpage>446</lpage>. <pub-id pub-id-type="doi">10.1016/S0140-6736(20)30367-6</pub-id></citation>
</ref>
<ref id="B126">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Loconte</surname> <given-names>R.</given-names></name> <name><surname>Orr&#x000F9;</surname> <given-names>G.</given-names></name> <name><surname>Tribastone</surname> <given-names>M.</given-names></name> <name><surname>Pietrini</surname> <given-names>P.</given-names></name> <name><surname>Sartori</surname> <given-names>G.</given-names></name></person-group> (<year>2023</year>). <source>Challenging ChatGPT &#x0201C;Intelligence&#x0201D; with Human Tools: A Neuropsychological Investigation on Prefrontal Functioning of a Large Language Model.</source></citation>
</ref>
<ref id="B127">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lombardi</surname> <given-names>A.</given-names></name> <name><surname>Diacono</surname> <given-names>D.</given-names></name> <name><surname>Amoroso</surname> <given-names>N.</given-names></name> <name><surname>Biecek</surname> <given-names>P.</given-names></name> <name><surname>Monaco</surname> <given-names>A.</given-names></name> <name><surname>Bellantuono</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>A robust framework to investigate the reliability and stability of explainable artificial intelligence markers of Mild Cognitive Impairment and Alzheimer&#x00027;s Disease</article-title>. <source>Brain Inform</source>. 9, 17. <pub-id pub-id-type="doi">10.1186/s40708-022-00165-5</pub-id></citation>
</ref>
<ref id="B128">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lund</surname> <given-names>B. D.</given-names></name> <name><surname>Wang</surname> <given-names>T.</given-names></name> <name><surname>Mannuru</surname> <given-names>N. R.</given-names></name> <name><surname>Nie</surname> <given-names>B.</given-names></name> <name><surname>Shimray</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT and a new academic reality: artificial intelligence-written research papers and the ethics of the large language models in scholarly publishing</article-title>. <source>J. Assoc. Inf. Sci. Technol</source>. <volume>74</volume>, <fpage>570</fpage>&#x02013;<lpage>581</lpage>. <pub-id pub-id-type="doi">10.1002/asi.24750</pub-id></citation>
</ref>
<ref id="B129">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>Z.</given-names></name> <name><surname>Mei</surname> <given-names>Y.</given-names></name> <name><surname>Su</surname> <given-names>Z.</given-names></name></person-group> (<year>2024</year>). <article-title>Understanding the benefits and challenges of using large language model-based conversational agents for mental well-being support</article-title>. <source>AMIA Annu. Symp. Proc</source>. <volume>2023</volume>, <fpage>1105</fpage>&#x02013;<lpage>1114</lpage>. <pub-id pub-id-type="pmid">38222348</pub-id></citation>
</ref>
<ref id="B130">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mao</surname> <given-names>C.</given-names></name> <name><surname>Xu</surname> <given-names>J.</given-names></name> <name><surname>Rasmussen</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Adekkanattu</surname> <given-names>P.</given-names></name> <name><surname>Pacheco</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>AD-BERT: Using pre-trained language model to predict the progression from mild cognitive impairment to Alzheimer&#x00027;s disease</article-title>. <source>J. Biomed. Inform</source>. 144, 104442. <pub-id pub-id-type="doi">10.1016/j.jbi.2023.104442</pub-id></citation>
</ref>
<ref id="B131">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Mauran</surname> <given-names>C.</given-names></name></person-group> (<year>2023</year>). <article-title>Google I/O 2023 unveils PaLM 2 large language model [WWW Document]</article-title>, in <source>Mashable</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://mashable.com/article/google-io-2023-palm2-ai-announcement">https://mashable.com/article/google-io-2023-palm2-ai-announcement</ext-link> (accessed 23 September, 2023).</citation>
</ref>
<ref id="B132">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>McKenzie</surname> <given-names>I.</given-names></name> <name><surname>Lyzhov</surname> <given-names>A.</given-names></name> <name><surname>Parrish</surname> <given-names>A.</given-names></name> <name><surname>Prabhu</surname> <given-names>A.</given-names></name> <name><surname>Mueller</surname> <given-names>A.</given-names></name> <name><surname>Kim</surname> <given-names>N.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Inverse-scaling/prize</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2306.09479</pub-id></citation>
</ref>
<ref id="B133">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>McKinzie</surname> <given-names>B.</given-names></name> <name><surname>Gan</surname> <given-names>Z.</given-names></name> <name><surname>Fauconnier</surname> <given-names>J.-P.</given-names></name> <name><surname>Dodge</surname> <given-names>S.</given-names></name> <name><surname>Zhang</surname> <given-names>B.</given-names></name> <name><surname>Dufter</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>MM1: methods, analysis and insights from multimodal LLM pre-training</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2403.09611</pub-id></citation>
</ref>
<ref id="B134">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Merkin</surname> <given-names>A.</given-names></name> <name><surname>Krishnamurthi</surname> <given-names>R.</given-names></name> <name><surname>Medvedev</surname> <given-names>O. N.</given-names></name></person-group> (<year>2022</year>). <article-title>Machine learning, artificial intelligence and the prediction of dementia</article-title>. <source>Curr. Opin. Psychiatry</source> <volume>35</volume>, <fpage>123</fpage>&#x02013;<lpage>129</lpage>. <pub-id pub-id-type="doi">10.1097/YCO.0000000000000768</pub-id><pub-id pub-id-type="pmid">34861656</pub-id></citation>
</ref>
<ref id="B135">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mesk&#x000F3;</surname> <given-names>B.</given-names></name> <name><surname>Topol</surname> <given-names>E. J.</given-names></name></person-group> (<year>2023</year>). <article-title>The imperative for regulatory oversight of large language models (or generative AI) in healthcare</article-title>. <source>NPJ Digit. Med</source>. <volume>6</volume>, <fpage>1</fpage>&#x02013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1038/s41746-023-00873-0</pub-id><pub-id pub-id-type="pmid">36596833</pub-id></citation>
</ref>
<ref id="B136">
<citation citation-type="web"><person-group person-group-type="author"><collab>Meta</collab></person-group> (<year>2023a</year>). <source>Introducing LLaMA: A Foundational, 65-Billion-Parameter Language Model</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://ai.meta.com/blog/large-language-model-llama-meta-ai/">https://ai.meta.com/blog/large-language-model-llama-meta-ai/</ext-link> (accessed 21 October, 2023).</citation>
</ref>
<ref id="B137">
<citation citation-type="web"><person-group person-group-type="author"><collab>Meta</collab></person-group> (<year>2023b</year>). <article-title>Meta and Microsoft Introduce the Next Generation of Llama</article-title>, in <source>Meta AI</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://ai.meta.com/blog/llama-2/">https://ai.meta.com/blog/llama-2/</ext-link> (accessed 20 December, 2023).</citation>
</ref>
<ref id="B138">
<citation citation-type="web"><person-group person-group-type="author"><collab>Meta</collab></person-group> (<year>2024</year>). <article-title>Introducing Meta Llama 3: The most capable openly available LLM to date</article-title>, in <source>Meta AI.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://ai.meta.com/blog/meta-llama-3/">https://ai.meta.com/blog/meta-llama-3/</ext-link> (accessed 19 April, 2024).</citation>
</ref>
<ref id="B139">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Min</surname> <given-names>B.</given-names></name> <name><surname>Ross</surname> <given-names>H.</given-names></name> <name><surname>Sulem</surname> <given-names>E.</given-names></name> <name><surname>Veyseh</surname> <given-names>A. P. B.</given-names></name> <name><surname>Nguyen</surname> <given-names>T. H.</given-names></name> <name><surname>Sainz</surname> <given-names>O.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Recent advances in natural language processing via large pre-trained language models: a survey</article-title>. <source>ACM Comput. Surv.</source> 56, <volume>30</volume>:<fpage>1</fpage>&#x02013;<lpage>30</lpage>:40. <pub-id pub-id-type="doi">10.1145/3605943</pub-id></citation>
</ref>
<ref id="B140">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miotto</surname> <given-names>M.</given-names></name> <name><surname>Rossberg</surname> <given-names>N.</given-names></name> <name><surname>Kleinberg</surname> <given-names>B.</given-names></name></person-group> (<year>2022</year>). <article-title>Who is GPT-3? An exploration of personality, values and demographics</article-title>, in <source>Proceedings of the Fifth Workshop on Natural Language Processing and Computational Social Science (NLP</source>&#x0002B;<italic>CSS)</italic> (Abu Dhabi: Association for Computational Linguistics), <fpage>218</fpage>&#x02013;<lpage>227</lpage>.</citation>
</ref>
<ref id="B141">
<citation citation-type="web"><person-group person-group-type="author"><collab>Mistral AI</collab></person-group> (<year>2023a</year>). <article-title>Mistral 7B</article-title>, in <source>Mistral AI</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://mistral.ai/news/announcing-mistral-7b/">https://mistral.ai/news/announcing-mistral-7b/</ext-link> (accessed 20 December, 2023).</citation>
</ref>
<ref id="B142">
<citation citation-type="web"><person-group person-group-type="author"><collab>Mistral AI</collab></person-group> (<year>2023b</year>). <article-title>Mixtral of experts</article-title>, in <source>Mistral AI.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://mistral.ai/news/mixtral-of-experts/">https://mistral.ai/news/mixtral-of-experts/</ext-link> (accessed 20 December, 2023).</citation>
</ref>
<ref id="B143">
<citation citation-type="web"><person-group person-group-type="author"><collab>Mistral AI</collab></person-group> (<year>2024</year>). <article-title>Au Large</article-title>, in <source>Mistral AI</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://mistral.ai/news/mistral-large/">https://mistral.ai/news/mistral-large/</ext-link> (accessed 17 March, 2024).</citation>
</ref>
<ref id="B144">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mitchell</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <article-title>How do we know how smart AI systems are?</article-title> <source>Science</source> <volume>381</volume>:<fpage>adj5957</fpage>. <pub-id pub-id-type="doi">10.1126/science.adj5957</pub-id><pub-id pub-id-type="pmid">37440662</pub-id></citation>
</ref>
<ref id="B145">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mitchell</surname> <given-names>M.</given-names></name> <name><surname>Krakauer</surname> <given-names>D. C.</given-names></name></person-group> (<year>2023</year>). <article-title>The debate over understanding in AI&#x00027;s large language models</article-title>. <source>Proc. Natl. Acad. Sci</source>. <volume>120</volume>:<fpage>e2215907120</fpage>. <pub-id pub-id-type="doi">10.1073/pnas.2215907120</pub-id><pub-id pub-id-type="pmid">36943882</pub-id></citation>
</ref>
<ref id="B146">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mitra</surname> <given-names>A.</given-names></name> <name><surname>Del Corro</surname> <given-names>L.</given-names></name> <name><surname>Mahajan</surname> <given-names>S.</given-names></name> <name><surname>Codas</surname> <given-names>A.</given-names></name> <name><surname>Simoes</surname> <given-names>C.</given-names></name> <name><surname>Agarwal</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Orca 2: teaching small language models how to reason</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2311.11045</pub-id></citation>
</ref>
<ref id="B147">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Mo</surname> <given-names>Y.</given-names></name> <name><surname>Baptista</surname> <given-names>E.</given-names></name></person-group> (<year>2023</year>). <source>China&#x00027;s Baidu unveils new Ernie AI version to rival GPT-4</source>. <publisher-loc>London</publisher-loc>: <publisher-name>Reuters</publisher-name>.</citation>
</ref>
<ref id="B148">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Mondini</surname> <given-names>S.</given-names></name> <name><surname>Madella</surname> <given-names>I.</given-names></name> <name><surname>Zangrossi</surname> <given-names>A.</given-names></name> <name><surname>Bigolin</surname> <given-names>A.</given-names></name> <name><surname>Tomasi</surname> <given-names>C.</given-names></name> <name><surname>Michieletto</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Cognitive reserve in dementia: implications for cognitive training</article-title>. <source>Front. Aging Neurosci</source>. 8.84. <pub-id pub-id-type="doi">10.3389/fnagi.2016.00084</pub-id></citation>
</ref>
<ref id="B149">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Morales-de-Jes&#x000FA;s</surname> <given-names>V.</given-names></name> <name><surname>G&#x000F3;mez-Adorno</surname> <given-names>H.</given-names></name> <name><surname>Somodevilla-Garc&#x000ED;a</surname> <given-names>M.</given-names></name> <name><surname>Vilari&#x000F1;o</surname> <given-names>D.</given-names></name></person-group> (<year>2021</year>). <article-title>Conversational System as assistant tool in reminiscence therapy for people with early-stage of Alzheimer&#x00027;s</article-title>. <source>Healthcare</source> <volume>9</volume>:<fpage>1036</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare9081036</pub-id><pub-id pub-id-type="pmid">34442173</pub-id></citation>
</ref>
<ref id="B150">
<citation citation-type="web"><person-group person-group-type="author"><collab>Mosaic AI Research Team</collab></person-group> (<year>2024</year>). <article-title>Introducing DBRX: a new state-of-the-art open LLM</article-title>, in <source>Mosaic AI Res.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm">https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm</ext-link> (accessed June 4, 2024).</citation>
</ref>
<ref id="B151">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murdoch</surname> <given-names>B. E.</given-names></name> <name><surname>Chenery</surname> <given-names>H. J.</given-names></name> <name><surname>Wilks</surname> <given-names>V.</given-names></name> <name><surname>Boyle</surname> <given-names>R. S.</given-names></name></person-group> (<year>1987</year>). <article-title>Language disorders in dementia of the Alzheimer type</article-title>. <source>Brain Lang</source>. <volume>31</volume>, <fpage>122</fpage>&#x02013;<lpage>137</lpage>. <pub-id pub-id-type="doi">10.1016/0093-934X(87)90064-2</pub-id><pub-id pub-id-type="pmid">2437993</pub-id></citation>
</ref>
<ref id="B152">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murley</surname> <given-names>A. G.</given-names></name> <name><surname>Coyle-Gilchrist</surname> <given-names>I.</given-names></name> <name><surname>Rouse</surname> <given-names>M. A.</given-names></name> <name><surname>Jones</surname> <given-names>P. S.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Wiggins</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Redefining the multidimensional clinical phenotypes of frontotemporal lobar degeneration syndromes</article-title>. <source>Brain</source> <volume>143</volume>, <fpage>1555</fpage>&#x02013;<lpage>1571</lpage>. <pub-id pub-id-type="doi">10.1093/brain/awaa097</pub-id><pub-id pub-id-type="pmid">32438414</pub-id></citation>
</ref>
<ref id="B153">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nandi</surname> <given-names>A.</given-names></name> <name><surname>Counts</surname> <given-names>N.</given-names></name> <name><surname>Chen</surname> <given-names>S.</given-names></name> <name><surname>Seligman</surname> <given-names>B.</given-names></name> <name><surname>Tortorice</surname> <given-names>D.</given-names></name> <name><surname>Vigo</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Global and regional projections of the economic burden of Alzheimer&#x00027;s disease and related dementias from 2019 to 2050: A value of statistical life approach</article-title>. <source>eClinicalMed.</source> <volume>51</volume>, <fpage>1</fpage>&#x02013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1016/j.eclinm.2022.101580</pub-id></citation>
</ref>
<ref id="B154">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nasr</surname> <given-names>M.</given-names></name> <name><surname>Carlini</surname> <given-names>N.</given-names></name> <name><surname>Hayase</surname> <given-names>J.</given-names></name> <name><surname>Jagielski</surname> <given-names>M.</given-names></name> <name><surname>Cooper</surname> <given-names>A. F.</given-names></name> <name><surname>Ippolito</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Scalable extraction of training data from (production) language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2311.17035</pub-id></citation>
</ref>
<ref id="B155">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Naveed</surname> <given-names>H.</given-names></name> <name><surname>Khan</surname> <given-names>A. U.</given-names></name> <name><surname>Qiu</surname> <given-names>S.</given-names></name> <name><surname>Saqib</surname> <given-names>M.</given-names></name> <name><surname>Anwar</surname> <given-names>S.</given-names></name> <name><surname>Usman</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>A comprehensive overview of large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2307.06435</pub-id></citation>
</ref>
<ref id="B156">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nguyen</surname> <given-names>T.</given-names></name> <name><surname>Li</surname> <given-names>X.</given-names></name></person-group> (<year>2020</year>). <article-title>Understanding public-stigma and self-stigma in the context of dementia: a systematic review of the global literature</article-title>. <source>Dementia</source> <volume>19</volume>, <fpage>148</fpage>&#x02013;<lpage>181</lpage>. <pub-id pub-id-type="doi">10.1177/1471301218800122</pub-id><pub-id pub-id-type="pmid">31920117</pub-id></citation>
</ref>
<ref id="B157">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nichols</surname> <given-names>E.</given-names></name> <name><surname>Steinmetz</surname> <given-names>J. D.</given-names></name> <name><surname>Vollset</surname> <given-names>S. E.</given-names></name> <name><surname>Fukutaki</surname> <given-names>K.</given-names></name> <name><surname>Chalek</surname> <given-names>J.</given-names></name> <name><surname>Abd-Allah</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Estimation of the global prevalence of dementia in 2019 and forecasted prevalence in 2050: an analysis for the Global Burden of Disease Study 2019</article-title>. <source>Lancet Public Health</source> <volume>7</volume>, <fpage>e105</fpage>&#x02013;<lpage>e125</lpage>. <pub-id pub-id-type="doi">10.1016/S2468-2667(21)00249-8</pub-id><pub-id pub-id-type="pmid">34998485</pub-id></citation>
</ref>
<ref id="B158">
<citation citation-type="web"><person-group person-group-type="author"><collab>Norwegian Consumer Council</collab></person-group> (<year>2023</year>). <source>Ghost in the Machine - Addressing the Consumer Harms of Generative AI</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://storage02.forbrukerradet.no/media/2023/06/generative-ai-rapport-2023.pdf">https://storage02.forbrukerradet.no/media/2023/06/generative-ai-rapport-2023.pdf</ext-link> (accessed February 22, 2024).</citation>
</ref>
<ref id="B159">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oh</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>A study on the case of using ChatGPT and learners&#x00027; perceptions in college liberal arts writing</article-title>. <source>Korean J. Gen. Educ</source>. <volume>17</volume>, <fpage>11</fpage>&#x02013;<lpage>23</lpage>. <pub-id pub-id-type="doi">10.46392/kjge.2023.17.3.11</pub-id></citation>
</ref>
<ref id="B160">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Olney</surname> <given-names>N. T.</given-names></name> <name><surname>Spina</surname> <given-names>S.</given-names></name> <name><surname>Miller</surname> <given-names>B. L.</given-names></name></person-group> (<year>2017</year>). <article-title>Frontotemporal dementia</article-title>. <source>Neurol. Clin</source>. <volume>35</volume>, <fpage>339</fpage>&#x02013;<lpage>374</lpage>. <pub-id pub-id-type="doi">10.1016/j.ncl.2017.01.008</pub-id><pub-id pub-id-type="pmid">28410663</pub-id></citation>
</ref>
<ref id="B161">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>O&#x00027;Malley</surname> <given-names>R. P. D.</given-names></name> <name><surname>Mirheidari</surname> <given-names>B.</given-names></name> <name><surname>Harkness</surname> <given-names>K.</given-names></name> <name><surname>Reuber</surname> <given-names>M.</given-names></name> <name><surname>Venneri</surname> <given-names>A.</given-names></name> <name><surname>Walker</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Fully automated cognitive screening tool based on assessment of speech and language</article-title>. <source>J. Neurol. Neurosurg. Psychiatry</source> 20:jnnp-2019-322517. <pub-id pub-id-type="doi">10.1002/alz.041980</pub-id></citation>
</ref>
<ref id="B162">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Omarov</surname> <given-names>B.</given-names></name> <name><surname>Zhumanov</surname> <given-names>Z.</given-names></name> <name><surname>Gumar</surname> <given-names>A.</given-names></name> <name><surname>Kuntunova</surname> <given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>Artificial intelligence enabled mobile chatbot psychologist using AIML and cognitive behavioral therapy</article-title>. <source>Int. J. Adv. Comput. Sci. Appl</source>. <volume>14</volume>:<fpage>616</fpage>. <pub-id pub-id-type="doi">10.14569/IJACSA.2023.0140616</pub-id></citation>
</ref>
<ref id="B163">
<citation citation-type="web"><person-group person-group-type="author"><collab>OpenAI</collab></person-group> (<year>2022</year>). <source>Introducing ChatGPT [WWW Document].</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://openai.com/blog/chatgpt">https://openai.com/blog/chatgpt</ext-link> (accessed 16 December, 2023).</citation>
</ref>
<ref id="B164">
<citation citation-type="journal"><person-group person-group-type="author"><collab>OpenAI</collab></person-group> (<year>2023</year>). <article-title>GPT-4 technical report</article-title>. <source>arXiv [Preprint]</source>. arXiv: 2303.08774v6.</citation>
</ref>
<ref id="B165">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Oscar</surname> <given-names>N.</given-names></name> <name><surname>Fox</surname> <given-names>P. A.</given-names></name> <name><surname>Croucher</surname> <given-names>R.</given-names></name> <name><surname>Wernick</surname> <given-names>R.</given-names></name> <name><surname>Keune</surname> <given-names>J.</given-names></name> <name><surname>Hooker</surname> <given-names>K.</given-names></name></person-group> (<year>2017</year>). <article-title>Machine learning, sentiment analysis, and tweets: an examination of Alzheimer&#x00027;s disease stigma on Twitter</article-title>. <source>J. Gerontol. Ser. B</source> <volume>72</volume>, <fpage>742</fpage>&#x02013;<lpage>751</lpage>. <pub-id pub-id-type="doi">10.1093/geronb/gbx014</pub-id></citation>
</ref>
<ref id="B166">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Ouyang</surname> <given-names>L.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Jiang</surname> <given-names>X.</given-names></name> <name><surname>Almeida</surname> <given-names>D.</given-names></name> <name><surname>Wainwright</surname> <given-names>C.</given-names></name> <name><surname>Mishkin</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Training language models to follow instructions with human feedback</article-title>. <source>Adv. Neural Inf. Process. Syst</source>. <volume>35</volume>, <fpage>27730</fpage>&#x02013;<lpage>27744</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://proceedings.neurips.cc/paper_files/paper/2022/file/b1efde53be364a73914f58805a001731-Paper-Conference.pdf">https://proceedings.neurips.cc/paper_files/paper/2022/file/b1efde53be364a73914f58805a001731-Paper-Conference.pdf</ext-link></citation>
</ref>
<ref id="B167">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pal</surname> <given-names>A.</given-names></name> <name><surname>Umapathi</surname> <given-names>L. K.</given-names></name> <name><surname>Sankarasubbu</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <source>Med-HALT: Medical Domain Hallucination Test for Large Language Models</source>.</citation>
</ref>
<ref id="B168">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>S. H.</given-names></name></person-group> (<year>2023</year>). <article-title>Use of generative artificial intelligence, including large language models such as ChatGPT, in scientific publications: policies of KJR and prominent authorities</article-title>. <source>Korean J. Radiol</source>. <volume>24</volume>, <fpage>715</fpage>&#x02013;<lpage>718</lpage>. <pub-id pub-id-type="doi">10.3348/kjr.2023.0643</pub-id><pub-id pub-id-type="pmid">37500572</pub-id></citation>
</ref>
<ref id="B169">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Parra</surname> <given-names>M. A.</given-names></name> <name><surname>Butler</surname> <given-names>S.</given-names></name> <name><surname>McGeown</surname> <given-names>W. J.</given-names></name> <name><surname>Brown Nicholls</surname> <given-names>L. A.</given-names></name> <name><surname>Robertson</surname> <given-names>D. J.</given-names></name></person-group> (<year>2019</year>). <article-title>Globalising strategies to meet global challenges: the case of ageing and dementia</article-title>. <source>J. Glob. Health</source> <volume>9</volume>:<fpage>020310</fpage>. <pub-id pub-id-type="doi">10.7189/jogh.09.020310</pub-id><pub-id pub-id-type="pmid">31777656</pub-id></citation>
</ref>
<ref id="B170">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Patel</surname> <given-names>F.</given-names></name> <name><surname>Thakore</surname> <given-names>R.</given-names></name> <name><surname>Nandwani</surname> <given-names>I.</given-names></name> <name><surname>Bharti</surname> <given-names>S. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Combating depression in students using an intelligent chatbot: a cognitive behavioral therapy</article-title>, in <source>2019 IEEE 16th India Council International Conference (INDICON)</source> (<publisher-loc>Rajkot</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>4</lpage>.</citation>
</ref>
<ref id="B171">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Paulus</surname> <given-names>J. K.</given-names></name> <name><surname>Kent</surname> <given-names>D. M.</given-names></name></person-group> (<year>2020</year>). <article-title>Predictably unequal: understanding and addressing concerns that algorithmic clinical prediction may increase health disparities</article-title>. <source>NPJ Digit. Med</source>. <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1038/s41746-020-0304-9</pub-id><pub-id pub-id-type="pmid">31934645</pub-id></citation>
</ref>
<ref id="B172">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Pichai</surname> <given-names>S.</given-names></name> <name><surname>Hassabis</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>Introducing Gemini: Google&#x00027;s most capable AI model yet</article-title>, in <source>The Keyword</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/ai/google-gemini-ai/">https://blog.google/technology/ai/google-gemini-ai/</ext-link> (accessed December 17, 2023).</citation>
</ref>
<ref id="B173">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Pichai</surname> <given-names>S.</given-names></name> <name><surname>Hassabis</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Our next-generation model: Gemini 1.5</article-title>, in <source>The Keyword</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/">https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/</ext-link> (accessed February 22, 2024).</citation>
</ref>
<ref id="B174">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pillette</surname> <given-names>L.</given-names></name> <name><surname>Moreau</surname> <given-names>G.</given-names></name> <name><surname>Normand</surname> <given-names>J.-M.</given-names></name> <name><surname>Perrier</surname> <given-names>M.</given-names></name> <name><surname>L&#x000E9;cuyer</surname> <given-names>A.</given-names></name> <name><surname>Cogn&#x000E9;</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <article-title>A systematic review of navigation assistance systems for people with dementia</article-title>. <source>IEEE Trans. Vis. Comput. Graph</source>. <volume>29</volume>, <fpage>2146</fpage>&#x02013;<lpage>2165</lpage>. <pub-id pub-id-type="doi">10.1109/TVCG.2022.3141383</pub-id><pub-id pub-id-type="pmid">35007194</pub-id></citation>
</ref>
<ref id="B175">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pl&#x000E1;cido</surname> <given-names>J.</given-names></name> <name><surname>de Almeida</surname> <given-names>C. A. B.</given-names></name> <name><surname>Ferreira</surname> <given-names>J. V.</given-names></name> <name><surname>de Oliveira Silva</surname> <given-names>F.</given-names></name> <name><surname>Monteiro-Junior</surname> <given-names>R. S.</given-names></name> <name><surname>Tangen</surname> <given-names>G. G.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Spatial navigation in older adults with mild cognitive impairment and dementia: a systematic review and meta-analysis</article-title>. <source>Exp. Gerontol</source>. <volume>165</volume>:<fpage>111852</fpage>. <pub-id pub-id-type="doi">10.1016/j.exger.2022.111852</pub-id><pub-id pub-id-type="pmid">35644416</pub-id></citation>
</ref>
<ref id="B176">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Prince</surname> <given-names>M.</given-names></name> <name><surname>Wimo</surname> <given-names>A.</given-names></name> <name><surname>Guerchet</surname> <given-names>A.</given-names></name> <name><surname>Ali</surname> <given-names>G.-C.</given-names></name> <name><surname>Wu</surname> <given-names>Y.-T.</given-names></name> <name><surname>Prina</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <source>World Alzheimer Report 2015 The Global Impact of Dementia: An Analysis of Prevalence, Incidence, Cost and Trends, Alzheimer&#x00027;s Disease International</source>. <publisher-loc>London</publisher-loc>: <publisher-name>Alzheimer&#x00027;s Disease International (ADI)</publisher-name>.</citation>
</ref>
<ref id="B177">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Qi</surname> <given-names>P.</given-names></name> <name><surname>Wu</surname> <given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>ChatGPT: a promising tool to combat social isolation and loneliness in older adults with mild cognitive impairment</article-title>. <source>NeurologyLive</source> 6. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.neurologylive.com/view/chatgpt-promising-tool-combat-social-isolation-loneliness-older-adults-mild-cognitive-impairment">https://www.neurologylive.com/view/chatgpt-promising-tool-combat-social-isolation-loneliness-older-adults-mild-cognitive-impairment</ext-link></citation>
</ref>
<ref id="B178">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Qiu</surname> <given-names>S.</given-names></name> <name><surname>Miller</surname> <given-names>M. I.</given-names></name> <name><surname>Joshi</surname> <given-names>P. S.</given-names></name> <name><surname>Lee</surname> <given-names>J. C.</given-names></name> <name><surname>Xue</surname> <given-names>C.</given-names></name> <name><surname>Ni</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Multimodal deep learning for Alzheimer&#x00027;s disease dementia assessment</article-title>. <source>Nat. Commun</source>. <volume>13</volume>:<fpage>3404</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-022-31037-5</pub-id><pub-id pub-id-type="pmid">35725739</pub-id></citation>
</ref>
<ref id="B179">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Raffaele</surname> <given-names>F.</given-names></name> <name><surname>Claudia</surname> <given-names>M.</given-names></name> <name><surname>John</surname> <given-names>H.</given-names></name></person-group> (<year>2019</year>). <article-title>Genetics and molecular mechanisms of frontotemporal lobar degeneration: an update and future avenues</article-title>. <source>Neurobiol. Aging</source> <volume>78</volume>, <fpage>98</fpage>&#x02013;<lpage>110</lpage>. <pub-id pub-id-type="doi">10.1016/j.neurobiolaging.2019.02.006</pub-id><pub-id pub-id-type="pmid">30925302</pub-id></citation>
</ref>
<ref id="B180">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Raghavan</surname> <given-names>M.</given-names></name> <name><surname>Barocas</surname> <given-names>S.</given-names></name> <name><surname>Kleinberg</surname> <given-names>J.</given-names></name> <name><surname>Levy</surname> <given-names>K.</given-names></name></person-group> (<year>2020</year>). <article-title>Mitigating bias in algorithmic hiring: evaluating claims and practices</article-title>, in <source>Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, FAT<sup>&#x0002A;</sup>&#x00027;20</source> (<publisher-loc>New York</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>469</fpage>&#x02013;<lpage>481</lpage>.</citation>
</ref>
<ref id="B181">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rai</surname> <given-names>H. K.</given-names></name> <name><surname>Kernaghan</surname> <given-names>D.</given-names></name> <name><surname>Schoonmade</surname> <given-names>L.</given-names></name> <name><surname>Egan</surname> <given-names>K. J.</given-names></name> <name><surname>Pot</surname> <given-names>A. M.</given-names></name></person-group> (<year>2022</year>). <article-title>Digital technologies to prevent social isolation and loneliness in dementia: a systematic review</article-title>. <source>J. Alzheimers Dis</source>. <volume>90</volume>, <fpage>513</fpage>&#x02013;<lpage>528</lpage>. <pub-id pub-id-type="doi">10.3233/JAD-220438</pub-id><pub-id pub-id-type="pmid">36120780</pub-id></citation>
</ref>
<ref id="B182">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Raile</surname> <given-names>P.</given-names></name></person-group> (<year>2024</year>). <article-title>The usefulness of ChatGPT for psychotherapists and patients</article-title>. <source>Humanit. Soc. Sci. Commun</source>. <volume>11</volume>, <fpage>1</fpage>&#x02013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1057/s41599-023-02567-0</pub-id></citation>
</ref>
<ref id="B183">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Raji</surname> <given-names>I. D.</given-names></name> <name><surname>Buolamwini</surname> <given-names>J.</given-names></name></person-group> (<year>2019</year>). <article-title>Actionable auditing: investigating the impact of publicly naming biased performance results of commercial AI products</article-title>, in <source>Proceedings of the (2019). AAAI/ACM Conference on AI, Ethics, and Society, AIES&#x00027;19</source> (<publisher-loc>New York</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>429</fpage>&#x02013;<lpage>435</lpage>.</citation>
</ref>
<ref id="B184">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Reisner</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>These 183,000 Books Are Fueling the Biggest Fight in Publishing and Tech</article-title>, in <source>The Atlantic</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.theatlantic.com/technology/archive/2023/09/books3-database-generative-ai-training-copyright-infringement/675363/">https://www.theatlantic.com/technology/archive/2023/09/books3-database-generative-ai-training-copyright-infringement/675363/</ext-link> (accessed 25 November, 2023).</citation>
</ref>
<ref id="B185">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Richardson</surname> <given-names>A.</given-names></name> <name><surname>Robbins</surname> <given-names>C. B.</given-names></name> <name><surname>Wisely</surname> <given-names>C. E.</given-names></name> <name><surname>Henao</surname> <given-names>R.</given-names></name> <name><surname>Grewal</surname> <given-names>D. S.</given-names></name> <name><surname>Fekrat</surname> <given-names>S.</given-names></name></person-group> (<year>2022</year>). <article-title>Artificial intelligence in dementia</article-title>. <source>Curr. Opin. Ophthalmol</source>. <volume>33</volume>, <fpage>425</fpage>&#x02013;<lpage>431</lpage>. <pub-id pub-id-type="doi">10.1097/ICU.0000000000000881</pub-id><pub-id pub-id-type="pmid">35916570</pub-id></citation>
</ref>
<ref id="B186">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rozado</surname> <given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>The political biases of ChatGPT</article-title>. <source>Soc. Sci</source>. 12, 148. <pub-id pub-id-type="doi">10.3390/socsci12030148</pub-id></citation>
</ref>
<ref id="B187">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ryu</surname> <given-names>H.</given-names></name> <name><surname>Kim</surname> <given-names>S.</given-names></name> <name><surname>Kim</surname> <given-names>D.</given-names></name> <name><surname>Han</surname> <given-names>S.</given-names></name> <name><surname>Lee</surname> <given-names>K.</given-names></name> <name><surname>Kang</surname> <given-names>Y.</given-names></name></person-group> (<year>2020</year>). <article-title>Simple and steady interactions win the healthy mentality: designing a chatbot service for the elderly</article-title>. <source>Proc. ACM Hum.-Comput. Interact.</source> 4, <volume>152</volume>:<fpage>1</fpage>&#x02013;<lpage>152</lpage>:25. <pub-id pub-id-type="doi">10.1145/3415223</pub-id></citation>
</ref>
<ref id="B188">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Saeidnia</surname> <given-names>H. R.</given-names></name> <name><surname>Kozak</surname> <given-names>M.</given-names></name> <name><surname>Lund</surname> <given-names>B. D.</given-names></name> <name><surname>Hassanzadeh</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <source>Evaluation of ChatGPT&#x00027;s Responses to Information Needs and Information Seeking of Dementia Patients</source>.</citation>
</ref>
<ref id="B189">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schneider</surname> <given-names>J. A.</given-names></name> <name><surname>Arvanitakis</surname> <given-names>Z.</given-names></name> <name><surname>Bang</surname> <given-names>W.</given-names></name> <name><surname>Bennett</surname> <given-names>D. A.</given-names></name></person-group> (<year>2007</year>). <article-title>Mixed brain pathologies account for most dementia cases in community-dwelling older persons</article-title>. <source>Neurology</source> <volume>69</volume>, <fpage>2197</fpage>&#x02013;<lpage>2204</lpage>. <pub-id pub-id-type="doi">10.1212/01.wnl.0000271090.28148.24</pub-id><pub-id pub-id-type="pmid">17568013</pub-id></citation>
</ref>
<ref id="B190">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname> <given-names>N.</given-names></name> <name><surname>Chakrabarti</surname> <given-names>S.</given-names></name> <name><surname>Grover</surname> <given-names>S.</given-names></name></person-group> (<year>2016</year>). <article-title>Gender differences in caregiving among family - caregivers of people with mental illnesses</article-title>. <source>World J. Psychiatry</source> <volume>6</volume>, <fpage>7</fpage>&#x02013;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.5498/wjp.v6.i1.7</pub-id><pub-id pub-id-type="pmid">27014594</pub-id></citation>
</ref>
<ref id="B191">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sheehan</surname> <given-names>B.</given-names></name></person-group> (<year>2012</year>). <article-title>Assessment scales in dementia</article-title>. <source>Ther. Adv. Neurol. Disord</source>. 5, 349. <pub-id pub-id-type="doi">10.1177/1756285612455733</pub-id></citation>
</ref>
<ref id="B192">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Singhal</surname> <given-names>K.</given-names></name> <name><surname>Azizi</surname> <given-names>S.</given-names></name> <name><surname>Tu</surname> <given-names>T.</given-names></name> <name><surname>Mahdavi</surname> <given-names>S. S.</given-names></name> <name><surname>Wei</surname> <given-names>J.</given-names></name> <name><surname>Chung</surname> <given-names>H. W.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Large language models encode clinical knowledge</article-title>. <source>Nature</source> <volume>620</volume>, <fpage>172</fpage>&#x02013;<lpage>180</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-023-06291-2</pub-id><pub-id pub-id-type="pmid">37438534</pub-id></citation>
</ref>
<ref id="B193">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sorin</surname> <given-names>V.</given-names></name> <name><surname>Brin</surname> <given-names>D.</given-names></name> <name><surname>Barash</surname> <given-names>Y.</given-names></name> <name><surname>Konen</surname> <given-names>E.</given-names></name> <name><surname>Charney</surname> <given-names>A.</given-names></name> <name><surname>Nadkarni</surname> <given-names>G.</given-names></name> <etal/></person-group>. (<year>2023</year>). <source>Large Language Models (LLMs) and Empathy &#x02013; A Systematic Review</source>.</citation>
</ref>
<ref id="B194">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stamate</surname> <given-names>D.</given-names></name> <name><surname>Smith</surname> <given-names>R.</given-names></name> <name><surname>Tsygancov</surname> <given-names>R.</given-names></name> <name><surname>Vorobev</surname> <given-names>R.</given-names></name> <name><surname>Langham</surname> <given-names>J.</given-names></name> <name><surname>Stahl</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Applying deep learning to predicting dementia and mild cognitive impairment</article-title>. <source>Artif. Intell. Appl. Innov</source>. <volume>584</volume>, <fpage>308</fpage>&#x02013;<lpage>319</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-49186-4_26</pub-id></citation>
</ref>
<ref id="B195">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Sundjaja</surname> <given-names>J. H.</given-names></name> <name><surname>Shrestha</surname> <given-names>R.</given-names></name> <name><surname>Krishan</surname> <given-names>K.</given-names></name></person-group> (<year>2024</year>). <source>McNemar And Mann-Whitney U Tests.</source> <publisher-loc>Treasure Island</publisher-loc>: <publisher-name>StatPearls Publishing</publisher-name>.</citation>
</ref>
<ref id="B196">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tay</surname> <given-names>K.-W.</given-names></name> <name><surname>Subramaniam</surname> <given-names>P.</given-names></name> <name><surname>Oei</surname> <given-names>T. P.</given-names></name></person-group> (<year>2019</year>). <article-title>Cognitive behavioural therapy can be effective in treating anxiety and depression in persons with dementia: a systematic review</article-title>. <source>Psychogeriatrics</source> <volume>19</volume>, <fpage>264</fpage>&#x02013;<lpage>275</lpage>. <pub-id pub-id-type="doi">10.1111/psyg.12391</pub-id><pub-id pub-id-type="pmid">30548731</pub-id></citation>
</ref>
<ref id="B197">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tian</surname> <given-names>S.</given-names></name> <name><surname>Jin</surname> <given-names>Q.</given-names></name> <name><surname>Yeganova</surname> <given-names>L.</given-names></name> <name><surname>Lai</surname> <given-names>P.-T.</given-names></name> <name><surname>Zhu</surname> <given-names>Q.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Opportunities and challenges for ChatGPT and large language models in biomedicine and health</article-title>. <source>Brief. Bioinform.</source> <volume>25</volume>:<fpage>bbad493</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbad493</pub-id></citation>
</ref>
<ref id="B198">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Toth</surname> <given-names>L.</given-names></name> <name><surname>Hoffmann</surname> <given-names>I.</given-names></name> <name><surname>Gosztolya</surname> <given-names>G.</given-names></name> <name><surname>Vincze</surname> <given-names>V.</given-names></name> <name><surname>Szatloczki</surname> <given-names>G.</given-names></name> <name><surname>Banreti</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>A speech recognition-based solution for the automatic detection of mild cognitive impairment from spontaneous speech</article-title>. <source>Curr. Alzheimer Res</source>. <volume>15</volume>, <fpage>130</fpage>&#x02013;<lpage>138</lpage>. <pub-id pub-id-type="doi">10.2174/1567205014666171121114930</pub-id><pub-id pub-id-type="pmid">29165085</pub-id></citation>
</ref>
<ref id="B199">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Touvron</surname> <given-names>H.</given-names></name> <name><surname>Lavril</surname> <given-names>T.</given-names></name> <name><surname>Izacard</surname> <given-names>G.</given-names></name> <name><surname>Martinet</surname> <given-names>X.</given-names></name> <name><surname>Lachaux</surname> <given-names>M.-A.</given-names></name> <name><surname>Lacroix</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>LLaMA: open and efficient foundation language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2302.13971</pub-id></citation>
</ref>
<ref id="B200">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Treder</surname> <given-names>M. S.</given-names></name> <name><surname>Shock</surname> <given-names>J. P.</given-names></name> <name><surname>Stein</surname> <given-names>D. J.</given-names></name> <name><surname>du Plessis</surname> <given-names>S.</given-names></name> <name><surname>Seedat</surname> <given-names>S.</given-names></name> <name><surname>Tsvetanov</surname> <given-names>K. A.</given-names></name></person-group> (<year>2021</year>). <article-title>Correlation constraints for regression models: controlling bias in brain age prediction</article-title>. <source>Front. Psychiatry</source> <volume>12</volume>:<fpage>615754</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyt.2021.615754</pub-id><pub-id pub-id-type="pmid">33679476</pub-id></citation>
</ref>
<ref id="B201">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tsoi</surname> <given-names>K. K. F.</given-names></name> <name><surname>Jia</surname> <given-names>P.</given-names></name> <name><surname>Dowling</surname> <given-names>N. M.</given-names></name> <name><surname>Titiner</surname> <given-names>J. R.</given-names></name> <name><surname>Wagner</surname> <given-names>M.</given-names></name> <name><surname>Capuano</surname> <given-names>A. W.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Applications of artificial intelligence in dementia research</article-title>. <source>Camb. Prisms Precis. Med</source>. <volume>1</volume>:<fpage>e9</fpage>. <pub-id pub-id-type="doi">10.1017/pcm.2022.10</pub-id></citation>
</ref>
<ref id="B202">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vaidyam</surname> <given-names>A. N.</given-names></name> <name><surname>Wisniewski</surname> <given-names>H.</given-names></name> <name><surname>Halamka</surname> <given-names>J. D.</given-names></name> <name><surname>Kashavan</surname> <given-names>M. S.</given-names></name> <name><surname>Torous</surname> <given-names>J. B.</given-names></name></person-group> (<year>2019</year>). <article-title>Chatbots and conversational agents in mental health: a review of the psychiatric landscape</article-title>. <source>Can. J. Psychiatry</source> <volume>64</volume>, <fpage>456</fpage>&#x02013;<lpage>464</lpage>. <pub-id pub-id-type="doi">10.1177/0706743719828977</pub-id><pub-id pub-id-type="pmid">30897957</pub-id></citation>
</ref>
<ref id="B203">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Veen</surname> <given-names>D.</given-names></name> <name><surname>Van Uden</surname> <given-names>C.</given-names></name> <name><surname>Blankemeier</surname> <given-names>L.</given-names></name> <name><surname>Delbrouck</surname> <given-names>J.-B.</given-names></name> <name><surname>Aali</surname> <given-names>A.</given-names></name> <name><surname>Bluethgen</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Adapted large language models can outperform medical experts in clinical text summarization</article-title>. <source>Nat. Med</source>. <volume>30</volume>, <fpage>1134</fpage>&#x02013;<lpage>1142</lpage>. <pub-id pub-id-type="doi">10.1038/s41591-024-02855-5</pub-id><pub-id pub-id-type="pmid">38413730</pub-id></citation>
</ref>
<ref id="B204">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Van Veen</surname> <given-names>D.</given-names></name> <name><surname>Van Uden</surname> <given-names>C.</given-names></name> <name><surname>Blankemeier</surname> <given-names>L.</given-names></name> <name><surname>Delbrouck</surname> <given-names>J. B.</given-names></name> <name><surname>Aali</surname> <given-names>A.</given-names></name> <name><surname>Bluethgen</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Clinical text summarization: adapting large language models can outperform human experts</article-title>. <source>Res. Square</source> rs.3.rs-3483777. <pub-id pub-id-type="doi">10.21203/rs.3.rs-3483777/v1</pub-id></citation>
</ref>
<ref id="B205">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A.</given-names></name> <name><surname>Shazeer</surname> <given-names>N.</given-names></name> <name><surname>Parmar</surname> <given-names>N.</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J.</given-names></name> <name><surname>Jones</surname> <given-names>L.</given-names></name> <name><surname>Gomez</surname> <given-names>A. N.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Attention is all you need</article-title>, in <source>Advances in Neural Information Processing Systems</source> (<publisher-loc>Red Hook, NY</publisher-loc>: <publisher-name>Curran Associates, Inc</publisher-name>), <fpage>5998</fpage>&#x02013;<lpage>6008</lpage>.</citation>
</ref>
<ref id="B206">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Victor</surname> <given-names>C. R.</given-names></name></person-group> (<year>2021</year>). <article-title>Is loneliness a cause or consequence of dementia? A public health analysis of the literature</article-title>. <source>Front. Psychol</source>. <volume>11</volume>:<fpage>612771</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyg.2020.612771</pub-id><pub-id pub-id-type="pmid">33679498</pub-id></citation>
</ref>
<ref id="B207">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Vincent</surname> <given-names>J.</given-names></name></person-group> (<year>2023</year>). <article-title>Google and Microsoft&#x00027;s chatbots are already citing one another in a misinformation shitshow</article-title>, in <source>The Verge</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.theverge.com/2023/3/22/23651564/google-microsoft-bard-bing-chatbots-misinformation">https://www.theverge.com/2023/3/22/23651564/google-microsoft-bard-bing-chatbots-misinformation</ext-link> (accessed 28 November, 2023).</citation>
</ref>
<ref id="B208">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Vinod</surname> <given-names>P.</given-names></name> <name><surname>Safar</surname> <given-names>S.</given-names></name> <name><surname>Mathew</surname> <given-names>D.</given-names></name> <name><surname>Venugopal</surname> <given-names>P.</given-names></name> <name><surname>Joly</surname> <given-names>L. M.</given-names></name> <name><surname>George</surname> <given-names>J.</given-names></name></person-group> (<year>2020</year>). <article-title>Fine-tuning the BERTSUMEXT model for clinical report summarization</article-title>, in <source>2020 International Conference for Emerging Technology (INCET)</source> (<publisher-loc>Belgaum</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>7</lpage>.</citation>
</ref>
<ref id="B209">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>von Werra</surname> <given-names>L.</given-names></name> <name><surname>Belkada</surname> <given-names>Y.</given-names></name> <name><surname>Mangrulkar</surname> <given-names>S.</given-names></name> <name><surname>Tunstall</surname> <given-names>L.</given-names></name> <name><surname>Dehaene</surname> <given-names>O.</given-names></name> <name><surname>Cuenca</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>The Falcon has landed in the Hugging Face ecosystem</article-title>, in <source>Hugging Face Blog.</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://huggingface.co/blog/falcon">https://huggingface.co/blog/falcon</ext-link> (accessed 20 December, 2023).</citation>
</ref>
<ref id="B210">
<citation citation-type="web"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>B.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>Enabling conversational interaction on mobile with LLMs</article-title>, in <source>Google Res. Blog</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://blog.research.google/2023/05/enabling-conversational-interaction-on.html">https://blog.research.google/2023/05/enabling-conversational-interaction-on.html</ext-link> (accessed January 1, 2024).</citation>
</ref>
<ref id="B211">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Zhong</surname> <given-names>W.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Mi</surname> <given-names>F.</given-names></name> <name><surname>Zeng</surname> <given-names>X.</given-names></name> <name><surname>Huang</surname> <given-names>W.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Aligning large language models with human: a survey</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2307.12966</pub-id></citation>
</ref>
<ref id="B212">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>A.</given-names></name> <name><surname>Haghtalab</surname> <given-names>N.</given-names></name> <name><surname>Steinhardt</surname> <given-names>J.</given-names></name></person-group> (<year>2023</year>). <article-title>Jailbroken: how does llm safety training fail?</article-title> <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2307.02483</pub-id></citation>
</ref>
<ref id="B213">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Schuurmans</surname> <given-names>D.</given-names></name> <name><surname>Bosma</surname> <given-names>M.</given-names></name> <name><surname>Ichter</surname> <given-names>B.</given-names></name> <name><surname>Xia</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Chain-of-thought prompting elicits reasoning in large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2201.11903</pub-id></citation>
</ref>
<ref id="B214">
<citation citation-type="web"><person-group person-group-type="author"><collab>WHO</collab></person-group> (<year>2023</year>). <source>Dementia [WWW Document].</source> Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.who.int/news-room/fact-sheets/detail/dementia">https://www.who.int/news-room/fact-sheets/detail/dementia</ext-link> (accessed November 18, 2023).</citation>
</ref>
<ref id="B215">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>C.</given-names></name> <name><surname>Lin</surname> <given-names>W.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Xie</surname> <given-names>W.</given-names></name></person-group> (<year>2023</year>). <article-title>PMC-LLaMA: towards building open-source language models for medicine</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2304.14454</pub-id></citation>
</ref>
<ref id="B216">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Ouyang</surname> <given-names>L.</given-names></name> <name><surname>Ziegler</surname> <given-names>D. M.</given-names></name> <name><surname>Stiennon</surname> <given-names>N.</given-names></name> <name><surname>Lowe</surname> <given-names>R.</given-names></name> <name><surname>Leike</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <source>Recursively Summarizing Books with Human Feedback</source>.</citation>
</ref>
<ref id="B217">
<citation citation-type="web"><person-group person-group-type="author"><collab>xAI</collab></person-group> (<year>2024a</year>). <article-title>Open Release of Grok-1</article-title>, in <source>xAI Blog</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://x.ai/blog/grok-os">https://x.ai/blog/grok-os</ext-link> (accessed April 6, 2024).</citation>
</ref>
<ref id="B218">
<citation citation-type="web"><person-group person-group-type="author"><collab>xAI</collab></person-group> (<year>2024b</year>). <article-title>Announcing Grok-1.5</article-title>, in <source>xAI Blog</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://x.ai/blog/grok-1.5">https://x.ai/blog/grok-1.5</ext-link> (accessed April 6, 2024).</citation>
</ref>
<ref id="B219">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yao</surname> <given-names>S.</given-names></name> <name><surname>Yu</surname> <given-names>D.</given-names></name> <name><surname>Zhao</surname> <given-names>J.</given-names></name> <name><surname>Shafran</surname> <given-names>I.</given-names></name> <name><surname>Griffiths</surname> <given-names>T. L.</given-names></name> <name><surname>Cao</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Tree of thoughts: deliberate problem solving with large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2305.10601</pub-id></citation>
</ref>
<ref id="B220">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yasunaga</surname> <given-names>M.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Pasupat</surname> <given-names>P.</given-names></name> <name><surname>Leskovec</surname> <given-names>J.</given-names></name> <name><surname>Liang</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Large language models as analogical reasoners</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2310.01714</pub-id></citation>
</ref>
<ref id="B221">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ye</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>T.</given-names></name> <name><surname>Zhang</surname> <given-names>A.</given-names></name> <name><surname>Hua</surname> <given-names>W.</given-names></name> <name><surname>Jia</surname> <given-names>W.</given-names></name></person-group> (<year>2023</year>). <article-title>Cognitive mirage: a review of hallucinations in large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2309.06794</pub-id></citation>
</ref>
<ref id="B222">
<citation citation-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L.</given-names></name> <name><surname>Negrinho</surname> <given-names>R.</given-names></name> <name><surname>Ghosh</surname> <given-names>A.</given-names></name> <name><surname>Jagannathan</surname> <given-names>V.</given-names></name> <name><surname>Hassanzadeh</surname> <given-names>H. R.</given-names></name> <name><surname>Schaaf</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Leveraging pretrained models for automatic summarization of doctor-patient conversations</article-title>, in <source>Findings of the Association for Computational Linguistics: EMNLP 2021</source> (<publisher-loc>Punta Cana</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>), <fpage>3693</fpage>&#x02013;<lpage>3712</lpage>.</citation>
</ref>
<ref id="B223">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Dong</surname> <given-names>Z.</given-names></name> <name><surname>Phillips</surname> <given-names>P.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Ji</surname> <given-names>G.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>Detection of subjects and brain regions related to Alzheimer&#x00027;s disease using 3D MRI scans based on eigenbrain and machine learning</article-title>. <source>Front. Comput. Neurosci</source>. <volume>9</volume>:<fpage>66</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2015.00066</pub-id><pub-id pub-id-type="pmid">26082713</pub-id></citation>
</ref>
<ref id="B224">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Cui</surname> <given-names>L.</given-names></name> <name><surname>Cai</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>L.</given-names></name> <name><surname>Fu</surname> <given-names>T.</given-names></name> <name><surname>Huang</surname> <given-names>X.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Siren&#x00027;s song in the ai ocean: a survey on hallucination in large language models</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2309.01219</pub-id></citation>
</ref>
<ref id="B225">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>H.</given-names></name> <name><surname>Chen</surname> <given-names>H.</given-names></name> <name><surname>Yang</surname> <given-names>F.</given-names></name> <name><surname>Liu</surname> <given-names>N.</given-names></name> <name><surname>Deng</surname> <given-names>H.</given-names></name> <name><surname>Cai</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Explainability for large language models: a survey</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.2309.01029</pub-id></citation>
</ref>
<ref id="B226">
<citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ziegler</surname> <given-names>D. M.</given-names></name> <name><surname>Stiennon</surname> <given-names>N.</given-names></name> <name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Brown</surname> <given-names>T. B.</given-names></name> <name><surname>Radford</surname> <given-names>A.</given-names></name> <name><surname>Amodei</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Fine-tuning language models from human preferences</article-title>. <source>arXiv [Preprint]</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1909.08593</pub-id></citation>
</ref>
</ref-list>
</back>
</article>