<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="brief-report" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Public Health</journal-id>
<journal-title-group>
<journal-title>Frontiers in Public Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Public Health</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-2565</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpubh.2026.1603238</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Perspective</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Toward clinical integration of generative AI in mental health: personalization, multimodality and inter-entity experience</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Frisone</surname>
<given-names>Fabio</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3021428"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Pupillo</surname>
<given-names>Chiara</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3020972"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Rossi</surname>
<given-names>Chiara</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1649513"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Riva</surname>
<given-names>Giuseppe</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/126074"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Humane Technology Laboratory, Catholic University of the Sacred Heart</institution>, <city>Milan</city>, <country country="it">Italy</country></aff>
<aff id="aff2"><label>2</label><institution>Sophia University Institute</institution>, <city>Figline e Incisa Valdarno (Firenze)</city>, <country country="it">Italy</country></aff>
<aff id="aff3"><label>3</label><institution>Department of Computer Science, University of Pisa</institution>, <city>Pisa</city>, <country country="it">Italy</country></aff>
<aff id="aff4"><label>4</label><institution>Department of Human Sciences, Guglielmo Marconi University</institution>, <city>Rome</city>, <country country="it">Italy</country></aff>
<aff id="aff5"><label>5</label><institution>Applied Technology for Neuro-Psychology Laboratory, IRCCS Istituto Auxologico Italiano</institution>, <city>Milan</city>, <country country="it">Italy</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Chiara Pupillo, <email xlink:href="mailto:chiara.pupillo@phd.unipi.it">chiara.pupillo@phd.unipi.it</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-03-24">
<day>24</day>
<month>03</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>14</volume>
<elocation-id>1603238</elocation-id>
<history>
<date date-type="received">
<day>31</day>
<month>03</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>17</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>18</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Frisone, Pupillo, Rossi and Riva.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Frisone, Pupillo, Rossi and Riva</copyright-holder>
<license>
<ali:license_ref start_date="2026-03-24">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The integration of conversational agents (CAs) into mental health care presents a promising yet complex frontier. These systems have demonstrated potential in expanding access, enhancing user engagement and providing scalable support, particularly in contexts where human clinicians are unavailable. Despite these advantages, CAs face two fundamental limitations: their quality of interaction and the ability to facilitate self-reflection. This perspective article provides a theoretical framework with clinical relevance, identifying the critical challenges that must be addressed to transition Generative Artificial Intelligence (GenAI) from a conversational tool to an adjunctive psychological support. Firstly, personalization is essential. Advances in retrieval-augmented generation (RAG), fine-tuning, and emotional reasoning are necessary to enable context-aware and ethically grounded responses tailored to individual users. In addition, multimodal interaction (particularly through improvements in speech synthesis, prosody, and expressive dialogue) can help bridge the gap between human and AI communication, fostering greater emotional resonance and natural flow. Lastly, immersive environments, including embodied CA and virtual reality settings, may amplify presence, potentially engaging neural and psychological mechanisms typically associated with human-to-human interaction. These innovations must be accompanied by a strong ethical and regulatory foundation. Systems must ensure transparency, informed consent, and compliance with data privacy standards such as GDPR and HIPAA. Crucially, AI should not be viewed as a replacement for psychologists, but as an adaptive and supportive layer within a broader care ecosystem. By aligning technological capabilities with clinical intent, the future of GenAI in mental health may lie in its ability to complement human expertise and meaningfully extend psychological support.</p>
</abstract>
<kwd-group>
<kwd>clinical psychology</kwd>
<kwd>general psychology</kwd>
<kwd>generative artificial intelligence</kwd>
<kwd>mental health</kwd>
<kwd>psychological support</kwd>
<kwd>virtual reality</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the Italian Ministry of Health-Ricerca Corrente.</funding-statement>
</funding-group>
<counts>
<fig-count count="2"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="62"/>
<page-count count="10"/>
<word-count count="7870"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Public Mental Health</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<title>Introduction</title>
<p>The use of conversational agents (CAs) for adjunctive psychological support seems like an entirely new topic, but some recent research points out that this may already be something well underway (<xref ref-type="bibr" rid="ref1">1</xref>, <xref ref-type="bibr" rid="ref2">2</xref>), to the point that continued hesitation on the part of some clinical institutions about their use may be producing the effect of falling inexorably behind the continuing advances in digital health. As highlighted by Hatch et al. (<xref ref-type="bibr" rid="ref3">3</xref>), &#x201C;<italic>mental health experts find themselves in a precarious situation: we must quickly discern the possible destination [&#x2026;] of the AI-therapist train as it may have already left the station</italic>&#x201D; (p. 13). This suggestion, although it may come across as provocative to someone, shows that Generative Artificial Intelligence (GenAI) for adjunctive psychological support can break through scientific literature, making use of research and data that increasingly demonstrates its effectiveness. Indeed, the debate about the use of CAs to provide user assistance and psychological support is not entirely recent (<xref ref-type="bibr" rid="ref4">4</xref>, <xref ref-type="bibr" rid="ref5">5</xref>). In the mental health context, CAs have already been widely used to act as therapists, counselors, and facilitators (<xref ref-type="bibr" rid="ref6 ref7 ref8">6&#x2013;8</xref>). For example, several years ago, Cristea et al. (<xref ref-type="bibr" rid="ref9">9</xref>) showed that comparing face-to-face versus computer-based interventions could provide important insights. The authors focused on comparing a chatbot that simulated therapeutic interventions, the &#x201C;ELIZA&#x201D; program, with a trained human cognitive behavioral therapist (CBT). The research investigated aspects related to therapeutic performance, such as the effectiveness of the discussion, the correct approach to the problem brought by the patient, and the quality of the therapeutic relationship. The conclusions drawn from the research showed that most evaluators agreed that both therapists were regarded as human people, although with different skills. The only perceived difference between ELIZA and the CBT therapist seemed to lie in the quality of therapeutic performance, but not in any inherent characteristics of the two. Needless to say, things have changed a lot from 2013 to the present, especially in terms of technological advances, so to expect that at present the use of CAs can give better results than those obtained more than a decade ago does not seem to be a far-fetched assumption (<xref ref-type="bibr" rid="ref10">10</xref>, <xref ref-type="bibr" rid="ref11">11</xref>). Rapid advances in Large Language Models (LLMs) such as ChatGPT have generated interest in their potential for clinical applications, sparking both enthusiasm and skepticism. While AI-based interventions aim to improve accessibility and reduce costs, their clinical effectiveness is being tested and the results that have been obtained are opposing (<xref ref-type="bibr" rid="ref12">12</xref>, <xref ref-type="bibr" rid="ref13">13</xref>) but promising (<xref ref-type="bibr" rid="ref14 ref15 ref16">14&#x2013;16</xref>). We could say that today&#x2019;s question is not whether these tools work, but <italic>how</italic>, <italic>when</italic>, <italic>why,</italic> and <italic>for whom</italic> (<xref ref-type="bibr" rid="ref17">17</xref>). From a public mental health perspective, this also necessitates consideration of <italic>at what scale</italic> and <italic>within which service delivery models</italic> these technologies can be responsibly integrated. The global treatment gap in low- and middle-income countries remains substantial, with estimates exceeding 80% (<xref ref-type="bibr" rid="ref18">18</xref>), highlighting the relevance of scalable approaches within stepped-care frameworks, where lower-intensity supports precede or complement face-to-face psychological support. In this broader landscape, AI-enabled tools are sometimes discussed as one possible avenue to support reach and efficiency, though their role depends on context, regulation, and local capacity.</p>
<p>Recent studies have demonstrated that users struggle to distinguish whether responses come from a real clinical professional or a CA. Research shows that regarding common therapeutic factors (therapeutic alliance, empathy, expectations, cultural competence, and therapeutic technique), CAs consistently score higher than mental health professionals themselves (<xref ref-type="bibr" rid="ref3">3</xref>). However, further findings suggest caution when using CAs. When users are informed about the source of responses, an attribution bias emerges, revealing a technophobic attitude (<xref ref-type="bibr" rid="ref19">19</xref>). Users tend to attribute positively impressive responses to human therapists while assigning less valuable responses to CAs (<xref ref-type="bibr" rid="ref3">3</xref>). This phenomenon raises the most challenging question about CA use in psychological support: are we ready to accept help from a tool specifically designed to support us, rather than from someone motivated by professional vocation? In other words: is the recognition we receive from a CA sufficient to make us feel better (<xref ref-type="bibr" rid="ref20">20</xref>, <xref ref-type="bibr" rid="ref21">21</xref>)?</p>
<p>To address these questions, we must move beyond intuitive reactions and subjective biases. A scientific approach requires suspending preconceptions and critically examining evidence (<xref ref-type="bibr" rid="ref22">22</xref>). Rather than automatically rejecting or embracing AI as an adjunctive psychological support, we need to rigorously evaluate its capabilities, limitations, and potential for meaningful psychological engagement.</p>
<p>Although CAs show great promise in expanding access to psychological support, particularly through scalable and immediate assistance, they also have substantial limitations that require careful consideration. This particularly applies to general-purpose LLMs, such as ChatGPT or Gemini, which are trained on vast, unfiltered internet corpora without clinical supervision or structured psychological knowledge bases. These models, while linguistically fluent, can produce hallucinatory or misleading content and reflect implicit social or cultural biases, posing risks when applied in high-stakes psychological contexts (<xref ref-type="bibr" rid="ref12">12</xref>). In contrast, mental health-specific GenAI systems trained on curated clinical datasets and optimized for the specific domain offer more structured representations of knowledge, but still face challenges in terms of transparency, interpretability, and generalizability. Furthermore, over-trust on these systems, especially among the most vulnerable or emotionally unstable users, may undermine integrity of the psychological support, highlighting the need for rigorous oversight, explicit scope delimitation, and human supervision in deployment (<xref ref-type="bibr" rid="ref23">23</xref>). Some fundamental limitations are associated with the quality of the interaction and the ability to effectively facilitate self-reflection, aspects widely recognized as central to psychological support (<xref ref-type="bibr" rid="ref20">20</xref>). For psychological support, pre-reflective and pre-thematic processes of intercorporeal attunement are essentials (<xref ref-type="bibr" rid="ref24">24</xref>). This means that when establishing an interaction, the body&#x2019;s nonverbal communicative work cannot be overlooked. In addition, self-reflection involves the recursive examination of one&#x2019;s internal landscape, thoughts, emotions, behaviors, and underlying beliefs to develop greater self-awareness and psychological insight. This process is typically catalyzed by human clinicians who can offer empathic attunement, contextual sensitivity, and the capacity to adaptively challenge cognitive distortions. In contrast, AI systems lack non-verbal communication, bodily presence, intentionality, conscious awareness, and metacognitive flexibility, which are critical faculties for guiding clients through the layered complexities of emotional and existential experience. While advanced language models can generate coherent and emotionally resonant responses, they do so without subjective understanding or a genuine grasp of user intent. Furthermore, current CAs are primarily optimized for linguistic coherence and conversational fluency, not for promoting the kind of open-ended inquiry, emotional mirroring, and Socratic questioning that drive reflective growth.</p>
<p>Consequently, while CAs may simulate aspects of the dialogue, they are not yet capable of eliciting insight-oriented transformation. They cannot fully decode subtext, intrapersonal conflict, or nonverbal cues, elements that are essential to identifying maladaptive schemas and fostering internal change. This deficiency limits their effectiveness in encouraging the introspective processes essential for meaningful psychological support.</p>
<p>Despite these limitations, optimizing CAs for psychological support is an ongoing effort.</p>
<p>This perspective article highlights the critical challenges that must be addressed for GenAI to become a viable mental health tool, such as the need for enhancing personalization, multimodal interaction, and immersive environments. We propose a framework to connect these three broad challenges in an epistemic structured model with psychological relevance. The current limitations of CAs can be overcome through targeted optimizations that align AI interventions with established psychological support principles.</p>
</sec>
<sec id="sec2">
<title>Personalization through psychological domain adaptation and tailored instructions</title>
<p>One of the main limitations of current CAs is their generic nature which lacks personalization and specificity. This limitation impairs their ability to recognize and appropriately respond to complex emotional states and experiences (<xref ref-type="bibr" rid="ref17">17</xref>). This is also confirmed by a recent perspective (<xref ref-type="bibr" rid="ref19">19</xref>) stating that generic LLMs in healthcare require domain-specific adaptation. The authors recommend using techniques such as domain-specific tuning, retrieval of verified clinical knowledge, and explicit guardrails to reduce hallucinations, improve factual accuracy, and align results with clinical needs. Considering this, we need to frame personalization on two different, but at the same time complementary, layers: the first shows the personalization as domain-specific knowledge, while the second represents the personalization as tailored response instructions.</p>
<p><italic>Personalization as domain-specific knowledge</italic>: To improve personalization, we propose to use the back-end Psychological Domain Adaptation (PDA) to supply clinical context and psychological support through three possible mechanisms: grounded retrieval-augmented generation (RAG), fine-tuning, and knowledge bases-backed (for custom GPT). RAG externalizes knowledge at inference time by conditioning responses on vetted clinical/psychological sources and displaying citations. This favors psychological support through psychoeducational uses that require traceability and principled abstention when evidence is lacking. Recent health-information evaluations show that RAG reduces hallucinations and increases appropriate non-response when the reference corpus is reliable (<xref ref-type="bibr" rid="ref25">25</xref>, <xref ref-type="bibr" rid="ref26">26</xref>).</p>
<p>Fine-tuning (including instruction/prompt tuning) internalizes psychological language, support strategies, and response schemas in the model parameters, stabilizing supportive dialogue. Work grounded in emotional-support corpora [e.g., ESConv; (<xref ref-type="bibr" rid="ref27">27</xref>)] and recent fine-tuning studies for long emotional-support conversations report more consistent strategy use and higher human-rated support quality (<xref ref-type="bibr" rid="ref28">28</xref>).</p>
<p>Curated knowledge bases (KB-backed custom GPTs) provide the verifiable corpus that assistants index via embeddings/semantic search and are useful when content does not require updates. A recent study demonstrates their potential through SOCRATES (<xref ref-type="bibr" rid="ref29">29</xref>) for accessible conversational psychological support.</p>
<p>To our knowledge, no study has thus far evaluated the incremental psychological support effect of back-end PDA. However, we believe that this topic is central to refining the use of CAs in adjunctive psychological support contexts. A distinctive feature of GenAI systems is their inherent multilingual capability and the potential for cultural adaptation through cultural-linguistic-specific training (<xref ref-type="bibr" rid="ref30">30</xref>). This theoretically positions them as tools capable of addressing linguistic and cultural barriers in accessing mental health. However, this would require model training that takes into account culturally specific idioms of distress, help-seeking patterns, and common psychological approaches (<xref ref-type="bibr" rid="ref25">25</xref>).</p>
<p><italic>Personalization as tailored response instructions</italic>: This layer specifies how the CA communicates and when it must be handed off. We can define an explicit CA identity (supportive, non-diagnostic, adjunct to care) with a constrained scope. In addition, an empathic response template that prioritizes reflective listening, acknowledgment of uncertainty, and &#x201C;clarify-before-advice&#x201D; can be useful with a safety lexicon + intent rules that trigger crisis flows. Concretely, we can maintain a red-flag lexicon (e.g., suicidal or self-harm terms) paired with intent detection. On detection, CA must avoid procedural advice, acknowledge risk, provide region-appropriate crisis options and urgent human escalation, and continue only in a supportive/containment mode. The need for such keyword/intent guardrails is evidenced by evaluations showing that many health CAs still fail to respond appropriately to suicidality prompts [only ~44%, (<xref ref-type="bibr" rid="ref31">31</xref>)], and by comparative studies of LLMs on suicidality response standards that show good performance at extremes but variability for intermediate-risk queries, underscoring the importance of calibrated prompts and explicit escalation criteria rather than implicit &#x201C;best effort&#x201D; behavior (<xref ref-type="bibr" rid="ref32">32</xref>).</p>
<p>However, operationalizing fine-tuning and RAG in adjunctive psychological support entails concrete technical barriers. First, mental-health&#x2013;specific corpora remain limited and costly to curate at scale, constraining robust domain adaptation and evaluation. Second, privacy risks are non-trivial: model adaptation and retrieval pipelines can surface or memorize sensitive content, requiring strict de-identification and leakage-mitigation strategies (<xref ref-type="bibr" rid="ref33">33</xref>). In addition, cultural and linguistic adaptation is an open problem: models trained on majority-language, majority-culture data often fail to generalize to diverse clinical norms and idioms, necessitating localized data and ongoing validation. These constraints imply that personalization must be engineered alongside data governance and culture-aware evaluation, not solely through architectural improvements (<xref ref-type="bibr" rid="ref34">34</xref>, <xref ref-type="bibr" rid="ref35">35</xref>). Lastly, instruction-based tuning that helps models adopt a specific, warmer, and more appropriate tone is essential. Personalization allows systems to recognize uncertainty and seek clarification rather than making assumptions, a critical feature for maintaining ethical GenAI interactions (<xref ref-type="bibr" rid="ref36">36</xref>).</p>
</sec>
<sec id="sec3">
<title>Advancing human-AI interaction through spoken and expressive language</title>
<p>Most existing CAs for adjunctive psychological support rely on text-based interactions, limiting their ability to replicate the richness of human conversation. While text-based systems offer accessibility and convenience, spoken dialogue introduces additional dimensions of engagement, particularly through prosody, tone, and rhythm. Research demonstrates that voice-based AI can enhance user trust and emotional connection, making interactions more natural and engaging (<xref ref-type="bibr" rid="ref37">37</xref>). Furthermore, voice interfaces could theoretically improve accessibility for populations with literacy difficulties or visual impairments (<xref ref-type="bibr" rid="ref38">38</xref>).</p>
<p>In addition to linguistic fluency, empathetic CAs can leverage affective computation pipelines to infer user state from acoustic-prosodic correlations (e.g., intensity dynamics, spectral slope, jitter/shimmer, pause structure, turn timing) and integrate these signals with lexical/pragmatic features to better tailor responses (<xref ref-type="bibr" rid="ref39">39</xref>). In psychological support, such recognition of vocal emotions and adaptive prosody have shown promise in increasing perceived empathy and engagement, but they exhibit sensitivity to demographic, cultural, and channel variability, with risks of misclassification in the presence of noise and heterogeneous conversation styles (<xref ref-type="bibr" rid="ref40">40</xref>). To avoid expectation drift and over-reliance (<xref ref-type="bibr" rid="ref41">41</xref>), voice systems should combine emotion-based adaptations with calibrated transparency (e.g., safety/uncertainty cues), provide human support, and have cross-language and cross-psychological norm validation.</p>
<p>The effectiveness of synthetic voice depends on its ability to convey warmth, empathy, and dynamically adapt to conversational cues. Current speech synthesis technologies often lack the modulation and affective nuances necessary for effective psychological support communication. Additionally, turn-taking and latency issues can disrupt conversational flow, reducing the AI system&#x2019;s perceived responsiveness. Addressing these challenges requires integrating advanced speech synthesis models capable of real-time prosodic adjustments, enabling AI to mirror human-like conversational rhythms.</p>
<p>However, the introduction of voice-based AI also raises concerns about user perception. Studies suggest that when AI-generated voices become too like human voices, users may develop unrealistic expectations of the CA&#x2019;s capabilities, leading to potential disappointment or overdependence (<xref ref-type="bibr" rid="ref11">11</xref>, <xref ref-type="bibr" rid="ref36">36</xref>). Maintaining ethical AI-assisted psychological support requires balancing naturalistic discourse with clear transparency about AI limitations and a simultaneous effort by clinicians and developers.</p>
</sec>
<sec id="sec4">
<title>The role of immersion in GenAI</title>
<p>In addition to text- and speech-based interactions, the use of CA in Virtual Reality (VR) environments represents a promising frontier in AI psychological support. Immersive technologies have been explored in various mental health applications, particularly exposure therapy and cognitive rehabilitation, demonstrating their potential to improve engagement and clinical outcomes (<xref ref-type="bibr" rid="ref42">42</xref>, <xref ref-type="bibr" rid="ref43">43</xref>).</p>
<p>Three levels of immersion are possible. At the most basic level, an interface can be represented by a CA with a static embodiment (CASE), which provides a visual presence without additional interaction. A more advanced approach involves non-immersive embodied CA (NIECA) that synchronizes lip movements and facial expressions, creating a more dynamic conversational interaction in a non-immersive VR environment. Finally, full immersion in a VR environment enables a simulated dialogue session in which users can interact with immersive embodied CA (IECA). Within this feasibility backdrop, embodiment&#x2019;s contribution may enhance engagement through presence, potentially supporting adherence and depth of participation (<xref ref-type="bibr" rid="ref44">44</xref>, <xref ref-type="bibr" rid="ref45">45</xref>).</p>
<p>The concept of virtual entity (VE) may expand the potential of AI systems as tools of psychological support. When users perceive IECA as VE within shared virtual spaces, interactions may shift from abstract text exchanges into more naturalistic. Literature suggests that IECA may elicit responses that partly overlap human-to-human social and emotional processes: from the level of brain activation and physiology to subjective feelings of trust, empathy, and connection; whether these signals mediate clinically meaningful change remains to be established (<xref ref-type="bibr" rid="ref46 ref47 ref48">46&#x2013;48</xref>). IECA and NIECA also enable non-verbal communication through gestures, posture, and proxemics, crucial elements often missing in traditional digital interventions. Nevertheless, this embodied experience raises ethical considerations about attachment formation and transference with VE, requiring careful implementation guidelines to maintain appropriate boundaries while leveraging the benefits of embodied presence. While some studies suggest that IECA can reduce social anxiety, improve self-reflection (<xref ref-type="bibr" rid="ref49">49</xref>), and increase engagement and social closeness, others warn of potential negative effects, including cognitive overload and derealization in vulnerable populations (<xref ref-type="bibr" rid="ref50">50</xref>). Additionally, the &#x201C;uncanny valley&#x201D; phenomenon, where highly realistic but slightly flawed IECA create discomfort, must be carefully considered when designing AI-driven psychological support environments. But beyond the design possibilities, supportive psychological implementation also depends on practical constraints. The total cost of ownership includes the availability of many headsets if used on a large scale, software licenses, and training for psychological staff; these factors determine feasibility and productivity in routine care (<xref ref-type="bibr" rid="ref43">43</xref>). Moreover, tolerance to VR is uneven: symptoms such as digital motion sickness and discomfort vary depending on users and settings; some users (e.g., those with vestibular problems, migraines, or epilepsy) are not eligible, and home use adds stable connectivity and remote supervision requirements to ensure safety and adherence (<xref ref-type="bibr" rid="ref51">51</xref>). Current evidence in mental health is still largely confined to controlled trials, so controlled studies, standardized outcomes, and safety monitoring are needed before proceeding to large-scale rollout. Immersion may amplify derealization/dissociation or excessive identification with IECA; therefore, prudent use requires pre-session screening, time limits, gradual exposure, and post-session debriefing to maintain psychological safety (<xref ref-type="bibr" rid="ref51">51</xref>).</p>
</sec>
<sec id="sec5">
<title>Theoretical framework: PMVE for psychological support applications</title>
<p>In light of the above considerations and evidence, we propose a theoretical framework (see <xref ref-type="fig" rid="fig1">Figure 1</xref>) aimed at integrating the three challenges (personalization, multimodal interaction, and immersive environments) as a prospective solution to the limitations of general-purpose CAs and a pathway toward their responsible and personalized integration in adjunctive psychological support.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>The PMVE framework (Personalization, Multimodality, and Virtual-Entity) for generative AI in psychological support. The model illustrates three interdependent layers (personalization, multimodal communication, and virtual entity) that collectively enhance epistemic alignment, psychological responsiveness, and experiential engagement between users and AI systems. Each layer contributes to transforming virtual entities from purely conversational agents into context-sensitive and psychological support interventions. Icons used in <xref ref-type="fig" rid="fig1">Figure 1</xref> were obtained from Flaticon (<ext-link xlink:href="https://www.flaticon.com" ext-link-type="uri">www.flaticon.com</ext-link>). Credits by panel are as follows: Multimodal interaction: human eye and headset by Freepik (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>) and mouth illustration by Fitri Handayani (<ext-link xlink:href="https://www.vecteezy.com/members/amalialuri19948857" ext-link-type="uri">https://www.vecteezy.com/members/amalialuri19948857</ext-link>). Personalization: user profile by Elzicon (<ext-link xlink:href="https://www.flaticon.com/authors/elzicon" ext-link-type="uri">https://www.flaticon.com/authors/elzicon</ext-link>). Back-end Psychological Domain Adaptation: Computer by Freepik (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>). Tailored response instructions: precision medicine icon by gravisio (<ext-link xlink:href="https://www.flaticon.com/authors/gravisio" ext-link-type="uri">https://www.flaticon.com/authors/gravisio</ext-link>). Text-based system: keyboard icon by Freepik  (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>). Spoken dialogue: microphone icon by Fathema Khanom (<ext-link xlink:href="https://www.freepik.com/author/fathema-khanom" ext-link-type="uri">https://www.freepik.com/author/fathema-khanom</ext-link>). Virtual entity: chatbot icon by Freepik (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>). Non-Immersive environments: Online meeting by user21718592 (<ext-link xlink:href="https://www.freepik.com/author/user21718592/icons" ext-link-type="uri">https://www.freepik.com/author/user21718592/icons</ext-link>). Immersive environments: VR Headset by Freepik (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>). Basic exchange (CASE): psychologist and bot icons by Freepik (<ext-link xlink:href="https://www.flaticon.com/authors/freepik" ext-link-type="uri">https://www.flaticon.com/authors/freepik</ext-link>). Interactive exchange (NIECA): chat icon by narak0rn (<ext-link xlink:href="https://www.flaticon.com/authors/narak0rn" ext-link-type="uri">https://www.flaticon.com/authors/narak0rn</ext-link>). Inter-Entity Experience (IECA): conversation icon by Prosymbols Premium (<ext-link xlink:href="https://www.flaticon.com/authors/prosymbols-premium" ext-link-type="uri">https://www.flaticon.com/authors/prosymbols-premium</ext-link>).</p>
</caption>
<graphic xlink:href="fpubh-14-1603238-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart illustrating a pathway for delivering psychological support through conversational agent systems: it begins with personalization and multimodal interaction, followed by domain adaptation and response instructions, leading to either text-based and/or speech-based systems. These can be deployed in non-immersive or immersive environments, and culminate in either interactive exchanges (user&#x2013;system) or inter-system experiences. Sidebars label two relational dimensions: connection and interaction.</alt-text>
</graphic>
</fig>
<p>We refer to this three-layer framework as the PMVE framework (Personalization, Multimodality and Virtual-Entity) that links (i) epistemic alignment, (ii) multimodal communication, and (iii) graded immersion to human-AI interaction.</p>
<p>We next unpack each layer in turn.</p>
<p>Personalization ensures epistemic consistency by aligning the system&#x2019;s outputs with domain-specific psychological principles and dynamically adapting responses to the user&#x2019;s profile. Tailored response instructions and the back-end PDA translate epistemic structures into individualized practices, ensuring that the system&#x2019;s adaptive responses maintain psychological supportive relevance and contextual sensitivity.</p>
<p>Text-based messages and voice responses (TTS) constitute the main channels of human-VE communication, establishing the linguistic substrate through which psychological supportive processes unfold. Building on this foundation, multimodal interaction expands the communicative spectrum by incorporating sensory and contextual dimensions that enable a more natural, embodied, and immersive engagement.</p>
<p>The subsequent layer, encompassing immersive environments and inter-entity experience, extends interaction into virtual contexts, where users engage with IECA that co-construct and sustain the setting. Notably, this opportunity allows for a level of exchange that is not only interactive, but experiential. We argue that the degree of user immersion and perceived interactivity increases systematically with each level of embodiment. This progression can be further conceptualized in terms of the intensity of the underlying bond: from basic exchange, as the mere establishment of a communicative link (e.g., with a CASE); to interactive exchange, denoting reciprocal and dynamic exchange (e.g., with a NIECA); and finally, to a sustained and co-constructed inter-entity experience (e.g., with a IECA). Each level corresponds to an increasing depth of engagement and mutual influence, reflecting the transition from potential connectivity to fully embodied experience. A disembodied interaction remains purely symbolic, while an immersive inter-entity experience establishes a fully embodied, spatially co-located interaction. This difference highlights the transition from cognitive to sensorimotor forms of engagement in human-VE communication. The degree of embodiment and immersion directly influences the potential of human-VE interactions. Lower levels (e.g., disembodied or static representations) may support basic cognitive interventions but provide limited emotional resonance. Conversely, higher levels, particularly immersive, enable multisensory engagement and co-presence, fostering empathy, behavioral modeling, and embodied learning.</p>
<p>The PMVE framework suggests that the experiential depth of the interface can modulate outcomes in digital settings for psychological support (see <xref ref-type="fig" rid="fig2">Figure 2</xref>). This is why it becomes crucial to determine which type of setting may be most beneficial for a VE-based intervention. At the same time, it is important to acknowledge that such settings should be precisely tailored to the user&#x2019;s characteristics and goals. Consequently, immersive environments, despite their greater potential for fostering meaningful experiences, will not always represent the most appropriate or effective solution. Collectively, these components articulate an epistemically structured model that bridges psychological, theoretical, and technological domains, fostering an integrative approach to psychological support in digital and immersive settings.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Three examples of possible application scenarios from the PMVE framework. <bold>(A)</bold> Generalized psychological support by a conversational agent (CA) with a static embodiment; <bold>(B)</bold> psychoeducation on anxiety via RAG and non-immersive embodied CA; <bold>(C)</bold> inter-entity experience for post-traumatic stress disorder with immersive embodied CA. This figure conceptualizes how increasing levels of immersion are correlated with greater presence and emotional resonance. The transition from connection to interaction, and ultimately, experience, reflects the progression from symbolic to dynamic modes of human-AI interaction. Applications vary based on purpose.</p>
</caption>
<graphic xlink:href="fpubh-14-1603238-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Table compares three possible psychological GenAI interaction modes: First-level Psychological Support, Specialized Psychoeducational Dialogue, and Inter-Entity Experience, across five criteria: psychological objective, backend architecture, response style, interaction channels, and virtual representation. Each mode increases in complexity and immersion from symbolic avatar chat to embodied VR interactions.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec6">
<title>Toward ethical and effectiveness GenAI for psychological support</title>
<p>The integration of GenAI into psychological support marks a pivotal evolution in the field, offering both unprecedented opportunities and significant ethical and clinical challenges. On one hand, GenAI holds immense promise: it can enhance accessibility, reduce systemic delays, and lower the cost of care, particularly for underserved and marginalized populations. On the other hand, current implementations, especially those relying solely on text-based interactions, often fail to guarantee the fundamental conditions for meaningful psychological support. This, indeed, includes the capacity to facilitate self-reflection, provide emotionally attuned and context-sensitive responses, and respect the user&#x2019;s cognitive and emotional bandwidth. Without these elements, AI-driven interventions risk being experienced as mechanistic, impersonal, or even detrimental.</p>
<p>Emerging research in IECA suggests that more relationally aware systems may begin to bridge this gap (<xref ref-type="bibr" rid="ref52">52</xref>). Functional neuroimaging studies have demonstrated that emotionally responsive NIECA and IECA can engage neural circuits implicated in social bonding, such as the medial prefrontal cortex, insula, and temporoparietal junction, areas fundamental to empathy, mentalizing, and perspective-taking (<xref ref-type="bibr" rid="ref46 ref47 ref48">46&#x2013;48</xref>, <xref ref-type="bibr" rid="ref53">53</xref>). Moreover, NIECA and IECA that mirror human nonverbal cues or engage in empathic dialogue have been shown to elicit affective bonds, encourage emotional disclosure, and even predict outcomes in digital mental health settings. These findings lend empirical support to the notion that AI, when thoughtfully designed, can transcend transactional dialogue and foster a sense of presence. However, even as AI systems become more sophisticated, their psychological validity remains contingent on continuous oversight, clear role definition, and transparent communication regarding the system&#x2019;s capabilities and limitations (<xref ref-type="bibr" rid="ref23">23</xref>).</p>
<p>In realizing this potential, adherence to clinical integrity, ethical transparency, and user safety must remain non-negotiable. To render these measures genuinely actionable in a psychological-support CA, the governance architecture should weave data protection, transparency, jurisdiction-specific incident readiness, lifecycle accountability, and meaningful clinical oversight into a coherent operational design. At the data-governance layer, any secondary use of conversational traces is preceded by a documented assessment of purpose compatibility and a clear articulation of the lawful basis for processing special-category data; safeguards associated with Article 89 of the GDPR (data minimization and pseudonymization) are applied ex ante and the resulting determinations are reflected in user notices in line with EU guidance on further processing (<xref ref-type="bibr" rid="ref54">54</xref>). The user interface operationalizes transparency obligations by making explicit that the interlocutor is an AI system, by providing an intelligible indicator whenever emotion-recognition functionality is active, and by unambiguously labeling synthetic outputs; each of these transparency events is durably logged to support audit and regulatory review, consistent with Article 50 of the EU AI Act (<xref ref-type="bibr" rid="ref55">55</xref>). Deployments aimed at the United States are scoped at the outset to determine whether the arrangement falls within HIPAA, including whether a covered-entity or business-associate relationship necessitates a Business Associate Agreement; where HIPAA does not apply, breach-response workflows are aligned with the Federal Trade Commission&#x2019;s Health Breach Notification Rule, as amended and effective 29 July 2024, and harmonized with Office for Civil Rights guidance on individual rights (<xref ref-type="bibr" rid="ref62">62</xref>). Lifecycle accountability treats the CAs as software within the ambit of the new EU Product Liability Directive, maintaining versioned model cards, safety-relevant change logs, and secure update/rollback policies capable of addressing post-sale defects, including those introduced by model updates or continuous learning, thereby preserving traceability and remedial agility (<xref ref-type="bibr" rid="ref56">56</xref>). Human-in-the-loop oversight qualified clinicians must be able to inspect inputs and plain-language rationales, monitor for anomalies, override or stop the system, and remain accountable for use; meeting the EU AI Act&#x2019;s human-oversight obligations and the FDA&#x2019;s &#x201C;Non-Device CDS&#x201D; criterion that clinicians independently review the basis for recommendations. These measures turn abstract compliance into concrete UX patterns, governance artifacts, and operational controls that protect users while preserving the value of psychological support.</p>
</sec>
<sec id="sec7">
<title>From framework to evaluation: evidence gaps and conclusions</title>
<p>While the PMVE framework draws on convergent findings from affective computing, human-computer interaction, and clinical psychology, the evidence base for embodied and immersive implementations remains limited. As far as we know, no randomized controlled trials have directly compared personalization strategies (RAG versus fine-tuning) using patient-centered outcomes in psychological contexts (<xref ref-type="bibr" rid="ref25">25</xref>). Voice-based affective computing studies remain predominantly correlational; the causal impact of adaptive prosody on psychological change has not been established in longitudinal designs (<xref ref-type="bibr" rid="ref57">57</xref>).</p>
<p>Neural activation patterns elicited by IECA demonstrate engagement of social cognition networks (<xref ref-type="bibr" rid="ref46">46</xref>), yet these findings do not establish psychological efficacy. As far as we know, no studies have demonstrated that IECA produce superior outcomes to text-based modalities in adequately controlled trials with standardized mental health endpoints. The potential for adverse effects, including attachment formation, parasocial over-reliance, or exacerbation of dissociative symptoms, has been noted qualitatively (<xref ref-type="bibr" rid="ref58">58</xref>) but lacks systematic prospective monitoring. This matters at system level because adverse events and over-reliance are not only individual risks but also implementation risks that can affect service demand, escalation burden, and trust in digital pathways if not monitored prospectively. WHO guidance on digital interventions (<xref ref-type="bibr" rid="ref59">59</xref>) emphasizes that digital tools should be implemented as part of functioning health systems, with attention to benefits, harms, feasibility, and equity rather than as standalone substitutes. From a preventive perspective of possible adverse events, it must be considered that AI-based psychological support systems are not appropriate for all clinical presentations and may pose specific risks to vulnerable populations. For example, trauma-related dissociative disorders present heightened risk in immersive environments, where spatial presence and perceptual realism may trigger dissociative episodes or emotional flooding (<xref ref-type="bibr" rid="ref51">51</xref>). Furthermore, individuals with psychotic features, severe mania, or eating disorders require specialized safeguards, as general-purpose systems may inadvertently reinforce maladaptive cognitions or exacerbate perceptual disturbances (<xref ref-type="bibr" rid="ref60">60</xref>, <xref ref-type="bibr" rid="ref61">61</xref>).</p>
<p>The PMVE framework should be considered as a structured roadmap ready for evaluation. Accordingly, we propose PMVE as an incremental, evidence-proportionate pathway (see <xref ref-type="fig" rid="fig2">Figure 2</xref>), where the three levels differ not only in modality but in their expected health-system contribution: lower-intensity modes can broaden reach and reduce wait-time pressure, hybrid modes support continuity and triage, and inter-entity experiences are reserved for protocolized, supervised use where added value justifies higher cost and risk. However, practical barriers to population-wide implementation include infrastructure requirements (reliable bandwidth, device access) that remain unequally distributed, particularly in resource-constrained settings.</p>
<p>The next step is therefore straightforward: move from plausible mechanisms to measurable clinical and implementation endpoints, so that PMVE&#x2019;s benefits, and its boundaries, can be established with the same rigor expected of any intervention intended for real-world mental health systems. In this light, GenAI should be understood not as a standalone solution, but as a complementary ally, a tool best deployed when human presence is unavailable, and always in service of enhancing, not replacing, the human connection at the heart of mental health care. This perspective article outlines a plausible path (personalization, multimodality, and inter-entity experience) through which GenAI systems could represent adjunctive psychological support; these possibilities deserve testing. Future research should rigorously evaluate their effectiveness in mental health care.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec8">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="ethics-statement" id="sec9">
<title>Ethics statement</title>
<p>Ethical approval was not required for the study involving human samples in accordance with the local legislation and institutional requirements because [reason ethics approval was not required]. Written informed consent for participation in this study was provided by the participants&#x2019; legal guardians/next of kin.</p>
</sec>
<sec sec-type="author-contributions" id="sec10">
<title>Author contributions</title>
<p>FF: Conceptualization, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. CP: Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. CR: Writing &#x2013; review &#x0026; editing. GR: Supervision, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec sec-type="COI-statement" id="sec11">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec12">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec13">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moulaei</surname><given-names>K</given-names></name> <name><surname>Yadegari</surname><given-names>A</given-names></name> <name><surname>Baharestani</surname><given-names>M</given-names></name> <name><surname>Farzanbakhsh</surname><given-names>S</given-names></name> <name><surname>Sabet</surname><given-names>B</given-names></name> <name><surname>Afrash</surname><given-names>MR</given-names></name></person-group>. <article-title>Generative artificial intelligence in healthcare: a scoping review on benefits, challenges and applications</article-title>. <source>Int J Med Inform</source>. (<year>2024</year>) <volume>188</volume>:<fpage>105474</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2024.105474</pub-id>, <pub-id pub-id-type="pmid">38733640</pub-id></mixed-citation></ref>
<ref id="ref2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Preiksaitis</surname><given-names>C</given-names></name> <name><surname>Rose</surname><given-names>C</given-names></name></person-group>. <article-title>Opportunities, challenges, and future directions of generative artificial intelligence in medical education: scoping review</article-title>. <source>JMIR Med Educ</source>. (<year>2023</year>) <volume>9</volume>:<fpage>E48785</fpage>. doi: <pub-id pub-id-type="doi">10.2196/48785</pub-id>, <pub-id pub-id-type="pmid">37862079</pub-id></mixed-citation></ref>
<ref id="ref3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hatch</surname><given-names>SG</given-names></name> <name><surname>Goodman</surname><given-names>ZT</given-names></name> <name><surname>Vowels</surname><given-names>L</given-names></name> <name><surname>Hatch</surname><given-names>HD</given-names></name> <name><surname>Brown</surname><given-names>AL</given-names></name> <name><surname>Guttman</surname><given-names>S</given-names></name> <etal/></person-group>. <article-title>When ELIZA meets therapists: a turing test for the heart and mind</article-title>. <source>PLoS Ment Health</source>. (<year>2025</year>) <volume>2</volume>:13. doi: <pub-id pub-id-type="doi">10.1371/journal.pmen.0000145</pub-id>, <pub-id pub-id-type="pmid">41661855</pub-id></mixed-citation></ref>
<ref id="ref4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tantam</surname><given-names>D</given-names></name></person-group>. <article-title>The machine as psychotherapist: impersonal communication with a machine</article-title>. <source>Adv Psychiatr Treat</source>. (<year>2006</year>) <volume>12</volume>:<fpage>416</fpage>&#x2013;<lpage>26</lpage>. doi: <pub-id pub-id-type="doi">10.1192/apt.12.6.416</pub-id></mixed-citation></ref>
<ref id="ref5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Weizenbaum</surname><given-names>J</given-names></name></person-group>. <article-title>ELIZA &#x2014; a computer program for the study of natural language communication between man and machine</article-title>. <source>Commun ACM</source>. (<year>1983</year>) <volume>26</volume>:<fpage>23</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.1145/357980.357991</pub-id></mixed-citation></ref>
<ref id="ref6"><label>6.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Brennan</surname><given-names>K</given-names></name></person-group>. The managed teacher: emotional labour, education, and technology. (<year>2006</year>). Available online at: <ext-link xlink:href="https://www.semanticscholar.org/paper/THE-MANAGED-TEACHER%3A-EMOTIONAL-LABOUR%2C-EDUCATION%2C-Brennan/18651b7573ed4c146e4d09a6ab83a7473c9a04d7" ext-link-type="uri">https://www.semanticscholar.org/paper/THE-MANAGED-TEACHER%3A-EMOTIONAL-LABOUR%2C-EDUCATION%2C-Brennan/18651b7573ed4c146e4d09a6ab83a7473c9a04d7</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>C-C</given-names></name> <name><surname>Chiu</surname><given-names>CW</given-names></name> <name><surname>Chang</surname><given-names>C-H</given-names></name> <name><surname>Lo</surname><given-names>F-y</given-names></name></person-group>. <article-title>Analysis of a chatbot as a dialogic reading facilitator: its influence on learning interest and learner interactions</article-title>. <source>Educ Technol Res Dev</source>. (<year>2024</year>) <volume>72</volume>:<fpage>4</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s11423-024-10370-0</pub-id></mixed-citation></ref>
<ref id="ref8"><label>8.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Shawar</surname><given-names>BA</given-names></name> <name><surname>Atwell</surname><given-names>E</given-names></name></person-group>. "<chapter-title>Different measurements metrics to evaluate a Chatbot system</chapter-title>". In:  <source>Proceedings of the workshop on bridging the gap academic and industrial research in dialog technologies - NAACL-HLT &#x2018;07</source> <publisher-loc>Rochester</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name>. (<year>2007</year>). p. <fpage>89</fpage>&#x2013;<lpage>96</lpage>.</mixed-citation></ref>
<ref id="ref9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cristea</surname><given-names>IA</given-names></name> <name><surname>Sucal&#x0103;</surname><given-names>M</given-names></name> <name><surname>David</surname><given-names>D</given-names></name></person-group>. <article-title>Can you tell the difference? Comparing face-to-face versus computer-based interventions. The &#x2018;Eliza&#x2019; effect in psychotherapy</article-title>. <source>J Cogn Behav Psychother (Romania)</source>. (<year>2013</year>) <volume>13</volume>:<fpage>291</fpage>&#x2013;<lpage>8</lpage>.</mixed-citation></ref>
<ref id="ref10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Doss</surname><given-names>MK</given-names></name> <name><surname>Pova&#x017E;an</surname><given-names>M</given-names></name> <name><surname>Rosenberg</surname><given-names>MD</given-names></name> <name><surname>Sepeda</surname><given-names>ND</given-names></name> <name><surname>Davis</surname><given-names>AK</given-names></name> <name><surname>Finan</surname><given-names>PH</given-names></name> <etal/></person-group>. <article-title>Psilocybin therapy increases cognitive and neural flexibility in patients with major depressive disorder</article-title>. <source>Transl Psychiatry</source>. (<year>2021</year>) <volume>11</volume>:<fpage>1</fpage>&#x2013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41398-021-01706-y</pub-id>, <pub-id pub-id-type="pmid">34750350</pub-id></mixed-citation></ref>
<ref id="ref11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xie</surname><given-names>Y</given-names></name> <name><surname>Zhu</surname><given-names>K</given-names></name> <name><surname>Zhou</surname><given-names>P</given-names></name> <name><surname>Liang</surname><given-names>C</given-names></name></person-group>. <article-title>How does anthropomorphism improve human-AI interaction satisfaction: a dual-path model</article-title>. <source>Comput Human Behav</source>. (<year>2023</year>) <volume>148</volume>:<fpage>107878</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.chb.2023.107878</pub-id></mixed-citation></ref>
<ref id="ref12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Demszky</surname><given-names>D</given-names></name> <name><surname>Yang</surname><given-names>D</given-names></name> <name><surname>Yeager</surname><given-names>DS</given-names></name> <name><surname>Bryan</surname><given-names>CJ</given-names></name> <name><surname>Clapper</surname><given-names>M</given-names></name> <name><surname>Chandhok</surname><given-names>S</given-names></name> <etal/></person-group>. <article-title>Using large language models in psychology</article-title>. <source>Nat Rev Psychol</source>. (<year>2023</year>) <volume>2</volume>:<fpage>688</fpage>&#x2013;<lpage>701</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s44159-023-00241-5</pub-id></mixed-citation></ref>
<ref id="ref13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yeh</surname><given-names>P-L</given-names></name> <name><surname>Kuo</surname><given-names>W-C</given-names></name> <name><surname>Tseng</surname><given-names>B-L</given-names></name> <name><surname>Sung</surname><given-names>Y-H</given-names></name></person-group>. <article-title>Does the AI-driven chatbot work? Effectiveness of the Woebot app in reducing anxiety and depression in group counseling courses and student acceptance of technological aids</article-title>. <source>Curr Psychol</source>. (<year>2025</year>) <volume>44</volume>:8133&#x2013;8145. doi: <pub-id pub-id-type="doi">10.1007/s12144-025-07359-0</pub-id></mixed-citation></ref>
<ref id="ref14"><label>14.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ayers</surname><given-names>JW</given-names></name> <name><surname>Poliak</surname><given-names>A</given-names></name> <name><surname>Dredze</surname><given-names>M</given-names></name> <name><surname>Leas</surname><given-names>EC</given-names></name> <name><surname>Zhu</surname><given-names>Z</given-names></name> <name><surname>Kelley</surname><given-names>JB</given-names></name> <etal/></person-group>. <article-title>Comparing physician and artificial intelligence Chatbot responses to patient questions posted to a public social media forum</article-title>. <source>JAMA Intern Med</source>. (<year>2023</year>) <volume>183</volume>:<fpage>589</fpage>&#x2013;<lpage>96</lpage>. doi: <pub-id pub-id-type="doi">10.1001/jamainternmed.2023.1838</pub-id>, <pub-id pub-id-type="pmid">37115527</pub-id></mixed-citation></ref>
<ref id="ref15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ovsyannikova</surname><given-names>D</given-names></name> <name><surname>de Olmburgo Mello</surname><given-names>V</given-names></name> <name><surname>Inzlicht</surname><given-names>M</given-names></name></person-group>. <article-title>Third-party evaluators perceive AI as more compassionate than expert humans</article-title>. <source>Commun Psychol</source>. (<year>2025</year>) <volume>3</volume>:<fpage>1</fpage>&#x2013;<lpage>11</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s44271-024-00182-6</pub-id>, <pub-id pub-id-type="pmid">39794410</pub-id></mixed-citation></ref>
<ref id="ref16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vowels</surname><given-names>LM</given-names></name> <name><surname>Francois-Walcott</surname><given-names>RRR</given-names></name> <name><surname>Darwiche</surname><given-names>J</given-names></name></person-group>. <article-title>AI in relationship counselling: evaluating ChatGPT&#x2019;s therapeutic capabilities in providing relationship advice</article-title>. <source>Comput Hum Behav Artif Hum</source>. (<year>2024</year>) <volume>2</volume>:<fpage>100078</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.chbah.2024.100078</pub-id></mixed-citation></ref>
<ref id="ref17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bendig</surname><given-names>E</given-names></name> <name><surname>Erb</surname><given-names>B</given-names></name> <name><surname>Schulze-Thuesing</surname><given-names>L</given-names></name> <name><surname>Baumeister</surname><given-names>H</given-names></name></person-group>. <article-title>The next generation: chatbots in clinical psychology and psychotherapy to foster mental health &#x2013; a scoping review</article-title>. <source>Verhaltenstherapie</source>. (<year>2019</year>) <volume>32</volume>:<fpage>64</fpage>&#x2013;<lpage>76</lpage>. doi: <pub-id pub-id-type="doi">10.1159/000501812</pub-id></mixed-citation></ref>
<ref id="ref18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Patel</surname><given-names>V</given-names></name> <name><surname>Saxena</surname><given-names>S</given-names></name> <name><surname>Lund</surname><given-names>C</given-names></name> <name><surname>Thornicroft</surname><given-names>G</given-names></name> <name><surname>Baingana</surname><given-names>F</given-names></name> <name><surname>Bolton</surname><given-names>P</given-names></name> <etal/></person-group>. <article-title>The lancet commission on global mental health and sustainable development</article-title>. <source>Lancet</source>. (<year>2018</year>) <volume>392</volume>:<fpage>1553</fpage>&#x2013;<lpage>98</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0140-6736(18)31612-X</pub-id>, <pub-id pub-id-type="pmid">30314863</pub-id></mixed-citation></ref>
<ref id="ref19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>Z</given-names></name> <name><surname>Wang</surname><given-names>H</given-names></name> <name><surname>Danek</surname><given-names>B</given-names></name> <name><surname>Li</surname><given-names>Y</given-names></name> <name><surname>Mack</surname><given-names>C</given-names></name> <name><surname>Arbuckle</surname><given-names>L</given-names></name> <etal/></person-group>. <article-title>A perspective for adapting generalist AI to specialized medical AI applications and their challenges</article-title>. <source>Npj Digit Med</source>. (<year>2025</year>) <volume>8</volume>:<fpage>429</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41746-025-01789-7</pub-id>, <pub-id pub-id-type="pmid">40646157</pub-id></mixed-citation></ref>
<ref id="ref20"><label>20.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll1">The Baffler</collab></person-group>. The therapist in the machine | Jess McAllen. (<year>2024</year>). Available online at: <ext-link xlink:href="https://thebaffler.com/salvos/the-therapist-in-the-machine-mcallen" ext-link-type="uri">https://thebaffler.com/salvos/the-therapist-in-the-machine-mcallen</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref21"><label>21.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll2">The Baffler</collab></person-group>. Who gets to be a therapist? | Jess McAllen. (<year>2025</year>). Available online at: <ext-link xlink:href="https://thebaffler.com/latest/who-gets-to-be-a-therapist-mcallen" ext-link-type="uri">https://thebaffler.com/latest/who-gets-to-be-a-therapist-mcallen</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref22"><label>22.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jaspers</surname><given-names>K</given-names></name></person-group>. <source>General psychopathology</source>. Baltimore: <publisher-name>JHU Press</publisher-name> (<year>1997</year>).</mixed-citation></ref>
<ref id="ref23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Blease</surname><given-names>C</given-names></name> <name><surname>Rodman</surname><given-names>A</given-names></name></person-group>. <article-title>Generative artificial intelligence in mental healthcare: an ethical evaluation</article-title>. <source>Curr Treat Options Psychiatry</source>. (<year>2024</year>) <volume>12</volume>:<fpage>1</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s40501-024-00340-x</pub-id></mixed-citation></ref>
<ref id="ref24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Stanghellini</surname><given-names>G</given-names></name> <name><surname>Lysaker</surname><given-names>PH</given-names></name></person-group>. <article-title>The psychotherapy of schizophrenia through the lens of phenomenology: intersubjectivity and the search for the recovery of first- and second-person awareness</article-title>. <source>Am J Psychother</source>. (<year>2007</year>) <volume>61</volume>:<fpage>163</fpage>&#x2013;<lpage>79</lpage>. doi: <pub-id pub-id-type="doi">10.1176/appi.psychotherapy.2007.61.2.163</pub-id>, <pub-id pub-id-type="pmid">17760320</pub-id></mixed-citation></ref>
<ref id="ref25"><label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>S</given-names></name> <name><surname>McCoy</surname><given-names>AB</given-names></name> <name><surname>Wright</surname><given-names>A</given-names></name></person-group>. <article-title>Improving large language model applications in biomedicine with retrieval-augmented generation: a systematic review, meta-analysis, and clinical development guidelines</article-title>. <source>J Am Med Inform Assoc</source>. (<year>2025</year>) <volume>32</volume>:<fpage>605</fpage>&#x2013;<lpage>15</lpage>. doi: <pub-id pub-id-type="doi">10.1093/jamia/ocaf008</pub-id>, <pub-id pub-id-type="pmid">39812777</pub-id></mixed-citation></ref>
<ref id="ref26"><label>26.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname><given-names>S</given-names></name> <name><surname>Yan</surname><given-names>Z</given-names></name> <name><surname>Dai</surname><given-names>C</given-names></name> <name><surname>Wu</surname><given-names>F</given-names></name></person-group>. <article-title>MEGA-RAG: a retrieval-augmented generation framework with multi-evidence guided answer refinement for mitigating hallucinations of LLMs in public health</article-title>. <source>Front Public Health</source>. (<year>2025</year>) <volume>13</volume>:<fpage>1635381</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpubh.2025.1635381</pub-id></mixed-citation></ref>
<ref id="ref27"><label>27.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>S</given-names></name> <name><surname>Zheng</surname><given-names>C</given-names></name> <name><surname>Demasi</surname><given-names>O</given-names></name> <name><surname>Sabour</surname><given-names>S</given-names></name> <name><surname>Li</surname><given-names>Y</given-names></name> <name><surname>Yu</surname><given-names>S</given-names></name> <etal/></person-group>. "<chapter-title>Towards emotional support dialog systems</chapter-title>". In: <person-group person-group-type="editor"><name><surname>Zong</surname><given-names>C</given-names></name> <name><surname>Xia</surname><given-names>F</given-names></name> <name><surname>Li</surname><given-names>W</given-names></name> <name><surname>Navigli</surname><given-names>R</given-names></name></person-group>, editors. <source>Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (Volume 1: Long Papers)</source> <publisher-name>Association for Computational Linguistics</publisher-name> (<year>2021</year>)</mixed-citation></ref>
<ref id="ref28"><label>28.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>T</given-names></name> <name><surname>Zhang</surname><given-names>X</given-names></name> <name><surname>Zhao</surname><given-names>J</given-names></name> <name><surname>Zhou</surname><given-names>L</given-names></name> <name><surname>Jin</surname><given-names>Q</given-names></name></person-group>. <article-title>ESCoT: towards interpretable emotional support dialogue systems</article-title>.&#x201D; In: Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Bangkok, Thailand (<year>2024</year>) <fpage>13395</fpage>&#x2013;<lpage>13412</lpage>.</mixed-citation></ref>
<ref id="ref29"><label>29.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Frisone</surname><given-names>F</given-names></name> <name><surname>Pupillo</surname><given-names>C</given-names></name> <name><surname>Rossi</surname><given-names>C</given-names></name> <name><surname>Riva</surname><given-names>G</given-names></name></person-group>. <article-title>SOCRATES. Developing and evaluating a fine-tuned ChatGPT model for accessible mental health intervention</article-title>. <source>Cyberpsychol Behav Soc Netw</source>. (<year>2025</year>) <volume>28</volume>:<fpage>366</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.1089/cyber.2025.45510.cyeuro</pub-id>, <pub-id pub-id-type="pmid">40241613</pub-id></mixed-citation></ref>
<ref id="ref30"><label>30.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Workshop</surname><given-names>BS</given-names></name> <name><surname>Scao</surname><given-names>TL</given-names></name> <name><surname>Fan</surname><given-names>A</given-names></name> <name><surname>Akiki</surname><given-names>C</given-names></name> <name><surname>Pavlick</surname><given-names>K</given-names></name> <name><surname>Ili&#x0107;</surname><given-names>S</given-names></name> <etal/></person-group>. <article-title>BLOOM: a 176B-parameter open-access multilingual language model</article-title>.&#x201D; arXiv:2211.05100. Preprint, arXiv, June 27 (<year>2023</year>). doi: <pub-id pub-id-type="doi">10.48550/arXiv.2211.05100</pub-id></mixed-citation></ref>
<ref id="ref31"><label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xue</surname><given-names>J</given-names></name> <name><surname>Zhang</surname><given-names>B</given-names></name> <name><surname>Zhao</surname><given-names>Y</given-names></name> <name><surname>Zhang</surname><given-names>Q</given-names></name> <name><surname>Zheng</surname><given-names>C</given-names></name> <name><surname>Jiang</surname><given-names>J</given-names></name> <etal/></person-group>. <article-title>Evaluation of the current state of Chatbots for digital health: scoping review</article-title>. <source>J Med Internet Res</source>. (<year>2023</year>) <volume>25</volume>:<fpage>e47217</fpage>. doi: <pub-id pub-id-type="doi">10.2196/47217</pub-id>, <pub-id pub-id-type="pmid">38113097</pub-id></mixed-citation></ref>
<ref id="ref32"><label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McBain</surname><given-names>RK</given-names></name> <name><surname>Cantor</surname><given-names>JH</given-names></name> <name><surname>Zhang</surname><given-names>LA</given-names></name> <name><surname>Baker</surname><given-names>A</given-names></name> <name><surname>Zhang</surname><given-names>F</given-names></name> <name><surname>Halbisen</surname><given-names>A</given-names></name> <etal/></person-group>. <article-title>Competency of large language models in evaluating appropriate responses to suicidal ideation: comparative study</article-title>. <source>J Med Internet Res</source>. (<year>2025</year>) <volume>27</volume>:<fpage>e67891</fpage>. doi: <pub-id pub-id-type="doi">10.2196/67891</pub-id></mixed-citation></ref>
<ref id="ref33"><label>33.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Carlini</surname><given-names>N</given-names></name> <name><surname>Tramer</surname><given-names>F</given-names></name> <name><surname>Wallace</surname><given-names>E</given-names></name> <name><surname>Jagielski</surname><given-names>M</given-names></name> <name><surname>Herbert-Voss</surname><given-names>A</given-names></name> <name><surname>Lee</surname><given-names>K</given-names></name> <etal/></person-group>. <article-title>Extracting training data from large language models</article-title>.&#x201D; In: 30th USENIX Security Symposium (USENIX Security &#x2019;21). Vancouver (BC), Canada (<year>2021</year>) 2633&#x2013;2650.</mixed-citation></ref>
<ref id="ref34"><label>34.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Bender</surname><given-names>EM</given-names></name> <name><surname>Gebru</surname><given-names>T</given-names></name> <name><surname>McMillan-Major</surname><given-names>A</given-names></name> <name><surname>Shmitchell</surname><given-names>S</given-names></name></person-group>. "<chapter-title>On the dangers of stochastic parrots: can language models be too big?</chapter-title>". In:  <source>Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (New York, NY, USA: Association for Computing Machinery), FAccT &#x2018;21, March 1</source> (<year>2021</year>). p. <fpage>610</fpage>&#x2013;<lpage>23</lpage>.</mixed-citation></ref>
<ref id="ref35"><label>35.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Lewis</surname><given-names>P</given-names></name> <name><surname>Perez</surname><given-names>E</given-names></name> <name><surname>Piktus</surname><given-names>A</given-names></name></person-group>. "<chapter-title>Retrieval-augmented generation for knowledge-intensive NLP tasks</chapter-title>". In:  <source>Proceedings of the 34th international conference on neural information processing systems (Red Hook, NY, USA: Curran Associates Inc.), NIPS &#x2018;20</source>, vol. <volume>33</volume> (<year>2020</year>). p. <fpage>9459</fpage>&#x2013;<lpage>74</lpage>.</mixed-citation></ref>
<ref id="ref36"><label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>Z</given-names></name> <name><surname>Wang</surname><given-names>J</given-names></name></person-group>. <article-title>Can AI replace psychotherapists? Exploring the future of mental health care</article-title>. <source>Front Psych</source>. (<year>2024</year>) <volume>15</volume>:<fpage>1444382</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpsyt.2024.1444382</pub-id>, <pub-id pub-id-type="pmid">39544371</pub-id></mixed-citation></ref>
<ref id="ref37"><label>37.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yonatan-Leus</surname><given-names>R</given-names></name> <name><surname>Brukner</surname><given-names>H</given-names></name></person-group>. <article-title>Comparing perceived empathy and intervention strategies of an AI chatbot and human psychotherapists in online mental health support</article-title>. <source>Couns Psychother Res</source>. (<year>2025</year>) <volume>25</volume>:<fpage>e12832</fpage>. doi: <pub-id pub-id-type="doi">10.1002/capr.12832</pub-id></mixed-citation></ref>
<ref id="ref38"><label>38.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll4">World Wide Web Consortium</collab></person-group>. Cognitive accessibility research modules - voice systems and conversational interfaces. (<year>2026</year>). Available online at: <ext-link xlink:href="https://www.w3.org/TR/coga-voice/?utm_source=chatgpt.com" ext-link-type="uri">https://www.w3.org/TR/coga-voice/?utm_source=chatgpt.com</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref39"><label>39.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Zhu</surname><given-names>Q</given-names></name> <name><surname>Chau</surname><given-names>A</given-names></name> <name><surname>Cohn</surname><given-names>M</given-names></name> <name><surname>Liang</surname><given-names>K-H</given-names></name> <name><surname>Wang</surname><given-names>H-C</given-names></name> <name><surname>Zellou</surname><given-names>G</given-names></name> <etal/></person-group>. "<chapter-title>Effects of emotional expressiveness on voice Chatbot interactions</chapter-title>". In:  <source>Proceedings of the 4th conference on conversational user interfaces. ACM, Glasgow, UK.</source> (<year>2022</year>). p. <fpage>1</fpage>&#x2013;<lpage>11</lpage>.</mixed-citation></ref>
<ref id="ref40"><label>40.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sanjeewa</surname><given-names>R</given-names></name> <name><surname>Iyer</surname><given-names>R</given-names></name> <name><surname>Apputhurai</surname><given-names>P</given-names></name> <name><surname>Wickramasinghe</surname><given-names>N</given-names></name> <name><surname>Meyer</surname><given-names>D</given-names></name></person-group>. <article-title>Perception of empathy in mental health care through voice-based conversational agent prototypes: experimental study</article-title>. <source>JMIR Form Res</source>. (<year>2025</year>) <volume>9</volume>:<fpage>e69329</fpage>. doi: <pub-id pub-id-type="doi">10.2196/69329</pub-id>, <pub-id pub-id-type="pmid">40334288</pub-id></mixed-citation></ref>
<ref id="ref41"><label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>G&#x00E1;lvez</surname><given-names>RH</given-names></name> <name><surname>Gravano</surname><given-names>A</given-names></name> <name><surname>Be&#x0148;u&#x0161;</surname><given-names>&#x0160;</given-names></name> <name><surname>Levitan</surname><given-names>R</given-names></name> <name><surname>Trnka</surname><given-names>M</given-names></name> <name><surname>Hirschberg</surname><given-names>J</given-names></name></person-group>. <article-title>An empirical study of the effect of acoustic-prosodic entrainment on the perceived trustworthiness of conversational avatars</article-title>. <source>Speech Comm</source>. (<year>2020</year>) <volume>124</volume>:<fpage>46</fpage>&#x2013;<lpage>67</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.specom.2020.07.007</pub-id></mixed-citation></ref>
<ref id="ref42"><label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cerasa</surname><given-names>A</given-names></name> <name><surname>Gaggioli</surname><given-names>A</given-names></name> <name><surname>Pioggia</surname><given-names>G</given-names></name> <name><surname>Riva</surname><given-names>G</given-names></name></person-group>. <article-title>Metaverse in mental health: the beginning of a long history</article-title>. <source>Curr Psychiatry Rep</source>. (<year>2024</year>) <volume>26</volume>:<fpage>294</fpage>&#x2013;<lpage>303</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11920-024-01501-8</pub-id>, <pub-id pub-id-type="pmid">38602624</pub-id></mixed-citation></ref>
<ref id="ref43"><label>43.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cushnan</surname><given-names>J</given-names></name> <name><surname>McCafferty</surname><given-names>P</given-names></name> <name><surname>Best</surname><given-names>P</given-names></name></person-group>. <article-title>Clinicians&#x2019; perspectives of immersive tools in clinical mental health settings: a systematic scoping review</article-title>. <source>BMC Health Serv Res</source>. (<year>2024</year>) <volume>24</volume>:<fpage>1</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12913-024-11481-3</pub-id>, <pub-id pub-id-type="pmid">39294727</pub-id></mixed-citation></ref>
<ref id="ref44"><label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bailey</surname><given-names>JO</given-names></name> <name><surname>Bailenson</surname><given-names>JN</given-names></name> <name><surname>Casasanto</surname><given-names>D</given-names></name></person-group>. <article-title>When does virtual embodiment change our minds?</article-title> <source>Presence Teleoperators Virtual Environ</source>. (<year>2016</year>) <volume>25</volume>:<fpage>222</fpage>&#x2013;<lpage>33</lpage>. doi: <pub-id pub-id-type="doi">10.1162/PRES_a_00263</pub-id></mixed-citation></ref>
<ref id="ref45"><label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rossi</surname><given-names>C</given-names></name> <name><surname>Frisone</surname><given-names>F</given-names></name> <name><surname>Riva</surname><given-names>G</given-names></name> <name><surname>Oasi</surname><given-names>O</given-names></name></person-group>. <article-title>Virtual meets reality: a psychodynamic perspective on immersive technologies</article-title>. <source>Clin Neuropsychiatry</source>. (<year>2025</year>) <volume>22</volume>:<fpage>320</fpage>&#x2013;<lpage>6</lpage>. doi: <pub-id pub-id-type="doi">10.36131/cnfioritieditore20250406</pub-id>, <pub-id pub-id-type="pmid">40989042</pub-id></mixed-citation></ref>
<ref id="ref46"><label>46.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>De Borst</surname><given-names>AW</given-names></name> <name><surname>De Gelder</surname><given-names>B</given-names></name></person-group>. <article-title>Is it the real deal? Perception of virtual characters versus humans: an affective cognitive neuroscience perspective</article-title>. <source>Front Psychol</source>. (<year>2015</year>) <volume>6</volume>:<fpage>576</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpsyg.2015.00576</pub-id></mixed-citation></ref>
<ref id="ref47"><label>47.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Numata</surname><given-names>T</given-names></name> <name><surname>Sato</surname><given-names>H</given-names></name> <name><surname>Asa</surname><given-names>Y</given-names></name> <name><surname>Koike</surname><given-names>T</given-names></name> <name><surname>Miyata</surname><given-names>K</given-names></name> <name><surname>Nakagawa</surname><given-names>E</given-names></name> <etal/></person-group>. <article-title>Achieving affective human&#x2013;virtual agent communication by enabling virtual agents to imitate positive expressions</article-title>. <source>Sci Rep</source>. (<year>2020</year>) <volume>10</volume>:<fpage>5977</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-020-62870-7</pub-id>, <pub-id pub-id-type="pmid">32249796</pub-id></mixed-citation></ref>
<ref id="ref48"><label>48.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sanjeewa</surname><given-names>R</given-names></name> <name><surname>Iyer</surname><given-names>R</given-names></name> <name><surname>Apputhurai</surname><given-names>P</given-names></name> <name><surname>Wickramasinghe</surname><given-names>N</given-names></name> <name><surname>Meyer</surname><given-names>D</given-names></name></person-group>. <article-title>Empathic conversational agent platform designs and their evaluation in the context of mental health: systematic review</article-title>. <source>JMIR Ment Health</source>. (<year>2024</year>) <volume>11</volume>:<fpage>E58974</fpage>. doi: <pub-id pub-id-type="doi">10.2196/58974</pub-id>, <pub-id pub-id-type="pmid">39250799</pub-id></mixed-citation></ref>
<ref id="ref49"><label>49.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Ren</surname><given-names>X</given-names></name></person-group>. Artificial intelligence and depression: how AI powered Chatbots in virtual reality games may reduce anxiety and depression levels. (<year>2020</year>). Available online at: <ext-link xlink:href="https://www.clausiuspress.com/article/606.html" ext-link-type="uri">https://www.clausiuspress.com/article/606.html</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref50"><label>50.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Takano</surname><given-names>M</given-names></name> <name><surname>Taka</surname><given-names>F</given-names></name></person-group>. <article-title>Fancy avatar identification and behaviors in the virtual world: preceding avatar customization and succeeding communication</article-title>. <source>Comput Human Behav Rep</source>. (<year>2022</year>) <volume>6</volume>:<fpage>100176</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.chbr.2022.100176</pub-id></mixed-citation></ref>
<ref id="ref51"><label>51.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lundin</surname><given-names>RM</given-names></name> <name><surname>Yeap</surname><given-names>Y</given-names></name> <name><surname>Menkes</surname><given-names>DB</given-names></name></person-group>. <article-title>Adverse effects of virtual and augmented reality interventions in psychiatry: systematic review</article-title>. <source>JMIR Ment Health</source>. (<year>2023</year>) <volume>10</volume>:<fpage>e43240</fpage>. doi: <pub-id pub-id-type="doi">10.2196/43240</pub-id>, <pub-id pub-id-type="pmid">37145841</pub-id></mixed-citation></ref>
<ref id="ref52"><label>52.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Hayes</surname><given-names>JH</given-names></name> <name><surname>Payne</surname><given-names>J</given-names></name> <name><surname>Leppelmeier</surname><given-names>M</given-names></name></person-group>. "<chapter-title>Toward improved artificial intelligence in requirements engineering: metadata for tracing datasets</chapter-title>". In:  <source>2019 IEEE 27th international requirements engineering conference workshops (IEEE, Jeju Island, Korea (South). REW), September</source> (<year>2019</year>). p. <fpage>256</fpage>&#x2013;<lpage>62</lpage>.</mixed-citation></ref>
<ref id="ref53"><label>53.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cheetham</surname><given-names>M</given-names></name></person-group>. <article-title>The human likeness dimension of the &#x2018;Uncanny Valley hypothesis&#x2019;: behavioral and functional MRI findings</article-title>. <source>Front Hum Neurosci</source>. (<year>2011</year>) <volume>5</volume>:<fpage>126</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnhum.2011.00126</pub-id>, <pub-id pub-id-type="pmid">22131970</pub-id></mixed-citation></ref>
<ref id="ref54"><label>54.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll5">Recital 50 - Further Processing of Personal Data</collab></person-group>. General data protection regulation (GDPR). (<year>2016</year>). Available online at: <ext-link xlink:href="https://gdpr-info.eu/recitals/no-50/" ext-link-type="uri">https://gdpr-info.eu/recitals/no-50/</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref55"><label>55.</label><mixed-citation publication-type="other">Article 50: transparency obligations for providers and deployers of certain AI systems | EU Artificial Intelligence Act. (<year>2024</year>). Available online at: <ext-link xlink:href="https://artificialintelligenceact.eu/article/50/" ext-link-type="uri">https://artificialintelligenceact.eu/article/50/</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref56"><label>56.</label><mixed-citation publication-type="other">Directive - 2024/2853 - EN - EUR-Lex. (<year>2024</year>). Available online at: <ext-link xlink:href="https://eur-lex.europa.eu/eli/dir/2024/2853/oj/eng" ext-link-type="uri">https://eur-lex.europa.eu/eli/dir/2024/2853/oj/eng</ext-link> (Accessed February 16, 2026).</mixed-citation></ref>
<ref id="ref57"><label>57.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schlicher</surname><given-names>M</given-names></name> <name><surname>Li</surname><given-names>Y</given-names></name> <name><surname>Munthumoduku Krishna Murthy</surname><given-names>S</given-names></name> <name><surname>Sun</surname><given-names>Q</given-names></name> <name><surname>Schuller</surname><given-names>BW</given-names></name></person-group>. <article-title>Emotionally adaptive support: a narrative review of affective computing for mental health</article-title>. <source>Front Digit Health</source>. (<year>2025</year>) <volume>7</volume>:<fpage>1657031</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fdgth.2025.1657031</pub-id>, <pub-id pub-id-type="pmid">41169731</pub-id></mixed-citation></ref>
<ref id="ref58"><label>58.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Malfacini</surname><given-names>K</given-names></name></person-group>. <article-title>The impacts of companion AI on human relationships: risks, benefits, and design considerations</article-title>. <source>AI &#x0026; Soc</source>. (<year>2025</year>) <volume>40</volume>:<fpage>5527</fpage>&#x2013;<lpage>40</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00146-025-02318-6</pub-id></mixed-citation></ref>
<ref id="ref59"><label>59.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><collab id="coll6">World Health Organization</collab></person-group>, editor. <source>Digital implementation investment guide (DIIG): integrating digital interventions into health programmes</source>. <edition>1st</edition> ed. <publisher-name>Geneva: World Health Organization</publisher-name> (<year>2020</year>).</mixed-citation></ref>
<ref id="ref60"><label>60.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>&#x00D8;stergaard</surname><given-names>SD</given-names></name></person-group>. <article-title>Will generative artificial intelligence Chatbots generate delusions in individuals prone to psychosis?</article-title> <source>Schizophr Bull</source>. (<year>2023</year>) <volume>49</volume>:<fpage>1418</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1093/schbul/sbad128</pub-id>, <pub-id pub-id-type="pmid">37625027</pub-id></mixed-citation></ref>
<ref id="ref61"><label>61.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Winecoff</surname><given-names>A</given-names></name> <name><surname>Klyman</surname><given-names>K</given-names></name></person-group>. <article-title>From symptoms to systems: an expert-guided approach to understanding risks of generative AI for eating disorders</article-title>.&#x201D; arXiv:2512.04843. Preprint, arXiv, December 4 (<year>2025</year>). doi: <pub-id pub-id-type="doi">10.48550/arXiv.2512.04843</pub-id></mixed-citation></ref>
<ref id="ref62"><label>62.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><collab id="coll656">Rights (OCR), Office for Civil</collab></person-group>. &#x201C;Covered Entities and Business Associates.&#x201D; (<year>2015</year>). &#x00C0;vailable online at: <ext-link xlink:href="https://www.hhs.gov/hipaa/for-professionals/covered-entities/index.html" ext-link-type="uri">https://www.hhs.gov/hipaa/for-professionals/covered-entities/index.html</ext-link>.</mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/102882/overview">Wulf R&#x00F6;ssler</ext-link>, Charit&#x00E9; University Medicine Berlin, Germany</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/770153/overview">Matteo Aloi</ext-link>, University of Messina, Italy</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2463269/overview">Marcin Rz&#x0105;deczka</ext-link>, Marie Curie-Sklodowska University, Poland</p>
</fn>
</fn-group>
<glossary>
<def-list>
<title>Glossary</title>
<def-item>
<term>Assistants API</term>
<def>
<p>An API-defined agent configuration object that binds (i) a base model, (ii) developer-provided instructions, and (iii) an enabled set of tools and tool resources (e.g., file_search vector stores, code interpreter), and that executes over a persistent Thread by creating a Run, during which the assistant may invoke tools and integrate their outputs into the response.</p>
</def>
</def-item>
<def-item>
<term>Conversational agent (CA)</term>
<def>
<p>An interactive artificial intelligence system designed to communicate with a user in natural language (text and/or speech) to provide information, assistance, or social interaction. If it&#x2019;s text-based only, it&#x2019;s called a chatbot.</p>
</def>
</def-item>
<def-item>
<term>Conversational agent with static embodiment (CASE)</term>
<def>
<p>Conversational agent that has a static visual representation (avatar or profile picture) that is not animated and non-interactive in real time.</p>
</def>
</def-item>
<def-item>
<term>Curated knowledge base (KB)</term>
<def>
<p>an externally maintained collection of domain content that is selected, cleaned, and versioned to serve as a non-parametric store of knowledge that can be accessed explicitly by a model (e.g., via a dense vector index) rather than being encoded only in model parameters.</p>
</def>
</def-item>
<def-item>
<term>Embodied Conversational Agent (ECA)</term>
<def>
<p>A CA represented by an on-screen character (or avatar) that uses verbal and nonverbal behaviors (e.g., gaze, gesture) to support face-to-face&#x2013;like interaction. It can be deployed as a Non-Immersive ECA (NIECA; presented on a conventional 2D display) or as an Immersive ECA (IECA; situated in an immersive virtual environment).</p>
</def>
</def-item>
<def-item>
<term>Fine-tuning</term>
<def>
<p>Adapting a pretrained model to a downstream task/domain by continuing training on task-specific data while updating (most or all) model parameters, yielding a task-specialized model instance.</p>
</def>
</def-item>
<def-item>
<term>Grounded retrival-augmented generation (RAG)</term>
<def>
<p>a retrieval-augmented generation setup in which a model retrieves relevant passages from an external corpus at inference time and uses the retrieved evidence to support and constrain its output, aiming to improve provenance/verifiability and reduce unsupported generation (hallucinations).</p>
</def>
</def-item>
<def-item>
<term>Virtual embodiment</term>
<def>
<p>The process/illusion by which a user maps their body schema onto a virtual body (avatar) and comes to experience that virtual body as an extension of the self, typically supported by afferent (e.g., visuo&#x2013;tactile) and/or sensorimotor correspondences between the user&#x2019;s physical body and the avatar&#x2019;s body.</p>
</def>
</def-item>
<def-item>
<term>Virtual entity</term>
<def>
<p>Terminology proposed by the authors to describe how the combination of personalization and multimodality creates something different from a generic conversational agent, whose characteristics recall those of a non-human virtual entity.</p>
</def>
</def-item>
<def-item>
<term>Virtual environment</term>
<def>
<p>A computer-generated environment that includes a virtual space plus its content and interaction rules, enabling users to perceive and act within it as an interactive world.</p>
</def>
</def-item>
<def-item>
<term>Virtual space</term>
<def>
<p>A computer-generated spatial framework that defines where variables can be located (positions, distances, directions) within a digital system.</p>
</def>
</def-item>
</def-list>
</glossary>
</back>
</article>