<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="research-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id><journal-title-group>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title></journal-title-group>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2026.1616955</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Study Protocol</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>SPEAK-SAFE: secure processing of electronic audio for knowledge in suicide assessment from therapeutic exchanges</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name><surname>Landau</surname><given-names>Christopher</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref>
<xref ref-type="author-notes" rid="an1"><sup>&#x2020;</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/3043666/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
<contrib contrib-type="author" equal-contrib="yes"><name><surname>Getty</surname><given-names>Patricia</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="an1"><sup>&#x2020;</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/3249072/overview" /><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
<contrib contrib-type="author"><name><surname>Gruler</surname><given-names>Caroline</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
<contrib contrib-type="author"><name><surname>Stadje</surname><given-names>Rebekka</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
<contrib contrib-type="author"><name><surname>Arampatzi</surname><given-names>Sofia</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Mandal</surname><given-names>Aishik</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Goel</surname><given-names>Anmol</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Gurevych</surname><given-names>Iryna</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Reif</surname><given-names>Andreas</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/30307/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Grimm</surname><given-names>Oliver</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/768976/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Psychiatry, Psychosomatic Medicine and Psychotherapy, University Hospital Frankfurt</institution>, <city>Frankfurt am Main</city>, <country country="de">Germany</country></aff>
<aff id="aff2"><label>2</label><institution>Ubiquitous Knowledge Processing Lab (UKP Lab), Department of Computer Science and Hessian Center for AI (hessian.AI), Technical University of Darmstadt</institution>, <city>Darmstadt</city>, <country country="de">Germany</country></aff>
<aff id="aff3"><label>3</label><institution>Fraunhofer Institute for Translational Medicine and Pharmacology ITMP</institution>, <city>Frankfurt am Main</city>, <country country="de">Germany</country></aff>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Christopher Landau <email xlink:href="mailto:c.landau@rz.uni-frankfurt.de">c.landau@rz.uni-frankfurt.de</email></corresp>
<fn fn-type="equal" id="an1"><label>&#x2020;</label><p>These authors share first authorship</p></fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-24"><day>24</day><month>02</month><year>2026</year></pub-date>
<pub-date publication-format="electronic" date-type="collection"><year>2026</year></pub-date>
<volume>8</volume><elocation-id>1616955</elocation-id>
<history>
<date date-type="received"><day>23</day><month>04</month><year>2025</year></date>
<date date-type="rev-recd"><day>27</day><month>01</month><year>2026</year></date>
<date date-type="accepted"><day>30</day><month>01</month><year>2026</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 Landau, Getty, Gruler, Stadje, Arampatzi, Mandal, Goel, Gurevych, Reif and Grimm.</copyright-statement>
<copyright-year>2026</copyright-year><copyright-holder>Landau, Getty, Gruler, Stadje, Arampatzi, Mandal, Goel, Gurevych, Reif and Grimm</copyright-holder><license><ali:license_ref start_date="2026-02-24">https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p></license>
</permissions>
<abstract><sec><title>Background</title>
<p>For therapists, the spoken word of their patients is among the most important foundations for clinical assessment. At the same time, it is hardly possible to monitor patients continuously and closely in sufficient numbers, for example, to ongoingly assess the risk of suicide in therapeutical conversations. Natural Language Processing (NLP) involves the use of Artificial Intelligence (AI) to analyze human language. Combining it with AI speech processing methods, we obtain multimodal methods which can automatically process large volumes of speech and language data to extract diagnostic information and therefore support individualized treatment plans. Thus, in NLP/multimodal methods, we see the opportunity to significantly improve patient care.</p>
</sec><sec><title>Methods</title>
<p>The SPEAK-SAFE project, implemented by clinicians and clinical researchers from the University hospital in Frankfurt in collaboration with the AI experts from the TU Darmstadt, aims to create the first German psychiatric corpus for evaluating and developing multimodal and NLP models to optimize diagnostic processes in psychiatric, psychosomatic, and psychotherapeutic care. Therefore, we will collect therapist-patient dialogues during therapy sessions. This sensitive data necessitates robust privacy. To meet this requirement, all collected data is pseudonymized, to ensure that no personal data is part of the evaluation and training of the AI models.</p>
</sec><sec><title>Discussion</title>
<p>During the implementation of our research project, we were faced with challenges regarding the security of patient privacy and the technical implementation of therapy recordings toreassure sufficient data quality for the data analysis. Therefore, in addition to improve the suicidality prediction with multimodal methods we will develop an end-to-end-workflow for further AI-research in the clinical context.</p>
<p><bold>Clinical Trial Registration</bold>: <ext-link ext-link-type="uri" xlink:href="https://drks.de/search/de/trial/DRKS00027878">https://drks.de/search/de/trial/DRKS00027878</ext-link>, identifier DRKS00027878.</p>
</sec>
</abstract>
<kwd-group>
<kwd>artificial intelligence</kwd>
<kwd>German speech dataset</kwd>
<kwd>natural language processing</kwd>
<kwd>suicide risk assessment</kwd>
<kwd>therapeutic dialogue analysis</kwd>
</kwd-group><funding-group><funding-statement>The author(s) declared that financial support was received for this work and/or its publication. The project is funded within the national research center for cybersecurity ATHENE [athene-center.de/en/] as part of the project PADAM (Privacy- Aware Domain-Adaptive Medical NLP).</funding-statement></funding-group><counts>
<fig-count count="1"/>
<table-count count="3"/><equation-count count="0"/><ref-count count="69"/><page-count count="12"/><word-count count="0"/></counts><custom-meta-group><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Digital Mental Health</meta-value></custom-meta></custom-meta-group>
</article-meta>
</front>
<body><sec id="s2" sec-type="intro"><title>Introduction</title>
<sec id="s2a"><title>Background and rationale</title>
<p>According to the Federal Statistical Office of Germany suicide led to more than 10,000 deaths in 2024 (<xref ref-type="bibr" rid="B1">1</xref>) and is the most common cause of death among young people between the age of 10&#x2013;25 in 2023 (<xref ref-type="bibr" rid="B2">2</xref>). In most suicide cases, individuals suffered from mental health disorders with rates ranging from 87,3&#x0025; (<xref ref-type="bibr" rid="B3">3</xref>) up to 96,8&#x0025; (<xref ref-type="bibr" rid="B4">4</xref>). Preventing suicides or suicide attempts is the goal of every psychiatric patient treatment and requires an ongoing suicide risk assessment. Thereby, most people committing suicide consult clinicians in the days or weeks prior to death (<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B6">6</xref>). However, predictive markers are currently not sufficiently able to predict suicide risk. Clinicians rely on spoken self-reports to gain insights into the patient&#x0027;s psychological state. Speech comprises a variety of aspects which carry information about someone&#x0027;s psychological state, like e.g., the content, speaking rhythm, loudness, pitch, pace, or coherence (<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B8">8</xref>). The diversity of content-related linguistic features that have been shown to be relevant for diagnostic assessment (<xref ref-type="bibr" rid="B9">9</xref>) and for the identification of suicidal ideation (<xref ref-type="bibr" rid="B10">10</xref>) is so extensive that a comprehensive evaluation of these features during an ongoing clinical conversation is unrealistic. It is therefore plausible that, along other factors, this linguistic complexity contributes to the limited consistency observed in suicide risk assessments (<xref ref-type="bibr" rid="B11">11</xref>). Further, some of the aspects of speech are subtle and hard to assess reliably by a human listener during a conversation, like e.g., jitter, shimmer, or harmonics-to-noise ratio, but important for the understanding of psychiatric illness&#x0027; like depression (<xref ref-type="bibr" rid="B12">12</xref>). Thus, to implement meaningful preventive measures, there is an urgent need for methods able to assess all important predictive markers related to suicide risk and mental disorders in general. Recent advances in Natural Language Processing (NLP), a field within artificial intelligence (AI), show promising results in mental disorder detection and an upward trend within this research area (<xref ref-type="bibr" rid="B13">13</xref>). Within NLP methods, linguistics and computer science enable large-scale analysis of language which are also applicable to the clinical psychological practice, where language plays a crucial role in patient care. NLP can be applied to various kinds of text data like social media posts (<xref ref-type="bibr" rid="B14">14</xref>&#x2013;<xref ref-type="bibr" rid="B16">16</xref>), interviews (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B18">18</xref>), and clinical notes (<xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B20">20</xref>). For the clinical psychiatric context, studies show promising results for the appliance of NLP methods. Regarding the field of suicide and suicide risk research, NLP models can successfully identify patients who have attempted suicide (<xref ref-type="bibr" rid="B20">20</xref>), predict suicidal ideation (<xref ref-type="bibr" rid="B17">17</xref>), and distinguish between suicidal and non-suicidal teenagers (<xref ref-type="bibr" rid="B21">21</xref>). Furthermore, suicidal language appears to be a stable classifier when compared to standard measurements. This stability is evident as suicidal language persists for up to 30 days after discharge from care, even when responses to other measurements change (<xref ref-type="bibr" rid="B18">18</xref>). Accordingly, there are multiple NLP-based methods which enable the exploration of previously unexamined linguistic features.</p>
<p>The aforementioned approaches have been somewhat overshadowed since the introduction of the Transformer architecture (<xref ref-type="bibr" rid="B22">22</xref>) and its fast development and huge impact, the evaluation of speech data has totally changed. Models have grown tremendously in a short period of time e.g., GPT3 is more than 100 times bigger than its predecessor GPT2 (<xref ref-type="bibr" rid="B23">23</xref>). Due to this rapid development, it is now possible to use even large amounts of speech data, such as clinical interviews, for evaluation with large language models (LLMs). Modern LLMs are not only ideal for general purposes, they also show capabilities when it comes to specific tasks in the domain of mental health (<xref ref-type="bibr" rid="B24">24</xref>). It could be demonstrated that LLMs like GPT-3.5 and GPT-4 are able to correctly identify patients with depression. However, they are known to show a considerable language bias and due to their nonspecialized training data, they are better at identifying control participants which could narrow their direct applicability for mental health applications (<xref ref-type="bibr" rid="B25">25</xref>). Models specialized on medical knowledge like Med-PaLM-2 can demonstrate human-like performance when predicting PHQ8 scores used for the assessment of the severity of depression (<xref ref-type="bibr" rid="B26">26</xref>). Besides spoken language many studies utilize written language like social media data and show promising results even with simple prompting strategies (<xref ref-type="bibr" rid="B27">27</xref>). Modern LLMs capabilities go way beyond this and offer numerous applications in addition to those mentioned before. Other studies indicate their usefulness in giving treatment suggestions (<xref ref-type="bibr" rid="B28">28</xref>) or creating simulated patients for psychotherapy training (<xref ref-type="bibr" rid="B29">29</xref>).</p>
<p>Apart from the content it is also important to take acoustic features accompanying the linguistic content in speech, also called prosodic and intonational features into account. These features are beyond the content related aspects of speech like phonemes and comprise patterns of pitch, duration, intensity, and rhythm, but are important in conveying meaning, emotion and nuance in speech (<xref ref-type="bibr" rid="B30">30</xref>). Recent studies focusing on acoustic features (<xref ref-type="bibr" rid="B31">31</xref>, <xref ref-type="bibr" rid="B32">32</xref>) confirm that detection and severity assessment of mental health conditions like depression can be achieved with high accuracy based solely on acoustic information, demonstrating that vocal markers provide a reliable proxy for underlying affective states. In addition to research focusing on one modality, there are studies that consider both linguistic and acoustic features in a single study (<xref ref-type="bibr" rid="B33">33</xref>).</p>
<p>With multimodal models, large scale language and acoustics data can be processed simultaneously to get a better prediction of e.g., suicidality (<xref ref-type="bibr" rid="B34">34</xref>) and depression (<xref ref-type="bibr" rid="B35">35</xref>). Those multifaceted insights can greatly enrich existing knowledge about diseases like depression or schizophrenia and thus improve patient diagnosis and care. Moreover, a multimodal approach enables a much more realistic evaluation of speech data sets and is therefore useful in many counseling situations.</p>
</sec>
</sec>
<sec id="s3"><title>Objectives</title>
<p>Considering the lack of linguistic diversity in datasets with a clear bias towards English (<xref ref-type="bibr" rid="B13">13</xref>, <xref ref-type="bibr" rid="B36">36</xref>) and social media data (<xref ref-type="bibr" rid="B24">24</xref>), we see an urgent need for further non-English speech datasets (<xref ref-type="bibr" rid="B37">37</xref>). To this day there is no sufficiently large German speech corpus from severely ill psychiatric patients. There have been several valuable efforts to advance the analysis of language in the context of psychotherapy and mental health within German-speaking populations. However, existing studies have primarily focused on outpatient samples with milder symptom severity (<xref ref-type="bibr" rid="B38">38</xref>) or relied on relatively small datasets (<xref ref-type="bibr" rid="B39">39</xref>). While these contributions are undoubtedly significant, they nevertheless leave an important gap regarding large-scale datasets from more severely affected patients in inpatient settings. Existing corpora, such as the Research and Teaching Corpus of Spoken German (FOLK) (<xref ref-type="bibr" rid="B40">40</xref>), are not specifically designed to meet the needs and peculiarities of psychiatric and psychotherapeutic contexts, thus limiting their utility for NLP and multimodal methods in these fields. Even from a global perspective, language data sets are rather rare compared to the multitude of existing languages (<xref ref-type="bibr" rid="B41">41</xref>).</p>
<p>We therefore aim to generate a large enough trans-diagnostic collection of audio data of hospitalized patients and patients treated in our outpatient clinic. A study by Parola et al. (<xref ref-type="bibr" rid="B42">42</xref>) has shown several language specific differences regarding reduced semantic coherence in schizophrenia, which indicates a limited generalizability of speech patterns across different languages.</p>
<p>With our study, we aim to fill this gap by applying recent advances in NLP and multimodal methods. Our goal is to help identify non-explicit content language patterns in clinical patients, which can serve as prognostic markers specifically for the German language. Thus, our primary aim is to assess the suicide risk in patients with depression, bipolar disorder, and psychotic disorders, such as schizophrenia, schizoaffective disorder and acute polymorphic psychotic disorder (according to ICD-10), as well as the psychopathological development (Clinical Global Impression, CGI). As a secondary outcome we want to assess the overall psychopathological condition of the patients. This includes standardized and validated questionnaires comprising the clinician&#x0027;s impression of the patients general psychosocial functioning (Global Assessment of Functioning, GAF), suicidal symptoms (Depressive Symptom Index&#x2014;Suicidality Subscale, DSI-SS) and suicidality (Suicidal Scale of the Mini-International) as well as electronic health records (EHR) of the patient&#x0027;s psychopharmacological medication. Generally, the standardized text of the patient&#x0027;s EHR will be accessible for the text-based analysis. These will be supplemented by the standardized assessment through a clinician, which will be used as the gold standard for comparison with the AI analysis.</p>
</sec>
<sec id="s4"><title>Trial design</title>
<p>The study is multicentric, data will be collected via the University hospital in Frankfurt and in the Private Clinic Dr. Amelung in K&#x00F6;nigstein and will be processed by data scientists from the TU Darmstadt. The raw data will be stored exclusively on protected local servers at the University hospital Frankfurt and only pseudonymized data will be transferred to Darmstadt.</p>
</sec>
<sec id="s5"><title>Methods: participants, interventions and outcomes</title>
<sec id="s5a"><title>Study setting</title>
<p>Data will be collected in the University hospital in Frankfurt a.M. Additionally, one clinician and project member will collect data at the private hospital Dr. Amelung in K&#x00F6;nigstein. Data access is regulated via a data security protocol.</p>
</sec>
<sec id="s5b"><title>Eligibility criteria</title>
<p>We will include patients with ICD-10 diagnoses F20.x, F23, F25, F30-33, aged between 18 and 65 years, who are legally competent, without any issues in voice production, and have a German speaking proficiency of at least B2 (lower level in German would contain too little information content of linguistic utterances). Additionally, psychotherapists employed at the University Hospital Frankfurt will be included for the therapeutic conversations.</p>
</sec>
<sec id="s5c"><title>Who will take informed consent?</title>
<p>Informed consent will be contained by members of the study team who are not involved in the patients care. The patients are addressed at least one day prior to study participation and informed about the study. One day later, they will be addressed again and sign the consent if interested in participation.</p>
</sec>
</sec>
<sec id="s6"><title>Interventions</title>
<sec id="s6a"><title>Explanation for the choice of comparators</title>
<p>n/a: No comparator treatment was conducted. All participants equally engaged in receiving therapy sessions and completing questionnaires.</p>
</sec>
<sec id="s6b"><title>Intervention description</title>
<p>The participation will consist of the recordings of the therapeutic clinical conversations that are carried out in the regular clinical practice. In addition, for our study a student assistant will place a recorder and microphones before the therapy session and pick it up afterwards. This student assistant will be absent during the therapy session and only install the recording hardware. Further, after the recorded therapy session the clinicians and participants are required to fill out questionnaires regarding the clinical state and suicide risk of the patient on the same day as the therapy session was recorded.</p>
</sec>
<sec id="s6c"><title>Criteria for discontinuing or modifying allocated interventions</title>
<p>Study participation will be discontinued if the patients or their therapist request to discontinue participation or if the patient&#x0027;s clinical treatment is completed.</p>
</sec>
<sec id="s6d"><title>Strategies to improve adherence to interventions</title>
<p>To maintain adherence as much as possible, we strive to integrate the participation process seamlessly into the clinical routine of both patients and therapists. The only additional effort required for participation is approximately 5&#x2005;min to fill out questionnaires after a recorded therapy session.</p>
</sec>
<sec id="s6f"><title>Provisions for post-trial care</title>
<p>n/a: we do not expect that patients suffer any harm from participation, since the recordings are conducted as part of their regular therapeutic care and participants additionally only fill out questionnaires.</p>
</sec>
<sec id="s6g"><title>Outcomes</title>
<p>As part of an initial exploratory analysis, the assessment of psychopathology by treating physicians is to be compared with the evaluation by a computer model. Therefore, using intraclass correlation coefficients or t- tests, the hypothesis &#x201C;There is no difference in the assessment of patients regarding the severity of illness and suicidality&#x201D; will be tested.</p>
</sec>
<sec id="s6h"><title>Participant timeline</title>
<p>In-patients and out-patients will regularly be screened for participation eligibility. If the eligibility criteria are met and in consultation with their treating clinicians, patients will be informed about their study and will have at least one day to consider participation and sign the informed consent. The same procedure applies for clinicians who participate in the study, while they only must consent once for the recording of therapy sessions with several patients who gave informed consent. From the time point on where patients agreed to participate, up to five of their regularly held therapy sessions will be recorded. After the therapy session, patients and therapists will fill out questionnaires about the psychopathology of the patient (approx. 5&#x2005;min. for each). They are asked to fill out the questionnaire on the same day the therapy session took place. <xref ref-type="table" rid="T1">Table&#x00A0;1</xref> summarizes the procedure and when the respective questionnaires are used.</p>
<table-wrap id="T1" position="float"><label>Table&#x00A0;1</label>
<caption><p>Content for the schedule of enrollment and participation.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left" rowspan="2">TIMEPOINT</th>
<th valign="top" align="center">Enrollment</th>
<th valign="top" align="center">Informed Consent</th>
<th valign="top" align="center" colspan="5">Recording and Questionnaires</th>
</tr>
<tr>
<th valign="top" align="center"><italic>&#x2212;t2</italic></th>
<th valign="top" align="center"><italic>&#x2212;t1</italic></th>
<th valign="top" align="center"><italic>t1</italic></th>
<th valign="top" align="center"><italic>t2</italic></th>
<th valign="top" align="center"><italic>t3</italic></th>
<th valign="top" align="center"><italic>t4</italic></th>
<th valign="top" align="center"><italic>t5</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="3"/>
<td valign="top" align="center" rowspan="3">Ongoing</td>
<td valign="top" align="center" rowspan="3">After confirmation by clinician</td>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="5">At least 1 week between each participation</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="5">Up to five therapy sessions can be recorded per participant</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="5">Questionnaires have to be filled out the same day as the recording</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="8"><italic>Study admission</italic></td>
</tr>
<tr>
<td valign="top" align="left">Eligibility screen</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
</tr>
<tr>
<td valign="top" align="left">Consultation with the treating clinician</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
</tr>
<tr>
<td valign="top" align="left">Informed consent</td>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
</tr>
<tr>
<td valign="top" align="left"><italic>Recording of therapy session</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="8"><italic>Patient reported outcomes (PRO)</italic></td>
</tr>
<tr>
<td valign="top" align="left"><italic>Beck&#x2019;s Depression Inventory (BDI-II)</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
<tr>
<td valign="top" align="left"><italic>DSI-SS-Scale</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
<tr>
<td valign="top" align="left">MINI C Suicidality Scale</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
</tr>
<tr>
<td valign="top" align="left"><italic>Demographics, Diagnoses and Concomitant Medication</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="8"><italic>Therapist Ratings</italic></td>
</tr>
<tr>
<td valign="top" align="left"><italic>Global Assessment of Functioning (GAF)</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
<tr>
<td valign="top" align="left"><italic>Clinical Global Impression (CGI)</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
<tr>
<td valign="top" align="left"><italic>Brief Psychiatric Rating Scale (BPRS)</italic></td>
<td valign="top" align="center"/>
<td valign="top" align="center"/>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
<td valign="top" align="center">X</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s6i"><title>Sample size</title>
<p>To determine the required sample size for comparing the accuracy of our models against the gold standard of a trained psychiatrist&#x0027;s evaluation, we conducted an <italic>a priori</italic> power analysis using G&#x002A;Power software (<xref ref-type="bibr" rid="B43">43</xref>).</p>
<p>We based our effect size estimate on the accuracies reported in previous studies that used machine learning techniques to predict suicidality from clinical data (<xref ref-type="bibr" rid="B44">44</xref>). Pestian et al. (<xref ref-type="bibr" rid="B45">45</xref>) reported an accuracy of 85&#x0025;, Bernert et al. (<xref ref-type="bibr" rid="B46">46</xref>) reported 90&#x0025; among 87 studies in their meta-analysis, and Barak-Corren et al. (<xref ref-type="bibr" rid="B47">47</xref>) reported 94&#x0025;. The average accuracy across these studies is 89.2&#x0025;. Assuming a gold standard accuracy of 100&#x0025;, we calculated an effect size (Cohen&#x0027;s g) of 0.11 using the formula: g&#x2009;&#x003D;&#x2009;(P&#x2014;Po)/&#x221A;[Po(1&#x2014;Po)], where <italic>P</italic> is the expected accuracy (0.892) and Po is the gold standard accuracy (1.0).</p>
<p>For our power analysis, we set the following parameters: a two-tailed test, an alpha level of 0.05, a power of 0.80, and a constant proportion of 0.5 (representing a balanced dataset). The results indicated that a total sample size of 162 is required to detect an effect size of 0.11 with 80&#x0025; power at a significance level of 0.05. The critical values for the lower and upper bounds of the acceptance region are 68 and 94, respectively. The actual power achieved with this sample size is 0.8046757, and the actual alpha level is 0.0491726.</p>
<p>In summary, to compare the accuracy of our models against the gold standard of a trained psychiatrist&#x0027;s evaluation with 80&#x0025; power and a 5&#x0025; significance level, we require a total sample size of <italic>n</italic>&#x2009;&#x003D;&#x2009;162, assuming an effect size of 0.11 based on the accuracies reported in previous comparable studies. As we expect a drop-out rate of 15&#x0025;&#x2013;20&#x0025; due to technical issues, quality issues etc., we expect to collect <italic>n</italic>&#x2009;&#x003D;&#x2009;200 interviews.</p>
</sec>
<sec id="s6j"><title>Recruitment</title>
<p>The participants are recruited from both inpatient and outpatient individuals at the Department of Psychiatry, Psychosomatics, and Psychotherapy at the University Hospital Frankfurt. Additionally, a physician collects data at the Dr. Amelung Private Clinic in K&#x00F6;nigstein. For ongoing enrollment, patients will be screened regularly for participation criteria by the study teams and by their treating clinicians. Study information will be given by members of the study team who are not involved with the patient&#x0027;s treatment to prevent participation out of a sense of obligation and to maintain voluntary participation. There will be no assignment of interventions since participation does not include differing intervention types.</p>
</sec>
</sec>
<sec id="s7"><title>Assignment of interventions: allocation</title>
<sec id="s7a"><title>Sequence generation</title>
<p>n/a: since participation comprises the recording of up to five regularly held therapy sessions and subsequent filling out of questionnaires which will be equally done by all study participants, sequence generation is not needed.</p>
</sec>
</sec>
<sec id="s8"><title>Data collection and management</title>
<sec id="s8a"><title>Plans for assessment and collection of outcomes</title>
<p>The recording of the therapy sessions and the questionnaire data will be collected on the same day. An overview of all applied study instruments is depicted in <xref ref-type="table" rid="T2">Table&#x00A0;2</xref>. After giving the patient the study information and obtaining informed consent, audio recordings are conducted during routine medical activities. Up to five of their therapy sessions with a typical duration of 30&#x2013;60&#x2005;min are recorded. Audio recordings of therapeutic sessions are captured using high-quality recording equipment with carefully positioned microphones to minimize background noise and ensure clear speech capture. Recordings are made using a Tascam DR-40X audio recorder. Recordings are captured in stereo mode using the recorder&#x0027;s external left/right inputs. The audio is stored as uncompressed WAV files with a sampling rate of 48&#x2005;kHz and a bit depth of 24 bit. The recorder input level is set to &#x2212;12&#x2005;dB, providing sufficient headroom to avoid clipping. Phantom power (48&#x2005;V) is enabled to supply the external microphones. Audio signals are acquired using two external condenser microphones (the t.bone CC 75) with a frequency response of 20&#x2005;Hz&#x2013;20&#x2005;kHz, corresponding to the standard human auditory range. Microphones are attached at the chest/collar area of the participants ensuring comparable distance to the mouth between the participants. Care is taken to ensure that the microphone grille did not come into contact with participants&#x0027; clothing and is oriented toward the participant&#x0027;s head, in order to minimize clothing-rub noise and to enable more accurate recordings. These settings were chosen based on previous test runs using different setups of microphones and recording environments/parameters. We transcribed and diarized the recordings and manually evaluated the quality of these transcriptions with their corresponding audio recordings to find the best setup.</p>
<table-wrap id="T2" position="float"><label>Table&#x00A0;2</label>
<caption><p>Overview of collected data.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left" colspan="2">Data category</th>
<th valign="top" align="center" colspan="2">Purpose of use</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Demographic data</td>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Control of the study<italic>&#x2019;</italic>s inclusion criteria</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="10">Clinical data</td>
<td valign="top" align="left">Clinical Global Impression (CGI)</td>
<td valign="top" align="left" rowspan="2">Dimensional measure of development of psychopathology</td>
<td valign="top" align="left" rowspan="10">Assessment of the patient<italic>&#x2019;</italic>s condition or psychopathology</td>
</tr>
<tr>
<td valign="top" align="left">Clinical Global Impression of Suicide Severity (CGI-SS)</td>
</tr>
<tr>
<td valign="top" align="left">Global Assessment of Functioning (GAF)</td>
<td valign="top" align="left">Assessment of the general psychosocial functioning level</td>
</tr>
<tr>
<td valign="top" align="left">MiNi C Suicidality Scale (MiNi)</td>
<td valign="top" align="left">Assessment of suicide risk and preceded suicidality</td>
</tr>
<tr>
<td valign="top" align="left">Suicidality Scale of the Depressive Symptom Index (DSI-SS)</td>
<td valign="top" align="left">Measuring instrument for recording suicidality</td>
</tr>
<tr>
<td valign="top" align="left">Beck<italic>&#x2019;</italic>s Depression Inventory (BDI- II)</td>
<td valign="top" align="left">Severity detection</td>
</tr>
<tr>
<td valign="top" align="left">Brief Psychiatric Rating Scale (BRPS)</td>
<td valign="top" align="left">Severity detection</td>
</tr>
<tr>
<td valign="top" align="left">Additional information obtained from medical record</td>
<td valign="top" align="left">Text based analysis</td>
</tr>
<tr>
<td valign="top" align="left">Medical history</td>
<td valign="top" align="left"/>
</tr>
<tr>
<td valign="top" align="left">Concomitant Medication</td>
<td valign="top" align="left"/>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Recording of the therapy sessions</td>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Analysis of speech features</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Additionally, data on general psychosocial functioning and the severity of the patient&#x0027;s psychopathological state are collected using questionnaires completed by both therapists and patients.</p>
<p>The Clinical Global Impression (CGI) is a short assessment of the clinician&#x0027;s view of the patient&#x0027;s global functioning and to track patient progress or treatment response over time and comprises three measures of which we use two: severity of illness, global improvement, and an efficacy index (which we leave out) (<xref ref-type="bibr" rid="B48">48</xref>). The instrument has been proven to be a readily understood and practical measurement tool which is well applicable in the clinical practice (<xref ref-type="bibr" rid="B49">49</xref>) and has been used in different mental disorders including affective and psychotic disorders (<xref ref-type="bibr" rid="B50">50</xref>, <xref ref-type="bibr" rid="B51">51</xref>). Additionally, we use the CGI-SS which is a modified version of the CGI to evaluate severity of and change in suicidality according to the version of Meltzer et al. (<xref ref-type="bibr" rid="B52">52</xref>). We only changed the time period from &#x201C;past 7 days&#x201D; to &#x201C;since the last assessment&#x201D;.</p>
<p>The Global Assessment of Functioning (GAF) scale is an established instrument for measuring illness severity in the clinical context and introduced as a rating scale within the DSM-III-R (<xref ref-type="bibr" rid="B53">53</xref>). The instrument has shown to be a valid and reliable measure of psychiatric illness (<xref ref-type="bibr" rid="B54">54</xref>).</p>
</sec>
<sec id="s8b"><title>Plans to promote participant retention and complete follow-up</title>
<p>To promote retention of the participating patients and their therapists, we are integrating the participation progress in everyday clinical care with minimal extra effort for the participation process. If therapists or patients nevertheless want to end their participation, there will be no consequences or disadvantages.</p>
</sec>
<sec id="s8c"><title>Data management</title>
<p>During the data collection phase, the raw audio recordings, patient data, and questionnaires will be collected. In this phase, only clinicians and student assistants from the University hospital in Frankfurt as well as one doctoral student from the Privatklinik Dr. Amelung in K&#x00F6;nigstein are involved. The analysis of the data is conducted by the Ubiquitous Knowledge Processing (UKP) Lab at TU Darmstadt using natural language processing and multimodal methods and is mostly carried out in Darmstadt exclusively with pseudonymized data. The data accessed by the research group from TU Darmstadt will only have pseudonymized files and data without any personal information about the participants. The data analysis is complemented by a standardized assessment of the course by an independent board-certified psychiatrist or experienced assistant in psychiatry, serving as a &#x201C;gold standard&#x201D; for comparison with the AI analysis. Thus, the pseudonymized audio files are listened to and classified by a board-certified psychiatrist based on clinical features. The gold standard assessment covers three different areas. It consists of the BPRS as a broad psychopathology questionnaire, the five-stage severity item of the CGI-SS combined with an open question, asking the experts to briefly justify their rating within 2&#x2013;3 sentences and the VTAS-R-Short Form, a five-item therapeutic alliance questionnaire (<xref ref-type="bibr" rid="B55">55</xref>).</p>
<p>The pseudonymized audio recordings will then be analyzed for acoustic and prosodic features using the open-source toolkit, openSMILE (<xref ref-type="bibr" rid="B56">56</xref>) and then transcribed to text with the freely available software WhisperX (<xref ref-type="bibr" rid="B57">57</xref>). Both datasets will then be analyzed on its own and later on fed to a multimodal neural network to generate a clinically relevant outcome score (e.g., suicidality prognosis score). <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref> shows an overview of our workflow.</p>
<fig id="F1" position="float"><label>Figure&#x00A0;1</label>
<caption><p>Schematic depiction of data collection and analysis workflow.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1616955-g001.tif"><alt-text content-type="machine-generated">Diagram showing a process where a microphone produces both a transcript and audio recording; text and acoustic features are extracted, processed through neural network layers, and combined to deliver a quantitative assessment output.</alt-text>
</graphic>
</fig>
<p>Both, audio track and therapy transcript have to be pseudonymized to ensure data privacy. The sessions are transcribed to text with the freely available software WhisperX (<xref ref-type="bibr" rid="B57">57</xref>). In order to pseudonymize the data, the members of the study team listen to the therapy session and check the transcript at the same time. In this way it is not only possible to delete personal data but also keep track of transcription errors. Furthermore, this also enables the correction of errors in speaker diarization which will be relevant for data analysis. Transcripts are also edited according to predefined rules to ensure consistency, including standardized notation for stuttering and speech disfluencies, clear indication of overlapping speech, flagging of incomprehensible sections, and documentation of interruptions such as third parties entering. By doing this we are left with not only a pseudonymized audio track but also with a high quality pseudonymized and diarized transcript. Afterwards, the transcripts are ready for use with e.g., locally operating LLMs to assess their capability to distinguish between patient groups and evaluate symptom severity. The audio tracks are ready for feature extraction with the open-source toolkit openSMILE (<xref ref-type="bibr" rid="B56">56</xref>). Extracted audio features will serve as the basis for subsequent calculations, e.g., correlations with specific symptoms as well as general severity scores. Details of the unimodal and multimodal analyses are provided in the Statistical Methods section.</p>
</sec>
<sec id="s8d"><title>Confidentiality</title>
<p>A dedicated data security concept for this study was developed in close collaboration with the local data protection officer, fully compliant with the EU General Data Protection Regulation [GDPR, Regulation (EU) 2016/679] and the German Federal Data Protection Act (Bundesdatenschutzgesetz, BDSG). The measures ensure both technical and organizational protection of personal data throughout the project. Key provisions include strict pseudonymization and, where feasible, anonymization of all personal data before processing. Access to sensitive data is restricted through a role-based authorization system, and all project staff are trained in privacy and IT security procedures. Data transfers are conducted via encrypted communication channels (TLS/SSL), and stored data is protected using state-of-the-art encryption algorithms. Routine system audits, monitoring, and regular penetration testing are implemented to detect and mitigate vulnerabilities. Redundant backups are encrypted and stored in physically secure environments, with access strictly controlled. Detailed data usage logs and audit trails are maintained to record all data access, modifications, and deletions. The network infrastructure is protected by firewalls and intrusion detection systems, and written informed consent is mandatory for all data subjects. These measures collectively ensure the confidentiality, integrity, and availability of research data, safeguarding data subjects&#x0027; rights according to EU and German law. Specifically, we implemented the following measures to ensure data security. For every enrolled patient a participation ID will be generated via SecuTrial, a software for managing and conducting clinical trials, offering data collection, management, and analysis with high security and quality standards. One secured document which is saved locally in the University hospital Frankfurt contains the assigned patient identities to the SecuTrial-IDs. All confidential data will be stored exclusively in the university hospital Frankfurt. The questionnaire data will be transferred to SecuTrial and stored under the assigned participant ID. The raw audio files and their transcripts are then pseudonymized by a member of the study group. This involves deleting elements that can be clearly attributed to the patient or the clinician, such as names, addresses, institutions, or other personal information. All files are pseudonymized in compliance with data protection regulations and stored under the assigned participants ID. Password-protected Excel file (256-bit key) with the association of real names is stored in the hospital network. Only designated study personnel with identifiable names have knowledge of this password. Access to the system is logged in a log file.</p>
</sec>
</sec>
<sec id="s9"><title>Statistical methods</title>
<sec id="s9a"><title>Statistical methods for primary and secondary outcomes</title>
<p>We will employ a zero-shot prompting approach with OpenAI&#x0027;s recently released GPT-OSS (Generative Pretrained Transformer&#x2014;Open Source Software) series models, which are available under the permissive Apache 2.0 license and have demonstrated strong performance on reasoning tasks relevant to clinical assessment. We have selected these models for several methodological and practical reasons:</p>
<p>First, their open-weight nature ensures full transparency and reproducibility of our analyses. Second, their mixture-of-experts architecture enables efficient deployment on clinical infrastructure with limited computational resources. Third, their documented performance on complex reasoning tasks, including healthcare-related queries, makes them well-suited for psychiatric assessment applications.</p>
<p>Our primary analytical comparison will involve two model sizes from the GPT-OSS family: GPT-OSS-120B (116.8 billion total parameters with 5.1 billion active parameters per token) and GPT-OSS-20B (20.9 billion total parameters with 3.6 billion active parameters per token) (<xref ref-type="bibr" rid="B58">58</xref>). This comparison directly addresses a critical question for clinical deployment: whether larger, more capable models provide sufficient performance advantages to justify their greater computational demands, or whether smaller, more efficient models can achieve comparable diagnostic accuracy. The GPT-OSS-120B model has been shown to achieve near-parity with state-of-the-art reasoning models on standard benchmarks while running efficiently on a single 80GB GPU, whereas GPT-OSS-20B delivers comparable performance to advanced compact models while requiring only 16GB of memory, making it feasible for deployment on standard clinical workstations or even edge devices.</p>
<p>For each model size, we will systematically evaluate two distinct inference configurations defined by the temperature hyperparameter, which controls the randomness of model outputs. We will test a conservative configuration (temperature&#x2009;&#x003D;&#x2009;0.1) designed to prioritize consistency and clinical reliability by producing more deterministic responses, and a rather liberal configuration (temperature&#x2009;&#x003D;&#x2009;0.8). In the NLP literature the effects of temperature are well represented but systematic clinical applications have yet to be done to our knowledge. Accuracy in multiple choice question answering tasks show high consistency in temperature ranging between 0 and 1, but an expansion to other problem types is recommended (<xref ref-type="bibr" rid="B59">59</xref>). This strategic selection of hyperparameters reflects the clinical tension between reproducibility critical for diagnostic applications where consistency across assessments is paramount and the risk of highly inconsistent outcomes that could potentially lead to incorrect diagnoses and treatment recommendations. In addition, we apply top-p (nucleus) sampling with a cumulative probability threshold of 0.9 to each analysis. This setting restricts sampling to the smallest set of tokens whose cumulative probability mass is at least 0.9, balancing output diversity with coherence.</p>
<p>This yields a 2&#x2009;&#x00D7;&#x2009;2 factorial design with four experimental conditions: 1. GPT-OSS-120B with temperature 0.1, 2. GPT-OSS-120B with temperature 0.8, 3. GPT-OSS-20B with temperature 0.1, and 4. GPT-OSS-20B with temperature 0.8. We will evaluate each condition&#x0027;s performance on primary outcomes including general diagnostic classification accuracy and suicidality detection, severity score prediction (correlation with standardized clinical measures), and clinical utility metrics such as sensitivity, specificity, and positive/negative predictive values. All model versions, hyperparameters, and inference code will be fully documented and made available to ensure complete reproducibility.</p>
<p>Concurrent with text-based analysis using GPT-OSS models, we will extract acoustic features from therapy session recordings using openSMILE (version 3.0) (<xref ref-type="bibr" rid="B56">56</xref>) with the eGeMAPSv02 feature set (<xref ref-type="bibr" rid="B60">60</xref>), which includes validated markers known to associate with psychiatric conditions. The eGeMAPSv02 feature set comprises 88 acoustic features encompassing spectral (Mel-frequency cepstral coefficients), glottal (fundamental frequency, jitter, shimmer), and prosodic characteristics (energy, speech rate, voice quality measures such as harmonics-to-noise ratio). These features will be extracted at the frame-level (10-millisecond intervals) and subsequently aggregated using statistical functions (mean, standard deviation, minimum, maximum, percentiles) to produce session-level representations comparable to clinical assessment timeframes.</p>
<p>Our overarching analytical strategy employs a three-tier statistical framework designed to characterize how textual, acoustic, and combined modalities contribute to clinical prediction:
<list list-type="simple">
<list-item>
<p>Tier 1: Unimodal Baseline Analysis. We will analyze each modality independently to establish baseline predictive performance before considering fusion. For the text modality, we will conduct two-way mixed-effects logistic regression with outcomes (diagnostic classification or severity degree) regressed on the fixed effects of model size and temperature, with random intercepts for participant and session. Effect sizes will be reported as odds ratios with 95&#x0025; confidence intervals. For the acoustic modality, we will conduct mixed-effects linear regression using a reduced set of the most clinically relevant acoustic features selected <italic>a priori</italic> based on prior literature associations with depression and suicidality, including pitch variability, energy measures, MFCC and jitter as predictors of continuously scored clinical outcomes (e.g., severity scores from standardized rating scales) (<xref ref-type="bibr" rid="B60">60</xref>, <xref ref-type="bibr" rid="B61">61</xref>).</p></list-item>
<list-item>
<p>Tier 2: Factorial Interaction Analysis for 2&#x2009;&#x00D7;&#x2009;2 Design. For text-based predictions, we will formally test the 2&#x2009;&#x00D7;&#x2009;2 factorial design using mixed-effects ANOVA to partition variance according to main effects (model size, temperature) and their interaction. Specifically, we will model primary outcome (e.g., predicted depression severity) as a function of fixed effects for model size (GPT-OSS-120B vs. 20B), temperature (0.1 vs. 0.8), and their interaction (Model Size&#x2009;&#x00D7;&#x2009;Temperature), with random intercepts for participants and sessions to account for the hierarchical data structure. This approach allows us to determine whether inference strategy recommendations (conservative vs. liberal temperature) can be applied consistently across model sizes, or whether optimization differs by model choice. If a significant Model Size&#x2009;&#x00D7;&#x2009;Temperature interaction emerges, it would suggest that clinical deployment requires model-specific tuning. Conversely, absence of interaction would support unified recommendations applicable regardless of computational constraints. We will report partial eta-squared (<italic>&#x03B7;</italic>&#x00B2;p) effect sizes with 95&#x0025; confidence intervals for each factorial term.</p></list-item>
<list-item>
<p>Tier 3: Multimodal fusion and Late-Fusion Integration. To evaluate how textual and acoustic information combine for clinical prediction, we will employ a structured late-fusion approach. For each of the four 2&#x2009;&#x00D7;&#x2009;2 factorial conditions, we will generate separate predictions from the GPT-OSS models (text modality) and from a logistic regression model trained on acoustic features (acoustic modality). We will then integrate these predictions through weighted averaging, with weights determined via cross-validation (80&#x0025; training, 20&#x0025; held-out validation) on a stratified subset of sessions. This approach maintains interpretability&#x2014;each modality&#x0027;s contribution can be quantified&#x2014;and avoids the architectural complexity that would introduce uncontrolled hyperparameter flexibility incompatible with pre-registration.</p></list-item>
</list>We will quantify multimodal benefit by comparing fusion performance (AUC, sensitivity/specificity at clinically relevant decision thresholds) against the best unimodal baseline. If fusion provides incremental improvement beyond single modalities, we will use mixed-effects logistic regression with interaction terms to test whether the fusion benefit depends on model size or temperature setting. This tests the hypothesis that acoustic information provides complementary information that enhances text-based predictions uniformly, or whether certain model configurations better leverage multimodal information.</p>
<p>Statistical rigor and assumptions testing. For all mixed-effects models, we will verify assumptions including normality of residuals (Shapiro&#x2013;Wilk test on standardized residuals), homogeneity of variance (Levene&#x0027;s test), and absence of multicollinearity (variance inflation factors&#x2009;&#x003C;&#x2009;5 for predictors). Violations of normality or homogeneity will be addressed through appropriate transformations (e.g., Box-Cox for acoustic features) or robust variance estimation. Missing data will be handled via maximum likelihood estimation under a missing-at-random assumption, standard for mixed-effects models.</p>
<p>Multiple comparisons within the 2&#x2009;&#x00D7;&#x2009;2 factorial framework will be controlled via Bonferroni correction applied within each outcome domain (text-based prediction, acoustic baseline, fusion). We will test all fixed effects (model size main effect, temperature main effect, Model Size&#x2009;&#x00D7;&#x2009;Temperature interaction) at <italic>&#x03B1;</italic>&#x2009;&#x003D;&#x2009;0.05 (two-tailed). <italic>post-hoc</italic> comparisons between specific factorial cells (e.g., testing whether GPT-OSS-120B at temperature 0.8 outperforms GPT-OSS-20B at temperature 0.1) will be pre-specified and corrected for multiple comparisons.</p>
<p>This simplified, focused approach ensures that our study adheres to pre-registration principles, prevents <italic>post-hoc</italic> methodological flexibility, and provides a reproducible framework that other research groups can directly replicate or build upon.</p>
</sec>
<sec id="s8e"><title>Plans to give access to the full protocol, participant level-data and statistical code</title>
<p>Access to the full protocol, participant-level data, and statistical code will be restricted and carefully controlled. Researchers interested in accessing these materials must submit a formal application, which will be subject to review by the Institutional Review Board (IRB) for local data processing. This process is designed to ensure the protection of sensitive information and maintain the integrity of the research. It is important to note that, in accordance with the decision of the data security board at the University of Frankfurt, no data can be shared outside of this controlled environment consisting of the study team in Frankfurt and selected members of the study team in Darmstadt.</p>
</sec>
</sec>
<sec id="s10"><title>Oversight and monitoring</title>
<sec id="s10a"><title>Composition of the coordinating center and trial steering committee</title>
<p>Director of Studies are PD Dr med. Oliver Grimm (OG) &#x0026; Prof. Dr. Andreas Reif (AR) from the Department of Psychiatry, Psychosomatics and Psychotherapy, University Hospital Frankfurt. Further, the UKP Lab headed by Prof. Dr. Iryna Gurevych (IG) will cooperate and is responsible for the processing of the data, feature extraction and building of an NLP/multimodal model. The trial steering committee consists of OG, AR, and IG, along with a senior statistician and an independent expert in psychiatry research. This committee will oversee the overall conduct of the study, monitor its progress, and make key decisions regarding the trial&#x0027;s direction and any necessary protocol amendments.</p>
</sec>
<sec id="s10b"><title>Composition of the data monitoring committee, its role and reporting structure</title>
<p>According to German law, no formal data monitoring committee is needed for a study without therapeutic intervention.</p>
</sec>
<sec id="s10e"><title>Plans for communicating important protocol amendments to relevant parties (e.g., trial participants, ethical committees)</title>
<p>Protocol modifications will firstly be sent to the central and local ethics committees. Once changes are approved, currently enrolled patients will be asked to sign the newest version of the informed consent form accordingly.</p>
</sec>
</sec>
<sec id="s11"><title>Dissemination plans</title>
<p>Study results will be communicated to the scientific community and general public via publications.</p>
</sec>
<sec id="s12" sec-type="discussion"><title>Discussion</title>
<p>The SPEAK-SAFE study protocol presents a novel approach to utilizing Natural Language Processing (NLP)/multimodal methods for analyzing therapist-patient conversations in psychiatric care. This discussion addresses key aspects of the study design, potential challenges, and implications for future research and clinical practice.</p>
<p>The sensitive nature of the conversational data collected in this study necessitates robust privacy measures. Our study represents an initial feasible approach to this matter. Future research should further address these issues to optimize data collection while ensuring adequate implementation of data security. Collaboration with institutions like ATHENE will be instrumental in developing innovative and secure solutions for mental health research.</p>
<p>The integration of cutting-edge cybersecurity measures not only enhances the robustness of our data protection protocols but also sets a precedent for future studies in this field. This interdisciplinary approach, combining mental health research with advanced cybersecurity practices, paves the way for more secure and efficient data handling in sensitive research areas.</p>
<p>In addition to the challenges arising from the strict data protection guidelines, recording therapy sessions presents several technical challenges that could impact data quality. There are indications that the quality of the final results is more negatively affected by poor audio quality rather than computational issues (<xref ref-type="bibr" rid="B62">62</xref>). Background noise, variations in speech volume, and acoustic properties of the recording environment are crucial influencing factors which may affect transcription accuracy and overall data quality in non-laboratory settings (<xref ref-type="bibr" rid="B63">63</xref>). To mitigate these issues, we have implemented quality assurance strategies, including the use of high-quality recording equipment and careful placement of microphones. Nevertheless, some degree of data loss or inaccuracy may be unavoidable, potentially impacting the AI model&#x0027;s performance.</p>
<p>What accuracy should be expected from an AI-based suicide evaluation? Large language models demonstrate strong performance in detecting depression and suicide risk from clinical narratives, with recent real-world validation substantially informing expectations for psychotherapy transcripts. In a recent cross-sectional study of 100 admission notes from a psychiatric hospital, a German-fine-tuned Llama-2 model achieved 87.5&#x0025; accuracy, 83.0&#x0025; sensitivity, and 91.8&#x0025; specificity in identifying suicidality status against expert ground truth, demonstrating robust performance across varied prompting strategies while preserving data privacy (<xref ref-type="bibr" rid="B64">64</xref>). To contextualize these capabilities, human interrater reliability in assessing mental health from dialogue provides an essential benchmark. Unstructured suicide risk assessment shows substantially lower concordance: psychotherapists demonstrate Gwet&#x0027;s AC&#x2081;&#x2009;&#x003D;&#x2009;0.44 with only 58&#x0025; identical risk classifications, while psychology students achieve AC&#x2081;&#x2009;&#x003D;&#x2009;0.35 (<xref ref-type="bibr" rid="B65">65</xref>). Unstructured psychiatric diagnosis using clinical interviews yields <italic>&#x03BA;</italic>&#x2009;&#x003D;&#x2009;0.24&#x2013;0.43 (fair at best), compared to <italic>&#x03BA;</italic>&#x2009;&#x003D;&#x2009;0.75 (substantial) when structured protocols guide assessment (<xref ref-type="bibr" rid="B66">66</xref>). This empirical gap is instructive: current clinical practice relies on unstructured judgment with relatively poor reliability, particularly for suicide risk estimation. Consequently, LLM-assisted assessment of suicidality in psychotherapy transcripts is already useful potentially when achieving only modest accuracy performance.</p>
<p>The dialogic nature of therapy sessions introduces complexities in transcription and analysis. Colloquialisms, pauses, overlapping speech, and non-verbal cues are integral to therapeutic communication but challenging to capture and analyze (<xref ref-type="bibr" rid="B67">67</xref>). Our multimodal approach aims to address these challenges by incorporating advanced speech recognition and sentiment analysis techniques. However, the nuances of therapeutic dialogue may still pose difficulties in accurate interpretation, necessitating ongoing refinement of our multimodal models.</p>
<p>The SPEAK-SAFE study utilizes both structured (questionnaires) and unstructured (conversation recordings) data. While structured data offers standardized, easily quantifiable information, unstructured data provides rich, contextual insights into patient-therapist interactions (<xref ref-type="bibr" rid="B68">68</xref>). Our decision to focus on conversational data stems from its potential to capture subtle linguistic markers of mental health states that may not be evident in structured assessments. Our aim is to predict occurring changes in the structured data, what we consider as ground truth, by using the rich, unstructured speech data.</p>
<p>A limitation of the current study design is the lack of long-term follow-up to assess the predictive accuracy of the NLP/multimodal model over time. This has been done so far by only few studies (<xref ref-type="bibr" rid="B36">36</xref>). While our approach provides valuable insights into immediate risk assessment, the dynamic nature of mental health conditions necessitates longitudinal evaluation. Future research should incorporate extended follow-up periods to validate the model&#x0027;s long-term predictive capabilities and assess its utility in tracking treatment progress.</p>
<p>The use of AI in mental healthcare raises important ethical considerations. Potential biases in NLP/multimodal models, particularly regarding cultural and linguistic diversity, must be carefully addressed (<xref ref-type="bibr" rid="B69">69</xref>). Moreover, the integration of AI-based tools in clinical decision-making necessitates a balance between technological assistance and human clinical judgment. Our study emphasizes the importance of human oversight in interpreting AI-generated insights, recognizing that these tools should augment rather than replace clinical expertise.</p>
<p>The SPEAK-SAFE study has significant potential to enhance mental health care. By providing clinicians with AI-assisted risk assessment tools, we aim to improve the accuracy and efficiency of suicide risk evaluation (<xref ref-type="bibr" rid="B34">34</xref>). This could lead to more timely interventions and personalized treatment strategies. However, the successful integration of such tools into clinical workflows requires careful consideration of user acceptance, training needs, and potential impacts on the therapeutic relationship.</p>
<p>Building on the SPEAK-SAFE study, future research should explore the application of our AI models to a broader range of mental health conditions. Additionally, investigating the integration of multimodal data, including physiological measures and social media activity, could enhance the model&#x0027;s predictive capabilities (<xref ref-type="bibr" rid="B13">13</xref>). Longitudinal studies examining the long-term impact of AI-assisted care on patient outcomes are also crucial for validating the clinical utility of this approach. We plan to augment our data collection with smartphone-based ecological momentary assessment (EMA) to allow outcome prediction.</p>
<p>In conclusion, the SPEAK-SAFE study represents a significant step towards leveraging AI in mental health care. While challenges remain in data collection, analysis, and ethical implementation, the potential benefits in improving risk assessment and patient care are substantial. Continued research and interdisciplinary collaboration will be essential in refining these approaches and translating them into clinical practice.</p>
</sec>
</body>
<back>
<sec id="s13"><title>Trial status</title>
<p>Current Protocol version 6, 08.04.2025. Recruitment began on 1st of November 2023. Data collection will continue until further notice without end date.</p>
</sec>
<sec id="s14" sec-type="ethics-statement"><title>Ethics statement</title>
<p>The studies involving humans were approved by Ethikkommission des Fachbereichs Medizin der Goethe-Universit&#x00E4;tc/o Universit&#x00E4;tsklinikum &#x00B7;Theodor-Stern-Kai 7 60590 Frankfurt am Main. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants&#x0027; legal guardians/next of kin. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.</p>
</sec>
<sec id="s15" sec-type="author-contributions"><title>Author contributions</title>
<p>CL: Writing &#x2013; review &#x0026; editing, Writing &#x2013; original draft. PG: Writing &#x2013; review &#x0026; editing, Writing &#x2013; original draft. CG: Writing &#x2013; original draft. RS: Writing &#x2013; original draft. SA: Writing &#x2013; review &#x0026; editing. AM: Writing &#x2013; review &#x0026; editing. AG: Writing &#x2013; review &#x0026; editing. IG: Writing &#x2013; review &#x0026; editing. AR: Writing &#x2013; review &#x0026; editing. OG: Writing &#x2013; review &#x0026; editing, Writing &#x2013; original draft.</p>
</sec>
<sec id="s16" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s17" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="s18" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><mixed-citation publication-type="book"><collab>Statistisches Bundesamt (Destatis)</collab>. <source>Zahl der Todesf&#x00E4;lle Aufgrund von Suizid 2024 Leicht Gestiegen</source>. <publisher-loc>Wiesbaden</publisher-loc>: <publisher-name>Destatis</publisher-name> (<year>2024</year>). <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Gesundheit/Todesursachen/suizid.html">https://www.destatis.de/DE/Themen/Gesellschaft-Umwelt/Gesundheit/Todesursachen/suizid.html</ext-link> <comment>(Accessed October 23, 2025)</comment>.</mixed-citation></ref>
<ref id="B2"><label>2.</label><mixed-citation publication-type="book"><collab>Statistisches Bundesamt (Destatis)</collab>. <source>Pr&#x00E4;ventionstag Gegen Suizid: Jeder 100. Todesfall in Deutschland ist ein Suizid</source>. <publisher-loc>Wiesbaden</publisher-loc>: <publisher-name>Destatis</publisher-name> (<year>2024</year>). <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://www.destatis.de/DE/Presse/Pressemitteilungen/2024/09/PD24_N046_23211.html">https://www.destatis.de/DE/Presse/Pressemitteilungen/2024/09/PD24_N046_23211.html</ext-link> <comment>(Accessed October 23, 2025)</comment>.</mixed-citation></ref>
<ref id="B3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arsenault-Lapierre</surname> <given-names>G</given-names></name> <name><surname>Kim</surname> <given-names>C</given-names></name> <name><surname>Turecki</surname> <given-names>G</given-names></name></person-group>. <article-title>Psychiatric diagnoses in 3275 suicides: a meta-analysis</article-title>. <source>BMC Psychiatry</source>. (<year>2004</year>) <volume>4</volume>:<fpage>37</fpage>. <pub-id pub-id-type="doi">10.1186/1471-244X-4-37</pub-id><pub-id pub-id-type="pmid">15527502</pub-id></mixed-citation></ref>
<ref id="B4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bertolote</surname> <given-names>JM</given-names></name> <name><surname>Fleischmann</surname> <given-names>A</given-names></name></person-group>. <article-title>Suicide and psychiatric diagnosis: a worldwide perspective</article-title>. <source>Off J World Psychiatr Assoc</source>. (<year>2002</year>) <volume>1</volume>(<issue>3</issue>):<fpage>181</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="B5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Leavey</surname> <given-names>G</given-names></name> <name><surname>Rosato</surname> <given-names>M</given-names></name> <name><surname>Galway</surname> <given-names>K</given-names></name> <name><surname>Hughes</surname> <given-names>L</given-names></name> <name><surname>Mallon</surname> <given-names>S</given-names></name> <name><surname>Rondon</surname> <given-names>J</given-names></name></person-group>. <article-title>Patterns and predictors of help-seeking contacts with health services and general practitioner detection of suicidality prior to suicide: a cohort analysis of suicides occurring over a two-year period</article-title>. <source>BMC Psychiatry</source>. (<year>2016</year>) <volume>16</volume>(<issue>1</issue>):<fpage>120</fpage>. <pub-id pub-id-type="doi">10.1186/s12888-016-0824-7</pub-id><pub-id pub-id-type="pmid">27138875</pub-id></mixed-citation></ref>
<ref id="B6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mann</surname> <given-names>JJ</given-names></name> <name><surname>Apter</surname> <given-names>A</given-names></name> <name><surname>Bertolote</surname> <given-names>J</given-names></name> <name><surname>Beautrais</surname> <given-names>A</given-names></name> <name><surname>Currier</surname> <given-names>D</given-names></name> <name><surname>Haas</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Suicide prevention strategies: a systematic review</article-title>. <source>JAMA</source>. (<year>2005</year>) <volume>294</volume>(<issue>16</issue>):<fpage>2064</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1001/jama.294.16.2064</pub-id><pub-id pub-id-type="pmid">16249421</pub-id></mixed-citation></ref>
<ref id="B7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Corcoran</surname> <given-names>CM</given-names></name> <name><surname>Cecchi</surname> <given-names>GA</given-names></name></person-group>. <article-title>Using language processing and speech analysis for the identification of psychosis and other disorders</article-title>. <source>Biol Psychiatry Cogn Neurosci Neuroimaging</source>. (<year>2020</year>) <volume>5</volume>(<issue>8</issue>):<fpage>770</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.bpsc.2020.06.004</pub-id><pub-id pub-id-type="pmid">32771179</pub-id></mixed-citation></ref>
<ref id="B8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cummins</surname> <given-names>N</given-names></name> <name><surname>Scherer</surname> <given-names>S</given-names></name> <name><surname>Krajewski</surname> <given-names>J</given-names></name> <name><surname>Schnieder</surname> <given-names>S</given-names></name> <name><surname>Epps</surname> <given-names>J</given-names></name> <name><surname>Quatieri</surname> <given-names>TF</given-names></name></person-group>. <article-title>A review of depression and suicide risk assessment using speech analysis</article-title>. <source>Speech Commun</source>. (<year>2015</year>) <volume>71</volume>:<fpage>10</fpage>&#x2013;<lpage>49</lpage>. <pub-id pub-id-type="doi">10.1016/j.specom.2015.03.004</pub-id></mixed-citation></ref>
<ref id="B9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Stade</surname> <given-names>EC</given-names></name> <name><surname>Ungar</surname> <given-names>L</given-names></name> <name><surname>Eichstaedt</surname> <given-names>JC</given-names></name> <name><surname>Sherman</surname> <given-names>G</given-names></name> <name><surname>Ruscio</surname> <given-names>AM</given-names></name></person-group>. <article-title>Depression and anxiety have distinct and overlapping language patterns: results from a clinical interview</article-title>. <source>J Psychopathol Clin Sci</source>. (<year>2023</year>) <volume>132</volume>(<issue>8</issue>):<fpage>972</fpage>&#x2013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1037/abn0000850</pub-id><pub-id pub-id-type="pmid">37471025</pub-id></mixed-citation></ref>
<ref id="B10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>TMH</given-names></name> <name><surname>Chen</surname> <given-names>J</given-names></name> <name><surname>Law</surname> <given-names>FOC</given-names></name> <name><surname>Li</surname> <given-names>CT</given-names></name> <name><surname>Chan</surname> <given-names>NY</given-names></name> <name><surname>Chan</surname> <given-names>JWY</given-names></name><etal/></person-group> <article-title>Detection of suicidal ideation in clinical interviews for depression using natural language processing and machine learning: cross-sectional study</article-title>. <source>JMIR Med Inform</source>. (<year>2023</year>) <volume>11</volume>(<issue>1</issue>):<fpage>e50221</fpage>. <pub-id pub-id-type="doi">10.2196/50221</pub-id><pub-id pub-id-type="pmid">38054498</pub-id></mixed-citation></ref>
<ref id="B11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Saini</surname> <given-names>P</given-names></name> <name><surname>While</surname> <given-names>D</given-names></name> <name><surname>Chantler</surname> <given-names>K</given-names></name> <name><surname>Windfuhr</surname> <given-names>K</given-names></name> <name><surname>Kapur</surname> <given-names>N</given-names></name></person-group>. <article-title>Assessment and management of suicide risk in primary care</article-title>. <source>Crisis</source>. (<year>2014</year>) <volume>35</volume>(<issue>6</issue>):<fpage>415</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1027/0227-5910/a000277</pub-id><pub-id pub-id-type="pmid">25234744</pub-id></mixed-citation></ref>
<ref id="B12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>DeSouza</surname> <given-names>DD</given-names></name> <name><surname>Robin</surname> <given-names>J</given-names></name> <name><surname>Gumus</surname> <given-names>M</given-names></name> <name><surname>Yeung</surname> <given-names>A</given-names></name></person-group>. <article-title>Natural language processing as an emerging tool to detect late-life depression</article-title>. <source>Front Psychiatry</source>. (<year>2021</year>) <volume>12</volume>:<fpage>719125</fpage>. <pub-id pub-id-type="doi">10.3389/fpsyt.2021.719125</pub-id><pub-id pub-id-type="pmid">34552519</pub-id></mixed-citation></ref>
<ref id="B13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>T</given-names></name> <name><surname>Schoene</surname> <given-names>AM</given-names></name> <name><surname>Ji</surname> <given-names>S</given-names></name> <name><surname>Ananiadou</surname> <given-names>S</given-names></name></person-group>. <article-title>Natural language processing applied to mental illness detection: a narrative review</article-title>. <source>Npj Digit Med</source>. (<year>2022</year>) <volume>5</volume>(<issue>1</issue>):<fpage>46</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-022-00589-7</pub-id><pub-id pub-id-type="pmid">35396451</pub-id></mixed-citation></ref>
<ref id="B14"><label>14.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alghamdi</surname> <given-names>NS</given-names></name> <name><surname>Hosni Mahmoud</surname> <given-names>HA</given-names></name> <name><surname>Abraham</surname> <given-names>A</given-names></name> <name><surname>Alanazi</surname> <given-names>SA</given-names></name> <name><surname>Garc&#x00ED;a-Hern&#x00E1;ndez</surname> <given-names>L</given-names></name></person-group>. <article-title>Predicting depression symptoms in an Arabic psychological forum</article-title>. <source>IEEE Access</source>. (<year>2020</year>) <volume>8</volume>:<fpage>57317</fpage>&#x2013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.2981834</pub-id></mixed-citation></ref>
<ref id="B15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Castillo-S&#x00E1;nchez</surname> <given-names>G</given-names></name> <name><surname>Marques</surname> <given-names>G</given-names></name> <name><surname>Dorronzoro</surname> <given-names>E</given-names></name> <name><surname>Rivera-Romero</surname> <given-names>O</given-names></name> <name><surname>Franco-Mart&#x00ED;n</surname> <given-names>M</given-names></name> <name><surname>De la Torre-D&#x00ED;ez</surname> <given-names>I</given-names></name></person-group>. <article-title>Suicide risk assessment using machine learning and social networks: a scoping review</article-title>. <source>J Med Syst</source>. (<year>2020</year>) <volume>44</volume>(<issue>12</issue>):<fpage>1</fpage>&#x2013;<lpage>15</lpage>. <pub-id pub-id-type="doi">10.1007/s10916-020-01669-5</pub-id></mixed-citation></ref>
<ref id="B16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Coppersmith</surname> <given-names>G</given-names></name> <name><surname>Leary</surname> <given-names>R</given-names></name> <name><surname>Crutchley</surname> <given-names>P</given-names></name> <name><surname>Fine</surname> <given-names>A</given-names></name></person-group>. <article-title>Natural language processing of social media as screening for suicide risk</article-title>. <source>Biomed Inform Insights</source>. (<year>2018</year>) <volume>10</volume>:<fpage>117822261879286</fpage>. <pub-id pub-id-type="doi">10.1177/1178222618792860</pub-id></mixed-citation></ref>
<ref id="B17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cook</surname> <given-names>BL</given-names></name> <name><surname>Progovac</surname> <given-names>AM</given-names></name> <name><surname>Chen</surname> <given-names>P</given-names></name> <name><surname>Mullin</surname> <given-names>B</given-names></name> <name><surname>Hou</surname> <given-names>S</given-names></name> <name><surname>Baca-Garcia</surname> <given-names>E</given-names></name></person-group>. <article-title>Novel use of natural language processing (NLP) to predict suicidal ideation and psychiatric symptoms in a text-based mental health intervention in Madrid</article-title>. <source>Comput Math Methods Med</source>. (<year>2016</year>) <volume>2016</volume>:<fpage>8708434</fpage>. <pub-id pub-id-type="doi">10.1155/2016/8708434</pub-id><pub-id pub-id-type="pmid">27752278</pub-id></mixed-citation></ref>
<ref id="B18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pestian</surname> <given-names>J</given-names></name> <name><surname>Santel</surname> <given-names>D</given-names></name> <name><surname>Sorter</surname> <given-names>M</given-names></name> <name><surname>Bayram</surname> <given-names>U</given-names></name> <name><surname>Connolly</surname> <given-names>B</given-names></name> <name><surname>Glauser</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>A machine learning approach to identifying changes in suicidal language</article-title>. <source>Suicide Life Threat Behav</source>. (<year>2020</year>) <volume>50</volume>(<issue>5</issue>):<fpage>939</fpage>&#x2013;<lpage>47</lpage>. <pub-id pub-id-type="doi">10.1111/sltb.12642</pub-id><pub-id pub-id-type="pmid">32484597</pub-id></mixed-citation></ref>
<ref id="B19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Carson</surname> <given-names>NJ</given-names></name> <name><surname>Mullin</surname> <given-names>B</given-names></name> <name><surname>Sanchez</surname> <given-names>MJ</given-names></name> <name><surname>Lu</surname> <given-names>F</given-names></name> <name><surname>Yang</surname> <given-names>K</given-names></name> <name><surname>Menezes</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Identification of suicidal behavior among psychiatrically hospitalized adolescents using natural language processing and machine learning of electronic health records</article-title>. <source>PLoS One</source>. (<year>2019</year>) <volume>14</volume>(<issue>2</issue>):<fpage>e0211116</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0211116</pub-id><pub-id pub-id-type="pmid">30779800</pub-id></mixed-citation></ref>
<ref id="B20"><label>20.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fernandes</surname> <given-names>AC</given-names></name> <name><surname>Dutta</surname> <given-names>R</given-names></name> <name><surname>Velupillai</surname> <given-names>S</given-names></name> <name><surname>Sanyal</surname> <given-names>J</given-names></name> <name><surname>Stewart</surname> <given-names>R</given-names></name> <name><surname>Chandran</surname> <given-names>D</given-names></name></person-group>. <article-title>Identifying suicide ideation and suicidal attempts in a psychiatric clinical research database using natural language processing</article-title>. <source>Sci Rep</source>. (<year>2018</year>) <volume>8</volume>(<issue>1</issue>):<fpage>7426</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-018-25773-2</pub-id><pub-id pub-id-type="pmid">29743531</pub-id></mixed-citation></ref>
<ref id="B21"><label>21.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pestian</surname> <given-names>JP</given-names></name> <name><surname>Grupp-Phelan</surname> <given-names>J</given-names></name> <name><surname>Bretonnel Cohen</surname> <given-names>K</given-names></name> <name><surname>Meyers</surname> <given-names>G</given-names></name> <name><surname>Richey</surname> <given-names>LA</given-names></name> <name><surname>Matykiewicz</surname> <given-names>P</given-names></name><etal/></person-group> <article-title>A controlled trial using natural language processing to examine the language of suicidal adolescents in the emergency department</article-title>. <source>Suicide Life Threat Behav</source>. (<year>2016</year>) <volume>46</volume>(<issue>2</issue>):<fpage>154</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1111/sltb.12180</pub-id><pub-id pub-id-type="pmid">26252868</pub-id></mixed-citation></ref>
<ref id="B22"><label>22.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A</given-names></name> <name><surname>Brain</surname> <given-names>G</given-names></name> <name><surname>Shazeer</surname> <given-names>N</given-names></name> <name><surname>Parmar</surname> <given-names>N</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J</given-names></name> <name><surname>Jones</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Attention is all you need</article-title>. In: <person-group person-group-type="editor"><name><surname>Guyon</surname> <given-names>I</given-names></name> <name><surname>Von Luxburg</surname> <given-names>U</given-names></name> <name><surname>Bengio</surname> <given-names>S</given-names></name> <name><surname>Wallach</surname> <given-names>H</given-names></name> <name><surname>Fergus</surname> <given-names>R</given-names></name> <name><surname>Vishwanathan</surname> <given-names>S</given-names></name><etal/></person-group>, editors. <source>31st Conference on Neural Information Processing Systems (NIPS). Advances in Neural Information Processing Systems. Vol. 30</source>. <publisher-loc>Long Beach, CA</publisher-loc>: <publisher-name>Curran Associates, Inc.</publisher-name> (<year>2017</year>). p. <fpage>5998</fpage>&#x2013;<lpage>6008</lpage>.</mixed-citation></ref>
<ref id="B23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kalyan</surname> <given-names>KS</given-names></name></person-group>. <article-title>A survey of GPT-3 family large language models including ChatGPT and GPT-4</article-title>. <source>Nat Lang Process J</source>. (<year>2024</year>) <volume>6</volume>:<fpage>100048</fpage>. <pub-id pub-id-type="doi">10.1016/j.nlp.2023.100048</pub-id></mixed-citation></ref>
<ref id="B24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hua</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>F</given-names></name> <name><surname>Yang</surname> <given-names>K</given-names></name> <name><surname>Li</surname> <given-names>Z</given-names></name> <name><surname>Na</surname> <given-names>H</given-names></name> <name><surname>Sheu</surname> <given-names>YH</given-names></name><etal/></person-group> <article-title>Large language models in mental health care: a scoping review</article-title>. <source>Curr Treat Options Psychiatry</source>. (<year>2025</year>) <volume>12</volume>(<issue>1</issue>). <pub-id pub-id-type="doi">10.1007/s40501-025-00363-y</pub-id><pub-id pub-id-type="pmid">40524733</pub-id></mixed-citation></ref>
<ref id="B25"><label>25.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Z</given-names></name> <name><surname>Deng</surname> <given-names>J</given-names></name> <name><surname>Zhou</surname> <given-names>J</given-names></name> <name><surname>Wu</surname> <given-names>J</given-names></name> <name><surname>Qian</surname> <given-names>T</given-names></name> <name><surname>Huang</surname> <given-names>M</given-names></name></person-group>. <article-title>Depression detection in clinical interviews with LLM-empowered structural element graph</article-title>. <conf-name>Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)</conf-name>; <publisher-name>Association for Computational Linguistics</publisher-name> (<year>2024</year>). p. <fpage>8181</fpage>&#x2013;<lpage>94</lpage>.</mixed-citation></ref>
<ref id="B26"><label>26.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Galatzer-Levy</surname> <given-names>IR</given-names></name> <name><surname>Mcduff</surname> <given-names>D</given-names></name> <name><surname>Natarajan</surname> <given-names>V</given-names></name> <name><surname>Karthikesalingam</surname> <given-names>A</given-names></name> <name><surname>Malgaroli</surname> <given-names>M</given-names></name></person-group>. <comment>The Capability of Large Language Models to Measure Psychiatric Functioning. arXiv (2023)</comment>.</mixed-citation></ref>
<ref id="B27"><label>27.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>X</given-names></name> <name><surname>Yao</surname> <given-names>B</given-names></name> <name><surname>Dong</surname> <given-names>Y</given-names></name> <name><surname>Gabriel</surname> <given-names>S</given-names></name> <name><surname>Yu</surname> <given-names>H</given-names></name> <name><surname>Hendler</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Mental-LLM: leveraging large language models for mental health prediction via online text data</article-title>. <conf-name>Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies</conf-name>; <publisher-name>Association for Computing Machinery</publisher-name> (<year>2024</year>).</mixed-citation></ref>
<ref id="B28"><label>28.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Levkovich</surname> <given-names>I</given-names></name></person-group>. <article-title>Evaluating diagnostic accuracy and treatment efficacy in mental health: a comparative analysis of large language model tools and mental health professionals</article-title>. <source>Eur J Investig Health Psychol Educ</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>9</fpage>. <pub-id pub-id-type="doi">10.3390/ejihpe15010009</pub-id><pub-id pub-id-type="pmid">39852192</pub-id></mixed-citation></ref>
<ref id="B29"><label>29.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Louie</surname> <given-names>R</given-names></name> <name><surname>Nandi</surname> <given-names>A</given-names></name> <name><surname>Fang</surname> <given-names>W</given-names></name> <name><surname>Chang</surname> <given-names>C</given-names></name> <name><surname>Brunskill</surname> <given-names>E</given-names></name> <name><surname>Yang</surname> <given-names>D</given-names></name></person-group>. <comment>Roleplay-doh: Enabling Domain-Experts to Create LLM-simulated Patients via Eliciting and Adhering to Principles. arXiv (July 14, 2024)</comment>.</mixed-citation></ref>
<ref id="B30"><label>30.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Frick</surname> <given-names>RW</given-names></name></person-group>. <article-title>Communicating emotion: the role of prosodic features</article-title>. <source>Psychol Bull</source>. (<year>1985</year>) <volume>97</volume>:<fpage>412</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.1037/0033-2909.97.3.412</pub-id></mixed-citation></ref>
<ref id="B31"><label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>C</given-names></name> <name><surname>Jiang</surname> <given-names>M</given-names></name> <name><surname>Gao</surname> <given-names>L</given-names></name> <name><surname>Xin</surname> <given-names>Y</given-names></name> <name><surname>Dong</surname> <given-names>Y</given-names></name></person-group>. <article-title>A novel study for depression detecting using audio signals based on graph neural network</article-title>. <source>Biomed Signal Process Control</source>. (<year>2024</year>) <volume>88</volume>:<fpage>105675</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2023.105675</pub-id></mixed-citation></ref>
<ref id="B32"><label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dumpala</surname> <given-names>SH</given-names></name> <name><surname>Dikaios</surname> <given-names>K</given-names></name> <name><surname>Rodriguez</surname> <given-names>S</given-names></name> <name><surname>Langley</surname> <given-names>R</given-names></name> <name><surname>Rempel</surname> <given-names>S</given-names></name> <name><surname>Uher</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Manifestation of depression in speech overlaps with characteristics used to represent and recognize speaker identity</article-title>. <source>Sci Rep</source>. (<year>2023</year>) <volume>13</volume>(<issue>1</issue>):<fpage>11155</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-023-35184-7</pub-id><pub-id pub-id-type="pmid">37429935</pub-id></mixed-citation></ref>
<ref id="B33"><label>33.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>Y</given-names></name> <name><surname>Zhu</surname> <given-names>C</given-names></name> <name><surname>Hao</surname> <given-names>S</given-names></name> <name><surname>Hong</surname> <given-names>R</given-names></name></person-group>. <comment>A Topic-Attentive Transformer-Based Model for Multimodal Depression Detection. arXiv (June 27, 2022)</comment>.</mixed-citation></ref>
<ref id="B34"><label>34.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Belouali</surname> <given-names>A</given-names></name> <name><surname>Gupta</surname> <given-names>S</given-names></name> <name><surname>Sourirajan</surname> <given-names>V</given-names></name> <name><surname>Yu</surname> <given-names>J</given-names></name> <name><surname>Allen</surname> <given-names>N</given-names></name> <name><surname>Alaoui</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Acoustic and language analysis of speech for suicidal ideation among US veterans</article-title>. <source>BioData Min</source>. (<year>2021</year>) <volume>14</volume>(<issue>1</issue>):<fpage>11</fpage>. <pub-id pub-id-type="doi">10.1186/s13040-021-00245-y</pub-id><pub-id pub-id-type="pmid">33531048</pub-id></mixed-citation></ref>
<ref id="B35"><label>35.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mao</surname> <given-names>K</given-names></name> <name><surname>Zhang</surname> <given-names>W</given-names></name> <name><surname>Wang</surname> <given-names>DB</given-names></name> <name><surname>Li</surname> <given-names>A</given-names></name> <name><surname>Jiao</surname> <given-names>R</given-names></name> <name><surname>Zhu</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Prediction of depression severity based on the prosodic and semantic features with bidirectional LSTM and time distributed CNN</article-title>. <source>IEEE Trans Affect Comput</source>. (<year>2023</year>) <volume>14</volume>(<issue>3</issue>):<fpage>2251</fpage>&#x2013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.1109/TAFFC.2022.3154332</pub-id></mixed-citation></ref>
<ref id="B36"><label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Malgaroli</surname> <given-names>M</given-names></name> <name><surname>Hull</surname> <given-names>TD</given-names></name> <name><surname>Zech</surname> <given-names>JM</given-names></name> <name><surname>Althoff</surname> <given-names>T</given-names></name></person-group>. <article-title>Natural language processing for mental health interventions: a systematic review and research framework</article-title>. <source>Transl Psychiatry</source>. (<year>2023</year>) <volume>13</volume>(<issue>1</issue>):<fpage>309</fpage>. <pub-id pub-id-type="doi">10.1038/s41398-023-02592-2</pub-id><pub-id pub-id-type="pmid">37798296</pub-id></mixed-citation></ref>
<ref id="B37"><label>37.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Joshi</surname> <given-names>P</given-names></name> <name><surname>Santy</surname> <given-names>S</given-names></name> <name><surname>Budhiraja</surname> <given-names>A</given-names></name> <name><surname>Bali</surname> <given-names>K</given-names></name> <name><surname>Choudhury</surname> <given-names>M</given-names></name></person-group>. <article-title>The state and fate of linguistic diversity and inclusion in the NLP world</article-title>. <conf-name>Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</conf-name>; <publisher-name>Association for Computational Linguistics</publisher-name> (<year>2020</year>). p. <fpage>6282</fpage>&#x2013;<lpage>93</lpage></mixed-citation></ref>
<ref id="B38"><label>38.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Eberhardt</surname> <given-names>ST</given-names></name> <name><surname>Schaffrath</surname> <given-names>J</given-names></name> <name><surname>Moggia</surname> <given-names>D</given-names></name> <name><surname>Schwartz</surname> <given-names>B</given-names></name> <name><surname>Jaehde</surname> <given-names>M</given-names></name> <name><surname>Rubel</surname> <given-names>JA</given-names></name><etal/></person-group> <article-title>Decoding emotions: exploring the validity of sentiment analysis in psychotherapy</article-title>. <source>Psychother Res</source>. (<year>2025</year>) <volume>35</volume>(<issue>2</issue>):<fpage>174</fpage>&#x2013;<lpage>89</lpage>. <pub-id pub-id-type="doi">10.1080/10503307.2024.2322522</pub-id><pub-id pub-id-type="pmid">38415369</pub-id></mixed-citation></ref>
<ref id="B39"><label>39.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Menne</surname> <given-names>F</given-names></name> <name><surname>D&#x00F6;rr</surname> <given-names>F</given-names></name> <name><surname>Schr&#x00E4;der</surname> <given-names>J</given-names></name> <name><surname>Tr&#x00F6;ger</surname> <given-names>J</given-names></name> <name><surname>Habel</surname> <given-names>U</given-names></name> <name><surname>K&#x00F6;nig</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>The voice of depression: speech features as biomarkers for major depressive disorder</article-title>. <source>BMC Psychiatry</source>. (<year>2024</year>) <volume>24</volume>(<issue>1</issue>):<fpage>794</fpage>. <pub-id pub-id-type="doi">10.1186/s12888-024-06253-6</pub-id><pub-id pub-id-type="pmid">39533239</pub-id></mixed-citation></ref>
<ref id="B40"><label>40.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Schmidt</surname> <given-names>T</given-names></name></person-group>. <article-title>The research and teaching corpus of spoken German-FOLK</article-title>. <conf-name>Proceedings of the Ninth International Conference on Language Resources and Evaluation</conf-name>; (<year>2014</year>). p. <fpage>383</fpage>&#x2013;<lpage>7</lpage>.</mixed-citation></ref>
<ref id="B41"><label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Latif</surname> <given-names>S</given-names></name> <name><surname>Qadir</surname> <given-names>J</given-names></name> <name><surname>Qayyum</surname> <given-names>A</given-names></name> <name><surname>Usama</surname> <given-names>M</given-names></name> <name><surname>Younis</surname> <given-names>S</given-names></name></person-group>. <article-title>Speech technology for healthcare: opportunities, challenges, and state of the art</article-title>. <source>IEEE Rev Biomed Eng</source>. (<year>2020</year>) <volume>14</volume>:<fpage>342</fpage>&#x2013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.1109/RBME.2020.3006860</pub-id></mixed-citation></ref>
<ref id="B42"><label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Parola</surname> <given-names>A</given-names></name> <name><surname>Lin</surname> <given-names>JM</given-names></name> <name><surname>Simonsen</surname> <given-names>A</given-names></name> <name><surname>Bliksted</surname> <given-names>V</given-names></name> <name><surname>Zhou</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Speech disturbances in schizophrenia: assessing cross-linguistic generalizability of NLP automated measures of coherence</article-title>. <source>Schizophr Res</source>. (<year>2023</year>) <volume>259</volume>:<fpage>59</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1016/j.schres.2022.07.002</pub-id><pub-id pub-id-type="pmid">35927097</pub-id></mixed-citation></ref>
<ref id="B43"><label>43.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Faul</surname> <given-names>A</given-names></name></person-group>. <article-title>Lang. Statistical power analyses using G&#x002A;power 3.1: tests for correlation and regression analyses</article-title>. <source>Behav Res Methods</source>. (<year>2009</year>) <volume>41</volume>(<issue>4</issue>):<fpage>1149</fpage>&#x2013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.3758/BRM.41.4.1149</pub-id><pub-id pub-id-type="pmid">19897823</pub-id></mixed-citation></ref>
<ref id="B44"><label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Linthicum</surname> <given-names>KP</given-names></name> <name><surname>Schafer</surname> <given-names>KM</given-names></name> <name><surname>Ribeiro</surname> <given-names>JD</given-names></name></person-group>. <article-title>Machine learning in suicide science: applications and ethics</article-title>. <source>Behav Sci Law</source>. (<year>2019</year>) <volume>37</volume>(<issue>3</issue>):<fpage>214</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1002/bsl.2392</pub-id><pub-id pub-id-type="pmid">30609102</pub-id></mixed-citation></ref>
<ref id="B45"><label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pestian</surname> <given-names>JP</given-names></name> <name><surname>Sorter</surname> <given-names>M</given-names></name> <name><surname>Connolly</surname> <given-names>B</given-names></name> <name><surname>Bretonnel Cohen</surname> <given-names>K</given-names></name> <name><surname>McCullumsmith</surname> <given-names>C</given-names></name> <name><surname>Gee</surname> <given-names>JT</given-names></name><etal/></person-group> <article-title>A machine learning approach to identifying the thought markers of suicidal subjects: a prospective multicenter trial</article-title>. <source>Suicide Life Threat Behav</source>. (<year>2017</year>) <volume>47</volume>(<issue>1</issue>):<fpage>112</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1111/sltb.12312</pub-id><pub-id pub-id-type="pmid">27813129</pub-id></mixed-citation></ref>
<ref id="B46"><label>46.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bernert</surname> <given-names>RA</given-names></name> <name><surname>Hilberg</surname> <given-names>AM</given-names></name> <name><surname>Melia</surname> <given-names>R</given-names></name> <name><surname>Kim</surname> <given-names>JP</given-names></name> <name><surname>Shah</surname> <given-names>NH</given-names></name> <name><surname>Abnousi</surname> <given-names>F</given-names></name></person-group>. <article-title>Artificial intelligence and suicide prevention: a systematic review of machine learning investigations</article-title>. <source>Int J Environ Res Public Health</source>. (<year>2020</year>) <volume>17</volume>(<issue>16</issue>):<fpage>1</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.3390/ijerph17165929</pub-id></mixed-citation></ref>
<ref id="B47"><label>47.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Barak-Corren</surname> <given-names>Y</given-names></name> <name><surname>Castro</surname> <given-names>VM</given-names></name> <name><surname>Javitt</surname> <given-names>S</given-names></name> <name><surname>Hoffnagle</surname> <given-names>AG</given-names></name> <name><surname>Dai</surname> <given-names>Y</given-names></name> <name><surname>Perlis</surname> <given-names>RH</given-names></name><etal/></person-group> <article-title>Predicting suicidal behavior from longitudinal electronic health records</article-title>. <source>Am J Psychiatry</source>. (<year>2017</year>) <volume>174</volume>(<issue>2</issue>):<fpage>154</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1176/appi.ajp.2016.16010077</pub-id><pub-id pub-id-type="pmid">27609239</pub-id></mixed-citation></ref>
<ref id="B48"><label>48.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Guy</surname> <given-names>W</given-names></name></person-group>. <source>ECDEU Assessment Manual for Psychopharmacology</source>. <publisher-loc>Rockville, MD</publisher-loc>: <publisher-name>US Department of Health, Education, and Welfare</publisher-name> (<year>1976</year>). p. <fpage>217</fpage>&#x2013;<lpage>22</lpage>.</mixed-citation></ref>
<ref id="B49"><label>49.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Busner</surname> <given-names>J</given-names></name> <name><surname>Targum</surname> <given-names>SD</given-names></name></person-group>. <article-title>The clinical global impressions scale: applying a research tool in clinical practice</article-title>. <source>Psychiatry Edgmont</source>. (<year>2007</year>) <volume>4</volume>:<fpage>28</fpage>&#x2013;<lpage>37</lpage>.<pub-id pub-id-type="pmid">20526405</pub-id></mixed-citation></ref>
<ref id="B50"><label>50.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hale</surname> <given-names>AS</given-names></name> <name><surname>Azorin</surname> <given-names>JM</given-names></name> <name><surname>Lemming</surname> <given-names>OM</given-names></name> <name><surname>M&#x00E6;hlum</surname> <given-names>E</given-names></name></person-group>. <article-title>Sertindole in the long-term treatment of schizophrenia</article-title>. <source>Int Clin Psychopharmacol</source>. (<year>2012</year>) <volume>27</volume>(<issue>4</issue>):<fpage>231</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1097/YIC.0b013e328354dcda</pub-id><pub-id pub-id-type="pmid">22609816</pub-id></mixed-citation></ref>
<ref id="B51"><label>51.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Morrens</surname> <given-names>J</given-names></name> <name><surname>Mathews</surname> <given-names>M</given-names></name> <name><surname>Popova</surname> <given-names>V</given-names></name> <name><surname>Borentain</surname> <given-names>S</given-names></name> <name><surname>Moro</surname> <given-names>BGM</given-names></name> <name><surname>Jamieson</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Use of clinical global impressions-severity (CGI-S) to assess response to antidepressant treatment in patients with treatment-resistant depression</article-title>. <source>Neuropsychiatr Dis Treat</source>. (<year>2022</year>) <volume>18</volume>:<fpage>1127</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.2147/NDT.S358367</pub-id><pub-id pub-id-type="pmid">35707064</pub-id></mixed-citation></ref>
<ref id="B52"><label>52.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Meltzer</surname> <given-names>HY</given-names></name> <name><surname>Alphs</surname> <given-names>L</given-names></name> <name><surname>Green</surname> <given-names>AI</given-names></name> <name><surname>Altamura</surname> <given-names>AC</given-names></name> <name><surname>Anand</surname> <given-names>R</given-names></name> <name><surname>Bertoldi</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Clozapine treatment for suicidality in schizophrenia: international suicide prevention trial (InterSePT)</article-title>. <source>JAMA Psychiatry</source>. (<year>2003</year>) <volume>60</volume>(<issue>1</issue>):<fpage>82</fpage>&#x2013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.1001/archpsyc.60.1.82</pub-id></mixed-citation></ref>
<ref id="B53"><label>53.</label><mixed-citation publication-type="book"><collab>American Psychiatric Association</collab>. <source>Diagnostic and Statistical Manual of Mental Disorders</source>. <publisher-loc>Washington DC</publisher-loc>: <publisher-name>American Psychiatric Association</publisher-name> (<year>1987</year>).</mixed-citation></ref>
<ref id="B54"><label>54.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jones</surname> <given-names>SH</given-names></name> <name><surname>Thornicroft</surname> <given-names>G</given-names></name> <name><surname>Coffey</surname> <given-names>M</given-names></name> <name><surname>Dunn</surname> <given-names>G</given-names></name></person-group>. <article-title>A brief mental health outcome scale. Reliability and validity of the global assessment of functioning (GAF)</article-title>. <source>Br J Psychiatry</source>. (<year>1995</year>) <volume>166</volume>:<fpage>654</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1192/bjp.166.5.654</pub-id><pub-id pub-id-type="pmid">7620753</pub-id></mixed-citation></ref>
<ref id="B55"><label>55.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shelef</surname> <given-names>K</given-names></name> <name><surname>Diamond</surname> <given-names>GM</given-names></name></person-group>. <article-title>Short form of the revised vanderbilt therapeutic alliance scale: development, reliability, and validity</article-title>. <source>Psychother Res</source>. (<year>2008</year>) <volume>18</volume>(<issue>4</issue>):<fpage>433</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1080/10503300701810801</pub-id><pub-id pub-id-type="pmid">18815995</pub-id></mixed-citation></ref>
<ref id="B56"><label>56.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Eyben</surname> <given-names>F</given-names></name> <name><surname>W&#x00F6;llmer</surname> <given-names>M</given-names></name> <name><surname>Schuller</surname> <given-names>B</given-names></name></person-group>. <article-title>Opensmile&#x2014;the Munich versatile and fast open-source audio feature extractor</article-title>. <conf-name>proceedings of the ACM Multimedia (MM)</conf-name>; <conf-loc>Florence, Italy</conf-loc>: <publisher-name>Association for Computing Machinery</publisher-name> (<year>2010</year>). p. <fpage>1459</fpage>&#x2013;<lpage>62</lpage></mixed-citation></ref>
<ref id="B57"><label>57.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Bain</surname> <given-names>M</given-names></name> <name><surname>Huh</surname> <given-names>J</given-names></name> <name><surname>Han</surname> <given-names>T</given-names></name> <name><surname>Zisserman</surname> <given-names>A</given-names></name></person-group>. <article-title>Whisperx: time-accurate speech transcription of long-form audio</article-title>. <conf-name>Proc INTERSPEECH 2023</conf-name> (<year>2023</year>). p. <fpage>4489</fpage>&#x2013;<lpage>93</lpage></mixed-citation></ref>
<ref id="B58"><label>58.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Agarwal</surname> <given-names>S</given-names></name> <name><surname>Ahmad</surname> <given-names>L</given-names></name> <name><surname>Ai</surname> <given-names>J</given-names></name> <name><surname>Altman</surname> <given-names>S</given-names></name> <name><surname>Applebaum</surname> <given-names>A</given-names></name> <name><surname>Arbus</surname> <given-names>E</given-names></name><etal/></person-group> <comment>gpt-oss-120b &#x0026; gpt-oss-20b Model Card. arXiv (August 8, 2025)</comment>.</mixed-citation></ref>
<ref id="B59"><label>59.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Renze</surname> <given-names>M</given-names></name></person-group>. <article-title>The effect of sampling temperature on problem solving in large language models</article-title>. In: <person-group person-group-type="editor"><name><surname>Al-Onaizan</surname> <given-names>Y</given-names></name> <name><surname>Bansal</surname> <given-names>M</given-names></name> <name><surname>Chen</surname> <given-names>YN</given-names></name></person-group>, editors. <source>Findings of the Association for Computational Linguistics: EMNLP 2024</source>. <publisher-loc>Miami, Florida, USA</publisher-loc>: <publisher-name>Association for Computational Linguistics</publisher-name> (<year>2024</year>). p. <fpage>7346</fpage>&#x2013;<lpage>56</lpage>. <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://aclanthology.org/2024.findings-emnlp.432/">https://aclanthology.org/2024.findings-emnlp.432/</ext-link> <comment>(Accessed January 23, 2026)</comment>.</mixed-citation></ref>
<ref id="B60"><label>60.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Eyben</surname> <given-names>F</given-names></name> <name><surname>Scherer</surname> <given-names>KR</given-names></name> <name><surname>Schuller</surname> <given-names>BW</given-names></name> <name><surname>Sundberg</surname> <given-names>J</given-names></name> <name><surname>Andre</surname> <given-names>E</given-names></name> <name><surname>Busso</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing</article-title>. <source>IEEE Trans Affect Comput</source>. (<year>2016</year>) <volume>7</volume>(<issue>2</issue>):<fpage>190</fpage>&#x2013;<lpage>202</lpage>. <pub-id pub-id-type="doi">10.1109/TAFFC.2015.2457417</pub-id></mixed-citation></ref>
<ref id="B61"><label>61.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Koops</surname> <given-names>S</given-names></name> <name><surname>Brederoo</surname> <given-names>SG</given-names></name> <name><surname>de Boer</surname> <given-names>JN</given-names></name> <name><surname>Nadema</surname> <given-names>FG</given-names></name> <name><surname>Voppel</surname> <given-names>AE</given-names></name> <name><surname>Sommer</surname> <given-names>IE</given-names></name></person-group>. <article-title>Speech as a biomarker for depression</article-title>. <source>CNS Neurol Disord&#x2014;Drug Targets</source>. (<year>2023</year>) <volume>22</volume>(<issue>2</issue>):<fpage>152</fpage>&#x2013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.2174/1871527320666211213125847</pub-id></mixed-citation></ref>
<ref id="B62"><label>62.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Flemotomos</surname> <given-names>N</given-names></name> <name><surname>Martinez</surname> <given-names>VR</given-names></name> <name><surname>Chen</surname> <given-names>Z</given-names></name> <name><surname>Singla</surname> <given-names>K</given-names></name> <name><surname>Ardulov</surname> <given-names>V</given-names></name> <name><surname>Peri</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Automated evaluation of psychotherapy skills using speech and language technologies</article-title>. <source>Behav Res Methods</source>. (<year>2022</year>) <volume>54</volume>(<issue>2</issue>):<fpage>690</fpage>&#x2013;<lpage>711</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-021-01623-4</pub-id><pub-id pub-id-type="pmid">34346043</pub-id></mixed-citation></ref>
<ref id="B63"><label>63.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zolnoori</surname> <given-names>M</given-names></name> <name><surname>Vergez</surname> <given-names>S</given-names></name> <name><surname>Xu</surname> <given-names>Z</given-names></name> <name><surname>Esmaeili</surname> <given-names>E</given-names></name> <name><surname>Zolnour</surname> <given-names>A</given-names></name> <name><surname>Anne Briggs</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>Decoding disparities: evaluating automatic speech recognition system performance in transcribing Black and White patient verbal communication with nurses in home healthcare</article-title>. <source>JAMIA Open</source>. (<year>2024</year>) <volume>7</volume>(<issue>4</issue>):<fpage>ooae130</fpage>. <pub-id pub-id-type="doi">10.1093/jamiaopen/ooae130</pub-id><pub-id pub-id-type="pmid">39659993</pub-id></mixed-citation></ref>
<ref id="B64"><label>64.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wiest</surname> <given-names>IC</given-names></name> <name><surname>Verhees</surname> <given-names>FG</given-names></name> <name><surname>Ferber</surname> <given-names>D</given-names></name> <name><surname>Zhu</surname> <given-names>J</given-names></name> <name><surname>Bauer</surname> <given-names>M</given-names></name> <name><surname>Lewitzka</surname> <given-names>U</given-names></name><etal/></person-group> <article-title>Detection of suicidality from medical text using privacy-preserving large language models</article-title>. <source>Br J Psychiatry</source>. (<year>2024</year>) <volume>225</volume>(<issue>6</issue>):<fpage>532</fpage>&#x2013;<lpage>7</lpage>. doi: <pub-id pub-id-type="doi">10.1192/bjp.2024.134</pub-id></mixed-citation></ref>
<ref id="B65"><label>65.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kolochowski</surname> <given-names>FD</given-names></name> <name><surname>Kreckeler</surname> <given-names>N</given-names></name> <name><surname>Forkmann</surname> <given-names>T</given-names></name> <name><surname>Teismann</surname> <given-names>T</given-names></name></person-group>. <article-title>Reliability of suicide risk estimates: a vignette study</article-title>. <source>Arch Suicide Res</source>. (<year>2025</year>) <volume>29</volume>(<issue>2</issue>):<fpage>544</fpage>&#x2013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1080/13811118.2024.2382709</pub-id><pub-id pub-id-type="pmid">39045846</pub-id></mixed-citation></ref>
<ref id="B66"><label>66.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miller</surname> <given-names>PR</given-names></name></person-group>. <article-title>Inpatient diagnostic assessments: 2. Interrater reliability and outcomes of structured vs. unstructured interviews</article-title>. <source>Psychiatry Res</source>. (<year>2001</year>) <volume>105</volume>(<issue>3</issue>):<fpage>265</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1016/S0165-1781(01)00318-3</pub-id><pub-id pub-id-type="pmid">11814545</pub-id></mixed-citation></ref>
<ref id="B67"><label>67.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Mayer</surname> <given-names>T</given-names></name> <name><surname>Warikoo</surname> <given-names>N</given-names></name> <name><surname>Grimm</surname> <given-names>O</given-names></name> <name><surname>Reif</surname> <given-names>A</given-names></name> <name><surname>Gurevych</surname> <given-names>I</given-names></name></person-group>. <comment>GDPR Compliant Collection of Therapist-Patient-Dialogues. arXiv. (November 22, 2022)</comment>.</mixed-citation></ref>
<ref id="B68"><label>68.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Hopkins</surname> <given-names>D</given-names></name> <name><surname>Rickwodd</surname> <given-names>DJ</given-names></name> <name><surname>Hallford</surname> <given-names>DJ</given-names></name> <name><surname>Watsford</surname> <given-names>C</given-names></name></person-group>. <comment>Structured Data vs. Unstructured Data in Machine Learning Prediction Models for Suicidal Behaviors: A Systematic Review and Meta-analysis.</comment></mixed-citation></ref>
<ref id="B69"><label>69.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Morley</surname> <given-names>J</given-names></name> <name><surname>Machado</surname> <given-names>CCV</given-names></name> <name><surname>Burr</surname> <given-names>C</given-names></name> <name><surname>Cowls</surname> <given-names>J</given-names></name> <name><surname>Joshi</surname> <given-names>I</given-names></name> <name><surname>Taddeo</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>The ethics of AI in health care: a mapping review</article-title>. <source>Soc Sci Med</source>. (<year>2020</year>) <volume>260</volume>:<fpage>113172</fpage>. <pub-id pub-id-type="doi">10.1016/j.socscimed.2020.113172</pub-id><pub-id pub-id-type="pmid">32702587</pub-id></mixed-citation></ref></ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/113659/overview">Christos A. Frantzidis</ext-link>, University of Lincoln, United Kingdom</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/493439/overview">Panteleimon Chriskos</ext-link>, Aristotle University of Thessaloniki, Greece</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3094806/overview">Abdulqahar Mukhtar Abubakar</ext-link>, Amrita Vishwa Vidyapeetham University, India</p></fn>
<fn fn-type="abbr" id="abbrev1"><p><bold>Abbreviations</bold> AI, artificial intelligence; NLP, natural language processing; CGI, clinical global impression; BPRS, brief psychiatric rating scale; BDI, Beck&#x0027;s depression inventory; GAF, global assessment of functioning; DSI-SS, depressive symptom index&#x2014;suicidality subscale; PRO, patient reported outcomes.</p></fn>
</fn-group>
</back>
</article>