<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="brief-report" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2025.1507159</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Digital Health</subject>
<subj-group>
<subject>Perspective</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>The role of trustworthy and reliable AI for multiple sclerosis</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes"><name><surname>Werthen-Brabants</surname><given-names>Lorin</given-names></name>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref><uri xlink:href="https://loop.frontiersin.org/people/2646066/overview"/><role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/><role content-type="https://credit.niso.org/contributor-roles/investigation/"/><role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/><role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/></contrib>
<contrib contrib-type="author"><name><surname>Dhaene</surname><given-names>Tom</given-names></name><uri xlink:href="https://loop.frontiersin.org/people/3008285/overview"/><role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/><role content-type="https://credit.niso.org/contributor-roles/supervision/"/><role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/></contrib>
<contrib contrib-type="author"><name><surname>Deschrijver</surname><given-names>Dirk</given-names></name><role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/><role content-type="https://credit.niso.org/contributor-roles/supervision/"/><role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/></contrib>
</contrib-group>
<aff><institution>SUMO Lab, IDLab, INTEC, Ghent University &#x2013; imec</institution>, <addr-line>Ghent</addr-line>, <country>Belgium</country></aff>
<author-notes>
<fn fn-type="edited-by"><p><bold>Edited by:</bold> Axel Faes, University of Hasselt, Belgium</p></fn>
<fn fn-type="edited-by"><p><bold>Reviewed by:</bold> Aonghus Lawlor, University College Dublin, Ireland</p></fn>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Lorin Werthen-Brabants <email>lorin.werthenbrabants@ugent.be</email></corresp>
</author-notes>
<pub-date pub-type="epub"><day>24</day><month>03</month><year>2025</year></pub-date>
<pub-date pub-type="collection"><year>2025</year></pub-date>
<volume>7</volume><elocation-id>1507159</elocation-id>
<history>
<date date-type="received"><day>08</day><month>10</month><year>2024</year></date>
<date date-type="accepted"><day>12</day><month>03</month><year>2025</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2025 Werthen-Brabants, Dhaene and Deschrijver.</copyright-statement>
<copyright-year>2025</copyright-year><copyright-holder>Werthen-Brabants, Dhaene and Deschrijver</copyright-holder><license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p></license>
</permissions>
<abstract>
<p>This paper investigates the importance of Trustworthy Machine Learning (ML) in the context of Multiple Sclerosis (MS) research and care. Due to the complex and individual nature of MS, the need for reliable and trustworthy ML models is essential. In this paper, key aspects of trustworthy ML, such as out-of-distribution generalization, explainability, uncertainty quantification and calibration are explored, highlighting their significance for healthcare applications. Challenges in integrating these ML tools into clinical workflows are addressed, discussing the difficulties in interpreting AI outputs, data diversity, and the need for comprehensive, quality data. It calls for collaborative efforts among researchers, clinicians, and policymakers to develop ML solutions that are technically sound, clinically relevant, and patient-centric.</p>
</abstract>
<kwd-group>
<kwd>artificial intelligence</kwd>
<kwd>multiple sclerosis</kwd>
<kwd>trustworthy AI</kwd>
<kwd>deep learning</kwd>
<kwd>uncertainty quantification</kwd>
</kwd-group><counts>
<fig-count count="0"/>
<table-count count="0"/><equation-count count="4"/><ref-count count="36"/><page-count count="6"/><word-count count="0"/></counts><custom-meta-wrap><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Informatics</meta-value></custom-meta></custom-meta-wrap>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>Machine Learning (ML) is increasingly applied to healthcare applications (<xref ref-type="bibr" rid="B1">1</xref>). While traditional statistical methods can help with biomarker discovery and recognizing trends and correlations, modern ML techniques such as Deep Learning (DL), are able to uncover complex correlations and provide better results than traditional, simpler techniques (<xref ref-type="bibr" rid="B2">2</xref>) due to their universal nature (<xref ref-type="bibr" rid="B3">3</xref>). Conversely, as these techniques become more complex, the need for <italic>reliable</italic> and <italic>trustworthy</italic> models increases (<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B5">5</xref>), especially within healthcare. However, building trust does not have a one-size-fits-all solution, resulting in many techniques to be developed to aid decision making.</p>
<p>For an end-user, be it a clinician or a patient, a model that is trustworthy is one that can provide certain guarantees on its predictions, explain its predictions, and provide a notion of uncertainty. For a complex disease such as Multiple Sclerosis (MS), the need for trustworthy models is especially pertinent, as its progression is non-trivially defined, and the decisions made to hinder its progression are important ones. A machine learning system that does not provide adequate reliability metrics, or trustworthy insights, will be less appealing to the end-user when there are high-stakes consequences. In recent years, the need for Trustworthy ML (TML) has also reached mainstream attention with the use of generative AI becoming more prevalent. For example, though Large Language Models have shown impressive results, they may still provide incorrect results, without any notion of uncertainty or trustworthiness (<xref ref-type="bibr" rid="B6">6</xref>). This is also known as the &#x201C;hallucination&#x201D; effect (<xref ref-type="bibr" rid="B7">7</xref>). Complex data and relationships warrant the use of trustworthiness techniques.</p>
<p>In the following Sections, we provide a summary of techniques present in Trustworthy ML (TML) (<xref ref-type="sec" rid="s2">Section 2</xref>), why TML is necessary for MS (<xref ref-type="sec" rid="s3a">Section 3.1</xref>), and the associated challenges (<xref ref-type="sec" rid="s3b">Section 3.2</xref>).</p>
</sec>
<sec id="s2"><label>2</label><title>Trustworthy machine learning</title>
<sec id="s2a"><label>2.1</label><title>Out-of-distribution generalization</title>
<p>The many ways in which MS progression can occur (different limbs, locations of lesion growth, etc.), makes the disease variable and patient specific. Therefore, training data will rarely contain enough data to cover the full extent of the ways progression can be observed. Furthermore, due to protocols changing regularly and equipment variability, <italic>concept</italic> or <italic>model drift</italic> (<xref ref-type="bibr" rid="B8">8</xref>) may pose a real issue when ML models are deployed in the real world. Model drift occurs when new data do not correspond to the data on which a model was trained. As a result, models must be continually adapted so changes in data distributions are captured.</p>
<p>These issues can be tackled by making use of techniques such as domain adaptation (<xref ref-type="bibr" rid="B9">9</xref>, <xref ref-type="bibr" rid="B10">10</xref>), a specific case of transfer learning (<xref ref-type="bibr" rid="B11">11</xref>), and synthetic data sampling such as SMOTE (<xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B13">13</xref>).</p>
<p>The concept of Out-of-Distribution Generalization can be elucidated by considering a concrete example within the MS context. Imagine an ML model trained on data from North American patients. When this model is applied to patients from different geographical regions with distinct genetic and environmental factors, its predictions may falter due to differences in disease manifestation. Domain adaptation techniques can help here by adjusting the model to account for these regional variations. Similarly, synthetic data sampling, like the aforementioned SMOTE technique, can artificially&#x2014;not necessarily in a representative way&#x2014;augment the dataset to include underrepresented samples in a given dataset, improving the model&#x2019;s robustness against a wide range of clinical scenarios. However, it must be stressed that data quality is key, and an underrepresented dataset can not fully capture the underlying factors to guarantee good out-of-distribution generalization.</p>
</sec>
<sec id="s2b"><label>2.2</label><title>Explainability and interpretability</title>
<p>A perfectly interpretable AI provides insights into the inner workings and decision process of an AI system. When it comes to the types of ML systems, they can broadly be divided into two categories: white-box models and black-box models.</p>
<sec id="s2b1"><label>2.2.1</label><title>White-box models</title>
<p>Models that are inherently explainable and interpretable. These are often simpler methods such as linear or logistic regression, the latter of which can be represented as a nomogram (<xref ref-type="bibr" rid="B14">14</xref>), a graphical representation of such models that visually convey the weight of different input variables. These models can be fully dissected, so there may be many ways of representing or explaining them.</p>
</sec>
<sec id="s2b2"><label>2.2.2</label><title>Black-box models</title>
<p>Models that can not be interpreted easily, and are regarded as a &#x201C;black box&#x201D; out of which little or no knowledge can be derived. However, there are techniques that can provide explainability when working with black-box models, such as making use of Shapley values (<xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B16">16</xref>) or making use of Deep Learning specific techniques (<xref ref-type="bibr" rid="B17">17</xref>) such as Layer-wise Relevance Propagation (<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B19">19</xref>). These are often post-hoc. In practice, these techniques will show a number of features and their importances expressed as a number. This could also be in the form of a heatmap. These feature importances may not always be as readily interpretable and may need training and education to comprehend adequately. Additionally, they do not necessarily <italic>explain</italic> why those features are important.</p>
<p>A classifier that may perform well in its evaluation metrics (sensitivity, specificity, ROC AUC, etc.) may still benefit from explainability methods. In particular, if models were to take into account many multimodal variables, the primary drivers of a given prediction may offer important insight for the user of the machine learning system.</p>
<p>Related to interpretable AI is explainable AI. Rather than being able to fully comprehend the inner workings of a model, an explainable AI model is able to be queried so that a reasonable explanation to the prediction is provided. Explainable AI can be viewed on different levels as well: Global, cohort, and local explainability. Global explainability provides information about the entire population or dataset. Due to the complex nature of the MS disease, valuable insights on a population level are scarce. Cohort explainability gives insight on subsets of the data, which can be more interesting when taking into account certain covariates. In this way, different groups of patients can be identified and correlations within these groups may offer more helpful insights than looking only at a global level. Lastly, local explainability provides insight on the model&#x2019;s output for a single input example. Every patient has a different profile, and therefore local explainability may help acquire insight into the prediction of the model for that specific patient or observation.</p>
</sec>
</sec>
<sec id="s2c"><label>2.3</label><title>Uncertainty quantification and calibration</title>
<sec id="s2c1"><label>2.3.1</label><title>Uncertainty quantification</title>
<p>In machine learning models, uncertainty plays a critical, yet understated role in understanding and interpreting predictions. Healthcare specifically can greatly benefit from uncertainty quantification, as it can add a layer of trust between the user and the model (<xref ref-type="bibr" rid="B20">20</xref>&#x2013;<xref ref-type="bibr" rid="B22">22</xref>). Two major sources of uncertainty are aleatoric and epistemic uncertainty (<xref ref-type="bibr" rid="B23">23</xref>).</p>
<sec id="s2c1a"><label>2.3.1.1</label><title>Aleatoric uncertainty</title>
<p>This type of irreducible uncertainty is inherent in the data itself. It cannot be reduced by adding more data and manifests as the noise within the data. An example of this uncertainty arises when using very few features. For example, a patient&#x2019;s blood pressure is a crucial health metric, but it exhibits natural variability within an individual due to various factors like stress, activity level, time of day, and even the way it is measured.</p>
<p>This uncertainty can be either homoscedastic, when it remains constant for all values (e.g., base noise of a sensor), or heteroscedastic, when it varies depending on the value of the sample.</p>
</sec>
<sec id="s2c1b"><label>2.3.1.2</label><title>Epistemic uncertainty.</title>
<p>Epistemic uncertainty arises from the model&#x2019;s limited knowledge. This reducible uncertainty is high when the model has insufficient data to characterize or capture the target variable. Increasing the size of the data set can help reduce epistemic uncertainty. An intuitive example can be demonstrated as follows: Say there are multiple experts for a single disease such as MS. These experts may disagree on a given prognosis, despite all of them being equally trained for such a task. Analogously, in a machine learning model predicting patient outcomes for MS, the model might exhibit high epistemic uncertainty if it has been trained on a limited or non-representative dataset. Just as the disagreement among experts might stem from variations in their individual experiences and interpretations, the model&#x2019;s uncertainty arises from its limited exposure to the diverse manifestations of the disease. By providing the model with more comprehensive data that captures a wider range of patient histories, symptoms, and outcomes, the epistemic uncertainty can be reduced, leading to more consistent and reliable predictions.</p>
<p>Applying uncertainty quantification in MS involves recognizing and managing the inherent unpredictability in patient responses and disease progression. For instance, a model expressing aleatoric uncertainty might show the variability in a patient&#x2019;s symptoms over time, acknowledging that certain aspects of MS progression cannot be predicted with complete precision. Epistemic uncertainty can be illustrated by a model&#x2019;s varying predictions based on different patient subgroups, reflecting limited knowledge about specific MS manifestations. To quantify and capture these uncertainties, techniques like Monte Carlo Dropout (MCD) (<xref ref-type="bibr" rid="B24">24</xref>) can be employed, providing a probabilistic understanding of a model&#x2019;s predictions and helping clinicians make informed decisions under uncertainty.</p>
<p>Uncertainty quantification has been applied to lesion detection in MRI images (<xref ref-type="bibr" rid="B25">25</xref>&#x2013;<xref ref-type="bibr" rid="B27">27</xref>), often making use of MCD or other methods of obtaining a model that can express uncertainty (<xref ref-type="bibr" rid="B28">28</xref>).</p>
</sec>
</sec>
<sec id="s2c2"><label>2.3.2</label><title>Calibration</title>
<p>A well-calibrated machine learning model is one in which the model&#x2019;s predicted probabilities closely match the probabilities observed in the actual data (<xref ref-type="bibr" rid="B29">29</xref>). Mathematically, this is represented as <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM1"><mml:mi>P</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>y</mml:mi><mml:mo fence="false" stretchy="false">|</mml:mo><mml:mrow><mml:mover><mml:mrow><mml:mspace width=".1em"/><mml:mi>p</mml:mi></mml:mrow><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>y</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula>. This equation signifies that the probability of an event <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM2"><mml:mi>y</mml:mi></mml:math></inline-formula> occurring, given that the model predicts it with probability <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM3"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula>, should ideally be <inline-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="IM4"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula> itself. As a practical example: a model that predicts the probability of 40&#x0025; disease progression for a patient will ideally be correct 40&#x0025; of the time of all patients who receive a similar prognosis. For methods such as neural networks, this is not often the case by default, and calibration needs to be improved. Additionally, calibration can also be applied to regressors that output a distribution, rather than a single value. In this case, the confidence interval (such as a 95&#x0025; confidence interval, for example) can be calibrated to ensure that it matches the observations.</p>
<p>The need for calibration is evident in the lack of information an uncalibrated classifier or regressor provides. Often, as is the case with neural networks, a neural network classifier will collapse to output probabilities close to 100&#x0025; or 0&#x0025; consistently, rather than providing accurate probability estimates (<xref ref-type="bibr" rid="B29">29</xref>). As a result, a user of such a system needs to blindly trust the classifier rather than being able to take the confidence of the classifier into account.</p>
</sec>
</sec>
</sec>
<sec id="s3" sec-type="discussion"><label>3</label><title>Discussion</title>
<sec id="s3a"><label>3.1</label><title>Why trustworthy ML is necessary for MS research</title>
<p>With the current knowledge of MS and performance of state-of-the-art machine learning models in the field, it stands to reason that there may not be a one-size-fits-all solution to detecting disease progression. Although other types of model (such as image classifiers) may perform very well and can reliably be used in most, if not all, cases, this may not be the case for MS. ML models for this purpose will likely be a tool to aid decision making, rather than a decision maker by itself. To that end, an ML model that just states &#x201C;yes&#x201D; or &#x201C;no&#x201D; is not sufficient. Rather, more information should be supplied to the user. A trustworthy version of this model will highlight parts of the input that contribute greatly to the prediction, show which global and cohort features are important, and also provide a notion of (un)certainty with the prediction. In this way, the user can:</p>
<p>
<list list-type="simple">
<list-item><label>&#x2022;</label>
<p>Select which predictions to trust and keep, both by using aleatoric and epistemic uncertainty as guides</p></list-item>
<list-item><label>&#x2022;</label>
<p>Analyze the subgroup in which the prediction fits</p></list-item>
<list-item><label>&#x2022;</label>
<p>Analyze the specific prediction and the features leading to the prediction</p></list-item>
</list>For MS research, the use and adoption of ML will be guided by advances in <italic>trustworthy</italic> ML. MS is a disease marked by its heterogeneity in symptoms, progression, and response to treatment, making reliable analysis of significant importance.</p>
<p>The ability of ML models to process and analyze different types of data&#x2014;from clinical observations to MRI images&#x2014;can lead to earlier detection and more precise monitoring of the disease&#x2019;s progression. However, the value of these insights depends on their explainability. Clinicians and patients must be able to understand and trust the model&#x2019;s predictions, necessitating a focus on explainable AI. For example, an ML model might identify subtle changes in brain lesions over time, but this information becomes clinically actionable only when it is presented in an understandable manner. Explainable models can elucidate the factors driving a prediction, thereby enhancing the clinician&#x2019;s ability to make informed treatment decisions.</p>
<p>Moreover, the integration of uncertainty quantification in ML models is particularly relevant for MS. Given the variability in how the disease presents and progresses, models that can express their confidence in predictions are invaluable. They provide clinicians with a more nuanced understanding of each prediction, facilitating more informed risk-benefit analyses when deciding on treatment plans. A model that indicates a high level of uncertainty in its prediction might prompt further testing or closer monitoring, whereas a prediction made with high confidence could lead to more decisive action.</p>
<p>The importance of trustworthy ML in MS research also extends to patient empowerment. Access to understandable and reliable ML-driven insights can foster better patient-clinician dialogues. When patients understand the basis for predictions about their condition, they are better positioned to make informed decisions about their treatment and lifestyle choices.</p>
</sec>
<sec id="s3b"><label>3.2</label><title>Challenges of trustworthy ML for MS</title>
<sec id="s3b1"><label>3.2.1</label><title>Integration of ML tools to aid clinical decisions</title>
<p>Integrating ML tools into existing clinical workflows presents another layer of complexity. For these tools to be adopted, they must fit into the highly regulated environment of healthcare. This integration involves designing user interfaces and metrics that are intuitive for clinicians, ensuring that ML predictions are presented in a way that complements decision-making processes rather than complicating them (<xref ref-type="bibr" rid="B30">30</xref>). Furthermore, imperfect data pose a problem during the training and prediction stages of an ML model. Data collection can be a laborious task, and in some cases the data cannot be accurately represented due to individual differences in disease expression. This rings especially true in the case of MS.</p>
</sec>
<sec id="s3b2"><label>3.2.2</label><title>Usability of uncertainty quantification and explainability techniques</title>
<p>As highlighted previously, UQ and explainability techniques have their merit, as they can highlight potential issues when making use of ML assisted decision systems. However, the end-user may not find much use in the way UQ results are represented in literature. Even explainabilty results have varying degrees of success concerning their usability (<xref ref-type="bibr" rid="B31">31</xref>). These techniques could benefit from user studies, as their usability hinges on the representation and, in turn, interpretation by the end-user. For example, rather than providing the clinician and/or patient with a numerical value signifying a &#x201C;trustworthiness&#x201D; score or certainty otherwise, larger trust could be gained by comparing the patient with other patients that have similar disease trajectories. This opacity can hinder trust and acceptance, especially in a high-stakes field like healthcare where understanding the &#x201C;why&#x201D; behind a diagnosis or prognosis is as crucial as the outcome itself (<xref ref-type="bibr" rid="B31">31</xref>).</p>
</sec>
<sec id="s3b3"><label>3.2.3</label><title>Out-of-distribution data, diverse data, available data</title>
<p>Data diversity and availability are critical factors that significantly influence the development and performance of ML models in MS research. MS is a disease with a highly variable clinical course and a wide range of symptoms that differ from patient to patient. This heterogeneity necessitates a rich and diverse dataset that captures the broad spectrum of the disease. After all, deep learning techniques are prone to overfitting, and may have performance below acceptable levels as a result (<xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B32">32</xref>). Initiatives such as MSBase (<xref ref-type="bibr" rid="B33">33</xref>, <xref ref-type="bibr" rid="B34">34</xref>) attempt to address the issue of out-of-distribution performance by providing multi-center data. The amount of data by itself may give the end-user a reason to trust a model, given enough diversity. Data quality is another concern, with issues such as missing values, inconsistent data entry, and the need for standardization across different data sources complicating the development of reliable ML models. Introducing diversity by including measurements that stray away from purely medical imaging or clinical data may also provide a new avenue of research, potentially discovering novel biomarkers. Future work should focus on developing models that can adapt to individual patient variations and incorporating emerging data types such as Motor Evoked Potentials (<xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B36">36</xref>) into ML models.</p>
</sec>
</sec>
</sec>
<sec id="s4" sec-type="conclusions"><label>4</label><title>Conclusion</title>
<p>This paper underscores the importance of trustworthiness in Machine Learning (ML) applications for Multiple Sclerosis (MS). Key aspects such as explainability, uncertainty quantification and calibration, and out-of-distribution generalization have been explored. Additionally, the challenges in integrating ML into clinical workflows and the hurdles posed by data diversity and availability have been discussed.</p>
<p>The authors urge the research community and healthcare providers to prioritize the development and implementation of trustworthy ML solutions for MS (and healthcare in general). There is an urgent need to foster partnerships between computer scientists, neurologists, and patients. This collaboration will ensure the development of ML solutions that are not only technically sound but also clinically relevant and patient-centric. Making comprehensive, high-quality data sets accessible while respecting privacy concerns is crucial. Initiatives should focus on standardizing data collection and sharing practices to aid in the development of more effective ML models. ML tools must be integrated into clinical workflows in a way that is intuitive and enhances decision-making processes. This involves designing user-friendly interfaces and ensuring that clinicians are adequately trained to use these tools effectively.</p>
</sec>
</body>
<back>
<sec id="s5" sec-type="data-availability"><title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec id="s6" sec-type="author-contributions"><title>Author contributions</title>
<p>LW-B: Conceptualization, Investigation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. TD: Funding acquisition, Supervision, Writing &#x2013; review &#x0026; editing. DD: Funding acquisition, Supervision, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec id="s7" sec-type="funding-information"><title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. This work has been supported by the Flemish Government under the &#x201C;Onderzoeksprogramma Artifici&#x00EB;le Intelligentie (AI) Vlaanderen&#x201D; program.</p>
</sec>
<sec id="s8" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s9" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec id="s10" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname><given-names>F</given-names></name><name><surname>Jiang</surname><given-names>Y</given-names></name><name><surname>Zhi</surname><given-names>H</given-names></name><name><surname>Dong</surname><given-names>Y</given-names></name><name><surname>Li</surname><given-names>H</given-names></name><name><surname>Ma</surname><given-names>S</given-names></name></person-group>, et al. <article-title>Artificial intelligence in healthcare: past, present and future</article-title>. <source>Stroke Vasc Neurol</source>. (<year>2017</year>) <volume>2</volume>:<fpage>230</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1136/svn-2017-000101</pub-id><pub-id pub-id-type="pmid">29507784</pub-id></citation></ref>
<ref id="B2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jumper</surname><given-names>J</given-names></name><name><surname>Evans</surname><given-names>R</given-names></name><name><surname>Pritzel</surname><given-names>A</given-names></name><name><surname>Green</surname><given-names>T</given-names></name><name><surname>Figurnov</surname><given-names>M</given-names></name><name><surname>Ronneberger</surname><given-names>O</given-names></name></person-group>, et al. <article-title>Highly accurate protein structure prediction with alphafold</article-title>. <source>Nature</source>. (<year>2021</year>) <volume>596</volume>:<fpage>583</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-021-03819-2</pub-id><pub-id pub-id-type="pmid">34265844</pub-id></citation></ref>
<ref id="B3"><label>3.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Cybenko</surname><given-names>G</given-names></name></person-group>. <article-title>Approximation by superpositions of a sigmoidal function</article-title>. <source>Math Control Signals Syst</source>. (<year>1989</year>) <volume>2</volume>:<fpage>303</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1007/BF02551274</pub-id></citation></ref>
<ref id="B4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wiens</surname><given-names>J</given-names></name><name><surname>Shenoy</surname><given-names>ES</given-names></name></person-group>. <article-title>Machine learning for healthcare: on the verge of a major shift in healthcare epidemiology</article-title>. <source>Clin Infect Dis</source>. (<year>2018</year>) <volume>66</volume>:<fpage>149</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.1093/cid/cix731</pub-id><pub-id pub-id-type="pmid">29020316</pub-id></citation></ref>
<ref id="B5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Thirunavukarasu</surname><given-names>AJ</given-names></name><name><surname>Ting</surname><given-names>DSJ</given-names></name><name><surname>Elangovan</surname><given-names>K</given-names></name><name><surname>Gutierrez</surname><given-names>L</given-names></name><name><surname>Tan</surname><given-names>TF</given-names></name><name><surname>Ting</surname><given-names>DSW</given-names></name></person-group>. <article-title>Large language models in medicine</article-title>. <source>Nat Med</source>. (<year>2023</year>) <volume>29</volume>:<fpage>1930</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1038/s41591-023-02448-8</pub-id><pub-id pub-id-type="pmid">37460753</pub-id></citation></ref>
<ref id="B6"><label>6.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sejnowski</surname><given-names>TJ</given-names></name></person-group>. <article-title>Large language models and the reverse turing test</article-title>. <source>Neural Comput</source>. (<year>2023</year>) <volume>35</volume>:<fpage>309</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1162/neco_a_01563</pub-id><pub-id pub-id-type="pmid">36746144</pub-id></citation></ref>
<ref id="B7"><label>7.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Shen</surname><given-names>Y</given-names></name><name><surname>Heacock</surname><given-names>L</given-names></name><name><surname>Elias</surname><given-names>J</given-names></name><name><surname>Hentel</surname><given-names>KD</given-names></name><name><surname>Reig</surname><given-names>B</given-names></name><name><surname>Shih</surname><given-names>G</given-names></name><etal/></person-group> <article-title>Data from: Chatgpt and other large language models are double-edged swords</article-title> (<year>2023</year>)</citation></ref>
<ref id="B8"><label>8.</label><citation citation-type="book"><person-group person-group-type="author"><name><surname>Tsymbal</surname><given-names>A</given-names></name></person-group>. <source>The Problem of Concept Drift: Definitions and Related Work</source>. Technical report (TCD-CS-2004-15). <publisher-loc>Dublin</publisher-loc>: <publisher-name>Computer Science Department, Trinity College Dublin</publisher-name> (<year>2004</year>). <comment>Vol. 106. p. 58</comment>.</citation></ref>
<ref id="B9"><label>9.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Farahani</surname><given-names>A</given-names></name><name><surname>Voghoei</surname><given-names>S</given-names></name><name><surname>Rasheed</surname><given-names>K</given-names></name><name><surname>Arabnia</surname><given-names>HR</given-names></name></person-group>. <article-title>A brief review of domain adaptation</article-title>. In: <source>Advances in Data Science and Information Engineering: Proceedings from ICDATA 2020 and IKE 2020</source> (<year>2021</year>). p. <fpage>877</fpage>&#x2013;<lpage>94</lpage>.</citation></ref>
<ref id="B10"><label>10.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Valverde</surname><given-names>S</given-names></name><name><surname>Salem</surname><given-names>M</given-names></name><name><surname>Cabezas</surname><given-names>M</given-names></name><name><surname>Pareto</surname><given-names>D</given-names></name><name><surname>Vilanova</surname><given-names>JC</given-names></name><name><surname>Rami&#x00F3;-Torrent&#x00E0;</surname><given-names>L</given-names></name></person-group>, et al. <article-title>One-shot domain adaptation in multiple sclerosis lesion segmentation using convolutional neural networks</article-title>. <source>NeuroImage Clin</source>. (<year>2019</year>) <volume>21</volume>:<fpage>101638</fpage>. <pub-id pub-id-type="doi">10.1016/j.nicl.2018.101638</pub-id><pub-id pub-id-type="pmid">30555005</pub-id></citation></ref>
<ref id="B11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Weiss</surname><given-names>K</given-names></name><name><surname>Khoshgoftaar</surname><given-names>TM</given-names></name><name><surname>Wang</surname><given-names>D</given-names></name></person-group>. <article-title>A survey of transfer learning</article-title>. <source>J Big Data</source>. (<year>2016</year>) <volume>3</volume>:<fpage>1</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1186/s40537-016-0043-6</pub-id></citation></ref>
<ref id="B12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Branco</surname><given-names>D</given-names></name><name><surname>Martino</surname><given-names>B</given-names></name><name><surname>Esposito</surname><given-names>A</given-names></name><name><surname>Tedeschi</surname><given-names>G</given-names></name><name><surname>Bonavita</surname><given-names>S</given-names></name><name><surname>Lavorgna</surname><given-names>L</given-names></name></person-group>. <article-title>Machine learning techniques for prediction of multiple sclerosis progression</article-title>. <source>Soft Comput</source>. (<year>2022</year>) <volume>26</volume>:<fpage>12041</fpage>&#x2013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1007/s00500-022-07503-z</pub-id></citation></ref>
<ref id="B13"><label>13.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chawla</surname><given-names>NV</given-names></name><name><surname>Bowyer</surname><given-names>KW</given-names></name><name><surname>Hall</surname><given-names>LO</given-names></name><name><surname>Kegelmeyer</surname><given-names>WP</given-names></name></person-group>. <article-title>Smote: synthetic minority over-sampling technique</article-title>. <source>J Artif Intell Res</source>. (<year>2002</year>) <volume>16</volume>:<fpage>321</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1613/jair.953</pub-id></citation></ref>
<ref id="B14"><label>14.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Kattan</surname><given-names>MW</given-names></name><name><surname>Marasco</surname><given-names>J</given-names></name></person-group>. <article-title>What is a real nomogram?</article-title> In: <source>Seminars in Oncology</source>. <publisher-name>Elsevier</publisher-name> (<year>2010</year>). Vol. <volume>37</volume>. p. <fpage>23</fpage>&#x2013;<lpage>6</lpage>.</citation></ref>
<ref id="B15"><label>15.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lundberg</surname><given-names>SM</given-names></name><name><surname>Lee</surname><given-names>SI</given-names></name></person-group>. <article-title>A unified approach to interpreting model predictions</article-title>. <source>Adv Neural Inform Process Syst</source>. (<year>2017</year>) <volume>30</volume>:<fpage>4768</fpage>&#x2013;<lpage>77</lpage>.</citation></ref>
<ref id="B16"><label>16.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Basu</surname><given-names>S</given-names></name><name><surname>Munafo</surname><given-names>A</given-names></name><name><surname>Ben-Amor</surname><given-names>AF</given-names></name><name><surname>Roy</surname><given-names>S</given-names></name><name><surname>Girard</surname><given-names>P</given-names></name><name><surname>Terranova</surname><given-names>N</given-names></name></person-group>. <article-title>Predicting disease activity in patients with multiple sclerosis: an explainable machine-learning approach in the mavenclad trials</article-title>. <source>CPT Pharmacom Syst Pharmacol</source>. (<year>2022</year>) <volume>11</volume>:<fpage>843</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.1002/psp4.12796</pub-id></citation></ref>
<ref id="B17"><label>17.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Chakraborty</surname><given-names>S</given-names></name><name><surname>Tomsett</surname><given-names>R</given-names></name><name><surname>Raghavendra</surname><given-names>R</given-names></name><name><surname>Harborne</surname><given-names>D</given-names></name><name><surname>Alzantot</surname><given-names>M</given-names></name><name><surname>Cerutti</surname><given-names>F</given-names></name><etal/></person-group> <article-title>Interpretability of deep learning models: a survey of results</article-title>. In: <source>2017 IEEE Smartworld, Ubiquitous Intelligence &#x0026; Computing, Advanced &#x0026; Trusted Computed, Scalable Computing &#x0026; Communications, Cloud &#x0026; Big Data Computing, Internet of People and Smart City Innovation (Smartworld/SCALCOM/UIC/ATC/CBDcom/IOP/SCI)</source>. IEEE (<year>2017</year>). p. <fpage>1</fpage>&#x2013;<lpage>6</lpage>.</citation></ref>
<ref id="B18"><label>18.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Creagh</surname><given-names>AP</given-names></name><name><surname>Lipsmeier</surname><given-names>F</given-names></name><name><surname>Lindemann</surname><given-names>M</given-names></name><name><surname>Vos</surname><given-names>MD</given-names></name></person-group>. <article-title>Interpretable deep learning for the remote characterisation of ambulation in multiple sclerosis using smartphones</article-title>. <source>Sci Rep</source>. (<year>2021</year>) <volume>11</volume>:<fpage>14301</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-021-92776-x</pub-id><pub-id pub-id-type="pmid">34253769</pub-id></citation></ref>
<ref id="B19"><label>19.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Montavon</surname><given-names>G</given-names></name><name><surname>Binder</surname><given-names>A</given-names></name><name><surname>Lapuschkin</surname><given-names>S</given-names></name><name><surname>Samek</surname><given-names>W</given-names></name><name><surname>M&#x00FC;ller</surname><given-names>KR</given-names></name></person-group>. <article-title>Layer-wise relevance propagation: an overview</article-title>. <source>Explainable AI</source>. (<year>2019</year>) <volume>11700</volume>:<fpage>193</fpage>&#x2013;<lpage>209</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-28954-6_10</pub-id></citation></ref>
<ref id="B20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Seoni</surname><given-names>S</given-names></name><name><surname>Jahmunah</surname><given-names>V</given-names></name><name><surname>Salvi</surname><given-names>M</given-names></name><name><surname>Barua</surname><given-names>PD</given-names></name><name><surname>Molinari</surname><given-names>F</given-names></name><name><surname>Acharya</surname><given-names>UR</given-names></name></person-group>. <article-title>Application of uncertainty quantification to artificial intelligence in healthcare: a review of last decade (2013&#x2013;2023)</article-title>. <source>Comput Biol Med</source>. (<year>2023</year>) <volume>165</volume>:<fpage>107441</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.107441</pub-id><pub-id pub-id-type="pmid">37683529</pub-id></citation></ref>
<ref id="B21"><label>21.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Begoli</surname><given-names>E</given-names></name><name><surname>Bhattacharya</surname><given-names>T</given-names></name><name><surname>Kusnezov</surname><given-names>D</given-names></name></person-group>. <article-title>The need for uncertainty quantification in machine-assisted medical decision making</article-title>. <source>Nat Mach Intell</source>. (<year>2019</year>) <volume>1</volume>:<fpage>20</fpage>&#x2013;<lpage>3</lpage>. <pub-id pub-id-type="doi">10.1038/s42256-018-0004-1</pub-id></citation></ref>
<ref id="B22"><label>22.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Lambert</surname><given-names>B</given-names></name><name><surname>Forbes</surname><given-names>F</given-names></name><name><surname>Doyle</surname><given-names>S</given-names></name><name><surname>Dehaene</surname><given-names>H</given-names></name><name><surname>Dojat</surname><given-names>M</given-names></name></person-group>. <article-title>Trustworthy clinical ai solutions: a unified review of uncertainty quantification in deep learning models for medical image analysis</article-title>. <source>Artif Intell Med</source>. (<year>2024</year>) <volume>150</volume>:<fpage>102830</fpage>. <pub-id pub-id-type="doi">10.1016/j.artmed.2024.102830</pub-id><pub-id pub-id-type="pmid">38553168</pub-id></citation></ref>
<ref id="B23"><label>23.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>H&#x00FC;llermeier</surname><given-names>E</given-names></name><name><surname>Waegeman</surname><given-names>W</given-names></name></person-group>. <article-title>Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods</article-title>. <source>Mach Learn</source>. (<year>2021</year>) <volume>110</volume>:<fpage>457</fpage>&#x2013;<lpage>506</lpage>. <pub-id pub-id-type="doi">10.1007/s10994-021-05946-3</pub-id></citation></ref>
<ref id="B24"><label>24.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Gal</surname><given-names>Y</given-names></name><name><surname>Ghahramani</surname><given-names>Z</given-names></name></person-group>. <article-title>Dropout as a bayesian approximation: representing model uncertainty in deep learning</article-title>. In: <source>International Conference on Machine Learning</source>. PMLR (<year>2016</year>). p. <fpage>1050</fpage>&#x2013;<lpage>9</lpage>.</citation></ref>
<ref id="B25"><label>25.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nair</surname><given-names>T</given-names></name><name><surname>Precup</surname><given-names>D</given-names></name><name><surname>Arnold</surname><given-names>DL</given-names></name><name><surname>Arbel</surname><given-names>T</given-names></name></person-group>. <article-title>Exploring uncertainty measures in deep networks for multiple sclerosis lesion detection and segmentation</article-title>. <source>Med Image Anal</source>. (<year>2020</year>) <volume>59</volume>:<fpage>101557</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2019.101557</pub-id><pub-id pub-id-type="pmid">31677438</pub-id></citation></ref>
<ref id="B26"><label>26.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Molchanova</surname><given-names>N</given-names></name><name><surname>Raina</surname><given-names>V</given-names></name><name><surname>Malinin</surname><given-names>A</given-names></name><name><surname>La Rosa</surname><given-names>F</given-names></name><name><surname>Muller</surname><given-names>H</given-names></name><name><surname>Gales</surname><given-names>M</given-names></name><etal/></person-group> <article-title>Novel structural-scale uncertainty measures and error retention curves: application to multiple sclerosis</article-title>. In: <source>2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI)</source>. IEEE (<year>2023</year>). p. <fpage>1</fpage>&#x2013;<lpage>5</lpage>.</citation></ref>
<ref id="B27"><label>27.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Tousignant</surname><given-names>A</given-names></name><name><surname>Lema&#x00EE;tre</surname><given-names>P</given-names></name><name><surname>Precup</surname><given-names>D</given-names></name><name><surname>Arnold</surname><given-names>DL</given-names></name><name><surname>Arbel</surname><given-names>T</given-names></name></person-group>. <article-title>Prediction of disease progression in multiple sclerosis patients using deep learning analysis of MRI data</article-title>. In: <source>International Conference on Medical Imaging with Deep Learning</source>. PMLR (<year>2019</year>). p. <fpage>483</fpage>&#x2013;<lpage>92</lpage>.</citation></ref>
<ref id="B28"><label>28.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Lambert</surname><given-names>B</given-names></name><name><surname>Forbes</surname><given-names>F</given-names></name><name><surname>Doyle</surname><given-names>S</given-names></name><name><surname>Tucholka</surname><given-names>A</given-names></name><name><surname>Dojat</surname><given-names>M</given-names></name></person-group>. <article-title>Fast uncertainty quantification for deep learning-based MR brain segmentation</article-title>. In: <source>EGC 2022-Conference Francophone Pour l&#x2019;Extraction et la Gestion des Connaissances</source>. (<year>2022</year>). p. <fpage>1</fpage>&#x2013;<lpage>12</lpage>.</citation></ref>
<ref id="B29"><label>29.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Guo</surname><given-names>C</given-names></name><name><surname>Pleiss</surname><given-names>G</given-names></name><name><surname>Sun</surname><given-names>Y</given-names></name><name><surname>Weinberger</surname><given-names>KQ</given-names></name></person-group>. <article-title>On calibration of modern neural networks</article-title>. In: <source>International Conference on Machine Learning</source>. PMLR (<year>2017</year>). p. <fpage>1321</fpage>&#x2013;<lpage>30</lpage>.</citation></ref>
<ref id="B30"><label>30.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Dabbs</surname><given-names>ADV</given-names></name><name><surname>Myers</surname><given-names>BA</given-names></name><name><surname>Mc Curry</surname><given-names>KR</given-names></name><name><surname>Dunbar-Jacob</surname><given-names>J</given-names></name><name><surname>Hawkins</surname><given-names>RP</given-names></name><name><surname>Begey</surname><given-names>A</given-names></name></person-group>, et al. <article-title>User-centered design and interactive health technologies for patients</article-title>. <source>Comput Inform Nurs</source>. (<year>2009</year>) <volume>27</volume>:<fpage>175</fpage>. <pub-id pub-id-type="doi">10.1097/NCN.0b013e31819f7c7c</pub-id><pub-id pub-id-type="pmid">19411947</pub-id></citation></ref>
<ref id="B31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rasheed</surname><given-names>K</given-names></name><name><surname>Qayyum</surname><given-names>A</given-names></name><name><surname>Ghaly</surname><given-names>M</given-names></name><name><surname>Al-Fuqaha</surname><given-names>A</given-names></name><name><surname>Razi</surname><given-names>A</given-names></name><name><surname>Qadir</surname><given-names>J</given-names></name></person-group>. <article-title>Explainable, trustworthy, and ethical machine learning for healthcare: a survey</article-title>. <source>Comput Biol Med</source>. (<year>2022</year>) <volume>149</volume>:<fpage>106043</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.106043</pub-id><pub-id pub-id-type="pmid">36115302</pub-id></citation></ref>
<ref id="B32"><label>32.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>M&#x00E5;rtensson</surname><given-names>G</given-names></name><name><surname>Ferreira</surname><given-names>D</given-names></name><name><surname>Granberg</surname><given-names>T</given-names></name><name><surname>Cavallin</surname><given-names>L</given-names></name><name><surname>Oppedal</surname><given-names>K</given-names></name><name><surname>Padovani</surname><given-names>A</given-names></name></person-group>, et al. <article-title>The reliability of a deep learning model in clinical out-of-distribution MRI data: a multicohort study</article-title>. <source>Med Image Anal</source>. (<year>2020</year>) <volume>66</volume>:<fpage>101714</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2020.101714</pub-id></citation></ref>
<ref id="B33"><label>33.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Butzkueven</surname><given-names>H</given-names></name><name><surname>Chapman</surname><given-names>J</given-names></name><name><surname>Cristiano</surname><given-names>E</given-names></name><name><surname>Grand&#x2019;Maison</surname><given-names>F</given-names></name><name><surname>Hoffmann</surname><given-names>M</given-names></name><name><surname>Izquierdo</surname><given-names>G</given-names></name></person-group>, et al. <article-title>Msbase: an international, online registry and platform for collaborative outcomes research in multiple sclerosis</article-title>. <source>Mult Scler J</source>. (<year>2006</year>) <volume>12</volume>:<fpage>769</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1177/1352458506070775</pub-id></citation></ref>
<ref id="B34"><label>34.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>De Brouwer</surname><given-names>E</given-names></name><name><surname>Becker</surname><given-names>T</given-names></name><name><surname>Werthen-Brabants</surname><given-names>L</given-names></name><name><surname>Dewulf</surname><given-names>P</given-names></name><name><surname>Iliadis</surname><given-names>D</given-names></name><name><surname>Dekeyser</surname><given-names>C</given-names></name></person-group>, et al. <article-title>Machine-learning-based prediction of disability progression in multiple sclerosis: an observational, international, multi-center study</article-title>. <source>PLoS Digit Health</source>. (<year>2024</year>) <volume>3</volume>:<fpage>e0000533</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000533</pub-id><pub-id pub-id-type="pmid">39052668</pub-id></citation></ref>
<ref id="B35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Rossini</surname><given-names>PM</given-names></name><name><surname>Rossi</surname><given-names>S</given-names></name></person-group>. <article-title>Clinical applications of motor evoked potentials</article-title>. <source>Electroencephalogr Clin Neurophysiol</source>. (<year>1998</year>) <volume>106</volume>:<fpage>180</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1016/S0013-4694(97)00097-7</pub-id><pub-id pub-id-type="pmid">9743275</pub-id></citation></ref>
<ref id="B36"><label>36.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yperman</surname><given-names>J</given-names></name><name><surname>Popescu</surname><given-names>V</given-names></name><name><surname>Van Wijmeersch</surname><given-names>B</given-names></name><name><surname>Becker</surname><given-names>T</given-names></name><name><surname>Peeters</surname><given-names>LM</given-names></name></person-group>. <article-title>Motor evoked potentials for multiple sclerosis, a multiyear follow-up dataset</article-title>. <source>Sci Data</source>. (<year>2022</year>) <volume>9</volume>:<fpage>207</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-022-01335-0</pub-id><pub-id pub-id-type="pmid">35577808</pub-id></citation></ref></ref-list>
</back>
</article>