<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="brief-report" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id><journal-title-group>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title></journal-title-group>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2026.1779041</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Perspective</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Operationalizing trustworthy artificial intelligence in clinical and operational workflows</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes"><name><surname>Khashu</surname><given-names>Kunal</given-names></name>
<xref ref-type="aff" rid="aff1"/>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2934672/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
</contrib-group>
<aff id="aff1"><institution>HCA Healthcare</institution>, <city>Nashville</city>, <state>TN</state>, <country country="us">United States</country></aff>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Kunal Khashu <email xlink:href="mailto:kunal.khashu@gmail.com">kunal.khashu@gmail.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-13"><day>13</day><month>02</month><year>2026</year></pub-date>
<pub-date publication-format="electronic" date-type="collection"><year>2026</year></pub-date>
<volume>8</volume><elocation-id>1779041</elocation-id>
<history>
<date date-type="received"><day>31</day><month>12</month><year>2025</year></date>
<date date-type="rev-recd"><day>16</day><month>01</month><year>2026</year></date>
<date date-type="accepted"><day>21</day><month>01</month><year>2026</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 Khashu.</copyright-statement>
<copyright-year>2026</copyright-year><copyright-holder>Khashu</copyright-holder><license><ali:license_ref start_date="2026-02-13">https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p></license>
</permissions>
<abstract>
<p>Artificial intelligence (AI) is increasingly deployed across healthcare systems to support clinical decision-making, optimize operational processes, and improve population health outcomes. Despite substantial investment and rapid advances in model performance, real-world adoption and sustained impact remain limited. A central barrier is the challenge of trust among clinicians, administrators, patients, and regulators in AI-enabled systems. While the concept of &#x201C;trustworthy AI&#x201D; is widely invoked, existing frameworks largely emphasize technical model properties and ethical principles without sufficient guidance for operational implementation. This paper argues that trustworthiness is not an intrinsic attribute of AI models but an emergent property of socio-technical systems in which AI is embedded. We propose a comprehensive framework for operationalizing trustworthy AI that shifts attention from model-centric validation to workflow-level design, governance, and continuous evaluation. By integrating decision-centered design, human&#x2013;AI role delineation, failure visibility, embedded accountability, and longitudinal performance monitoring, the framework provides a pragmatic foundation for deploying AI systems that are safe, equitable, and sustainable in both clinical and operational contexts.</p>
</abstract>
<kwd-group>
<kwd>artificial intelligence (AI)</kwd>
<kwd>digital health (DH)</kwd>
<kwd>digital transformation</kwd>
<kwd>healthcare administration</kwd>
<kwd>responsible AI</kwd>
</kwd-group><funding-group><funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement></funding-group><counts>
<fig-count count="1"/>
<table-count count="0"/><equation-count count="0"/><ref-count count="9"/><page-count count="8"/><word-count count="5486"/></counts><custom-meta-group><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Technology Implementation</meta-value></custom-meta></custom-meta-group>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>Artificial intelligence has become a central pillar of contemporary digital health strategy. Applications now span diagnostic imaging, clinical risk prediction, triage, workforce scheduling, capacity management, revenue cycle optimization, and population health surveillance. Numerous studies report impressive algorithmic performance in retrospective evaluations, often exceeding human benchmarks in narrowly defined tasks. Yet, translation of these advances into consistent improvements in patient outcomes, operational efficiency, or clinician experience has been uneven.</p>
<p>This gap between technical promise and real-world impact has prompted increasing scrutiny of how AI systems are designed, deployed, and governed in healthcare settings. Concerns related to bias, transparency, safety, accountability, and unintended consequences have catalyzed the emergence of the concept of <italic>trustworthy AI</italic>. Global organizations, including the World Health Organization (WHO) and international research consortia, have articulated ethical and operational principles intended to guide responsible AI development and deployment (<xref ref-type="bibr" rid="B1">1</xref>).</p>
<p>Despite this progress, health systems continue to face practical challenges in translating abstract principles into operational reality. Many AI initiatives fail not because models are inaccurate, but because they are misaligned with clinical workflows, poorly integrated into decision processes, or insufficiently governed once deployed (<xref ref-type="bibr" rid="B2">2</xref>). As a result, clinicians may distrust AI outputs, override recommendations inconsistently, or disengage from AI tools altogether. The effectiveness of artificial intelligence in healthcare depends on its ability to augment human expertise rather than replace clinical judgment, requiring careful integration into workflows that preserve professional autonomy and accountability (<xref ref-type="bibr" rid="B3">3</xref>).</p>
<p>This paper contends that these challenges stem from an overly narrow, model-centric conception of trustworthiness. We argue that trustworthiness must be understood as a system-level property that emerges from the interaction between AI technologies, human actors, organizational processes, and governance structures. Building on existing guidelines and socio-technical theory, we propose a framework for operationalizing trustworthy AI that emphasizes workflow integration, decision accountability, and continuous oversight across the AI lifecycle.</p>
</sec>
<sec id="s2"><label>2</label><title>The evolution of trustworthy AI in healthcare</title>
<sec id="s2a"><label>2.1</label><title>From technical performance to ethical principles</title>
<p>Early AI research in healthcare focused primarily on predictive accuracy, sensitivity, and specificity. As models became more complex and influential, concerns emerged regarding bias, explainability, and generalizability. In response, scholars and policymakers advanced ethical principles intended to constrain harmful applications of AI, often drawing on broader discussions of AI ethics across industries.</p>
<p>The WHO&#x0027;s guidance on the ethics and governance of AI for health articulates six core principles, including human autonomy, well-being, transparency, accountability, inclusiveness, and sustainability (<xref ref-type="bibr" rid="B1">1</xref>).</p>
<p>These frameworks represent an important shift from narrow technical evaluation toward a more holistic understanding of AI&#x0027;s societal and clinical implications. However, they remain largely aspirational, offering limited direction on how principles should be operationalized within specific workflows or organizational contexts.</p>
</sec>
<sec id="s2b"><label>2.2</label><title>Limitations of principle-based approaches</title>
<p>Principle-based frameworks are intentionally broad to accommodate diverse healthcare systems and use cases. Yet this generality also limits their operational utility. Health systems implementing AI must make concrete decisions about workflow integration, user interfaces, escalation pathways, and governance responsibilities; areas that are only indirectly addressed by high-level ethical guidance.</p>
<p>Moreover, principles such as fairness or transparency may conflict in practice. For example, increasing model complexity may improve accuracy but reduce interpretability; aggressive bias mitigation may alter performance across subpopulations in unintended ways. Without explicit mechanisms for managing these trade-offs, health systems are left to navigate trustworthiness challenges on an <italic>ad hoc</italic> basis (<xref ref-type="bibr" rid="B4">4</xref>).</p>
<p>These limitations suggest the need for frameworks that bridge ethical aspirations with operational realities, explicitly addressing how AI systems function within everyday clinical and administrative work.</p>
</sec>
</sec>
<sec id="s3"><label>3</label><title>Trustworthiness as a socio-technical property</title>
<sec id="s3a"><label>3.1</label><title>A shift in conceptualization</title>
<p>Traditional evaluations of AI trustworthiness focus on properties of the algorithm itself. In contrast, socio-technical theory emphasizes that system behavior arises from the interaction between technical components and social context. From this perspective, an AI model cannot be trustworthy in isolation; its trustworthiness depends on how it is used, interpreted, monitored, and governed.</p>
<p>This framing aligns with empirical observations that trust failures often occur despite technically sound models. For instance, clinicians may distrust accurate models if recommendations conflict with clinical intuition or workflow constraints. Conversely, poorly performing models may be trusted excessively if embedded in authoritative interfaces or institutional mandates.</p>
<p>Explainability has been identified as a necessary but insufficient condition for trustworthy AI, as understanding must be situated within clinical context, professional responsibility, and governance structures (<xref ref-type="bibr" rid="B5">5</xref>).</p>
<p>Understanding trustworthiness as emergent shifts analytical focus from pre-deployment validation to ongoing system performance, user behavior, and organizational accountability.</p>
</sec>
<sec id="s3b"><label>3.2</label><title>Implications for AI lifecycle management</title>
<p>If trustworthiness is emergent, it cannot be fully established at the point of deployment. Instead, it must be continuously produced through monitoring, feedback, and adaptation. This implies that responsibility for trustworthiness extends beyond data scientists to include clinicians, operational leaders, informatics teams, and governance bodies (<xref ref-type="bibr" rid="B6">6</xref>).</p>
<p>Operationalizing trustworthy AI therefore requires explicit attention to workflow design, decision authority, failure management, and longitudinal evaluation; domains often underemphasized in technical AI research.</p>
</sec>
</sec>
<sec id="s4"><label>4</label><title>A framework for operationalizing trustworthy AI</title>
<p>We propose a framework organized around five interdependent domains: decision-centered design, human-AI role delineation, failure visibility, embedded governance, and continuous evaluation. Together, these domains translate high-level trust principles into operational practices.</p>
<sec id="s4a"><label>4.1</label><title>Decision-centered design</title>
<p>Trustworthy AI systems should be designed around the <italic>decisions</italic> they are intended to influence, rather than around predictive tasks alone. This requires explicit articulation of the decision context, including:
<list list-type="order">
<list-item>
<p>The clinical or operational decision to be supported</p></list-item>
<list-item>
<p>The user responsible for the decision</p></list-item>
<list-item>
<p>The expected relationship between AI output and human judgment</p></list-item>
</list>Without this clarity, AI systems risk becoming either prescriptive tools that override human expertise or advisory tools that are routinely ignored (<xref ref-type="bibr" rid="B7">7</xref>).</p>
<p>Decision-centered design also requires alignment with organizational incentives and constraints. For example, an AI system designed to reduce length of stay may conflict with clinician priorities related to patient safety unless decision logic and accountability are explicitly defined.</p>
</sec>
<sec id="s4b"><label>4.2</label><title>Human-AI role delineation</title>
<p>Ambiguity regarding human and AI roles is a common source of implementation failure. Trustworthy deployment requires clear delineation of tasks that are automated, tasks that are augmented, and tasks that remain exclusively human.</p>
<p>Role delineation should be stable and visible. Frequent, unexplained changes in AI behavior or scope can erode clinician confidence and increase cognitive burden. Importantly, delineation should include explicit override mechanisms, enabling clinicians to contest AI recommendations without penalty.</p>
<p>This approach mitigates automation bias while preserving the potential benefits of decision support.</p>
</sec>
<sec id="s4c"><label>4.3</label><title>Failure visibility and safe degradation</title>
<p>Healthcare environments are characterized by uncertainty, incomplete data, and evolving conditions. Trustworthy AI systems must therefore anticipate failure rather than assume perfect performance.</p>
<p>Failure visibility mechanisms include confidence scores, uncertainty estimates, and alerts for out-of-distribution inputs (<xref ref-type="bibr" rid="B4">4</xref>). When performance degrades, systems should default to safe modes that preserve human control rather than continuing to generate misleading outputs.</p>
<p>Opacity during failure is particularly damaging to trust. Systems that fail silently may continue to influence decisions long after their reliability has diminished.</p>
</sec>
<sec id="s4d"><label>4.4</label><title>Embedded governance and accountability</title>
<p>Operational trustworthiness depends on governance structures that define responsibility across the AI lifecycle. Governance should encompass:
<list list-type="bullet">
<list-item>
<p>Pre-deployment risk assessment and ethical review</p></list-item>
<list-item>
<p>Clear ownership of deployed models</p></list-item>
<list-item>
<p>Processes for monitoring performance, bias, and drift</p></list-item>
<list-item>
<p>Authority to modify, suspend, or retire systems when necessary</p></list-item>
</list>The European Commission&#x0027;s AI evidence pathway emphasizes that governance must be continuous and evidence-driven, rather than episodic or compliance-focused (<xref ref-type="bibr" rid="B6">6</xref>).</p>
<p>Embedding governance into routine operations ensures that trustworthiness is actively maintained rather than retrospectively audited.</p>
</sec>
<sec id="s4e"><label>4.5</label><title>Continuous evaluation beyond accuracy</title>
<p>Post-deployment evaluation should extend beyond traditional performance metrics to capture system-level impact. Relevant measures include:
<list list-type="bullet">
<list-item>
<p>User adoption and override rates</p></list-item>
<list-item>
<p>Effects on clinician workload and cognitive load</p></list-item>
<list-item>
<p>Downstream clinical and operational outcomes</p></list-item>
<list-item>
<p>Differential effects across demographic and socioeconomic groups</p></list-item>
</list>Such evaluation aligns with the FUTURE-AI emphasis on deployability and real-world impact, recognizing that accuracy alone is insufficient to justify continued use</p>
</sec>
<sec id="s4f"><label>4.6</label><title>The decision-governed trustworthy AI (DG-TAI) framework</title>
<p>To consolidate the operational principles described above, we refer to the proposed model as the Decision-Governed Trustworthy AI (DG-TAI) framework (<xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>). The naming is intentional and descriptive rather than proprietary. It reflects the central premise that trustworthiness in healthcare AI is governed not at the level of models or predictions, but at the level of <italic>decisions</italic> specifically, how algorithmic outputs are interpreted, acted upon, overridden, and monitored within real-world clinical and operational workflows.</p>
<fig id="F1" position="float"><label>Figure&#x00A0;1</label>
<caption><p>Conceptual architecture of the decision-governed trustworthy AI (DG-TAI) framework. The framework positions clinical and operational decisions not algorithmic predictions as the central locus of trust. AI model outputs are embedded within workflows that define decision authority, human&#x2013;AI role delineation, and override mechanisms. Trustworthiness emerges at runtime through governance structures that ensure failure visibility, accountability, and continuous performance evaluation across clinical, operational, and equity dimensions.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1779041-g001.tif"><alt-text content-type="machine-generated">Flowchart illustrating AI deployment in healthcare. At the top is \"Healthcare Context\" influencing the entire process. Steps in order: \"AI Model Outputs,\" \"Workflow Integration,\" \"Decision Governance (Core Layer)\" highlighted in blue, \"Runtime Governance and Failure Visibility,\" \"System-Level Outcomes,\" and \"Continuous Feedback Loop.\" Each step describes specific functions, such as predictions, integration, decision authority, monitoring, and outcome evaluations within healthcare AI systems.</alt-text>
</graphic>
</fig>
<p>The DG-TAI framework does not introduce new ethical principles or technical requirements. Instead, it integrates established trustworthiness constructs into a decision-centric governance architecture that emphasizes accountability, failure visibility, and longitudinal oversight. In doing so, it provides a unifying structure that links abstract trust principles to operational mechanisms that can be implemented, evaluated, and adapted over time.</p>
<p>This framing is particularly relevant in healthcare settings, where AI systems rarely operate autonomously and instead shape decisions distributed across clinicians, administrators, and multidisciplinary teams. By making decision governance the organizing unit of analysis, the DG-TAI framework addresses a persistent gap between model validation and real-world system behavior.</p>
</sec>
<sec id="s4g"><label>4.7</label><title>Conceptual architecture of the DG-TAI framework</title>
<p>As illustrated in <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>, the DG-TAI framework is organized around five interdependent layers: model outputs, workflow integration, decision execution, governance mechanisms, and observed outcomes.</p>
<p>At the foundational layer, AI models generate predictions or recommendations. These outputs are not treated as decisions in themselves, but as inputs to structured workflows that define how information is surfaced to users, what actions are permissible, and where discretion resides. This distinction is critical: while models produce estimates, responsibility for action remains human and institutionally governed.</p>
<p>The workflow layer specifies human AI role delineation, including which decisions are automated, augmented, or reserved exclusively for human judgment. Importantly, this layer also defines override pathways and escalation mechanisms, ensuring that AI systems support rather than constrain.</p>
<p>Decisions executed within workflows then interact with governance structures that operate continuously rather than episodically. These structures include performance monitoring, bias surveillance, incident reporting, and authority to intervene when systems deviate from expected behavior. Governance is thus embedded into routine operations rather than treated as external compliance.</p>
<p>Finally, outcomes such as clinical, operational, experiential, and equity-related feed back into governance processes, enabling iterative refinement. Trustworthiness is therefore conceptualized as a dynamic system property that is produced and maintained through ongoing interaction between technology, humans, and institutions.</p>
</sec>
</sec>
<sec id="s5"><label>5</label><title>Trustworthy AI in operational workflows</title>
<p>While much of the trustworthy AI literature focuses on clinical decision support, operational AI systems such as those used for staffing, scheduling, or capacity management, pose comparable risks. These systems influence care indirectly by shaping resource availability and prioritization, yet their effects may be less visible to frontline clinicians.</p>
<p>Operational AI systems require the same trust principles as clinical tools, with particular emphasis on transparency and governance. Because operational decisions often affect large populations, small biases or errors can produce systemic inequities if left unchecked (<xref ref-type="bibr" rid="B1">1</xref>).</p>
<p>Embedding trustworthy AI into operational workflows therefore requires cross-functional governance and explicit mechanisms for contestability and oversight.</p>
</sec>
<sec id="s6"><label>6</label><title>Implications for health systems and policy</title>
<p>Operationalizing trustworthy AI demands organizational investment beyond technical development. Health systems must cultivate interdisciplinary expertise spanning informatics, clinical leadership, operations, and ethics. Training programs should emphasize not only how AI works, but how it should be questioned, monitored, and governed.</p>
<p>For policymakers and regulators, the framework suggests a shift from static pre-approval criteria toward evidence of real-world performance and governance maturity. Trustworthiness should be evaluated longitudinally, with attention to how AI systems behave under changing conditions.</p>
<p>Importantly, trustworthy AI should be framed not as a constraint on innovation but as an enabler of scale. Systems that fail transparently and are governed responsibly are more likely to earn sustained trust and adoption.</p>
</sec>
<sec id="s7" sec-type="discussion"><label>7</label><title>Discussion</title>
<p>This paper advances a socio-technical framework for operationalizing trustworthy artificial intelligence (AI) in healthcare by shifting emphasis from algorithmic properties to system-level behavior within clinical and operational workflows. While existing guidelines articulate important ethical and technical principles, their translation into practice remains inconsistent. The framework proposed here addresses this gap by identifying operational mechanisms through which trustworthiness can be enacted, observed, and sustained over time.</p>
<p>A central contribution of this work is the explicit reframing of trustworthiness as an emergent property rather than a static attribute. This perspective aligns with empirical observations that AI systems may perform well in development environments yet fail to generate value or even introduce risk; once deployed in real-world settings. By foregrounding workflow integration, decision authority, and governance, the framework highlights dimensions of AI performance that are often invisible in traditional validation studies but critical to real-world success.</p>
<sec id="s7a"><label>7.1</label><title>Implications for clinical practice</title>
<p>For clinicians, the framework underscores the importance of clarity in how AI systems intersect with professional judgment. Ambiguity regarding whether AI outputs are advisory, suggestive, or prescriptive has been shown to undermine trust and contribute to both over-reliance and disengagement. Decision-centered design and explicit human-AI role delineation help mitigate these risks by aligning AI recommendations with clinicians&#x2019; mental models and accountability structures.</p>
<p>The emphasis on failure visibility is particularly salient in clinical contexts, where uncertainty is inherent and stakes are high. Systems that explicitly communicate confidence and limitations may initially appear less decisive, yet they are more likely to foster durable trust. Importantly, such transparency supports appropriate skepticism rather than blind reliance, reinforcing professional autonomy rather than displacing it.</p>
</sec>
<sec id="s7b"><label>7.2</label><title>Implications for operational and administrative AI</title>
<p>Operational AI systems such as those used for staffing, scheduling, capacity management, and resource allocation present distinct trust challenges that are often underappreciated. Unlike clinical decision support tools, operational systems frequently exert indirect influence on patient care, making their effects less visible and their failures harder to detect. The framework&#x0027;s emphasis on embedded governance and continuous evaluation is therefore particularly relevant in operational domains.</p>
<p>By extending trustworthy AI principles beyond bedside decision-making, this work contributes to a more comprehensive understanding of how AI shapes healthcare delivery at scale. Operational trustworthiness requires not only technical robustness but also transparency, contestability, and alignment with organizational values, particularly with respect to equity and workforce well-being.</p>
</sec>
<sec id="s7c"><label>7.3</label><title>Governance, accountability, and organizational maturity</title>
<p>A recurring theme in this framework is the centrality of governance. Trustworthy AI cannot be sustained through <italic>ad hoc</italic> oversight or <italic>post hoc</italic> audits alone. Instead, it requires clearly defined ownership, escalation pathways, and authority to intervene when systems underperform or behave unexpectedly. These requirements imply a level of organizational maturity that many health systems are still developing.</p>
<p>From this perspective, trustworthiness is as much an institutional capability as it is a technical one. Health systems that lack data governance infrastructure, interdisciplinary oversight committees, or mechanisms for continuous monitoring may struggle to operationalize trustworthy AI, regardless of model quality. Conversely, organizations that invest in governance capacity may be better positioned to deploy AI safely and at scale.</p>
</sec>
<sec id="s7d"><label>7.4</label><title>Equity and unintended consequences</title>
<p>Equity considerations cut across all dimensions of trustworthy AI. While fairness is often framed as a statistical property of models, inequities may also arise from workflow design, differential access to AI-enabled services, or uneven adoption across clinical settings. The framework&#x0027;s focus on system-level evaluation encourages attention to these downstream effects.</p>
<p>Continuous monitoring of differential outcomes across demographic and socioeconomic groups is therefore essential. Importantly, such monitoring should not be limited to clinical outcomes alone but should also encompass operational consequences, such as differential wait times, staffing burdens, or access to specialized services. By embedding equity considerations into routine evaluation, health systems can move beyond compliance toward proactive risk mitigation.</p>
</sec>
<sec id="s7e"><label>7.5</label><title>Limitations and future research</title>
<p>This work is conceptual in nature and does not present empirical validation of the proposed framework. While the framework is grounded in existing literature and consensus guidance, its practical utility will depend on context-specific implementation and empirical testing. Future research should examine how different elements of the framework influence trust, adoption, and outcomes across diverse healthcare settings.</p>
<p>Potential research directions include comparative studies of governance models, evaluations of failure visibility mechanisms, and longitudinal analyses of trust dynamics following AI deployment. Additionally, there is a need for validated metrics that capture workflow-level trustworthiness, complementing existing technical performance measures.</p>
</sec>
<sec id="s7f"><label>7.6</label><title>Conceptual contribution and positioning relative to prior frameworks</title>
<p>The World Health Organization (<xref ref-type="bibr" rid="B1">1</xref>) guidance establishes a foundational ethical and human-rights framework for artificial intelligence in health, emphasizing principles such as transparency, accountability, equity, and protection of human autonomy. However, it intentionally remains high-level and does not prescribe concrete operational mechanisms for implementation, monitoring, or value realization. The FUTURE-AI framework (<xref ref-type="bibr" rid="B8">8</xref>) advances the field by translating trustworthy AI into six core principles and 30 consensus-based best practices across the AI lifecycle, offering substantially more practical guidance for healthcare AI development and deployment. Nevertheless, FUTURE-AI focuses primarily on system-level practices and stops short of defining enterprise governance structures, decision rights, or performance management mechanisms. In contrast, the DG-TAI framework extends both WHO and FUTURE-AI by functioning as an execution-oriented governance operating model that embeds trustworthy AI principles into organizational structures, lifecycle controls, and measurable outcomes. DG-TAI explicitly links ethics and trustworthiness to accountability, continuous monitoring, and value realization, enabling health systems to operationalize trustworthy AI at scale rather than treating it solely as a design-time or compliance consideration.</p>
<p>While numerous frameworks have articulated principles for trustworthy or responsible AI, the DG-TAI framework contributes to the literature in three distinct ways.</p>
<p>First, it repositions trustworthiness as a runtime system property rather than a design-time attribute. Existing frameworks largely emphasize pre-deployment validation and ethical compliance. In contrast, DG-TAI foregrounds post-deployment behavior, recognizing that trust is shaped by how systems perform under real-world conditions, including uncertainty, drift, and organizational change.</p>
<p>Second, the framework elevates decisions as the primary unit of governance, rather than models or predictions. This shift has practical and ethical significance. Decisions are observable, auditable, and attributable, making them more amenable to governance than abstract model properties. By anchoring trustworthiness in decision execution, the framework aligns AI oversight with established clinical and administrative accountability structures.</p>
<p>Third, DG-TAI explicitly extends trustworthy AI considerations to operational workflows, including staffing, scheduling, capacity management, and resource allocation. While prior work has focused predominantly on clinical decision support, operational AI systems often exert broader systemic influence and may generate inequities that are less visible but equally consequential. The framework treats clinical and operational AI symmetrically, applying the same trust, governance, and evaluation principles across both domains.</p>
<p>Importantly, the contribution of this work lies not in introducing new ethical norms, but in integrating existing principles into a coherent operational architecture. This synthesis enables health systems to move from aspirational trustworthiness to implementable, monitorable practice.</p>
</sec>
<sec id="s7g"><label>7.7</label><title>Implications for evaluation and future research</title>
<p>The DG-TAI framework suggests several directions for future empirical research. First, it motivates the development of metrics that assess trustworthiness at the decision and workflow levels, complementing traditional model-centric performance measures. Second, it invites comparative evaluation of governance models, examining how different ownership and escalation structures influence trust, adoption, and outcomes. Third, it highlights the need for longitudinal studies that examine how trust evolves over time as AI systems and organizational contexts change.</p>
<p>By offering a decision-governed lens, the framework provides a foundation for testable hypotheses without presupposing specific technologies or use cases. As such, it is designed to remain applicable across clinical domains, operational functions, and health system contexts.</p>
</sec>
<sec id="s7h"><label>7.8</label><title>Operationalizing DG-TAI in practice: a pilot-to-scale use case</title>
<p>Consider a large, multi-hospital healthcare system seeks to deploy an AI-enabled workforce forecasting and staffing optimization solution to address persistent nurse shortages, reduce reliance on premium labor, and improve patient safety. Applying the DG-TAI framework, the organization begins by establishing a formal governance structure that defines decision rights and accountability across HR, clinical operations, IT, data governance, and compliance. Furthermore, Executive sponsors, model owners, data stewards, and a trust or ethics review function are identified to oversee workforce equity, clinician well-being, and regulatory considerations.</p>
<p>Rather than proceeding directly to enterprise-wide rollout, DG-TAI requires deployment as a controlled pilot within a limited set of units or facilities. During the design and pre-deployment phase, historical staffing, census, acuity, and scheduling data supporting the pilot are subjected to rigorous data quality, representativeness, and bias assessments. The model is evaluated for differential impact across shifts, units, and demographic proxies, and technical risks such as robustness and generalizability are assessed in the context of the pilot environment. Advancement beyond the pilot is contingent on meeting predefined trust and performance thresholds, not solely on predictive accuracy.</p>
<p>During the pilot phase, DG-TAI operationalizes continuous monitoring through a defined set of key performance indicators spanning technical, ethical, and operational dimensions. These include forecast accuracy and drift, override and exception rates, staffing equity indicators, overtime hours, vacancy coverage, and clinician experience proxies. Governance forums review these metrics at regular intervals, enabling rapid identification of unintended consequences and informed decision-making about model refinement.</p>
<p>Only after the pilot demonstrates sustained performance, acceptable equity impacts, and operational value does DG-TAI authorize scaled deployment across additional units or the broader enterprise. Post-scale, monitoring and governance mechanisms remain active, with formal escalation and remediation pathways triggered by threshold breaches or contextual changes. Insights from the pilot and early scale phases are systematically captured and fed back into the organization&#x0027;s AI maturity model, strengthening future implementations. This pilot-to-scale approach illustrates how DG-TAI operationalizes trustworthy AI as a phased, accountable process, bridging ethical intent, real-world execution, and sustained value realization in complex healthcare settings.</p>
</sec>
</sec>
<sec id="s8" sec-type="conclusions"><label>8</label><title>Conclusion</title>
<p>As artificial intelligence becomes increasingly embedded in healthcare delivery, the question of trust has emerged as a defining challenge. This paper argues that trustworthy AI cannot be reduced to a checklist of model attributes or ethical principles. Instead, trustworthiness must be understood as a dynamic, system-level achievement that depends on how AI is integrated into clinical and operational workflows, how responsibilities are allocated, and how failures are anticipated and managed.</p>
<p>By proposing a framework that operationalizes trustworthiness through decision-centered design, human&#x2013;AI role delineation, failure visibility, embedded governance, and continuous evaluation, this work offers a pragmatic pathway from principle to practice. The framework does not seek to replace existing ethical or technical guidance but to complement it by addressing the operational realities of AI deployment in healthcare.</p>
<p>Ultimately, the sustainability of AI in healthcare will depend not on whether systems are merely accurate or innovative, but on whether they are reliable, transparent, and aligned with professional values and organizational goals. Health systems that invest in the socio-technical foundations of trustworthy AI may be better positioned to realize its potential while minimizing unintended harm. In this sense, trustworthiness should be viewed not as a constraint on innovation, but as a prerequisite for scale, resilience, and long-term impact.</p>
<p>Future research should empirically test elements of the framework, examining how different governance models, role delineations, or evaluation strategies influence trust, adoption, and outcomes across settings.</p>
</sec>
</body>
<back>
<sec id="s10" sec-type="author-contributions"><title>Author contributions</title>
<p>KK: Writing &#x2013; review &#x0026; editing, Writing &#x2013; original draft.</p>
</sec>
<sec id="s12" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s13" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="s14" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><mixed-citation publication-type="book"><collab>World Health Organization</collab>. <source>Ethics and Governance of Artificial Intelligence for Health: WHO Guidance</source>. <publisher-loc>Geneva</publisher-loc>: <publisher-name>World Health Organization</publisher-name> (<year>2021</year>). <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://www.who.int/publications/i/item/9789240029200">https://www.who.int/publications/i/item/9789240029200</ext-link></mixed-citation></ref>
<ref id="B2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kelly</surname> <given-names>CJ</given-names></name> <name><surname>Karthikesalingam</surname> <given-names>A</given-names></name> <name><surname>Suleyman</surname> <given-names>M</given-names></name> <name><surname>Corrado</surname> <given-names>G</given-names></name> <name><surname>King</surname> <given-names>D</given-names></name></person-group>. <article-title>Key challenges for delivering clinical impact with artificial intelligence</article-title>. <source>BMC Med</source>. (<year>2019</year>) <volume>17</volume>:<fpage>195</fpage>. <pub-id pub-id-type="doi">10.1186/s12916-019-1426-2</pub-id><pub-id pub-id-type="pmid">31665002</pub-id></mixed-citation></ref>
<ref id="B3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Topol</surname> <given-names>EJ</given-names></name></person-group>. <article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title>. <source>Nat Med</source>. (<year>2019</year>) <volume>25</volume>(<issue>1</issue>):<fpage>44</fpage>&#x2013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.1038/s41591-018-0300-7</pub-id><pub-id pub-id-type="pmid">30617339</pub-id></mixed-citation></ref>
<ref id="B4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alelyani</surname> <given-names>T</given-names></name> <name><surname>Qureshi</surname> <given-names>R</given-names></name> <name><surname>Gill</surname> <given-names>SS</given-names></name> <name><surname>Buyya</surname> <given-names>R</given-names></name> <name><surname>Ranjan</surname> <given-names>R</given-names></name></person-group>. <article-title>A validated framework for responsible artificial intelligence in healthcare</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>:<fpage>25266</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-25266-z</pub-id><pub-id pub-id-type="pmid">40651987</pub-id></mixed-citation></ref>
<ref id="B5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Amann</surname> <given-names>J</given-names></name> <name><surname>Blasimme</surname> <given-names>A</given-names></name> <name><surname>Vayena</surname> <given-names>E</given-names></name> <name><surname>Frey</surname> <given-names>D</given-names></name> <name><surname>Madai</surname> <given-names>VI</given-names></name></person-group>. <article-title>Explainability for artificial intelligence in healthcare: a multidisciplinary perspective</article-title>. <source>BMC Med Inform Decis Mak</source>. (<year>2020</year>) <volume>20</volume>:<fpage>310</fpage>. <pub-id pub-id-type="doi">10.1186/s12911-020-01332-6</pub-id><pub-id pub-id-type="pmid">33256715</pub-id></mixed-citation></ref>
<ref id="B6"><label>6.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Griesinger</surname><given-names>CB</given-names></name> <name><surname>Reina</surname><given-names>V</given-names></name> <name><surname>Panidis</surname><given-names>D</given-names></name> <name><surname>Chassaigne</surname><given-names>H</given-names></name></person-group>. <source>AI Evidence Pathway for Operationalising Trustworthy AI in Health: An Ontology Unfolding Ethical Principles into Translational and Fundamental Concepts</source>. <publisher-loc>Luxembourg</publisher-loc>: <publisher-name>Publications Office of the European Union</publisher-name> (<year>2025</year>). <pub-id pub-id-type="doi">10.2760/8107037</pub-id></mixed-citation></ref>
<ref id="B7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moreno-S&#x00E1;nchez</surname> <given-names>PA</given-names></name> <name><surname>Ser</surname> <given-names>JD</given-names></name> <name><surname>van Gils</surname> <given-names>M</given-names></name> <name><surname>Hernesniemi</surname> <given-names>J</given-names></name></person-group>. <article-title>A design framework for operationalizing trustworthy artificial intelligence in healthcare: requirements, tradeoffs and challenges for its clinical adoption</article-title>. <source>Inf Fusion</source>. (<year>2026</year>). <volume>127</volume>(<issue>Part B</issue>):<fpage>103812</fpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2025.103812</pub-id></mixed-citation></ref>
<ref id="B8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lekadir</surname> <given-names>K</given-names></name> <name><surname>Frangi</surname> <given-names>AF</given-names></name> <name><surname>Porras</surname> <given-names>AR</given-names></name> <name><surname>Glocker</surname> <given-names>B</given-names></name> <name><surname>Cintas</surname> <given-names>C</given-names></name> <name><surname>Langlotz</surname> <given-names>CP</given-names></name><etal/></person-group> <article-title>FUTURE-AI: international consensus guideline for trustworthy and deployable artificial intelligence in healthcare</article-title>. <source>BMJ</source>. (<year>2025</year>) <volume>388</volume>:<fpage>e081554</fpage>. <pub-id pub-id-type="doi">10.1136/bmj-2024-081554</pub-id><pub-id pub-id-type="pmid">39909534</pub-id></mixed-citation></ref></ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2378614/overview">Tom Karl Schaal</ext-link>, West Saxon University of Applied Sciences of Zwickau, Germany</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/988479/overview">Gulzar Alam</ext-link>, Dalian Medical University, China</p></fn>
</fn-group>
</back>
</article>