<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="brief-report" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurol.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Neurology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurol.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1664-2295</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fneur.2026.1767832</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Brief Research Report</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Causal dynamic decision-making for robotic systems in non-Markovian high-difficulty surgery</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Na</surname>
<given-names>Guo</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2618706"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Minghui</surname>
<given-names>Tan</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3390563"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tiantian</surname>
<given-names>Li</given-names>
</name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yang</surname>
<given-names>Liu</given-names>
</name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Qinjian</surname>
<given-names>Zhang</given-names>
</name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1964954"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yuanxin</surname>
<given-names>Li</given-names>
</name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tianlei</surname>
<given-names>Xu</given-names>
</name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1415857"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Fuchun</surname>
<given-names>Sun</given-names>
</name>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/639726"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>School of Computer and Communication Engineering, University of Science and Technology Beijing</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>College of Information and Electrical Engineering, China Agricultural University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>School of Mechanical and Electrical Engineering, Beijing Information Science and Technology University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<aff id="aff4"><label>4</label><institution>Beijing Tsinghua Changgung Hospital, Tsinghua University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<aff id="aff5"><label>5</label><institution>Institute of Intelligent Healthcare, Tsinghua University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<aff id="aff6"><label>6</label><institution>Department of Computer Science and Technology, Tsinghua University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Guo Na, <email xlink:href="mailto:guona2023@ustb.edu.cn">guona2023@ustb.edu.cn</email>; Sun Fuchun, <email xlink:href="mailto:fcsun@mail.tsinghua.edu.cn">fcsun@mail.tsinghua.edu.cn</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-20">
<day>20</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>17</volume>
<elocation-id>1767832</elocation-id>
<history>
<date date-type="received">
<day>15</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>06</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Na, Minghui, Tiantian, Yang, Qinjian, Yuanxin, Tianlei and Fuchun.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Na, Minghui, Tiantian, Yang, Qinjian, Yuanxin, Tianlei and Fuchun</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-20">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Markov assumption-based surgical decision models cannot account for the time-varying, irregular effects of high-risk intraoperative anomalies such as sudden hemorrhage or inadvertent instrument loss, making them inadequate for specialized procedures like neurosurgery and spinal interventions. To overcome the non-Markovian limitations of conventional surgical process modeling, this study develops a causal modeling framework based on Vector Autoregression (VAR) and Granger causality analysis. The framework constructs a causal chain (original gesture <inline-formula>
<mml:math id="M1">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> &#x2192; abnormal event <inline-formula>
<mml:math id="M2">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> &#x2192; recovery action <inline-formula>
<mml:math id="M3">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>) to enable intelligent response and adaptive decision-making. Validation was performed on a large-scale synthetic dataset containing 10,000 samples (including anomaly, positive, and negative cases), and evaluated using accuracy, F1-score, and recall metrics. Experimental results show the proposed method achieves 95.60% accuracy in causal inference, maintaining stability at 10,000 samples with an F1 score of 95.77%. Notably, recall (95.88%) slightly exceeds precision (95.34%), reflecting the clinical principle of prioritizing safety. The framework effectively captures non-Markovian temporal correlations induced by abnormal events, overcoming key limitations of traditional approaches. Its design is not procedure-specific, providing a versatile and generalizable pathway for enhancing autonomous decision-making in surgical robots across diverse clinical applications.</p>
</abstract>
<kwd-group>
<kwd>causal inference</kwd>
<kwd>dynamic decision-making</kwd>
<kwd>Granger causality</kwd>
<kwd>non-Markov processes</kwd>
<kwd>surgical robotics</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the Youth Faculty Interdisciplinary Research Project of the University of Science and Technology Beijing (Grant No. FRF-IDRY-24-021).</funding-statement>
</funding-group>
<counts>
<fig-count count="3"/>
<table-count count="1"/>
<equation-count count="8"/>
<ref-count count="27"/>
<page-count count="9"/>
<word-count count="6618"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Artificial Intelligence in Neurology</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>Decision-making is the core prerequisite for surgical robots to achieve autonomous operation. Recent advances in embodied intelligence (<xref ref-type="bibr" rid="ref1">1</xref>) have propelled the field beyond traditional leader-follower control toward systems capable of environmental perception and autonomous decision-making. The early STAR robot, developed by the Johns Hopkins University team, utilized near-infrared fluorescence imaging for millimeter-level tissue perfusion identification (<xref ref-type="bibr" rid="ref2">2</xref>, <xref ref-type="bibr" rid="ref3">3</xref>); while the more recent SRT-H system employs miniature cameras and precision control to achieve submillimeter accuracy in vascular clamping (<xref ref-type="bibr" rid="ref4">4</xref>, <xref ref-type="bibr" rid="ref5">5</xref>). And frameworks like VPPV have demonstrated &#x201C;zero-shot transfer&#x201D; of operational strategies for multi-task decision-making (<xref ref-type="bibr" rid="ref1">1</xref>).</p>
<p>However, existing surgical decision-making methods predominantly focus on conventional laparoscopic procedures such as cholecystectomy, with a primary emphasis on automating localized manipulations such as suturing and knot-tying. Systematic research into task-level autonomous decision-making for complex, high-risk surgical specialties including neurosurgery and spinal surgery remains notably scarce. Task-level surgical video data are often confounded by numerous variables such as instrument slippage (as visually illustrated for normal vs. abnormal scenarios <xref ref-type="fig" rid="fig1">Figures 1b</xref>,<xref ref-type="fig" rid="fig1">c</xref>) (<xref ref-type="bibr" rid="ref6">6</xref>) and tissue deformation (<xref ref-type="bibr" rid="ref7">7</xref>), which exhibit distinct non-Markovian temporal characteristics. This renders conventional temporal models inadequate for capturing long-range causal dependencies, thereby constraining the adaptability and reliability of current systems in complex surgical settings: Deterministic methods such as behavior trees (<xref ref-type="bibr" rid="ref8">8</xref>, <xref ref-type="bibr" rid="ref9">9</xref>), behavior networks (<xref ref-type="bibr" rid="ref10">10</xref>), or task priority graphs (<xref ref-type="bibr" rid="ref11">11</xref>) are too rigid to adapt to unexpected events, while probabilistic models based on Markov Decision Processes (MDPs) (<xref ref-type="bibr" rid="ref12">12</xref>, <xref ref-type="bibr" rid="ref13">13</xref>) fail to capture long-range causal relationships. Although the &#x201C;Transformer-based planning + imitation/reinforcement learning-based control&#x201D; paradigm (<xref ref-type="bibr" rid="ref4">4</xref>, <xref ref-type="bibr" rid="ref14">14</xref>) is mainstream, traditional reinforcement learning often optimizes statistical correlations (<xref ref-type="bibr" rid="ref15">15</xref>), leading to performance drops during Sim2Real transfer or sensor failure.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Clinical value and schematic representation of the causal dynamic decision-making framework for robotic surgery in non-Markovian high-difficulty scenarios. <bold>(a)</bold> Surgical application value of the causal dynamic decision-making framework; <bold>(b)</bold> Surgical suturing process under normal conditions; <bold>(c)</bold> Surgical suturing procedures in abnormal situations (such as needle drops); <bold>(d)</bold> Syntax diagram for surgical tasks (<xref ref-type="bibr" rid="ref26">26</xref>); <bold>(e)</bold> Execution errors under abnormal circumstances (<xref ref-type="bibr" rid="ref27">27</xref>).</p>
</caption>
<graphic xlink:href="fneur-17-1767832-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Figure composed of five panels labeled a through e. Panel a shows a circular schematic centered on a robotic causal dynamic decision-making framework, surrounded by icons representing surgical benefits such as improved precision, reduced complications, and standardized procedures. Panels b, c, and d show flowcharts of gesture transitions, with increasing complexity from b to d and various nodes labeled with gesture codes such as G1, G6, and Knot. Panel e is a vertical bar chart comparing numbers of gestures with different error types (color-coded) for six gesture categories, labeled G1 through Knot, with a legend identifying error types.</alt-text>
</graphic>
</fig>
<p>Causal reasoning, which transforms correlational strategies into causal ones (<xref ref-type="bibr" rid="ref16">16</xref>, <xref ref-type="bibr" rid="ref17">17</xref>), provides new approaches to address partial observability (POMDP) and confounding factors in robotic perception. Work on Local Causal Models (LCMs) (<xref ref-type="bibr" rid="ref18">18</xref>) and structural causal models (<xref ref-type="bibr" rid="ref19">19</xref>, <xref ref-type="bibr" rid="ref20">20</xref>) has shown promise in improving sample efficiency and enabling zero-shot transfer in generic robotic tasks. However, these general causal inference methods are ill-suited for surgical scenarios: LCMs rely on locally factored dynamics and stationary environment assumptions, making them incapable of addressing abrupt, non-stationary abnormal events intraoperatively; structural causal models prioritize static causal graph construction, which hinders their ability to capture dynamic, long-term temporal dependencies arising from intraoperative disruptions. Consequently, they cannot be readily applied to surgical scenarios characterized by non-stationary state spaces and unforeseen anomalies. Surgical applications require models that enable safe, priority-driven dynamic causal inference under clinical logic constraints. Therefore, a solution specifically designed for the unique needs of surgical scenarios is urgently needed.</p>
<p>Inspired by work on local confounding detection (<xref ref-type="bibr" rid="ref21">21</xref>) and adaptive methods (<xref ref-type="bibr" rid="ref22">22</xref>) in non-stationary environments, this study models surgical robot decision-making as a sequential reasoning problem constrained by clinical logic. We propose a causal dynamic reasoning framework tailored to non-Markovian surgical environments, employing Granger causality inference (<xref ref-type="bibr" rid="ref23">23</xref>, <xref ref-type="bibr" rid="ref24">24</xref>) as the core mechanism. By integrating it with the clinical logic of &#x201C;surgical gesture &#x2013; abnormal event &#x2013; recovery action&#x201D; and optimizing the lag order of the VAR model (<xref ref-type="bibr" rid="ref25">25</xref>), the framework achieves safe, priority-based detection of dynamic abnormal events and supports autonomous decision-making.</p>
<p>This approach yields a framework with inherent generalization capabilities. By integrating the core Granger causality mechanism with transferable clinical logic (as schematically depicted in <xref ref-type="fig" rid="fig1">Figure 1a</xref>), this framework is not only directly applicable to diverse surgical scenarios such as neurosurgery and spinal procedures but also effectively addresses the non-Markovian dynamics and unexpected anomalies common in clinical practice.</p>
</sec>
<sec sec-type="methods" id="sec2">
<label>2</label>
<title>Method</title>
<sec id="sec3">
<label>2.1</label>
<title>Framework overview and causal analysis</title>
<sec id="sec4">
<label>2.1.1</label>
<title>Dynamic causal characteristic analysis of surgical adverse events</title>
<p>Surgical abnormal events (such as instrument drops or tissue damage) interrupt predefined operational workflows (see the surgical task syntax diagram in <xref ref-type="fig" rid="fig1">Figure 1d</xref>), and their dynamic adjustment strategies rely on accurate understanding of the &#x201C;preceding actions - anomaly type - response operations&#x201D; causal chain. Such anomalous events exhibit pronounced temporal long-range dependencies in non-Markovian environments and may be categorized into two primary types: (1) Procedural errors, manifested as omitted steps or sequential errors; (2) Execution errors, including positioning deviations and instrument loss of control, refer to failures in single-step operations. In minimally invasive surgical environments, such execution errors are particularly common due to constrained visual fields and anatomical variations, directly impacting operational effectiveness and procedural continuity.</p>
<p>Taking the suturing task as an example (as shown in <xref ref-type="fig" rid="fig1">Figure 1e</xref>), the handling logic for execution errors exhibits significant causal correlation characteristics. For simplification, it is assumed that each type of error occurs only once and can be successfully addressed:<list list-type="bullet">
<list-item>
<p>Needle tip positioning abnormality (G2): When positioning inaccuracy occurs, the system must repeat the positioning operation (G2) until successful before proceeding with tissue puncture (G3);</p>
</list-item>
<list-item>
<p>Equipment Fall Incident (E1): Should instrument drop be detected during the left-hand suture pulling (G6) process, the retrieval operation shall take precedence, thereafter, return to G6 to continue the subsequent procedure;</p>
</list-item>
<list-item>
<p>Equipment Out of Bounds Incident (E3): When the right hand is tightening the suture (G9), the instrument moves beyond the field of view, it is necessary to first adjust the instrument&#x2019;s position, then decide whether to continue G9 or switch to other operations based on the actual situation.</p>
</list-item>
</list></p>
<p>Notably, these recovery pathways are dynamically determined by the interaction between anomaly type and the current operational context, emphasizing the requirement for autonomous systems to conduct real-time, context-aware causal reasoning in unstructured surgical environment.</p>
</sec>
<sec id="sec5">
<label>2.1.2</label>
<title>Dynamic inference framework for surgical procedures based on Granger causality testing</title>
<p>To establish a dynamic decision-making mechanism suitable for the non-Markovian surgical environments described above, this paper proposes a three-stage causal dynamic reasoning framework based on Granger causality testing. The overall implementation pathway of this framework is illustrated in <xref ref-type="fig" rid="fig2">Figure 2</xref>, which outlines the process from surgical video structuring to Granger causality verification. This framework is designed to identify causal relationships between surgical phases and actions, providing a basis for the dynamic decision-making of surgical robots. It comprises the following core components:<list list-type="simple">
<list-item>
<p>1. Structured representation of surgical video streams</p>
</list-item>
</list></p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Implementation pipeline of the causal dynamic decision-making framework for surgical robots. Firstly, the key features of the surgical video are extracted to construct a characterization framework, and then the VAR model is used for temporal correlation modeling, and finally the causal timing is verified by Granger causality test, so as to accurately identify the causal relationship and provide theoretical support for the autonomous decision-making of surgical robots.</p>
</caption>
<graphic xlink:href="fneur-17-1767832-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart illustrating a three-step analytical process: Step 1 organizes raw surgical video streams into structured temporal sequence data with phase and action dimensions; Step 2 builds a vector autoregression model using these sequences for temporal correlation; Step 3 applies Granger causality tests to model results and sequence data, comparing restricted and unrestricted models using F-statistics to identify causal relationships and mission interruptions.</alt-text>
</graphic>
</fig>
<p>Using deep learning models, key surgical phases and action markers are extracted from surgical videos to construct a structured representation framework. This approach addresses the challenge of modeling phase-action associations in complex scenarios, providing a robust data foundation for subsequent causal analysis.<list list-type="simple">
<list-item>
<p>2. Construction of vector autoregressive (VAR) models to capture temporal dynamics</p>
</list-item>
</list></p>
<p>A Vector Autoregression (VAR) model is applied to capture temporal correlations. Compared to traditional Markov models, the VAR model more effectively captures long-range dynamic dependencies inherent in non-Markovian surgical processes, enabling more comprehensive temporal analysis.<list list-type="simple">
<list-item>
<p>3. Dynamic decision based on granger causal testing</p>
</list-item>
</list></p>
<p>Granger causality testing is employed to validate genuine causal relationships by evaluating predictive capability differences. This method distinguishes between mere temporal correlations and true causal associations, providing a reliable autonomous decision-making strategy for surgical robots.</p>
</sec>
</sec>
<sec id="sec6">
<label>2.2</label>
<title>Structured characterization of surgical procedures</title>
<p>This study employs a &#x201C;phase-action&#x201D; coding method to convert continuous surgical videos into structured temporal sequences: the phase dimension is defined by the continuous execution of standard operational states together with sudden abnormal events; while the action dimension represents the executable operations of the robotic system at the current moment. This integrated approach aligns with surgical cognitive patterns, clearly distinguishing routine operations from abnormal events.</p>
<p>Let there be <inline-formula>
<mml:math id="M4">
<mml:mi>N</mml:mi>
</mml:math>
</inline-formula> standard operational states <inline-formula>
<mml:math id="M5">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, M random events <inline-formula>
<mml:math id="M6">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, W executable operations <inline-formula>
<mml:math id="M7">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, which are defined as follows:</p>
<p>At any time <italic>t</italic>:</p>
<p><inline-formula>
<mml:math id="M8">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2208;</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">}</mml:mo>
</mml:math>
</inline-formula> denotes whether the <italic>i</italic>-th standard operational state is active. If <inline-formula>
<mml:math id="M9">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:math>
</inline-formula>=&#x2018;<italic>Needle Holding</italic>&#x2019;, <inline-formula>
<mml:math id="M10">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:math>
</inline-formula> indicates that the step &#x2018;<italic>Needle Holding</italic>&#x2019; is being performed, and <inline-formula>
<mml:math id="M11">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:math>
</inline-formula> indicates it is not.</p>
<p><inline-formula>
<mml:math id="M12">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2208;</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">}</mml:mo>
</mml:math>
</inline-formula> denotes whether the <italic>j</italic>-th random event occurs. If <inline-formula>
<mml:math id="M13">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:math>
</inline-formula>= &#x2018;<italic>Needle Dropping</italic>&#x2019;, <inline-formula>
<mml:math id="M14">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:math>
</inline-formula> indicates that the abnormal event &#x2018;<italic>Needle Dropping</italic>&#x2019; has occurred, and <inline-formula>
<mml:math id="M15">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:math>
</inline-formula> indicates it has not.</p>
<p><inline-formula>
<mml:math id="M16">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>W</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2208;</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">}</mml:mo>
</mml:math>
</inline-formula> denotes whether the <italic>k</italic>-th executable action is triggered or recommended by the system. If <inline-formula>
<mml:math id="M17">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:math>
</inline-formula>= &#x2018;<italic>Needle Retrieval</italic>&#x2019;, <inline-formula>
<mml:math id="M18">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:math>
</inline-formula> indicates that the action is triggered, and <inline-formula>
<mml:math id="M19">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
</mml:math>
</inline-formula> indicates it is not.</p>
<p>Accordingly, the surgical phase and action at time <italic>t</italic> can be described as:<disp-formula id="E1">
<mml:math id="M20">
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">[</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>N</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>M</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:math>
</disp-formula><disp-formula id="E2">
<mml:math id="M21">
<mml:mi>Z</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">[</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>W</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msup>
</mml:math>
</disp-formula></p>
<p>Where, <inline-formula>
<mml:math id="M22">
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> represents the stage variable at time <inline-formula>
<mml:math id="M23">
<mml:mi>t</mml:mi>
</mml:math>
</inline-formula>, integrating both standard operational states and random events. <inline-formula>
<mml:math id="M24">
<mml:mi>Z</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> denotes the action variable at time <inline-formula>
<mml:math id="M25">
<mml:mi>t</mml:mi>
</mml:math>
</inline-formula>, comprising <inline-formula>
<mml:math id="M26">
<mml:mi>W</mml:mi>
</mml:math>
</inline-formula> executable operational states <inline-formula>
<mml:math id="M27">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>W</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>.</p>
</sec>
<sec id="sec7">
<label>2.3</label>
<title>Construction of vector autoregression (VAR) models</title>
<p>The core objective of this study is to learn a dynamic decision function from surgical video time-series data that can accurately describe the mapping from historical phase information to the current execution action. This function represents the conditional probability distribution of the system taking a specific action given the historical surgical context, which can be generally expressed as:<disp-formula id="E3">
<mml:math id="M28">
<mml:mi>P</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>Z</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2223;</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</disp-formula></p>
<p>Where <inline-formula>
<mml:math id="M29">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> denotes the length of the historical time window influencing the current decision. This conditional probability distribution essentially defines a state transition process, mapping the sequence of past <inline-formula>
<mml:math id="M30">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> phase states <inline-formula>
<mml:math id="M31">
<mml:mo stretchy="true">{</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">}</mml:mo>
</mml:math>
</inline-formula> to the action variable <inline-formula>
<mml:math id="M32">
<mml:mi>Z</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> at time t.</p>
<p>To characterize dynamic interactions within the &#x2018;phase&#x2019; variable <inline-formula>
<mml:math id="M33">
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> and its potential influence on the &#x2018;action&#x2019; variable <inline-formula>
<mml:math id="M34">
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>, this study employs a Vector Autoregressive (VAR) model as the core temporal modeling tool. The VAR model enables simultaneous estimation of the interdependencies among multiple endogenous variables and captures the persistent effects of historical states on the current system through the introduction of lag terms. This provides an ideal structural foundation for subsequently identifying the Granger causal relationships between the &#x201C;phase&#x201D; and &#x201C;action&#x201D; variables.</p>
<p>The VAR-based phase-action model can be expressed as follows:<disp-formula id="E4">
<mml:math id="M35">
<mml:mi>Z</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mi>f</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>;</mml:mo>
<mml:mi>&#x0398;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:mi>&#x03B5;</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</disp-formula></p>
<p>Where, <inline-formula>
<mml:math id="M36">
<mml:mi>f</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mo>&#x22C5;</mml:mo>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> is a linear function, <inline-formula>
<mml:math id="M37">
<mml:mi>&#x0398;</mml:mi>
</mml:math>
</inline-formula> is the parameter matrix to be estimated, and <inline-formula>
<mml:math id="M38">
<mml:mi>&#x03B5;</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> is a random disturbance term. This model represents action decisions as a linear combination of historical phase variables, providing interpretable parameter estimates for subsequent analysis.</p>
<p>More specifically, each action variable (in Section 2.2) can be written as:<disp-formula id="E5">
<mml:math id="M39">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>M</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mi>&#x03B4;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>&#x03B5;</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</disp-formula></p>
<p>Where, <inline-formula>
<mml:math id="M40">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mspace width="0.33em"/>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>W</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> denotes the k-th action at time <inline-formula>
<mml:math id="M41">
<mml:mi>t</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M42">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> is the lag order, representing the length of historical information influencing the current action, <inline-formula>
<mml:math id="M43">
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> represents the influence of the i-th state at lag <inline-formula>
<mml:math id="M44">
<mml:mi>q</mml:mi>
</mml:math>
</inline-formula> on action <inline-formula>
<mml:math id="M45">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M46">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> <inline-formula>
<mml:math id="M47">
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>is the value of the i-th phase variable at past time <inline-formula>
<mml:math id="M48">
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M49">
<mml:msubsup>
<mml:mi>&#x03B4;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> represents the influence of the j-th random event at lag <inline-formula>
<mml:math id="M50">
<mml:mi>q</mml:mi>
</mml:math>
</inline-formula> on action <inline-formula>
<mml:math id="M51">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M52">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> <inline-formula>
<mml:math id="M53">
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>is the value of the j-th random event at past time <inline-formula>
<mml:math id="M54">
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M55">
<mml:msub>
<mml:mi>&#x03B5;</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> is a random noise term at time <inline-formula>
<mml:math id="M56">
<mml:mi>t</mml:mi>
</mml:math>
</inline-formula>.</p>
<p>This formulation captures both the influence of historical phase states and abnormal events on current actions, providing a clear, interpretable framework for modeling surgical decision-making. These actions may be triggered by abnormal events (e.g., &#x201C;Bleeding&#x201D; triggering &#x201C;Electrocoagulation&#x201D;) or arise naturally from routine surgical logic (e.g., the &#x201C;Suturing&#x201D; state leading to &#x201C;Knot Tying&#x201D;). Granger causality, which will be introduced in Section 2.4, is later incorporated into the time-series modeling framework to identify and remove spurious correlations, enabling the construction of a decision model with explicit causal interpretability and strong generalization.</p>
</sec>
<sec id="sec8">
<label>2.4</label>
<title>Dynamic decision based on Granger causal testing</title>
<p>Building upon the VAR-based phase-action model introduced in Section 2.3, Granger causality tests are employed to identify temporal causal relationships between surgical phases and robotic actions. Herein, the &#x201C;phase component&#x201D; refers to <inline-formula>
<mml:math id="M57">
<mml:msub>
<mml:mi>V</mml:mi>
<mml:mi mathvariant="italic">ij</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>, which denotes the standard operational state variable <inline-formula>
<mml:math id="M58">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> or individual random event variable <inline-formula>
<mml:math id="M59">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> extracted from the integrated stage variable <inline-formula>
<mml:math id="M60">
<mml:mi>V</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> (defined in Section 2.2). In a non-Markovian surgical environment, if incorporating historical information from past phase variables <inline-formula>
<mml:math id="M61">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> and random events <inline-formula>
<mml:math id="M62">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:mo>&#x22EF;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> significantly improves the prediction of the current action <inline-formula>
<mml:math id="M63">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>, these variables are considered Granger causes of <inline-formula>
<mml:math id="M64">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula>. The testing process, detailed in Algorithm 1, is performed independently for each potential (phase component, action component) causal pair.
<statement id="algo1" content-type="algorithm">
<label>Algorithm 1</label>
<p>Element-wise Granger causality testing in temporal sequences.<graphic xlink:href="fneur-17-1767832-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart outlines a statistical testing procedure using surgical video data, detailing inputs, output, and stepwise methods for encoding observational data, building restricted and unrestricted models, calculating sums of squares, computing an F statistic, comparing to a critical value, and determining the presence or absence of causal links.</alt-text>
</graphic></p>
</statement>
<list list-type="simple">
<list-item>
<p>(1) Restricted model: Uses only the historical data of action <inline-formula>
<mml:math id="M65">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> itself for prediction, disregarding the influence of historical phase information, which can be expressed as follows:<disp-formula id="E6">
<mml:math id="M66">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mover accent="true">
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">&#x00AF;</mml:mo>
</mml:mover>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mover accent="true">
<mml:mi>&#x03B2;</mml:mi>
<mml:mo stretchy="true">&#x00AF;</mml:mo>
</mml:mover>
<mml:mi>k</mml:mi>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>&#x03B5;</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</disp-formula></p>
</list-item>
</list></p>
<p>Here, <inline-formula>
<mml:math id="M67">
<mml:msub>
<mml:mover accent="true">
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">&#x00AF;</mml:mo>
</mml:mover>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:math>
</inline-formula> denotes the constant term,<inline-formula>
<mml:math id="M68">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> represents the lag order, <inline-formula>
<mml:math id="M69">
<mml:msubsup>
<mml:mover accent="true">
<mml:mi>&#x03B2;</mml:mi>
<mml:mo stretchy="true">&#x00AF;</mml:mo>
</mml:mover>
<mml:mi>k</mml:mi>
<mml:mi>q</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> signifies the autoregressive coefficient of <inline-formula>
<mml:math id="M70">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> at lag q.<list list-type="simple">
<list-item>
<p>(2) Unrestricted model: Incorporates lagged terms of all historical phase variables to explicitly capture the respective causal influences of past states and events on the current action.<disp-formula id="E7">
<mml:math id="M71">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>&#x03B1;</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mi>&#x03B2;</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mo>+</mml:mo>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>p</mml:mi>
</mml:munderover>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>M</mml:mi>
</mml:munderover>
<mml:msubsup>
<mml:mi>&#x03B4;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>&#x03B5;</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
</disp-formula></p>
</list-item>
</list></p>
<p>The coefficient <inline-formula>
<mml:math id="M72">
<mml:msubsup>
<mml:mi>&#x03B3;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M73">
<mml:msubsup>
<mml:mi>&#x03B4;</mml:mi>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>q</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> represent the average effect of the state <inline-formula>
<mml:math id="M74">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M75">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>q</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> has on the action <inline-formula>
<mml:math id="M76">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> at lag q, e.g., the driving effect of a G6 operation on the picking up needle action, or the triggering effect of dropping needle on the picking up needle action.<list list-type="simple">
<list-item>
<p>(3) Causal significance <inline-formula>
<mml:math id="M77">
<mml:mi>F</mml:mi>
</mml:math>
</inline-formula></p>
</list-item>
</list></p>
<p>The likelihood ratio statistic is used to assess whether the predictive improvement of the unrestricted model relative to the restricted model is statistically significant, which can be expressed as follows:<disp-formula id="E8">
<mml:math id="M78">
<mml:mi>F</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi mathvariant="italic">ur</mml:mi>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>/</mml:mo>
<mml:mi>Q</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi mathvariant="italic">ur</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula></p>
<p>Here, <inline-formula>
<mml:math id="M79">
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M80">
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi mathvariant="italic">ur</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> represent the sums of squared residuals of the restricted and unrestricted models, respectively; <inline-formula>
<mml:math id="M81">
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> reflects the prediction error without causal variables, while <inline-formula>
<mml:math id="M82">
<mml:mi mathvariant="italic">SS</mml:mi>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mi mathvariant="italic">ur</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> reflects the prediction error after introducing causal variables. <inline-formula>
<mml:math id="M83">
<mml:mi>n</mml:mi>
</mml:math>
</inline-formula> denotes the sample size, <inline-formula>
<mml:math id="M84">
<mml:mi>Q</mml:mi>
</mml:math>
</inline-formula> represents the total number of lagged terms of the causal variables, <inline-formula>
<mml:math id="M85">
<mml:mi>p</mml:mi>
</mml:math>
</inline-formula> is the lag order (Q&#x202F;=&#x202F;2p).</p>
<p>Based on the results of the Granger causality test, a dynamic correction mechanism for surgical procedures is established. When the system detects an abnormal event, it automatically triggers the corresponding corrective action. For example, &#x201C;pick up needle + re-execute G6&#x201D; can be determined by integrating its causal association &#x201C;dropped needle&#x201D; with the preceding state &#x201C;G6 operation.&#x201D; Unlike conventional predefined workflows, this mechanism generates adaptive response strategies through data-driven causal inference, enabling process reconstruction in non-sequential or unexpected surgical scenarios. The corresponding pseudocode is presented as follows:</p>
<statement id="algo2" content-type="algorithm">
<label>Algorithm 2</label>
<p>Dynamic correction mechanism.
<graphic xlink:href="fneur-17-1767832-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart diagram describes a process for generating recovery action commands based on surgical video streams, VAR models, and historical data, with steps for state retrieval, abnormal event identification, F-statistic computation, Granger-causality check, and differentiated outputs for event- or state-triggered corrections.</alt-text>
</graphic></p>
</statement>
</sec>
</sec>
<sec sec-type="results" id="sec9">
<label>3</label>
<title>Results</title>
<p>This section employs synthetic data to simulate clinically abnormal scenarios, thereby quantitatively evaluating the causal reasoning framework&#x2019;s capability to dynamically model the sequential logic of surgical procedures.</p>
<sec id="sec10">
<label>3.1</label>
<title>Dataset</title>
<p>Due to the scarcity of annotated data on abnormal events during actual surgical procedures, coupled with ethical and privacy constraints, this study employs synthetic data to validate a surgical process modeling approach based on Granger causality testing. The dataset design focuses on the causal relationship between &#x201C;abnormal events&#x201D; and &#x201C;countermeasures,&#x201D; comprising three core temporal variables:</p>
<p><inline-formula>
<mml:math id="M86">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>(original gesture, e.g., G6 &#x201C;left hand pulling thread&#x201D;).</p>
<p><inline-formula>
<mml:math id="M87">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> (abnormal event, e.g., &#x201C;dropped needle&#x201D;).</p>
<p><inline-formula>
<mml:math id="M88">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>(recovery action, e.g., &#x201C;pick up needle + re-execute G6&#x201D;).</p>
<p>A variable value of 1 indicates that an event has occurred, while 0 indicates that it has not.</p>
<p>This study employs both positive and negative examples as samples. Positive examples are constructed based on clinical logic, establishing explicit causal relationships such as &#x201C;needle drop &#x2192; needle retrieval&#x201D; and &#x201C;inaccurate positioning &#x2192; repositioning.&#x201D; This ensures that <inline-formula>
<mml:math id="M89">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> is driven by historical data from <inline-formula>
<mml:math id="M90">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> and <inline-formula>
<mml:math id="M91">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, thereby simulating the anomaly handling procedures encountered in actual surgical procedures. Negative examples are generated by randomly producing sequences of unrelated variables (<inline-formula>
<mml:math id="M92">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M93">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M94">
<mml:msub>
<mml:mi>Z</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>)to eliminate temporal correlation interference, thereby validating the model&#x2019;s discriminative capability in scenarios lacking genuine causal relationships.</p>
<p>To enhance the clinical plausibility of synthetic data, three senior consultants reviewed all predefined causal relationships, yielding a Kappa consistency score of 0.92. Furthermore, to better simulate real surgical environments, we introduced clinical noise in 20% of samples (specifically simulating gesture recognition bias caused by tissue occlusion) to bolster both data authenticity and model robustness. In this study, two datasets of different scales were constructed: one containing 5,000 samples (2,500 positive and 2,500 negative), and the other containing 10,000 samples (5,000 positive and 5,000 negative). The goal is to evaluate the model&#x2019;s stability and generalization ability under varying data scales.</p>
</sec>
<sec id="sec11">
<label>3.2</label>
<title>Experimental setup</title>
<p>When constructing the VAR model, the selection of the lag order <italic>p</italic> is crucial to model performance. To effectively capture the influence of prior actions on the current recovery operation while avoiding excessive noise, <italic>p</italic> must be properly determined. Considering that intraoperative abnormal events typically span about two action steps from occurrence to response (for example, after a &#x201C;needle drop,&#x201D; the surgeon must first &#x201C;pick up the needle&#x201D; and then &#x201C;rethread&#x201D;), this study preliminarily sets the lag order to <italic>p</italic>&#x202F;=&#x202F;2.</p>
<p>To further optimize the selection of the lag order and improve both reproducibility and transparency, the Akaike Information Criterion (AIC) was adopted for validation. As a widely recognized information-theoretic metric for model selection, the AIC effectively balances goodness of fit with model parsimony. Specifically, lower AIC values indicate a more optimal trade-off between capturing meaningful temporal dependencies and minimizing the risk of overfitting, thereby ensuring the model remains both robust and computationally efficient. By comparing AIC values under different <italic>p</italic> settings, the results show that when <italic>p</italic>&#x202F;=&#x202F;2, the model achieves the smallest residual (AIC&#x202F;=&#x202F;&#x2212;3.2), outperforming <italic>p</italic>&#x202F;=&#x202F;1 (AIC&#x202F;=&#x202F;&#x2212;2.8) and <italic>p</italic>&#x202F;=&#x202F;3 (AIC&#x202F;=&#x202F;&#x2212;2.9). Therefore, the optimal lag order was finally determined to be <italic>p</italic>&#x202F;=&#x202F;2, ensuring an accurate representation of the surgical dynamics while avoiding overfitting.</p>
<p>The experimental parameters are set as follows: the sample size <italic>n</italic>&#x202F;=&#x202F;5,000, the lag order <italic>p</italic>&#x202F;=&#x202F;2, the generation probabilities for the variables <inline-formula>
<mml:math id="M95">
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> (normal gesture) and <inline-formula>
<mml:math id="M96">
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> (abnormal event) are 0.3 and 0.2, respectively, and the weighting coefficient for the lagged term <inline-formula>
<mml:math id="M97">
<mml:mi>S</mml:mi>
</mml:math>
</inline-formula> and that for the lagged term <inline-formula>
<mml:math id="M98">
<mml:mi>E</mml:mi>
</mml:math>
</inline-formula> are both 0.5. The regression coefficients (including <inline-formula>
<mml:math id="M99">
<mml:msub>
<mml:mi>&#x03B1;</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M100">
<mml:msub>
<mml:mi>&#x03B2;</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, <inline-formula>
<mml:math id="M101">
<mml:msub>
<mml:mi>&#x03B3;</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>, and <inline-formula>
<mml:math id="M102">
<mml:msub>
<mml:mi>&#x03B4;</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
</mml:math>
</inline-formula>) are estimated using Ordinary Least Squares (OLS), with a total of Q&#x202F;=&#x202F;4 lag terms.</p>
</sec>
<sec id="sec12">
<label>3.3</label>
<title>Experimental results</title>
<p><xref ref-type="table" rid="tab1">Table 1</xref> and <xref ref-type="fig" rid="fig3">Figure 3</xref> demonstrate the performance of the proposed method across varying sample sizes. The model achieved an accuracy exceeding 95.6% on both 5,000 and 10,000 samples, with the Matthews correlation coefficient stabilizing at 0.912. This indicates excellent discriminative capability and generalization performance.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Analysis of experimental results.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Positive example\Negative example</th>
<th align="center" valign="top">Accuracy rate</th>
<th align="center" valign="top">Precision rate</th>
<th align="center" valign="top">Recall rate</th>
<th align="center" valign="top">F1 score</th>
<th align="center" valign="top">MCC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">2,500\2,500</td>
<td align="center" valign="middle">95.60%</td>
<td align="center" valign="middle">95.34%</td>
<td align="center" valign="middle">95.88%</td>
<td align="center" valign="middle">95.60%</td>
<td align="center" valign="middle">0.912</td>
</tr>
<tr>
<td align="left" valign="middle">5,000\5,000</td>
<td align="center" valign="middle">95.66%</td>
<td align="center" valign="middle">94.96%</td>
<td align="center" valign="middle">96.44%</td>
<td align="center" valign="middle">95.77%</td>
<td align="center" valign="middle">0.912</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Confusion matrix analysis for different sample sizes. <bold>(a)</bold> The results of the model at a scale of 5,000 samples (2,500 positives and 2,500 negatives); <bold>(b)</bold> the results of the model at a scale of 10,000 samples (5,000 positives and 5,000 negatives).</p>
</caption>
<graphic xlink:href="fneur-17-1767832-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two confusion matrix graphics labeled &#x201C;a&#x201D; and &#x201C;b&#x201D; display actual versus predicted classifications with blue color scales. Image a shows values: true positive 2,397, false negative 103, false positive 117, true negative 2,383. Image b shows values: true positive 4,822, false negative 178, false positive 256, true negative 4,744. Both matrices use axes labeled actual positive example, actual negative example, positive prediction, and negative prediction.</alt-text>
</graphic>
</fig>
<p>It is worth noting that the model&#x2019;s recall consistently exceeds its precision, indicating a preference for minimizing the rate of missed true causative events&#x2014;that is, reducing false negatives. In safety-critical surgical scenarios, this bias toward &#x201C;better to err on the side of false positives than false negatives&#x201D; aligns with the safety-first clinical principle. It helps ensure all critical anomalies are effectively identified and trigger appropriate response mechanisms.</p>
</sec>
</sec>
<sec sec-type="discussion" id="sec13">
<label>4</label>
<title>Discussion</title>
<p>In high-complexity surgical scenarios such as neurosurgery and spinal surgery, confounding factors including instrument slippage, tissue deformation (<xref ref-type="bibr" rid="ref6">6</xref>, <xref ref-type="bibr" rid="ref7">7</xref>), and visual field occlusion embedded in task-level video data exhibit distinct characteristics of cross-phase correlation and dynamic coupling over the temporal dimension. The non-Markovian nature shaped by these long-range causal dependencies profoundly underscores the inherent complexity of unstructured surgical environments, emerging as a key bottleneck that hinders surgical robots from achieving high-level autonomous decision-making. However, traditional temporal modeling methods are constrained by short-range dependency assumptions (<xref ref-type="bibr" rid="ref8">8</xref>, <xref ref-type="bibr" rid="ref12">12</xref>), impeding their ability to effectively capture such long-range causal structures; meanwhile, existing systems based on deterministic rules or Markov Decision Processes (MDPs) fail to dynamically respond to and adaptively adjust for process disruptions induced by abnormal events, primarily due to their rigid architectures and inadequate state modeling. Ultimately, this severely compromises the reliability, robustness, and clinical applicability of these systems in real-world complex surgical settings.</p>
<p>Inspired by local causality discovery and non-stationary adaptive learning theories (<xref ref-type="bibr" rid="ref21">21</xref>, <xref ref-type="bibr" rid="ref22">22</xref>), this study proposes a dynamic reasoning framework based on Granger causality testing. This framework translates the clinical logic of &#x201C;surgical gesture&#x2013;abnormal event&#x2013;recovery action&#x201D; into computable temporal causal hypotheses. Through vector autoregressive modeling and Granger significance testing, it achieves data-driven identification of abnormal causal chains and dynamic decision-making.</p>
<p>Experimental results demonstrate that the framework achieves excellent and stable performance on synthetic datasets. The model demonstrated accuracy exceeding 95.6% and a MCC value of 0.912 across both 5,000 and 10,000 sample sizes. Notably, when the sample size increased to 10,000 instances, the F1 score remained at 95.77%, attesting to its exceptional stability and robust capability to distinguish genuine causality from coincidental correlations. The core advantage of this approach lies in constructing causal chains linking &#x201C;original gesture-abnormal event-recovery action&#x201D; (e.g., &#x201C;dropping needle &#x2192; picking up needle &#x2192; resuming threading&#x201D;). This not only captures the direct association between &#x201C;abnormal event &#x2192; countermeasure&#x201D; but also quantifies the dynamic influence of historical states on current decisions through the VAR model&#x2019;s lagged term coefficients. Consequently, it overcomes the traditional Markov model&#x2019;s reliance on &#x201C;fixed temporal sequences&#x201D;. Crucially, the model exhibits a persistent tendency for recall (95.88%) to exceed precision (95.34%), reflecting a cautious bias toward prioritizing &#x201C;better to report a false positive than miss a true positive.&#x201D; In safety-critical surgical scenarios, this design philosophy&#x2014;which minimizes false negatives (i.e., missed detection of abnormal causal pairs)&#x2014;aligns closely with the safety-first clinical principle. It effectively ensures that all critical anomalies (such as &#x201C;dropped needles&#x201D;) are identified and trigger appropriate responses, thereby significantly enhancing the system&#x2019;s inherent safety. The confusion matrix further confirms that the model&#x2019;s false negative rate consistently remains below 5%, demonstrating substantial application potential in surgical anomaly modeling and autonomous response.</p>
<p>Despite these initial advances, several limitations persist that warrant further investigation. First, model validation currently relies on synthetic data; while high performance is achieved under simplified assumptions, the framework may encounter generalization challenges in real surgical environments. Inherent complexities including continuous tissue deformation, variable instrument&#x2013;tissue interactions such as context-dependent friction or adhesion, and instrument occlusion introduce unmodeled dynamic noise that is not fully captured in synthetic datasets, potentially undermining the framework&#x2019;s reliability during clinical transition. To bridge the gap between simulation and actual clinical practice, we plan to collaborate with clinical institutions to collect and annotate real surgical videos across specialties such as neurosurgery and spinal surgery that include intraoperative anomalies. Establishing a high-quality dataset requires satisfying two core criteria, namely that the video library must cover an extensive spectrum of surgical scenarios to ensure robust representativeness, and that comprehensive documentation of surgical phases, potential anomalous events, and their respective mitigation strategies must be integrated. To meet these requirements, experienced surgeons will conduct systematic reviews of surgical footage to generate descriptive narratives, which will then be parsed and structured by large language models to refine classification systems for standard operative states and abnormal events. Following the formulation of rigorous annotation protocols, the labeling will be independently completed by multiple professionals, with inter-annotator consistency verified via the Kappa coefficient to guarantee the dataset&#x2019;s reliability and generalizability. Furthermore, the current Vector Autoregressive (VAR) model may struggle to capture the extended temporal dependencies inherent in complex procedures. Consequently, future work will explore incorporating advanced architectures such as Transformers to enhance the modeling of long-range causal relationships and further improve decision-making robustness.</p>
<p>In summary, this study addresses the critical issue of long-range dependencies triggered by abnormal events in non-Markovian surgical environments through a novel causal dynamic reasoning framework. By integrating Granger causality testing with vector autoregressive models, the proposed method successfully constructs interpretable &#x201C;surgical gesture-abnormal event-recovery action&#x201D; chains, enabling autonomous reasoning and dynamic decision-making. The framework&#x2019;s high accuracy, stability, and inherent bias toward safety&#x2014;evidenced by recall consistently exceeding precision&#x2014;provide a robust and clinically-aligned foundation for handling surgical disruptions. The insights and architecture presented here mark a significant step forward in propelling surgical robots from programmed execution toward cognitive, adaptive decision-making.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec14">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="author-contributions" id="sec15">
<title>Author contributions</title>
<p>GN: Conceptualization, Methodology, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. TM: Formal analysis, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. LT: Investigation, Software, Writing &#x2013; review &#x0026; editing. LYa: Data curation, Writing &#x2013; review &#x0026; editing. ZQ: Resources, Writing &#x2013; review &#x0026; editing. LYu: Data curation, Writing &#x2013; review &#x0026; editing. XT: Data curation, Writing &#x2013; review &#x0026; editing. SF: Funding acquisition, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>The authors thank the senior clinical consultants who reviewed the causal relationships in the dataset and the staff of collaborating institutions who supported the study with technical assistance.</p>
</ack>
<sec sec-type="COI-statement" id="sec16">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec17">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec18">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Long</surname><given-names>Y</given-names></name> <name><surname>Lin</surname><given-names>A</given-names></name> <name><surname>Kwok</surname><given-names>DHC</given-names></name> <name><surname>Zhang</surname><given-names>L</given-names></name> <name><surname>Yang</surname><given-names>Z</given-names></name> <name><surname>Shi</surname><given-names>K</given-names></name> <etal/></person-group>. <article-title>Surgical embodied intelligence for generalized task autonomy in laparoscopic robot-assisted surgery</article-title>. <source>Sci Robot</source>. (<year>2025</year>) <volume>10</volume>:<fpage>eadt3093</fpage>. doi: <pub-id pub-id-type="doi">10.1126/scirobotics.adt3093</pub-id>, <pub-id pub-id-type="pmid">40668896</pub-id></mixed-citation></ref>
<ref id="ref2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shademan</surname><given-names>A</given-names></name> <name><surname>Decker</surname><given-names>RS</given-names></name> <name><surname>Opfermann</surname><given-names>JD</given-names></name> <name><surname>Leonard</surname><given-names>S</given-names></name> <name><surname>Krieger</surname><given-names>A</given-names></name> <name><surname>Kim</surname><given-names>PCW</given-names></name></person-group>. <article-title>Supervised autonomous robotic soft tissue surgery</article-title>. <source>Sci Transl Med</source>. (<year>2016</year>) <volume>8</volume>:<fpage>337ra64</fpage>. doi: <pub-id pub-id-type="doi">10.1126/scitranslmed.aad9398</pub-id>, <pub-id pub-id-type="pmid">27147588</pub-id></mixed-citation></ref>
<ref id="ref3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shakir</surname><given-names>T</given-names></name> <name><surname>Atraszkiewicz</surname><given-names>D</given-names></name> <name><surname>Hassouna</surname><given-names>M</given-names></name> <name><surname>Pampiglione</surname><given-names>T</given-names></name> <name><surname>Chand</surname><given-names>M</given-names></name></person-group>. <article-title>Beyond diagnosis: how advanced imaging technologies are shaping modern surgery</article-title>. <source>Artif Intell Surg</source>. (<year>2025</year>) <volume>5</volume>:<fpage>270</fpage>&#x2013;<lpage>82</lpage>. doi: <pub-id pub-id-type="doi">10.20517/ais.2024.79</pub-id></mixed-citation></ref>
<ref id="ref4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname><given-names>JWB</given-names></name> <name><surname>Chen</surname><given-names>JT</given-names></name> <name><surname>Hansen</surname><given-names>P</given-names></name> <name><surname>Shi</surname><given-names>LX</given-names></name> <name><surname>Goldenberg</surname><given-names>A</given-names></name> <name><surname>Schmidgall</surname><given-names>S</given-names></name> <etal/></person-group>. <article-title>SRT-H: a hierarchical framework for autonomous surgery via language-conditioned imitation learning</article-title>. <source>Sci Robot.</source> (<year>2025</year>) <volume>10</volume>:<fpage>eadt5254</fpage>. doi: <pub-id pub-id-type="doi">10.1126/scirobotics.adt5254</pub-id>, <pub-id pub-id-type="pmid">40632876</pub-id></mixed-citation></ref>
<ref id="ref5"><label>5.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Haworth</surname><given-names>J</given-names></name> <name><surname>Biswas</surname><given-names>R</given-names></name> <name><surname>Opfermann</surname><given-names>J</given-names></name> <name><surname>Kam</surname><given-names>M</given-names></name> <name><surname>Wang</surname><given-names>Y</given-names></name> <name><surname>Pantalone</surname><given-names>D</given-names></name> <etal/></person-group> (<year>2024</year>). Autonomous robotic system with optical coherence tomography guidance for vascular anastomosis. arXiv. Available online at: <ext-link xlink:href="https://arxiv.org/abs/2410.07493" ext-link-type="uri">https://arxiv.org/abs/2410.07493</ext-link> (Accessed July 5, 2025)</mixed-citation></ref>
<ref id="ref6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tapper</surname><given-names>A</given-names></name> <name><surname>Leale</surname><given-names>D</given-names></name> <name><surname>Megahan</surname><given-names>G</given-names></name> <name><surname>Nacker</surname><given-names>K</given-names></name> <name><surname>Killinger</surname><given-names>K</given-names></name> <name><surname>Hafron</surname><given-names>J</given-names></name></person-group>. <article-title>Robotic instrument failure&#x2014;a critical analysis of cause and quality improvement strategies</article-title>. <source>Urology</source>. (<year>2019</year>) <volume>131</volume>:<fpage>125</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.urology.2019.02.052</pub-id>, <pub-id pub-id-type="pmid">31158353</pub-id></mixed-citation></ref>
<ref id="ref7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Monji-Azad</surname><given-names>S</given-names></name> <name><surname>Kinz</surname><given-names>M</given-names></name> <name><surname>Kothari</surname><given-names>S</given-names></name> <name><surname>Khanna</surname><given-names>R</given-names></name> <name><surname>Mihan</surname><given-names>AC</given-names></name> <name><surname>M&#x00E4;nnel</surname><given-names>D</given-names></name> <etal/></person-group>. <article-title>DefTransNet: a transformer-based method for non-rigid point cloud registration in the simulation of soft tissue deformation</article-title>. <source>Meas Sci Technol</source>. (<year>2025</year>) <volume>36</volume>:<fpage>076006</fpage>. doi: <pub-id pub-id-type="doi">10.1088/1361-6501/ade613</pub-id></mixed-citation></ref>
<ref id="ref8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Colledanchise</surname><given-names>M</given-names></name> <name><surname>Parasuraman</surname><given-names>R</given-names></name> <name><surname>&#x00D6;gren</surname><given-names>P</given-names></name></person-group>. <article-title>Learning of behavior trees for autonomous agents</article-title>. <source>IEEE Trans Games</source>. (<year>2019</year>) <volume>11</volume>:<fpage>183</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TG.2018.2816806</pub-id></mixed-citation></ref>
<ref id="ref9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hu</surname><given-names>D</given-names></name> <name><surname>Gong</surname><given-names>Y</given-names></name> <name><surname>Hannaford</surname><given-names>B</given-names></name> <name><surname>Seibel</surname><given-names>EJ</given-names></name></person-group>. <article-title>Semi-autonomous simulated brain tumor ablation with RAVENII surgical robot using behavior tree</article-title>. <source>Proc IEEE Int Conf Robot Autom</source>. (<year>2015</year>):<fpage>3868</fpage>&#x2013;<lpage>75</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ICRA.2015.7139738</pub-id></mixed-citation></ref>
<ref id="ref10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hayes</surname><given-names>B</given-names></name> <name><surname>Scassellati</surname><given-names>B</given-names></name></person-group>. <article-title>Autonomously constructing hierarchical task networks for planning and human-robot collaboration</article-title>. <source>Proc IEEE Int Conf Robot Autom</source>. (<year>2016</year>):<fpage>5469</fpage>&#x2013;<lpage>76</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ICRA.2016.7487760</pub-id></mixed-citation></ref>
<ref id="ref11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pardowitz</surname><given-names>M</given-names></name> <name><surname>Knoop</surname><given-names>S</given-names></name> <name><surname>Dillmann</surname><given-names>R</given-names></name> <name><surname>Zollner</surname><given-names>RD</given-names></name></person-group>. <article-title>Incremental learning of tasks from user demonstrations, past experiences, and vocal comments</article-title>. <source>IEEE Trans Cybern</source>. (<year>2007</year>) <volume>37</volume>:<fpage>322</fpage>&#x2013;<lpage>32</lpage>. doi: <pub-id pub-id-type="doi">10.1109/tsmcb.2006.886951</pub-id>, <pub-id pub-id-type="pmid">17416160</pub-id></mixed-citation></ref>
<ref id="ref12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>J</given-names></name> <name><surname>Dridi</surname><given-names>M</given-names></name> <name><surname>El Moudni</surname><given-names>A</given-names></name></person-group>. <article-title>Scheduling elective surgeries with Markov decision process and approximate dynamic programming</article-title>. <source>IFAC-PapersOnLine</source>. (<year>2019</year>) <volume>52</volume>:<fpage>1831</fpage>&#x2013;<lpage>6</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ifacol.2019.11.468</pub-id></mixed-citation></ref>
<ref id="ref13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alison</surname><given-names>B</given-names></name> <name><surname>Brown</surname><given-names>L</given-names></name></person-group>. <article-title>BJS.03 Markov decision analysis of treatment pathways for resectable malignancy of the oesophagus or gastrooesophageal junction</article-title>. <source>Br J Surg</source>. (<year>2021</year>) <volume>108</volume>:<fpage>znab310.005</fpage>. doi: <pub-id pub-id-type="doi">10.1093/bjs/znab310.005</pub-id></mixed-citation></ref>
<ref id="ref14"><label>14.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Kim</surname><given-names>JW</given-names></name> <name><surname>Zhao</surname><given-names>TZ</given-names></name> <name><surname>Schmidgall</surname><given-names>S</given-names></name> <name><surname>Deguet</surname><given-names>A</given-names></name> <name><surname>Kobilarov</surname><given-names>M</given-names></name> <name><surname>Finn</surname><given-names>C</given-names></name> <etal/></person-group> (<year>2024</year>). Surgical robot transformer (SRT): imitation learning for surgical tasks. arXiv Available online at: <ext-link xlink:href="https://arxiv.org/abs/2407.12998" ext-link-type="uri">https://arxiv.org/abs/2407.12998</ext-link> (Accessed July 8, 2025)</mixed-citation></ref>
<ref id="ref15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zeng</surname><given-names>Y</given-names></name> <name><surname>Cai</surname><given-names>R</given-names></name> <name><surname>Sun</surname><given-names>F</given-names></name> <name><surname>Huang</surname><given-names>L</given-names></name> <name><surname>Hao</surname><given-names>Z</given-names></name></person-group>. <article-title>A survey on causal reinforcement learning</article-title>. <source>IEEE Trans Neural Netw Learn Syst</source>. (<year>2025</year>) <volume>36</volume>:<fpage>5942</fpage>&#x2013;<lpage>62</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TNNLS.2024.3403001</pub-id>, <pub-id pub-id-type="pmid">40030342</pub-id></mixed-citation></ref>
<ref id="ref16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bareinboim</surname><given-names>E</given-names></name> <name><surname>Pearl</surname><given-names>J</given-names></name></person-group>. <article-title>Causal inference and the data-fusion problem</article-title>. <source>Proc Natl Acad Sci USA</source>. (<year>2016</year>) <volume>113</volume>:<fpage>7345</fpage>&#x2013;<lpage>52</lpage>. doi: <pub-id pub-id-type="doi">10.1073/pnas.1510507113</pub-id>, <pub-id pub-id-type="pmid">27382148</pub-id></mixed-citation></ref>
<ref id="ref17"><label>17.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Dasgupta</surname><given-names>I</given-names></name> <name><surname>Wang</surname><given-names>J</given-names></name> <name><surname>Chiappa</surname><given-names>S</given-names></name> <name><surname>Mitrovic</surname><given-names>J</given-names></name> <name><surname>Ortega</surname><given-names>P</given-names></name> <name><surname>Raposo</surname><given-names>D</given-names></name> <etal/></person-group> (<year>2019</year>). Causal reasoning from meta-reinforcement learning. arXiv. Available online at: <ext-link xlink:href="https://arxiv.org/abs/1901.08162" ext-link-type="uri">https://arxiv.org/abs/1901.08162</ext-link> (Accessed July 8, 2025)</mixed-citation></ref>
<ref id="ref18"><label>18.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Pitis</surname><given-names>S</given-names></name> <name><surname>Creager</surname><given-names>E</given-names></name> <name><surname>Garg</surname><given-names>A</given-names></name></person-group>. "<chapter-title>Counterfactual data augmentation using locally factored dynamics</chapter-title>" In: <person-group person-group-type="editor"><name><surname>Larochelle</surname><given-names>H</given-names></name> <name><surname>Ranzato</surname><given-names>M</given-names></name> <name><surname>Hadsell</surname><given-names>R</given-names></name> <name><surname>Balcan</surname><given-names>MF</given-names></name> <name><surname>Lin</surname><given-names>H</given-names></name></person-group>, editors. <source>Advances in neural information processing systems 33</source> (<year>2020</year>). <fpage>3976</fpage>&#x2013;<lpage>90</lpage>.</mixed-citation></ref>
<ref id="ref19"><label>19.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Buesing</surname><given-names>L</given-names></name> <name><surname>Weber</surname><given-names>T</given-names></name> <name><surname>Zwols</surname><given-names>Y</given-names></name> <name><surname>Racaniere</surname><given-names>S</given-names></name> <name><surname>Guez</surname><given-names>A</given-names></name> <name><surname>Lespiau</surname><given-names>JB</given-names></name> <etal/></person-group> (<year>2018</year>). Woulda, coulda, shoulda: counterfactually-guided policy search. arXiv. Available online at: <ext-link xlink:href="https://arxiv.org/abs/1811.06272" ext-link-type="uri">https://arxiv.org/abs/1811.06272</ext-link> (Accessed October 10, 2025).</mixed-citation></ref>
<ref id="ref20"><label>20.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Mozifian</surname><given-names>M</given-names></name> <name><surname>Zhang</surname><given-names>A</given-names></name> <name><surname>Pineau</surname><given-names>J</given-names></name> <name><surname>Meger</surname><given-names>D</given-names></name></person-group>. (<year>2020</year>). Intervention design for effective Sim2Real transfer. arXiv. Available online at: <ext-link xlink:href="https://arxiv.org/abs/2012.02055" ext-link-type="uri">https://arxiv.org/abs/2012.02055</ext-link> (Accessed July 10, 2025)</mixed-citation></ref>
<ref id="ref21"><label>21.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Seitzer</surname><given-names>M</given-names></name> <name><surname>Sch&#x00F6;lkopf</surname><given-names>B</given-names></name> <name><surname>Martius</surname><given-names>G</given-names></name></person-group>. "<chapter-title>Causal influence detection for improving efficiency in reinforcement learning</chapter-title>" In: <person-group person-group-type="editor"><name><surname>Ranzato</surname><given-names>M</given-names></name> <name><surname>Beygelzimer</surname><given-names>A</given-names></name> <name><surname>Dauphin</surname><given-names>Y</given-names></name> <name><surname>Liang</surname><given-names>PS</given-names></name> <name><surname>Vaughan</surname><given-names>JW</given-names></name></person-group>, editors. <source>Advances in neural information processing systems 34</source> (<year>2021</year>). <fpage>22905</fpage>&#x2013;<lpage>18</lpage>.</mixed-citation></ref>
<ref id="ref22"><label>22.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Feng</surname><given-names>F</given-names></name> <name><surname>Huang</surname><given-names>B</given-names></name> <name><surname>Zhang</surname><given-names>K</given-names></name> <name><surname>Magliacane</surname><given-names>S</given-names></name></person-group>. "<chapter-title>Factored adaptation for non-stationary reinforcement learning</chapter-title>" In: <person-group person-group-type="editor"><name><surname>Koyejo</surname><given-names>S</given-names></name> <name><surname>Mohamed</surname><given-names>S</given-names></name> <name><surname>Agarwal</surname><given-names>A</given-names></name> <name><surname>Belgrave</surname><given-names>D</given-names></name> <name><surname>Cho</surname><given-names>K</given-names></name> <name><surname>Oh</surname><given-names>A</given-names></name></person-group>, editors. <source>Advances in neural information processing systems. Red Hook, NY, USA: Curran Associates, Inc.</source> (<year>2022</year>) 35, <fpage>31957</fpage>&#x2013;<lpage>71</lpage>.</mixed-citation></ref>
<ref id="ref23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Granger</surname><given-names>CWJ</given-names></name></person-group>. <article-title>Testing for causality: a personal viewpoint</article-title>. <source>J Econ Dyn Control</source>. (<year>1980</year>) <volume>2</volume>:<fpage>329</fpage>&#x2013;<lpage>52</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0165-1889(80)90069-X</pub-id></mixed-citation></ref>
<ref id="ref24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Granger</surname><given-names>CWJ</given-names></name></person-group>. <article-title>Investigating causal relations by econometric models and cross-spectral methods</article-title>. <source>Econometrica</source>. (<year>1969</year>) <volume>37</volume>:<fpage>424</fpage>&#x2013;<lpage>38</lpage>. doi: <pub-id pub-id-type="doi">10.2307/1912791</pub-id></mixed-citation></ref>
<ref id="ref25"><label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Okada</surname><given-names>T</given-names></name> <name><surname>Okamoto</surname><given-names>S</given-names></name> <name><surname>Yamada</surname><given-names>Y</given-names></name></person-group>. <article-title>Affective dynamics: causality modeling of temporally evolving perceptual and affective responses</article-title>. <source>IEEE Trans Affect Comput</source>. (<year>2022</year>) <volume>13</volume>:<fpage>628</fpage>&#x2013;<lpage>39</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TAFFC.2019.2942931</pub-id></mixed-citation></ref>
<ref id="ref26"><label>26.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shao</surname><given-names>R</given-names></name> <name><surname>Lan</surname><given-names>X</given-names></name></person-group>. <article-title>Adversarial auto-encoder for unsupervised deep domain adaptation</article-title>. <source>IET Image Process</source>. (<year>2019</year>) <volume>13</volume>:<fpage>2772</fpage>&#x2013;<lpage>7</lpage>. doi: <pub-id pub-id-type="doi">10.1049/iet-ipr.2018.6687</pub-id></mixed-citation></ref>
<ref id="ref27"><label>27.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>H</given-names></name> <name><surname>Liu</surname><given-names>H</given-names></name> <name><surname>Fu</surname><given-names>H</given-names></name> <name><surname>Xu</surname><given-names>Y</given-names></name> <name><surname>Shu</surname><given-names>H</given-names></name> <name><surname>Niu</surname><given-names>K</given-names></name> <etal/></person-group>. <article-title>A generic fundus image enhancement network boosted by frequency self-supervised representation learning</article-title>. <source>Med Image Anal</source>. (<year>2023</year>) <volume>90</volume>:<fpage>102945</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2023.102945</pub-id>, <pub-id pub-id-type="pmid">37703674</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1992972/overview">Keping Yu</ext-link>, Hosei University, Japan</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3321110/overview">Heng Li</ext-link>, Shenzhen University of Advanced Technology, China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3321172/overview">Xuquan Ji</ext-link>, Beihang University, China</p>
</fn>
</fn-group>
</back>
</article>