<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="review-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Digit. Health</journal-id><journal-title-group>
<journal-title>Frontiers in Digital Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Digit. Health</abbrev-journal-title></journal-title-group>
<issn pub-type="epub">2673-253X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fdgth.2026.1633888</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Systematic Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Bibliometric analysis of deep learning for surgical instrument segmentation, detection and tracking in minimally invasive surgery</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Yousef</surname><given-names>Mahmoud</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/2977723/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role></contrib>
<contrib contrib-type="author"><name><surname>Aly</surname><given-names>Kareem Essam</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role></contrib>
<contrib contrib-type="author"><name><surname>Ahmed</surname><given-names>Mariam</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Ahmed</surname><given-names>Fatimaelzahraa Ali</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role></contrib>
<contrib contrib-type="author"><name><surname>Al Jalham</surname><given-names>Khalid</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author" corresp="yes"><name><surname>Balakrishnan</surname><given-names>Shidin</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x2020;</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/2637460/overview" /><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role></contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Weill Cornell Medicine Qatar</institution>, <city>Doha</city>, <country country="qa">Qatar</country></aff>
<aff id="aff2"><label>2</label><institution>College of Medicine, Qatar University</institution>, <city>Doha</city>, <country country="qa">Qatar</country></aff>
<aff id="aff3"><label>3</label><institution>Department of Surgery, Hamad Medical Corporation</institution>, <city>Doha</city>, <country country="qa">Qatar</country></aff>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Shidin Balakrishnan <email xlink:href="mailto:sbalakrishnan1@hamad.qa">sbalakrishnan1@hamad.qa</email></corresp>
<fn fn-type="other" id="fn001"><label>&#x2020;</label><p>ORCID Shidin Balakrishnan <uri xlink:href="https://orcid.org/0000-0001-6361-4980">orcid.org/0000-0001-6361-4980</uri></p></fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-27"><day>27</day><month>02</month><year>2026</year></pub-date>
<pub-date publication-format="electronic" date-type="collection"><year>2026</year></pub-date>
<volume>8</volume><elocation-id>1633888</elocation-id>
<history>
<date date-type="received"><day>23</day><month>05</month><year>2025</year></date>
<date date-type="rev-recd"><day>21</day><month>01</month><year>2026</year></date>
<date date-type="accepted"><day>27</day><month>01</month><year>2026</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 Yousef, Aly, Ahmed, Ahmed, Al Jalham and Balakrishnan.</copyright-statement>
<copyright-year>2026</copyright-year><copyright-holder>Yousef, Aly, Ahmed, Ahmed, Al Jalham and Balakrishnan</copyright-holder><license><ali:license_ref start_date="2026-02-27">https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p></license>
</permissions>
<abstract><sec><title>Background</title>
<p>Deep learning (DL) methods for surgical video analysis have expanded rapidly in minimally invasive surgery (MIS). However, a structured bibliometric overview focused on DL-based surgical instrument segmentation, detection, and tracking is lacking. The objective of this review is to systematically map the research landscape with this focus, by examining publication trends, influential authors, institutions, and countries, collaboration networks, keyword co-occurrence patterns, and the thematic trajectory of the discipline.</p>
</sec><sec><title>Methods</title>
<p>We performed a bibliometric analysis of original research articles on DL-based surgical instrument segmentation/detection/tracking in laparoscopic or robotic MIS, published between 2017 and 2024. Searches were conducted in six databases namely PubMed, Scopus, IEEE Xplore, Embase, Medline, and Web of Science. Records were de-duplicated in EndNote and analyzed using the Bibliometrix R package, with co-authorship, co-citation, and keyword networks visualized in VOSviewer. Citation counts were extracted from each study&#x0027;s respective database and interpreted cautiously given the influence of publication age.</p>
</sec><sec><title>Results</title>
<p>We included 217 articles. Annual output increased from 2017 to a peak in 2023, indicating sustained growth in DL research for MIS instrument analysis. The most productive countries included the United States and France, with major institutional contributions from the University of Strasbourg and Furtwangen University. Keyword analysis indicated continued dominance of convolutional neural networks alongside emerging themes including transformer-based architectures, multimodal learning, and real-time intraoperative applications.</p>
</sec><sec><title>Conclusions</title>
<p>This bibliometric study characterizes the evolution, leading contributors, collaboration patterns, and thematic trajectories of DL-based instrument segmentation/detection/tracking in MIS. While these findings can inform research prioritization and collaboration, this study does not evaluate clinical effectiveness. Future work should prioritize explainable and efficient real-time models, standardized annotation protocols, and broader global partnerships to support responsible clinical translation.</p>
</sec>
</abstract>
<kwd-group>
<kwd>bibliometric analysis</kwd>
<kwd>minimally invasive surgery</kwd>
<kwd>surgical instrument detection and tracking</kwd>
<kwd>surgical instrument segmentation</kwd>
<kwd>surgical video analysis</kwd>
</kwd-group><funding-group><award-group id="gs1"><funding-source id="sp1"><institution-wrap><institution>Qatar National Library</institution><institution-id institution-id-type="doi" vocab="open-funder-registry" vocab-identifier="10.13039/open_funder_registry">10.13039/100019779</institution-id></institution-wrap></funding-source></award-group><funding-statement>The author(s) declared that financial support was received for this work and/or its publication. Research reported in this publication was supported by the Qatar Research Development and Innovation Council (QRDI) grant number ARG01-0522-230266.</funding-statement></funding-group><counts>
<fig-count count="12"/>
<table-count count="7"/><equation-count count="0"/><ref-count count="172"/><page-count count="22"/><word-count count="0"/></counts><custom-meta-group><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Health Informatics</meta-value></custom-meta></custom-meta-group>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>Minimally invasive surgery (MIS) represents a broad class of operative techniques that reduce surgical tissue damage by utilizing advanced intraoperative visualization. MIS includes laparoscopy and robotic-assisted surgery. In MIS, an endoscope is inserted into the patient&#x0027;s body through a small incision to give a clear and magnified view of the surgical site. Robotic-assisted surgery allows surgeons to control instruments from a console without direct interaction with the patient. The number of laparoscopic surgeries performed each year has exceeded 15 million (<xref ref-type="bibr" rid="B1">1</xref>) and robotic surgeries have represented over 15&#x0025; of cases in fields like general surgery (<xref ref-type="bibr" rid="B2">2</xref>) and over 85&#x0025; for certain procedures such as prostatectomies (<xref ref-type="bibr" rid="B3">3</xref>). Since these procedures use cameras, surgeries are often recorded. This has led to the development of publicly available surgical datasets, including benchmark resources from the MICCAI Endoscopic Vision (EndoVIS) challenges and datasets like HeiChole, that have helped facilitate advancements in the field (<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B5">5</xref>).</p>
<p>Computer vision and machine learning, subsets of Artificial Intelligence (AI), have played an important role in turning raw surgical video footage into processed and annotated datasets that can then be used for research and in clinical settings. Recently, advancements in deep learning (DL) techniques has enabled substantial improvements in modelling MIS, especially coupled with the growth of surgical data volume. Computer vision tasks like detecting objects, identifying instruments, segmenting crucial anatomical areas and recognition of surgical phase in surgical videos rely heavily on DL models. The exceptional performance of DL in various applications suggests active potential for automation in surgery, aiming to provide real-time insights during operations and supports surgeons in making better decisions. Furthermore, the development and release of open-source foundation models have been transformative, enabling researchers to collaborate and implement DL into surgical practices rapidly. For example, DL has been applied in instrument segmentation, surgical phase recognition, or anatomical landmark detection. For example, models like EndoNet (<xref ref-type="bibr" rid="B6">6</xref>) and SV-RCNet (<xref ref-type="bibr" rid="B7">7</xref>), as well as learning-based approaches for surgical phase recognition based on temporal modeling of instrument usage and video context (<xref ref-type="bibr" rid="B8">8</xref>) have all contributed significantly in advancing real-time workflow recognition and tool classification in laparoscopic videos. Other promising innovations include the use of Vision Transformers (ViTs) (<xref ref-type="bibr" rid="B9">9</xref>, <xref ref-type="bibr" rid="B10">10</xref>) for surgical tool detection or Generative Adversarial Networks (GANs) (<xref ref-type="bibr" rid="B11">11</xref>, <xref ref-type="bibr" rid="B12">12</xref>) for enhancing video quality.</p>
<p>These technical advancements suggest a growing importance of DL in MIS, yet the literature has not been systematically analyzed. These gaps leave significant questions unanswered regarding the intellectual structure of the field, the collaborative networks driving its growth, and the emerging trends that may shape its future. Similar bibliometric approaches have been used in other specialized fields to elucidate research impact and thematic evolution (<xref ref-type="bibr" rid="B13">13</xref>). By providing a structured, data-driven overview of impact and collaboration patterns, such analyses play a critical role in guiding future research priorities, identifying opportunities for meaningful collaboration, and supporting the coordinated advancement of the field in a coherent and informed manner.</p>
<p>This study aims to address the aforementioned gap by conducting a detailed bibliometric analysis of current research on DL in MIS, gathered from six major databases. It analyzes the most cited studies in the field, specifically those using deep learning for instrument segmentation, detection, and tracking in MIS. We included only papers that developed DL models due to their superior performance on complex tasks, such semantic segmentation, to: (1) examine publication trends and the growth trajectory of the field, (2) identify influential authors, institutions, and countries driving advancements in DL for MIS, (3) analyze collaboration networks and citation patterns to understand the field&#x0027;s intellectual and social structure, (4) explore keyword co-occurrence and emerging research topics, and (5) propose future research directions based on identified trends.</p>
<p>This bibliometric review provides a structured overview of the evolution of research on DL in MIS. By analyzing trends in publications, citations, collaboration patterns, and research topics, the study offers insight into the organization and progression of the field. Unlike clinical studies or meta-analyses that primarily assess technical performance or clinical accuracy, this analysis is intended to inform researchers about broader research patterns and gaps in the literature, offering a foundation for future work by supporting informed study design and collaboration.</p>
</sec>
<sec id="s1a"><label>2</label><title>Methodology</title>
<p>This bibliometric analysis was conducted in accordance with established methodological conventions used in previously published bibliometric studies across various clinical domains, ensuring methodological rigor (<xref ref-type="bibr" rid="B13">13</xref>&#x2013;<xref ref-type="bibr" rid="B21">21</xref>). Accordingly, this study has been designed and reported in accordance with the BIBLIO checklist for bibliometric analyses (completed checklist provided as Supplementary File S1).</p>
<sec id="s1b"><label>2.1</label><title>Search strategy</title>
<p>A comprehensive search was performed across six databases namely PubMed, Scopus, IEEE Xplore, Embase, Medline, and Web of Science using keywords like &#x201C;robotic-assisted surgery,&#x201D; &#x201C;minimally invasive surgery,&#x201D; &#x201C;deep learning,&#x201D; and &#x201C;computer vision&#x201D;, for the time period 2017&#x2013;2024. We limited the analysis to 2017&#x2013;2024 for two reasons. First, 2017 corresponds to the earliest year in which deep learning&#x2013;based approaches for surgical instrument analysis in minimally invasive/video-guided surgery were consistently represented in major indexing services, coinciding with the emergence of widely used community benchmarks and challenges (e.g., EndoNet-era laparoscopic video recognition and the EndoVis challenge series (<xref ref-type="bibr" rid="B6">6</xref>). Second, 2024 was the most recent complete publication year at the time of our search. We did not include 2025 because the year was ongoing during data collection and indexing and citation accrual would be incomplete.</p>
<p>We included papers that focused on laparoscopic or robotic-assisted surgeries besides utilizing DL models such as Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, Generative Adversarial Networks (GANs), or Vision Transformers (ViT). Because instrument-related surgical video studies are not consistently labeled with instrument-specific keywords in titles and abstracts, we intentionally used a high-sensitivity search strategy focused on MIS&#x2009;&#x002B;&#x2009;deep learning&#x2009;&#x002B;&#x2009;video/endoscopy terms. Specificity was then enforced during screening using predefined eligibility criteria requiring a direct link to instrument segmentation, detection, or tracking.</p>
<p>The search strategy is shown in <xref ref-type="table" rid="T1">Table&#x00A0;1</xref> below:</p>
<table-wrap id="T1" position="float"><label>Table&#x00A0;1</label>
<caption><p>Search strategy.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Concept</th>
<th valign="top" align="center">Keywords and MeSH Terms</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="3">Minimally Invasive Surgery</td>
<td valign="top" align="left">Keywords: laparoscop&#x002A; OR &#x201C;partial nephrectomy&#x201D; OR endoscop&#x002A; OR adrenalectomy OR cholecystectomy OR splenectomy OR nephrectomy OR &#x201C;general surgery&#x201D; OR &#x201C;minimally invasive surgical procedur&#x002A;&#x201D; OR &#x201C;minimally invasive surgery&#x201D; OR &#x201C;surgical procedur&#x002A;&#x201D; OR &#x201C;operation room&#x201D; OR &#x201C; surgery robot&#x002A;&#x201D; OR &#x201C;surgical robot&#x002A;&#x201D;</td>
</tr>
<tr>
<td valign="top" align="left">MeSH term: &#x201C;Minimally Invasive Surgical Procedures&#x201D;[Mesh] &#x201C;Surgical Procedures, Operative&#x201D;[Mesh]</td>
</tr>
<tr>
<td valign="top" align="left">Search: &#x201C;Minimally Invasive Surgical Procedures&#x201D;[Mesh] OR &#x201C;Surgical Procedures, Operative&#x201D;[Mesh] OR laparoscop&#x002A; OR &#x201C;partial nephrectomy&#x201D; OR endoscop&#x002A; OR adrenalectomy OR cholecystectomy OR splenectomy OR nephrectomy OR &#x201C;general surgery&#x201D; OR &#x201C;minimally invasive surgical procedur&#x002A;&#x201D; OR &#x201C;minimally invasive surgery&#x201D; OR &#x201C;surgical procedur&#x002A;&#x201D; OR &#x201C;operation room&#x201D; OR &#x201C; surgery robot&#x002A;&#x201D; OR &#x201C;surgical robot&#x002A;&#x201D;</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="3">Deep Learning</td>
<td valign="top" align="left">Keywords:&#x201C;artificial intelligence&#x201D; OR &#x201C;machine learning&#x201D; OR &#x201C;Image-guided Surgery&#x201D; OR &#x201C;deep learning&#x201D; OR &#x201C;artificial neural network&#x002A;&#x201D; OR &#x201C;neural network&#x002A;&#x201D; OR &#x201C;convolutional neural network&#x002A;&#x201D;</td>
</tr>
<tr>
<td valign="top" align="left">MeSH terms: &#x201C;Artificial Intelligence&#x201D;[Mesh]</td>
</tr>
<tr>
<td valign="top" align="left">Search: &#x201C;Artificial Intelligence&#x0022;[Mesh] OR &#x201C;artificial intelligence&#x201D; OR &#x201C;machine learning&#x201D; OR &#x201C;Image-guided Surgery&#x201D; OR &#x201C;deep learning&#x201D; OR &#x201C;artificial neural network&#x002A;&#x201D; OR &#x201C;neural network&#x002A;&#x201D; OR &#x201C;convolutional neural network&#x002A;&#x0022;</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="2">Video</td>
<td valign="top" align="left">Keywords: video&#x002A; OR &#x201C;Video Recording&#x002A;&#x201D; OR &#x201C;Video-Assisted Surgery&#x201D; MeSH term: &#x201C;Video Recording&#x201D;[Mesh]</td>
</tr>
<tr>
<td valign="top" align="left">Search: &#x201C;Video Recording&#x201D;[Mesh] OR video&#x002A; OR &#x201C;Video Recording&#x002A;&#x201D; OR &#x201C;Video-Assisted Surgery&#x201D;</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s1c"><label>2.2</label><title>Screening &#x0026; inclusion/exclusion criteria</title>
<p>A thorough process was implemented to ensure the inclusion of eligible studies. Firstly, the search results were exported in RIS format and imported into Endnote 21.3 (Clarivate Analytics, Philadelphia, PA, USA). Duplicate records were then removed, after which all retrieved studies underwent independent title and abstract screening by two reviewers, followed by full-text assessment by two reviewers to determine eligibility. Discrepancies were resolved by consensus. This strategy ensured high sensitivity at the search stage while maintaining strict specificity through subsequent screening based on predefined eligibility criteria.</p>
<p>Our exclusion criteria included research focusing solely on conventional image processing or non-DL machine learning models. Additionally, studies that were applied to ex-vivo testing or simulation kits were not considered. Secondary research like reviews and meta-analyses were excluded to maintain the focus on original studies. Furthermore, studies focusing on workflow recognition without a direct connection to surgical tool detection were also excluded, though those incorporating tool detection as part of surgical phase identification were included.</p>
</sec>
<sec id="s1d"><label>2.3</label><title>Data extraction</title>
<p>Data were extracted clearly and systematically from the 217 included papers. The extracted metadata included the title, list of authors, year of publication, citation count, title of the journal, publisher, institutional affiliations of authors, and keywords. Citation counts were extracted from each study&#x0027;s respective indexing database. We report absolute citation counts as primary bibliometric indicators. Because citations accumulate over time and are influenced by publication age and indexing practices, we also report citations per year (total citations divided by years since publication) for the most-cited studies and interpret cross-country and cross-institution comparisons cautiously. Field-normalized metrics such as field-weighted citation impact (FWCI) were not used because they were not consistently available across all retrieved records. Papers with 50 citation counts or above were analyzed critically, by extracting the DL methodology, type of procedure and application. This approach helped us understand the correlation between the popularity of these papers and the direction of research. Later, the metadata of the 217 papers were analyzed using specific bibliometric software application. It allows the identification of important patterns of research, the identification of influential authors, and the study of relationships in the structure.</p>
</sec>
<sec id="s1e"><label>2.4</label><title>Bibliometric analysis techniques</title>
<p>After duplicates removal, the RIS file containing bibliometric data of the 217 studies was imported into Bibliometrix (<xref ref-type="bibr" rid="B22">22</xref>) a bibliometrics analysis tool driven by R programming language (version 4.1.3; R Foundation for Statistical Computing, Vienna, Austria). The data in this file was processed in a systematic manner to create tables and figures representing major trends and emergent patterns in the research landscape. The mapping of the co-authorship and citation patterns was created using the VOSviewer 1.6.20 (<xref ref-type="bibr" rid="B23">23</xref>) software package developed at Leiden University by Van Eck. Hence, these visualizations allowed deep insights into the relations among countries, authors, and institutions and thus shed light on the collaborative nature of research in DL in MIS.</p>
</sec>
</sec>
<sec id="s2" sec-type="results"><label>3</label><title>Results</title>
<sec id="s2a"><label>3.1</label><title>Growth of publications over time</title>
<p>The utilization of DL in the segmentation of surgical instruments for MIS has seen substantial advancement over the last eight years (<xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>). The initial articles on surgical data recognition via deep learning first appeared in 2017, starting with eight papers (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B24">24</xref>&#x2013;<xref ref-type="bibr" rid="B30">30</xref>). The sector experienced consistent progress, resulting in a linear growth trajectory start in 2021. During this interval, an average of 35 papers were published per year over two consecutive years, indicating that the publication rate had reached a plateau. Nonetheless, this plateau was momentary, as the field resumed its upward trend, ultimately reaching a peak in 2023 and exceeding 50 publications.</p>
<fig id="F1" position="float"><label>Figure&#x00A0;1</label>
<caption><p>Total number of articles published per year, displaying the growth of publications in the field.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g001.tif"><alt-text content-type="machine-generated">Line graph showing the number of articles published from 2017 to 2023. The number starts under 10 in 2017, rises steeply to about 30 in 2019, then gradually increases to 50 by 2023.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2b"><label>3.2</label><title>Distribution by publication type</title>
<p>Since 2017, 217 papers on the application of DL for surgical instrument recognition in MIS have been published, with an average citation around 20 per paper. These include 128 journal articles and 89 conference papers (<xref ref-type="table" rid="T2">Table&#x00A0;2</xref>). Different journals and conferences showed interest in this topic, resulting in 93 various sources. The average number of authors per article was 6.27, with 25.11&#x0025; of them involving international co-authorships.</p>
<table-wrap id="T2" position="float"><label>Table&#x00A0;2</label>
<caption><p>Summary of relevant characteristics of publications including general information, document contents, authors, and document types.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Characteristic</th>
<th valign="top" align="center">Value</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">General information</td>
</tr>
<tr>
<td valign="top" align="left">Timespan</td>
<td valign="top" align="center">2017&#x2013;2024</td>
</tr>
<tr>
<td valign="top" align="left">Number of Various Sources (Journals, Conferences, etc.)</td>
<td valign="top" align="center">93</td>
</tr>
<tr>
<td valign="top" align="left">Number of Documents</td>
<td valign="top" align="center">217</td>
</tr>
<tr>
<td valign="top" align="left">Annual Growth Rate &#x0025;</td>
<td valign="top" align="center">9.43</td>
</tr>
<tr>
<td valign="top" align="left">Average Years since Publication</td>
<td valign="top" align="center">3.06</td>
</tr>
<tr>
<td valign="top" align="left">Average citations per doc</td>
<td valign="top" align="center">19.7</td>
</tr>
<tr>
<td valign="top" align="left">References</td>
<td valign="top" align="center">6,179</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Document contents</td>
</tr>
<tr>
<td valign="top" align="left">Keywords Plus (ID)</td>
<td valign="top" align="center">1,501</td>
</tr>
<tr>
<td valign="top" align="left">Author&#x0027;s Keywords (DE)</td>
<td valign="top" align="center">482</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Authors</td>
</tr>
<tr>
<td valign="top" align="left">Authors</td>
<td valign="top" align="center">974</td>
</tr>
<tr>
<td valign="top" align="left">Authors of single-authored docs</td>
<td valign="top" align="center">5</td>
</tr>
<tr>
<td valign="top" align="left">Single-authored docs</td>
<td valign="top" align="center">5</td>
</tr>
<tr>
<td valign="top" align="left">Average Number of Co-Authors per Doc</td>
<td valign="top" align="center">6.27</td>
</tr>
<tr>
<td valign="top" align="left">International co-authorships (&#x0025;)</td>
<td valign="top" align="center">25.11</td>
</tr>
<tr>
<td valign="top" align="left" style="background-color:#d9d9d9" colspan="2">Document types</td>
</tr>
<tr>
<td valign="top" align="left">Articles</td>
<td valign="top" align="center">128</td>
</tr>
<tr>
<td valign="top" align="left">Conference Papers</td>
<td valign="top" align="center">89</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2c"><label>3.3</label><title>Distribution of publications by country</title>
<p>Geographical analysis reveals that many countries have made significant contributions to the field (<xref ref-type="fig" rid="F2">Figure 2</xref>). The United States leads in the total number of publications (35) (<xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B31">31</xref>&#x2013;<xref ref-type="bibr" rid="B64">64</xref>), followed by China (25) (<xref ref-type="bibr" rid="B44">44</xref>&#x2013;<xref ref-type="bibr" rid="B68">68</xref>) and Germany (24) (<xref ref-type="bibr" rid="B24">24</xref>, <xref ref-type="bibr" rid="B65">65</xref>&#x2013;<xref ref-type="bibr" rid="B87">87</xref>) Notably, France also stands out with 19 publications (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B29">29</xref>, <xref ref-type="bibr" rid="B88">88</xref>&#x2013;<xref ref-type="bibr" rid="B103">103</xref>), and the United Kingdom (<xref ref-type="bibr" rid="B104">104</xref>&#x2013;<xref ref-type="bibr" rid="B121">121</xref>) and Japan (<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B102">102</xref>&#x2013;<xref ref-type="bibr" rid="B117">117</xref>) make considerable contributions with 18 and 17 publications, respectively. The data indicates that a large portion of publications, totaling 88, originate from Europe, with substantially fewer contributions from regions such as Latin America and Africa. Countries in East Asia, including Hong Kong, South Korea, Japan, Vietnam, and China, represent a significant proportion of contributions from that region.</p>
<fig id="F2" position="float"><label>Figure&#x00A0;2</label>
<caption><p>Number of publications by country. The country of publication is determined by the geographic location of the primary authors&#x0027; institution.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g002.tif"><alt-text content-type="machine-generated">World map highlighting the number of publications by country. The United States, Canada, China, and some European countries are colored. Dark purple indicates the highest number of publications, while yellow represents the lowest.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2d"><label>3.4</label><title>Distribution of publications by institution</title>
<p>A total of 370 institutions contributed to at least one article relevant to the production of the 217 included papers (<xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref>). The distribution of contributing institutions spans several countries, with notable contributions from France, Germany, the United States, the United Kingdom, and China. France&#x0027;s contributions were primarily driven by institutions like the University of Strasbourg (86 articles). German institutions like Furtwangen University (42 articles) and the German Cancer Research Center (DKFZ) (9 articles) also featured prominently. In the United States, Stanford University (17 articles) and Johns Hopkins University (16 articles), among others, were notable contributors. Additionally, it is relevant to note that two institutions with 11 articles were not included in the figure, namely the University of Texas Southwestern Medical Center and the Shanghai Jiao Tong University.</p>
<fig id="F3" position="float"><label>Figure&#x00A0;3</label>
<caption><p>Number of article contributions by institution, regarding the top twenty institutions in the field. An article contribution for an institution is counted if at least one of the authors on the source belongs to that respective institution.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g003.tif"><alt-text content-type="machine-generated">Bar chart showing article contributions from various institutions. University of Strasbourg has the highest contributions, around 100. Furtwangen University, The Chinese University of Hong Kong, and National Cancer Center Hospital East follow. Other institutions have fewer contributions, with Massachusetts General Hospital and German Cancer Center contributing the least.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2e"><label>3.5</label><title>Distribution of publications by author</title>
<p>A total of 974 authors have contributed to articles in this field, however, a few prominent authors emerge (<xref ref-type="table" rid="T3">Table&#x00A0;3</xref>). Padoy (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B53">53</xref>, <xref ref-type="bibr" rid="B69">69</xref>, <xref ref-type="bibr" rid="B73">73</xref>, <xref ref-type="bibr" rid="B88">88</xref>&#x2013;<xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B93">93</xref>, <xref ref-type="bibr" rid="B95">95</xref>, <xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B99">99</xref>, <xref ref-type="bibr" rid="B101">101</xref>, <xref ref-type="bibr" rid="B103">103</xref>, <xref ref-type="bibr" rid="B122">122</xref>) has the most at 17 publications from the University of Strasbourg, followed closely by Stoyanov (<xref ref-type="bibr" rid="B53">53</xref>, <xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B102">102</xref>, <xref ref-type="bibr" rid="B106">106</xref>, <xref ref-type="bibr" rid="B109">109</xref>, <xref ref-type="bibr" rid="B110">110</xref>, <xref ref-type="bibr" rid="B113">113</xref>, <xref ref-type="bibr" rid="B114">114</xref>, <xref ref-type="bibr" rid="B116">116</xref>&#x2013;<xref ref-type="bibr" rid="B121">121</xref>, <xref ref-type="bibr" rid="B123">123</xref>) from University College of London with 15. Next, Jalal (<xref ref-type="bibr" rid="B65">65</xref>, <xref ref-type="bibr" rid="B66">66</xref>, <xref ref-type="bibr" rid="B68">68</xref>, <xref ref-type="bibr" rid="B70">70</xref>, <xref ref-type="bibr" rid="B71">71</xref>, <xref ref-type="bibr" rid="B74">74</xref>&#x2013;<xref ref-type="bibr" rid="B77">77</xref>, <xref ref-type="bibr" rid="B80">80</xref>, <xref ref-type="bibr" rid="B82">82</xref>, <xref ref-type="bibr" rid="B86">86</xref>, <xref ref-type="bibr" rid="B117">117</xref>) from Furtwangen and Mutter (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B88">88</xref>, <xref ref-type="bibr" rid="B90">90</xref>, <xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B94">94</xref>, <xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B99">99</xref>, <xref ref-type="bibr" rid="B101">101</xref>, <xref ref-type="bibr" rid="B103">103</xref>) from Strasbourg, have 13 and 10 publications, respectively. Heng (<xref ref-type="bibr" rid="B25">25</xref>, <xref ref-type="bibr" rid="B63">63</xref>, <xref ref-type="bibr" rid="B106">106</xref>, <xref ref-type="bibr" rid="B124">124</xref>&#x2013;<xref ref-type="bibr" rid="B129">129</xref>). from the Chinese University of Hong Kong contributed 9 publications, while Jin (<xref ref-type="bibr" rid="B106">106</xref>, <xref ref-type="bibr" rid="B124">124</xref>, <xref ref-type="bibr" rid="B126">126</xref>, <xref ref-type="bibr" rid="B130">130</xref>&#x2013;<xref ref-type="bibr" rid="B133">133</xref>), Moeller (<xref ref-type="bibr" rid="B65">65</xref>, <xref ref-type="bibr" rid="B66">66</xref>, <xref ref-type="bibr" rid="B68">68</xref>, <xref ref-type="bibr" rid="B70">70</xref>, <xref ref-type="bibr" rid="B71">71</xref>, <xref ref-type="bibr" rid="B74">74</xref>, <xref ref-type="bibr" rid="B76">76</xref>, <xref ref-type="bibr" rid="B77">77</xref>), Mascagni, (<xref ref-type="bibr" rid="B53">53</xref>, <xref ref-type="bibr" rid="B88">88</xref>, <xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B95">95</xref>, <xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B99">99</xref>, <xref ref-type="bibr" rid="B122">122</xref>), Neumuth (<xref ref-type="bibr" rid="B65">65</xref>, <xref ref-type="bibr" rid="B68">68</xref>, <xref ref-type="bibr" rid="B71">71</xref>, <xref ref-type="bibr" rid="B74">74</xref>&#x2013;<xref ref-type="bibr" rid="B78">78</xref>), and Nwoye (<xref ref-type="bibr" rid="B69">69</xref>, <xref ref-type="bibr" rid="B88">88</xref>&#x2013;<xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B99">99</xref>, <xref ref-type="bibr" rid="B101">101</xref>) have 8 article contributions each.</p>
<table-wrap id="T3" position="float"><label>Table&#x00A0;3</label>
<caption><p>Number of publications for the ten most published authors.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Author</th>
<th valign="top" align="center">Publications</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Padoy N.</td>
<td valign="top" align="center">17</td>
</tr>
<tr>
<td valign="top" align="left">Stoyanov D.</td>
<td valign="top" align="center">15</td>
</tr>
<tr>
<td valign="top" align="left">Jalal N.A.</td>
<td valign="top" align="center">13</td>
</tr>
<tr>
<td valign="top" align="left">Mutter D.</td>
<td valign="top" align="center">10</td>
</tr>
<tr>
<td valign="top" align="left">Heng P.A.</td>
<td valign="top" align="center">9</td>
</tr>
<tr>
<td valign="top" align="left">Jin Y.</td>
<td valign="top" align="center">8</td>
</tr>
<tr>
<td valign="top" align="left">Mascagni P.</td>
<td valign="top" align="center">8</td>
</tr>
<tr>
<td valign="top" align="left">Moeller K.</td>
<td valign="top" align="center">8</td>
</tr>
<tr>
<td valign="top" align="left">Neumuth T.</td>
<td valign="top" align="center">8</td>
</tr>
<tr>
<td valign="top" align="left">Nwoye C.</td>
<td valign="top" align="center">8</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2f"><label>3.6</label><title>Distribution of citations by country</title>
<p>The analysis of the published studies showed that the countries with the highest number of citations were also the countries with the largest number of contributions (<xref ref-type="fig" rid="F4">Figure&#x00A0;4</xref>). France has more than 1,000 citations, making it the country with the highest citation count, followed by Hong Kong (424). United Kingdom and China have the same number of citations at 273, followed by Japan at 254. Most of the top 10 most cited countries have over 100 citations, however, South Korea (97) and Canada (60) fall below this.</p>
<fig id="F4" position="float"><label>Figure&#x00A0;4</label>
<caption><p>Number of citations by country. The country of citation is determined by the geographic location of the primary authors&#x0027; institution.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g004.tif"><alt-text content-type="machine-generated">World map showing citation counts by country, using a gradient from yellow (250 citations) to purple (1,000 citations). North America, selected European countries, China, and Japan have color-coded citations. Gray indicates no data.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2g"><label>3.7</label><title>Most frequently used keywords</title>
<p>The top five most occurring keywords were &#x201C;Laparoscopy&#x201D; (223), &#x201C;Deep Learning&#x201D; (206), &#x201C;Convolutional Neural Network&#x201D; (115), &#x201C;Surgical Equipment&#x201D; (101), and &#x201C;Robotic Surgery&#x201D; (96) (<xref ref-type="table" rid="T4">Table&#x00A0;4</xref>). Further, &#x201C;Videorecording&#x201D; (86) and &#x201C;Endoscopic Surgery&#x201D; (62) were frequently mentioned.</p>
<table-wrap id="T4" position="float"><label>Table&#x00A0;4</label>
<caption><p>Number of keyword occurrences in the included publications, showing the top ten.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Keyword</th>
<th valign="top" align="center">Occurrences</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Laparoscopy</td>
<td valign="top" align="center">223</td>
</tr>
<tr>
<td valign="top" align="left">Deep Learning</td>
<td valign="top" align="center">206</td>
</tr>
<tr>
<td valign="top" align="left">Convolutional Neural Network</td>
<td valign="top" align="center">115</td>
</tr>
<tr>
<td valign="top" align="left">Surgical Equipment</td>
<td valign="top" align="center">101</td>
</tr>
<tr>
<td valign="top" align="left">Robotic Surgery</td>
<td valign="top" align="center">96</td>
</tr>
<tr>
<td valign="top" align="left">Videorecording</td>
<td valign="top" align="center">86</td>
</tr>
<tr>
<td valign="top" align="left">Endoscopic Surgery</td>
<td valign="top" align="center">62</td>
</tr>
<tr>
<td valign="top" align="left">Cataract Surgery</td>
<td valign="top" align="center">61</td>
</tr>
<tr>
<td valign="top" align="left">Transplant Surgery</td>
<td valign="top" align="center">52</td>
</tr>
<tr>
<td valign="top" align="left">Workflow</td>
<td valign="top" align="center">48</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2h"><label>3.8</label><title>Emerging trends in keywords over time</title>
<p>The timeline (<xref ref-type="fig" rid="F5">Figure&#x00A0;5A</xref>) clearly shows a growing trend in research activity of the keywords, especially from 2017 onwards, which is consistent with the number of article publications displayed in <xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>. For example, mentions of &#x201C;deep learning&#x201D; increased from just 3 occurrences in 2017 to 74 in 2021, reaching 206 by the end of 2023. Similarly, &#x201C;laparoscopy&#x201D; grew significantly during the same period. <xref ref-type="fig" rid="F5">Figure&#x00A0;5B</xref> illustrates the total occurrences of the top keywords. Although &#x201C;Laparoscopy&#x201D; dominates in frequency, other specific surgeries categorized under it, such as &#x201C;Endoscopic Surgery,&#x201D; &#x201C;Transplant Surgery,&#x201D; and &#x201C;Cataract Surgery,&#x201D; are also frequently found.</p>
<fig id="F5" position="float"><label>Figure&#x00A0;5</label>
<caption><p><bold>(A)</bold> Total number of keyword occurrences in the included publications from 2017 to 2024. <bold>(B)</bold> Total number of keyword occurrences in the included publications from 2017 to 2024.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g005.tif"><alt-text content-type="machine-generated">Line chart and radar chart display keyword occurrences from 2017 to 2024. The line chart shows an upward trend in laparoscopic surgery and other medical-related terms. The radar chart highlights the total values, with laparoscopic surgery having the highest occurrence. Keywords include deep learning, robotic surgery, and endoscopic surgery.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2i"><label>3.9</label><title>Distribution of publications by journal</title>
<p>Although there were 93 sources in the form of journals and conference proceedings, a few specific journals stood out in terms of total number of publications. The journal with the highest number of articles published was the International Journal of Computer Assisted Radiology and Surgery (<xref ref-type="bibr" rid="B24">24</xref>, <xref ref-type="bibr" rid="B36">36</xref>, <xref ref-type="bibr" rid="B38">38</xref>, <xref ref-type="bibr" rid="B41">41</xref>, <xref ref-type="bibr" rid="B54">54</xref>, <xref ref-type="bibr" rid="B56">56</xref>, <xref ref-type="bibr" rid="B67">67</xref>, <xref ref-type="bibr" rid="B73">73</xref>, <xref ref-type="bibr" rid="B78">78</xref>, <xref ref-type="bibr" rid="B83">83</xref>, <xref ref-type="bibr" rid="B90">90</xref>, <xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B101">101</xref>, <xref ref-type="bibr" rid="B109">109</xref>, <xref ref-type="bibr" rid="B110">110</xref>, <xref ref-type="bibr" rid="B117">117</xref>, <xref ref-type="bibr" rid="B119">119</xref>, <xref ref-type="bibr" rid="B126">126</xref>, <xref ref-type="bibr" rid="B134">134</xref>&#x2013;<xref ref-type="bibr" rid="B144">144</xref>), with 27 articles and an impact factor (IF) of 2.3. Lecture Notes in Computer Science (<xref ref-type="bibr" rid="B25">25</xref>, <xref ref-type="bibr" rid="B58">58</xref>, <xref ref-type="bibr" rid="B69">69</xref>, <xref ref-type="bibr" rid="B84">84</xref>, <xref ref-type="bibr" rid="B99">99</xref>, <xref ref-type="bibr" rid="B103">103</xref>, <xref ref-type="bibr" rid="B104">104</xref>, <xref ref-type="bibr" rid="B118">118</xref>, <xref ref-type="bibr" rid="B121">121</xref>, <xref ref-type="bibr" rid="B130">130</xref>, <xref ref-type="bibr" rid="B145">145</xref>&#x2013;<xref ref-type="bibr" rid="B153">153</xref>) came next with 19 articles published followed by Medical Image Analysis (<xref ref-type="bibr" rid="B88">88</xref>, <xref ref-type="bibr" rid="B89">89</xref>, <xref ref-type="bibr" rid="B91">91</xref>, <xref ref-type="bibr" rid="B93">93</xref>, <xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B102">102</xref>, <xref ref-type="bibr" rid="B113">113</xref>, <xref ref-type="bibr" rid="B124">124</xref>, <xref ref-type="bibr" rid="B127">127</xref>, <xref ref-type="bibr" rid="B154">154</xref>, <xref ref-type="bibr" rid="B155">155</xref>) with a high IF of 10.9. Surgical Endoscopy (<xref ref-type="bibr" rid="B31">31</xref>, <xref ref-type="bibr" rid="B40">40</xref>, <xref ref-type="bibr" rid="B42">42</xref>, <xref ref-type="bibr" rid="B50">50</xref>, <xref ref-type="bibr" rid="B53">53</xref>, <xref ref-type="bibr" rid="B95">95</xref>, <xref ref-type="bibr" rid="B156">156</xref>&#x2013;<xref ref-type="bibr" rid="B159">159</xref>) and IEEE Transactions on Medical Imaging (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>, <xref ref-type="bibr" rid="B32">32</xref>, <xref ref-type="bibr" rid="B34">34</xref>, <xref ref-type="bibr" rid="B106">106</xref>, <xref ref-type="bibr" rid="B114">114</xref>, <xref ref-type="bibr" rid="B120">120</xref>, <xref ref-type="bibr" rid="B125">125</xref>, <xref ref-type="bibr" rid="B160">160</xref>, <xref ref-type="bibr" rid="B161">161</xref>) both had 10 publications; however, they had IFs of 2.4 and 10.6, respectively. The next 5 journals all had fewer than 10 articles in each and IFs of less than 5.</p>
</sec>
<sec id="s2j"><label>3.10</label><title>Top cited papers</title>
<p>The ten most cited articles using DL for surgical tool detection in MIS is shown in <xref ref-type="table" rid="T5">Table&#x00A0;5</xref>. Most of the highly cited articles are published by IEEE, with a few contributions published by Elsevier, JAMA and Springer. The most cited paper is &#x201C;EndoNet: A Deep Architecture for Recognition Tasks on Laparoscopic Videos,&#x201D; (<xref ref-type="bibr" rid="B6">6</xref>) authored by Twinanda (2017, 924 citations) which is published in IEEE. The paper proposed a CNN model performs surgical tools classification and workflow recognition. This paper is one of the first papers in the field of instruments segmentation using DL model. Shvets&#x0027; &#x201C;Automatic Instrument Segmentation in Robot-Assisted Surgery using Deep Learning&#x201D; (<xref ref-type="bibr" rid="B60">60</xref>) (2018, 433 citations), which explored binary, semantic and tools&#x2019; parts segmentation using CNN blocks in an encoder-decoder fashion. The model was more specific compared to other models as it was developed for robotic-assisted surgeries. Jin&#x0027;s 2018 article, &#x201C;Tool Detection and Operative Skill Assessment in Surgical Videos Using Region-Based Convolutional Neural Networks&#x201D; (<xref ref-type="bibr" rid="B61">61</xref>), with 327 citations, utilizes instrument segmentation to assess the surgeons&#x0027; skill level for training purposes.</p>
<table-wrap id="T5" position="float"><label>Table&#x00A0;5</label>
<caption><p>The ten most cited articles in the field, alongside their year of publication, primary author, total citations, and publisher.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="center"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Title</th>
<th valign="top" align="center">Year</th>
<th valign="top" align="center">Primary author</th>
<th valign="top" align="center">Total citations</th>
<th valign="top" align="center">Publisher</th>
<th valign="top" align="center">Application</th>
<th valign="top" align="center">Dl algorithms</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">EndoNet: A Deep Architecture for Recognition Tasks on Laparoscopic Videos (<xref ref-type="bibr" rid="B6">6</xref>)</td>
<td valign="top" align="left">(2017)</td>
<td valign="top" align="left">Twinada</td>
<td valign="top" align="center">924</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Tools classifications and workflow recognition</td>
<td valign="top" align="left">CNNs</td>
</tr>
<tr>
<td valign="top" align="left">Automatic Instrument Segmentation in Robot-Assisted Surgery using Deep Learning (<xref ref-type="bibr" rid="B60">60</xref>)</td>
<td valign="top" align="left">(2018)</td>
<td valign="top" align="left">Shvets</td>
<td valign="top" align="center">433</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Binary, semantic, tools&#x2019; parts segmentation</td>
<td valign="top" align="left">CNNs with encoder decoder architecture</td>
</tr>
<tr>
<td valign="top" align="left">Tool Detection and Operative Skill Assessment in Surgical Videos Using Region-Based Convolutional Neural Networks (<xref ref-type="bibr" rid="B61">61</xref>)</td>
<td valign="top" align="left">(2018)</td>
<td valign="top" align="left">Jin</td>
<td valign="top" align="center">327</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Multiclass classification and detection using bounding box</td>
<td valign="top" align="left">CNNs</td>
</tr>
<tr>
<td valign="top" align="left">SV-RCNet: Workflow Recognition From Surgical Videos Using Recurrent Convolutional Network (<xref ref-type="bibr" rid="B7">7</xref>)</td>
<td valign="top" align="left">(2018)</td>
<td valign="top" align="left">Jin</td>
<td valign="top" align="center">274</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Classification of instruments and workflow recognition</td>
<td valign="top" align="left">Recurrent convolution neural networks</td>
</tr>
<tr>
<td valign="top" align="left">Multi-task recurrent convolutional network with correlation loss for surgical video analysis (<xref ref-type="bibr" rid="B127">127</xref>)</td>
<td valign="top" align="left">(2020)</td>
<td valign="top" align="left">Jin</td>
<td valign="top" align="center">179</td>
<td valign="top" align="left">El Selvier ScienceDirect</td>
<td valign="top" align="left">Workflow recognition and instrument classification</td>
<td valign="top" align="left">Recurrent convolutional network, spatio-temporal features, very deep residual network, long short-term memory.</td>
</tr>
<tr>
<td valign="top" align="left">Machine and deep learning for workflow recognition during surgery (<xref ref-type="bibr" rid="B162">162</xref>)</td>
<td valign="top" align="left">(2019)</td>
<td valign="top" align="left">Padoy</td>
<td valign="top" align="center">148</td>
<td valign="top" align="left">Taylor and Francis</td>
<td valign="top" align="left">Surgical workflow recognition and tool detection</td>
<td valign="top" align="left">CNNs and LSTM</td>
</tr>
<tr>
<td valign="top" align="left">Weakly supervised convolutional LSTM approach for tool tracking in laparoscopic videos (<xref ref-type="bibr" rid="B101">101</xref>)</td>
<td valign="top" align="left">(2019)</td>
<td valign="top" align="left">Nwoye</td>
<td valign="top" align="center">144</td>
<td valign="top" align="left">Springer Link</td>
<td valign="top" align="left">Real-time binary surgical tool tracking</td>
<td valign="top" align="left">CNN&#x2009;&#x002B;&#x2009;Convolutional LSTM (ConvLSTM)</td>
</tr>
<tr>
<td valign="top" align="left">Articulated Multi-Instrument 2-D Pose Estimation Using Fully Convolutional Networks (<xref ref-type="bibr" rid="B120">120</xref>)</td>
<td valign="top" align="left">(2018)</td>
<td valign="top" align="left">Du</td>
<td valign="top" align="center">140</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Instruments joints detection</td>
<td valign="top" align="left">FCN</td>
</tr>
<tr>
<td valign="top" align="left">Deep Learning Based Robotic Tool Detection and Articulation Estimation With Spatio-Temporal Layers (<xref ref-type="bibr" rid="B163">163</xref>)</td>
<td valign="top" align="left">(2019)</td>
<td valign="top" align="left">Colleoni</td>
<td valign="top" align="center">115</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Surgical instrument joint detection and localization</td>
<td valign="top" align="left">Three dimensional CNN</td>
</tr>
<tr>
<td valign="top" align="left">Assessment of Automated Identification of Phases in Videos of Cataract Surgery Using Machine Learning and Deep Learning Techniques (<xref ref-type="bibr" rid="B164">164</xref>)</td>
<td valign="top" align="left">(2019)</td>
<td valign="top" align="left">Yu</td>
<td valign="top" align="center">106</td>
<td valign="top" align="left">JAMA</td>
<td valign="top" align="left">Instrument classification and workflow recognition</td>
<td valign="top" align="left">CNN-RNN</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="TF1"><p>The number of citations were verified using Google Scholar.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>Moreover, Du&#x0027;s &#x201C;Articulated Multi-Instrument 2-D Pose Estimation Using Fully Convolutional Networks&#x201D; (<xref ref-type="bibr" rid="B120">120</xref>) (2018, 40 citations), focuses on pose estimation in laparoscopic surgery, relying on segmentation for accurate tool positioning. The accurate positioning was implemented through the detection of instruments joints. Colleoni has also developed a three-dimensional CNN that detects and localize surgical tools joints, published under &#x201C;Deep Learning Based Robotic Tool Detection and Articulation Estimation With Spatio-Temporal Layers&#x201D; (2019, 115 citations). Jin&#x0027;s (<xref ref-type="bibr" rid="B163">163</xref>) &#x201C;Multi-task recurrent convolutional network with correlation loss for surgical video analysis&#x201D; (<xref ref-type="bibr" rid="B127">127</xref>) (2020, 179 citations), and Yu&#x0027;s &#x201C;Assessment of Automated Identification of Phases in Videos of Cataract Surgery Using Machine Learning and Deep Learning Techniques&#x201D; (<xref ref-type="bibr" rid="B164">164</xref>) (2019, 106 citations) utilized surgical tool detection in workflow recognition. Nwoye&#x0027;s &#x201C;Weakly supervised convolutional LSTM approach for tool tracking in laparoscopic videos&#x201D; (<xref ref-type="bibr" rid="B101">101</xref>) (2019, 144 citations) utilized instrument segmentation to achieve accurate and reliable tool detection and tracking.</p>
</sec>
<sec id="s2k"><label>3.11</label><title>Top recent cited papers</title>
<p>The 3 most cited articles from 2020 onwards in surgical video analysis are shown in <xref ref-type="table" rid="T6">Table&#x00A0;6</xref>. Out of the 3 articles, 2 were published in ScienceDirect Elsevier and 1 in IEEE. The most highly cited article, Jin, &#x201C;Multi-task recurrent convolutional network with correlation loss for surgical video analysis,&#x201D; (<xref ref-type="bibr" rid="B127">127</xref>) was published by ScienceDirect (Elsevier), with 179 citations. The paper talks about surgical video analysis, focusing on tool detection and phase recognition. Jin and his co-authors were able to achieve such task through the utilization of CNN and RNNs to maintain the temporal elements. Next, Kitaguchi published the second most cited article during the last five years at 99, titled &#x201C;Automated laparoscopic colorectal surgery workflow recognition using artificial intelligence&#x201D; (<xref ref-type="bibr" rid="B165">165</xref>). The paper introduces a new dataset &#x201C;LapSig300&#x201D; which encompasses 300 intraoperative videos that were collected from 19 high-volume centers. This novel work describes the dataset components such as surgical workflows that were classified into nine phases, 3 actions and 5 tools. Other notable contributions during this recent timeframe include Jin&#x0027;s (2021) &#x201C;Temporal Memory Relation Network for Workflow Recognition&#x201D; (<xref ref-type="bibr" rid="B125">125</xref>) with 81 citations. The work presents a novel network that detects the surgical phase based on the presence of surgical tools by utilizing LSTM that preserves temporal information.</p>
<table-wrap id="T6" position="float"><label>Table&#x00A0;6</label>
<caption><p>The three most cited articles in the field from 2020 onwards, alongside their year of publication, primary author, total citations, and publisher.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="center"/>
<col align="center"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Title</th>
<th valign="top" align="center">Year</th>
<th valign="top" align="center">Primary author</th>
<th valign="top" align="center">Total citations</th>
<th valign="top" align="center">Publisher</th>
<th valign="top" align="center">Application</th>
<th valign="top" align="center">DL Algorithms</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Multi-task recurrent convolutional network with correlation loss for surgical video analysis (<xref ref-type="bibr" rid="B127">127</xref>)</td>
<td valign="top" align="left">(2020)</td>
<td valign="top" align="center">Jin</td>
<td valign="top" align="center">179</td>
<td valign="top" align="left">ScienceDirect ElSevier</td>
<td valign="top" align="left">Tool detection and phase recognition tasks</td>
<td valign="top" align="left">RNN and CNN</td>
</tr>
<tr>
<td valign="top" align="left">Automated laparoscopic colorectal surgery workflow recognition using artificial intelligence: Experimental research (<xref ref-type="bibr" rid="B165">165</xref>)</td>
<td valign="top" align="left">(2020)</td>
<td valign="top" align="center">Kitaguchi</td>
<td valign="top" align="center">99</td>
<td valign="top" align="left">ScienceDirect ElSevier</td>
<td valign="top" align="left">Tool detection, segmentation and surgical phase recognition</td>
<td valign="top" align="left">CNN</td>
</tr>
<tr>
<td valign="top" align="left">Temporal Memory Relation Network for Workflow Recognition From Surgical Video (<xref ref-type="bibr" rid="B125">125</xref>)</td>
<td valign="top" align="left">(2021)</td>
<td valign="top" align="center">Jin</td>
<td valign="top" align="center">81</td>
<td valign="top" align="left">IEEE</td>
<td valign="top" align="left">Tool presence and phase recognition</td>
<td valign="top" align="left">ResNet and LSTM</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="TF2"><p>The number of citations were verified using Google Scholar.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>Additionally, a growing subset of recent papers explores Transformer-based models and multimodal learning approaches, signaling a shift from purely convolutional or recurrent architectures toward methods capable of integrating video, textual, and even sensor data for more robust surgical context understanding.</p>
</sec>
</sec>
<sec id="s3" sec-type="discussion"><label>4</label><title>Discussion</title>
<p>This bibliometric analysis demonstrates sustained growth and increasing global engagement in deep learning methods for surgical instrument perception in MIS. Instrument detection, segmentation, and tracking are key enabling technologies for broader surgical data science applications, including workflow recognition, skill assessment, and context-aware intraoperative assistance. The literature initially focused on CNN-based approaches and benchmark challenge datasets (e.g., EndoNet-era laparoscopic video recognition and the EndoVis series), but recent work increasingly explores transformer architectures and large-scale pretrained models. For example, Segment Anything&#x2013;style models have begun to be adapted for robotic instrument segmentation [e.g., Surgical-DeSA M (<xref ref-type="bibr" rid="B10">10</xref>)], highlighting the emerging role of foundation-model paradigms in MIS. At the same time, progress remains dependent on high-quality annotated datasets and standardized benchmarking (e.g., EndoVis, HeiChole, Cholec80, CaDIS), alongside rigorous evaluation under domain shift, occlusion, smoke/bleeding artifacts, and strict real-time constraints (<xref ref-type="bibr" rid="B113">113</xref>). Overall, these trends suggest a maturing field moving toward more generalizable and clinically deployable solutions.</p>
<p>The sustained research interest seen in 2024 suggests that this field will likely continue to keep evolving, bringing further innovations to clinical practices. The field of DL in MIS is rapidly advancing, with a diverse range of countries, institutions, and authors contributing to its growth. Nevertheless, researchers are collaborating not only on a local level between institutions but also from different continents. The increasing number of publications and citations reflects the high relevance of this research in enhancing surgical outcomes. As more collaborative efforts arise and the technology advances, DL has the potential to transform surgical practices, enhancing both training and the precision and safety of procedures. The geographic distribution of publications shows that many countries are making significant contributions. There is a noticeable difference in how productive and impactful these contributions are</p>
<sec id="s3a"><label>4.1</label><title>Leading countries in the field</title>
<p>A comparative analysis of the total number of citations and the average citations per article offers valuable information into the balance between research volume and individual article influence in each country. <xref ref-type="fig" rid="F6">Figure&#x00A0;6</xref> highlights France as an outlier, with the highest citation count despite not leading in the number of publications. This indicates that France produces a high volume of influential research as the average citations per article is high (<xref ref-type="fig" rid="F7">Figure&#x00A0;7</xref>). Hong Kong follows with fewer total citations compared to France, but still significant in number. Importantly, Hong Kong stands out with the highest average citations per article, indicating that while the number of publications may be lower, the impact of each publication is very high. This shows that the research from institutions in Hong Kong is highly targeted and impactful within its academic field, possibly reflecting a focus on quality over quantity. The UK shows a moderate number of total citations, only slightly behind Hong Kong. Its average citations per article are similarly high, reflecting that although the influence may be less, it is consistent throughout publications. China and Japan display similar trends with moderate total citations. Interestingly, their average citations per article are lower, implying that while these countries may publish frequently, the individual impact of each publication tends to be smaller. Despite the large quantities of research, the overall influence of each article may not be as pronounced.</p>
<fig id="F6" position="float"><label>Figure&#x00A0;6</label>
<caption><p>Comparison of the total number of citations and the average citations per article across various countries in DL research for surgical instrument segmentation in MIS. The gray bars represent the total number of citations (left axis), while the blue bars indicate the average citations per article (right axis).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g006.tif"><alt-text content-type="machine-generated">Bar chart comparing the number of total citations and average citations per article across ten countries: France, Hong Kong, United Kingdom, China, Japan, United States, Germany, Italy, South Korea, and Canada. Total citations are denoted by blue bars, while average citations per article are in gray. France, Hong Kong, and United Kingdom have the highest total citations, while China, Japan, and United States show significant average citations per article.</alt-text>
</graphic>
</fig>
<fig id="F7" position="float"><label>Figure&#x00A0;7</label>
<caption><p>Total citation vs. number of publications for each country.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g007.tif"><alt-text content-type="machine-generated">Scatter plot showing \"Publications vs Citations by Country.\" The y-axis represents total citations, and the x-axis indicates the number of publications. Countries like the United States and China have many publications but varying citations, while France has high citations with fewer publications. Other countries like Hong Kong, Japan, and Germany vary in both citations and publications. Each country is marked with a blue dot labeled accordingly.</alt-text>
</graphic>
</fig>
<p>Despite the reputation of the US in global research, the total number of citations is somewhat lower than expected, considering it has the highest number of publications per country. It has a modest average citation rate per article, showing that the country produces a steady stream of research, but its per-article influence is not as strong as other countries. This may warrant a deeper look into the quality or visibility of US publications in this research area. With both a low total citation count and a modest average citation rate, Germany&#x0027;s overall contribution appears less prominent. However, its consistent output may indicate steady participation in the field without significant high-impact publications. The citations on publications from Italy portray an interesting pattern. Though not among the highest in total citations, Italy&#x0027;s publications have a strong average citation per article, suggesting that its research, while less frequent, tends to be highly impactful. This suggests a focus on producing fewer but higher-quality or more widely recognized studies. Finally, both South Korea and Canada are on the lower end in terms of total citations and average citations per article. This shows that their contributions are either newer or less widely recognized in this field. This could also reflect limited participation or a developing research presence in the field.</p>
<p>However, on an interpretative note, we need to be cognizant of the fact that citation-based indicators reflect visibility and uptake of publications and are influenced by publication age, venue, and citation practices. Accordingly, differences in total citations and average citations per article should not be interpreted as direct measures of methodological quality or clinical effectiveness.</p>
</sec>
<sec id="s3b"><label>4.2</label><title>Country co-authorship</title>
<p>In terms of geographical distribution, the United States occupies a central position in the co-authorship network, characterized by a high number of international collaboration links. Many countries demonstrate frequent co-authorship with the United States, reflecting its central role in the network. France, despite a lower publication volume, shows a comparatively higher average citation rate per article, suggesting a strong impact of its research output. China has also shown rapid growth in both publications and citations, indicating its increasing influence in the field.</p>
<p>Further, institutions such as the University of Strasbourg, Furtwangen University, and Johns Hopkins University are at the forefront of DL research in MIS. This highlights how these institutions can drive the collection of information. Institutions from Europe, North America, and Asia account for the majority of publications, indicating broad geographic representation across these regions. However, collaboration involving institutions from underrepresented regions appears less frequent in the co-authorship network. Additionally, author and institutional collaboration networks demonstrate frequent connections within certain clusters, particularly between European, American, and East Asian countries. Authors like Stoyanov, Padoy, and Heng are central figures in this collaborative environment, driving forward impactful research. There is no doubt still room for growth in terms of incorporating more diverse contributors from underrepresented regions, which could lead to new perspectives and novel advancements.</p>
<p>The collaboration network in these publications, which focuses on several countries, displays a thorough, but unevenly distributed, structure of global cooperation (<xref ref-type="fig" rid="F8">Figure&#x00A0;8</xref>). The United States demonstrates the highest number of co-authorship links with countries in Europe and East Asia. Several European countries, including the United Kingdom, Germany, and France, show frequent co-authorship links with institutions based in the United States. Similarly, East Asian countries like China, Japan, and South Korea also exhibit strong collaborative research both within the region and with the US. Furthermore, while other countries, like India, Australia, and Canada make significant contributions, their collaborative networks are not as extensive or dense as those seen in Europe and East Asia, indicating that there is room for growth and integration.</p>
<fig id="F8" position="float"><label>Figure&#x00A0;8</label>
<caption><p>Co-authorship collaboration network of publications by country. The size of each country&#x0027;s circle represents the number of publications they have contributed, and the thickness of the lines represents the number of collaborations between the countries. The minimum number of article contributions for a country to be included in this network was set to 2. VOSviewer was used in the production of this figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g008.tif"><alt-text content-type="machine-generated">Network visualization showing international connections between various countries, represented by nodes in different colors. Larger nodes like United States, China, and Germany indicate more connections. Lines illustrate the relationships between nodes, highlighting countries like the United Kingdom, Japan, and France as key connectors.</alt-text>
</graphic>
</fig>
<p>Despite active research clusters in Europe, North America, and East Asia, participation from regions such as Africa and Latin America remains limited, highlighting the need for more inclusive global collaborations and funding initiatives to broaden the field&#x0027;s impact. There are a variety of reasons why certain authors or countries may be isolated. It is possible that the author&#x0027;s work might not be widely disseminated or accessible to the broader research community. This can be the result of publication in less prominent journals, or even limited participation in international conferences and networks. Additionally, if the author&#x0027;s work spans across multiple disciplines, it might not fit neatly into the established research networks of any single field, leading to isolation within both publication and citation networks. Further, geographical or institutional factors might limit the author&#x0027;s opportunities for collaboration. Over time, as the field develops, there may be an increase in both collaborations and citations across authors from different countries and institutions over time.</p>
</sec>
<sec id="s3c"><label>4.3</label><title>Author collaboration and co-citation networks</title>
<p>The co-citation network represents the relationships between authors based on how often they are cited together in these scientific publications (<xref ref-type="fig" rid="F9">Figure&#x00A0;9</xref>). Padoy appears to be a central figure with a lot of connections, which suggests that he is highly influential and widely cited across different studies. Additionally, Stoyanov and Jannin also appear to be very important in their clusters, demonstrating substantial impact and regular co-citation with other authors. Some authors like Jalal, who were found to be isolated in publications networks, were also found to be isolated in the citation networks. Inter-cluster connections suggest that influential authors are not only leaders within their specific areas but also contribute to interdisciplinary research, bridging gaps between different fields and fostering collaboration.</p>
<fig id="F9" position="float"><label>Figure&#x00A0;9</label>
<caption><p>Author co-citation network of the included publications. The size of each author&#x0027;s circle represents the number of citations they have received, and the thickness of the lines represents the number of co-citations between the authors. The minimum number of citations for an author to be included in this network was set to 10. VOSviewer was used in the production of this figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g009.tif"><alt-text content-type="machine-generated">Network visualization showing connections between authors using colored nodes and lines. Nodes represent authors, sized by influence, with colors indicating different clusters: green, red, blue, and yellow. Lines denote collaborations. Generated with VOSviewer.</alt-text>
</graphic>
</fig>
<p>Stoyanov plays a central role in collaborations, followed by Padoy and Heng, who act as bridges between various researchers (<xref ref-type="fig" rid="F10">Figure 10</xref>). Author collaborations indicate high-impact research, with broadening opportunities for further collaboration. Given his extensive connections with other scholars, Stoyanov appears to have a crucial role in promoting cooperation both inside and outside of his cluster. Moreover, Padoy and Heng exhibit a noteworthy collaborative influence, serving as a bridge between various researchers. Certain researchers, like Loukas and Ito, seem to be more isolated with fewer connections.</p>
<fig id="F10" position="float"><label>Figure&#x00A0;10</label>
<caption><p>Co-authorship collaboration network of the publications in the field. The size of each author&#x0027;s circle represents the number of publications they have contributed, and the thickness of the lines represents the number of collaborations between the authors. The minimum number of article contributions for an author to be included in this network was set to 3. VOSviewer was used in the production of this figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g010.tif"><alt-text content-type="machine-generated">A network visualization showing clusters of nodes labeled with names, connected by lines. The clusters are color-coded: green, purple, red, blue, and yellow, representing different groups or themes. Each cluster contains several nodes connected within and to other clusters. The visualization is created using VOSviewer, indicated by the logo in the bottom left corner.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3d"><label>4.4</label><title>Keyword co-occurrence analysis</title>
<p>The frequent occurrence of terms like &#x201C;Deep Learning,&#x201D; &#x201C;Laparoscopy,&#x201D; and &#x201C;Convolutional Neural Networks&#x201D; highlights the focus on laparoscopic surgeries as a key area for DL (<xref ref-type="table" rid="T4">Table&#x00A0;4</xref>). From these keywords, it is evident that the use of video data as opposed to images for training DL models is most common. Additionally, most DL models utilize CNNs in their architectures, given that it is the most frequent algorithm in the keyword analysis. This indicates that they have the DL algorithm with the most extensive application in image recognition and segmentation tasks. Additionally, <xref ref-type="table" rid="T5">Tables&#x00A0;5</xref>, <xref ref-type="table" rid="T7">7</xref> illustrate CNNs&#x0027; popularity as they are widely utilized by the top cited papers in the field. The growing trend in the use of video-based data for DL training, as opposed to static images, is indicative of the shift towards its applications in surgery. This suggests that the future of DL in MIS may lie in improving intraoperative decision-making by providing real-time analysis and feedback to surgeons. This significant increase also aligns with the rapid technological advancements in DL algorithms and their increasing use in medical imaging and surgical robotics. It is clear that this is a developing field, as indicated by the rapid growth of publications and keyword use. By analyzing keywords in these research papers, important information can be acquired about research trends and focus areas in the field of DL use in MIS.</p>
<table-wrap id="T7" position="float"><label>Table&#x00A0;7</label>
<caption><p>Top ten journals with the highest number of articles published in the field, alongside their impact factors from the 2022 journal impact factors by <ext-link ext-link-type="uri" xlink:href="https://annualreviews.org">annualreviews.org</ext-link>.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="center"/>
<col align="center"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Source</th>
<th valign="top" align="center">Number of Articles</th>
<th valign="top" align="center">IF 2022</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">International Journal of Computer Assisted Radiology and Surgery</td>
<td valign="top" align="center">27</td>
<td valign="top" align="center">2.3</td>
</tr>
<tr>
<td valign="top" align="left">Lecture Notes in Computer Science</td>
<td valign="top" align="center">19</td>
<td valign="top" align="center">1.1</td>
</tr>
<tr>
<td valign="top" align="left">Medical Image Analysis</td>
<td valign="top" align="center">11</td>
<td valign="top" align="center">10.9</td>
</tr>
<tr>
<td valign="top" align="left">IEEE Transactions on Medical Imaging</td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">10.6</td>
</tr>
<tr>
<td valign="top" align="left">Surgical Endoscopy</td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">2.4</td>
</tr>
<tr>
<td valign="top" align="left">Current Directions in Biomedical Engineering</td>
<td valign="top" align="center">9</td>
<td valign="top" align="center">0.5</td>
</tr>
<tr>
<td valign="top" align="left">IEEE Robotics and Automation Letters</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">4.6</td>
</tr>
<tr>
<td valign="top" align="left">International Journal of Medical Robotics and Computer Assisted Surgery</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">2.3</td>
</tr>
<tr>
<td valign="top" align="left">Computer Methods in Biomechanics and Biomedical Engineering: Imaging and Visualization</td>
<td valign="top" align="center">4</td>
<td valign="top" align="center">1.3</td>
</tr>
<tr>
<td valign="top" align="left">Computer Methods and Programs in Biomedicine</td>
<td valign="top" align="center">3</td>
<td valign="top" align="center">4.9</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="TF3"><p>Conference proceedings were excluded from this table.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>The keyword co-occurrence network in <xref ref-type="fig" rid="F11">Figure&#x00A0;11</xref> provides insights into key technical terms and their relationships in DL. The central term, &#x201C;DL,&#x201D; is closely linked with &#x201C;image segmentation,&#x201D; &#x201C;computer vision,&#x201D; &#x201C;CNN,&#x201D; and &#x201C;AI,&#x201D; reflecting the focus on DL techniques for image analysis in surgery. The green cluster highlights terms such as &#x201C;image segmentation,&#x201D; &#x201C;image processing,&#x201D; and &#x201C;semantics,&#x201D; which highlights some of the most commonly used techniques by DL algorithms in the processing of surgical videos. These models are critical for accurately tracking surgical instruments and automating workflow recognition. The blue cluster features &#x201C;CNN&#x201D; and &#x201C;LSTM&#x201D; (long short-term memory networks), suggesting that both spatial and temporal features are handled through a combination of neural networks. These tools enhance real-time surgical decision-making by identifying various phases of surgery. Lastly, the purple cluster emphasizes &#x201C;object recognition&#x201D; and &#x201C;instrument detection,&#x201D; addressing the importance of these tasks in developing a function DL algorithm. Recent works are using transformers, however, as it is till new to the field of surgical video analysis it less frequent given that the timeframe is since 2017, where CNNs were dominating the field.</p>
<fig id="F11" position="float"><label>Figure&#x00A0;11</label>
<caption><p>Data science and deep learning keyword co-occurrence network of the included publications. The size of each keyword&#x0027;s circle represents the number of publications they have occurred in, and the thickness of the lines represents the number of co-occurrences between the keywords. The minimum number of occurrences for a keyword to be included in this network was set to 5. VOSviewer was used in the production of this figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g011.tif"><alt-text content-type="machine-generated">A network visualization map shows interconnections among terms related to deep learning (dl), convolutional neural networks (cnn), and various computational topics like computer vision and image segmentation. Nodes are colored differently to represent clusters of related terms, with lines indicating their relationships.</alt-text>
</graphic>
</fig>
<p>The keyword co-occurrence network in <xref ref-type="fig" rid="F12">Figure&#x00A0;12</xref> is more focused on surgical and medical terminology, illustrating their relationships to MIS. The central node, &#x201C;surgical equipment,&#x201D; is closely connected to multiple terms such as &#x201C;robotic surgery,&#x201D; &#x201C;endoscopic surgery,&#x201D; and &#x201C;laparoscopic cholecystectomy,&#x201D; indicating the DL algorithms used had a particular focus on instrumentation in all these types of surgeries. The green cluster, which includes terms like &#x201C;surgical instrument&#x201D; and &#x201C;surgical tools,&#x201D; highlights the focus on instrument development and refinement, which is essential for enhancing precision and outcomes in MIS procedures. The connection to &#x201C;transplant surgery&#x201D; shows that advancements in surgical tools are not just limited to one type of surgery but also seem to have broad applications across different fields. In the red cluster, terms such as &#x201C;medical imaging&#x201D; and &#x201C;neurosurgery&#x201D; emphasize the integration of imaging techniques in surgeries, especially for complex procedures like brain surgery. This cluster demonstrates the critical role of imaging in guiding surgical procedures and improving surgical accuracy. The link to &#x201C;surgical phase recognition&#x201D; suggests the growing importance of recognizing and automating phases of surgery using advanced tools and DL models.</p>
<fig id="F12" position="float"><label>Figure&#x00A0;12</label>
<caption><p>Surgery related keyword co-occurrence network of the included publications. The size of each keyword&#x0027;s circle represents the number of publications they have occurred in, and the thickness of the lines represents the number of co-occurrences between the keywords. The minimum number of occurrences for a keyword to be included in this network was set to 5. VOSviewer was used in the production of this figure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fdgth-08-1633888-g012.tif"><alt-text content-type="machine-generated">Network diagram visualizing connections related to laparoscopic cholecystectomy. Central node linked to terms like transplant surgery, robotic surgery, and education. Lines represent relationships, varying in color and thickness. Created with VOSviewer.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3e"><label>4.5</label><title>Popular deep learning models</title>
<p>DL is significantly contributing to the surgical domain, particularly through convolutional neural networks (CNNs). CNNs are highly effective for processing image data, making them popular for the integration with other algorithms, such as recurrent neural networks (RNNs) and long short-term memory networks (LSTMs), which preserve temporal features. This combination is beneficial when maintaining the temporal information across different frames is critical, like surgical workflow recognition.</p>
<p>Recurrent networks are extensively utilized in tracking applications, such as tool tracking, where they maintain continuity by connecting subsequent frames and preventing the loss of recognized or split items. However, tasks that are within the spatial components only can be achieved using CNNs only like classification, localization, or segmentation of surgical instruments. Notably, Encoder-decoder architecture is suitable for generating segmentation masks, including binary, semantic, or tool parts. In this network, the encoder identifies and captures important features, then the decoder reconstructs them into an abstract form.</p>
<p>Three-dimensional CNNs are used to identify surgical tool joints as they harness data from the third dimension. Moreover, numerous articles utilize existing CNN-based models such as YOLOv3 and ResNet, illustrating the adaptability of CNNs across diverse architectural frameworks to tackle unique issues in surgical operations.</p>
</sec>
<sec id="s3f"><label>4.6</label><title>Detection, segmentation and tracking applications</title>
<p>Within our corpus, instrument perception research in MIS can be grouped into three closely related tasks: detection (tool presence and coarse localization), segmentation (pixel-level delineation), and tracking/pose estimation (temporal continuity and articulation). Earlier high-impact studies frequently emphasized detection and tool presence classification, often as components of workflow analysis and training pipelines (<xref ref-type="bibr" rid="B166">166</xref>) where coarse spatial cues are sufficient, and were commonly evaluated on public laparoscopic video datasets such as Cholec80 (<xref ref-type="bibr" rid="B50">50</xref>).</p>
<p>More recent publications increasingly focus on segmentation, driven by the need for precise spatial understanding and by the maturation of benchmark datasets and challenges. Studies published include approaches that leverage transformer backbones and foundation-model paradigms for robust instrument segmentation under occlusion and scene variability, such as for example, adaptations of Segment Anything&#x2013;style models for robotic instrument segmentation on EndoVis benchmarks [e.g., Surgical-DeSAM (<xref ref-type="bibr" rid="B10">10</xref>)]. Segmentation outputs are also being integrated into downstream applications such as objective performance assessment and context-aware assistance (<xref ref-type="bibr" rid="B167">167</xref>).</p>
<p>Tracking and pose estimation extend detection and segmentation across video frames and support higher-level scene understanding by enforcing temporal consistency. Recent spatio-temporal architectures have been used to track instrument motion (<xref ref-type="bibr" rid="B31">31</xref>) and articulation (<xref ref-type="bibr" rid="B163">163</xref>) and to support phase/step recognition, with benchmarking on datasets such as Cholec80 and EndoVis. Together, these complementary task families suggest a continued shift toward unified, real-time models that jointly model spatial detail and temporal dynamics.</p>
</sec>
<sec id="s3g"><label>4.7</label><title>Content analysis of research topics and subfields within surgical video analysis</title>
<p>The top ten most cited papers show the use of DL algorithms for a wide variety of purposes, including in MIS (<xref ref-type="table" rid="T5">Table&#x00A0;5</xref>). Twinada&#x0027;s &#x201C;EndoNet: A Deep Architecture for Recognition Tasks on Laparoscopic Videos&#x201D; (<xref ref-type="bibr" rid="B6">6</xref>) introduced a very popular convolutional neural network architecture named EndoNet. Although it focuses on surgical phase recognition in a variety of surgeries, it has only been applied to laparoscopic surgeries in this study. Instrument segmentation is crucial for this process, enabling the accurate tracking of instruments and allowing the algorithm to recognize the current phase of the surgery. Similarly, Jin&#x0027;s 2018 article &#x201C;SV-RCNet: Workflow Recognition From Surgical Videos Using Recurrent Convolutional Network&#x201D; (<xref ref-type="bibr" rid="B7">7</xref>) also focuses on workflow recognition, specifically in laparoscopic cholecystectomy videos. Jin&#x0027;s approach combines a convolutional neural network and a recurrent neural network, allowing for better utilization of both the temporal and visual features present in surgical videos. Padoy&#x0027;s 2019 (<xref ref-type="bibr" rid="B162">162</xref>) and Dergachyova&#x0027;s 2016 articles also focus on similar objective, using neural networks and other DL algorithms to identify the various stages of surgical procedures and the utilization of endoscopic videos for the training of these algorithms.</p>
<p>Six out of the 10 most cited articles were published in IEEE, likely because the field was in its early stages of development and the research was more technical in nature. At that time, the focus was on algorithms and specific technology, which was not clinically applicable yet. As the field progressed and deep learning applications became more integrated into surgical practice, the research gained clinical relevance. This has led to an increase in publications in medical journals rather than the more technical journals like the ones published by IEEE. This shift reflects the growing importance of these technologies to surgeons as they transitioned from experimental to practical, clinical tools.</p>
</sec>
</sec>
<sec id="s4"><label>5</label><title>Limitations of this study</title>
<p>Despite the comprehensive nature of this bibliometric analysis, several limitations should be acknowledged. First, restricting the search strategy to six major databases may have led to the exclusion of relevant studies indexed elsewhere or published in non-English languages. As with any bibliometric analysis, the observed patterns are shaped by the structure and indexing practices of the selected databases and therefore reflect trends within the indexed academic literature rather than the entirety of research activity in this domain. Additionally, the reliance on specific search terms (e.g., &#x201C;deep learning,&#x201D; &#x201C;robotic-assisted surgery,&#x201D; &#x201C;minimally invasive surgery&#x201D;) could have omitted certain papers that used alternative terminologies or did not explicitly mention deep learning in the title or abstract. Second, citation-based metrics can be influenced by factors such as varying publication times and self-citations, potentially skewing the apparent impact of certain studies. Third, conference proceedings that are often pivotal for rapidly evolving fields like deep learning may be underrepresented if they were not consistently indexed. Moreover, limiting the scope to original research involving deep-learning models for instrument segmentation, detection, or tracking may overlook valuable insights from review articles, meta-analyses, or purely simulation-based studies. Moreover, bibliometric analyses characterize quantitative and relational patterns in the literature and do not assess the methodological quality, technical rigor, or clinical relevance of individual studies. Importantly, while co-authorship and co-citation networks illuminate collaboration patterns, they do not capture the quality of these partnerships or the nuances of data sharing and methodological reproducibility across different research groups. Finally, because this is a bibliometric analysis, it characterizes research activity and influence rather than technical performance or patient outcomes; therefore, implications for clinical practice should be interpreted indirectly.</p>
</sec>
<sec id="s5"><label>6</label><title>Future direction and challenges</title>
<p>Building on the current trends observed in DL for MIS, several avenues appear especially promising. As the field transitions from experimental algorithms to clinical deployment, explainable AI and regulatory frameworks will become essential for ensuring transparency and patient safety. Real-time analysis potentially facilitated by edge computing and high-efficiency models can significantly advance intraoperative guidance, allowing surgeons to receive immediate, context-aware feedback.</p>
<p>An emerging direction is the development of agentic AI solutions for the operating room in the form of software agents that can maintain a structured representation of the evolving surgical context (e.g., surgical phase, instruments, anatomy, and patient status) and plan multi-step assistance actions. Rather than producing single-task predictions in isolation, an agentic system could integrate outputs from perception modules (instrument/scene segmentation and tracking), workflow recognition, device telemetry, and peri-operative data streams. This could potentially enable proactive, context-aware decision support such as tailored safety-check prompts, anomaly-detection alerts, and guideline-consistent &#x201C;next-step&#x201D; suggestions, while also streamlining routine tasks such as intraoperative documentation and equipment-readiness checks. However, because such agents may initiate recommendations without explicit prompting, safe deployment will require careful human-in-the-loop design, transparent uncertainty reporting, audit trails, rigorous prospective validation, and clear governance around cybersecurity, privacy, and accountability to ensure these systems support (rather than replace) clinician judgement.</p>
<p>Moreover, multimodal learning (combining visual data with text, sensor data, or electronic health records) stands to enrich surgical decision-making, while foundation models pretrained on large-scale datasets could be fine-tuned for MIS-specific tasks. Recent work illustrates how language&#x2013;vision pretraining can be adapted for surgical scene understanding [e.g., CLIP-based approaches for surgical scene segmentation (<xref ref-type="bibr" rid="B168">168</xref>)]. In parallel, there is growing momentum toward holistic surgical scene segmentation that jointly delineate instruments and relevant anatomy which may better support context-aware assistance and downstream analytics than tool-only models (<xref ref-type="bibr" rid="B169">169</xref>, <xref ref-type="bibr" rid="B170">170</xref>). Concurrently, there is also a growing need for standardized datasets and labeling protocols, which will enhance reproducibility and reduce the fragmentation currently seen in surgical video annotation efforts. Beyond intraoperative deployment, these perception capabilities can also be leveraged in simulation-based education and skills assessment&#x2014;where interface design and serious-game/gamified training environments may benefit from reliable instrument/scene understanding (<xref ref-type="bibr" rid="B171">171</xref>, <xref ref-type="bibr" rid="B172">172</xref>). On the international front, broadening collaboration across underrepresented regions could diversify research perspectives and spur innovations that are globally relevant. Finally, continued integration with robotic platforms could pave the way for semi-autonomous or even fully autonomous surgical maneuvers, provided safety and ethical concerns are rigorously addressed.</p>
</sec>
<sec id="s6" sec-type="conclusions"><label>7</label><title>Conclusions</title>
<p>This bibliometric analysis demonstrates the rapid growth and increasing maturity of deep learning research in minimally invasive surgery. Our analysis revealed that research in this was field was globally distributed but concentrated in certain regions, institutions, and authors. Early advances in instrument detection and surgical workflow analysis were followed by sustained increases in publication volume and citation impact, particularly across Europe, North America, and East Asia. Several regions, including France, Hong Kong, and the United Kingdom, exhibit high citation influence relative to output, indicating that impact is driven by research quality and network position rather than volume alone. The rising prominence of deep learning and laparoscopy-related keywords reflects continued emphasis on video-based intraoperative analysis. Structural limitations were also noticed, including fragmented datasets and inconsistent benchmarking, which constrain reproducibility and cross-study comparability. Future progress in DL for MIS will depend on clinically deployable, explainable, and real-time methods supported by standardized data and broader collaboration. Continued integration with multimodal inputs and surgical platforms will be critical to ensure safe and effective translation into practice.</p>
<p>Overall, this review delineates how deep learning research in minimally invasive surgery has evolved and where influence and structural limitations persist. Emerging architectures like Transformers, the growing emphasis on explainability, and the adoption of large, community-driven datasets suggest that the field is poised for another phase of innovation. By mapping the existing research landscape and identifying influential studies, contributors, and collaborations, it provides an empirical basis to guide future research directions, optimize inter-institutional partnerships, and drive meaningful contributions to the field of DL in MIS.</p>
</sec>
</body>
<back>
<sec id="s7" sec-type="data-availability"><title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec id="s8" sec-type="author-contributions"><title>Author contributions</title>
<p>MY: Writing &#x2013; review &#x0026; editing, Formal analysis, Investigation, Methodology, Writing &#x2013; original draft, Data curation. KEA: Writing &#x2013; review &#x0026; editing, Formal analysis, Methodology. MA: Formal analysis, Methodology, Writing &#x2013; review &#x0026; editing. FA: Conceptualization, Writing &#x2013; review &#x0026; editing, Supervision, Writing &#x2013; original draft. KAJ: Writing &#x2013; original draft, Conceptualization, Supervision, Writing &#x2013; review &#x0026; editing. SB: Funding acquisition, Project administration, Supervision, Writing &#x2013; review &#x0026; editing, Conceptualization.</p>
</sec>
<ack><title>Acknowledgments</title>
<p>The authors would like to acknowledge the support of the Surgical Research Section at Hamad Medical Corporation for the conduct of this research.</p>
</ack>
<sec id="s11" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s12" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="s13" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s10"><title>Author disclaimer</title>
<p>The content is solely the responsibility of the authors and does not necessarily represent the official views of Qatar Research Development and Innovation Council.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Blencowe</surname> <given-names>NS</given-names></name> <name><surname>Waldon</surname> <given-names>R</given-names></name> <name><surname>Vipond</surname> <given-names>MN</given-names></name></person-group>. <article-title>Management of patients after laparoscopic procedures</article-title>. <source>Br Med J</source>. (<year>2018</year>) <volume>360</volume>:<fpage>k120</fpage>. <pub-id pub-id-type="doi">10.1136/bmj.k120</pub-id></mixed-citation></ref>
<ref id="B2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sheetz</surname> <given-names>KH</given-names></name> <name><surname>Claflin</surname> <given-names>J</given-names></name> <name><surname>Dimick</surname> <given-names>JB</given-names></name></person-group>. <article-title>Trends in the adoption of robotic surgery for common surgical procedures</article-title>. <source>JAMA Netw Open</source>. (<year>2020</year>) <volume>3</volume>:<fpage>e1918911</fpage>. <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2019.18911</pub-id><pub-id pub-id-type="pmid">31922557</pub-id></mixed-citation></ref>
<ref id="B3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Von Ahlen</surname> <given-names>C</given-names></name> <name><surname>Geissler</surname> <given-names>A</given-names></name> <name><surname>Vogel</surname> <given-names>J</given-names></name></person-group>. <article-title>Comparison of the effectiveness of open, laparoscopic, and robotic-assisted radical prostatectomies based on complication rates: a retrospective observational study with administrative data from Switzerland</article-title>. <source>BMC Urol</source><italic>.</italic> (<year>2024</year>) <volume>24</volume>:<fpage>215</fpage>. <pub-id pub-id-type="doi">10.1186/s12894-024-01597-3</pub-id><pub-id pub-id-type="pmid">39375695</pub-id></mixed-citation></ref>
<ref id="B4"><label>4.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Allan</surname> <given-names>M</given-names></name> <name><surname>Shvets</surname> <given-names>A</given-names></name> <name><surname>Kurmann</surname> <given-names>T</given-names></name> <name><surname>Zhang</surname> <given-names>Z</given-names></name> <name><surname>Duggal</surname> <given-names>R</given-names></name> <name><surname>Su</surname> <given-names>Y-H</given-names></name><etal/></person-group> <comment>2017 Robotic Instrument Segmentation Challenge</comment>. (<year>2019</year>). <pub-id pub-id-type="doi">10.48550/ARXIV.1902.06426</pub-id></mixed-citation></ref>
<ref id="B5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bodenstedt</surname> <given-names>S</given-names></name> <name><surname>Rivoir</surname> <given-names>D</given-names></name> <name><surname>Jenke</surname> <given-names>A</given-names></name> <name><surname>Wagner</surname> <given-names>M</given-names></name> <name><surname>Breucha</surname> <given-names>M</given-names></name> <name><surname>M&#x00FC;ller-Stich</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Active learning using deep Bayesian networks for surgical workflow analysis</article-title>. <source>Int J CARS</source>. (<year>2019</year>) <volume>14</volume>:<fpage>1079</fpage>&#x2013;<lpage>87</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-01963-9</pub-id></mixed-citation></ref>
<ref id="B6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Twinanda</surname> <given-names>AP</given-names></name> <name><surname>Shehata</surname> <given-names>S</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name> <name><surname>Marescaux</surname> <given-names>J</given-names></name> <name><surname>de Mathelin</surname> <given-names>M</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>Endonet: a deep architecture for recognition tasks on laparoscopic videos</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2017</year>) <volume>36</volume>:<fpage>86</fpage>&#x2013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2016.2593957</pub-id><pub-id pub-id-type="pmid">27455522</pub-id></mixed-citation></ref>
<ref id="B7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Yu</surname> <given-names>L</given-names></name> <name><surname>Qin</surname> <given-names>J</given-names></name> <name><surname>Fu</surname> <given-names>C-W</given-names></name><etal/></person-group> <article-title>SV-RCNet: workflow recognition from surgical videos using recurrent convolutional network</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2018b</year>) <volume>37</volume>:<fpage>1114</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2017.2787657</pub-id></mixed-citation></ref>
<ref id="B8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>Machine and deep learning for workflow recognition during surgery</article-title>. <source>Minim Invasive Ther Allied Technol MITAT Off J Soc Minim Invasive Ther</source><italic>.</italic> (<year>2019</year>) <volume>28</volume>:<fpage>82</fpage>&#x2013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1080/13645706.2019.1584116</pub-id></mixed-citation></ref>
<ref id="B9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kiyasseh</surname> <given-names>D</given-names></name> <name><surname>Ma</surname> <given-names>R</given-names></name> <name><surname>Haque</surname> <given-names>TF</given-names></name> <name><surname>Miles</surname> <given-names>BJ</given-names></name> <name><surname>Wagner</surname> <given-names>C</given-names></name> <name><surname>Donoho</surname> <given-names>DA</given-names></name><etal/></person-group> <article-title>A vision transformer for decoding surgeon activity from surgical videos. Nat</article-title>. <source>Biomed Eng</source>. (<year>2023</year>) <volume>7</volume>:<fpage>780</fpage>&#x2013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1038/s41551-023-01010-8</pub-id></mixed-citation></ref>
<ref id="B10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sheng</surname> <given-names>Y</given-names></name> <name><surname>Bano</surname> <given-names>S</given-names></name> <name><surname>Clarkson</surname> <given-names>MJ</given-names></name> <name><surname>Islam</surname> <given-names>M</given-names></name></person-group>. <article-title>Surgical-DeSAM: decoupling SAM for instrument segmentation in robotic surgery</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2024</year>) <volume>19</volume>:<fpage>1267</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-024-03163-6</pub-id><pub-id pub-id-type="pmid">38758289</pub-id></mixed-citation></ref>
<ref id="B11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Marzullo</surname> <given-names>A</given-names></name> <name><surname>Moccia</surname> <given-names>S</given-names></name> <name><surname>Catellani</surname> <given-names>M</given-names></name> <name><surname>Calimeri</surname> <given-names>F</given-names></name> <name><surname>Momi</surname> <given-names>ED</given-names></name></person-group>. <article-title>Towards realistic laparoscopic image generation using image-domain translation</article-title>. <source>Comput Methods Programs Biomed</source><italic>.</italic> (<year>2021</year>) <volume>200</volume>:<fpage>105834</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2020.105834</pub-id><pub-id pub-id-type="pmid">33229016</pub-id></mixed-citation></ref>
<ref id="B12"><label>12.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>R</given-names></name> <name><surname>Zhang</surname><given-names>D</given-names></name> <name><surname>Li</surname><given-names>Q</given-names></name> <name><surname>Zhou</surname><given-names>X-Y</given-names></name> <name><surname>Lo</surname><given-names>B</given-names></name></person-group>. <article-title>Real-time surgical environment enhancement for robot-assisted minimally invasive surgery based on super-resolution</article-title>. <source>arXiv</source> <comment>[Preprint]</comment>. <source>arXiv:2011.04003</source> (<year>2020</year>). <pub-id pub-id-type="doi">10.48550/ARXIV.2011.04003</pub-id></mixed-citation></ref>
<ref id="B13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>D&#x2019;Andria Ursoleo</surname> <given-names>J</given-names></name> <name><surname>Losiggio</surname> <given-names>R</given-names></name> <name><surname>Monaco</surname> <given-names>F</given-names></name></person-group>. <article-title>The first 100 years of the British journal of anaesthesia: a bibliometric analysis of the top-cited articles</article-title>. <source>Br J Anaesth</source>. (<year>2024</year>) <volume>132</volume>:<fpage>789</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1016/j.bja.2024.01.016</pub-id></mixed-citation></ref>
<ref id="B14"><label>14.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Akmal</surname> <given-names>M</given-names></name> <name><surname>Hasnain</surname> <given-names>N</given-names></name> <name><surname>Rehan</surname> <given-names>A</given-names></name> <name><surname>Iqbal</surname> <given-names>U</given-names></name> <name><surname>Hashmi</surname> <given-names>S</given-names></name> <name><surname>Fatima</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>Glioblastome Multiforme: a bibliometric analysis</article-title>. <source>World Neurosurg</source>. (<year>2020</year>) <volume>136</volume>:<fpage>270</fpage>&#x2013;<lpage>82</lpage>. <pub-id pub-id-type="doi">10.1016/j.wneu.2020.01.027</pub-id><pub-id pub-id-type="pmid">31953095</pub-id></mixed-citation></ref>
<ref id="B15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bidwe</surname> <given-names>RV</given-names></name> <name><surname>Mishra</surname> <given-names>S</given-names></name> <name><surname>Patil</surname> <given-names>S</given-names></name> <name><surname>Shaw</surname> <given-names>K</given-names></name> <name><surname>Vora</surname> <given-names>DR</given-names></name> <name><surname>Kotecha</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>Deep learning approaches for video compression: a bibliometric analysis</article-title>. <source>Big Data Cogn Comput</source>. (<year>2022</year>) <volume>6</volume>:<fpage>44</fpage>. <pub-id pub-id-type="doi">10.3390/bdcc6020044</pub-id></mixed-citation></ref>
<ref id="B16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ursoleo</surname> <given-names>JD</given-names></name> <name><surname>Bottussi</surname> <given-names>A</given-names></name> <name><surname>Christiansen</surname> <given-names>S</given-names></name> <name><surname>Sullivan</surname> <given-names>DR</given-names></name> <name><surname>Vranas</surname> <given-names>KC</given-names></name> <name><surname>Rosa</surname> <given-names>WE</given-names></name><etal/></person-group> <article-title>Mapping the science of palliative care: a bibliometric analysis of the top 100 cited articles</article-title>. <source>Palliat Med</source>. (<year>2025a</year>) <volume>39</volume>:<fpage>922</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1177/02692163251362560</pub-id></mixed-citation></ref>
<ref id="B17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>D&#x2019;Andria Ursoleo</surname> <given-names>J</given-names></name> <name><surname>Cal&#x00EC;</surname> <given-names>C</given-names></name> <name><surname>Losiggio</surname> <given-names>R</given-names></name> <name><surname>Limone</surname> <given-names>V</given-names></name> <name><surname>Mucci</surname> <given-names>E</given-names></name> <name><surname>Monaco</surname> <given-names>F</given-names></name></person-group>. <article-title>Spiritual care in palliative medicine and End of life: a bibliometric network analysis</article-title>. <source>J Palliat Med</source>. (<year>2025b</year>) <volume>28</volume>:<fpage>265</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1089/jpm.2024.0007</pub-id></mixed-citation></ref>
<ref id="B18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>D&#x2019;Andria Ursoleo</surname> <given-names>J</given-names></name> <name><surname>Losiggio</surname> <given-names>R</given-names></name> <name><surname>Licheri</surname> <given-names>M</given-names></name> <name><surname>Barucco</surname> <given-names>G</given-names></name> <name><surname>Lazzari</surname> <given-names>S</given-names></name> <name><surname>Faustini</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Minimal invasive extracorporeal circulation: a bibliometric network analysis of the global scientific output</article-title>. <source>Perfusion</source>. (<year>2025c</year>) <volume>40</volume>:<fpage>1176</fpage>&#x2013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1177/02676591241269729</pub-id></mixed-citation></ref>
<ref id="B19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>D</given-names></name> <name><surname>Yu</surname> <given-names>D</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Yang</surname> <given-names>R</given-names></name></person-group>. <article-title>A bibliometric analysis of PROTAC from 2001 to 2021</article-title>. <source>Eur J Med Chem</source><italic>.</italic> (<year>2022</year>) <volume>244</volume>:<fpage>114838</fpage>. <pub-id pub-id-type="doi">10.1016/j.ejmech.2022.114838</pub-id><pub-id pub-id-type="pmid">36274273</pub-id></mixed-citation></ref>
<ref id="B20"><label>20.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Z</given-names></name> <name><surname>Maimaiti</surname> <given-names>Z</given-names></name> <name><surname>Fu</surname> <given-names>J</given-names></name> <name><surname>Chen</surname> <given-names>J-Y</given-names></name> <name><surname>Xu</surname> <given-names>C</given-names></name></person-group>. <article-title>Global research landscape on artificial intelligence in arthroplasty: a bibliometric analysis</article-title>. <source>Digit Health</source>. (<year>2023</year>) <volume>9</volume>:<fpage>20552076231184048</fpage>. <pub-id pub-id-type="doi">10.1177/20552076231184048</pub-id><pub-id pub-id-type="pmid">37361434</pub-id></mixed-citation></ref>
<ref id="B21"><label>21.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Shi</surname> <given-names>J</given-names></name> <name><surname>Shi</surname> <given-names>S</given-names></name> <name><surname>Bo</surname> <given-names>R</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Hu</surname> <given-names>Y</given-names></name></person-group>. <article-title>Bibliometric analysis on the progress of chronic heart failure</article-title>. <source>Curr Probl Cardiol</source>. (<year>2022</year>) <volume>47</volume>:<fpage>101213</fpage>. <pub-id pub-id-type="doi">10.1016/j.cpcardiol.2022.101213</pub-id><pub-id pub-id-type="pmid">35525461</pub-id></mixed-citation></ref>
<ref id="B22"><label>22.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aria</surname> <given-names>M</given-names></name> <name><surname>Cuccurullo</surname> <given-names>C</given-names></name></person-group>. <article-title>Bibliometrix: an R-tool for comprehensive science mapping analysis</article-title>. <source>J Informetr</source><italic>.</italic> (<year>2017</year>) <volume>11</volume>:<fpage>959</fpage>&#x2013;<lpage>75</lpage>. <pub-id pub-id-type="doi">10.1016/j.joi.2017.08.007</pub-id></mixed-citation></ref>
<ref id="B23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Van Eck</surname> <given-names>NJ</given-names></name> <name><surname>Waltman</surname> <given-names>L</given-names></name></person-group>. <article-title>Software survey: VOSviewer, a computer program for bibliometric mapping</article-title>. <source>Scientometrics</source>. (<year>2010</year>) <volume>84</volume>:<fpage>523</fpage>&#x2013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1007/s11192-009-0146-3</pub-id><pub-id pub-id-type="pmid">20585380</pub-id></mixed-citation></ref>
<ref id="B24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sahu</surname> <given-names>M</given-names></name> <name><surname>Mukhopadhyay</surname> <given-names>A</given-names></name> <name><surname>Szengel</surname> <given-names>A</given-names></name> <name><surname>Zachow</surname> <given-names>S</given-names></name></person-group>. <article-title>Addressing multi-label imbalance problem of surgical tool detection using CNN</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2017</year>) <volume>12</volume>:<fpage>1013</fpage>&#x2013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-017-1565-x</pub-id><pub-id pub-id-type="pmid">28357628</pub-id></mixed-citation></ref>
<ref id="B25"><label>25.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>X</given-names></name> <name><surname>Yu</surname> <given-names>L</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Qin</surname> <given-names>J</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name><etal/></person-group> <article-title>AGNet: attention-guided network for surgical tool presence detection</article-title>. In: <person-group person-group-type="editor"><name><surname>Cardoso</surname> <given-names>MJ</given-names></name> <name><surname>Arbel</surname> <given-names>T</given-names></name> <name><surname>Carneiro</surname> <given-names>G</given-names></name> <name><surname>Syeda-Mahmood</surname> <given-names>T</given-names></name> <name><surname>Tavares</surname> <given-names>JMRS</given-names></name> <name><surname>Moradi</surname> <given-names>M</given-names></name></person-group>, editors. <source>Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2017</year>). p. <fpage>186</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-67558-9_22</pub-id></mixed-citation></ref>
<ref id="B26"><label>26.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S</given-names></name> <name><surname>Raju</surname> <given-names>A</given-names></name> <name><surname>Huang</surname> <given-names>J</given-names></name></person-group>. <article-title>Deep learning based multi-label classification for surgical tool presence detection in laparoscopic videos</article-title>. In: <article-title>2017 IEEE 14th international symposium on biomedical imaging (ISBI 2017) 620&#x2013;3</article-title>; <conf-loc>IEEE, Melbourne, Australia</conf-loc>; (<year>2017</year>). <pub-id pub-id-type="doi">10.1109/ISBI.2017.7950597</pub-id></mixed-citation></ref>
<ref id="B27"><label>27.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Mishra</surname> <given-names>K</given-names></name> <name><surname>Sathish</surname> <given-names>R</given-names></name> <name><surname>Sheet</surname> <given-names>D</given-names></name></person-group>. <article-title>Learning latent temporal connectionism of deep residual visual abstractions for identifying surgical tools in laparoscopy procedures</article-title>. In: <conf-name>2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2233&#x2013;40</conf-name>; <publisher-name>IEEE</publisher-name><conf-loc>; Honolulu, HI, USA</conf-loc>; (<year>2017</year>). <pub-id pub-id-type="doi">10.1109/CVPRW.2017.277</pub-id></mixed-citation></ref>
<ref id="B28"><label>28.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Leibetseder</surname> <given-names>A</given-names></name> <name><surname>Primus</surname> <given-names>MJ</given-names></name> <name><surname>Petscharnig</surname> <given-names>S</given-names></name> <name><surname>Schoeffmann</surname> <given-names>K</given-names></name></person-group>. <article-title>Real-time image-based smoke detection in endoscopic videos</article-title>. In: <conf-name>Proceedings of the on thematic workshops of ACM multimedia 2017 296&#x2013;304</conf-name>; <publisher-name>ACM, Mountain View</publisher-name><conf-loc>; California USA</conf-loc>; (<year>2017</year>). <pub-id pub-id-type="doi">10.1145/3126686.3126690</pub-id></mixed-citation></ref>
<ref id="B29"><label>29.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Al Hajj</surname> <given-names>H</given-names></name> <name><surname>Lamard</surname> <given-names>M</given-names></name> <name><surname>Conze</surname> <given-names>P-H</given-names></name> <name><surname>Roychowdhury</surname> <given-names>S</given-names></name> <name><surname>Hu</surname> <given-names>X</given-names></name> <name><surname>Mar&#x0161;alkait&#x0117;</surname> <given-names>G</given-names></name><etal/></person-group> <article-title>CATARACTS: challenge on automatic tool annotation for cataRACT surgery</article-title>. <source>Med Image Anal</source>. (<year>2019</year>) <volume>52</volume>:<fpage>24</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2018.11.008</pub-id><pub-id pub-id-type="pmid">30468970</pub-id></mixed-citation></ref>
<ref id="B30"><label>30.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Attia</surname> <given-names>M</given-names></name> <name><surname>Hossny</surname> <given-names>M</given-names></name> <name><surname>Nahavandi</surname> <given-names>S</given-names></name> <name><surname>Asadi</surname> <given-names>H</given-names></name></person-group>. <article-title>Surgical tool segmentation using a hybrid deep CNN-RNN auto encoder-decoder</article-title>. In: <conf-name>2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC) 3373&#x2013;8</conf-name>; <publisher-name>IEEE</publisher-name>; <conf-loc>Banff, AB</conf-loc>; (<year>2017</year>). <pub-id pub-id-type="doi">10.1109/SMC.2017.8123151</pub-id></mixed-citation></ref>
<ref id="B31"><label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hegde</surname> <given-names>SR</given-names></name> <name><surname>Namazi</surname> <given-names>B</given-names></name> <name><surname>Iyengar</surname> <given-names>N</given-names></name> <name><surname>Cao</surname> <given-names>S</given-names></name> <name><surname>Desir</surname> <given-names>A</given-names></name> <name><surname>Marques</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Automated segmentation of phases, steps, and tasks in laparoscopic cholecystectomy using deep learning</article-title>. <source>Surg Endosc</source>. (<year>2024</year>) <volume>38</volume>:<fpage>158</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-023-10482-3</pub-id><pub-id pub-id-type="pmid">37945709</pub-id></mixed-citation></ref>
<ref id="B32"><label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ban</surname> <given-names>Y</given-names></name> <name><surname>Eckhoff</surname> <given-names>JA</given-names></name> <name><surname>Ward</surname> <given-names>TM</given-names></name> <name><surname>Hashimoto</surname> <given-names>DA</given-names></name> <name><surname>Meireles</surname> <given-names>OR</given-names></name> <name><surname>Rus</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Concept graph neural networks for surgical video understanding</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2024</year>) <volume>43</volume>:<fpage>264</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2023.3299518</pub-id><pub-id pub-id-type="pmid">37498757</pub-id></mixed-citation></ref>
<ref id="B33"><label>33.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Wu</surname> <given-names>Z</given-names></name> <name><surname>Dai</surname> <given-names>J</given-names></name> <name><surname>Morgan</surname> <given-names>TN</given-names></name> <name><surname>Garbens</surname> <given-names>A</given-names></name> <name><surname>Kominsky</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Evaluating robotic-assisted partial nephrectomy surgeons with fully convolutional segmentation and multi-task attention networks</article-title>. <source>J Robotic Surg</source>. (<year>2023</year>) <volume>17</volume>:<fpage>2323</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1007/s11701-023-01657-0</pub-id></mixed-citation></ref>
<ref id="B34"><label>34.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lou</surname> <given-names>A</given-names></name> <name><surname>Tawfik</surname> <given-names>K</given-names></name> <name><surname>Yao</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Noble</surname> <given-names>J</given-names></name></person-group>. <article-title>Min-max similarity: a contrastive semi-supervised deep learning network for surgical tools segmentation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2023</year>) <volume>42</volume>:<fpage>2832</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2023.3266137</pub-id><pub-id pub-id-type="pmid">37037256</pub-id></mixed-citation></ref>
<ref id="B35"><label>35.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>King</surname> <given-names>D</given-names></name> <name><surname>Adidharma</surname> <given-names>L</given-names></name> <name><surname>Peng</surname> <given-names>H</given-names></name> <name><surname>Moe</surname> <given-names>K</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Yang</surname> <given-names>Z</given-names></name><etal/></person-group> <article-title>Automatic summarization of endoscopic skull base surgical videos through object detection and hidden markov modeling</article-title>. <source>Comput Med Imaging Graph</source>. (<year>2023</year>) <volume>108</volume>:<fpage>102248</fpage>. <pub-id pub-id-type="doi">10.1016/j.compmedimag.2023.102248</pub-id><pub-id pub-id-type="pmid">37315397</pub-id></mixed-citation></ref>
<ref id="B36"><label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fischer</surname> <given-names>E</given-names></name> <name><surname>Jawed</surname> <given-names>KJ</given-names></name> <name><surname>Cleary</surname> <given-names>K</given-names></name> <name><surname>Balu</surname> <given-names>A</given-names></name> <name><surname>Donoho</surname> <given-names>A</given-names></name> <name><surname>Thompson Gestrich</surname> <given-names>W</given-names></name><etal/></person-group> <article-title>A methodology for the annotation of surgical videos for supervised machine learning applications</article-title>. <source>Int J CARS</source>. (<year>2023</year>) <volume>18</volume>:<fpage>1673</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-023-02923-0</pub-id></mixed-citation></ref>
<ref id="B37"><label>37.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hasan</surname> <given-names>SMK</given-names></name> <name><surname>Simon</surname> <given-names>RA</given-names></name> <name><surname>Linte</surname> <given-names>CA</given-names></name></person-group>. <article-title>Inpainting surgical occlusion from laparoscopic video sequences for robot-assisted interventions</article-title>. <source>J Med Imaging</source>. (<year>2023</year>) <volume>10</volume>:<fpage>045002</fpage>. <pub-id pub-id-type="doi">10.1117/1.JMI.10.4.045002</pub-id></mixed-citation></ref>
<ref id="B38"><label>38.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ding</surname> <given-names>H</given-names></name> <name><surname>Wu</surname> <given-names>JY</given-names></name> <name><surname>Li</surname> <given-names>Z</given-names></name> <name><surname>Unberath</surname> <given-names>M</given-names></name></person-group>. <article-title>Rethinking causality-driven robot tool segmentation with temporal constraints</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2023</year>) <volume>18</volume>:<fpage>1009</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-023-02872-8</pub-id><pub-id pub-id-type="pmid">37027082</pub-id></mixed-citation></ref>
<ref id="B39"><label>39.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nespolo</surname> <given-names>RG</given-names></name> <name><surname>Cole</surname> <given-names>E</given-names></name> <name><surname>Wang</surname> <given-names>D</given-names></name> <name><surname>Yi</surname> <given-names>D</given-names></name> <name><surname>Leiderman</surname> <given-names>YI</given-names></name></person-group>. <article-title>A platform for tracking surgeon and observer gaze as a surrogate for attention in ophthalmic surgery</article-title>. <source>Ophthalmol Sci</source>. (<year>2023</year>) <volume>3</volume>:<fpage>100246</fpage>. <pub-id pub-id-type="doi">10.1016/j.xops.2022.100246</pub-id><pub-id pub-id-type="pmid">36748062</pub-id></mixed-citation></ref>
<ref id="B40"><label>40.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Eckhoff</surname> <given-names>JA</given-names></name> <name><surname>Ban</surname> <given-names>Y</given-names></name> <name><surname>Rosman</surname> <given-names>G</given-names></name> <name><surname>M&#x00FC;ller</surname> <given-names>DT</given-names></name> <name><surname>Hashimoto</surname> <given-names>DA</given-names></name> <name><surname>Witkowski</surname> <given-names>E</given-names></name><etal/></person-group> <article-title>TEsonet: knowledge transfer in surgical phase recognition from laparoscopic sleeve gastrectomy to the laparoscopic part of ivor&#x2013;lewis esophagectomy</article-title>. <source>Surg Endosc</source>. (<year>2023</year>) <volume>37</volume>:<fpage>4040</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-023-09971-2</pub-id><pub-id pub-id-type="pmid">36932188</pub-id></mixed-citation></ref>
<ref id="B41"><label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>B</given-names></name> <name><surname>Goel</surname> <given-names>B</given-names></name> <name><surname>Sarhan</surname> <given-names>MH</given-names></name> <name><surname>Goel</surname> <given-names>VK</given-names></name> <name><surname>Abukhalil</surname> <given-names>R</given-names></name> <name><surname>Kalesan</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Surgical workflow recognition with temporal convolution and transformer for action segmentation</article-title>. <source>Int J CARS</source>. (<year>2022a</year>) <volume>18</volume>:<fpage>785</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-022-02811-z</pub-id></mixed-citation></ref>
<ref id="B42"><label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>JH</given-names></name> <name><surname>Goodman</surname> <given-names>ED</given-names></name> <name><surname>Dawes</surname> <given-names>AJ</given-names></name> <name><surname>Gahagan</surname> <given-names>JV</given-names></name> <name><surname>Esquivel</surname> <given-names>MM</given-names></name> <name><surname>Liebert</surname> <given-names>CA</given-names></name><etal/></person-group> <article-title>Using AI and computer vision to analyze technical proficiency in robotic surgery</article-title>. <source>Surg Endosc</source>. (<year>2023</year>) <volume>37</volume>:<fpage>3010</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-022-09781-y</pub-id><pub-id pub-id-type="pmid">36536082</pub-id></mixed-citation></ref>
<ref id="B43"><label>43.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yeh</surname> <given-names>H-H</given-names></name> <name><surname>Jain</surname> <given-names>AM</given-names></name> <name><surname>Fox</surname> <given-names>O</given-names></name> <name><surname>Sebov</surname> <given-names>K</given-names></name> <name><surname>Wang</surname> <given-names>SY</given-names></name></person-group>. <article-title>Phacotrainer: deep learning for cataract surgical videos to track surgical tools</article-title>. <source>Transl Vis Sci Technol</source>. (<year>2023</year>) <volume>12</volume>:<fpage>23</fpage>. <pub-id pub-id-type="doi">10.1167/tvst.12.3.23</pub-id><pub-id pub-id-type="pmid">36947046</pub-id></mixed-citation></ref>
<ref id="B44"><label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>B</given-names></name> <name><surname>Sturgeon</surname> <given-names>D</given-names></name> <name><surname>Shankar</surname> <given-names>AR</given-names></name> <name><surname>Goel</surname> <given-names>VK</given-names></name> <name><surname>Barker</surname> <given-names>J</given-names></name> <name><surname>Ghanem</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Surgical instrument recognition for instrument usage documentation and surgical video library indexing</article-title>. <source>Comput Methods Biomech Biomed Eng Imaging Vis</source>. (<year>2023b</year>) <volume>11</volume>:<fpage>1064</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1080/21681163.2022.2152371</pub-id></mixed-citation></ref>
<ref id="B45"><label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Markarian</surname> <given-names>N</given-names></name> <name><surname>Kugener</surname> <given-names>G</given-names></name> <name><surname>Pangal</surname> <given-names>DJ</given-names></name> <name><surname>Unadkat</surname> <given-names>V</given-names></name> <name><surname>Sinha</surname> <given-names>A</given-names></name> <name><surname>Zhu</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Validation of machine learning&#x2013;based automated surgical instrument annotation using publicly available intraoperative video</article-title>. <source>Oper Neurosurg</source>. (<year>2022</year>) <volume>23</volume>:<fpage>235</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.1227/ons.0000000000000274</pub-id><pub-id pub-id-type="pmid">35972087</pub-id></mixed-citation></ref>
<ref id="B46"><label>46.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mao</surname> <given-names>G</given-names></name> <name><surname>Lubelski</surname> <given-names>D</given-names></name> <name><surname>Zakaria</surname> <given-names>HM</given-names></name> <name><surname>Theodore</surname> <given-names>N</given-names></name></person-group>. <article-title>Image-guided minimally invasive surgery for treatment of the Bertolotti syndrome&#x2014;a case study: 2-dimensional operative video</article-title>. <source>Oper Neurosurg</source><italic>.</italic> (<year>2022</year>) <volume>22</volume>:<fpage>e222</fpage>&#x2013;<lpage>3</lpage>. <pub-id pub-id-type="doi">10.1227/ons.0000000000000132</pub-id><pub-id pub-id-type="pmid">35426882</pub-id></mixed-citation></ref>
<ref id="B47"><label>47.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Matton</surname> <given-names>N</given-names></name> <name><surname>Qalieh</surname> <given-names>A</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Annadanam</surname> <given-names>A</given-names></name> <name><surname>Thibodeau</surname> <given-names>A</given-names></name> <name><surname>Li</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Analysis of cataract surgery instrument identification performance of convolutional and recurrent neural network ensembles leveraging BigCat</article-title>. <source>Trans Vis Sci Tech</source>. (<year>2022</year>) <volume>11</volume>:<fpage>1</fpage>. <pub-id pub-id-type="doi">10.1167/tvst.11.4.1</pub-id></mixed-citation></ref>
<ref id="B48"><label>48.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Leifman</surname> <given-names>G</given-names></name> <name><surname>Aides</surname> <given-names>A</given-names></name> <name><surname>Golany</surname> <given-names>T</given-names></name> <name><surname>Freedman</surname> <given-names>D</given-names></name> <name><surname>Rivlin</surname> <given-names>E</given-names></name></person-group>. <article-title>Pixel-accurate segmentation of surgical tools based on bounding box annotations</article-title>. In: <conf-name>2022 26th International Conference on Pattern Recognition (ICPR) 5096&#x2013;103</conf-name>; <publisher-name>IEEE</publisher-name>; <conf-loc>Montreal, QC, Canada (Tor)</conf-loc>; (<year>2022</year>). <pub-id pub-id-type="doi">10.1109/ICPR56361.2022.9956530</pub-id></mixed-citation></ref>
<ref id="B49"><label>49.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Z</given-names></name> <name><surname>Simon</surname> <given-names>R</given-names></name> <name><surname>Linte</surname> <given-names>CA</given-names></name></person-group>. <article-title>A weakly supervised learning approach for surgical instrument segmentation from laparoscopic video sequences</article-title>. In: <person-group person-group-type="editor"><name><surname>Linte</surname> <given-names>CA</given-names></name> <name><surname>Siewerdsen</surname> <given-names>JH</given-names></name></person-group>, editors. <source>Medical Imaging 2022: Image-Guided Procedures, Robotic Interventions, and Modeling, Volume 60</source>. <publisher-loc>San Diego, United States</publisher-loc>: <publisher-name>SPIE</publisher-name> (<year>2022</year>). p. <fpage>10.1117/12.2610778</fpage>. <pub-id pub-id-type="doi">10.1117/12.2610778</pub-id></mixed-citation></ref>
<ref id="B50"><label>50.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Namazi</surname> <given-names>B</given-names></name> <name><surname>Sankaranarayanan</surname> <given-names>G</given-names></name> <name><surname>Devarajan</surname> <given-names>V</given-names></name></person-group>. <article-title>A contextual detector of surgical tools in laparoscopic videos using deep learning</article-title>. <source>Surg Endosc</source>. (<year>2022</year>) <volume>36</volume>:<fpage>679</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-021-08336-x</pub-id><pub-id pub-id-type="pmid">33559057</pub-id></mixed-citation></ref>
<ref id="B51"><label>51.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yeh</surname> <given-names>H-H</given-names></name> <name><surname>Jain</surname> <given-names>AM</given-names></name> <name><surname>Fox</surname> <given-names>O</given-names></name> <name><surname>Wang</surname> <given-names>SY</given-names></name></person-group>. <article-title>Phacotrainer: a multicenter study of deep learning for activity recognition in cataract surgical videos</article-title>. <source>Transl Vis Sci Technol</source>. (<year>2021</year>) <volume>10</volume>:<fpage>23</fpage>. <pub-id pub-id-type="doi">10.1167/tvst.10.13.23</pub-id><pub-id pub-id-type="pmid">34784415</pub-id></mixed-citation></ref>
<ref id="B52"><label>52.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lin</surname> <given-names>S</given-names></name> <name><surname>Qin</surname> <given-names>F</given-names></name> <name><surname>Peng</surname> <given-names>H</given-names></name> <name><surname>Bly</surname> <given-names>RA</given-names></name> <name><surname>Moe</surname> <given-names>KS</given-names></name> <name><surname>Hannaford</surname> <given-names>B</given-names></name></person-group>. <article-title>Multi-Frame feature aggregation for real-time instrument segmentation in endoscopic video</article-title>. <source>IEEE Robot Autom Lett</source>. (<year>2021</year>) <volume>6</volume>:<fpage>6773</fpage>&#x2013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1109/LRA.2021.3096156</pub-id></mixed-citation></ref>
<ref id="B53"><label>53.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Meireles</surname> <given-names>OR</given-names></name> <name><surname>Rosman</surname> <given-names>G</given-names></name> <name><surname>Altieri</surname> <given-names>MS</given-names></name> <name><surname>Carin</surname> <given-names>L</given-names></name> <name><surname>Hager</surname> <given-names>G</given-names></name> <name><surname>Madani</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>SAGES Consensus recommendations on an annotation framework for surgical video</article-title>. <source>Surg Endosc</source>. (<year>2021</year>) <volume>35</volume>:<fpage>4918</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-021-08578-9</pub-id><pub-id pub-id-type="pmid">34231065</pub-id></mixed-citation></ref>
<ref id="B54"><label>54.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kelly</surname> <given-names>JD</given-names></name> <name><surname>Petersen</surname> <given-names>A</given-names></name> <name><surname>Lendvay</surname> <given-names>TS</given-names></name> <name><surname>Kowalewski</surname> <given-names>TM</given-names></name></person-group>. <article-title>Bidirectional long short-term memory for surgical skill classification of temporally segmented tasks</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2020</year>) <volume>15</volume>:<fpage>2079</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02269-x</pub-id><pub-id pub-id-type="pmid">33000365</pub-id></mixed-citation></ref>
<ref id="B55"><label>55.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jain</surname> <given-names>S</given-names></name> <name><surname>Lee</surname> <given-names>S</given-names></name> <name><surname>Barber</surname> <given-names>SR</given-names></name> <name><surname>Chang</surname> <given-names>EH</given-names></name> <name><surname>Son</surname> <given-names>Y-J</given-names></name></person-group>. <article-title>Virtual reality based hybrid simulation for functional endoscopic sinus surgery</article-title>. <source>IISE Trans Healthc Syst Eng</source><italic>.</italic> (<year>2020</year>) <volume>10</volume>:<fpage>127</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1080/24725579.2019.1692263</pub-id></mixed-citation></ref>
<ref id="B56"><label>56.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>TS</given-names></name> <name><surname>O&#x0027;Brien</surname> <given-names>M</given-names></name> <name><surname>Zafar</surname> <given-names>S</given-names></name> <name><surname>Hager</surname> <given-names>GD</given-names></name> <name><surname>Sikder</surname> <given-names>S</given-names></name> <name><surname>Vedula</surname> <given-names>SS</given-names></name></person-group>. <article-title>Objective assessment of intraoperative technical skill in capsulorhexis using videos of cataract surgery</article-title>. <source>Int J CARS</source>. (<year>2019</year>) <volume>14</volume>:<fpage>1097</fpage>&#x2013;<lpage>105</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-01956-8</pub-id></mixed-citation></ref>
<ref id="B57"><label>57.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>McTaggart</surname> <given-names>MI</given-names></name> <name><surname>Higgins</surname> <given-names>WE</given-names></name></person-group>. <article-title>Robust video-frame classification for bronchoscopy</article-title>. In: <person-group person-group-type="editor"><name><surname>Fei</surname> <given-names>B</given-names></name> <name><surname>Linte</surname> <given-names>CA</given-names></name></person-group>, editors. <source>Medical Imaging 2019: Image-Guided Procedures, Robotic Interventions, and Modeling, Volume 61</source>. <publisher-loc>San Diego, United States</publisher-loc>: <publisher-name>SPIE</publisher-name>, (<year>2019</year>). p. <fpage>10.1117/12.2507290</fpage>. <pub-id pub-id-type="doi">10.1117/12.2507290</pub-id></mixed-citation></ref>
<ref id="B58"><label>58.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S</given-names></name> <name><surname>Xu</surname> <given-names>Z</given-names></name> <name><surname>Yan</surname> <given-names>C</given-names></name> <name><surname>Huang</surname> <given-names>J</given-names></name></person-group>. <article-title>Graph convolutional nets for tool presence detection in surgical videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Chung</surname> <given-names>ACS</given-names></name> <name><surname>Gee</surname> <given-names>JC</given-names></name> <name><surname>Yushkevich</surname> <given-names>PA</given-names></name> <name><surname>Bao</surname> <given-names>S</given-names></name></person-group>, editors. <source>Information Processing in Medical Imaging, Volume 11492</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2019</year>). p. <fpage>467</fpage>&#x2013;<lpage>78</lpage>.</mixed-citation></ref>
<ref id="B59"><label>59.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>El-Haddad</surname> <given-names>MT</given-names></name> <name><surname>Malone</surname> <given-names>JD</given-names></name> <name><surname>Hoang</surname> <given-names>NT</given-names></name> <name><surname>Tao</surname> <given-names>YK</given-names></name></person-group>. <article-title>Deep-learning based automated instrument tracking and adaptive-sampling of intraoperative OCT for video-rate volumetric imaging of ophthalmic surgical maneuvers</article-title>. In: <person-group person-group-type="editor"><name><surname>Izatt</surname> <given-names>JA</given-names></name> <name><surname>Fujimoto</surname> <given-names>JG</given-names></name></person-group>. <source>Optical Coherence Tomography and Coherence Domain Optical Methods in Biomedicine XXIII, 57</source>. <publisher-loc>San Francisco, United States</publisher-loc>: <publisher-name>SPIE</publisher-name> (<year>2019</year>). p. <fpage>10.1117/12.2511827</fpage>. <pub-id pub-id-type="doi">10.1117/12.2511827</pub-id></mixed-citation></ref>
<ref id="B60"><label>60.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Shvets</surname> <given-names>AA</given-names></name> <name><surname>Rakhlin</surname> <given-names>A</given-names></name> <name><surname>Kalinin</surname> <given-names>AA</given-names></name> <name><surname>Iglovikov</surname> <given-names>VI</given-names></name></person-group>. <article-title>Automatic instrument segmentation in robot-assisted surgery using deep learning</article-title>. In: <conf-name>2018 17th IEEE international conference on machine learning and applications (ICMLA)</conf-name>; <conf-date>624&#x2013;628</conf-date>; (<year>2018</year>). <pub-id pub-id-type="doi">10.1109/ICMLA.2018.00100</pub-id></mixed-citation></ref>
<ref id="B61"><label>61.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>A</given-names></name> <name><surname>Yeung</surname> <given-names>S</given-names></name> <name><surname>Jopling</surname> <given-names>J</given-names></name> <name><surname>Krause</surname> <given-names>J</given-names></name> <name><surname>Azagury</surname> <given-names>D</given-names></name> <name><surname>Milstein</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Tool detection and operative skill assessment in surgical videos using region-based convolutional neural networks</article-title>. In: <source>2018 IEEE Winter Conference on Applications of Computer Vision (WACV)</source> (<year>2018a</year>). p. <fpage>691</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/WACV.2018.00081</pub-id></mixed-citation></ref>
<ref id="B62"><label>62.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>B</given-names></name> <name><surname>Fung</surname> <given-names>A</given-names></name> <name><surname>Torabi</surname> <given-names>M</given-names></name> <name><surname>Barker</surname> <given-names>J</given-names></name> <name><surname>Foley</surname> <given-names>G</given-names></name> <name><surname>Abukhalil</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>C-ECT: online surgical phase recognition with cross-enhancement causal transformer</article-title>. In: <source>2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), (Cartagena, Colombia: IEEE)</source> (<year>2023a</year>). p. <fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.1109/ISBI53787.2023.10230841</pub-id></mixed-citation></ref>
<ref id="B63"><label>63.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>M</given-names></name> <name><surname>Ye</surname> <given-names>M</given-names></name> <name><surname>Rafii-Tari</surname> <given-names>H</given-names></name></person-group>. <article-title>Automatic biopsy tool presence and episode recognition in robotic bronchoscopy using a multi-task vision transformer network</article-title>. In: <conf-name>2022 International Conference on Robotics and Automation (ICRA)</conf-name>; <conf-date>7349&#x2013;55</conf-date>; <conf-loc>Philadelphia, PA, USA</conf-loc>: <publisher-name>IEEE</publisher-name> (<year>2022</year>). <pub-id pub-id-type="doi">10.1109/ICRA46639.2022.9811982</pub-id></mixed-citation></ref>
<ref id="B64"><label>64.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>JW</given-names></name> <name><surname>Zhang</surname> <given-names>P</given-names></name> <name><surname>Gehlbach</surname> <given-names>P</given-names></name> <name><surname>Iordachita</surname> <given-names>I</given-names></name> <name><surname>Kobilarov</surname> <given-names>M</given-names></name></person-group>. <article-title>Towards autonomous eye surgery by combining deep imitation learning with optimal control</article-title>. <source>Proc Mach Learn Res</source><italic>.</italic> (<year>2021</year>) <volume>155</volume>:<fpage>2347</fpage>&#x2013;<lpage>58</lpage>.<pub-id pub-id-type="pmid">34712957</pub-id></mixed-citation></ref>
<ref id="B65"><label>65.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Aldeen Jalal</surname> <given-names>N</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>A comparative evaluation of spatial pooling methods for surgical tool detection</article-title>. <source>Curr Dir Biomed Eng</source>. (<year>2023</year>) <volume>9</volume>:<fpage>214</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2023-1054</pub-id></mixed-citation></ref>
<ref id="B66"><label>66.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arabian</surname> <given-names>H</given-names></name> <name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Krueger-Ziolek</surname> <given-names>S</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>P-CSEM: an attention module for improved laparoscopic surgical tool detection</article-title>. <source>Sensors</source>. (<year>2023</year>) <volume>23</volume>:<fpage>7257</fpage>. <pub-id pub-id-type="doi">10.3390/s23167257</pub-id><pub-id pub-id-type="pmid">37631791</pub-id></mixed-citation></ref>
<ref id="B67"><label>67.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Reiter</surname> <given-names>W</given-names></name></person-group>. <article-title>Domain generalization improves end-to-end object detection for real-time surgical tool detection</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2022</year>) <volume>18</volume>:<fpage>939</fpage>&#x2013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-022-02823-9</pub-id><pub-id pub-id-type="pmid">36581742</pub-id></mixed-citation></ref>
<ref id="B68"><label>68.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Alshirbaji</surname> <given-names>TA</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Arabian</surname> <given-names>H</given-names></name> <name><surname>Laufer</surname> <given-names>B</given-names></name> <name><surname>Krueger-Ziolek</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Laparoscopic video analysis using temporal, attention, and multi-feature fusion based-approaches</article-title>. <source>Sensors</source>. (<year>2023</year>) <volume>23</volume>:<fpage>1958</fpage>. <pub-id pub-id-type="doi">10.3390/s23041958</pub-id><pub-id pub-id-type="pmid">36850554</pub-id></mixed-citation></ref>
<ref id="B69"><label>69.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Yamlahi</surname> <given-names>A</given-names></name> <name><surname>Tran</surname> <given-names>TN</given-names></name> <name><surname>Godau</surname> <given-names>P</given-names></name> <name><surname>Schellenberg</surname> <given-names>M</given-names></name> <name><surname>Michael</surname> <given-names>D</given-names></name> <name><surname>Smidt</surname> <given-names>F-H</given-names></name><etal/></person-group> <article-title>Self-distillation for surgical action recognition</article-title>. In: <person-group person-group-type="editor"><name><surname>Greenspan</surname> <given-names>H</given-names></name> <name><surname>Madabhushi</surname> <given-names>A</given-names></name> <name><surname>Mousavi</surname> <given-names>P</given-names></name> <name><surname>Salcudean</surname> <given-names>S</given-names></name> <name><surname>Duncan</surname> <given-names>J</given-names></name> <name><surname>Syeda-Mahmood</surname> <given-names>T</given-names></name><etal/></person-group> editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2023</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name> (<year>2023</year>). p. <fpage>637</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-43996-4_61</pub-id></mixed-citation></ref>
<ref id="B70"><label>70.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arabian</surname> <given-names>H</given-names></name> <name><surname>Dalla</surname> <given-names>FA</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Alshirbaji</surname> <given-names>TA</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>Attention networks for improving surgical tool classification in laparoscopic videos</article-title>. <source>Curr Dir Biomed Eng</source>. (<year>2022</year>) <volume>8</volume>:<fpage>676</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2022-1172</pub-id></mixed-citation></ref>
<ref id="B71"><label>71.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Aldeen Jalal</surname> <given-names>N</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>Neural network classification of surgical tools in gynecological videos</article-title>. <source>Curr Dir Biomed Eng</source><italic>.</italic> (<year>2022</year>) <volume>8</volume>:<fpage>644</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2022-1164</pub-id></mixed-citation></ref>
<ref id="B72"><label>72.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Philipp</surname> <given-names>M</given-names></name> <name><surname>Alperovich</surname> <given-names>A</given-names></name> <name><surname>Gutt-Will</surname> <given-names>M</given-names></name> <name><surname>Mathis</surname> <given-names>A</given-names></name> <name><surname>Saur</surname> <given-names>S</given-names></name> <name><surname>Raabe</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Dynamic CNNs using uncertainty to overcome domain generalization for surgical instrument localization</article-title>. In: <source>2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)</source>. <publisher-loc>Waikoloa, HI, USA</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2022</year>). p. <fpage>1727</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1109/WACV51458.2022.00179</pub-id></mixed-citation></ref>
<ref id="B73"><label>73.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aspart</surname> <given-names>F</given-names></name> <name><surname>Bolmgren</surname> <given-names>JL</given-names></name> <name><surname>Lavanchy</surname> <given-names>JL</given-names></name> <name><surname>Beldi</surname> <given-names>G</given-names></name> <name><surname>Woods</surname> <given-names>MS</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>Clipassistnet: bringing real-time safety feedback to operating rooms</article-title>. <source>Int J CARS</source>. (<year>2022</year>) <volume>17</volume>:<fpage>5</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-021-02441-x</pub-id></mixed-citation></ref>
<ref id="B74"><label>74.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alshirbaji</surname> <given-names>TA</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>Assessing generalisation capabilities of CNN models for surgical tool classification</article-title>. <source>Curr Dir Biomed Eng</source><italic>.</italic> (<year>2021</year>) <volume>7</volume>:<fpage>476</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2021-2121</pub-id></mixed-citation></ref>
<ref id="B75"><label>75.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>M&#x00F6;ller</surname> <given-names>K</given-names></name></person-group>. <article-title>A deep learning spatial-temporal framework for detecting surgical tools in laparoscopic videos</article-title>. <source>Biomed Signal Process Control</source>. (<year>2021</year>) <volume>68</volume>:<fpage>102801</fpage>. <pub-id pub-id-type="doi">10.1016/j.bspc.2021.102801</pub-id></mixed-citation></ref>
<ref id="B76"><label>76.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Alshirbaji</surname> <given-names>TA</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>A deep learning framework for recognising surgical phases in laparoscopic videos</article-title>. <source>IFAC Pap</source>. (<year>2021</year>) <volume>54</volume>:<fpage>334</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.ifacol.2021.10.278</pub-id></mixed-citation></ref>
<ref id="B77"><label>77.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Docherty</surname> <given-names>PD</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name> <name><surname>Moeller</surname> <given-names>K</given-names></name></person-group>. <article-title>Surgical tool detection in laparoscopic videos by modeling temporal dependencies between adjacent frames</article-title>. In: <person-group person-group-type="editor"><name><surname>Jarm</surname> <given-names>T</given-names></name> <name><surname>Cvetkoska</surname> <given-names>A</given-names></name> <name><surname>Mahni&#x010D;-Kalamiza</surname> <given-names>S</given-names></name> <name><surname>Miklavcic</surname> <given-names>D</given-names></name></person-group>, editors. <source>8th European Medical and Biological Engineering Conference</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2021</year>). p. <fpage>1045</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-64610-3_117</pub-id></mixed-citation></ref>
<ref id="B78"><label>78.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bieck</surname> <given-names>R</given-names></name> <name><surname>Heuermann</surname> <given-names>K</given-names></name> <name><surname>Pirlich</surname> <given-names>M</given-names></name> <name><surname>Neumann</surname> <given-names>J</given-names></name> <name><surname>Neumuth</surname> <given-names>T</given-names></name></person-group>. <article-title>Language-based translation and prediction of surgical navigation steps for endoscopic wayfinding assistance in minimally invasive surgery</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2020</year>) <volume>15</volume>:<fpage>2089</fpage>&#x2013;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02264-2</pub-id><pub-id pub-id-type="pmid">33037490</pub-id></mixed-citation></ref>
<ref id="B79"><label>79.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ivantsits</surname> <given-names>M</given-names></name> <name><surname>Tautz</surname> <given-names>L</given-names></name> <name><surname>S&#x00FC;ndermann</surname> <given-names>S</given-names></name> <name><surname>Wamala</surname> <given-names>I</given-names></name> <name><surname>Kempfert</surname> <given-names>J</given-names></name> <name><surname>Kuehne</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>DL-based segmentation of endoscopic scenes for mitral valve repair</article-title>. <source>Curr Dir Biomed Eng</source>. (<year>2020</year>) <volume>6</volume>:<fpage>20200017</fpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2020-0017</pub-id></mixed-citation></ref>
<ref id="B80"><label>80.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tamer Abdulbaki</surname> <given-names>A</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>M&#x00F6;ller</surname> <given-names>K</given-names></name></person-group>. <article-title>A convolutional neural network with a two-stage LSTM model for tool presence detection in laparoscopic videos</article-title>. <source>Curr Dir Biomed Eng</source><italic>.</italic> (<year>2020</year>) <volume>6</volume>:<fpage>20200002</fpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2020-0002</pub-id></mixed-citation></ref>
<ref id="B81"><label>81.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sahu</surname> <given-names>M</given-names></name> <name><surname>Szengel</surname> <given-names>A</given-names></name> <name><surname>Mukhopadhyay</surname> <given-names>A</given-names></name> <name><surname>Zachow</surname> <given-names>S</given-names></name></person-group>. <article-title>Surgical phase recognition by learning phase transitions</article-title>. <source>Curr Dir Biomed Eng</source>. (<year>2020</year>) <volume>6</volume>:<fpage>20200037</fpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2020-0037</pub-id></mixed-citation></ref>
<ref id="B82"><label>82.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>Alshirbaji</surname> <given-names>TA</given-names></name> <name><surname>M&#x00F6;ller</surname> <given-names>K</given-names></name></person-group>. <article-title>Predicting surgical phases using CNN-NARX neural network</article-title>. <source>Curr Dir Biomed Eng</source><italic>.</italic> (<year>2019</year>) <volume>5</volume>:<fpage>405</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2019-0102</pub-id></mixed-citation></ref>
<ref id="B83"><label>83.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bodenstedt</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Active learning using deep Bayesian networks for surgical workflow analysis</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2019</year>) <volume>14</volume>:<fpage>1079</fpage>&#x2013;<lpage>87</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-01963-9</pub-id><pub-id pub-id-type="pmid">30968355</pub-id></mixed-citation></ref>
<ref id="B84"><label>84.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Funke</surname> <given-names>I</given-names></name> <name><surname>Bodenstedt</surname> <given-names>S</given-names></name> <name><surname>Oehme</surname> <given-names>F</given-names></name> <name><surname>Von Bechtolsheim</surname> <given-names>F</given-names></name> <name><surname>Weitz</surname> <given-names>J</given-names></name> <name><surname>Speidel</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Using 3D convolutional neural networks to learn spatiotemporal features for automatic surgical gesture recognition in video</article-title>. In: <person-group person-group-type="editor"><name><surname>Shen</surname> <given-names>D</given-names></name> <name><surname>Liu</surname> <given-names>T</given-names></name> <name><surname>Peters</surname> <given-names>TM</given-names></name> <name><surname>Staib</surname> <given-names>LH</given-names></name> <name><surname>Essert</surname> <given-names>C</given-names></name> <name><surname>Zhou</surname> <given-names>S</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2019</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2019</year>). p. <fpage>467</fpage>&#x2013;<lpage>75</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-32254-0_52</pub-id></mixed-citation></ref>
<ref id="B85"><label>85.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Prellberg</surname> <given-names>J</given-names></name> <name><surname>Kramer</surname> <given-names>O</given-names></name></person-group>. <article-title>Multi-label classification of surgical tools with convolutional neural networks</article-title>. In: <conf-name>2018 International Joint Conference on Neural Networks (IJCNN)</conf-name>; <conf-date>1-8</conf-date>; <conf-loc>Rio de Janeiro</conf-loc>; <publisher-name>IEEE</publisher-name>; (<year>2018</year>). <pub-id pub-id-type="doi">10.1109/IJCNN.2018.8489647</pub-id></mixed-citation></ref>
<ref id="B86"><label>86.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdulbaki Alshirbaji</surname> <given-names>T</given-names></name> <name><surname>Jalal</surname> <given-names>NA</given-names></name> <name><surname>M&#x00F6;ller</surname> <given-names>K</given-names></name></person-group>. <article-title>Surgical tool classification in laparoscopic videos using convolutional neural network</article-title>. <source>Curr Dir Biomed Eng</source>. (<year>2018</year>) <volume>4</volume>:<fpage>407</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1515/cdbme-2018-0097</pub-id></mixed-citation></ref>
<ref id="B87"><label>87.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Petscharnig</surname> <given-names>S</given-names></name> <name><surname>Schoffmann</surname> <given-names>K</given-names></name> <name><surname>Benois-Pineau</surname> <given-names>J</given-names></name> <name><surname>Chaabouni</surname> <given-names>S</given-names></name> <name><surname>Keckstein</surname> <given-names>J</given-names></name></person-group>. <article-title>Early and late fusion of temporal information for classification of surgical actions in laparoscopic gynecology</article-title>. In: <conf-name>2018 IEEE 31st International Symposium on Computer-Based Medical Systems (CBMS)</conf-name>; <conf-date>369&#x2013;74</conf-date>; <conf-loc>Karlstad</conf-loc>; <publisher-name>IEEE</publisher-name>; (<year>2018</year>). <pub-id pub-id-type="doi">10.1109/CBMS.2018.00071</pub-id></mixed-citation></ref>
<ref id="B88"><label>88.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Sharma</surname> <given-names>S</given-names></name> <name><surname>Murali</surname> <given-names>A</given-names></name> <name><surname>Alapatt</surname> <given-names>D</given-names></name> <name><surname>Vardazaryan</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Cholectriplet2022: show me a tool and tell me the triplet &#x2014; an endoscopic vision challenge for surgical action triplet detection</article-title>. <source>Med Image Anal</source>. (<year>2023b</year>) <volume>89</volume>:<fpage>102888</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.102888</pub-id></mixed-citation></ref>
<ref id="B89"><label>89.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ramesh</surname> <given-names>S</given-names></name> <name><surname>Srivastav</surname> <given-names>V</given-names></name> <name><surname>Alapatt</surname> <given-names>D</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Murali</surname> <given-names>A</given-names></name> <name><surname>Sestini</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Dissecting self-supervised learning methods for surgical computer vision</article-title>. <source>Med Image Anal</source>. (<year>2023</year>) <volume>88</volume>:<fpage>102844</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.102844</pub-id><pub-id pub-id-type="pmid">37270898</pub-id></mixed-citation></ref>
<ref id="B90"><label>90.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname> <given-names>S</given-names></name> <name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>Rendezvous in time: an attention-based temporal fusion approach for surgical triplet recognition</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2023</year>) <volume>18</volume>:<fpage>1053</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-023-02914-1</pub-id><pub-id pub-id-type="pmid">37097518</pub-id></mixed-citation></ref>
<ref id="B91"><label>91.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Alapatt</surname> <given-names>D</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Vardazaryan</surname> <given-names>A</given-names></name> <name><surname>Xia</surname> <given-names>F</given-names></name> <name><surname>Zhao</surname> <given-names>Z</given-names></name><etal/></person-group> <article-title>Cholectriplet2021: a benchmark challenge for surgical action triplet recognition</article-title>. <source>Med Image Anal</source>. (<year>2023a</year>) <volume>86</volume>:<fpage>102803</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.102803</pub-id></mixed-citation></ref>
<ref id="B92"><label>92.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Madad Zadeh</surname> <given-names>S</given-names></name> <name><surname>Fran&#x00E7;ois</surname> <given-names>T</given-names></name> <name><surname>Comptour</surname> <given-names>A</given-names></name> <name><surname>Canis</surname> <given-names>M</given-names></name> <name><surname>Bourdel</surname> <given-names>N</given-names></name> <name><surname>Bartoli</surname> <given-names>A</given-names></name></person-group>. <article-title>SurgAI3.8K: a labeled dataset of gynecologic organs in laparoscopy with application to automatic augmented reality surgical guidance</article-title>. <source>J Minim Invasive Gynecol</source>. (<year>2023</year>) <volume>30</volume>:<fpage>397</fpage>&#x2013;<lpage>405</lpage>. <pub-id pub-id-type="doi">10.1016/j.jmig.2023.01.012</pub-id><pub-id pub-id-type="pmid">36720429</pub-id></mixed-citation></ref>
<ref id="B93"><label>93.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sestini</surname> <given-names>L</given-names></name> <name><surname>Rosa</surname> <given-names>B</given-names></name> <name><surname>De Momi</surname> <given-names>E</given-names></name> <name><surname>Ferrigno</surname> <given-names>G</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>FUN-SIS: a fully unsupervised approach for surgical instrument segmentation</article-title>. <source>Med Image Anal</source><italic>.</italic> (<year>2023</year>) <volume>85</volume>:<fpage>102751</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.102751</pub-id><pub-id pub-id-type="pmid">36716700</pub-id></mixed-citation></ref>
<ref id="B94"><label>94.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Takeuchi</surname> <given-names>M</given-names></name> <name><surname>Collins</surname> <given-names>T</given-names></name> <name><surname>Ndagijimana</surname> <given-names>A</given-names></name> <name><surname>Kawakubo</surname> <given-names>H</given-names></name> <name><surname>Kitagawa</surname> <given-names>Y</given-names></name> <name><surname>Marescaux</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Automatic surgical phase recognition in laparoscopic inguinal hernia repair with artificial intelligence</article-title>. <source>Hernia</source>. (<year>2022</year>) <volume>26</volume>:<fpage>1669</fpage>&#x2013;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1007/s10029-022-02621-x</pub-id><pub-id pub-id-type="pmid">35536371</pub-id></mixed-citation></ref>
<ref id="B95"><label>95.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mascagni</surname> <given-names>P</given-names></name> <name><surname>Alapatt</surname> <given-names>D</given-names></name> <name><surname>Laracca</surname> <given-names>GG</given-names></name> <name><surname>Guerriero</surname> <given-names>L</given-names></name> <name><surname>Spota</surname> <given-names>A</given-names></name> <name><surname>Fiorillo</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Multicentric validation of EndoDigest: a computer vision platform for video documentation of the critical view of safety in laparoscopic cholecystectomy</article-title>. <source>Surg Endosc</source>. (<year>2022a</year>) <volume>36</volume>:<fpage>8379</fpage>&#x2013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-022-09112-1</pub-id></mixed-citation></ref>
<ref id="B96"><label>96.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Gonzalez</surname> <given-names>C</given-names></name> <name><surname>Seeliger</surname> <given-names>B</given-names></name> <name><surname>Mascagni</surname> <given-names>P</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Rendezvous: attention mechanisms for the recognition of surgical action triplets in endoscopic videos</article-title>. <source>Med Image Anal</source>. (<year>2022</year>) <volume>78</volume>:<fpage>102433</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2022.102433</pub-id><pub-id pub-id-type="pmid">35398658</pub-id></mixed-citation></ref>
<ref id="B97"><label>97.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huaulm&#x00E9;</surname> <given-names>A</given-names></name> <name><surname>Sarikaya</surname> <given-names>D</given-names></name> <name><surname>Le Mut</surname> <given-names>K</given-names></name> <name><surname>Despinoy</surname> <given-names>F</given-names></name> <name><surname>Long</surname> <given-names>Y</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name><etal/></person-group> <article-title>MIcro-surgical anastomose workflow recognition challenge report</article-title>. <source>Comput Methods Programs Biomed</source>. (<year>2021</year>) <volume>212</volume>:<fpage>106452</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2021.106452</pub-id></mixed-citation></ref>
<ref id="B98"><label>98.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ramesh</surname> <given-names>S</given-names></name> <name><surname>Dall&#x0027;Alba</surname> <given-names>D</given-names></name> <name><surname>Gonzalez</surname> <given-names>C</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Mascagni</surname> <given-names>P</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Multi-task temporal convolutional networks for joint recognition of surgical phases and steps in gastric bypass procedures</article-title>. <source>Int J CARS</source>. (<year>2021</year>) <volume>16</volume>:<fpage>1111</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-021-02388-z</pub-id></mixed-citation></ref>
<ref id="B99"><label>99.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Gonzalez</surname> <given-names>C</given-names></name> <name><surname>Yu</surname> <given-names>T</given-names></name> <name><surname>Mascagni</surname> <given-names>P</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name> <name><surname>Marescaux</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Recognition of instrument-tissue interactions in endoscopic videos via action triplets</article-title>. In: <person-group person-group-type="editor"><name><surname>Martel</surname> <given-names>AL</given-names></name> <name><surname>Abolmaesumi</surname> <given-names>P</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name> <name><surname>Mateus</surname> <given-names>D</given-names></name> <name><surname>Zuluaga</surname> <given-names>MA</given-names></name> <name><surname>Zhou</surname> <given-names>SK</given-names></name><etal/></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2020</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2020</year>). <fpage>364</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-59716-0_35</pub-id></mixed-citation></ref>
<ref id="B100"><label>100.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rosa</surname> <given-names>B</given-names></name> <name><surname>Bordoux</surname> <given-names>V</given-names></name> <name><surname>Nageotte</surname> <given-names>F</given-names></name></person-group>. <article-title>Combining differential kinematics and optical flow for automatic labeling of Continuum robots in minimally invasive surgery</article-title>. <source>Front Robot AI</source>. (<year>2019</year>) <volume>6</volume>:<fpage>86</fpage>. <pub-id pub-id-type="doi">10.3389/frobt.2019.00086</pub-id><pub-id pub-id-type="pmid">33501101</pub-id></mixed-citation></ref>
<ref id="B101"><label>101.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nwoye</surname> <given-names>CI</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name> <name><surname>Marescaux</surname> <given-names>J</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>Weakly supervised convolutional LSTM approach for tool tracking in laparoscopic videos</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2019</year>) <volume>14</volume>:<fpage>1059</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-01958-6</pub-id><pub-id pub-id-type="pmid">30968356</pub-id></mixed-citation></ref>
<ref id="B102"><label>102.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Al Hajj</surname> <given-names>H</given-names></name> <name><surname>Lamard</surname><given-names>M</given-names></name> <name><surname>Conze</surname><given-names>P-H</given-names></name> <name><surname>Roychowdhury</surname><given-names>S</given-names></name> <name><surname>Hu</surname><given-names>X</given-names></name> <name><surname>Mar&#x0161;alkait&#x0117;</surname><given-names>G</given-names></name><etal/></person-group> <article-title>CATARACTS: challenge on automatic tool annotation for CATARACT surgery</article-title>. <source>Med Image Anal</source><italic>.</italic> (<year>2019</year>) <volume>52</volume>:<fpage>24</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2018.11.008</pub-id><pub-id pub-id-type="pmid">30468970</pub-id></mixed-citation></ref>
<ref id="B103"><label>103.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Vardazaryan</surname> <given-names>A</given-names></name> <name><surname>Mutter</surname> <given-names>D</given-names></name> <name><surname>Marescaux</surname> <given-names>J</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name></person-group>. <article-title>Weakly- Supervised learning for tool localization in laparoscopic videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Stoyanov</surname> <given-names>D</given-names></name> <name><surname>Taylor</surname> <given-names>Z</given-names></name> <name><surname>Balocco</surname> <given-names>S</given-names></name> <name><surname>Sznitman</surname> <given-names>R</given-names></name> <name><surname>Martel</surname> <given-names>A</given-names></name> <name><surname>Maier-Hein</surname> <given-names>L</given-names></name><etal/></person-group>, editors. <source>Intravascular Imaging and Computer Assisted Stenting and Large-Scale Annotation of Biomedical Data and Expert Label Synthesis</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2018</year>). p. <fpage>169</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-01364-6_19</pub-id></mixed-citation></ref>
<ref id="B104"><label>104.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Huber</surname> <given-names>M</given-names></name> <name><surname>Ourselin</surname> <given-names>S</given-names></name> <name><surname>Bergeles</surname> <given-names>C</given-names></name> <name><surname>Vercauteren</surname> <given-names>T</given-names></name></person-group>. <article-title>Deep homography prediction for endoscopic camera motion imitation learning</article-title>. In <person-group person-group-type="editor"><name><surname>Greenspan</surname> <given-names>H</given-names></name> <name><surname>Madabhushi</surname> <given-names>A</given-names></name> <name><surname>Mousavi</surname> <given-names>P</given-names></name> <name><surname>Salcudean</surname> <given-names>S</given-names></name> <name><surname>Duncan</surname> <given-names>J</given-names></name> <name><surname>Syeda-Mahmood</surname> <given-names>T</given-names></name><etal/></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2023</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name> (<year>2023</year>). p. <fpage>217</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-43996-4_21</pub-id></mixed-citation></ref>
<ref id="B105"><label>105.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lam</surname> <given-names>K</given-names></name> <name><surname>Lo</surname> <given-names>FP-W</given-names></name> <name><surname>An</surname> <given-names>Y</given-names></name> <name><surname>Darzi</surname> <given-names>A</given-names></name> <name><surname>Kinross</surname> <given-names>JM</given-names></name> <name><surname>Purkayastha</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Deep learning for instrument detection and assessment of operative skill in surgical videos</article-title>. <source>IEEE Trans Med Robot Bionics</source>. (<year>2022</year>) <volume>4</volume>:<fpage>1068</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1109/TMRB.2022.3214377</pub-id></mixed-citation></ref>
<ref id="B106"><label>106.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Yu</surname> <given-names>Y</given-names></name> <name><surname>Chen</surname> <given-names>C</given-names></name> <name><surname>Zhao</surname> <given-names>Z</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Exploring intra- and inter-video relation for surgical semantic scene segmentation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2022</year>) <volume>41</volume>:<fpage>2991</fpage>&#x2013;<lpage>3002</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2022.3177077</pub-id><pub-id pub-id-type="pmid">35604967</pub-id></mixed-citation></ref>
<ref id="B107"><label>107.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>S</given-names></name> <name><surname>Ali</surname> <given-names>S</given-names></name> <name><surname>Goldsmith</surname> <given-names>L</given-names></name> <name><surname>Turney</surname> <given-names>B</given-names></name> <name><surname>Rittscher</surname> <given-names>J</given-names></name></person-group>. <article-title>Multi-class motion-based semantic segmentation for ureteroscopy and laser lithotripsy</article-title>. <source>Comput Med Imaging Graph</source><italic>.</italic> (<year>2022</year>) <volume>101</volume>:<fpage>102112</fpage>. <pub-id pub-id-type="doi">10.1016/j.compmedimag.2022.102112</pub-id><pub-id pub-id-type="pmid">36030620</pub-id></mixed-citation></ref>
<ref id="B108"><label>108.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tukra</surname> <given-names>S</given-names></name> <name><surname>Marcus</surname> <given-names>HJ</given-names></name> <name><surname>Giannarou</surname> <given-names>S</given-names></name></person-group>. <article-title>See-through vision with unsupervised scene occlusion reconstruction</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source><italic>.</italic> (<year>2022</year>) <volume>1</volume>:<comment>10.1109/TPAMI.2021.3058410</comment>. <pub-id pub-id-type="doi">10.1109/TPAMI.2021.3058410</pub-id></mixed-citation></ref>
<ref id="B109"><label>109.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kadkhodamohammadi</surname> <given-names>A</given-names></name> <name><surname>Luengo</surname> <given-names>I</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>PATG: position-aware temporal graph networks for surgical phase recognition on laparoscopic videos</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2022</year>) <volume>17</volume>:<fpage>849</fpage>&#x2013;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-022-02600-8</pub-id><pub-id pub-id-type="pmid">35353299</pub-id></mixed-citation></ref>
<ref id="B110"><label>110.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Bano</surname> <given-names>S</given-names></name> <name><surname>Page</surname> <given-names>A-S</given-names></name> <name><surname>Deprest</surname> <given-names>J</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name> <name><surname>Vasconcelos</surname> <given-names>F</given-names></name></person-group>. <article-title>Large-scale surgical workflow segmentation for laparoscopic sacrocolpopexy</article-title>. <source>Int J CARS</source>. (<year>2022b</year>) <volume>17</volume>:<fpage>467</fpage>&#x2013;<lpage>77</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-021-02544-5</pub-id></mixed-citation></ref>
<ref id="B111"><label>111.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>FX</given-names></name> <name><surname>Al Moubayed</surname> <given-names>N</given-names></name> <name><surname>Shum</surname> <given-names>HPH</given-names></name></person-group>. <article-title>Towards graph representation learning based surgical workflow anticipation</article-title>. In: <conf-name>2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI)</conf-name>; <conf-date>01&#x2013;4</conf-date>; <conf-loc>Ioannina, Greece</conf-loc>; <publisher-name>IEEE</publisher-name>; <article-title>(</article-title><year>2022</year>). <pub-id pub-id-type="doi">10.1109/BHI56158.2022.9926801</pub-id></mixed-citation></ref>
<ref id="B112"><label>112.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Shao</surname> <given-names>M</given-names></name> <name><surname>Clark</surname> <given-names>J</given-names></name> <name><surname>Huson</surname> <given-names>D</given-names></name> <name><surname>Hardeberg</surname> <given-names>J</given-names></name></person-group>. <article-title>Real-time style transfer for videos to enhance the realism of simulation of laparoscopic surgeries</article-title>. In: <conf-name>2022 10th European Workshop on Visual Information Processing (EUVIP)</conf-name>; <conf-date>1&#x2013;6</conf-date>; <publisher-name>IEEE</publisher-name>; <conf-loc>Lisbon, Portugal</conf-loc>; (<year>2022</year>). <pub-id pub-id-type="doi">10.1109/EUVIP53989.2022.9922706</pub-id></mixed-citation></ref>
<ref id="B113"><label>113.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Grammatikopoulou</surname> <given-names>M</given-names></name> <name><surname>Flouty</surname> <given-names>E</given-names></name> <name><surname>Kadkhodamohammadi</surname> <given-names>A</given-names></name> <name><surname>Quellec</surname> <given-names>G</given-names></name> <name><surname>Chow</surname> <given-names>A</given-names></name> <name><surname>Nehme</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>CaDIS: cataract dataset for surgical RGB-image segmentation</article-title>. <source>Med Image Anal</source>. (<year>2021</year>) <volume>71</volume>:<fpage>102053</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.102053</pub-id><pub-id pub-id-type="pmid">33864969</pub-id></mixed-citation></ref>
<ref id="B114"><label>114.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garcia-Peraza-Herrera</surname> <given-names>LC</given-names></name> <name><surname>Fidon</surname> <given-names>L</given-names></name> <name><surname>D&#x2019;Ettorre</surname> <given-names>C</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name> <name><surname>Vercauteren</surname> <given-names>T</given-names></name> <name><surname>Ourselin</surname> <given-names>S</given-names></name></person-group>. <article-title>Image compositing for segmentation of surgical tools without manual annotations</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2021</year>) <volume>40</volume>:<fpage>1450</fpage>&#x2013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2021.3057884</pub-id><pub-id pub-id-type="pmid">33556005</pub-id></mixed-citation></ref>
<ref id="B115"><label>115.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname> <given-names>C</given-names></name> <name><surname>Singh</surname> <given-names>H</given-names></name> <name><surname>Orihuela-Espina</surname> <given-names>F</given-names></name> <name><surname>Darzi</surname> <given-names>A</given-names></name> <name><surname>Sodergren</surname> <given-names>MH</given-names></name></person-group>. <article-title>Visual gaze patterns reveal surgeons&#x2019; ability to identify risk of bile duct injury during laparoscopic cholecystectomy</article-title>. <source>HPB (Oxford)</source>. (<year>2021</year>) <volume>23</volume>:<fpage>715</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.1016/j.hpb.2020.09.007</pub-id><pub-id pub-id-type="pmid">32988756</pub-id></mixed-citation></ref>
<ref id="B116"><label>116.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Colleoni</surname> <given-names>E</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Robotic instrument segmentation with image-to-image translation</article-title>. <source>IEEE Robot Autom Lett</source><italic>.</italic> (<year>2021</year>) <volume>6</volume>:<fpage>935</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1109/LRA.2021.3056354</pub-id></mixed-citation></ref>
<ref id="B117"><label>117.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ravasio</surname> <given-names>CS</given-names></name> <name><surname>Pissas</surname> <given-names>T</given-names></name> <name><surname>Bloch</surname> <given-names>E</given-names></name> <name><surname>Flores</surname> <given-names>B</given-names></name> <name><surname>Jalali</surname> <given-names>S</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Learned optical flow for intra-operative tracking of the retinal fundus</article-title>. <source>Int J CARS</source>. (<year>2020</year>) <volume>15</volume>:<fpage>827</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02160-9</pub-id></mixed-citation></ref>
<ref id="B118"><label>118.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Colleoni</surname> <given-names>E</given-names></name> <name><surname>Edwards</surname> <given-names>P</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Synthetic and real inputs for tool segmentation in robotic surgery</article-title>. In: <person-group person-group-type="editor"><name><surname>Martel</surname> <given-names>AL</given-names></name> <name><surname>Abolmaesumi</surname> <given-names>P</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name> <name><surname>Mateus</surname> <given-names>D</given-names></name> <name><surname>Zuluaga</surname> <given-names>MA</given-names></name> <name><surname>Zhou</surname> <given-names>SK</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2020</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2020</year>). p. <fpage>700</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-59716-0_67</pub-id></mixed-citation></ref>
<ref id="B119"><label>119.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fuentes-Hurtado</surname> <given-names>F</given-names></name> <name><surname>Kadkhodamohammadi</surname> <given-names>A</given-names></name> <name><surname>Flouty</surname> <given-names>E</given-names></name> <name><surname>Barbarisi</surname> <given-names>S</given-names></name> <name><surname>Luengo</surname> <given-names>I</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Easylabels: weak labels for scene segmentation in laparoscopic videos</article-title>. <source>Int J CARS</source>. (<year>2019</year>) <volume>14</volume>:<fpage>1247</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-02003-2</pub-id></mixed-citation></ref>
<ref id="B120"><label>120.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Du</surname> <given-names>X</given-names></name> <name><surname>Kurmann</surname> <given-names>T</given-names></name> <name><surname>Chang</surname> <given-names>P-L</given-names></name> <name><surname>Allan</surname> <given-names>M</given-names></name> <name><surname>Ourselin</surname> <given-names>S</given-names></name> <name><surname>Sznitman</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Articulated multi-instrument 2-D pose estimation using fully convolutional networks</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2018</year>) <volume>37</volume>:<fpage>1276</fpage>&#x2013;<lpage>87</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2017.2787672</pub-id><pub-id pub-id-type="pmid">29727290</pub-id></mixed-citation></ref>
<ref id="B121"><label>121.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zisimopoulos</surname> <given-names>O</given-names></name> <name><surname>Flouty</surname> <given-names>E</given-names></name> <name><surname>Luengo</surname> <given-names>I</given-names></name> <name><surname>Giataganas</surname> <given-names>P</given-names></name> <name><surname>Nehme</surname> <given-names>J</given-names></name> <name><surname>Chow</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Deepphase: surgical phase recognition in CATARACTS videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Frangi</surname> <given-names>AF</given-names></name> <name><surname>Schnabel</surname> <given-names>JA</given-names></name> <name><surname>Davatzikos</surname> <given-names>C</given-names></name> <name><surname>Alberola-L&#x00F3;pez</surname> <given-names>C</given-names></name> <name><surname>Fichtinger</surname> <given-names>G</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2018</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2018</year>). p. <fpage>265</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00937-3_31</pub-id></mixed-citation></ref>
<ref id="B122"><label>122.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mascagni</surname> <given-names>P</given-names></name> <name><surname>Alapatt</surname> <given-names>D</given-names></name> <name><surname>Sestini</surname> <given-names>L</given-names></name> <name><surname>Altieri</surname> <given-names>MS</given-names></name> <name><surname>Madani</surname> <given-names>A</given-names></name> <name><surname>Watanabe</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Computer vision in surgery: from potential to clinical value. Npj digit</article-title>. <source>Med</source>. (<year>2022b</year>) <volume>5</volume>:<fpage>163</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-022-00707-5</pub-id></mixed-citation></ref>
<ref id="B123"><label>123.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Colleoni</surname> <given-names>E</given-names></name> <name><surname>Moccia</surname> <given-names>S</given-names></name> <name><surname>Du</surname> <given-names>X</given-names></name> <name><surname>De Momi</surname> <given-names>E</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Deep learning based robotic tool detection and articulation estimation with spatio-temporal layers</article-title>. <source>IEEE Robot Autom Lett</source><italic>.</italic> (<year>2019</year>) <volume>4</volume>:<fpage>2714</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1109/LRA.2019.2917163</pub-id></mixed-citation></ref>
<ref id="B124"><label>124.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>X</given-names></name> <name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name></person-group>. <article-title>Semi-supervised learning with progressive unlabeled data excavation for label-efficient surgical workflow recognition</article-title>. <source>Med Image Anal</source><italic>.</italic> (<year>2021</year>) <volume>73</volume>:<fpage>102158</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.102158</pub-id><pub-id pub-id-type="pmid">34325149</pub-id></mixed-citation></ref>
<ref id="B125"><label>125.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Long</surname> <given-names>Y</given-names></name> <name><surname>Chen</surname> <given-names>C</given-names></name> <name><surname>Zhao</surname> <given-names>Z</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name></person-group>. <article-title>Temporal memory relation network for workflow recognition from surgical video</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2021</year>) <volume>40</volume>:<fpage>1911</fpage>&#x2013;<lpage>23</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2021.3069471</pub-id><pub-id pub-id-type="pmid">33780335</pub-id></mixed-citation></ref>
<ref id="B126"><label>126.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>X</given-names></name> <name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name></person-group>. <article-title>LRTD: long-range temporal dependency based active learning for surgical workflow recognition</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2020</year>) <volume>15</volume>:<fpage>1573</fpage>&#x2013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02198-9</pub-id><pub-id pub-id-type="pmid">32588246</pub-id></mixed-citation></ref>
<ref id="B127"><label>127.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>H</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Qin</surname> <given-names>J</given-names></name> <name><surname>Fu</surname> <given-names>C-W</given-names></name><etal/></person-group> <article-title>Multi-task recurrent convolutional network with correlation loss for surgical video analysis</article-title>. <source>Med Image Anal</source>. (<year>2020</year>) <volume>59</volume>:<fpage>101572</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2019.101572</pub-id><pub-id pub-id-type="pmid">31639622</pub-id></mixed-citation></ref>
<ref id="B128"><label>128.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>Q</given-names></name> <name><surname>Yang</surname> <given-names>R</given-names></name> <name><surname>Yang</surname> <given-names>S</given-names></name> <name><surname>Ni</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Jiang</surname> <given-names>Z</given-names></name><etal/></person-group> <article-title>Development and validation of a deep-learning based assistance system for enhancing laparoscopic control level</article-title>. <source>Rob Comput Surg</source>. (<year>2023</year>) <volume>19</volume>:<fpage>e2449</fpage>. <pub-id pub-id-type="doi">10.1002/rcs.2449</pub-id></mixed-citation></ref>
<ref id="B129"><label>129.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zou</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>W</given-names></name> <name><surname>Wang</surname> <given-names>J</given-names></name> <name><surname>Tao</surname> <given-names>R</given-names></name> <name><surname>Zheng</surname> <given-names>G</given-names></name></person-group>. <article-title>ARST: auto-regressive surgical transformer for phase recognition from laparoscopic videos</article-title>. <source>Comput Methods Biomech Biomed Eng Imaging Vis</source><italic>.</italic> (<year>2023</year>) <volume>11</volume>:<fpage>1012</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1080/21681163.2022.2145238</pub-id></mixed-citation></ref>
<ref id="B130"><label>130.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Cheng</surname> <given-names>K</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Heng</surname> <given-names>P-A</given-names></name><etal/></person-group> <article-title>Incorporating temporal prior from motion flow for instrument segmentation in minimally invasive surgery video</article-title>. In: <person-group person-group-type="editor"><name><surname>Shen</surname> <given-names>D</given-names></name> <name><surname>Liu</surname> <given-names>T</given-names></name> <name><surname>Peters</surname> <given-names>TM</given-names></name> <name><surname>Staib</surname> <given-names>LH</given-names></name> <name><surname>Essert</surname> <given-names>C</given-names></name> <name><surname>Zhou</surname> <given-names>S</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2019</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2019</year>). p. <fpage>440</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-32254-0_49</pub-id></mixed-citation></ref>
<ref id="B131"><label>131.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Kim</surname> <given-names>M</given-names></name> <name><surname>Jin</surname> <given-names>S</given-names></name></person-group>. <article-title><italic>Real-time detection and tracking of surgical instrument based on YOLOv5 and DeepSORT&#x002A;</italic></article-title>. In: <conf-name>2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN)</conf-name>; <conf-date>1758&#x2013;63</conf-date>; <publisher-name>IEEE</publisher-name>; <conf-loc>Busan, Korea, Republic of</conf-loc>; (<year>2023</year>). <pub-id pub-id-type="doi">10.1109/RO-MAN57019.2023.10309495</pub-id></mixed-citation></ref>
<ref id="B132"><label>132.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Jin</surname> <given-names>Y</given-names></name> <name><surname>Zhu</surname> <given-names>L</given-names></name></person-group>. <article-title>Dynamic interactive relation capturing via scene graph learning for robotic surgical report generation</article-title>. In: <conf-name>2023 IEEE International Conference on Robotics and Automation (ICRA)</conf-name>; <conf-date>2702&#x2013;9</conf-date>; <publisher-name>IEEE</publisher-name><article-title>;</article-title> <conf-loc>London, United Kingdom</conf-loc>; (<year>2023</year>). <pub-id pub-id-type="doi">10.1109/ICRA48891.2023.10160647</pub-id></mixed-citation></ref>
<ref id="B133"><label>133.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kitaguchi</surname> <given-names>D</given-names></name> <name><surname>Fujino</surname> <given-names>T</given-names></name> <name><surname>Takeshita</surname> <given-names>N</given-names></name> <name><surname>Hasegawa</surname> <given-names>H</given-names></name> <name><surname>Mori</surname> <given-names>K</given-names></name> <name><surname>Ito</surname> <given-names>M</given-names></name></person-group>. <article-title>Limited generalizability of single deep neural network for surgical instrument segmentation in different surgical environments</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>:<fpage>12575</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-16923-8</pub-id><pub-id pub-id-type="pmid">35869249</pub-id></mixed-citation></ref>
<ref id="B134"><label>134.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ata</surname> <given-names>YM</given-names></name> <name><surname>Al-Jassim</surname> <given-names>FA</given-names></name> <name><surname>Alabassi</surname> <given-names>K</given-names></name> <name><surname>Albakr</surname> <given-names>A</given-names></name> <name><surname>Ismail</surname> <given-names>T</given-names></name> <name><surname>Al Jalham</surname> <given-names>K</given-names></name></person-group>. <article-title>Pelvic lipomatosis&#x2014;a rare diagnosis and a challenging management: a case report and literature review</article-title>. <source>J Surg Case Rep</source>. (<year>2024</year>) <volume>2024</volume>:<fpage>rjae777</fpage>. <pub-id pub-id-type="doi">10.1093/jscr/rjae777</pub-id><pub-id pub-id-type="pmid">39678478</pub-id></mixed-citation></ref>
<ref id="B135"><label>135.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khalil</surname> <given-names>IA</given-names></name> <name><surname>Alani</surname> <given-names>A</given-names></name> <name><surname>Al Saeedi</surname> <given-names>A</given-names></name> <name><surname>Shehadeh</surname> <given-names>H</given-names></name> <name><surname>Ahmed</surname> <given-names>T</given-names></name> <name><surname>Al-Jalham</surname> <given-names>K</given-names></name></person-group>. <article-title>Life-threatening abdominal compartment syndrome as a complication of supine super mini percutaneous nephrolithotomy, the first case report and literature review</article-title>. <source>Urol Case Rep</source>. (<year>2021</year>) <volume>36</volume>:<fpage>101578</fpage>. <pub-id pub-id-type="doi">10.1016/j.eucr.2021.101578</pub-id><pub-id pub-id-type="pmid">33537210</pub-id></mixed-citation></ref>
<ref id="B136"><label>136.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>F</given-names></name> <name><surname>Sun</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>J</given-names></name></person-group>. <article-title>Surgical smoke removal via residual swin transformer network</article-title>. <source>Int J Comput Assist Radiol Surg</source><italic>.</italic> (<year>2023</year>) <volume>18</volume>:<fpage>1417</fpage>&#x2013;<lpage>27</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-023-02835-z</pub-id><pub-id pub-id-type="pmid">36683136</pub-id></mixed-citation></ref>
<ref id="B137"><label>137.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hayoz</surname> <given-names>M</given-names></name> <name><surname>Hahne</surname> <given-names>C</given-names></name> <name><surname>Gallardo</surname> <given-names>M</given-names></name> <name><surname>Candinas</surname> <given-names>D</given-names></name> <name><surname>Kurmann</surname> <given-names>T</given-names></name> <name><surname>Allan</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Learning how to robustly estimate camera pose in endoscopic videos</article-title>. <source>Int J CARS</source>. (<year>2023</year>) <volume>18</volume>:<fpage>1185</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-023-02919-w</pub-id></mixed-citation></ref>
<ref id="B138"><label>138.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Doughty</surname> <given-names>M</given-names></name> <name><surname>Ghugre</surname> <given-names>NR</given-names></name></person-group>. <article-title>HMD-EgoPose: head-mounted display-based egocentric marker-less tool and hand pose estimation for augmented surgical guidance</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2022</year>) <volume>17</volume>:<fpage>2253</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-022-02688-y</pub-id><pub-id pub-id-type="pmid">35701681</pub-id></mixed-citation></ref>
<ref id="B139"><label>139.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bamba</surname> <given-names>Y</given-names></name> <name><surname>Ogawa</surname> <given-names>S</given-names></name> <name><surname>Itabashi</surname> <given-names>M</given-names></name> <name><surname>Shindo</surname> <given-names>H</given-names></name> <name><surname>Kameoka</surname> <given-names>S</given-names></name> <name><surname>Okamoto</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Object and anatomical feature recognition in surgical video images based on a convolutional neural network</article-title>. <source>Int J CARS</source>. (<year>2021</year>) <volume>16</volume>:<fpage>2045</fpage>&#x2013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-021-02434-w</pub-id></mixed-citation></ref>
<ref id="B140"><label>140.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tanzi</surname> <given-names>L</given-names></name> <name><surname>Piazzolla</surname> <given-names>P</given-names></name> <name><surname>Porpiglia</surname> <given-names>F</given-names></name> <name><surname>Vezzetti</surname> <given-names>E</given-names></name></person-group>. <article-title>Real-time deep learning semantic segmentation during intra-operative surgery for 3D augmented reality assistance</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2021</year>) <volume>16</volume>:<fpage>1435</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-021-02432-y</pub-id><pub-id pub-id-type="pmid">34165672</pub-id></mixed-citation></ref>
<ref id="B141"><label>141.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Loukas</surname> <given-names>C</given-names></name> <name><surname>Frountzas</surname> <given-names>M</given-names></name> <name><surname>Schizas</surname> <given-names>D</given-names></name></person-group>. <article-title>Patch-based classification of gallbladder wall vascularity from laparoscopic images using deep learning</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2021</year>) <volume>16</volume>:<fpage>103</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02285-x</pub-id><pub-id pub-id-type="pmid">33146850</pub-id></mixed-citation></ref>
<ref id="B142"><label>142.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Gao</surname> <given-names>X</given-names></name></person-group>. <article-title>Object extraction via deep learning-based marker-free tracking framework of surgical instruments for laparoscope-holder robots</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2020</year>) <volume>15</volume>:<fpage>1335</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-020-02214-y</pub-id><pub-id pub-id-type="pmid">32577985</pub-id></mixed-citation></ref>
<ref id="B143"><label>143.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nakazawa</surname> <given-names>A</given-names></name> <name><surname>Harada</surname> <given-names>K</given-names></name> <name><surname>Mitsuishi</surname> <given-names>M</given-names></name> <name><surname>Jannin</surname> <given-names>P</given-names></name></person-group>. <article-title>Real-time surgical needle detection using region-based convolutional neural networks</article-title>. <source>Int J Comput Assist Radiol Surg</source>. (<year>2020</year>) <volume>15</volume>:<fpage>41</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-019-02050-9</pub-id><pub-id pub-id-type="pmid">31422553</pub-id></mixed-citation></ref>
<ref id="B144"><label>144.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bhattarai</surname> <given-names>A</given-names></name> <name><surname>Alsadoon</surname> <given-names>A</given-names></name> <name><surname>Prasad</surname> <given-names>PWC</given-names></name> <name><surname>Pham</surname> <given-names>L</given-names></name> <name><surname>Haddad</surname> <given-names>S</given-names></name> <name><surname>Hsu</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>A novel multiple communication paths for surgical telepresence videos delivery of the maxilla area in oral and maxillofacial surgery</article-title>. <source>Int J CARS</source>. (<year>2019</year>) <volume>14</volume>:<fpage>873</fpage>&#x2013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1007/s11548-018-01904-y</pub-id></mixed-citation></ref>
<ref id="B145"><label>145.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>G</given-names></name> <name><surname>Shankar</surname> <given-names>S</given-names></name> <name><surname>Pinisetty</surname> <given-names>S</given-names></name></person-group>. <article-title>Automated surgical procedure assistance framework using deep learning and formal runtime monitoring</article-title>. In: <person-group person-group-type="editor"><name><surname>Dang</surname> <given-names>T</given-names></name> <name><surname>Stolz</surname> <given-names>V</given-names></name></person-group>, editors. <source>Runtime Verification</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2022</year>). <fpage>25</fpage>&#x2013;<lpage>44</lpage>. <comment>vol. 13498</comment>.</mixed-citation></ref>
<ref id="B146"><label>146.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Valderrama</surname> <given-names>N</given-names></name> <name><surname>Ruiz Puentes</surname> <given-names>P</given-names></name> <name><surname>Hern&#x00E1;ndez</surname> <given-names>I</given-names></name> <name><surname>Ayobi</surname> <given-names>N</given-names></name> <name><surname>Verlyck</surname> <given-names>M</given-names></name> <name><surname>Santander</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Towards holistic surgical scene understanding</article-title>. In: <person-group person-group-type="editor"><name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Fletcher</surname> <given-names>PT</given-names></name> <name><surname>Speidel</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>S</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2022</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name> (<year>2022</year>). p. <fpage>442</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-16449-1_42</pub-id></mixed-citation></ref>
<ref id="B147"><label>147.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Schmidt</surname> <given-names>A</given-names></name> <name><surname>Mohareri</surname> <given-names>O</given-names></name> <name><surname>DiMaio</surname> <given-names>S</given-names></name> <name><surname>Salcudean</surname> <given-names>SE</given-names></name></person-group>. <article-title>Recurrent implicit neural graph for deformable tracking in endoscopic videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Fletcher</surname> <given-names>PT</given-names></name> <name><surname>Speidel</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>S</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention&#x2014;MICCAI 2022</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name> (<year>2022</year>). <fpage>478</fpage>&#x2013;<lpage>88</lpage>. <comment>vol. 13434</comment>.</mixed-citation></ref>
<ref id="B148"><label>148.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Fathollahi</surname> <given-names>M</given-names></name> <name><surname>Sarhan</surname> <given-names>MH</given-names></name> <name><surname>Pena</surname> <given-names>R</given-names></name> <name><surname>DiMonte</surname> <given-names>L</given-names></name> <name><surname>Gupta</surname> <given-names>A</given-names></name> <name><surname>Ataliwala</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Video-Based surgical skills assessment using long term tool tracking</article-title>. In: <person-group person-group-type="editor"><name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Dou</surname> <given-names>Q</given-names></name> <name><surname>Fletcher</surname> <given-names>PT</given-names></name> <name><surname>Speidel</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>S</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2022</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name> (<year>2022</year>). p. <fpage>541</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-031-16449-1_52</pub-id></mixed-citation></ref>
<ref id="B149"><label>149.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Kalia</surname> <given-names>M</given-names></name> <name><surname>Aleef</surname> <given-names>TA</given-names></name> <name><surname>Navab</surname> <given-names>N</given-names></name> <name><surname>Black</surname> <given-names>P</given-names></name> <name><surname>Salcudean</surname> <given-names>SE</given-names></name><etal/></person-group> <article-title>Co-generation and segmentation for generalized surgical instrument segmentation on unlabelled data</article-title>. In: <person-group person-group-type="editor"><name><surname>De Bruijne</surname> <given-names>M</given-names></name> <name><surname>Cattin</surname> <given-names>PC</given-names></name> <name><surname>Cotin</surname> <given-names>S</given-names></name> <name><surname>Padoy</surname> <given-names>N</given-names></name> <name><surname>Speidel</surname> <given-names>S</given-names></name> <name><surname>Zheng</surname> <given-names>Y</given-names></name></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention &#x2013; MICCAI 2021</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2021</year>). p. <fpage>403</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-87202-1_39</pub-id></mixed-citation></ref>
<ref id="B150"><label>150.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Zhao</surname> <given-names>Z</given-names></name> <name><surname>Shi</surname> <given-names>P</given-names></name> <name><surname>Hu</surname> <given-names>S</given-names></name></person-group>. <article-title>An efficient one-stage detector for real-time surgical tools detection in robot-assisted surgery</article-title>. In: <person-group person-group-type="editor"><name><surname>Papie&#x017C;</surname> <given-names>BW</given-names></name> <name><surname>Yaqub</surname> <given-names>M</given-names></name> <name><surname>Jiao</surname> <given-names>J</given-names></name> <name><surname>Namburete</surname> <given-names>AIL</given-names></name> <name><surname>Noble</surname> <given-names>JA</given-names></name></person-group>, editors. <source>Medical Image Understanding and Analysis</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2021</year>) <fpage>18</fpage>&#x2013;<lpage>29</lpage>. <comment>vol. 12722</comment>.</mixed-citation></ref>
<ref id="B151"><label>151.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Hachiuma</surname> <given-names>R</given-names></name> <name><surname>Shimizu</surname> <given-names>T</given-names></name> <name><surname>Saito</surname> <given-names>H</given-names></name> <name><surname>Kajita</surname> <given-names>H</given-names></name> <name><surname>Takatsume</surname> <given-names>Y</given-names></name></person-group>. <article-title>Deep selection: a fully supervised camera selection network for surgery recordings</article-title>. In: <person-group person-group-type="editor"><name><surname>Martel</surname><given-names>AL</given-names></name> <name><surname>Abolmaesumi</surname><given-names>P</given-names></name> <name><surname>Stoyanov</surname><given-names>D</given-names></name> <name><surname>Mateus</surname><given-names>D</given-names></name> <name><surname>Zuluaga</surname><given-names>MA</given-names></name> <name><surname>Zhou</surname><given-names>SK</given-names></name><etal/></person-group>, editors. <source>Medical Image Computing and Computer Assisted Intervention&#x2014;MICCAI 2020</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2020</year>). <fpage>419</fpage>&#x2013;<lpage>28</lpage>. <comment>vol. 12263</comment>.</mixed-citation></ref>
<ref id="B152"><label>152.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Sokolova</surname> <given-names>N</given-names></name> <name><surname>Schoeffmann</surname> <given-names>K</given-names></name> <name><surname>Taschwer</surname> <given-names>M</given-names></name> <name><surname>Putzgruber-Adamitsch</surname> <given-names>D</given-names></name> <name><surname>El-Shabrawi</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Evaluating the generalization performance of instrument classification in cataract surgery videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Ro</surname> <given-names>YM</given-names></name> <name><surname>Cheng</surname> <given-names>W-H</given-names></name> <name><surname>Kim</surname> <given-names>J</given-names></name> <name><surname>Chu</surname> <given-names>W-T</given-names></name> <name><surname>Cui</surname> <given-names>P</given-names></name> <name><surname>Choi</surname> <given-names>J-W</given-names></name></person-group>, editors. <source>MultiMedia Modeling</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2020</year>). p. <fpage>626</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-37734-2_51</pub-id></mixed-citation></ref>
<ref id="B153"><label>153.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Primus</surname> <given-names>MJ</given-names></name> <name><surname>Putzgruber-Adamitsch</surname> <given-names>D</given-names></name> <name><surname>Taschwer</surname> <given-names>M</given-names></name> <name><surname>M&#x00FC;nzer</surname> <given-names>B</given-names></name> <name><surname>El-Shabrawi</surname> <given-names>Y</given-names></name> <name><surname>B&#x00F6;sz&#x00F6;rmenyi</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Frame-Based classification of operation phases in cataract surgery videos</article-title>. In: <person-group person-group-type="editor"><name><surname>Schoeffmann</surname> <given-names>K</given-names></name> <name><surname>Chalidabhongse</surname> <given-names>TH</given-names></name> <name><surname>Ngo</surname> <given-names>CW</given-names></name> <name><surname>Aramvith</surname> <given-names>S</given-names></name> <name><surname>O&#x0027;Connor</surname> <given-names>NE</given-names></name> <name><surname>Ho</surname> <given-names>Y-S</given-names></name><etal/></person-group>, editors. <source>MultiMedia Modeling</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2018</year>). p. <fpage>241</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-73603-7_20</pub-id></mixed-citation></ref>
<ref id="B154"><label>154.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hasan Md</surname> <given-names>K</given-names></name> <name><surname>Calvet</surname> <given-names>L</given-names></name> <name><surname>Rabbani</surname> <given-names>N</given-names></name></person-group><person-group person-group-type="author"><name><surname>, Bartoli</surname> <given-names>A</given-names></name></person-group>, <article-title>Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry</article-title>. <source>Med Image Anal</source><italic>.</italic> (<year>2021</year>) <volume>70</volume>:<fpage>101994</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.101994</pub-id><pub-id pub-id-type="pmid">33611053</pub-id></mixed-citation></ref>
<ref id="B155"><label>155.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname> <given-names>X</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Shi</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>L</given-names></name> <name><surname>Wang</surname> <given-names>S</given-names></name></person-group>. <article-title>ST-ITEF: spatio-temporal intraoperative task estimating framework to recognize surgical phase and predict instrument path based on multi-object tracking in keratoplasty</article-title>. <source>Med Image Anal</source><italic>.</italic> (<year>2024</year>) <volume>91</volume>:<fpage>103026</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.103026</pub-id><pub-id pub-id-type="pmid">37976868</pub-id></mixed-citation></ref>
<ref id="B156"><label>156.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ping</surname> <given-names>L</given-names></name> <name><surname>Wang</surname> <given-names>Z</given-names></name> <name><surname>Yao</surname> <given-names>J</given-names></name> <name><surname>Gao</surname> <given-names>J</given-names></name> <name><surname>Yang</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Application and evaluation of surgical tool and tool tip recognition based on convolutional neural network in multiple endoscopic surgical scenarios</article-title>. <source>Surg Endosc</source>. (<year>2023</year>) <volume>37</volume>:<fpage>7376</fpage>&#x2013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-023-10323-3</pub-id><pub-id pub-id-type="pmid">37580576</pub-id></mixed-citation></ref>
<ref id="B157"><label>157.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>De Backer</surname> <given-names>P</given-names></name> <name><surname>Eckhoff</surname> <given-names>JA</given-names></name> <name><surname>Simoens</surname> <given-names>J</given-names></name> <name><surname>M&#x00FC;ller</surname> <given-names>DT</given-names></name> <name><surname>Allaeys</surname> <given-names>C</given-names></name> <name><surname>Creemers</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Multicentric exploration of tool annotation in robotic surgery: lessons learned when starting a surgical artificial intelligence project</article-title>. <source>Surg Endosc</source>. (<year>2022</year>) <volume>36</volume>:<fpage>8533</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-022-09487-1</pub-id><pub-id pub-id-type="pmid">35941310</pub-id></mixed-citation></ref>
<ref id="B158"><label>158.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gu&#x00E9;don</surname> <given-names>ACP</given-names></name> <name><surname>Meij</surname> <given-names>SEP</given-names></name> <name><surname>Osman</surname> <given-names>KNMMH</given-names></name> <name><surname>Kloosterman</surname> <given-names>HA</given-names></name> <name><surname>Van Stralen</surname> <given-names>KJ</given-names></name> <name><surname>Grimbergen</surname> <given-names>MCM</given-names></name><etal/></person-group> <article-title>Deep learning for surgical phase recognition using endoscopic videos</article-title>. <source>Surg Endosc</source>. (<year>2021</year>) <volume>35</volume>:<fpage>6150</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-020-08110-5</pub-id></mixed-citation></ref>
<ref id="B159"><label>159.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tokuyasu</surname> <given-names>T</given-names></name> <name><surname>Iwashita</surname> <given-names>Y</given-names></name> <name><surname>Matsunobu</surname> <given-names>Y</given-names></name> <name><surname>Kamiyama</surname> <given-names>T</given-names></name> <name><surname>Ishikake</surname> <given-names>M</given-names></name> <name><surname>Sakaguchi</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Development of an artificial intelligence system using deep learning to indicate anatomical landmarks during laparoscopic cholecystectomy</article-title>. <source>Surg Endosc</source>. (<year>2021</year>) <volume>35</volume>:<fpage>1651</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1007/s00464-020-07548-x</pub-id><pub-id pub-id-type="pmid">32306111</pub-id></mixed-citation></ref>
<ref id="B160"><label>160.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bian</surname> <given-names>G-B</given-names></name> <name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Li</surname> <given-names>Z</given-names></name> <name><surname>Fu</surname> <given-names>P</given-names></name> <name><surname>Yue</surname> <given-names>W-Q</given-names></name><etal/></person-group> <article-title>Motion decoupling network for intra-operative motion estimation under occlusion</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2023</year>) <volume>42</volume>:<fpage>2924</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2023.3268774</pub-id><pub-id pub-id-type="pmid">37079409</pub-id></mixed-citation></ref>
<ref id="B161"><label>161.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yue</surname> <given-names>W</given-names></name> <name><surname>Liao</surname> <given-names>H</given-names></name> <name><surname>Xia</surname> <given-names>Y</given-names></name> <name><surname>Lam</surname> <given-names>V</given-names></name> <name><surname>Luo</surname> <given-names>J</given-names></name> <name><surname>Wang</surname> <given-names>Z</given-names></name></person-group>. <article-title>Cascade multi-level transformer network for surgical workflow analysis</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2023</year>) <volume>42</volume>:<fpage>2817</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2023.3265354</pub-id><pub-id pub-id-type="pmid">37037257</pub-id></mixed-citation></ref>
<ref id="B162"><label>162.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Padoy</surname><given-names>N</given-names></name></person-group>. <article-title>Machine and deep learning for workflow recognition during surgery</article-title>. <source>Minim Invasive Ther Allied Technol.</source> (<year>2019</year>) <volume>28</volume>:<fpage>82</fpage>&#x2013;<lpage>90</lpage>. <pub-id pub-id-type="doi">10.1080/13645706.2019.1584116</pub-id><pub-id pub-id-type="pmid">30849261</pub-id></mixed-citation></ref>
<ref id="B163"><label>163.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Colleoni</surname> <given-names>E</given-names></name> <name><surname>Moccia</surname> <given-names>S</given-names></name> <name><surname>Du</surname> <given-names>X</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <source>Deep Learning Based Robotic Tool Detection and Articulation Estimation with Spatio-Temporal Layers</source>. <publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE Journals &#x0026; Magazine&#x007C;IEEE Xplore</publisher-name>. (<year>2019</year>). <pub-id pub-id-type="doi">10.1109/LRA.2019.2917163</pub-id></mixed-citation></ref>
<ref id="B164"><label>164.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kitaguchi</surname><given-names>D</given-names></name> <name><surname>Takeshita</surname><given-names>N</given-names></name> <name><surname>Matsuzaki</surname><given-names>H</given-names></name> <name><surname>Oda</surname><given-names>T</given-names></name> <name><surname>Watanabe</surname><given-names>M</given-names></name> <name><surname>Mori</surname><given-names>K</given-names></name><etal/></person-group> <article-title>Automated laparoscopic colorectal surgery workflow recognition using artificial intelligence: experimental research</article-title>. <source>Int J Surg.</source> (<year>2020</year>) <volume>79</volume>:<fpage>88</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijsu.2020.05.015</pub-id><pub-id pub-id-type="pmid">32413503</pub-id></mixed-citation></ref>
<ref id="B165"><label>165.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname><given-names>F</given-names></name> <name><surname>Silva Croso</surname><given-names>G</given-names></name> <name><surname>Kim</surname><given-names>TS</given-names></name> <name><surname>Song</surname><given-names>Z</given-names></name> <name><surname>Parker</surname><given-names>F</given-names></name> <name><surname>Hager</surname><given-names>GD</given-names></name><etal/></person-group> Assessment of automated identification of phases in videos of cataract surgery using machine learning and deep learning techniques. JAMA Netw Open. (2019) 2:e191860. <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2019.1860</pub-id></mixed-citation></ref>
<ref id="B166"><label>166.</label><mixed-citation publication-type="book"><article-title>Novel applications of deep learning in surgical training</article-title>. In: <person-group person-group-type="editor"><name><surname>De Pablos</surname><given-names>PO</given-names></name> <name><surname>Zhang</surname><given-names>X</given-names></name></person-group>, editors. <source>Artificial Intelligence, Big Data, Blockchain and 5G for the Digital Transformation of the Healthcare Industry</source>. <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Academic Press</publisher-name> (<year>2024</year>). <fpage>301</fpage>&#x2013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1016/B978-0-443-21598-8.00021-X</pub-id></mixed-citation></ref>
<ref id="B167"><label>167.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Wu</surname> <given-names>Z</given-names></name> <name><surname>Morgan</surname> <given-names>TN</given-names></name> <name><surname>Garbens</surname> <given-names>A</given-names></name></person-group>. <article-title>Evaluating robotic-assisted partial nephrectomy surgeons with fully convolutional segmentation and multi-task attention networks</article-title>. <source>J Robot Surg</source>. (<year>2023</year>) <volume>17</volume>:<fpage>2323</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1007/s11701-023-01657-0</pub-id><pub-id pub-id-type="pmid">37368225</pub-id></mixed-citation></ref>
<ref id="B168"><label>168.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ahmed</surname><given-names>FA</given-names></name> <name><surname>Arsalan</surname><given-names>M</given-names></name> <name><surname>Al-Ali</surname><given-names>A</given-names></name> <name><surname>Al-Jalham</surname><given-names>K</given-names></name> <name><surname>Balakrishnan</surname><given-names>S</given-names></name></person-group>. <article-title>CLIP-RL: surgical scene segmentation using contrastive language-vision pretraining &#x0026; reinforcement learning</article-title>. <source>arXiv</source> [Preprint]. <source>arXiv:2507.04317</source> (<year>2025</year>). <pub-id pub-id-type="doi">10.48550/ARXIV.2507.04317</pub-id></mixed-citation></ref>
<ref id="B169"><label>169.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Abdel-Ghani</surname> <given-names>M</given-names></name> <name><surname>Ali</surname> <given-names>M</given-names></name> <name><surname>Ali</surname> <given-names>M</given-names></name> <name><surname>Ahmed</surname> <given-names>F</given-names></name> <name><surname>Arsalan</surname> <given-names>M</given-names></name> <name><surname>Al-Ali</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>FASL-Seg: anatomy and tool segmentation of surgical scenes</article-title>. In: <person-group person-group-type="editor"><name><surname>Lynce</surname> <given-names>I</given-names></name> <name><surname>Murano</surname> <given-names>N</given-names></name> <name><surname>Vallati</surname> <given-names>M</given-names></name> <name><surname>Villata</surname> <given-names>S</given-names></name> <name><surname>Chesani</surname> <given-names>F</given-names></name> <name><surname>Milano</surname> <given-names>M</given-names></name><etal/></person-group> editors. <source>Frontiers in Artificial Intelligence and Applications</source>. <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>IOS Press</publisher-name> (<year>2025</year>). <pub-id pub-id-type="doi">10.3233/FAIA250908</pub-id></mixed-citation></ref>
<ref id="B170"><label>170.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ahmed</surname> <given-names>F</given-names></name> <name><surname>Abdel-Ghani</surname> <given-names>M</given-names></name> <name><surname>Arsalan</surname> <given-names>M</given-names></name> <name><surname>Ali</surname> <given-names>M</given-names></name> <name><surname>Al-Ali</surname> <given-names>A</given-names></name> <name><surname>Balakrishnan</surname> <given-names>S</given-names></name></person-group>. <article-title>Surg-SegFormer: a dual transformer-based model for holistic surgical scene segmentation</article-title>. In: <source>2025 IEEE 21st International Conference on Automation Science and Engineering (CASE)</source>. <publisher-loc>Los Angeles, CA, USA</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2025</year>). p. <fpage>1304</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/CASE58245.2025.11163962</pub-id></mixed-citation></ref>
<ref id="B171"><label>171.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Kharbech</surname><given-names>S</given-names></name> <name><surname>Abinahed</surname><given-names>J</given-names></name> <name><surname>Aboumarzouk</surname><given-names>O</given-names></name> <name><surname>Ansari</surname><given-names>WE</given-names></name> <name><surname>Al Ansari</surname><given-names>A</given-names></name> <name><surname>Balakrishnan</surname><given-names>S</given-names></name></person-group>. <article-title>Digital tools and innovative healthcare solutions: serious games and gamification in surgical training and patient care</article-title>. In: <person-group person-group-type="author"><name><surname>De Pablos</surname><given-names>PO</given-names></name> <name><surname>Zhang</surname><given-names>X</given-names></name></person-group>, editors. <source>Artificial Intelligence, Big Data, Blockchain and 5G for the Digital Transformation of the Healthcare Industry</source>. <publisher-loc>Amsterdam</publisher-loc>: <publisher-name>Academic Press</publisher-name> (<year>2024</year>). <fpage>321</fpage>&#x2013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1016/B978-0-443-21598-8.00007-5</pub-id></mixed-citation></ref>
<ref id="B172"><label>172.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shabir</surname> <given-names>D</given-names></name> <name><surname>Anbatawi</surname> <given-names>M</given-names></name> <name><surname>Padhan</surname> <given-names>J</given-names></name> <name><surname>Balakrishnan</surname> <given-names>S</given-names></name> <name><surname>Al-Ansari</surname> <given-names>A</given-names></name> <name><surname>Abinahed</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Evaluation of user-interfaces for controlling movements of virtual minimally invasive surgical instruments</article-title>. <source>Robot Comput Surg</source>. (<year>2022</year>) <volume>18</volume>:<fpage>e2414</fpage>. <pub-id pub-id-type="doi">10.1002/rcs.2414</pub-id></mixed-citation></ref></ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1118780/overview">Sophia Bano</ext-link>, University College London, United Kingdom</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2819164/overview">Jacopo D&#x0027;Andria Ursoleo</ext-link>, San Raffaele Scientific Institute (IRCCS), Italy</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3295201/overview">Mansoor Ali</ext-link>, Monterrey Institute of Technology and Higher Education (ITESM), Mexico</p></fn>
</fn-group>
</back>
</article>