<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="editorial">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2025.1740331</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Editorial</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Editorial: Advances and challenges in AI-driven visual intelligence: bridging theory and practice</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Huang</surname> <given-names>Bo</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2229689"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhang</surname> <given-names>Dawei</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2411012"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Qiao</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2850468"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>College of Optoelectronic Engineering, Chongqing University (CQU)</institution>, <city>Chongqing</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Computer Science and Technology, Zhejiang Normal University</institution>, <city>Jinhua</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>National Center for Applied Mathematics, Chongqing Normal University</institution>, <city>Chongqing</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Bo Huang, <email xlink:href="mailto:huangbo0326@cqu.edu.cn">huangbo0326@cqu.edu.cn</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-11-27">
<day>27</day>
<month>11</month>
<year>2025</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>8</volume>
<elocation-id>1740331</elocation-id>
<history>
<date date-type="received">
<day>05</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>17</day>
<month>11</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2025 Huang, Zhang and Liu.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Huang, Zhang and Liu</copyright-holder>
<license>
<ali:license_ref start_date="2025-11-27">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<kwd-group>
<kwd>AI-driven visual intelligence</kwd>
<kwd>adversarial attacks</kwd>
<kwd>semi-supervised learning</kwd>
<kwd>3D reconstruction</kwd>
<kwd>non-invasive diagnostic</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the National Natural Science Foundation of China (Nos. 62306049, 92471207 and W2421089), the General Program of Chongqing Natural Science Foundation (No. CSTB2023NSCQ-MSX0665), and the Fundamental Research Funds for the Central Universities (No. 2024CDJXY008).</funding-statement>
</funding-group>
<counts>
<fig-count count="1"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="0"/>
<page-count count="3"/>
<word-count count="1197"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Machine Learning and Artificial Intelligence</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
<notes notes-type="frontiers-research-topic">
<p><bold>Editorial on the Research Topic</bold> <ext-link xlink:href="https://www.frontiersin.org/research-topics/67594/advances-and-challenges-in-ai-driven-visual-intelligence-bridging-theory-and-practice" ext-link-type="uri">Advances and challenges in AI-driven visual intelligence: bridging theory and practice</ext-link></p></notes>
</front>
<body>
<p>Visual intelligence has become a fundamental research area in artificial intelligence, driven by rapid advances in deep learning and increasing demands from practical applications. This Research Topic brings together cutting-edge research that addresses both the remarkable progress and the persistent challenges in implementing AI-driven visual intelligence systems in real-world applications. The overview of this Research Topic is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The six articles published in this collection exemplify the breadth and depth of current research, spanning from fundamental security concerns to practical applications across diverse fields.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Overview of this Research Topic.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1740331-g0001.tif">
<alt-text content-type="machine-generated">Flowchart of AI-Driven Visual Intelligence research, detailing six articles under four main topics: Security &#x00026; Robustness, Semi-Supervised Learning, 3D Reconstruction, and Non-Invasive Diagnostic. Key challenges include security, data efficiency, domain integration, and real-world deployment.</alt-text>
</graphic>
</fig>
<sec id="s1">
<title>Robustness against adversarial attacks</title>
<p>As AI systems increasingly operate in dynamic, open-world scenarios, their vulnerability to adversarial manipulation presents significant concerns. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frai.2025.1621025">Pi et al.</ext-link> investigate test-time poisoning attacks targeting open-world test-time training (OWTTT) models. Their findings reveal that these models can be compromised with merely 100 queries using single-step query-based attack strategies. More critically, the affected models show limited recovery capability even when subsequently exposed to normal samples. This work serves as a crucial reminder that as we pursue adaptive AI systems, we must simultaneously address their security vulnerabilities to ensure safe deployment in critical applications.</p></sec>
<sec id="s2">
<title>Semi-supervised learning frameworks</title>
<p>The scarcity of labeled training data remains a persistent bottleneck in visual intelligence applications. Two articles tackle this limitation through innovative semi-supervised approaches. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frai.2025.1638772">Yu et al.</ext-link> propose an enhanced YOLOv8 framework that embeds the CBAM attention mechanism in high-level networks while incorporating Mean Teacher semi-supervised learning to address data labeling challenges. Their method achieves robust detection of micron-scale defects in industrial polymer films. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frai.2025.1645877">Guo et al.</ext-link> introduce a weight-aware semi-supervised self-ensembling framework (WSSL) for interior decoration style classification. By employing an adaptive weighting module based on truncated Gaussian functions, their approach selectively leverages reliable unlabeled data while mitigating confirmation bias from unreliable pseudo-labels. Both works demonstrate that semi-supervised learning can substantially reduce annotation costs without compromising performance.</p></sec>
<sec id="s3">
<title>3D reconstruction and cultural heritage</title>
<p>The preservation of cultural landscapes presents unique challenges that require accurate 3D reconstruction capabilities. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fenvs.2025.1556042">Chen et al.</ext-link> develop DGA-Net, which combines deep feature extraction with graph structure representation for reconstructing historical garden landscapes. The architecture incorporates attention mechanisms to emphasize ecologically significant features, enabling more accurate restoration planning and cultural heritage protection. This work exemplifies how AI-driven visual intelligence can contribute to environmental conservation and the preservation of cultural heritage sites.</p></sec>
<sec id="s4">
<title>Non-invasive diagnostic applications</title>
<p>Two articles demonstrate the potential of visual intelligence for accessible healthcare and agricultural monitoring. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fdata.2025.1557600">Navarro-Cabrera et al.</ext-link> apply DenseNet169 for detecting iron deficiency anemia in university students through fingernail image analysis. Their approach achieves 71.08% accuracy with 74.09% AUC, offering a cost-effective alternative to conventional blood tests, particularly relevant for developing regions. <ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frai.2025.1648867">Aron&#x000E9;s et al.</ext-link> present App2, a mobile application combining CNN and SVM architectures for apple leaf disease detection. The system achieves 95% accuracy on clear images and maintains 80% performance under real-field conditions after model adaptation. Notably, their inclusion of validation filters to verify leaf presence reduces false detections, demonstrating attention to practical deployment considerations. Both works highlight how AI-based visual intelligence can address accessibility gaps in healthcare and agriculture.</p></sec>
<sec id="s5">
<title>Open challenges and future directions</title>
<p>The collected works reveal several persistent challenges in visual intelligence research. First, the security of adaptive systems requires greater attention, particularly as models operate in adversarial environments. The difficulty in recovering from poisoning attacks suggests that defensive mechanisms must be incorporated at the architectural level. Second, efficient learning from limited labeled data continues to demand innovation, with semi-supervised and self-supervised methods showing promise but requiring careful handling of pseudo-label reliability. Third, the integration of domain knowledge with deep learning architectures remains an underexplored avenue that could enhance both performance and interpretability. Finally, the gap between laboratory performance and real-world deployment&#x02014;exemplified by the accuracy drop in field conditions&#x02014;indicates that robustness to environmental variations needs greater emphasis during model development.</p>
<p>The progress documented in this Research Topic demonstrates that effective visual intelligence systems require more than algorithmic sophistication. Success depends equally on addressing practical constraints including computational efficiency, data availability, security vulnerabilities, and deployment accessibility. As the field advances, bridging the gap between theoretical capabilities and practical requirements will necessitate continued collaboration across disciplines and careful consideration of real-world operating conditions.</p></sec>
</body>
<back>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>BH: Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. DZ: Writing &#x02013; review &#x00026; editing. QL: Writing &#x02013; review &#x00026; editing.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s8">
<title>Generative AI statement</title>
<p>The author(s) declare that no Gen AI was used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited and reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/152842/overview">Rashid Ibrahim Mehmood</ext-link>, Islamic University of Madinah, Saudi Arabia</p>
</fn>
</fn-group>
</back>
</article>