<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Microbiol.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Microbiology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Microbiol.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1664-302X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmicb.2026.1770528</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>EMDS-7-FSCIL: a benchmark for Few-Shot Class-Incremental Learning in environmental microorganism recognition</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Zhou</surname> <given-names>Jinyi</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<uri xlink:href="https://loop.frontiersin.org/people/3369067"/>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Zhang</surname> <given-names>Yinuo</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Xu</surname> <given-names>Sihang</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<uri xlink:href="https://loop.frontiersin.org/people/3227117"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Hu</surname> <given-names>Yangfan</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<uri xlink:href="https://loop.frontiersin.org/people/3321877"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Zeng</surname> <given-names>Jinhui</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Yin</surname> <given-names>Yimin</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<uri xlink:href="https://loop.frontiersin.org/people/3321318"/>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>School of Intelligent Manufacturing, Hunan First Normal University</institution>, <city>Changsha</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Rocket Force University of Engineering</institution>, <city>Xi&#x00027;an</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>School of Computer Science and Mathematics, Fujian University of Technology</institution>, <city>Fuzhou</city>, <country country="cn">China</country></aff>
<aff id="aff4"><label>4</label><institution>College of Meteorology and Oceanography, National University of Defense Technology</institution>, <city>Changsha</city>, <country country="cn">China</country></aff>
<aff id="aff5"><label>5</label><institution>School of Mathematics and Statistics, Hunan First Normal University</institution>, <city>Changsha</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Yimin Yin, <email xlink:href="mailto:yinyimin1@126.com">yinyimin1@126.com</email></corresp>
<fn fn-type="equal" id="fn001"><label>&#x02020;</label><p>These authors have contributed equally to this work</p></fn></author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-10">
<day>10</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>17</volume>
<elocation-id>1770528</elocation-id>
<history>
<date date-type="received">
<day>18</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>14</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>19</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2026 Zhou, Zhang, Xu, Hu, Zeng and Yin.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Zhou, Zhang, Xu, Hu, Zeng and Yin</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-10">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Deep learning&#x02013;based environmental microorganism recognition in a dynamic world demands models capable of recognizing novel classes in new tasks. However, it is hindered by data scarcity, high annotation costs, and the plasticity&#x02013;stability dilemma. Few-Shot Class-Incremental Learning (FSCIL) aims to address these challenges, yet a dedicated benchmark for environmental microorganism recognition remains absent. To bridge this gap, we establish the first FSCIL benchmark for environmental microorganism recognition and propose a unified evaluation protocol on the EMDS-7 dataset. We systematically reproduce 10 representative FSCIL methods: CEC, FACT, SAVC, PFR, ADBS, Comp, TEEN, Limit, BiDist, and CLOSER, and conduct comprehensive comparative experiments under a consistent implementation setting. We report multidimensional evaluation metrics, including per-session accuracy, average accuracy across sessions, and performance drop rate to quantify long-term performance degradation, along with thorough performance analyses. Our results reveal that SAVC and FACT achieve the highest overall accuracy, while PFR demonstrates more stable performance at the cost of a lower accuracy ceiling. In contrast, CLOSER and BiDist exhibit substantially weaker performance. Overall, our benchmark reveals that FSCIL methods effective on generic image benchmarks do not directly transfer to environmental microorganism recognition tasks, necessitating task-specific adaptations. This work provides a reproducible foundational platform to enable fair comparisons and accelerate future research on FSCIL for environmental microorganism recognition.</p></abstract>
<kwd-group>
<kwd>class-incremental learning</kwd>
<kwd>deep learning</kwd>
<kwd>EMDS-7 dataset</kwd>
<kwd>environmental microorganism</kwd>
<kwd>Few-Shot Learning</kwd>
</kwd-group>
<funding-group>
  <funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the Excellent Youth Project of Hunan Provincial Education Department (No. 24B0865).</funding-statement>
</funding-group>
<counts>
<fig-count count="6"/>
<table-count count="2"/>
<equation-count count="0"/>
<ref-count count="27"/>
<page-count count="11"/>
<word-count count="6461"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Systems Microbiology</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Environmental microorganisms are vital components of ecosystems, contributing to nutrient cycling, water purification, and overall ecological stability (<xref ref-type="bibr" rid="B19">Yin et al., 2024</xref>). At the same time, they serve as sensitive bioindicators for assessing water quality, detecting pollution sources, and monitoring ecological balance (<xref ref-type="bibr" rid="B12">Pluym et al., 2024</xref>). Accurate recognition of these microorganisms is therefore essential for understanding environmental dynamics and guiding sustainable management practices. However, traditional recognition methods&#x02013;such as manual microscopic examination, biochemical assays, and molecular sequencing&#x02013;are time-consuming, labor-intensive, and costly (<xref ref-type="bibr" rid="B1">Arbefeville et al., 2024</xref>), and they often require substantial expert involvement. Recent advances in computer vision and deep learning have enabled Automated Microorganism Detection and Recognition (AMDR), allowing models to analyze morphological and textural features in microscopic images. Representative work also includes task-specific architectures (e.g., LCU-Net; <xref ref-type="bibr" rid="B21">Zhang J. et al., 2021</xref>) designed for environmental microorganism microscopy segmentation. These approaches significantly improve recognition efficiency and objectivity. In real-world environmental microbial image classification and monitoring, deep learning models are typically static and can only recognize categories they were trained on. However, monitoring tasks are diverse, and environments change continuously, so new categories that are not included in the training set will inevitably appear in the field, and these new categories often have very few samples and are difficult to annotate. For example, in long-term online monitoring of environmental microbes (illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref>), as time evolves, the model needs to learn categories it has not encountered before. Ideally, a model should be able to rapidly adapt to new categories under sparse-sample conditions while not forgetting previously learned ones. Few-Shot Class-Incremental Learning (FSCIL) is well-suited to this requirement.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Illustration of the FSCIL-based environmental microorganism recognition process. As time progresses, microbial communities in real-world monitoring scenarios continuously evolve and new microorganism classes emerge. Given the high cost of data collection, and annotation, the model should continuously learn novel classes from only a few labeled examples while maintaining performance on previously learned classes.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0001.tif">
<alt-text content-type="machine-generated">Diagram depicting an incremental learning process for a model. It shows a sequence of three stages. Initially, the model is trained on base classes with sufficient labeled samples, enhancing recognition for base classes. In subsequent stages, new classes with limited labeled samples are introduced twice. After each introduction, the model undergoes incremental learning. It recognizes base plus three classes and then base plus six classes. The diagram emphasizes the model's evolving recognition ability over time, depicted by a timeline at the bottom.</alt-text>
</graphic>
</fig>
<p>These challenges align closely with the goals of FSCIL, which aims to extend a trained model to additional, previously unseen categories using only a few labeled examples, while retaining knowledge of previously learned classes (<xref ref-type="bibr" rid="B23">Zhang et al., 2025</xref>). The central technical difficulty, from an FSCIL perspective, is balancing stability and plasticity: mitigating catastrophic forgetting of old classes while remaining adaptable to new data (<xref ref-type="bibr" rid="B6">De Lange et al., 2021</xref>). This challenge is particularly relevant to environmental microorganism monitoring. Deep models for these tasks are typically trained in a static, closed-set manner and, as a result, can only reliably operate on the categories observed during training. In practice, monitoring tasks are diverse and may involve categories that are absent from the original training set, for which labeled samples are often scarce. Therefore, an ideal system would incorporate such categories under sparse supervision without degrading performance on previously learned ones. Despite this need, progress has been hindered by the lack of standardized benchmarks that reflect real-world ecological imaging complexity and support incremental, realistic evaluation protocols (<xref ref-type="bibr" rid="B11">Pana&#x000EF;otis et al., 2025</xref>).</p>
<p>To address this limitation, we adopt the Environmental Microorganism Dataset Seventh Version (EMDS-7) (<xref ref-type="bibr" rid="B18">Yang et al., 2023</xref>) as the foundation for constructing the first benchmark dedicated to FSCIL in environmental microorganism recognition. EMDS-7 is a publicly available microscopic dataset designed for object detection and ecological analysis of microorganisms. It contains 2,365 images across 41 microorganism categories and 13,216 annotated objects collected from natural lakes and rivers in Shenyang, China. The dataset exhibits notable ecological authenticity, capturing realistic conditions such as overlapping microorganisms, uneven staining, illumination variations, and background impurities. Compared with earlier EMDS versions, EMDS-7 expands taxonomic diversity and provides richer annotations, making it well-suited for studying both class-incremental and few-shot learning scenarios.</p>
<p>Building on EMDS-7, this work constructs an FSCIL evaluation protocol that partitions the dataset into base and incremental sessions to emulate real-world monitoring conditions in which new classes emerge under severe annotation scarcity. Under this framework, we evaluate representative FSCIL baselines, including rehearsal-based iCaRL (<xref ref-type="bibr" rid="B13">Rebuffi et al., 2017</xref>) and graph-adaptive classifier evolution CEC (<xref ref-type="bibr" rid="B20">Zhang C. et al., 2021</xref>). We further include forward-compatible training via FACT (<xref ref-type="bibr" rid="B25">Zhou D. et al., 2022</xref>) and contrastive regularization via SAVC (<xref ref-type="bibr" rid="B14">Song et al., 2023</xref>), and compare against a recent decision-boundary adaptation strategy via ADBS (<xref ref-type="bibr" rid="B9">Li et al., 2025</xref>). The benchmark reveals several intrinsic challenges of EMDS-7, such as significant intra-class variation and small inter-class differences, which complicate the stability-plasticity trade-off during continual adaptation.</p>
<p>In summary, this study introduces the first benchmark for FSCIL in environmental microorganism recognition, leveraging EMDS-7 as a realistic and challenging evaluation platform. By bridging environmental image analysis and adaptive deep learning, the benchmark provides a standardized foundation for testing FSCIL algorithms under ecologically plausible, data-scarce conditions. <xref ref-type="fig" rid="F2">Figure 2</xref> summarizes the end-to-end workflow of our benchmark construction and evaluation.</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>Overall workflow for constructing the EMDS-7 FSCIL benchmark. This figure presents the experimental setup and benchmark results of few-shot class-incremental learning on the EMDS-7 dataset. The base session contains 20 classes, followed by incremental sessions under a 3-way 3-shot setting; the incremental classes contain 21 classes. Multiple representative methods are compared in terms of average accuracy and performance degradation, providing a comprehensive evaluation of their classification capability and stability in environmental microorganism FSCIL scenarios.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0002.tif">
<alt-text content-type="machine-generated">Flowchart showing four sections: &#x0201C;Problem&#x0201D; states FSCIL lacks a benchmark for environmental microorganism image recognition. &#x0201C;Dataset&#x0201D; includes the EMDS-7 Dataset with images of microorganisms like Oscillatoria and Chaenea. &#x0201C;Experiment&#x0201D; lists methods (CEC, FACT, etc.) and training settings with 20 classes in base sessions and 3-way 3-shot in incremental sessions. &#x0201C;Benchmark&#x0201D; shows method performance with AA and PD scores.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2">
<label>2</label>
<title>Related work</title>
<sec>
<label>2.1</label>
<title>Few-Shot Learning</title>
<p>Few-Shot Learning (FSL) aims to enable models to recognize novel categories using only a few labeled samples per class (<xref ref-type="bibr" rid="B16">Tsoumplekas et al., 2024</xref>). This paradigm is particularly important in scientific domains such as environmental monitoring and biological taxonomy, where data collection is expensive and sample acquisition is constrained (<xref ref-type="bibr" rid="B3">Bennequin et al., 2022</xref>; <xref ref-type="bibr" rid="B10">Lim et al., 2025</xref>).</p>
<p>Existing FSL methods mainly follow three approaches. First, data augmentation-based methods expand limited datasets by generating or transforming samples (e.g., via semantic priors or adversarial generation). Second, meta-learning-based methods (e.g., MAML-style learning-to-learn) optimize for rapid task adaptation across episodes (<xref ref-type="bibr" rid="B7">Hospedales et al., 2021</xref>). Finally, graph-based methods explicitly model relationships between support and query samples for relational reasoning (<xref ref-type="bibr" rid="B27">Zhou X. et al., 2022</xref>).</p>
<p>Although these methods have achieved success on standard benchmarks such as Omniglot and miniImageNet, they are mostly evaluated on static, clean, and well-separated tasks, which can deviate from realistic fine-grained and evolving deployments (<xref ref-type="bibr" rid="B3">Bennequin et al., 2022</xref>). This differs substantially from real-world scenarios (e.g., environmental microbiology) where new morphotypes continually emerge, and the learner must adapt under extreme sample scarcity over time; consequently, conventional FSL alone is often insufficient for continuously evolving environments (<xref ref-type="bibr" rid="B10">Lim et al., 2025</xref>).</p>
</sec>
<sec>
<label>2.2</label>
<title>Class-Incremental Learning</title>
<p>Class-Incremental Learning (CIL) enables machine learning models, most commonly neural network, to recognize an expanding set of classes introduced progressively over time. Unlike FSL, CIL typically assumes that each newly introduced class is supported by a relatively sufficient number of training samples, while requiring the model to preserve discriminative performance on previously learned classes as new ones are acquired. The central challenge of CIL is mitigating catastrophic forgetting, a phenomenon in which the model&#x00027;s recognition performance on earlier classes degrades significantly as it learns new ones.</p>
<p>Existing CIL approaches can be broadly categorized into several methodological paradigms. Exemplar-based replay methods (e.g., iCaRL; <xref ref-type="bibr" rid="B13">Rebuffi et al., 2017</xref> ) retain a small set of historical samples and leverage knowledge distillation to maintain decision boundaries for previously learned classes. Constraint or regularization-based methods (e.g., LUCIR; <xref ref-type="bibr" rid="B8">Hou et al., 2019</xref> ) restrict updates to network parameters or feature representations in order to enforce representational consistency across incremental learning stages. Representation-level distillation methods perform feature-space distillation to ensure the stability of the embedding space throughout the incremental learning process. Although these methods have demonstrated strong performance on large-scale natural image benchmarks, they generally rely on the availability of sufficient, clean, and well-curated labeled data at each incremental stage and assume well-curated visual domains. Such assumptions limit their direct applicability to scientific domains such as environmental microbiome analysis, where samples from novel classes are often extremely scarce and are further characterized by substantial noise and high intra-class variability.</p>
</sec>
<sec>
<label>2.3</label>
<title>Few-Shot Class-Incremental Learning</title>
<p>FSCIL combines the strengths of FSL and CIL, enabling models to continually learn new classes using only a small number of labeled samples while retaining previously acquired knowledge (<xref ref-type="bibr" rid="B23">Zhang et al., 2025</xref>). This task was formally defined by, who established a benchmark protocol and highlighted the compounded challenges introduced by simultaneous data scarcity and class-incremental expansion.</p>
<p>Representative FSCIL methods address this challenge from multiple perspectives: TOPIC (<xref ref-type="bibr" rid="B15">Tao et al., 2020</xref>) introduces a topology-preserving mechanism (via neural gas) to stabilize the feature-space structure across incremental stages; CEC proposes continuously evolving classifiers to mitigate forgetting and maintain coherent decision boundaries (<xref ref-type="bibr" rid="B20">Zhang C. et al., 2021</xref>) Meta-learning-based frameworks (e.g., MetaFSCIL; <xref ref-type="bibr" rid="B4">Chi et al., 2022</xref>) and forward-compatible training (e.g., FACT; <xref ref-type="bibr" rid="B25">Zhou D. et al., 2022</xref>) enhance adaptability by improving cross-session generalization and by calibrating representations or classifiers for future updates.</p>
<p>These methods have achieved significant progress on widely used benchmark datasets, which are CUB-200-2011, miniImageNet, and CIFAR-100. However, most FSCIL baselines are still primarily benchmarked on general-purpose natural-image datasets under relatively controlled acquisition conditions. Such settings can differ substantially from scientific or ecological microscopy imagery, where fine-grained morphology, non-uniform illumination, imaging noise, background impurities, and skewed class distributions are common and can stress robustness and stability. The demonstration of the FSCIL workflow is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
</sec>
<sec>
<label>2.4</label>
<title>Environmental microorganism recognition</title>
<p>Environmental microorganism recognition has progressed rapidly from classical pipelines based on hand-crafted descriptors (<xref ref-type="bibr" rid="B5">Ciranni et al., 2024</xref>) to modern end-to-end learning frameworks (<xref ref-type="bibr" rid="B22">Zhang et al., 2023</xref>). Recent advances increasingly leverage deep convolutional networks and vision transformers, often coupled with metric learning and attention mechanisms to better capture fine-grained morphology. To reduce reliance on exhaustive annotations, self-supervised and contrastive pretraining, semi-weakly supervised learning, and domain adaptation have become common strategies for extracting transferable visual representations from large collections of unlabeled or heterogeneously labeled microbial images (<xref ref-type="bibr" rid="B2">Bendidi et al., 2024</xref>). In parallel, data-centric solutions such as strong augmentation, synthetic sample generation, and uncertainty-aware sampling further improve robustness under class imbalance and imaging variability&#x02013;conditions that are especially prevalent in environmental microbiology (<xref ref-type="bibr" rid="B11">Pana&#x000EF;otis et al., 2025</xref>).</p>
<p>Despite these methodological improvements, environmental monitoring imposes a fundamental operational constraint: models are typically trained in a static, closed-set manner and therefore can only recognize categories present in the training data, while real deployments often encounter samples from categories that were not included during training due to task diversity and incomplete coverage. Moreover, when such out-of-training categories are encountered, only a small number of labeled instances are usually available, making adaptation under sparse supervision essential. In contrast, many standard CIL settings implicitly assume that each incremental phase provides a non-trivial amount of training data, which can be unrealistic for environmental monitoring scenarios where newly encountered categories may be represented by only a handful of specimens. Conversely, FSL is designed for extreme data scarcity but does not, by itself, address the sequential acquisition of new categories and the need to preserve performance on previously learned classes (<xref ref-type="bibr" rid="B17">Wang et al., 2020</xref>).</p>
<p>This mismatch creates a critical gap for real-world environmental microorganism recognition: systems must simultaneously (i) learn novel microbial classes from very few images and (ii) retain prior knowledge over a long horizon without catastrophic forgetting (<xref ref-type="bibr" rid="B26">Zhou et al., 2024</xref>). FSCIL directly targets this dual requirement, making it particularly important for environmental applications where pronounced intra-class variability and subtle inter-class differences are common. Consequently, positioning FSCIL as a core paradigm&#x02013;and establishing a dedicated FSCIL benchmark for environmental microbial imagery&#x02013;is essential for aligning advanced AI recognition capabilities with the dynamic, long-term monitoring demands of environmental science.</p></sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Benchmark Construction</title>
<p>To establish a standardized, reproducible FSCIL evaluation platform for environmental microorganism image recognition, this study constructs a comprehensive FSCIL benchmark based on the EMDS-7 dataset. This section provides a detailed description of the data preprocessing pipeline, incremental task design, few-shot sampling strategy, and standardized evaluation protocols for assessing algorithm performance. The benchmark is designed to reflect practical ecological monitoring settings, where models trained on a fixed set of categories may encounter previously unseen categories with only a few labeled examples and must incorporate them without degrading performance on previously learned categories. Example images from the EMDS-7 dataset are shown in <xref ref-type="fig" rid="F3">Figure 3</xref>.</p>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>Some examples from the EMDS-7 dataset. Representative microscopic images of environmental microorganisms from the EMDS-7 dataset are illustrated. The dataset covers diverse microorganism categories with significant variations in morphology, scale, and texture. Meanwhile, several categories exhibit high inter-class similarity and complex visual patterns, which introduce fine-grained recognition challenges and increase the difficulty of FSCIL for environmental microorganism image recognition.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0003.tif">
<alt-text content-type="machine-generated">Microscopic images of various algae and microorganisms are arranged in a grid with labels: Oscillatoria, Ankistrodesmus, Chlorella, Gomphonema, Pinnularia, Actinastrum, Cocconeis, Tribonema, Microcystis, Tetraedron, Brachionus, Chaneae, Golenkinia, Spirulina, Cosmarium, Navicula, Scenedesmus, Pediastrum, and Sphaerocystis. Each image shows distinct shapes and structures of the organisms under a microscope.</alt-text>
</graphic>
</fig>
<sec>
<label>3.1</label>
<title>Dataset preprocessing</title>
<p>All 41 categories are partitioned into a base session and multiple incremental sessions. The base session includes 20 categories, with 25 samples per category, for fully supervised representation learning to establish a robust feature space for subsequent increments. The remaining 21 categories are organized into seven incremental sessions, each introducing three novel categories under a 3-way 3-shot setting (<xref ref-type="bibr" rid="B24">Zhang et al., 2024</xref>), which mirrors the sparse, progressive inclusion of previously unseen categories encountered in practical microbial monitoring. The choice of the 3-shot incremental configuration follows the standard FSCIL protocol introduced by (<xref ref-type="bibr" rid="B24">Zhang et al., 2024</xref>), where each new class is represented by only three support samples to simulate extreme data scarcity during deployment. To reduce evaluation bias, the category grouping is constructed to maintain taxonomic diversity and comparable difficulty across sessions. The sampling ratio used in the base session follows the widely adopted 60%&#x02013;40% train&#x02013;test division described by <xref ref-type="bibr" rid="B15">Tao et al. (2020)</xref> for CIFAR-100 and miniImageNet, ensuring that EMDS-7 is aligned with canonical FSCIL dataset construction principles and remains directly comparable to prior work. Overall, this session design follows standard FSCIL protocols and enables controlled yet realistic assessment of how models adapt to new categories while preserving prior knowledge.</p>
<p>In this benchmark, EMDS-7 is converted into an image-level classification setting. Specifically, each original image is treated as one classification sample, without performing bounding-box cropping or object-level filtering. This design choice avoids introducing object-detection&#x02013;dependent biases and ensures that all methods receive identical, unaltered visual evidence. All images are normalized and resized to a fixed resolution of 224 &#x000D7; 224 before being used for training and evaluation. This unified preprocessing avoids introducing design-dependent biases and ensures a standardized input format across all FSCIL methods. To ensure reproducibility, all processed images, session-wise class partitions, and fixed train/test split lists are included in the released benchmark resources.</p>
</sec>
<sec>
<label>3.2</label>
<title>Training and Evaluation Protocol</title>
<p>We follow a widely adopted and challenging FSCIL evaluation protocol to ensure rigor and comparability (<xref ref-type="bibr" rid="B13">Rebuffi et al., 2017</xref>). For fair comparison, all baselines use an ImageNet-pretrained ResNet-18 backbone. In the base stage, the model is trained on all images from the base categories for 100 epochs using SGD with an initial learning rate of 0.1 and standard learning rate decay. During incremental learning, we employ a frozen-backbone setting: backbone parameters remain fixed across sessions, and only the classifier is updated using the current session&#x00027;s 3-shot samples (<xref ref-type="bibr" rid="B20">Zhang C. et al., 2021</xref>). The base-stage training configuration (100 epochs with an initial learning rate of 0.1) follows widely adopted settings in existing FSCIL benchmarks and the original implementations of representative methods, and provides stable convergence on EMDS-7 without altering the relative performance ranking among methods. After each session, the model is evaluated on a cumulative test set containing all seen classes. We report <italic>Average Accuracy</italic> (AA), defined as the mean accuracy across all test sessions (<xref ref-type="bibr" rid="B15">Tao et al., 2020</xref>), and <italic>Performance Drop</italic> (PD), measured as the difference in accuracy between the base and final sessions, to quantify long-term stability and forgetting. This evaluation protocol corresponds to the Session-ID setting, where performance is assessed cumulatively after each incremental session, ensuring a consistent and transparent measurement of class-incremental behavior. All results are averaged over five random runs and summarized with performance evolution curves and tables.</p>
<p>We adopt the Session-ID evaluation protocol, where after each incremental sessions, the model is evaluated on a cumulative test set containing all classes observed up to that session. The Session-ID accuracy is defined as the top-1 classification accuracy on this cumulative set. Average Accuracy (AA) is computed as the mean of Session-ID accuracies across all sessions, and Performance Drop (PD) is defined as the difference between the base-session accuracy and the final-session accuracy. These metrics jointly quantify overall effectiveness and long-horizon stability under class-incremental learning.</p></sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experimental results and analysis</title>
<p>In this section, we compare representative FSCIL methods on the EMDS-7 dataset under the Session-ID evaluation protocol. We report both session-wise accuracies and aggregated metrics to characterize each method&#x00027;s overall effectiveness and long-horizon stability. As shown in Table 1, we provide session-wise accuracies (Sessions 0&#x02013;7) together with AA and PD. <xref ref-type="fig" rid="F4">Figures 4</xref>&#x02013;<xref ref-type="fig" rid="F6">6</xref> visualize the corresponding accuracy curves. For a concise summary, <xref ref-type="table" rid="T2">Table 2</xref> reports an overall ranking based on AA (higher is better) and PD (lower is better), along with each method&#x00027;s core mechanism.</p>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>The comparison of the results of CEC, FACT, SAVC, and PPR on the EMDS-7 dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0004.tif">
<alt-text content-type="machine-generated">Four line graphs show accuracy percentages across eight session IDs for different models. The top left graph shows CEC (CVPR2021) with declining accuracy from 85.40% to 60.90%. The top right graph shows FACT (CVPR2022) with accuracy dropping from 91.60% to 66.60%. The bottom left graph shows SAVC (CVPR2023), decreasing from 92.00% to 66.20%. The bottom right graph shows PFR (PR2024) with accuracy decreasing from 76.90% to 56.80%. Each graph displays consistent downward trends.</alt-text>
</graphic>
</fig>
<sec>
<label>4.1</label>
<title>Experimental setup</title>
<sec>
<label>4.1.1</label>
<title>Compared methods</title>
<p>We reproduce and compare 10 representative FSCIL methods spanning major design routes: classifier evolution (CEC), forward-compatible training (FACT), virtual class contrastive learning (SAVC), frequency-domain decomposition (PFR), adaptive decision boundary (ADBS), compositional learning (Comp), training-free calibration (TEEN), meta-task calibration (Limit), bilateral knowledge distillation (BiDist), and inter-class distance minimization (CLOSER). Following the visualization protocol in <xref ref-type="fig" rid="F4">Figures 4</xref>&#x02013;<xref ref-type="fig" rid="F6">6</xref>, we group the accuracy curves into three panels for clearer trend inspection: (CEC, FACT, SAVC, PFR) in <xref ref-type="fig" rid="F4">Figure 4</xref>, (ADBS, Comp, TEEN, and Limit) in <xref ref-type="fig" rid="F5">Figure 5</xref>, and (BiDist, CLOSER) in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>The comparison of the results of ADBS, Comp, TEEN, and Limit on the EMDS-7 dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0005.tif">
<alt-text content-type="machine-generated">Two line graphs compare accuracy over session IDs for BiDist and CLOSER methods. The BiDist graph shows accuracy decreasing from 46.7% at session 1 to 28.1% at session 8. The CLOSER graph displays accuracy dropping from 70.9% at session 1 to 12.7% at session 8.</alt-text>
</graphic>
</fig>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>The comparison of the results of BiDist and CLOSER on the EMDS-7 dataset.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmicb-17-1770528-g0006.tif">
<alt-text content-type="machine-generated">Four line graphs compare accuracy over eight sessions for different models. ADBS shows a decline from 90.20% to 64.00%, Comp drops from 87.00% to 57.60%, TEEN decreases from 88.30% to 62.10%, and Limit falls from 80.30% to 54.30%. Each graph demonstrates a downward trend in accuracy across sessions.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec>
<label>4.2</label>
<title>Quantitative results</title>
<sec>
<label>4.2.1</label>
<title>Overall comparison</title>
<p>As reported in <xref ref-type="table" rid="T1">Table 1</xref>, SAVC and FACT achieve the best overall <italic>AA</italic> on EMDS-7, reaching 80.4 and 79.9%, respectively. ADBS and TEEN follow next, with both methods attaining 77.1% <italic>AA</italic>. CEC shows mid-level performance with an <italic>AA</italic> of 73.5%, while Comp and Limit yield moderate <italic>AA</italic> values of 69.5 and 68.7%. Among the top and mid-performing methods, PFR exhibits the smallest <italic>PD</italic>, at 20.1%, but its <italic>AA</italic> remains lower at 67.2%, indicating stronger long-horizon stability but a limited separability ceiling. Compared with the above methods, BiDist operates at much lower accuracy, with an <italic>AA</italic> of 37.8% and a <italic>PD</italic> of 18.6%. CLOSER performs worst overall, with an <italic>AA</italic> of 22.2% and a substantially larger drop, where <italic>PD</italic> reaches 58.2%, suggesting markedly different failure patterns from the higher-performing group.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Performance comparison of methods in the Session ID format.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="left"><bold>Venue</bold></th>
<th valign="top" align="center" colspan="8"><bold>Session ID</bold></th>
<th valign="top" align="center"><bold>AA &#x02191;</bold></th>
<th valign="top" align="center"><bold>PD &#x02193;</bold></th>
</tr>
<tr>
<th/>
<th/>
<th valign="top" align="center"><bold>0</bold></th>
<th valign="top" align="center"><bold>1</bold></th>
<th valign="top" align="center"><bold>2</bold></th>
<th valign="top" align="center"><bold>3</bold></th>
<th valign="top" align="center"><bold>4</bold></th>
<th valign="top" align="center"><bold>5</bold></th>
<th valign="top" align="center"><bold>6</bold></th>
<th valign="top" align="center"><bold>7</bold></th>
<th/>
<th/>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">CEC</td>
<td valign="top" align="left">CVPR2021</td>
<td valign="top" align="center">85.4</td>
<td valign="top" align="center">82.3</td>
<td valign="top" align="center">77.1</td>
<td valign="top" align="center">75.0</td>
<td valign="top" align="center">74.1</td>
<td valign="top" align="center">67.4</td>
<td valign="top" align="center">66.2</td>
<td valign="top" align="center">60.9</td>
<td valign="top" align="center">73.5</td>
<td valign="top" align="center">24.5</td>
</tr>
<tr>
<td valign="top" align="left">FACT</td>
<td valign="top" align="left">CVPR2022</td>
<td valign="top" align="center">91.6</td>
<td valign="top" align="center">89.0</td>
<td valign="top" align="center">83.4</td>
<td valign="top" align="center">81.3</td>
<td valign="top" align="center">80.5</td>
<td valign="top" align="center">73.8</td>
<td valign="top" align="center">73.3</td>
<td valign="top" align="center">66.6</td>
<td valign="top" align="center">79.9</td>
<td valign="top" align="center">25.0</td>
</tr>
<tr>
<td valign="top" align="left">SAVC</td>
<td valign="top" align="left">CVPR2023</td>
<td valign="top" align="center">92.9</td>
<td valign="top" align="center">89.0</td>
<td valign="top" align="center">84.7</td>
<td valign="top" align="center">82.7</td>
<td valign="top" align="center">81.7</td>
<td valign="top" align="center">73.3</td>
<td valign="top" align="center">73.7</td>
<td valign="top" align="center">66.2</td>
<td valign="top" align="center">80.4</td>
<td valign="top" align="center">26.7</td>
</tr>
<tr>
<td valign="top" align="left">PFR</td>
<td valign="top" align="left">PR2024</td>
<td valign="top" align="center">76.9</td>
<td valign="top" align="center">75.1</td>
<td valign="top" align="center">70.1</td>
<td valign="top" align="center">68.0</td>
<td valign="top" align="center">67.4</td>
<td valign="top" align="center">62.0</td>
<td valign="top" align="center">61.7</td>
<td valign="top" align="center">56.8</td>
<td valign="top" align="center">67.2</td>
<td valign="top" align="center">20.1</td>
</tr>
<tr>
<td valign="top" align="left">ADBS</td>
<td valign="top" align="left">AAAI2025</td>
<td valign="top" align="center">90.2</td>
<td valign="top" align="center">85.8</td>
<td valign="top" align="center">80.0</td>
<td valign="top" align="center">78.5</td>
<td valign="top" align="center">77.1</td>
<td valign="top" align="center">69.8</td>
<td valign="top" align="center">71.1</td>
<td valign="top" align="center">64.0</td>
<td valign="top" align="center">77.1</td>
<td valign="top" align="center">26.2</td>
</tr>
<tr>
<td valign="top" align="left">Comp</td>
<td valign="top" align="left">ICML2024</td>
<td valign="top" align="center">80.7</td>
<td valign="top" align="center">77.2</td>
<td valign="top" align="center">75.8</td>
<td valign="top" align="center">71.1</td>
<td valign="top" align="center">69.2</td>
<td valign="top" align="center">64.3</td>
<td valign="top" align="center">60.1</td>
<td valign="top" align="center">57.6</td>
<td valign="top" align="center">69.5</td>
<td valign="top" align="center">23.1</td>
</tr>
<tr>
<td valign="top" align="left">TEEN</td>
<td valign="top" align="left">NIPS2023</td>
<td valign="top" align="center">88.3</td>
<td valign="top" align="center">87.1</td>
<td valign="top" align="center">82.4</td>
<td valign="top" align="center">80.1</td>
<td valign="top" align="center">75.3</td>
<td valign="top" align="center">72.0</td>
<td valign="top" align="center">69.2</td>
<td valign="top" align="center">62.1</td>
<td valign="top" align="center">77.1</td>
<td valign="top" align="center">26.2</td>
</tr>
<tr>
<td valign="top" align="left">Limit</td>
<td valign="top" align="left">TPAMI2024</td>
<td valign="top" align="center">80.3</td>
<td valign="top" align="center">77.2</td>
<td valign="top" align="center">75.0</td>
<td valign="top" align="center">70.1</td>
<td valign="top" align="center">66.6</td>
<td valign="top" align="center">65.2</td>
<td valign="top" align="center">61.0</td>
<td valign="top" align="center">54.3</td>
<td valign="top" align="center">68.7</td>
<td valign="top" align="center">26.0</td>
</tr>
<tr>
<td valign="top" align="left">BiDist</td>
<td valign="top" align="left">CVPR2023</td>
<td valign="top" align="center">46.7</td>
<td valign="top" align="center">44.5</td>
<td valign="top" align="center">42.0</td>
<td valign="top" align="center">40.0</td>
<td valign="top" align="center">38.4</td>
<td valign="top" align="center">32.1</td>
<td valign="top" align="center">30.7</td>
<td valign="top" align="center">28.1</td>
<td valign="top" align="center">37.8</td>
<td valign="top" align="center">18.6</td>
</tr>
<tr>
<td valign="top" align="left">CLOSER</td>
<td valign="top" align="left">ECCV2024</td>
<td valign="top" align="center">70.9</td>
<td valign="top" align="center">18.2</td>
<td valign="top" align="center">17.2</td>
<td valign="top" align="center">16.3</td>
<td valign="top" align="center">14.9</td>
<td valign="top" align="center">14.3</td>
<td valign="top" align="center">13.3</td>
<td valign="top" align="center">12.7</td>
<td valign="top" align="center">22.2</td>
<td valign="top" align="center">58.2</td>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Performance ranking based on <italic>AA</italic> and <italic>PD</italic>. (<italic>In</italic> %).</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Rank</bold></th>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="left"><bold>Core mechanism</bold></th>
<th valign="top" align="center"><bold><inline-formula><mml:math id="M1"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mrow><mml:mstyle mathvariant="script"><mml:mi>A</mml:mi></mml:mstyle></mml:mrow></mml:mrow><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="center"><bold><inline-formula><mml:math id="M2"><mml:mstyle mathvariant="bold"><mml:msub><mml:mrow><mml:mrow><mml:mstyle mathvariant="script"><mml:mi>A</mml:mi></mml:mstyle></mml:mrow></mml:mrow><mml:mrow><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub></mml:mstyle></mml:math></inline-formula></bold></th>
<th valign="top" align="center"><bold>PD (&#x02193;)</bold></th>
<th valign="top" align="center"><bold>AA (&#x02191;)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1</td>
<td valign="top" align="left">SAVC</td>
<td valign="top" align="left">Virtual class contrastive learning</td>
<td valign="top" align="center">92.9</td>
<td valign="top" align="center">66.2</td>
<td valign="top" align="center">26.7</td>
<td valign="top" align="center">80.4</td>
</tr>
<tr>
<td valign="top" align="left">2</td>
<td valign="top" align="left">FACT</td>
<td valign="top" align="left">Forward compatible training</td>
<td valign="top" align="center">91.6</td>
<td valign="top" align="center">66.6</td>
<td valign="top" align="center">25.0</td>
<td valign="top" align="center">79.9</td>
</tr>
<tr>
<td valign="top" align="left">3</td>
<td valign="top" align="left">ADBS</td>
<td valign="top" align="left">Adaptive decision boundary</td>
<td valign="top" align="center">90.2</td>
<td valign="top" align="center">64.0</td>
<td valign="top" align="center">26.2</td>
<td valign="top" align="center">77.1</td>
</tr>
<tr>
<td valign="top" align="left">4</td>
<td valign="top" align="left">TEEN</td>
<td valign="top" align="left">Training-free calibration</td>
<td valign="top" align="center">88.3</td>
<td valign="top" align="center">62.1</td>
<td valign="top" align="center">26.2</td>
<td valign="top" align="center">77.1</td>
</tr>
<tr>
<td valign="top" align="left">5</td>
<td valign="top" align="left">CEC</td>
<td valign="top" align="left">Classifier graph evolution</td>
<td valign="top" align="center">85.4</td>
<td valign="top" align="center">60.9</td>
<td valign="top" align="center">24.5</td>
<td valign="top" align="center">73.5</td>
</tr>
<tr>
<td valign="top" align="left">6</td>
<td valign="top" align="left">Comp</td>
<td valign="top" align="left">Compositional learning</td>
<td valign="top" align="center">80.7</td>
<td valign="top" align="center">57.6</td>
<td valign="top" align="center">23.1</td>
<td valign="top" align="center">69.5</td>
</tr>
<tr>
<td valign="top" align="left">7</td>
<td valign="top" align="left">LIMIT</td>
<td valign="top" align="left">Meta-task calibration</td>
<td valign="top" align="center">80.3</td>
<td valign="top" align="center">54.3</td>
<td valign="top" align="center">26.0</td>
<td valign="top" align="center">68.7</td>
</tr>
<tr>
<td valign="top" align="left">8</td>
<td valign="top" align="left">PFR</td>
<td valign="top" align="left">Frequency domain decomposition</td>
<td valign="top" align="center">76.9</td>
<td valign="top" align="center">56.8</td>
<td valign="top" align="center">20.1</td>
<td valign="top" align="center">67.2</td>
</tr>
<tr>
<td valign="top" align="left">9</td>
<td valign="top" align="left">CLOSER</td>
<td valign="top" align="left">Inter-class distance minimization</td>
<td valign="top" align="center">70.9</td>
<td valign="top" align="center">12.7</td>
<td valign="top" align="center">58.1</td>
<td valign="top" align="center">22.2</td>
</tr>
<tr>
<td valign="top" align="left">10</td>
<td valign="top" align="left">BiDist</td>
<td valign="top" align="left">Bilateral knowledge distillation</td>
<td valign="top" align="center">46.7</td>
<td valign="top" align="center">28.1</td>
<td valign="top" align="center">37.8</td>
<td valign="top" align="center">18.6</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
<sec>
<label>4.2.2</label>
<title>Session-wise trends</title>
<p>The accuracy curves in <xref ref-type="fig" rid="F4">Figures 4</xref>&#x02013;<xref ref-type="fig" rid="F6">6</xref> further clarify how different design choices affect long-horizon behavior. In <xref ref-type="fig" rid="F4">Figure 4</xref>, SAVC and FACT start from very high base-session accuracy and maintain strong early-to-mid performance, whereas PFR exhibits a more gradual, near-linear decay, consistent with a stability-oriented design. In <xref ref-type="fig" rid="F5">Figure 5</xref>, ADBS and TEEN remain competitive into later sessions with distinct stabilization patterns, while Comp and Limit show a more pronounced decline toward the final sessions. <xref ref-type="fig" rid="F6">Figure 6</xref> highlights the substantially lower accuracy regimes of BiDist and CLOSER. BiDist degrades smoothly across sessions, with accuracy decreasing from 46.7 to 28.1%. By contrast, CLOSER collapses immediately after the first incremental session, dropping from 70.9 to 18.2%, and then remains at a persistently low level thereafter.</p>
</sec>
</sec>
<sec>
<label>4.3</label>
<title>Results analysis and discussion</title>
<sec>
<label>4.3.1</label>
<title>Accuracy-stability trade-off on EMDS-7</title>
<p><xref ref-type="table" rid="T1">Table 1</xref> and <xref ref-type="fig" rid="F4">Figures 4</xref>&#x02013;<xref ref-type="fig" rid="F6">6</xref> reveal a clear accuracy-stability trade-off on EMDS-7. This trade-off reflects a fundamental interaction between FSCIL method design principles and the intrinsic characteristics of environmental microorganism imagery, rather than isolated empirical variations across methods. Representation-oriented methods such as SAVC and FACT reach the highest accuracy ceiling, with <italic>AA</italic> values of 80.4 and 79.9%, but they also incur relatively large performance drops across sessions, where <italic>PD</italic> rises to 26.7 and 25.0%. In contrast, PFR shows a smaller drop, with <italic>PD</italic> measured at 20.1%, suggesting better long-horizon stability, yet its overall <italic>AA</italic> remains lower at 67.2%, reflecting limited separability in this fine-grained setting. From a methodological perspective, this pattern indicates that on EM imagery, higher <italic>AA</italic> primarily depends on strong base-session discriminability induced by representation learning mechanisms, whereas lower <italic>PD</italic> benefits more from stabilization strategies that explicitly constrain cumulative drift.</p></sec>
<sec>
<label>4.3.2</label>
<title>Top performers: SAVC and FACT</title>
<p>SAVC achieves the highest base-session accuracy of 92.9% and remains strong through Sessions 1&#x02013;4, where performance decreases from 89.0 to 81.7%. This trend indicates that contrastive learning in virtual classes substantially improves fine-grained separability. From a method-design standpoint, SAVC emphasizes aggressive feature discrimination during the base session, which is particularly effective for EMDS-7 where subtle morphological cues dominate class differentiation. However, SAVC exhibits a pronounced drop between Sessions 4 and 5, from 81.7 to 73.3%, and converges to 66.2% in the final session, consistent with the sharper late-stage decline shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. Under dense near-neighbor class structures and extremely low-shot increments, such highly discriminative feature geometry can accumulate boundary crowding and prototype drift, leading to increasingly severe old&#x02013;new conflicts in later sessions.</p>
<p>FACT starts from a similarly high base accuracy and reaches 91.6%, then degrades more smoothly in early-to-mid sessions, with accuracy decreasing from 89.0 to 80.5%. This aligns with its forward-compatible design that delays interference. FACT explicitly attempts to reserve decision space for future classes, which moderates early interference compared with more rigid representation shaping. Nevertheless, FACT still drops to 66.6% at the final session, implying that space reservation alone cannot fully resolve decision boundary overlap induced by tightly clustered fine-grained classes under long incremental chains.</p></sec>
<sec>
<label>4.3.3</label>
<title>Second tier: ADBS and TEEN</title>
<p>ADBS and TEEN reach the same <italic>AA</italic> value of 77.1% but exhibit different stabilization patterns in <xref ref-type="fig" rid="F5">Figure 5</xref>. ADBS produces more controllable curves and shows a mild rebound at Session 6, rising to 71.1% from 69.8% at Session 5. This behavior is consistent with adaptive boundary adjustment that mitigates old-class dominance and reduces old-new conflicts. From a methodological perspective, ADBS dynamically reshapes decision boundaries, allowing limited plasticity while preventing excessive drift, which is advantageous in long-horizon incremental settings. TEEN maintains strong early performance, reaching 87.1% at Session 1, suggesting that training-free calibration can alleviate the initial attraction of novel classes to old ones. However, TEEN continues to decline along the long incremental chain and drops to 62.1% at Session 7, suggesting that prototype-level post-processing alone may accumulate calibration errors in the highly ambiguous EMDS-7 space.</p></sec>
<sec>
<label>4.3.4</label>
<title>Stable but ceiling-limited methods</title>
<p>CEC achieves an <italic>AA</italic> of 73.5% and performs reasonably well in early sessions, but its accuracy drops to 60.9% in the final session. This behavior is consistent with error accumulation in classifier evolution or relationship propagation under dense near-neighbor structures. Comp and Limit operate in a similar mid-range regime, yielding <italic>AA</italic> values of 69.5 and 68.7%, respectively, and neither shows a clear advantage in terms of ceiling or stability on EMDS-7. Comp demonstrates some robustness, with <italic>PD</italic> reaching 23.1%, but its overall separability remains limited, whereas Limit does not outperform other methods on either <italic>AA</italic> or <italic>PD</italic>. PFR, although achieving a lower <italic>AA</italic>, exhibits improved stability, with <italic>PD</italic> measured at 20.1%. From a method-design perspective, frequency-domain decomposition suppresses certain forms of cumulative drift by stabilizing texture-dominant cues in EM imagery, but this comes at the cost of reduced spatial discriminability, limiting the achievable accuracy ceiling.</p></sec>
<sec>
<label>4.3.5</label>
<title>Low-ceiling vs. failure-like behaviors</title>
<p>BiDist and CLOSER exhibit markedly different non-top-performing behaviors (<xref ref-type="fig" rid="F6">Figure 6</xref>). BiDist shows the smallest <italic>PD</italic> in our comparison, with a drop of 18.6%, while its overall <italic>AA</italic> remains substantially lower at 37.8%, indicating a relatively stable yet low-ceiling regime. This pattern is consistent with a distillation-dominated optimization that preserves earlier decision structure but constrains the model&#x00027;s capacity to carve out discriminative regions for newly introduced classes under the 3-way 3-shot setting.</p>
<p>More generally, distillation-dominated methods emphasize stability by preserving historical decision boundaries, but this inductive bias inevitably limits plasticity when discriminative cues between novel classes are subtle and data is extremely scarce. As a result, such methods tend to trade accuracy ceiling for improved long-term retention on EMDS-7.</p>
<p>In contrast, CLOSER undergoes a sharp degradation immediately after the first incremental session, falling from 70.9 to 18.2%, and remains at a very low level thereafter; consequently, it yields a large <italic>PD</italic> of 58.2% together with a low <italic>AA</italic> of 22.2%. Such a trajectory suggests that inter-class distance minimization may become incompatible with EMDS-7&#x00027;s compact, fine-grained class geometry when incremental evidence is extremely scarce. From a methodological perspective, CLOSER explicitly enforces geometric constraints on feature space by minimizing inter-class distances, implicitly assuming reliable and well-separated class prototypes.</p>
<p>However, EMDS-7 exhibits subtle inter-class differences and substantial intra-class variability, leading to highly compact feature distributions across categories. Under the extremely low-shot (3-shot) incremental setting, class prototypes estimated from only a few samples are subject to high uncertainty. When strong geometric constraints are imposed under such conditions, prototype instability can be amplified, resulting in severe decision-boundary overlap and rapid performance collapse.</p>
<p>Beyond these two methods, similar method&#x02013;data interactions consistently emerge across different FSCIL method families. Representation-driven approaches (e.g., SAVC and FACT) favor discriminability and achieve higher accuracy ceilings, whereas stabilization-oriented methods (e.g., ADBS, TEEN, and PFR) prioritize controlling drift and achieve improved long-horizon stability. Constraint-heavy designs, including aggressive distillation or geometric shaping, are particularly fragile under extreme data scarcity. Since all methods are reproduced using public implementations under a unified setup, these observations should be interpreted primarily as empirical signals of method-domain compatibility rather than definitive general conclusions.</p></sec>
<sec>
<label>4.3.6</label>
<title>Summary</title>
<p>Overall, EMDS-7 highlights that achieving high <italic>AA</italic> relies on strong fine-grained discriminability in representation learning, such as SAVC and FACT, while achieving lower <italic>PD</italic> benefits more from lightweight stabilization and calibration, such as ADBS, TEEN, and PFR. From a method-design perspective, this indicates that balancing representation quality and controlled plasticity is critical for FSCIL in environmental microorganism recognition. In near-neighbor classes, with 3-shot increments, and in long-horizon sessions, overly aggressive geometric shaping or distillation constraints may introduce instability and lead to substantial degradation.</p></sec></sec></sec>
<sec id="s5">
<label>5</label>
<title>Conclusions and prospects</title>
<p>In this work, EMDS-7 is introduced to evaluate 10 representative methods under a unified Session-ID protocol systematically, which is the first FSCIL benchmark for environmental microbiology. The results show that environmental microbial images pose distinct challenges, with high intra-class diversity, subtle inter-class differences, and noisy backgrounds, which lead to clear method-dependent behaviors that differ from those on natural-image benchmarks.</p>
<p>A key takeaway is that strong performance on generic FSCIL datasets does not necessarily transfer to EMDS-7; rather, the effectiveness of a method depends on how well its underlying mechanism matches the dataset&#x00027;s compact, fine-grained class geometry. In our experiments, representation-driven approaches such as SAVC and FACT achieve the best overall performance, while calibration-based stabilizers (e.g., ADBS and TEEN) remain highly competitive by mitigating old&#x02013;new conflicts over long session chains. Frequency-domain stabilization methods (e.g., PFR) improve long-term stability but are constrained by a lower separability ceiling, thereby highlighting an explicit accuracy&#x02013;stability trade-off. Frequency-domain stabilization (e.g., PFR) improves stability but is constrained by a lower separability ceiling, highlighting an accuracy&#x02013;stability trade-off.</p>
<p>Overall, EMDS-7 provides a solid experimental foundation and points to promising directio ns, including the need for task-specific FSCIL adaptations such as incorporating domain priors and improving stability&#x02013;plasticity control in fine-grained, data-scarce settings.</p></sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>JZh: Writing &#x02013; original draft. YZ: Writing &#x02013; original draft. SX: Writing &#x02013; review &#x00026; editing. YH: Writing &#x02013; review &#x00026; editing. JZe: Writing &#x02013; review &#x00026; editing. YY: Writing &#x02013; review &#x00026; editing.</p>
</sec>
<ack><title>Acknowledgments</title><p>We thank the maintainers of the EMDS-7 dataset for making the data publicly available and supporting reproducible research. We also thank our colleagues for helpful discussions and feedback on benchmark design, experimental protocol, and result interpretation. We acknowledge the computing resources and technical support provided by our institution that enabled the experiments reported in this work.</p></ack>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s11">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fmicb.2026.1770528/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fmicb.2026.1770528/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_2.pdf" id="SM2" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_3.pdf" id="SM3" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_4.pdf" id="SM4" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_5.pdf" id="SM5" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
<supplementary-material xlink:href="Data_Sheet_6.pdf" id="SM6" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arbefeville</surname> <given-names>S. S.</given-names></name> <name><surname>Timbrook</surname> <given-names>T. T.</given-names></name> <name><surname>Garner</surname> <given-names>C. D.</given-names></name></person-group> (<year>2024</year>). <article-title>Evolving strategies in microbe identification&#x02013;a comprehensive review of biochemical, maldi-tof ms and molecular testing methods</article-title>. <source>J. Antimicrobial Chemotherapy</source>, 79(Supplement_1):i2-i8. doi: <pub-id pub-id-type="doi">10.1093/jac/dkae275</pub-id><pub-id pub-id-type="pmid">39298363</pub-id></mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bendidi</surname> <given-names>I.</given-names></name> <name><surname>Bardes</surname> <given-names>A.</given-names></name> <name><surname>Cohen</surname> <given-names>E.</given-names></name> <name><surname>Lamiable</surname> <given-names>A.</given-names></name> <name><surname>Bollot</surname> <given-names>G.</given-names></name> <name><surname>Genovesio</surname> <given-names>A.</given-names></name></person-group> (<year>2024</year>). <article-title>Exploring self-supervised learning biases for microscopy image representation</article-title>. <source>Biol. Imag</source>. <volume>4</volume>:<fpage>e12</fpage>. doi: <pub-id pub-id-type="doi">10.1017/S2633903X2400014X</pub-id><pub-id pub-id-type="pmid">39776611</pub-id></mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Bennequin</surname> <given-names>E.</given-names></name> <name><surname>Tami</surname> <given-names>M.</given-names></name> <name><surname>Toubhans</surname> <given-names>A.</given-names></name> <name><surname>Hudelot</surname> <given-names>C.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;Few-shot image classification benchmarks are too far from reality: build back better with semantic task sampling,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>4767</fpage>&#x02013;<lpage>4776</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPRW56347.2022.00523</pub-id></mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Chi</surname> <given-names>Z.</given-names></name> <name><surname>Gu</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>H.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Yu</surname> <given-names>Y.</given-names></name> <name><surname>Tang</surname> <given-names>J.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;MetaFSCIL: a meta-learning approach for few-shot class incremental learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>New Orleans, LA</publisher-name>), <fpage>14166</fpage>&#x02013;<lpage>14175</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR52688.2022.01377</pub-id></mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ciranni</surname> <given-names>M.</given-names></name> <name><surname>Murino</surname> <given-names>V.</given-names></name> <name><surname>Odone</surname> <given-names>F.</given-names></name> <name><surname>Pastore</surname> <given-names>V. P.</given-names></name></person-group> (<year>2024</year>). <article-title>Computer vision and deep learning meet plankton: milestones and future directions</article-title>. <source>Image Vision Comput</source>. <volume>143</volume>:<fpage>104934</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.imavis.2024.104934</pub-id></mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>De Lange</surname> <given-names>M.</given-names></name> <name><surname>Aljundi</surname> <given-names>R.</given-names></name> <name><surname>Masana</surname> <given-names>M.</given-names></name> <name><surname>Parisot</surname> <given-names>S.</given-names></name> <name><surname>Jia</surname> <given-names>X.</given-names></name> <name><surname>Leonardis</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>&#x0201C;A continual learning survey: defying forgetting in classification tasks,&#x0201D;</article-title> in <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, Vol. 44 (Palo Alto, CA: AAAI), <fpage>3366</fpage>&#x02013;<lpage>3385</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TPAMI.2021.3057446</pub-id><pub-id pub-id-type="pmid">33544669</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Hospedales</surname> <given-names>T.</given-names></name> <name><surname>Antoniou</surname> <given-names>A.</given-names></name> <name><surname>Micaelli</surname> <given-names>P.</given-names></name> <name><surname>Storkey</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Meta-learning in neural networks: a survey,&#x0201D;</article-title> in <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, Vol. 44 (<publisher-loc>Basel</publisher-loc>: <publisher-name>MDPI</publisher-name>), <fpage>5149</fpage>&#x02013;<lpage>5169</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TPAMI.2021.3079209</pub-id><pub-id pub-id-type="pmid">33974543</pub-id></mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Hou</surname> <given-names>S.</given-names></name> <name><surname>Pan</surname> <given-names>X.</given-names></name> <name><surname>Loy</surname> <given-names>C. C.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name> <name><surname>Lin</surname> <given-names>D.</given-names></name></person-group> (<year>2019</year>). <article-title>&#x0201C;Learning a unified classifier incrementally via rebalancing,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>Long Beach, CA</publisher-name>), <fpage>831</fpage>&#x02013;<lpage>839</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR.2019.00092</pub-id></mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Tan</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>S.</given-names></name> <name><surname>Cheng</surname> <given-names>H.</given-names></name> <name><surname>Dong</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>L.</given-names></name></person-group> (<year>2025</year>). <article-title>&#x0201C;Adaptive decision boundary for few-shot class-incremental learning,&#x0201D;</article-title> in <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>, Vol. <volume>39</volume>, <fpage>18359</fpage>&#x02013;<lpage>18367</lpage>. doi: <pub-id pub-id-type="doi">10.1609/aaai.v39i17.34020</pub-id></mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>J. Y.</given-names></name> <name><surname>Lim</surname> <given-names>K. M.</given-names></name> <name><surname>Lee</surname> <given-names>C. P.</given-names></name> <name><surname>Tan</surname> <given-names>Y. X.</given-names></name></person-group> (<year>2025</year>). <article-title>A review of few-shot image classification: approaches, datasets and research trends</article-title>. <source>Neurocomputing</source> <volume>649</volume>:<fpage>130774</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neucom.2025.130774</pub-id></mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pana&#x000EF;otis</surname> <given-names>T.</given-names></name> <name><surname>Amblard</surname> <given-names>E.</given-names></name> <name><surname>Boniface-Chang</surname> <given-names>G.</given-names></name> <name><surname>Dulac-Arnold</surname> <given-names>G.</given-names></name> <name><surname>Woodward</surname> <given-names>B.</given-names></name> <name><surname>Irisson</surname> <given-names>J.-O.</given-names></name></person-group> (<year>2025</year>). <article-title>Benchmark of plankton images classification: emphasizing features extraction over classifier complexity</article-title>. <source>Earth System Sci. Data Discus</source>. <volume>2025</volume>, <fpage>1</fpage>&#x02013;<lpage>35</lpage>.</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pluym</surname> <given-names>T.</given-names></name> <name><surname>Waegenaar</surname> <given-names>F.</given-names></name> <name><surname>De Gusseme</surname> <given-names>B.</given-names></name> <name><surname>Boon</surname> <given-names>N.</given-names></name></person-group> (<year>2024</year>). <article-title>Microbial drinking water monitoring now and in the future</article-title>. <source>Microb. Biotechnol</source>. <volume>17</volume>:<fpage>e14532</fpage>. doi: <pub-id pub-id-type="doi">10.1111/1751-7915.14532</pub-id><pub-id pub-id-type="pmid">39051617</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Rebuffi</surname> <given-names>S. A.</given-names></name> <name><surname>Kolesnikov</surname> <given-names>A.</given-names></name> <name><surname>Sperl</surname> <given-names>G.</given-names></name> <name><surname>Lampert</surname> <given-names>C. H.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;iCaRL: Incremental classifier and representation learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>Honolulu, HI</publisher-name>), <fpage>2001</fpage>&#x02013;<lpage>2010</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR.2017.587</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>Z.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Shi</surname> <given-names>Y.</given-names></name> <name><surname>Peng</surname> <given-names>P.</given-names></name> <name><surname>Yuan</surname> <given-names>L.</given-names></name> <name><surname>Tian</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;Learning with fantasy: Semantic-aware virtual contrastive constraint for few-shot class-incremental learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>Vancouver, BC</publisher-name>), <fpage>24183</fpage>&#x02013;<lpage>24192</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR52729.2023.02316</pub-id></mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Tao</surname> <given-names>X.</given-names></name> <name><surname>Hong</surname> <given-names>X.</given-names></name> <name><surname>Chang</surname> <given-names>X.</given-names></name> <name><surname>Dong</surname> <given-names>S.</given-names></name> <name><surname>Wei</surname> <given-names>X.</given-names></name> <name><surname>Gong</surname> <given-names>Y.</given-names></name></person-group> (<year>2020</year>). <article-title>&#x0201C;Few-Shot class-incremental learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>Seattle, WA</publisher-name>), <fpage>12183</fpage>&#x02013;<lpage>12192</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR42600.2020.01220</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tsoumplekas</surname> <given-names>G.</given-names></name> <name><surname>Li</surname> <given-names>V.</given-names></name> <name><surname>Sarigiannidis</surname> <given-names>P.</given-names></name> <name><surname>Argyriou</surname> <given-names>V.</given-names></name></person-group> (<year>2024</year>). <article-title>A complete survey on contemporary methods, emerging paradigms and hybrid approaches for few-shot learning</article-title>. <source>arXiv</source> [preprint]. arXiv:2402.03017. doi: <pub-id pub-id-type="doi">10.48550/arXiv.2402.03017</pub-id></mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Yao</surname> <given-names>Q.</given-names></name> <name><surname>Kwok</surname> <given-names>J. T.</given-names></name> <name><surname>Ni</surname> <given-names>L. M.</given-names></name></person-group> (<year>2020</year>). <article-title>Generalizing from a few examples: a survey on few-shot learning</article-title>. <source>ACM Comput. Surv</source>. <volume>53</volume>, <fpage>1</fpage>&#x02013;<lpage>34</lpage>. doi: <pub-id pub-id-type="doi">10.1145/3386252</pub-id></mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>H.</given-names></name> <name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Zhao</surname> <given-names>X.</given-names></name> <name><surname>Cai</surname> <given-names>B.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Ma</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>EMDS-7: environmental microorganism image dataset seventh version for multiple object detection evaluation</article-title>. <source>Front. Microbiol</source>. <volume>14</volume>:<fpage>1084312</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmicb.2023.1084312</pub-id><pub-id pub-id-type="pmid">36891388</pub-id></mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yin</surname> <given-names>Q.</given-names></name> <name><surname>He</surname> <given-names>K.</given-names></name> <name><surname>Collins</surname> <given-names>G.</given-names></name> <name><surname>De Vrieze</surname> <given-names>J.</given-names></name> <name><surname>Wu</surname> <given-names>G.</given-names></name></person-group> (<year>2024</year>). <article-title>Microbial strategies driving low concentration substrate degradation for sustainable remediation solutions</article-title>. <source>npj Clean Water</source> <volume>7</volume>:<fpage>52</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41545-024-00348-z</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>C.</given-names></name> <name><surname>Song</surname> <given-names>N.</given-names></name> <name><surname>Lin</surname> <given-names>G.</given-names></name> <name><surname>Zheng</surname> <given-names>Y.</given-names></name> <name><surname>Pan</surname> <given-names>P.</given-names></name> <name><surname>Xu</surname> <given-names>Y.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Few-Shot Incremental learning with continually evolved classifiers,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>, 12455&#x02013;12464. doi: <pub-id pub-id-type="doi">10.1109/CVPR46437.2021.01227</pub-id></mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Kosov</surname> <given-names>S.</given-names></name> <name><surname>Grzegorzek</surname> <given-names>M.</given-names></name> <name><surname>Shirahama</surname> <given-names>K.</given-names></name> <name><surname>Jiang</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>LCU-Net: a novel low-cost U-Net for environmental microorganism image segmentation</article-title>. <source>Pattern Recognit</source>. <volume>115</volume>:<fpage>107885</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.patcog.2021.107885</pub-id></mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Yin</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Grzegorzek</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <article-title>Applications of artificial neural networks in microorganism image analysis: a comprehensive review from conventional multilayer perceptron to popular convolutional neural network and potential visual transformer</article-title>. <source>Artifi. Intell. Rev</source>. <volume>56</volume>, <fpage>1013</fpage>&#x02013;<lpage>1070</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10462-022-10192-7</pub-id><pub-id pub-id-type="pmid">35528112</pub-id></mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Liu</surname> <given-names>L.</given-names></name> <name><surname>Silven</surname> <given-names>O.</given-names></name> <name><surname>Pietik&#x000E4;inen</surname> <given-names>M.</given-names></name> <name><surname>Hu</surname> <given-names>D.</given-names></name></person-group> (<year>2025</year>). <article-title>&#x0201C;Few-shot class-incremental learning for classification and object detection: A survey,&#x0201D;</article-title> in <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source> (<publisher-loc>IEEE</publisher-loc>). doi: <pub-id pub-id-type="doi">10.1109/TPAMI.2025.3529038</pub-id><pub-id pub-id-type="pmid">40031007</pub-id></mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Zhao</surname> <given-names>P.</given-names></name> <name><surname>Zhao</surname> <given-names>Y.</given-names></name> <name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Hu</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). &#x0201C;Few-shot class-incremental learning for retinal disease recognition,&#x0201D; <italic>IEEE Journal of Biomedical and Health Informatics</italic> (IEEE). doi: <pub-id pub-id-type="doi">10.1109/JBHI.2024.3457915</pub-id></mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>F.</given-names></name> <name><surname>Ye</surname> <given-names>H.</given-names></name> <name><surname>Ma</surname> <given-names>L.</given-names></name> <name><surname>Pu</surname> <given-names>S.</given-names></name> <name><surname>Zhan</surname> <given-names>D.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;Forward compatible Few-Shot Class-Incremental Learning,&#x0201D;</article-title> in <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source> (<publisher-loc>IEEE</publisher-loc>: <publisher-name>New Orleans, LA</publisher-name>), <fpage>9046</fpage>&#x02013;<lpage>9056</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR52688.2022.00884</pub-id></mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>Q.</given-names></name> <name><surname>Qi</surname> <given-names>Z.</given-names></name> <name><surname>Ye</surname> <given-names>H.</given-names></name> <name><surname>Zhan</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name></person-group> (<year>2024</year>). <article-title>&#x0201C;Class-incremental learning: A survey,&#x0201D;</article-title> in <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source> (<publisher-loc>IEEE</publisher-loc>). doi: <pub-id pub-id-type="doi">10.1109/TPAMI.2024.3429383</pub-id><pub-id pub-id-type="pmid">39012754</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>Wei</surname> <given-names>Q.</given-names></name></person-group> (<year>2022</year>). <article-title>Few-shot fine-grained image classification via GNN</article-title>. <source>Sensors</source> <volume>22</volume>:<fpage>7640</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s22197640</pub-id><pub-id pub-id-type="pmid">36236743</pub-id></mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3126996/overview">Jia Guo</ext-link>, Queen Mary University of London, United Kingdom</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1126466/overview">Zhi-Feng Pang</ext-link>, Henan University, China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3325386/overview">Kai Shao</ext-link>, Henan University of Science and Technology, China</p>
</fn>
</fn-group>
</back>
</article>