<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="EN" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Med.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Medicine</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Med.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-858X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fmed.2026.1730353</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Attention-enhanced SAM with PBFO tuning: advancing glioma MRI segmentation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Alhatamleh</surname> <given-names>Salem</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/3093003/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Abu Mhanna</surname> <given-names>Hamad Yahia</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2900716/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Amin</surname> <given-names>Mohammad</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2898364/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Alishwait</surname> <given-names>Amal</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Latayfeh</surname> <given-names>Mohammad</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/3251925/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Mohammad</surname> <given-names>Qutaiba</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Khouqeer</surname> <given-names>Ghada A.</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/3243673/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Alrefai</surname> <given-names>Abdullah</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Alanazi</surname> <given-names>Sitah</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Sandougah</surname> <given-names>Kholoud J.</given-names></name>
<xref ref-type="aff" rid="aff6"><sup>6</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/3117595/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Computer Science, Faculty of Information Technology and Computer Sciences, Yarmouk University</institution>, <city>Irbid</city>, <country country="jo">Jordan</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Medical Imaging, Faculty of Allied Medical Sciences, Isra University</institution>, <city>Amman</city>, <country country="jo">Jordan</country></aff>
<aff id="aff3"><label>3</label><institution>Faculty of Medicine, Yarmouk University</institution>, <city>Irbid</city>, <country country="jo">Jordan</country></aff>
<aff id="aff4"><label>4</label><institution>Department of Physics, College of Science, Imam Mohammad Ibn Saud Islamic University (IMSIU)</institution>, <city>Riyadh</city>, <country country="sa">Saudi Arabia</country></aff>
<aff id="aff5"><label>5</label><institution>Faculty of Medicine, Jordan University of Science and Technology</institution>, <city>Irbid</city>, <country country="jo">Jordan</country></aff>
<aff id="aff6"><label>6</label><institution>Department of Radiology, College of Medicine, Imam Mohammad Ibn Saud Islamic University (IMSIU)</institution>, <city>Riyadh</city>, <country country="sa">Saudi Arabia</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Kholoud J. Sandougah, <email xlink:href="mailto:Ksandougah@imamu.edu.sa">Ksandougah@imamu.edu.sa</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-19">
<day>19</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>13</volume>
<elocation-id>1730353</elocation-id>
<history>
<date date-type="received">
<day>05</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>26</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Alhatamleh, Abu Mhanna, Amin, Alishwait, Latayfeh, Mohammad, Khouqeer, Alrefai, Alanazi and Sandougah.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Alhatamleh, Abu Mhanna, Amin, Alishwait, Latayfeh, Mohammad, Khouqeer, Alrefai, Alanazi and Sandougah</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-19">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>The segmentation of brain tumor MRI images is one of the most challenging tasks because of the variability and complexity associated with tumor tissues. This study introduces PoSAM-ULTRA, an improved segmentation framework designed to enhance the accuracy and robustness of brain tumor segmentation.</p>
</sec>
<sec>
<title>Methods</title>
<p>PoSAM-ULTRA employs the Polar-Bear Foraging Optimisation (PBFO) algorithm for hyperparameter tuning and utilizes an improved Segment Anything Model as its backbone. The framework is based on a ResNet-34 encoder modified to accept a four-channel input (RGB + prior information). Multi-scale feature extraction is performed via DownBlocks, while discriminative feature learning is enhanced using the Convolutional Block Attention Module (CBAM). Attention Gates are incorporated to ensure effective skip connections, and a multistage decoder is used for robust upsampling and feature integration. The model was evaluated on a dataset from the Integrative Genomic Analysis of Diffuse Lower Grade Gliomas (LGG) and compared with UNet, UNet++, and nnUNet.</p>
</sec>
<sec>
<title>Results</title>
<p>The proposed PoSAM-ULTRA model outperformed the baseline models, achieving superior performance with a Dice score of 91.4%, IoU of 88.9%, Accuracy of 99.8%, Precision of 95.2%, and Recall of 93.3%.</p>
</sec>
<sec>
<title>Discussion</title>
<p>The obtained results demonstrate the robustness and reliability of PoSAM-ULTRA in handling the complexity of brain tumor MRI segmentation, highlighting its effectiveness for challenging medical image segmentation tasks.</p>
</sec>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>glioma segmentation</kwd>
<kwd>lower grade gliomas</kwd>
<kwd>medical image analysis</kwd>
<kwd>MRI imaging</kwd>
<kwd>segment anything model (SAM)</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number: IMSIUDDRSP2601).</funding-statement>
</funding-group>
<counts>
<fig-count count="9"/>
<table-count count="3"/>
<equation-count count="36"/>
<ref-count count="50"/>
<page-count count="17"/>
<word-count count="11547"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Precision Medicine</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Brain tumors are not the most prevalent cancers in the world, but they are among the ones with the highest burden based on their adverse impact on neurological function, cognitive deterioration, and overall morbidity and mortality. Their neoplasias are heterogeneous with tumors of varying malignant potential, biological behavior, and clinical course. According to the latest figures provided by the central brain tumor registry of the United States (CBTRUS), over 1.3 million people nationwide have been diagnosed with a primary brain or Central nervous system (CNS) tumor, and up to 85% are non-malignant. However, malignant tumors of the brain, particularly Glioblastoma Multiforme (GBM), have some of the poorest survival rates among all types of cancer (<xref ref-type="bibr" rid="B1">1</xref>). Age plays an essential role in the histological, biological, and prognostic profile of the brain tumor: Glioblastoma has a median age of occurrence at 64 years, with the highest incidence in people aged 75&#x2013;84 (<xref ref-type="bibr" rid="B2">2</xref>). The term &#x201C;low-grade glioma (LGGs)&#x201D; generally refers to diffusive gliomas of the brain that are more aggressive than Grade I but less aggressive than Grade IV (glioblastomas). These include WHO Grade II (&#x201C;low grade&#x201D;) and Grade III (&#x201C;anaplastic&#x201D; or high-intermediate grade) gliomas (<xref ref-type="bibr" rid="B3">3</xref>). Traditionally, gliomas have been classified based on histopathology (cell shape, mitoses, etc.) (<xref ref-type="bibr" rid="B4">4</xref>). And more recently (since about 2016 and as applied by WHO CNS5, 2021), molecular markers are required components of classification (<xref ref-type="bibr" rid="B5">5</xref>).</p>
<p>Imaging is often the very first step raising suspicion of a brain tumor, especially when patients come in with neurological symptoms such as seizures, headache, or changes in cognitive functions. In the case of WHO Grade II and III diffuse gliomas, which are also referred to as LGGs, magnetic resonance imaging (MRI) is the primary method not only for diagnosis but also for monitoring over a long period (<xref ref-type="bibr" rid="B3">3</xref>). Along with the current advances in imaging and molecular diagnostics, the separation of lower-grade gliomas (LGG) from high-grade gliomas (HGG) or other brain lesions continues to be a major clinical hurdle. Accurate segmentation (delineation of tumors on medical images) becomes critical for the diagnosis, progression monitoring, and treatment planning of LGGs. Underestimating the tumor size might lead to incomplete resection and quicker recurrence, while overestimation could mean that healthy brain tissue, especially around the eloquent areas, is destroyed (<xref ref-type="bibr" rid="B5">5</xref>). Manual segmentation is not only a very lengthy process but also highly subjective, particularly when different observers are involved (<xref ref-type="bibr" rid="B5">5</xref>). The differences in tumor delineation may pose a risk to the reproducibility of study results.</p>
<p>CNNs in capturing the spatial characteristics and context information which is why they are very suitable for the detection of non-cancerous tumors in delicate and challenging areas such as the brain. Although AI is a great help, it not only saves the time that is usually taken to perform the task manually but also the orderly and consistent results which are the essentials in crucial health decisions and long-term monitoring. The availability of public datasets like BraTS (Brain Tumor Seg) has greatly accelerated the adoption and validation of CNN-based models. Thus, the AI industry is the one that takes the major and fast-growing part in medical imaging studies. Over the last 10 years, the focus of research shifted toward the segmentation of small brain tumors in simple gliomas. This was achieved mainly due to the drastic increase in the availability of annotated data sets and the improvement of skills. Numerous studies have proven that AI tools, predominantly employing CNNs &#x0026; U-Net architectures, can be extremely accurate. They are often on par or superior to human performance in specific cases. These tools not only excel in marking tumor boundaries but also in differentiating small parts such as the tumor core, edema, and hot areas. Additionally, recent studies have applied a great number of MRI modalities (T1, T1c, T2, FLAIR) &#x0026; high-resolution techniques. Moreover, 3D modeling has been used for these processes. This has contributed to the understanding of the blend and variety in LGGs. Many such initiatives have already been tested in older medical databases, and more are now being tested in new ones. This suggests that they are indeed effective in real-world applications. They assist in treatment and monitoring.</p>
<p>Brain tumor segmentation in magnetic resonance imaging (MRI) is a fundamental step in supporting accurate diagnosis and effective treatment planning. However, the high variability in tumor shape and size, coupled with the noise in medical images, makes segmentation a significant challenge for traditional models. Hence, the a need to develop more accurate and flexible models capable of handling these complexities. This study aims to develop a new model capable of improving the accuracy and reliability of brain tumor segmentation by combining advanced optimization strategies with modern deep learning architectures. The core concepts of PoSAM-ULTRA in glioma MRI image segmentation aim to achieve functional integration and adaptation of existing components, centered around three key innovations: First, the SAM model backbone has been re-engineered to accept four-channel inputs for co-processing RGB and tumor-related information, enriching feature context and improving tumor boundary definition. Second, an enhanced, attention-coordinated feature fusion mechanism has been introduced, incorporating CBAM modules and attention gates into multiscale encoding and interpretation blocks to enhance the differentiation of low-contrast and irregular glioma regions while maintaining structural consistency. Third, the PBFO hyperparameter fusion process represents a novel optimization strategy within medical SAM segmentation pathways, providing more stable convergence and improved autogeneralization. This combination of three elements results in a coherent and specialized enhancement for glioma segmentation. This will contribute to enhancing overall performance compared to popular models such as UNet, UNet++, and nnUNet.</p>
<p>The main contributions of this research can be summarized as follows:</p>
<list list-type="bullet">
<list-item><p>Proposing the PoSAM-ULTRA model, which combines the Polar-Bear Foraging Optimization (PBFO) algorithm to fine-tune hyperparameters with the Segment Anything Model (SAM) as the underlying segmentation architecture.</p></list-item>
<list-item><p>Designing an improved multi-level feature extraction, incorporating attention mechanisms such as CBAM and Attention Gates to optimize the use of lateral connections and eliminate irrelevant information.</p></list-item>
<list-item><p>Adopting a deep supervision mechanism (Deep Supervision) with a complex loss algorithm (BCE, Dice, Focal) to enhance the learning process and achieve a better balance between sensitivity and accuracy.</p></list-item>
<list-item><p>Evaluating the model on the LGG dataset demonstrated its superiority over well-known models in brain image segmentation, achieving the highest performance in Dice, IoU, precision, and recall.</p></list-item>
</list>
<p>Section 2 discusses the techniques used, providing a comprehensive explanation of the dataset, the structure of the proposed approach, and the training methods. Section 3 analyzes the data and evaluates the effectiveness of the proposed PoSAM-ULTRA model in brain tumor segmentation tests. Section 4 reviews key studies in brain tumor segmentation and compares them to the proposed approach. Section 5 concludes with conclusions and recommendations for further research.</p>
</sec>
<sec id="S2">
<label>2</label>
<title>Methodology</title>
<p>In this study, we propose PoSAM-ULTRA, a novel deep learning framework for brain tumor segmentation. The method integrates the Segment Anything Model (SAM) as a backbone with a ResNet34-based 4-channel encoder, enhanced by CBAM attention modules, extended skip connections, and deep supervision through auxiliary outputs, as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. On the performance side, PBFO hyperparameter search is performed, greatly minimizing tuning expenses. The dataset used is the Integrative Genomic Analysis of Diffuse Lower-Grade Gliomas (LGG) collection composed of paired MRI scans and tumor masks. Image preprocessing is done by resizing, normalizing, and augmenting for better robustness. Then the final models are trained for 50 epochs using the configurations derived from the optimization technique. Performance measurement itself takes place with an extended list of evaluation criteria, namely, Dice, IoU, Precision, Recall, Accuracy. Baselines of U-Net, U-Net++, and nnU-Net-lite serve as comparison points for PoSAM-ULTR.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>Overview of the proposed PoSAM-ULTRA framework for glioma segmentation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g001.tif">
<alt-text content-type="machine-generated">Flowchart illustrating the PoSAM-ULTRA framework for brain tumor MRI analysis, including data preprocessing steps, improved SAM algorithm, hyperparameter optimization, evaluation metrics, and visual comparison of ground truth and predicted tumor masks on brain MRI images.</alt-text>
</graphic>
</fig>
<sec id="S2.SS1">
<label>2.1</label>
<title>Dataset</title>
<p>The LGG Segmentation Dataset is a publicly accessible database of MRI brain scans for lower-grade gliomas, originally released by Mazurowski et al. (<xref ref-type="bibr" rid="B6">6</xref>). It involves MR images with manually provisioned FLAIR abnormality segmentation masks that allow for the exact delineation of chromatic abnormality regions. The dataset includes data from 110 patients from The Cancer Genome Atlas (TCGA) LGG collection, all of whom have FLAIR sequences and associated genomic cluster information. In addition to the images, patient metadata and genomic tumor cluster information are included in a separate CSV file for integrated studies of imaging and genomic characteristics. The MRI scans are from The Cancer Imaging Archive (TCIA), providing a set of standardized and publicly available imaging data. This data has been used to investigate the associations between tumor shape characteristics, as determined by deep learning, and a tumor&#x2019;s genomic subtype, as well as the effect size of these associations on patient outcomes. <xref ref-type="fig" rid="F2">Figure 2</xref> shows an example from the LGG Segmentation Dataset.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>Brain MRI images with ground truth tumor masks.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g002.tif">
<alt-text content-type="machine-generated">Grid layout displaying six brain MRI scans in the top row, each showing different tumor shapes and locations, and corresponding tumor segmentation masks as white regions on black backgrounds in the bottom row.</alt-text>
</graphic>
</fig>
</sec>
<sec id="S2.SS2">
<label>2.2</label>
<title>Data preprocessing</title>
<p>Preprocessing medical images is of utmost importance to maintain data in a standardized form, ensuring all properties of robustness and suitability for deep-learning models. In PoSAM-ULTRA, the entire pipeline of preprocessing steps was designed, encompassing resizing, normalization, data augmentation, and dataset partitioning, of MRI scans and their respective tumor masks of the LGG dataset (<xref ref-type="bibr" rid="B7">7</xref>). These steps ensure standard input dimensions, minimize intensity variability between subjects, increase training examples, and allow for unbiased evaluation. Resizing since an MRI scan naturally differs in resolution and aspect ratio, both images and their respective masks were resized to a uniform resolution of 256 &#x00D7; 256 pixels guaranteeing compatibility of all inputs with the encoder-decoder-style PoSAM-ULTRA, while simultaneously providing adequate spatial detailing for precise detections of tumor boundaries. <italic>I</italic> represent the original MRI image (<xref ref-type="bibr" rid="B8">8</xref>). One gets the corresponding ground mask resized with the same transformation to maintain spatial alignment, as shown in <xref ref-type="disp-formula" rid="S2.E1">Equation 1</xref>.</p>
<disp-formula id="S2.E1">
<mml:math id="M1">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>I</mml:mi>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>256</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(1)</label></disp-formula>
<p>Normalized medical images often suffer from varying intensity distributions due to differences in acquisition protocols and patient-specific contrast variations (<xref ref-type="bibr" rid="B9">9</xref>). To stabilize the training process and accelerate convergence, pixel values were normalized to the range [0, 1] using min-max scaling <italic>I</italic><sub><italic>min</italic></sub> and <italic>I</italic><sub><italic>max</italic></sub> denote the minimum and maximum pixel intensities of the image, respectively. This normalization ensures that the input dynamic range is consistent across all subjects, allowing the network to focus on structural differences rather than intensity biases, as shown in <xref ref-type="disp-formula" rid="S2.E2">Equation 2</xref>.</p>
<disp-formula id="S2.E2">
<mml:math id="M2">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>-</mml:mo>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(2)</label></disp-formula>
<p>Data augmentation to improve generalization and reduce overfitting, extensive on-the-fly augmentation was applied during training (<xref ref-type="bibr" rid="B10">10</xref>). Augmentation increases the diversity of the dataset by simulating variations in patient anatomy and acquisition conditions. The following transformations were randomly applied. Rotation random angular rotations up to &#x00B1; 20, simulating changes in patient orientation. Flipping: Random horizontal and vertical flips; please enforce symmetrical invariance (<xref ref-type="bibr" rid="B11">11</xref>). Scaling/Zooming small random rescaling of images, making the training robust toward changing tumor sizes. Intensity Adjustment changes brightness and contrast to simulate changes in imaging conditions. Formally, the augmented dataset is defined as <italic>D</italic> is the original dataset, <italic>M</italic> is the mask, and &#x1D4AF; is the distribution of transformations applied. This guarantees that the model sees more diverse tumor appearances, which is critical for robust segmentation of low-grade gliomas, as shown in <xref ref-type="disp-formula" rid="S2.E3">Equation 3</xref>.</p>
<disp-formula id="S2.E3">
<mml:math id="M3">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>I</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>I</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x2208;</mml:mo>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x223C;</mml:mo>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(3)</label></disp-formula>
<p>The dataset was divided into 80% training (3,143 images), 10% validation (393 images), and 10% testing (393 images) to ensure a fair and unbiased assessment of performance. The training set <italic>D</italic><sub><italic>train</italic></sub> is used to optimize model weights via backpropagation. The validation set <italic>D</italic><sub><italic>val</italic></sub> is used for hyperparameter tuning and observing model convergence. The test set <italic>D</italic><sub><italic>test</italic></sub> is reserved for the final evaluation to get a measure of conscious ability as shown in <xref ref-type="disp-formula" rid="S2.E4">Equation 4</xref>.</p>
<disp-formula id="S2.E4">
<mml:math id="M4">
<mml:mrow>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mi>D</mml:mi>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x222A;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x222A;</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo rspace="5.8pt">|</mml:mo>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+1.7pt">
<mml:mn>0.8</mml:mn>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo rspace="5.8pt">|</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mpadded width="+6.7pt">
<mml:mn>0.1</mml:mn>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo rspace="5.8pt">|</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mn>0.1</mml:mn>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(4)</label></disp-formula>
<p>To generate the four-channel inputs for the SAM backbone, a prior-information map is incorporated to highlight glioma probability regions. This prior-map is generated via a lightweight preprocessing pipeline that begins with skull-stripping using HD-BET and image intensity normalization using a Z-score. A primary tumor probability map is then produced through unsupervised clustering (K-means with &#x0024;k = 3&#x0024;), where the cluster with the highest average intensity is considered the primary tumor candidate. This mask is refined using morphological closures and a median filter to reduce noise, and then min&#x2013;max normalization to fit the range [0, 1]. This prior-channel approach has proven to provide meaningful spatial semantics, confirming that it provides additional structural cues that identify potential tumor regions without replacing the detailed learning process of the model.</p>
<p>A brain tumor dataset is, generally, characterized by extreme class imbalance, for tumor-region image pixels make up a much smaller proportion of the total image than do background pixels. To confront this, the weights of the loss function were optimized through PBFO to maintain an equilibrium between Dice loss and binary cross-entropy (BCE). Thus, the model is made to be sensitive to the tumor of small structures, while at the same time avoiding any chance of its overbearing influence by background pixels. The preprocessing pipeline allowed for the standardization, diversification, and balancing of the dataset from which PoSAM-ULTRA can draw worthy inputs. These would have helped directly in accurate tumor boundary detection, patient-to-patient generalization, and reduction of overfitting of the model so that it can perform well in segmentation against the baseline methods.</p>
</sec>
<sec id="S2.SS3">
<label>2.3</label>
<title>Improved segment anything model</title>
<p>The method integrates Segment Anything Model (SAM) as a backbone with a ResNet34-based 4-channel encoder, enhanced by CBAM attention modules, extended skip connections, and deep supervision through auxiliary outputs. To optimize performance, we employ the Polar-Bear Foraging Optimization (PBFO) algorithm for efficient hyperparameter search (<xref ref-type="fig" rid="F3">Figure 3</xref>).</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>Architecture overview of the six-stage improved SAM-based segmentation framework (PoSAM-ULTRA).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g003.tif">
<alt-text content-type="machine-generated">Flowchart diagram of a convolutional neural network architecture for segmentation tasks, detailing six distinct stages: patch embedding, hierarchical feature encoding, attention-enhanced downsampling, bottleneck fusion, decoding with skip connections, and refinement plus deep supervision. Color-coded blocks represent operations such as convolutions, batch normalization, ReLU activation, multi-head self-attention, pooling, skip connections, and masking, with arrows indicating the data flow and feature map transitions through the model leading to a final tumor mask output.</alt-text>
</graphic>
</fig>
<sec id="S2.SS3.SSS1">
<label>2.3.1</label>
<title>Segment anything model</title>
<p>The Segment Anything Model (SAM) serves as the primary backbone of PoSAM-ULTRA, providing powerful feature extraction for MRI brain tumor images (<xref ref-type="bibr" rid="B12">12</xref>). SAM is designed to capture both global and local structures, allowing the model to identify fine-grained tumor regions while preserving spatial context. Given an input image <italic>I</italic> &#x2208; &#x211D;<sup>C&#x2003;H&#x2003;W</sup>, SAM computes hierarchical feature maps <italic>F</italic><sup>(<italic>l</italic>)</sup> at multiple layers <italic>l</italic> using convolutional operations, as shown in <xref ref-type="disp-formula" rid="S2.E5">Equation 5</xref>.</p>
<disp-formula id="S2.E5">
<mml:math id="M5">
<mml:mrow>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mi>W</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x002A;</mml:mo>
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>-</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:msup>
<mml:mi>b</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mi>I</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(5)</label></disp-formula>
<p>Where &#x002A; denotes convolution, <italic>W</italic><sup>(<italic>l</italic>)</sup> and <italic>b</italic><sup>(<italic>l</italic>)</sup> are the weights and biases of layer <italic>l</italic>, whereas &#x03C3; is a nonlinear activation function like ReLU and GELU (<xref ref-type="bibr" rid="B13">13</xref>). In previous layers, the feature maps assimilate gradual low-texture details and in deeper layers, they integrate high-level semantic information. SAM incorporates an attention mechanism to improve feature relevance by emphasizing salient regions in the data. Let M<sub><italic>c</italic></sub>(F) and M<sub><italic>s</italic></sub>(F) be the channel and spatial attention maps, and the attended features are, as shown in <xref ref-type="disp-formula" rid="S2.E6">Equation 6</xref>.</p>
<disp-formula id="S2.E6">
<mml:math id="M6">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mi>F</mml:mi>
</mml:mpadded>
<mml:mo>&#x00D7;</mml:mo>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x00D7;</mml:mo>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(6)</label></disp-formula>
<p>This selective weighting enables the backbone to focus on tumor areas and discourage irrelevant areas from the background. Down sampling across layers is applied to reduce spatial resolution and to allow for receptor field growth by way of max pooling or stride convolutions (<xref ref-type="bibr" rid="B14">14</xref>). Finally, SAM further generates multi-level feature maps {<italic>F</italic><sup>(<italic>l</italic><sub>1</sub>)</sup>,<italic>F</italic><sup>(<italic>l</italic><sub>2</sub>)</sup>,<italic>F</italic><sup>(<italic>l</italic><sub>3</sub>)</sup>} that are fed into the Res34LikeEncoder4ch and downstream decoder, thereby providing a rich representation for further segmentation processing, as shown in <xref ref-type="disp-formula" rid="S2.E7">Equation 7</xref>.</p>
<disp-formula id="S2.E7">
<mml:math id="M7">
<mml:mrow>
<mml:mpadded width="+6.6pt">
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>D</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>w</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo mathsize="142%" stretchy="false">&#x2032;</mml:mo>
<mml:mo>&#x2063;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
<mml:mo rspace="10.8pt">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mo mathsize="142%" stretchy="false">&#x2032;</mml:mo>
<mml:mo>&#x2063;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
<mml:mo rspace="7.5pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(7)</label></disp-formula>
</sec>
<sec id="S2.SS3.SSS2">
<label>2.3.2</label>
<title>Hierarchical 4-channel feature encoder</title>
<p>The encoder from PoSAM-ULTRA is a four-channel convolutional neural network inspired by ResNet34 that extracts hierarchical features from multimodal MRI data (<xref ref-type="bibr" rid="B15">15</xref>). The encoder input, <italic>X</italic> &#x2208; &#x211D;<sup>4H&#x2003;W</sup>), has three MRI modalities and one prior probability map as channels, each giving complementary structural and tumor-specific information. The stem layer applies a 7 &#x00D7; 7 convolution plus batch normalization and ReLU activation, as shown in <xref ref-type="disp-formula" rid="S2.E8">Equation 8</xref>.</p>
<disp-formula id="S2.E8">
<mml:math id="M8">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>7</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(8)</label></disp-formula>
<p>A max-pooling operation halves each spatial dimension for a first-level feature map <italic>F<sub>1</sub></italic>. Then the encoder receives <italic>F<sub>1</sub></italic> into four hierarchical DownBlocks, wherein each DownBlock entails two consecutive 3 &#x00D7; 3 convolutions with batch normalization and ReLU activation, and optionally a stride for downsampling to reduce spatial resolution (<xref ref-type="bibr" rid="B16">16</xref>). The four levels of feature maps <italic>C</italic><sub>1</sub>,<italic>C</italic><sub>2</sub>,<italic>C</italic><sub>3</sub>,<italic>C</italic><sub>4</sub> are produced and denote increasingly abstract and semantically rich representations of the input. These hierarchical features will be used later in the skipping connections to the decoder for localizing tumor regions with high accuracy. The encoder can thus be formulated mathematically as follows, as shown in <xref ref-type="disp-formula" rid="S2.E9">Equations 9</xref>, <xref ref-type="disp-formula" rid="S2.E10">10</xref>.</p>
<disp-formula id="S2.E9">
<mml:math id="M9">
<mml:mrow>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>3</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mi>l</mml:mi>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">&#x22EF;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(9)</label></disp-formula>
<disp-formula id="S2.E10">
<mml:math id="M10">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>E</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>d</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>X</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(10)</label></disp-formula>
<p>Each <italic>C</italic><sub><italic>i</italic></sub> &#x2208; &#x211D;<sup><italic>F</italic><sub><italic>i</italic></sub> <italic>H</italic><sub><italic>i</italic></sub> <italic>W</italic><sub><italic>i</italic></sub></sup> corresponds to increasingly abstract and semantically rich representation of the input data. These hierarchical features are then passed to the decoder through skip connections, allowing correct localization of tumor areas. Hence, with this hierarchical design, PoSAM-ULTRA can effectively balance global context with local tumor localization for the challenging lower-grade glioma segmentation task.</p>
</sec>
<sec id="S2.SS3.SSS3">
<label>2.3.3</label>
<title>Attention-enhanced downblock</title>
<p>The Attention-Enhanced DownBlock forms the fundamental processing unit of the encoder in PoSAM-ULTRA, integrating three components a basic convolutional block conv_block, progressive spatial reduction through the DownBlock, and adaptive feature refinement via the Convolutional Block Attention Module (CBAM) (<xref ref-type="bibr" rid="B17">17</xref>). Ultimately, these features allow for effective multi-scale feature learning while also improving the network ability to emphasize the tumor-relevant areas. At its core, each DownBlock begins with a convolutional block that consists of three sequential operations: convolution, followed by batch normalization and then ReLU activation. Given the input feature map <italic>F</italic> &#x2208; &#x211D;<sup>C&#x2003;H&#x2003;W</sup>, it performs a basic but effective operation that adapts the local spatial patterns and then prepares the local spatial pattern features for a step of another layer of hierarchy even deeper in the neural network, as shown in <xref ref-type="disp-formula" rid="S2.E11">Equation 11</xref>.</p>
<disp-formula id="S2.E11">
<mml:math id="M11">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>3</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(11)</label></disp-formula>
<p>The DownBlock stacks two conv blocks consecutively and optionally downsamples by stride to reduce spatial resolution and increase feature channels. The idea is to let the network alternately capture higher semantic information, where <italic>s</italic>1 means feature refinement at the same resolution and <italic>s</italic>2 implies spatial down sampling of 2 times. A hierarchical stacking of DownBlocks across encoder stages maintains a balance between fine-grained localization and contextual abstraction, as shown in <xref ref-type="disp-formula" rid="S2.E12">Equation 12</xref>.</p>
<disp-formula id="S2.E12">
<mml:math id="M12">
<mml:mrow>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>L</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+6.6pt">
<mml:mn>3</mml:mn>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(12)</label></disp-formula>
<p>To increase the encoder representational capacity in PoSAM-ULTRA CBAM is embedded within the convolutional layers (<xref ref-type="bibr" rid="B18">18</xref>). CBAM applies channel attention and then spatial attention, which lets the network adaptively refine the features by concentrating on the most discriminating regions of brain tumor MRI scans. Channel attention: Given an intermediate feature map <italic>X</italic> &#x2208; &#x211D;<sup>C&#x2003;H&#x2003;W</sup>, CBAM first infers a channel attention map by aggregating spatial information using both global average pooling and global max pooling, and W<sub>0</sub> &#x0026; W<sub>1</sub> are learnable weights, &#x03C3; is the ReLU activation, and &#x03C3; is the sigmoid function. The refined feature is obtained, as shown in <xref ref-type="disp-formula" rid="S2.E13">Equations 13</xref>, <xref ref-type="disp-formula" rid="S2.E14">14</xref>.</p>
<disp-formula id="S2.E13">
<mml:math id="M13">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="7.5pt">=</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B4;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>g</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B4;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>M</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(13)</label></disp-formula>
<disp-formula id="S2.E14">
<mml:math id="M14">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>F</mml:mi>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mtext>F</mml:mtext>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(14)</label></disp-formula>
<p>Spatial attention emphasizes tumor-relevant regions within each feature map. Average pooling and max pooling are applied along the channel axis, and the two outputs are concatenated to form a spatial descriptor, and <italic>f</italic><sup>7&#x2003;7</sup> denotes a convolution with kernel size 7 &#x00D7; 7. The spatially refined feature is then computed (<xref ref-type="bibr" rid="B19">19</xref>). The output of CBAM is the sequentially refined feature map. By combining channel-wise and spatial-wise attention, CBAM enables PoSAM-ULTRA to suppress irrelevant background signals and highlight tumor boundaries, which is particularly beneficial for segmenting low-contrast glioma regions, as shown in <xref ref-type="disp-formula" rid="S2.E15">Equations 15</xref>, <xref ref-type="disp-formula" rid="S2.E16">16</xref>.</p>
<disp-formula id="S2.E15">
<mml:math id="M15">
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msup>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>7</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mi>A</mml:mi>
<mml:mi>v</mml:mi>
<mml:mi>g</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>;</mml:mo>
<mml:mo>;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mi>a</mml:mi>
<mml:mi>x</mml:mi>
<mml:mi>P</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>o</mml:mi>
<mml:mi>l</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
<label>(15)</label></disp-formula>
<disp-formula id="S2.E16">
<mml:math id="M16">
<mml:mrow>
<mml:mpadded width="+6.6pt">
<mml:msub>
<mml:mover accent="true">
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">^</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo>=</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>M</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msup>
<mml:mi>F</mml:mi>
<mml:mo>&#x2032;</mml:mo>
</mml:msup>
<mml:mo rspace="5.8pt" stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi>c</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(16)</label></disp-formula>
<p>Combining convolutional feature extraction with progressive downsampling and CBAM-based adaptive attention turns the Attention-Enhanced DownBlock into the most effective discriminant for PoSAM-ULTRA in differentiating tumors from neighboring brain tissues. This design choice places more emphasis on global context relative to local boundary precision, of utmost importance for glioma segmentation. Progressive downsampling reduces computations by shrinking the feature map sizes at deeper layers.</p>
</sec>
<sec id="S2.SS3.SSS4">
<label>2.3.4</label>
<title>Decoder with extended skip connections</title>
<p>The decoder in PoSAM-ULTRA performs the task of reconstructing the high-resolution segmentation mask from the multi-scale hierarchical features extracted by the encoder (<xref ref-type="bibr" rid="B20">20</xref>). Different from common decoders, skip connections extended beyond commonplace lengths are used to retain fine spatial details from the encoder, which is crucial in drawing tumor boundaries accurately in MRI scans. At a particular stage of the decoder, the provided lower-resolution feature map <inline-formula><mml:math id="INEQ15"><mml:mrow><mml:msubsup><mml:mi>F</mml:mi><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mi>&#x211D;</mml:mi><mml:mrow><mml:mpadded width="+6.6pt"><mml:mrow><mml:mtext>C</mml:mtext></mml:mrow></mml:mpadded><mml:mo>&#x2062;</mml:mo><mml:mpadded width="+6.6pt"><mml:mrow><mml:mtext>H</mml:mtext></mml:mrow></mml:mpadded><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mtext>W</mml:mtext></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> is first upsampled, normally, by bilinear interpolation and transposed convolution (<xref ref-type="bibr" rid="B21">21</xref>), to the spatial dimensions of the corresponding encoder feature map <italic>C<sub>l</sub></italic>, as shown in <xref ref-type="disp-formula" rid="S2.E17">Equation 17</xref>.</p>
<disp-formula id="S2.E17">
<mml:math id="M17">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msubsup>
<mml:mi>F</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>F</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(17)</label></disp-formula>
<p>The upsampled feature map is then concatenated with the encoder feature maps from all previous stages through extended skip connections. This concatenation enables the decoder to benefit from both the low-level spatial information and the high-level semantic feature information at the same time, which is useful for improving localization of small and irregular tumors, as shown in <xref ref-type="disp-formula" rid="S2.E18">Equation 18</xref>.</p>
<disp-formula id="S2.E18">
<mml:math id="M18">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msubsup>
<mml:mi>F</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>3</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:msubsup>
<mml:mi>F</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>l</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mrow>
<mml:mi>l</mml:mi>
<mml:mo>-</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">&#x22EF;</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(18)</label></disp-formula>
<p>In the decoder, attention integration of the CBAM by (<xref ref-type="bibr" rid="B22">22</xref>) attention applied in the encoder is similarly implemented in a refinement step in the decoder on each concatenated feature block to enable attention to tumor-related regions and suppression of features that are unrelated to the tumor in the background (<xref ref-type="bibr" rid="B23">23</xref>). Progressive reconstruction then involves the decoder performing multiple stages of up sampling, concatenation, and attention refinement until the feature map reaches the original input resolution HW. The final segmentation output <italic>S</italic> &#x2208; &#x211D;<sup>1H&#x2003;W</sup> is obtained through a 11convolution followed by a sigmoid activation, as shown in <xref ref-type="disp-formula" rid="S2.E19">Equations 19</xref>, <xref ref-type="disp-formula" rid="S2.E20">20</xref>.</p>
<disp-formula id="S2.E19">
<mml:math id="M19">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msubsup>
<mml:mover accent="true">
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">^</mml:mo>
</mml:mover>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>B</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>A</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>M</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>F</mml:mi>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(19)</label></disp-formula>
<disp-formula id="S2.E20">
<mml:math id="M20">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mi>S</mml:mi>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>1</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mover accent="true">
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">^</mml:mo>
</mml:mover>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(20)</label></disp-formula>
<p>The decoder can reconstruct a segmentation mask that is spatially precise and informed semantically owing to extended skip connections with CBAM refinement at every stage. Consequently, this design enables PoSAM-ULTRA to accurately segment the lower-grade glioma, capturing delicate boundaries of the tumors that standard encoder-decoder architectures lose.</p>
</sec>
<sec id="S2.SS3.SSS5">
<label>2.3.5</label>
<title>Deep supervision</title>
<p>In PoSAM-ULTRA, deep supervision introduces additional outputs to represent intermediate layers of the decoder to support gradient flow and to stabilize learning (<xref ref-type="bibr" rid="B24">24</xref>). With intermediate predictions, the model receives more influences during backpropagation, leading to improved learning of fine-grained tumor structures and reducing vanishing gradient issues in deeper networks. Say <inline-formula><mml:math id="INEQ19"><mml:mpadded width="+5pt"><mml:msubsup><mml:mi>F</mml:mi><mml:mi>d</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup></mml:mpadded></mml:math></inline-formula>denotes the feature map at an intermediate decoder layer <italic>l</italic>. The auxiliary output is created by applying 11 convolution followed with a sigmoid activation, as shown in <xref ref-type="disp-formula" rid="S2.E21">Equation 21</xref>.</p>
<disp-formula id="S2.E21">
<mml:math id="M21">
<mml:mrow>
<mml:mpadded width="+8.3pt">
<mml:msubsup>
<mml:mi>S</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mi mathvariant="normal">&#x03C3;</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi>v</mml:mi>
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:mn>1</mml:mn>
</mml:mpadded>
<mml:mo rspace="5.8pt">&#x00D7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mover accent="true">
<mml:mi>F</mml:mi>
<mml:mo stretchy="false">^</mml:mo>
</mml:mover>
<mml:mi>d</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(21)</label></disp-formula>
<p>PoSAM-ULTRA uses two auxiliary outputs aux1 and aux2 (<xref ref-type="bibr" rid="B25">25</xref>), from different decoder stages. Both auxiliary outputs are penalty terms in the final loss function which is composed of the main segmentation loss &#x2112;<sub><italic>main</italic></sub>, and losses &#x2112;<sub><italic>aux</italic>1</sub> &#x0026; &#x2112;<sub><italic>aux</italic>2</sub> from the auxiliary outputs, where &#x03BB;<sub>1</sub> and &#x03BB;<sub>2</sub> weight how much each auxiliary loss contributes to the total loss. Typically, each loss is penalized by the Dice loss or the binary cross-entropy loss, as shown in <xref ref-type="disp-formula" rid="S2.E22">Equation 22</xref>.</p>
<disp-formula id="S2.E22">
<mml:math id="M22">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi class="ltx_font_mathcaligraphic">&#x2112;</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi class="ltx_font_mathcaligraphic">&#x2112;</mml:mi>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03BB;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi class="ltx_font_mathcaligraphic">&#x2112;</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x03BB;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x2062;</mml:mo>
<mml:msub>
<mml:mi class="ltx_font_mathcaligraphic">&#x2112;</mml:mi>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(22)</label></disp-formula>
<p>The deep supervision used in PoSAM-ULTRA feeds the learning of useful representations of both high-level semantic features and low-level spatial details into the entire network, allowing tumor segmentation to better identify small or low-contrast regions that are quite risky for conventional architectures.</p>
</sec>
</sec>
<sec id="S2.SS4">
<label>2.4</label>
<title>Hyperparameter with polar-bear foraging optimization</title>
<p>PoSAM-ULTRA applies the Polar-Bear Foraging Optimization (PBFO) algorithm to fine-tune hyperparameters vital to the system such as the learning rate, batch size, dropout ratio, weight decay, and the number of filters per encoder block (<xref ref-type="bibr" rid="B26">26</xref>). Each candidate hyperparameter configuration is encoded as a vector, where <italic>h</italic><sub><italic>ij</italic></sub> stands for the <italic>j</italic>-th hyperparameter (<xref ref-type="fig" rid="F4">Figure 4</xref>).</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>PBFO algorithm workflow for automated hyperparameter tuning.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g004.tif">
<alt-text content-type="machine-generated">Flowchart outlining the process of hyperparameter optimization using Polar-Bear Foraging Optimization (PBFO), detailing steps from defining the search space, population initialization, iteration, fitness evaluation, updating best solutions, exploitation and exploration phases, boundary handling, and decision conditions for stopping, ending with returning the best hyperparameters.</alt-text>
</graphic>
</fig>
<p>Given iteration <italic>t</italic>, the population of candidate solutions is denoted; altogether, there are <italic>m</italic> candidates. This initial population is randomly generated within a predefined bound. <italic>U</italic>(<italic>a</italic>,<italic>b</italic>) stands for the uniform distribution (<xref ref-type="bibr" rid="B27">27</xref>). For instance, the learning rate may be picked in the range [10<sup>&#x2212;5</sup>,10<sup>&#x2212;2</sup>], batch size in the range [8,64] and dropout rate in the range [0.1,0.6]. Each candidate solution then undergoes evaluation against a fitness function that reflects the quality of the segmentation. For PoSAM-ULTRA, the Dice similarity coefficient is primarily used as a metric, as shown in <xref ref-type="disp-formula" rid="S2.E23">Equations 23</xref>&#x2013;<xref ref-type="disp-formula" rid="S2.E26">26</xref>.</p>
<disp-formula id="S2.E23">
<mml:math id="M23">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">&#x22EF;</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(23)</label></disp-formula>
<disp-formula id="S2.E24">
<mml:math id="M24">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>P</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mn>1</mml:mn>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mn>2</mml:mn>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mn>3</mml:mn>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">&#x22EF;</mml:mi>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>m</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(24)</label></disp-formula>
<disp-formula id="S2.E25">
<mml:math id="M25">
<mml:mrow>
<mml:msubsup>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mn>0</mml:mn>
</mml:msubsup>
<mml:mo>&#x223C;</mml:mo>
<mml:mrow>
<mml:mi>U</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(25)</label></disp-formula>
<disp-formula id="S2.E26">
<mml:math id="M26">
<mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo rspace="5.8pt">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>Y</mml:mi>
<mml:mo stretchy="false">^</mml:mo>
</mml:mover>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>Y</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(26)</label></disp-formula>
<p><inline-formula><mml:math id="INEQ29"><mml:mrow><mml:mover accent="true"><mml:mi>Y</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover><mml:mo>&#x2062;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi></mml:msubsup><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:math></inline-formula> represents the forecasted segmentation resulting from training the model with hyperparameters <inline-formula><mml:math id="INEQ30"><mml:msubsup><mml:mi>H</mml:mi><mml:mi>i</mml:mi><mml:mi>t</mml:mi></mml:msubsup></mml:math></inline-formula>. This enhances segmentation accuracy during optimization by ensuring that candidates that have high segmentation accuracy will be selected (<xref ref-type="bibr" rid="B26">26</xref>). PBFO alternates between a global roaming phase of exploration and a local hunting phase of exploitation. The exploration stage gives polar bears the chance to search through large portions of hyperparameter space to minimize the risk of premature convergence. The update of candidate solutions is defined as follows, as shown in <xref ref-type="disp-formula" rid="S2.E27">Equation 27</xref>.</p>
<disp-formula id="S2.E27">
<mml:math id="M27">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B1;</mml:mi>
<mml:mo>&#x22C5;</mml:mo>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x22C5;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>-</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B2;</mml:mi>
<mml:mo>&#x22C5;</mml:mo>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x22C5;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>j</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>-</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>k</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(27)</label></disp-formula>
<p>Here, <inline-formula><mml:math id="I1"><mml:mrow><mml:msubsup><mml:mi>H</mml:mi><mml:mrow><mml:mi>b</mml:mi><mml:mi>e</mml:mi><mml:mi>x</mml:mi><mml:mi>t</mml:mi></mml:mrow><mml:mi>t</mml:mi></mml:msubsup></mml:mrow></mml:math></inline-formula> refers to the best solution discovered thus far, and <inline-formula><mml:math id="INEQ31"><mml:msubsup><mml:mi>H</mml:mi><mml:mi>j</mml:mi><mml:mi>t</mml:mi></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="INEQ32"><mml:msubsup><mml:mi>H</mml:mi><mml:mi>k</mml:mi><mml:mi>t</mml:mi></mml:msubsup></mml:math></inline-formula> refer to two randomly chosen distinct solutions. Vectors of scale, <italic>R</italic><sub>1</sub>,<italic>R</italic><sub>2</sub><italic>U</italic>(0,1), are sampled from a Uniform (0,1) distribution, with parameters &#x03B1;,&#x03B2; controlling the degree of exploration. The definition emphasizes the counterbalance between convergence to the best solution and the diversity created due to interaction between individuals in the population. When an interesting area is located, PBFO proceeds toward exploitation, imitating local hunting and improving candidate solutions about present best hyperparameters (<xref ref-type="bibr" rid="B28">28</xref>), controlled by the parameter &#x03B3; &#x2208; (0,1), which denotes a fine degree of adjustment on local search. These three parameters, &#x03B1;,&#x03B2;, and &#x03B3;, dynamically balance exploration and exploitation over iterations; exploration is prominent during the initial iterations (with &#x03B1;,&#x03B2; large), and, as the later iterations proceed, exploitation begins to dominate (with &#x03B3; increasing), thus preventing premature convergence and enhancing the chances to find globally optimal hyperparameter. The optimization terminates when the number of iterations reaches <italic>T</italic><sub><italic>max</italic></sub> or when improvement in fitness becomes lower than some threshold &#x2208;, and the last optimal hyperparameter configuration is given by, as shown in <xref ref-type="disp-formula" rid="S2.E28">Equations 28</xref>&#x2013;<xref ref-type="disp-formula" rid="S2.E30">30</xref>.</p>
<disp-formula id="S2.E28">
<mml:math id="M28">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x03B3;</mml:mi>
<mml:mo>&#x22C5;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>-</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(28)</label></disp-formula>
<disp-formula id="S2.E29">
<mml:math id="M29">
<mml:mrow>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>+</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>-</mml:mo>
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msubsup>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo rspace="5.8pt">|</mml:mo>
</mml:mrow>
<mml:mo rspace="5.8pt">&lt;</mml:mo>
<mml:mo>&#x2208;</mml:mo>
</mml:mrow>
</mml:math>
<label>(29)</label></disp-formula>
<disp-formula id="S2.E30">
<mml:math id="M30">
<mml:mrow>
<mml:mpadded width="+3.3pt">
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mpadded>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>g</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>m</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+5pt">
<mml:mi>x</mml:mi>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:msub>
<mml:mi>H</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(30)</label></disp-formula>
<p>With the help of PBFO demonstrated in <xref ref-type="table" rid="T1">Table 1</xref>, PoSAM-ULTRA meaningfully searches the hyperparameter space, speeds up model training, decreases computation cost, and improves the segmentation results on lower-grade glioma MRI datasets. This principled strategy assures the model will achieve robust learning and a high level of accuracy without any manual trial-and-error.</p>
<table-wrap position="float" id="T1">
<label>TABLE 1</label>
<caption><p>Hyperparameter search space for PBFO-based model optimization.</p></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="center">Hyperparameter</th>
<th valign="top" align="center">Search range</th>
<th valign="top" align="center">Description</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">Learning rate</td>
<td valign="top" align="center">[110<sup>&#x2212;5</sup>,110<sup>&#x2212;2</sup>]</td>
<td valign="top" align="center">Controls gradient descent step size; critical for stable convergence.</td>
</tr>
<tr>
<td valign="top" align="center">Batch size</td>
<td valign="top" align="center">8, 16, 32, 64</td>
<td valign="top" align="center">The number of samples per training iteration affects memory and gradient stability.</td>
</tr>
<tr>
<td valign="top" align="center">Weight decay (L2 reg.)</td>
<td valign="top" align="center">[110<sup>&#x2212;6</sup>,110<sup>&#x2212;2</sup>]</td>
<td valign="top" align="center">Acts against large weight values that promote overfitting.</td>
</tr>
<tr>
<td valign="top" align="center">Dropout rate</td>
<td valign="top" align="center">[0.1, 0.5]</td>
<td valign="top" align="center">The probability that a neuron is randomly dropped to help with generalization.</td>
</tr>
<tr>
<td valign="top" align="center">Number of filters</td>
<td valign="top" align="center">32, 64, 128, 256</td>
<td valign="top" align="center">Initial channel depth-control network capacity.</td>
</tr>
<tr>
<td valign="top" align="center">Optimizer type</td>
<td valign="top" align="center">Adam, SGD, RMSprop</td>
<td valign="top" align="center">Train optimizers are chosen depending on stability and convergence.</td>
</tr>
<tr>
<td valign="top" align="center">Loss function weights</td>
<td valign="top" align="center">[0.1, 0.9]</td>
<td valign="top" align="center">Weights balancing the contributions of Dice, cross, and boundary loss.</td>
</tr>
<tr>
<td valign="top" align="center">Learning rate decay step</td>
<td valign="top" align="center">5, 10, 20</td>
<td valign="top" align="center">The frequency at which the learning rate schedules update.</td>
</tr>
<tr>
<td valign="top" align="center">Momentum (for SGD)</td>
<td valign="top" align="center">[0.8, 0.99]</td>
<td valign="top" align="center">Controls the smoothing of updates in stochastic gradient descent.</td>
</tr>
</tbody>
</table></table-wrap>
</sec>
<sec id="S2.SS5">
<label>2.5</label>
<title>Baseline models</title>
<p>To put PoSAM-ULTRA to a thorough test, we have run comparative experiments against three widely used brain tumor segmentation models: U-Net, U-Net++, and nnU-Net-lite. It is worth mentioning that these are all baseline models, and they have had their hyperparameters tuned by the Polar-Bear Foraging Optimization (PBFO) algorithm. Hence, while they served as competing baseline methods, all their performances were evaluated under the best parameter settings, making the final comparison fair and unbiased.</p>
<p>U-Net is a classical encoder-decoder network with symmetric skip connections (<xref ref-type="bibr" rid="B29">29</xref>). While it allows efficient segmentation, it simply lacks any attention mechanism and multi-modal feature integration, rendering it incapable of capturing refined tumor boundaries.</p>
<p>U-Net++ suggests that nested skip connections and dense skip connections improve gradient flow and multi-scale feature fusion (<xref ref-type="bibr" rid="B30">30</xref>). However, since the architecture &#x201C;squeeze&#x201D; is reliant on nested skip connections, it is possible that this architecture does not emphasize areas related to tumor-specific regions in lower-quality MRIs.</p>
<p>nnU-Net-lite is an adaptive lighter version of nnU-Net that automatically adjusts the layer depth of the architecture to meet the dataset&#x2019;s requirements (<xref ref-type="bibr" rid="B31">31</xref>). With PBFO based hyperparameter optimization, the nnU-Net achieves a solid baseline performance, but the model&#x2019;s architecture will prefer low computation given the shallowness of the network and the attention implemented within the model may not fully resolve lower contrast sections of the MRIs that require high computation use.</p>
</sec>
<sec id="S2.SS6">
<label>2.6</label>
<title>Model training and evaluation</title>
<p>PoSAM-ULTRA and the baseline networks (U-Net, U-Net++, and nnU-Net-lite) were trained on LGG, with hyperparameters optimized via the dynamic Polar-Bear Foraging Optimization (PBFO) technique. PBFO instantiates the best learning rate, batch size, optimizer type, and loss weighting to allow for fair and correct comparisons between models. The dataset was split into 80% training, 10% validation, and 10% test datasets. Preprocessing included resizing and normalization in addition to data augmentation (rotation, flipping, scaling, intensity variation) to increase robustness and minimize overfitting. At training time, PBFO manages the optimization of the weighting between Dice loss and binary cross-entropy to handle the class imbalance of tumor and non-tumor regions. It propagates early stopping and checkpoint saving for the best validation performance. For each evaluation metric, values were computed on the test set to measure segmentation and tumor localization performance. Dice Coefficient measures the overlap between the regions of tumors predicted and the ground truth. Very high values of Dice represent very accurate segmentations of tumor boundaries, which is important for accurate localization (<xref ref-type="bibr" rid="B32">32</xref>). Intersection over Union quantifies the measure of intersection and union of predicted and true tumor regions (<xref ref-type="bibr" rid="B33">33</xref>). This complements Dice by giving a negative score if the prediction over-segments the tumor; thus, it is useful to assess if the segmentation covers the tumor very precisely, as shown in <xref ref-type="disp-formula" rid="S2.E31">Equations 31</xref>, <xref ref-type="disp-formula" rid="S2.E32">32</xref>.</p>
<disp-formula id="S2.E31">
<mml:math id="M31">
<mml:mrow>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+3.3pt">
<mml:mi>e</mml:mi>
</mml:mpadded>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2229;</mml:mo>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi>P</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mi>G</mml:mi>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(31)</label></disp-formula>
<disp-formula id="S2.E32">
<mml:math id="M32">
<mml:mrow>
<mml:mrow>
<mml:mi>I</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+3.3pt">
<mml:mi>U</mml:mi>
</mml:mpadded>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2229;</mml:mo>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x222A;</mml:mo>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">|</mml:mo>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(32)</label></disp-formula>
<p>Accuracy measures the percentage of total pixels correctly classified throughout the entire image and therefore provides insight into a model&#x2019;s intrinsic accuracy in distinguishing tumors from non-tumor tissues. Precision is the percentage of pixels predicted as tumors that are tumors. When precision is very high there will be a comparatively low number of false positives, meaning a significantly lower overestimation of tumor volume. Recall measures the percentage of true tumor pixels that are classified positively. High recall is especially important for detecting small tumor regions during early diagnosis. Having recoveries with an inverse relation to false positives, the F1-score finds the harmonic means of Precision and Recall, resulting in a single measure of segmentation accuracy. This measure is pertinent when tumor regions are small or imbalanced against the background. An F1, therefore, means that the model detects tumor pixels well with comparatively fewer missed tumor regions, as shown in <xref ref-type="disp-formula" rid="S2.E33">Equations 33</xref>&#x2013;<xref ref-type="disp-formula" rid="S2.E36">36</xref>.</p>
<disp-formula id="S2.E33">
<mml:math id="M33">
<mml:mrow>
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>u</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+3.3pt">
<mml:mi>y</mml:mi>
</mml:mpadded>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mpadded width="+5pt">
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mpadded>
</mml:mrow>
</mml:math>
<label>(33)</label></disp-formula>
<disp-formula id="S2.E34">
<mml:math id="M34">
<mml:mrow>
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>r</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>o</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+3.3pt">
<mml:mi>n</mml:mi>
</mml:mpadded>
</mml:mrow>
<mml:mo rspace="5.8pt">=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(34)</label></disp-formula>
<disp-formula id="S2.E35">
<mml:math id="M35">
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>l</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo rspace="9.1pt" stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x2062;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(35)</label></disp-formula>
<disp-formula id="S2.E36">
<mml:math id="M36">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>p</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>e</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>c</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mpadded width="+6.6pt">
<mml:mi>y</mml:mi>
</mml:mpadded>
<mml:mo>&#x2062;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo>+</mml:mo>
<mml:mrow>
<mml:mi>F</mml:mi>
<mml:mo>&#x2062;</mml:mo>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(36)</label></disp-formula>
<p>By optimizing all training hyperparameters via PBFO and splitting the data into training, validation, and test sets, PoSAM-ULTRA achieves robust tumor detection, precise boundary segmentation, and faster convergence, outperforming baseline models on all evaluation metrics.</p>
</sec>
</sec>
<sec id="S3" sec-type="results">
<label>3</label>
<title>Results</title>
<p>The proposed PoSAM-ULTRA performance model was evaluated on the LGG segmentation dataset. The implementation was performed using the following libraries: OpenCV, Pillow, NumPy, Matplotlib, and TensorBoard for visualization. The entire experiment was run using Visual Studio Code (VS Code) running on Windows 11 Pro. A powerful machine with an Intel Core i7-12700K processor, 16 GB of RAM, and an NVIDIA RTX 4060 Ti graphics card was used to train and evaluate the model, providing sufficient computational resources to quickly and efficiently develop and test the model. A qualitative and quantitative evaluation of the proposed method, as well as a comprehensive evaluation of its modules, is presented here.</p>
<p>As seen in <xref ref-type="table" rid="T2">Table 2</xref>, the experimental results indicate that the proposed PoSAM-ULTRA model performs better than others, reaching 91.4% on the Dice, 88.9% on the IoU, and 99.8% on the accuracy, 95.2% on precision, 93.3% on recall, and 99.9% on specificity. In fact, the comparative analysis shows PoSAM-ULTRA making considerable advances over the baseline models, with the mentioned scores being the highest for the Dice, IoU, and accuracy metrics. Moreover, the results obtained are far better than those produced by the changes made in the UNet architecture, which obtained 83.5% Dice, 80.5% IoU, 99.6% accuracy, 89.7% precision, 88.4% recall, and 99.8% specificity, respectively.</p>
<table-wrap position="float" id="T2">
<label>TABLE 2</label>
<caption><p>Model performance analysis using the LGG segmentation dataset.</p></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="center">Model</th>
<th valign="top" align="center">Dice (DSC)</th>
<th valign="top" align="center">IoU</th>
<th valign="top" align="center">Accuracy</th>
<th valign="top" align="center">Precision</th>
<th valign="top" align="center">Recall (sensitivity)</th>
<th valign="top" align="center">Specificity</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">PoSAM-ULTRA</td>
<td valign="top" align="center">91.4%</td>
<td valign="top" align="center">88.9%</td>
<td valign="top" align="center">99.8%</td>
<td valign="top" align="center">95.2%</td>
<td valign="top" align="center">93.3%</td>
<td valign="top" align="center">99.9%</td>
</tr>
<tr>
<td valign="top" align="center">UNet</td>
<td valign="top" align="center">83.5%</td>
<td valign="top" align="center">80.5%</td>
<td valign="top" align="center">99.6%</td>
<td valign="top" align="center">89.7%</td>
<td valign="top" align="center">88.4%</td>
<td valign="top" align="center">99.8%</td>
</tr>
<tr>
<td valign="top" align="center">UNet++-light</td>
<td valign="top" align="center">82.1%</td>
<td valign="top" align="center">79.6%</td>
<td valign="top" align="center">99.5%</td>
<td valign="top" align="center">94.3%</td>
<td valign="top" align="center">83.7%</td>
<td valign="top" align="center">99.8%</td>
</tr>
<tr>
<td valign="top" align="center">nnUNet-lite</td>
<td valign="top" align="center">84.8%</td>
<td valign="top" align="center">81.7%</td>
<td valign="top" align="center">99.6%</td>
<td valign="top" align="center">93.7%</td>
<td valign="top" align="center">87.1%</td>
<td valign="top" align="center">99.8%</td>
</tr>
</tbody>
</table></table-wrap>
<p>PoSAM-ULTRA has an edge over the more sophisticated architectures. The UNet++-light, in spite of its high connectivity patterns, got only 82.1% Dice and 79.6% IoU which are considerable improvements of 9.3 and 9.3%, respectively when compared to PoSAM-ULTRA. Likewise, the nnUNet-lite model, acclaimed for its automatic preprocessing and augmentation techniques, managed to get 84.8% Dice and 81.7% IoU, still below the proposed method&#x2019;s performance. The comparison underscores the power of the combination of the Polar Bear Foraging Optimization (PBFO) algorithm and the Segment Anything Model (SAM) foundation. The Dice coefficient, an indicator of the overlap between predicted and actual segmentations, provides PoSAM-ULTRA with first place with a score of 91.4%, which is very high accuracy value for delineating glioma areas or regions of tumor. The IoU metric supports this with a score of 88.9%, which means predicted and actual tumor borders were/are extremely well spatially bandied. The precision score of 95.2% means the model nearly never makes false positives, which is extremely important in the medical realm as over-segmentation may lead to surgery otherwise deemed unnecessary. The recall number of 93.3% concludes that the model is very good at locating almost all the tumor areas and, thus, it does not add to the problem of under-diagnosis.</p>
<p>PoSAM-ULTRA achieved a pixel-wise accuracy of 0.9979. In terms of boundary-based metrics, the model obtained an HD95 of 4.28, ASSD of 1.28, and MSD of 0.008. The structural similarity between predicted and ground-truth masks was also high, with SSIM = 0.923.</p>
<p>The overall training, which can be seen in <xref ref-type="fig" rid="F5">Figure 5</xref>, points out that the suggested PoSAM-ULTRA has quick convergence, high stability, and better generalization than the other architectures. These characteristics indeed justify the advantages brought by incorporating the PBFO algorithm into the SAM foundation, as reflected in the consistent leading position of all key indicators. It should be noted that the &#x201C;Test&#x201D; label marked at the end of each curve indicates the model&#x2019;s final evaluation on the held-out test set after running 50 epochs. This serves to visually show the final performance on unseen data and hence verifies that the various reported quantitative metrics were from independent test samples and not subjects involved in either training or validation subsets. According to <xref ref-type="fig" rid="F5">Figure 5</xref>, PoSAM-ULTRA exhibits a very smooth and fast convergence curve, achieving a Dice score higher than 0.90 within the first 10 epochs and reaching up to 0.95 in the final epochs. Convergence of the baseline models is comparatively slower with some noticeable fluctuations in both the training and validation phases. A narrow gap in the PoSAM-ULTRA training and validation curves testifies to its good generalization ability, with resistance to overfitting.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p>Comparative training and validation curves of the Dice of PoSAM-ULTRA and baseline models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g005.tif">
<alt-text content-type="machine-generated">Four line charts compare training and validation Dice scores for PoSAM-ULTRA, UNet, UNet++-Light, and nnUNet-lite models over fifty epochs. PoSAM-ULTRA achieves the highest and most stable scores, while UNet++-Light and nnUNet-lite show greater fluctuations and lower validation scores. All charts display blue lines for training and orange lines for validation, with steady improvements but visible gaps between training and validation performances.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F6">Figure 6</xref> shows the IoU metric with a similar trend; PoSAM-ULTRA also improves consistently and smoothly in the overall trend, up to a plateau of about 0.89, outperforming UNet and nnUNet-lite, which settle down at below 0.82. This result indicates that the proposed model provides more exact and spatially coherent boundaries for segmentation of glioma regions due to superior optimization of feature representations.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>Comparative training and validation curves of the IoU of PoSAM-ULTRA and baseline models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g006.tif">
<alt-text content-type="machine-generated">Four line charts compare intersection over union (IoU) for train and validation data across epochs for PoSAM-ULTRA, UNet, UNet++-Light, and nnUNet-lite models. PoSAM-ULTRA shows the highest IoU stability and scores. Each chart tracks IoU as epochs progress from zero to fifty, with both train and validation lines labeled.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F7">Figure 7</xref> further supports this observation by illustrating that PoSAM-ULTRA maintains perfect stability, with accuracy values &#x003E; 0.997 throughout training. Highly fitted alignment between training and validation curves underlines the well-balanced learning dynamics and the robustness of network performance across different partitions of data.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption><p>Comparative training and validation curves of the accuracy of PoSAM-ULTRA and baseline models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g007.tif">
<alt-text content-type="machine-generated">Four line charts compare accuracy over fifty epochs for PoSAM-ULTRA, UNet, UNet++-Light, and nnUNet-lite models. Each chart displays training and validation accuracy curves, with PoSAM-ULTRA achieving the highest and most stable accuracy throughout.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F8">Figure 8</xref> contains a visual assessment that is very detailed and indicates segmentation performance in four representative clinical cases from the Lower-Grade Gliomas dataset. Every line reveals a complete case featuring the original MRI scan, ground truth tumor annotation, and segmentation outputs from PoSAM-ULTRA together with three baseline models (UNet, UNet++-light, and nnUNet-lite). The tumor regions are marked in green and quantitative metrics show the segmented tumor area in pixels as well as the corresponding Dice coefficient for each model. This visualization allows for direct evaluation of segmentation quality, boundary accuracy, and each model&#x2019;s capability to cope with different tumor characteristics, such as size, location, and morphological complexity.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption><p>Qualitative comparison of segmentation results on lower-grade glioma cases.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g008.tif">
<alt-text content-type="machine-generated">Four rows, each representing a brain MRI scan case, are shown with six columns: the original image, ground truth tumor segmentation, and results from four different segmentation models. Each column displays segmented tumor regions in green, with associated area and Dice score values.</alt-text>
</graphic>
</fig>
<p>Over the course of the testing, PoSAM-ULTRA demonstrated its excellent segmentation quality in all the cases presented with Dice scores between 0.897 and 0.964 and the estimate of the tumorous area being very close to the actual one. The segmentation masks were not only giving the accurate boundary but also they had smooth contours and at the same time they kept being reliable regardless of the tumor&#x2019;s size or location. On the other hand, the baseline models suffer from different limitations, which are, among others, over-segmentation (UNet), under-segmentation (nnUNet-lite), and inconsistent performance with occasional complete failure in detecting tumors (UNet++-light showing 0.000 Dice in Case 3). These qualitative results are in line with the quantitative findings and thus, PoSAM-ULTRA&#x2019;s practical advantages for clinical applications where accurate tumor delineation is critical for diagnosis and treatment planning have been demonstrated.</p>
<p>In <xref ref-type="fig" rid="F9">Figure 9</xref>, the performance of glioma segmentation done by PoSAM-ULTRA (green masks) and by three baseline models: UNet, UNet++-light, and nnUNet-lite (blue masks) is qualitatively compared in four selected cases of brain MRI. The first column shows the original MRI image, the second column shows the ground truth annotation, and the last columns show the predictions from all the models with their respective isolated tumor masks. A visual check shows that PoSAM-ULTRA results have better boundary precision and morphological accuracy, no matter the size, location, and shape of the tumors, and that their results come very close to the expert annotations in every case. The baselines produce segmentations that are more or less reasonable, but they also differ from each other in boundary sharpness and shape fidelity, especially in the areas where it is difficult because of low contrast or irregular margins. This qualitative analysis confirms the quantitative inferior position of PoSAM-ULTRA (Dice: 91.49%, IoU: 88.94%) showing that the fusion of SAM backbone, attention mechanisms, and PBFO optimization has a direct effect on the clinically noticeable segmentation quality increase for surgical planning, radiotherapy targeting, and treatment monitoring.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption><p>Qualitative comparison of segmentation results across different models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fmed-13-1730353-g009.tif">
<alt-text content-type="machine-generated">Comparison chart displays rows of brain MRI images showing original scans, ground truth tumor annotations, and segmentation predictions from four models: PoSAM-ULTRA, UNet, UNet++-light, and nnUNet-lite. Tumor regions are highlighted in different colors for visual comparison.</alt-text>
</graphic>
</fig>
</sec>
<sec id="S4" sec-type="discussion">
<label>4</label>
<title>Discussion</title>
<sec id="S4.SS1">
<label>4.1</label>
<title>Comparison with other studies</title>
<p>Proposed study (<xref ref-type="bibr" rid="B34">34</xref>) a new network framework called MUNet, which combines the advantages of UNet and Mamba to achieve accurate and efficient segmentation of brain tumors, the results showed that MUNet achieved superior segmentation accuracy and excellent generalization capabilities when validated on TCGA-LGG segmentation datasets, with DSC, IOU, and accuracy reaching 0.702, 0.581, and 0.981, respectively. This paper aims (<xref ref-type="bibr" rid="B35">35</xref>) to study the effectiveness of the Segment Anything Model (SAM) in the context of brain MRI, the researchers used the TCGA-LGG dataset, and the results showed that when they applied the SAM model, they were able to achieve 0.87 higher IOU scores.</p>
<p>Researchers (<xref ref-type="bibr" rid="B36">36</xref>) aimed to improve the segmentation of low-grade gliomas in MRI by combining a 2D U-Net and a WAO-based voting mechanism. The proposed model achieved an accuracy of 0.9978 and an IOU of 0.8573 on the TCGA-LGG dataset. A hybrid deep learning model combining U-Net and SegNet was proposed for automatic segmentation of low-grade gliomas (LGG) from MRI images (<xref ref-type="bibr" rid="B37">37</xref>), and was trained on the TCGA-LGG dataset, where it achieved an average DSC of 83 and 85.7%, outperforming each model individually. Researchers (<xref ref-type="bibr" rid="B38">38</xref>) presented an improved U-Net-based segmentation algorithm to improve brain tumor segmentation from MRI images, and after training it on the TCGA-LGG dataset, the proposed method achieved an accuracy of 99.13%, DSC of 0.4596, and IOU of 0.3132.</p>
<p>A lightweight parallel 3D U-Net (LATUP-Net) was designed for brain tumor segmentation from MRI images, ensuring fast and efficient treatment, their proposed model achieved DSC rates of 88.41, 83.82, and 73.67% for the entire tumor, tumor core, and enhanced tumor on the BraTS 2020 dataset, while these rates on the BraTS 2021 dataset were 90.29, 89.54, and 83.92%, respectively (<xref ref-type="bibr" rid="B39">39</xref>). In this study (<xref ref-type="bibr" rid="B40">40</xref>) the researchers introduced a hybrid model, the 3D ResAttU-Net-Swin, which integrates a residual U-Net, an attention mechanism, and a Swin transformer, the researchers tested their hybrid model on both the BraTS 2020 and BraTS 2019 datasets, the model got a DSC of 88.27% and an IOU of 79.93% on BraTS 2020, and an average DSC of 89.20% and an average IOU of 81.40% on BraTS 2019. In this research paper (<xref ref-type="bibr" rid="B41">41</xref>), the VSA-GCNN-BiGRU model was applied to segment and classify brain tumor MRI images to improve diagnostic accuracy, the researchers relied on the BraTS datasets of 2019, 2020, and 2021, the results showed that their proposed model achieved outstanding performance on the BraTS 2019 dataset, with an accuracy of 99.98%, while it achieved an accuracy of 98.2% and a DSC of 40% on the BraTS 2020 dataset, when the proposed model was evaluated on the BraTS 2021 dataset, it achieved an accuracy of 97.6% and a DSC of 98.6%. <xref ref-type="table" rid="T3">Table 3</xref> shows a comparison with previous studies on brain tumor segmentation.</p>
<table-wrap position="float" id="T3">
<label>TABLE 3</label>
<caption><p>Comparison of the results of the proposed study with previous studies for the brain tumor segmentation.</p></caption>
<table cellspacing="5" cellpadding="5" frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="center">Year</th>
<th valign="top" align="center">References</th>
<th valign="top" align="center">Dataset</th>
<th valign="top" align="center">Methodology</th>
<th valign="top" align="center">Limitations</th>
<th valign="top" align="center">Accuracy</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B34">34</xref>)</td>
<td valign="top" align="center">TCGA-LGG</td>
<td valign="top" align="center">MUNet (U-Net, mamba)</td>
<td valign="top" align="center">Overfitting problem, poor performance in complex areas</td>
<td valign="top" align="center">DSC:70.2%,IOU:58.15, accuracy: 98.1%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B35">35</xref>)</td>
<td valign="top" align="center">TCGA-LGG</td>
<td valign="top" align="center">Segment anything model (SAM)</td>
<td valign="top" align="center">Reliance on the IOU scale alone, lack of clinical assessment</td>
<td valign="top" align="center">IOU: 87%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B36">36</xref>)</td>
<td valign="top" align="center">TCGA-LGG</td>
<td valign="top" align="center">U-Net, WAO-based on refinement</td>
<td valign="top" align="center">The model has not been tested on an external dataset</td>
<td valign="top" align="center">IOU: 85.73%<break/> accuracy: 99.78%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B38">38</xref>)</td>
<td valign="top" align="center">TCGA-LGG</td>
<td valign="top" align="center">U-Net</td>
<td valign="top" align="center">Reliance on one type of MRI data</td>
<td valign="top" align="center">DSC:45.96%, IOU:31.32%,<break/> accuracy: 99.13%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B39">39</xref>)</td>
<td valign="top" align="center">BraTS 2020, BraTS 2021</td>
<td valign="top" align="center">LATUP-Net</td>
<td valign="top" align="center">Limited dataset size (approximately 1,200 patients), lack of computing resources</td>
<td valign="top" align="center">BraTS 2020: : DSC<break/> 88.41, 83.82, 73.67%.<break/> BraTS 2021:<break/> DSC<break/> 90.29, 89.54, 83.92%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B40">40</xref>)</td>
<td valign="top" align="center">BraTS2019<break/> BraTS2020</td>
<td valign="top" align="center">(3D) ResAttU-Net-Swin</td>
<td valign="top" align="center">Poor performance in the ET (Enhancing Tumor) region, increased model complexity</td>
<td valign="top" align="center">BraTS 2020: DSC 88.27%,<break/> IOU 79.93%;<break/> BraTS 2019: DSC 89.20%,<break/> IOU : 81.40%</td>
</tr>
<tr>
<td valign="top" align="center">2025</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B41">41</xref>)</td>
<td valign="top" align="center">BraTS2019<break/> BraTS2020</td>
<td valign="top" align="center">VSA-GCNN-BiGRU</td>
<td valign="top" align="center">The need for big data, long training period</td>
<td valign="top" align="center">BraTS 2020: DSC : 40%,<break/> accuracy 98.2%;<break/> BraTS 2021: DSC :98.6%,<break/> accuracy : 97.6%</td>
</tr>
<tr>
<td valign="top" align="center">2023</td>
<td valign="top" align="center">(<xref ref-type="bibr" rid="B37">37</xref>)</td>
<td valign="top" align="center">TCGA-LGG</td>
<td valign="top" align="center">Hybrid U-Net, SegNet</td>
<td valign="top" align="center">Lack of data and poor generalization</td>
<td valign="top" align="center">DSC: 83, 85.7%</td>
</tr>
</tbody>
</table></table-wrap>
<p>The three studies collectively highlight recent advances in deep-learning&#x2013;based medical image analysis, particularly for improving brain MRI segmentation and classification. The first study (<xref ref-type="bibr" rid="B42">42</xref>) introduces a dual-channel encoder&#x2013;decoder network enhanced with compound attention mechanisms to more accurately segment Multiple Sclerosis (MS) lesions, achieving a notably high Dice score (0.73) and outperforming prior approaches. The second work (<xref ref-type="bibr" rid="B43">43</xref>) focuses on brain-tumor classification, leveraging transfer learning with augmented MRI data and demonstrating how pretrained models can significantly boost diagnostic performance when datasets are limited. Meanwhile, the third study (<xref ref-type="bibr" rid="B44">44</xref>) presents a supervised segmentation method designed to identify salient brain tissues in MRI images, forming an essential foundation for automated tissue analysis and downstream clinical tasks. The optimized U-Net presented in a recent study (<xref ref-type="bibr" rid="B45">45</xref>) achieved remarkably high performance (Dice 92.54%, IoU 90.42%) while maintaining rapid inference time, demonstrating its suitability for real-time clinical applications. Including this reference provides valuable context by highlighting complementary efficiency-focused advancements in deep learning&#x2013;based medical image segmentation. Together, these studies illustrate a clear progression from precise lesion segmentation to robust tumor classification, to general tissue-segmentation frameworks demonstrating how modern deep learning models continue to enhance diagnostic accuracy and support real-world clinical decision-making.</p>
<p>As illustrated in <xref ref-type="table" rid="T3">Table 3</xref>, several recent studies on glioma segmentation continue to face recurring limitations that affect their clinical reliability and generalization ability. Many U-Net&#x2013;based or transformer-enhanced architectures, such as MUNet (<xref ref-type="bibr" rid="B34">34</xref>) and the WAO-refined U-Net (<xref ref-type="bibr" rid="B36">36</xref>), exhibit overfitting and poor performance in complex or low-contrast tumor regions, particularly when trained on limited single-source datasets. Other methods, including SAM-based segmentation (<xref ref-type="bibr" rid="B35">35</xref>), rely heavily on specific evaluation scales and often lack clinical validation or multi-perspective assessment. Several works also suffer from restricted modality usage, as seen in Obayya et al. (<xref ref-type="bibr" rid="B38">38</xref>), or do not perform external dataset testing, which limits their generalization capability.</p>
<p>PoSAM-ULTRA is designed to directly tackle these limitations. First, the integration of CBAM, Attention Gates, and multi-scale DownBlocks strengthens tumor-boundary awareness, addressing the poor delineation issues observed in prior CNN-only architectures. Second, the PBFO-driven hyperparameter optimization reduces the tendency toward overfitting by systematically exploring configurations that favor stable, generalizable convergence&#x2014;unlike models that rely on static or manually tuned settings. Third, the use of a four-channel input, which incorporates prior tumor-likelihood information, enhances robustness across scans with varying intensity distributions, a limitation noted in studies that depend on single MRI modalities.</p>
<p>While we did not conduct an additional external test run in this revision, the comparative evidence in <xref ref-type="table" rid="T3">Table 3</xref> demonstrates strong relative generalization: PoSAM-ULTRA surpasses methods known to struggle with dataset shift and maintains performance levels closer to architectures evaluated on heterogeneous clinical datasets (e.g., BraTS 2019&#x2013;2021). This analysis supports the conclusion that PoSAM-ULTRA not only addresses key limitations in the literature but also exhibits improved resilience when benchmarked against a broad spectrum of contemporary segmentation techniques.</p>
</sec>
<sec id="S4.SS2">
<label>4.2</label>
<title>Computational efficiency and practical considerations</title>
<p>Addressing the need for practical clinical deployment, the PoSAM-ULTRA framework was designed to achieve a favorable balance between high segmentation accuracy and computational efficiency. Our architectural choice of a ResNet-34 encoder offers a strong but relatively lightweight backbone, keeping the overall parameter count manageable. Furthermore, the specialized components CBAM and Attention Gates are implemented to strategically enhance feature discriminative power and optimize information flow across skip connections, which ultimately leads to performance gains without proportional overhead in model size. Critically, the Deep Supervision strategy, integral to stabilizing the training process and accelerating convergence, is disabled entirely during the inference phase. This ensures that PoSAM-ULTRA maintains a high segmentation speed, making it suitable for clinical environments where timely processing is essential, thereby demonstrating a strong commitment to balancing high clinical performance with operational resource efficiency.</p>
</sec>
<sec id="S4.SS3">
<label>4.3</label>
<title>Integration of the proposed method into the hospital</title>
<p>The clinical application of PoSAM-ULTRA entails the usage of DICOM protocol support for the integration with hospital PACS systems, thus allowing the automated retrieval and processing of multi-sequence MRI data on GPU-enabled servers. The suggested workflow would be a combination of both automated and manual, wherein, within 1&#x2013;2 min per case, the system would create preliminary segmentation masks for the radiologist to review and refine through an interactive interface. After the installation, validation on institution-specific data (200&#x2013;300 cases) is most important to determine the performance across local MRI protocols, along with regulatory compliance (FDA/CE marking), patient consent protocols, and comprehensive staff training. Monitoring segmentation quality and user feedback indisputably assures continuous clinical utility, thereby presenting PoSAM-ULTRA as a smart assistant that not only improves upon diagnostic accuracy but also lightens the radiologist&#x2019;s workload.</p>
<p>Although PoSAM-ULTRA achieves strong quantitative results, qualitative error inspection remains a crucial component of its clinical usability. During internal validation, the model demonstrated occasional under-segmentation in cases where tumor margins blended gradually with surrounding tissue, making the boundary inherently ambiguous. Similarly, mild over-segmentation was observed in regions exhibiting edema-like textures that share intensity patterns with tumor tissue. Motion-corrupted scans also led to small, localized discrepancies due to artifact-induced distortions in structural information.</p>
<p>These patterns are consistent with the known challenges of glioma segmentation reported in clinical imaging literature. When integrated into the hospital workflow, such errors are efficiently managed through the radiologist-in-the-loop design, where the system generates preliminary masks and the specialist performs rapid verification and refinement. This approach preserves clinical safety while still providing a significant reduction in manual annotation time. Continuous monitoring of these edge cases further supports iterative improvement of the model across diverse imaging protocols.</p>
</sec>
<sec id="S4.SS4">
<label>4.4</label>
<title>Limitations of this study and further work</title>
<p>There are numerous limitations to this study, and they are examiners for future research. The LGG dataset, albeit helpful for the measurement of performance, might not capture the entire variety of the higher-grade tumors and their different imaging protocols in various institutions, thus validation on bigger datasets like BraTS is needed. The present attention on whole tumor segmentation needs to be increased to sub-region delineation, which can contribute to better radiotherapy and surgical planning. Computational power may prove to be a hurdle in the case of resource-limited areas, thus indicating the need for model compression techniques. The model&#x2019;s low interpretability points out the necessity of merging explainable AI methods and uncertainty quantification to gain clinical trust. Future research will have to include temporal modeling for the longitudinal assessment, multi-modal architectures that combine clinical data (molecular biomarkers, patient history), and extended PBFO optimization for neural architecture search and meta-learning to allow quick adaptation to different healthcare settings.</p>
</sec>
</sec>
<sec id="S5" sec-type="conclusion">
<label>5</label>
<title>Conclusion</title>
<p>This study presents a novel deep learning framework for glioma MRI image segmentation. PoSAM-ULTRA, a novel deep learning framework for glioma MRI segmentation, is presented. It combines the Segment Anything model with Polar-Bear foraging optimization to fine-tune hyperparameters. The proposed architecture uses a modified ResNet-34 encoder with four-channel input processing, multi-scale feature extraction, and attention mechanisms such as CBAM and attention gates to enhance discriminative ability. In an evaluation on the LGG dataset, PoSAM-ULTRA demonstrates better performance than previously defined baselines (UNet, UNet++, and nnUNet), achieving a Dice coefficient of 91.4%, an IoU of 88.9%, a precision of 95.2%, and a recall of 93.3%. These results demonstrate the success of advanced segmentation architectures, coupled with optimization algorithms, in medical image analysis. The integration of deep supervision, dropout regularization, and complex loss functions improves training stability and generalizability across various tumor types. While the study acknowledges limitations imposed by diverse datasets, computational requirements, and the necessity of subregion segmentation, the proposed framework still lays a solid foundation for clinical deployment with proper validation and infrastructure integration.</p>
</sec>
</body>
<back>
<sec id="S6" sec-type="data-availability">
<title>Data availability statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found at: <ext-link ext-link-type="uri" xlink:href="https://www.kaggle.com/datasets/mateuszbuda/lgg-mri-segmentation">https://www.kaggle.com/datasets/mateuszbuda/lgg-mri-segmentation</ext-link>.</p>
</sec>
<sec id="S7" sec-type="author-contributions">
<title>Author contributions</title>
<p>SAlh: Methodology, Resources, Software, Writing &#x2013; original draft. HA: Supervision, Validation, Visualization, Writing &#x2013; original draft. MA: Project administration, Software, Supervision, Writing &#x2013; original draft. AAli: Formal analysis, Investigation, Software, Visualization, Writing &#x2013; review &#x0026; editing. ML: Conceptualization, Data curation, Visualization, Writing &#x2013; review &#x0026; editing. QM: Data curation, Investigation, Validation, Writing &#x2013; review &#x0026; editing. GK: Formal analysis, Project administration, Supervision, Writing &#x2013; original draft. AAlr: Data curation, Formal analysis, Resources, Writing &#x2013; review &#x0026; editing. SAla: Conceptualization, Project administration, Supervision, Writing &#x2013; review &#x0026; editing. KS: Funding acquisition, Project administration, Resources, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number: IMSIU-DDRSP2601).</p>
</ack>
<sec id="S9" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="S10" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="S11" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Neff</surname> <given-names>C</given-names></name> <name><surname>Price</surname> <given-names>M</given-names></name> <name><surname>Cioffi</surname> <given-names>G</given-names></name> <name><surname>Kruchko</surname> <given-names>C</given-names></name> <name><surname>Waite</surname> <given-names>K</given-names></name> <name><surname>Barnholtz-Sloan</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Complete prevalence of primary malignant and nonmalignant brain tumors in comparison to other cancers in the United States.</article-title> <source><italic>Cancer</italic>.</source> (<year>2023</year>) <volume>129</volume>:<fpage>2514</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1002/cncr.34837</pub-id> <pub-id pub-id-type="pmid">37199898</pub-id></mixed-citation></ref>
<ref id="B2">
<label>2.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Tamimi</surname> <given-names>A</given-names></name> <name><surname>Juweid</surname> <given-names>M.</given-names></name></person-group> <source><italic>Epidemiology and Outcome of Glioblastoma.</italic></source> <publisher-loc>Brisbane, AU</publisher-loc>: <publisher-name>Codon Publications</publisher-name> (<year>2017</year>). <pub-id pub-id-type="doi">10.15586/codon.glioblastoma.2017.ch8</pub-id> <pub-id pub-id-type="pmid">29251870</pub-id></mixed-citation></ref>
<ref id="B3">
<label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Picca</surname> <given-names>A</given-names></name> <name><surname>Berzero</surname> <given-names>G</given-names></name> <name><surname>Sanson</surname> <given-names>M</given-names></name></person-group>. <article-title>Current therapeutic approaches to diffuse grade II and III gliomas.</article-title> <source><italic>Ther Adv Neurol Disord.</italic></source> (<year>2018</year>) <volume>11</volume>:<fpage>1756285617752039</fpage>. <pub-id pub-id-type="doi">10.1177/1756285617752039</pub-id> <pub-id pub-id-type="pmid">29403544</pub-id></mixed-citation></ref>
<ref id="B4">
<label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Forst</surname> <given-names>D</given-names></name> <name><surname>Nahed</surname> <given-names>B</given-names></name> <name><surname>Loeffler</surname> <given-names>J</given-names></name> <name><surname>Batchelor</surname> <given-names>T</given-names></name></person-group>. <article-title>Low-grade gliomas.</article-title> <source><italic>Oncologist.</italic></source> (<year>2014</year>) <volume>19</volume>:<fpage>403</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1634/theoncologist.2013-0345</pub-id> <pub-id pub-id-type="pmid">24664484</pub-id></mixed-citation></ref>
<ref id="B5">
<label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Youssef</surname> <given-names>G</given-names></name> <name><surname>Miller</surname> <given-names>J</given-names></name></person-group>. <article-title>Lower grade gliomas.</article-title> <source><italic>Curr Neurol Neurosci Rep.</italic></source> (<year>2020</year>) <volume>20</volume>:<fpage>21</fpage>. <pub-id pub-id-type="doi">10.1007/s11910-020-01040-8</pub-id> <pub-id pub-id-type="pmid">32444979</pub-id></mixed-citation></ref>
<ref id="B6">
<label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mazurowski</surname> <given-names>M</given-names></name> <name><surname>Clark</surname> <given-names>K</given-names></name> <name><surname>Czarnek</surname> <given-names>N</given-names></name> <name><surname>Shamsesfandabadi</surname> <given-names>P</given-names></name> <name><surname>Peters</surname> <given-names>K</given-names></name> <name><surname>Saha</surname> <given-names>A</given-names></name></person-group>. <article-title>Radiogenomics of lower-grade glioma: algorithmically-assessed tumor shape is associated with tumor genomic subtypes and patient outcomes in a multi-institutional study with The cancer genome atlas data.</article-title> <source><italic>J Neurooncol.</italic></source> (<year>2017</year>) <volume>133</volume>:<fpage>27</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1007/s11060-017-2420-1</pub-id> <pub-id pub-id-type="pmid">28470431</pub-id></mixed-citation></ref>
<ref id="B7">
<label>7.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Groza</surname> <given-names>V</given-names></name> <name><surname>Tuchinov</surname> <given-names>B</given-names></name> <name><surname>Amelina</surname> <given-names>E</given-names></name> <name><surname>Pavlovskiy</surname> <given-names>E</given-names></name> <name><surname>Tolstokulakov</surname> <given-names>N</given-names></name> <name><surname>Amelin</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Brain tumor segmentation and associated uncertainty evaluation using multi-sequences MRI mixture data preprocessing.</article-title> In: Crimi A, Bakas S, editors. <source><italic>International MICCAI Brainlesion Workshop.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2020</year>). p. <fpage>148</fpage>&#x2013;<lpage>57</lpage>.</mixed-citation></ref>
<ref id="B8">
<label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fajar</surname> <given-names>A</given-names></name> <name><surname>Sarno</surname> <given-names>R</given-names></name> <name><surname>Fatichah</surname> <given-names>C</given-names></name> <name><surname>Fahmi</surname> <given-names>A</given-names></name></person-group>. <article-title>Reconstructing and resizing 3D images from DICOM files.</article-title> <source><italic>J King Saud Univer Comput Inf Sci</italic>.</source> (<year>2022</year>) <volume>34</volume>:<fpage>3517</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1016/j.jksuci.2020.12.004</pub-id></mixed-citation></ref>
<ref id="B9">
<label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Drozdzal</surname> <given-names>M</given-names></name> <name><surname>Chartrand</surname> <given-names>G</given-names></name> <name><surname>Vorontsov</surname> <given-names>E</given-names></name> <name><surname>Shakeri</surname> <given-names>M</given-names></name> <name><surname>Di Jorio</surname> <given-names>L</given-names></name> <name><surname>Tang</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Learning normalized inputs for iterative estimation in medical image segmentation.</article-title> <source><italic>Med Image Anal.</italic></source> (<year>2018</year>) <volume>44</volume>:<fpage>1</fpage>&#x2013;<lpage>13</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2017.11.005</pub-id> <pub-id pub-id-type="pmid">29169029</pub-id></mixed-citation></ref>
<ref id="B10">
<label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Akram</surname> <given-names>MT</given-names></name> <name><surname>Asghar</surname> <given-names>S</given-names></name> <name><surname>Shahid</surname> <given-names>AR</given-names></name></person-group>. <article-title>Effective data augmentation for brain tumor segmentation.</article-title> <source><italic>Int J Imaging Syst Technol.</italic></source> (<year>2023</year>) <volume>33</volume>:<fpage>1247</fpage>&#x2013;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1002/ima.22853</pub-id></mixed-citation></ref>
<ref id="B11">
<label>11.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>S</given-names></name> <name><surname>Zhang</surname> <given-names>H</given-names></name> <name><surname>Valiant</surname> <given-names>G</given-names></name> <name><surname>R&#x00E9;</surname> <given-names>C</given-names></name></person-group>. <article-title>On the generalization effects of linear transformations in data augmentation.</article-title> in <source><italic>Proceedings of the International Conference on Machine Learning.</italic></source> <publisher-name>PMLR</publisher-name> (<year>2020</year>). p. <fpage>10410</fpage>&#x2013;<lpage>20</lpage>.</mixed-citation></ref>
<ref id="B12">
<label>12.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Kirillov</surname> <given-names>A</given-names></name> <name><surname>Mintun</surname> <given-names>E</given-names></name> <name><surname>Ravi</surname> <given-names>N</given-names></name> <name><surname>Mao</surname> <given-names>H</given-names></name> <name><surname>Rolland</surname> <given-names>C</given-names></name> <name><surname>Gustafson</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Segment anything.</article-title> in<source> <italic>Proceedings of the IEEE/CVF International Conference on Computer Vision.</italic></source> <publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2023</year>). p. <fpage>4015</fpage>&#x2013;<lpage>26</lpage>.</mixed-citation></ref>
<ref id="B13">
<label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Osco</surname> <given-names>LP</given-names></name> <name><surname>Wu</surname> <given-names>Q</given-names></name> <name><surname>de Lemos</surname> <given-names>L</given-names></name> <name><surname>Gon&#x00E7;alves</surname> <given-names>W</given-names></name> <name><surname>Li</surname> <given-names>J</given-names></name> <name><surname>Marcato</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>The segment anything model (sam) for remote sensing applications: from zero to one shot.</article-title> <source><italic>Int J Appl Earth Observ Geoinformation</italic>.</source> (<year>2023</year>) <volume>124</volume>:<fpage>103540</fpage>. <pub-id pub-id-type="doi">10.1016/j.jag.2023.103540</pub-id></mixed-citation></ref>
<ref id="B14">
<label>14.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Deng</surname> <given-names>X</given-names></name> <name><surname>Lu</surname> <given-names>Y</given-names></name></person-group>. <article-title>Segment anything model (sam) for medical image segmentation: A preliminary review.</article-title> in <source><italic>Proceedings of the 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM).</italic></source> <publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2023</year>). p. <fpage>4187</fpage>&#x2013;<lpage>94</lpage>.</mixed-citation></ref>
<ref id="B15">
<label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Q</given-names></name> <name><surname>Guo</surname> <given-names>S</given-names></name> <name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Guo</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Hierarchical channel-spatial encoding for communication-efficient collaborative learning.</article-title> <source><italic>Adv Neural Inf Process Syst</italic>.</source> (<year>2022</year>) <volume>35</volume>:<fpage>5788</fpage>&#x2013;<lpage>801</lpage>.</mixed-citation></ref>
<ref id="B16">
<label>16.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Su</surname> <given-names>Y-C</given-names></name> <name><surname>Liu</surname> <given-names>T-J</given-names></name> <name><surname>Liuy</surname> <given-names>K-H</given-names></name></person-group>. <article-title>Multi-scale wavelet frequency channel attention for remote sensing image segmentation.</article-title> in <source><italic>Proceedings of the 2022 IEEE 14th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP).</italic></source> <publisher-loc>Piscataway, NJ</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2022</year>). p. <fpage>1</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="B17">
<label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lv</surname> <given-names>C</given-names></name> <name><surname>Zhang</surname> <given-names>E</given-names></name> <name><surname>Qi</surname> <given-names>G</given-names></name> <name><surname>Li</surname> <given-names>F</given-names></name> <name><surname>Huo</surname> <given-names>J</given-names></name></person-group>. <article-title>A lightweight parallel attention residual network for tile defect recognition.</article-title> <source><italic>Sci Rep.</italic></source> (<year>2024</year>) <volume>14</volume>:<fpage>21872</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-70570-9</pub-id> <pub-id pub-id-type="pmid">39300076</pub-id></mixed-citation></ref>
<ref id="B18">
<label>18.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Woo</surname> <given-names>S</given-names></name> <name><surname>Park</surname> <given-names>J</given-names></name> <name><surname>Lee</surname> <given-names>J-Y</given-names></name> <name><surname>Kweon</surname> <given-names>IS</given-names></name></person-group>. <article-title>Cbam: Convolutional block attention module.</article-title> in <source><italic>Proceedings of the European Conference on Computer Vision (ECCV).</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>3</fpage>&#x2013;<lpage>19</lpage>.</mixed-citation></ref>
<ref id="B19">
<label>19.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Pang</surname> <given-names>B</given-names></name></person-group>. <article-title>Classification of images using EfficientNet CNN model with convolutional block attention module (CBAM) and spatial group-wise enhance module (SGE).</article-title> in <source><italic>Proceedings of the International Conference on imAge, Signal Processing, and Pattern Recognition (ISPP 2022).</italic></source> <publisher-name>SPIE</publisher-name> (<year>2022</year>). p. <fpage>34</fpage>&#x2013;<lpage>41</lpage>.</mixed-citation></ref>
<ref id="B20">
<label>20.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xiang</surname> <given-names>T</given-names></name> <name><surname>Zhang</surname> <given-names>C</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Song</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>D</given-names></name> <name><surname>Huang</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Towards bi-directional skip connections in encoder-decoder architectures and beyond.</article-title> <source><italic>Med Image Anal.</italic></source> (<year>2022</year>) <volume>78</volume>:<fpage>102420</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2022.102420</pub-id> <pub-id pub-id-type="pmid">35334445</pub-id></mixed-citation></ref>
<ref id="B21">
<label>21.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Couturier</surname> <given-names>R</given-names></name> <name><surname>Perrot</surname> <given-names>G</given-names></name> <name><surname>Salomon</surname> <given-names>M</given-names></name></person-group>. <article-title>Image denoising using a deep encoder-decoder network with skip connections.</article-title> in <source><italic>Proceedings of the International Conference on Neural Information Processing.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>554</fpage>&#x2013;<lpage>565</lpage>.</mixed-citation></ref>
<ref id="B22">
<label>22.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>P</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Chen</surname> <given-names>J</given-names></name></person-group>. <article-title>Enhanced rain removal network with convolutional block attention module (CBAM): a novel approach to image de-raining.</article-title> <source><italic>EURASIP J Adv Signal Process</italic>.</source> (<year>2025</year>) <volume>9</volume>:<fpage>2025</fpage>. <pub-id pub-id-type="doi">10.1186/s13634-025-01212-z</pub-id></mixed-citation></ref>
<ref id="B23">
<label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mao</surname> <given-names>X</given-names></name> <name><surname>Shen</surname> <given-names>C</given-names></name> <name><surname>Yang</surname> <given-names>Y.-B</given-names></name></person-group>. <article-title>Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections.</article-title> <source><italic>Adv Neural Inf Process Syst</italic>.</source> (<year>2016</year>) <volume>29</volume>.</mixed-citation></ref>
<ref id="B24">
<label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>G</given-names></name> <name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Qian</surname> <given-names>J</given-names></name> <name><surname>Zhang</surname> <given-names>J</given-names></name> <name><surname>Yin</surname> <given-names>X</given-names></name> <name><surname>Cui</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>DSEU-net: a novel deep supervision SEU-net for medical ultrasound image segmentation.</article-title> <source><italic>Expert Syst Appl</italic>.</source> (<year>2023</year>) <volume>223</volume>:<fpage>119939</fpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2023.119939</pub-id></mixed-citation></ref>
<ref id="B25">
<label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mishra</surname> <given-names>D</given-names></name> <name><surname>Chaudhury</surname> <given-names>S</given-names></name> <name><surname>Sarkar</surname> <given-names>M</given-names></name> <name><surname>Soin</surname> <given-names>A</given-names></name></person-group>. <article-title>Ultrasound image segmentation: a deeply supervised network with attention to boundaries.</article-title> <source><italic>IEEE Trans Biomed Eng</italic>.</source> (<year>2018</year>) <volume>66</volume>:<fpage>1637</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1109/TBME.2018.2877577</pub-id> <pub-id pub-id-type="pmid">30346279</pub-id></mixed-citation></ref>
<ref id="B26">
<label>26.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Preethi</surname> <given-names>B. M</given-names></name> <name><surname>Lekha</surname> <given-names>J</given-names></name> <name><surname>Seethalakshmy</surname> <given-names>A</given-names></name> <name><surname>Gokul</surname> <given-names>S</given-names></name></person-group>. <article-title>Adaptive feature selection for brain tumor classification in MRI images using genetic algorithm polar bear optimization and SVM.</article-title> <source><italic>J Comput Anal Appl.</italic></source> (<year>2024</year>) <volume>33</volume>:<fpage>356</fpage>&#x2013;<lpage>75</lpage></mixed-citation></ref>
<ref id="B27">
<label>27.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Po&#x0142;ap</surname> <given-names>D</given-names></name> <name><surname>Wo&#x017A;niak</surname> <given-names>M</given-names></name></person-group>. <article-title>Polar bear optimization algorithm: Meta-heuristic with fast population movement and dynamic birth and death mechanism.</article-title> <source><italic>Symmetry.</italic></source> (<year>2017</year>) <volume>9</volume>:<fpage>203</fpage>. <pub-id pub-id-type="doi">10.3390/sym9100203</pub-id></mixed-citation></ref>
<ref id="B28">
<label>28.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thamizhselvi</surname> <given-names>P</given-names></name> <name><surname>Muneera</surname> <given-names>N</given-names></name></person-group>. <article-title>An intensified polar bear optimization (IPBO)-Neoteric quantum neural network (NQNN) Classification model for breast cancer diagnosis.</article-title> <source><italic>Int J Comadem</italic>.</source> (<year>2023</year>) <volume>26</volume>.</mixed-citation></ref>
<ref id="B29">
<label>29.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ronneberger</surname> <given-names>O</given-names></name> <name><surname>Fischer</surname> <given-names>P</given-names></name> <name><surname>Brox</surname> <given-names>T</given-names></name></person-group>. <article-title>U-net: Convolutional networks for biomedical image segmentation.</article-title> in <source><italic>Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2015</year>). p. <fpage>234</fpage>&#x2013;<lpage>41</lpage>.</mixed-citation></ref>
<ref id="B30">
<label>30.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Micallef</surname> <given-names>N</given-names></name> <name><surname>Seychell</surname> <given-names>D</given-names></name> <name><surname>Bajada</surname> <given-names>C. J</given-names></name></person-group>. <article-title>Exploring the u-net++ model for automatic brain tumor segmentation.</article-title> <source><italic>IEEE Access</italic>.</source> (<year>2021</year>) <volume>9</volume>;<fpage>125523</fpage>&#x2013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2021.3111131</pub-id></mixed-citation></ref>
<ref id="B31">
<label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Isensee</surname> <given-names>F</given-names></name> <name><surname>Jaeger</surname> <given-names>P</given-names></name> <name><surname>Kohl</surname> <given-names>S</given-names></name> <name><surname>Petersen</surname> <given-names>J</given-names></name> <name><surname>Maier-Hein</surname> <given-names>K</given-names></name></person-group>. <article-title>nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation.</article-title> <source><italic>Nat Methods.</italic></source> (<year>2021</year>) <volume>18</volume>:<fpage>203</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-020-01008-z</pub-id> <pub-id pub-id-type="pmid">33288961</pub-id></mixed-citation></ref>
<ref id="B32">
<label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dice</surname> <given-names>LR</given-names></name></person-group>. <article-title>Measures of the amount of ecologic association between species.</article-title> <source><italic>Ecology</italic>.</source> (<year>1945</year>) <volume>26</volume>:<fpage>297</fpage>&#x2013;<lpage>302</lpage>. <pub-id pub-id-type="doi">10.2307/1932409</pub-id></mixed-citation></ref>
<ref id="B33">
<label>33.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Rahman</surname> <given-names>MA</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name></person-group>. <article-title>Optimizing intersection-over-union in deep neural networks for image segmentation.</article-title> In: Bebis, G.<etal/> editors. <source><italic>International Symposium on Visual Computing.</italic></source> <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>234</fpage>&#x2013;<lpage>44</lpage>.</mixed-citation></ref>
<ref id="B34">
<label>34.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>L</given-names></name> <name><surname>Dong</surname> <given-names>Q</given-names></name> <name><surname>Lin</surname> <given-names>D</given-names></name> <name><surname>Tian</surname> <given-names>C</given-names></name> <name><surname>L&#x00FC;</surname> <given-names>X</given-names></name></person-group>. <article-title>MUNet: a novel framework for accurate brain tumor segmentation combining UNet and mamba networks.</article-title> <source><italic>Front Comput Neurosci.</italic></source> (<year>2025</year>) <volume>19</volume>:<fpage>1513059</fpage>. <pub-id pub-id-type="doi">10.3389/fncom.2025.1513059</pub-id> <pub-id pub-id-type="pmid">39944950</pub-id></mixed-citation></ref>
<ref id="B35">
<label>35.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kaur</surname> <given-names>P</given-names></name> <name><surname>Kaushik</surname> <given-names>A</given-names></name> <name><surname>Jyotsna</surname> <given-names>I.</given-names></name> <name><surname>Singhal</surname> <given-names>A.</given-names></name> <name><surname>Pandey</surname> <given-names>R</given-names></name> <name><surname>Singhal</surname> <given-names>R</given-names></name></person-group>. <article-title>Advancing brain MRI segmentation using segment anything model.</article-title> <source><italic>Proc Comput Sci.</italic></source> (<year>2025</year>) <volume>206</volume>;<fpage>110</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1016/j.procs.2025.03.183</pub-id></mixed-citation></ref>
<ref id="B36">
<label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Christakakis</surname> <given-names>P</given-names></name> <name><surname>Protopapadakis</surname> <given-names>E</given-names></name></person-group>. <article-title>Post-heuristic cancer segmentation refinement over MRI images and deep learning models.</article-title> <source><italic>AI.</italic></source> (<year>2025</year>) <volume>6</volume>:<fpage>9</fpage>. <pub-id pub-id-type="doi">10.3390/ai6090212</pub-id></mixed-citation></ref>
<ref id="B37">
<label>37.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jlassi</surname> <given-names>A</given-names></name> <name><surname>Elbedoui</surname> <given-names>K</given-names></name> <name><surname>Barhoumi</surname> <given-names>W</given-names></name></person-group>. <article-title>Brain tumor segmentation of lower-grade glioma across MRI images using hybrid convolutional neural networks.</article-title> In: ICAART Editorial Board editor. <source><italic>International Conference on Agents and Artificial Intelligence, Science and Technology Publications.</italic></source> <publisher-name>LDA</publisher-name> (<year>2023</year>). p. <fpage>454</fpage>&#x2013;<lpage>65</lpage>. <pub-id pub-id-type="doi">10.5220/0011895900003393</pub-id></mixed-citation></ref>
<ref id="B38">
<label>38.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Obayya</surname> <given-names>M</given-names></name> <name><surname>Alshuhail</surname> <given-names>A</given-names></name> <name><surname>Mahmood</surname> <given-names>K</given-names></name> <name><surname>Alanazi</surname> <given-names>H</given-names></name> <name><surname>Alqahtani</surname> <given-names>M</given-names></name> <name><surname>Aljehane</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>A novel U-net model for brain tumor segmentation from MRI images.</article-title> <source><italic>Alexandria Eng J.</italic></source> (<year>2025</year>) <volume>126</volume>;<fpage>220</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1016/j.aej.2025.04.051</pub-id>.</mixed-citation></ref>
<ref id="B39">
<label>39.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alwadee</surname> <given-names>E</given-names></name> <name><surname>Sun</surname> <given-names>X</given-names></name> <name><surname>Qin</surname> <given-names>Y</given-names></name> <name><surname>Langbein</surname> <given-names>F</given-names></name></person-group>. <article-title>LATUP-Net: a lightweight 3D attention U-Net with parallel convolutions for brain tumor segmentation.</article-title> <source><italic>Comput Biol Med.</italic></source> (<year>2025</year>) <volume>184</volume>:<fpage>109353</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.109353</pub-id> <pub-id pub-id-type="pmid">39577347</pub-id></mixed-citation></ref>
<ref id="B40">
<label>40.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Angona</surname> <given-names>TM</given-names></name> <name><surname>Mondal</surname> <given-names>MRH</given-names></name></person-group>. <article-title>An attention based residual U-Net with swin transformer for brain MRI segmentation.</article-title> <source><italic>Array</italic>.</source> (<year>2025</year>) <volume>25</volume>:<fpage>100376</fpage>. <pub-id pub-id-type="doi">10.1016/j.array.2025.100376</pub-id>.</mixed-citation></ref>
<ref id="B41">
<label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pratap Joshi</surname> <given-names>K</given-names></name> <name><surname>Gowda</surname> <given-names>VB</given-names></name> <name><surname>Bidare Divakarachari</surname> <given-names>P</given-names></name> <name><surname>Siddappa Parameshwarappa</surname> <given-names>P</given-names></name> <name><surname>Patra</surname> <given-names>RK</given-names></name></person-group>. <article-title>VSA-GCNN: attention guided graph neural networks for brain tumor segmentation and classification.</article-title> <source><italic>Big Data Cogn Comput.</italic></source> (<year>2025</year>) <volume>9</volume>:<fpage>29</fpage>. <pub-id pub-id-type="doi">10.3390/bdcc9020029</pub-id>.</mixed-citation></ref>
<ref id="B42">
<label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ghosal</surname> <given-names> P</given-names></name> <name><surname>Roy</surname> <given-names> A</given-names></name> <name><surname>Agarwal</surname> <given-names> R</given-names></name> <name><surname>Purkayastha</surname> <given-names> K</given-names></name> <name><surname>Sharma</surname> <given-names> AL</given-names></name> <name><surname>Kumar</surname> <given-names>A</given-names></name></person-group>. <article-title>Compound attention embedded dual channel encoder-decoder for ms lesion segmentation from brain MRI.</article-title> <source><italic>Multimed Tools Appl</italic>.</source> (<year>2025</year>) <volume>84</volume>:<fpage>31139</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1007/s11042-024-20416-3</pub-id></mixed-citation></ref>
<ref id="B43">
<label>43.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>H</given-names></name> <name><surname>Wan</surname> <given-names>Y</given-names></name> <name><surname>Xu</surname> <given-names>H</given-names></name> <name><surname>Chen</surname> <given-names>L</given-names></name> <name><surname>Xiao</surname> <given-names>J</given-names></name> <name><surname>Ran</surname> <given-names>Q</given-names></name></person-group>. <article-title>Swinbtc: transfer learning to brain tumor classification for healthcare electronics using augmented MR images.</article-title> <source><italic>IEEE Trans Consumer Electronics</italic>.</source> (<year>2025</year>) <volume>2025</volume>;1. <pub-id pub-id-type="doi">10.1109/TCE.2025.3527061</pub-id></mixed-citation></ref>
<ref id="B44">
<label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ghosal</surname> <given-names>P</given-names></name> <name><surname>Chowdhury</surname> <given-names>T</given-names></name> <name><surname>Kumar</surname> <given-names>A</given-names></name> <name><surname>Bhadra</surname> <given-names>A</given-names></name> <name><surname>Chakraborty</surname> <given-names>J</given-names></name> <name><surname>Nandi</surname> <given-names>D</given-names></name></person-group>. <article-title>MhURI:a supervised segmentation approach to leverage salient brain tissues in magnetic resonance images.</article-title> <source><italic>Comput Methods Programs Biomed.</italic></source> (<year>2021</year>) <volume>200</volume>:<fpage>105841</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2020.105841</pub-id> <pub-id pub-id-type="pmid">33221057</pub-id></mixed-citation></ref>
<ref id="B45">
<label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Amri</surname> <given-names>Y</given-names></name> <name><surname>Ben Slama</surname> <given-names>A</given-names></name> <name><surname>Mbarki</surname> <given-names>Z</given-names></name> <name><surname>Selmi</surname> <given-names>R</given-names></name> <name><surname>Trabelsi</surname> <given-names>H</given-names></name></person-group>. <article-title>Automatic glioma segmentation based on efficient U-net model using MRI images.</article-title> <source><italic>Intell Based Med.</italic></source> (<year>2025</year>) <volume>11</volume>:<fpage>100216</fpage>. <pub-id pub-id-type="doi">10.1016/j.ibmed.2025.100216</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/59387/overview">Nasser Kashou</ext-link>, Kash Global Tech, United States</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1550000/overview">Palash Ghosal</ext-link>, Sikkim Manipal University, India</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3038569/overview">Yessine Amri</ext-link>, B&#x00E9;chir-Hamza Children&#x2019;s Hospital, Tunisia</p></fn>
</fn-group>
</back>
</article>