<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Environ. Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Environmental Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Environ. Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-665X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1657984</article-id>
<article-id pub-id-type="doi">10.3389/fenvs.2025.1657984</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Segmentation of arctic coastal shoreline and bluff edges using optical satellite imagery and deep learning</article-title>
<alt-title alt-title-type="left-running-head">Bagavathyraj et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fenvs.2025.1657984">10.3389/fenvs.2025.1657984</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Bagavathyraj</surname>
<given-names>Harshavardhini</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3119902"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal Analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Vargas Zesati</surname>
<given-names>Sergio</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3155190"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Fuentes</surname>
<given-names>Olac</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal Analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Peterson</surname>
<given-names>Sasha</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1473301"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tweedie</surname>
<given-names>Craig E.</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3249346"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x26; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/">Writing &#x2013; review and editing</role>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<institution>Department of Computer Science, University of Texas at El Paso</institution>, <city>El Paso</city>, <state>TX</state>, <country country="US">United States</country>
</aff>
<aff id="aff2">
<label>2</label>
<institution>Biological Sciences Department, University of Texas at El Paso</institution>, <city>El Paso</city>, <state>TX</state>, <country country="US">United States</country>
</aff>
<aff id="aff3">
<label>3</label>
<institution>Environmental Science and Engineering Program, University of Texas at El Paso</institution>, <city>El Paso</city>, <state>TX</state>, <country country="US">United States</country>
</aff>
<author-notes>
<corresp id="c001">
<label>&#x2a;</label>Correspondence: Harshavardhini Bagavathyraj, <email xlink:href="mailto:hbagavathyr@miners.utep.edu">hbagavathyr@miners.utep.edu</email>
</corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-21">
<day>21</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>13</volume>
<elocation-id>1657984</elocation-id>
<history>
<date date-type="received">
<day>02</day>
<month>07</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>29</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>06</day>
<month>11</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Bagavathyraj, Vargas Zesati, Fuentes, Peterson and Tweedie.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Bagavathyraj, Vargas Zesati, Fuentes, Peterson and Tweedie</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-21">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The Arctic coastline spans multiple countries, supports Indigenous livelihoods, and plays a vital role in the Arctic system. Rapid climate change is accelerating permafrost thaw, sea-level rise, and coastal erosion, underscoring the need for decision making to be informed by accurate delineation of tundra shoreline (instantaneous water line) and bluff edge (vegetation&#x2013;slope boundary) position and change trends. To address this need, we compared two image segmentation approaches for mapping Arctic land and water interfaces from high-resolution satellite imagery. (1) U-Net, a supervised convolutional neural network trained on expert-annotated scenes, and (2) Differentiable Feature Clustering (DifFeat), an unsupervised model applied in a minimally supervised manner via expert-guided cluster selection. The shoreline and bluff edge boundaries were derived from the segmented land and water masks using an automated interface extraction approach. DifFeat achieved higher segmentation accuracy, with IoU values of 0.95 (water) and 0.92 (land), compared to U-Net&#x2019;s 0.58 and 0.50, respectively. U-Net produced reliable results and benefited from infrared and vegetation spectral indices, but required extensive annotation and showed limited generalization to UAV imagery. DifFeat achieved superior results without manual annotation, reducing the dependence on labeled data and completing training 99.87% faster than U-Net. These findings highlight the complementary strengths of supervised and semi-supervised models for Arctic coastal mapping, with DifFeat offering a scalable, label-efficient solution for long-term coastal-change monitoring. Future work will integrate elevation data to further improve bluff edge feature detection.</p>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>segmentation</kwd>
<kwd>shoreline</kwd>
<kwd>bluff edge</kwd>
<kwd>U-Net</kwd>
<kwd>differentiable feature clustering</kwd>
<kwd>satellite imagery</kwd>
<kwd>unsupervised learning</kwd>
</kwd-group>
<funding-group>
<funding-statement>The authors declare that financial support was received for the research and/or publication of this article. This research was supported by the Office of Polar Programs at the US National Science Foundation (Grant Nos: 1656026, 1836861, 1927373, 2318378, and 2322664), and the National Aeronautics and Space Administration (Grant No: 80NSSC21K1164).</funding-statement>
</funding-group>
<counts>
<fig-count count="6"/>
<table-count count="2"/>
<equation-count count="0"/>
<ref-count count="35"/>
<page-count count="15"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Environmental Informatics and Remote Sensing</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>During the past 40&#xa0;years, the Arctic has warmed nearly four times faster than the global average (<xref ref-type="bibr" rid="B28">Rantanen et al., 2022</xref>). Much of the Arctic coastal zone is underlain by permafrost, a landform that remains frozen for two or more years (<xref ref-type="bibr" rid="B22">National Snow and Ice Data Center, 2023</xref>), which stabilizes landscapes and stores large amounts of carbon (<xref ref-type="bibr" rid="B32">Schuur et al., 2015</xref>). As permafrost thaws, it disrupts the Arctic carbon balance and reduces the ocean&#x2019;s ability to absorb atmospheric carbon dioxide, amplifying global climate feedbacks (<xref ref-type="bibr" rid="B15">Jones et al., 2018</xref>; <xref ref-type="bibr" rid="B25">Nielsen et al., 2024</xref>). Thawing also destabilizes the ice-rich coastal bluffs, triggering rapid erosion and land loss. Erosion rates have doubled over the past half century (<xref ref-type="bibr" rid="B24">Nielsen et al., 2022</xref>) and are projected to increase two to three times by 2,100 (<xref ref-type="bibr" rid="B24">Nielsen et al., 2022</xref>). This process threatens local infrastructure, cultural heritage, and indigenous subsistence practices by reshaping ecosystems and disrupting food webs (<xref ref-type="bibr" rid="B5">Cassidy et al., 2024</xref>; <xref ref-type="bibr" rid="B25">Nielsen et al., 2024</xref>; <xref ref-type="bibr" rid="B11">Fritz et al., 2017</xref>; <xref ref-type="bibr" rid="B16">Juma et al., 2025</xref>). <xref ref-type="fig" rid="F1">Figure 1C</xref> shows an example of bluff collapse, where entire large sections of the coastal landscape can erode to the ocean. When combined with sea level rise and permafrost subsidence, total land loss can be six to eight times greater than erosion alone (<xref ref-type="bibr" rid="B7">Creel et al., 2024</xref>). However, long-term understanding of Arctic coastal change remains limited: approximately 86% of the Arctic coastline lacks positional data (i.e., spatial and temporal information of landscape features) (<xref ref-type="bibr" rid="B15">Jones et al., 2018</xref>). As erosion accelerates and social vulnerability increases, so does the need for scalable and automated methods to monitor Arctic coastal change (<xref ref-type="bibr" rid="B24">Nielsen et al., 2022</xref>; <xref ref-type="bibr" rid="B14">Irrgang et al., 2022</xref>).</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>
<bold>(A)</bold> Base map showing WorldView-2 imagery captured in July 2010, highlighting the approximate 11&#xa0;km-long shoreline study area (outlined in purple) adjacent to Elson Lagoon, near the I&#xf1;upiat village of Utqia&#x121;vik on the North slope of Alaska. <bold>(B)</bold> A UAV orthoimage highlights a coastal section with manually annotated shoreline (blue) and bluff edge (orange) lines used as reference. <bold>(C)</bold> Ground-based image showing a heavily eroded coastal tundra section along Elson Lagoon.</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g001.tif">
<alt-text content-type="machine-generated">(A) Base map showing WorldView-2 imagery captured in July 2010, highlighting the approximate 11 km-long shoreline study area (outlined in purple) adjacent to Elson Lagoon, near the I&#x00F1;upiat village of Utqia3vik on the North slope of Alaska. (B) A UAV orthoimage highlights a coastal section with manually annotated shoreline (blue) and bluff edge (orange) lines used as reference. (C) Ground-based image showing a heavily eroded coastal tundra section along Elson Lagoon.</alt-text>
</graphic>
</fig>
<p>Although this study employs deep learning, previously used methods to identify coastal features have traditionally relied on field surveys, manual mapping, or digitization in Geographic Information Systems (GIS). These approaches are resource-intensive, prone to human error, and particularly difficult to scale in remote Arctic regions where accessibility and field conditions are challenging (<xref ref-type="bibr" rid="B15">Jones et al., 2018</xref>; <xref ref-type="bibr" rid="B14">Irrgang et al., 2022</xref>). In recent years, high-resolution satellite imagery has become more accessible through improvements in spatial resolution, revisit frequency, and the expansion of open-access platforms such as Landsat and Sentinel (<xref ref-type="bibr" rid="B12">Gabarr&#xf3; et al., 2023</xref>; <xref ref-type="bibr" rid="B3">Beamish et al., 2020</xref>; <xref ref-type="bibr" rid="B33">Wenzl et al., 2024</xref>). However, images alone cannot extract detailed geospatial features without automated analytical techniques. Machine learning approaches have been increasingly applied to classify land and water regions and to identify shoreline positions from imagery (<xref ref-type="bibr" rid="B4">Bengoufa et al., 2021</xref>). Models such as Random Forests (RF) (<xref ref-type="bibr" rid="B1">Aryal et al., 2021</xref>; <xref ref-type="bibr" rid="B19">McAllister et al., 2022</xref>; <xref ref-type="bibr" rid="B4">Bengoufa et al., 2021</xref>), Support Vector Machines (SVM) (<xref ref-type="bibr" rid="B19">McAllister et al., 2022</xref>; <xref ref-type="bibr" rid="B4">Bengoufa et al., 2021</xref>), Decision Trees, K-Nearest Neighbors (KNN) (<xref ref-type="bibr" rid="B19">McAllister et al., 2022</xref>), and XGBoost (<xref ref-type="bibr" rid="B9">Efimova et al., 2020</xref>) have been applied in this context. Although effective for broad land&#x2013;water classification, these models often lack the spatial precision needed to delineate complex Arctic coastlines or capture subtle geomorphic transitions. To overcome these limitations, deep learning approaches leverage spatial context and hierarchical feature extraction to provide improved boundary accuracy and generalization for large-scale Arctic coastal mapping (<xref ref-type="bibr" rid="B31">Scala et al., 2024</xref>; <xref ref-type="bibr" rid="B29">Ronneberger et al., 2015</xref>).</p>
<p>Deep learning methods have shown strong performance in land&#x2013;water segmentation and shoreline mapping, outperforming traditional machine learning approaches in spatial precision, robustness to scene variability, and generalization across diverse landscapes (<xref ref-type="bibr" rid="B1">Aryal et al., 2021</xref>; <xref ref-type="bibr" rid="B6">Clark et al., 2022</xref>). Among these, convolutional neural networks (CNNs), particularly U-Net models (<xref ref-type="bibr" rid="B1">Aryal et al., 2021</xref>; <xref ref-type="bibr" rid="B27">Philipp et al., 2022</xref>; <xref ref-type="bibr" rid="B34">Yang et al., 2020</xref>; <xref ref-type="bibr" rid="B13">Heidler et al., 2021</xref>; <xref ref-type="bibr" rid="B8">Dang et al., 2022</xref>; <xref ref-type="bibr" rid="B26">Park and Song, 2024</xref>) have proven effective at capturing both fine spatial detail and broader contextual patterns, making them well suited for delineating land&#x2013;water interfaces even in settings with limited or partially annotated training data (<xref ref-type="bibr" rid="B1">Aryal et al., 2021</xref>). However, most U-Net&#x2013;based shoreline studies have been developed and validated on temperate or tropical coasts, with limited adaptation to Arctic environments. Applying such models in the Arctic presents additional challenges, including short ice-free seasons that limit the availability of optical imagery, persistent cloud cover that reduces the number of usable satellite scenes (<xref ref-type="bibr" rid="B24">Nielsen et al., 2022</xref>; <xref ref-type="bibr" rid="B2">Bartsch et al., 2020</xref>), and complex permafrost-related geomorphologies such as polygonal tundra, thaw lake basins, and ice wedge networks that can introduce spectral and textural noise in classifications (<xref ref-type="bibr" rid="B2">Bartsch et al., 2020</xref>; <xref ref-type="bibr" rid="B24">Nielsen et al., 2022</xref>; <xref ref-type="bibr" rid="B14">Irrgang et al., 2022</xref>). Although pixel-wise segmentation of land and water enables accurate shoreline detection, Arctic coastal monitoring also requires delineation of bluff edge boundaries, an equally important but often overlooked component for quantifying land loss and erosion dynamics (<xref ref-type="bibr" rid="B15">Jones et al., 2018</xref>). In this study, the shoreline is defined as the instantaneous water line, and the bluff edge as the transition from vegetated tundra to a non-vegetated slope or cliff, as shown in <xref ref-type="fig" rid="F1">Figure 1B</xref> (blue: shoreline; orange: bluff edge).</p>
<p>Although supervised deep learning models like U-Net (<xref ref-type="bibr" rid="B29">Ronneberger et al., 2015</xref>) have demonstrated success in semantic segmentation tasks using partially labeled data, their application to Arctic land&#x2013;water segmentation and boundary extraction remains limited. Performing detailed land and water segmentation that can support accurate shoreline and bluff edge delineation typically requires large, high-quality annotated datasets, an effort that is time-consuming and difficult to scale in the Arctic. In contrast, unsupervised or minimally supervised deep learning approaches for Arctic coastal mapping remain largely unexplored, despite their potential to reduce reliance on labeled data. Recent foundation models such as the Segment Anything Model (SAM) (<xref ref-type="bibr" rid="B18">Kirillov et al., 2023</xref>) demonstrate zero-shot segmentation without task-specific labels, but were not considered here due to their reliance on per-image prompting (points, boxes, or masks), high computational demands for high-resolution satellite imagery, and training on natural image datasets that differ from the narrow, elongated geomorphic features typical of Arctic coasts. In addition, SAM&#x2019;s outputs can be sensitive to scale and orientation, reducing consistency when applied to such linear coastal forms. These constraints motivate the exploration of alternative minimally supervised segmentation methods, such as Differentiable Feature Clustering (DifFeat) (<xref ref-type="bibr" rid="B17">Kim et al., 2020</xref>) that can produce consistent land&#x2013;water delineations from which geomorphic boundaries are derived without per-image manual prompting.</p>
<p>To address the limitations and challenges of different deep learning methods, we investigate two approaches for semantic segmentation of Arctic land and water surfaces. The models are trained on WorldView-2 satellite imagery of Elson Lagoon (<xref ref-type="fig" rid="F1">Figure 1</xref>) and are also evaluated on high-resolution UAV imagery of the same site, enabling assessment of cross-sensor generalization across differing spatial resolutions. We apply U-Net, a supervised convolutional neural network trained on manually annotated satellite scenes, and DifFeat (<xref ref-type="bibr" rid="B17">Kim et al., 2020</xref>), an unsupervised deep learning algorithm trained on a single unlabeled image, with minimal supervision introduced through expert-guided cluster selection of land and water classes on one image tile. To our knowledge, this represents the first application of an unsupervised deep learning framework with minimal expert supervision for Arctic coastal mapping. From the resulting land and water segmentation, shoreline and bluff edge boundaries are automatically derived using an interface extraction approach; these geomorphic boundaries serve as critical indicators of permafrost-driven land loss and coastal change. This framework enables a direct comparison between supervised and minimally supervised models, examining trade-offs in segmentation accuracy, annotation effort, and cross-sensor generalization. Specifically, this study aims to (1) evaluate both models for accurate land and water segmentation that supports reliable shoreline and bluff edge extraction, (2) compare supervised and minimally supervised approaches in terms of performance and generalization, and (3) assess their potential for larger-scale Arctic coastal monitoring applications.</p>
</sec>
<sec sec-type="methods" id="s2">
<label>2</label>
<title>Methodology</title>
<p>We implemented two deep learning-based approaches, U-Net and Differentiable Feature Clustering (DifFeat), to perform semantic segmentation of land and water surfaces using high resolution optical imagery. The following subsections outline dataset preparation, and the preprocessing pipeline and interface extraction approach, and describe the implementation and inference of the two models.</p>
<sec id="s2-1">
<label>2.1</label>
<title>Dataset source</title>
<p>For segmentation experiments, we used five pan-sharpened WorldView-2 (WV2) images (red, green, blue (RGB) and infrared1 (IR1) bands only) with less than 20% cloud coverage, acquired between May and September in multiple years from 2010 to 2024. The imagery was provided by the Polar Geospatial Center (PGC) at the University of Minnesota through their cooperative agreement with the US National Science Foundations&#x2019; Office of Polar Programs. The spatial resolution of the WV2 images is approximately 0.5&#xa0;m of ground sample distance (GSD), providing sufficient detail for land&#x2013;water segmentation (<xref ref-type="bibr" rid="B10">European Space Agency, 2025</xref>). Each image is approximately 18,000&#x2013;25,000 pixels in height and 17,000&#x2013;25,000 pixels in width. The off-nadir angles for the five images were 11&#xb0;, 25&#xb0;, 31.6&#xb0;, 42.9&#xb0;, and 44.2&#xb0;, reflecting differences in satellite viewing geometry which may influence shoreline and bluff visibility.</p>
<p>The geographic region was constrained to Elson Lagoon, a shallow embayment near the I&#xf1;upiat community of Utqia&#x121;vik (<xref ref-type="fig" rid="F1">Figure 1</xref>). Elson Lagoon was selected because Arctic Lagoons are geomorphologically diverse and ecologically significant, representing approximately 44% of the Alaskan Beaufort Sea coastline and offering a strong analog for Arctic coastal systems more broadly (<xref ref-type="bibr" rid="B21">Miller et al., 2021</xref>). Additionally, coastal erosion has been monitored along the Elson Lagoon shoreline for over 8&#xa0;decades, providing a valuable resource of <italic>in situ</italic> data that can be used to validate newly developed machine learning-based approaches. The imagery captures diverse geomorphic elements, land-water interfaces, coastal bluffs, and permafrost-related landscape features, which are crucial for studying the impacts of climate change and coastal erosion in the Arctic. The dataset also reflects the unique imaging challenges of the region, including low contrast, mixed spectral textures between the shoreline, vegetation, and lagoon water, as well as seasonal variability. These complexities make the data particularly valuable for the Arctic more broadly, as they represent common issues encountered in Arctic satellite image analysis.</p>
</sec>
<sec id="s2-2">
<label>2.2</label>
<title>Annotation</title>
<p>The five WV2 scenes from Elson Lagoon were manually annotated using ArcGIS Pro (version 3.5) to delineate land and water regions. These annotations served as ground-truth labels for the supervised model training. Annotation was performed by an expert with direct field experience at the site and prior <italic>in situ</italic> data collection over multiple seasons. The imagery selected for annotation featured ideal environmental conditions: minimal cloud cover, no snow or ice, clear visibility of land&#x2013;water transitions, and favorable sensor viewing angles. The labeling scheme focused on identifying only land and water classes. Areas not belonging to either category, such as beaches, sloping bluffs, tundra surfaces, or small inland water bodies, were left unannotated and thus implicitly treated as background. This structure allows for clear delineation of the main surface types while preserving the natural transitions needed for subsequent boundary extraction.</p>
<p>Annotations were created as polygonal masks, enabling a consistent delineation of spatial regions. Polygon boundaries were drawn along visually identifiable transitions based on surface color, texture, and shading. In areas with low contrast or shadowing, expert interpretation was used, guided by site-specific knowledge and prior field observations, to approximate the most probable land&#x2013;water separation. Despite favorable conditions, challenges such as spectral confusion between shallow water and wet sediment, vegetation gradients along bluffs, and surface moisture occasionally complicated boundary placement. To ensure consistency, detailed labeling guidelines were developed and applied uniformly in all five scenes, each requiring approximately 2.5&#x2013;3.5&#xa0;h to annotate. <xref ref-type="fig" rid="F2">Figures 2A,B</xref> shows examples of the imagery and corresponding annotated polygons, where land and water features are visualized in tan and teal, respectively. <xref ref-type="sec" rid="s11">Supplementary Figure S1</xref> provides additional visualization of the manual annotations overlaid on WV2 imagery, illustrating the spatial characteristics and delineation detail of the digitized features.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Illustration of the DifFeat workflow for identifying target features: <bold>(A)</bold> False-color input RGB image slice, <bold>(B)</bold> Corresponding manually annotated water (blue) and land (green) polygon, <bold>(C)</bold> Segmentation map generated by the DifFeat model, <bold>(D)</bold> Target boundaries extracted from the segmentation map, and <bold>(E)</bold> Final predictions overlaid on the input image slice.</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g002.tif">
<alt-text content-type="machine-generated">Illustration of the DifFeat workflow for identifying target features: (A) False-color input RGB image slice, (B) Corresponding manually annotated water (blue) and land (green) polygon, (C) Segmentation map generated by the DifFeat model, (D) Target boundaries extracted from the segmentation map, and (E) Final predictions overlaid on the input image slice.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s2-3">
<label>2.3</label>
<title>Data preparation</title>
<p>The satellite imagery used in this study underwent multiple pre-processing steps to ensure compatibility with the U-Net and DifFeat models. These steps standardized spatial alignment, improved feature separability, and prepared the data for model input. The main steps are described below.</p>
<sec id="s2-3-1">
<label>2.3.1</label>
<title>Image and shapefile matching</title>
<p>Each WV2 satellite image was paired with its corresponding manually annotated land and water polygons. All raster images and shapefiles were maintained in a consistent projected coordinate reference system (NAD83 (2011) datum, UTM zone 4N, EPSG:26904) to ensure spatial alignment across datasets. Because the annotations were created directly on the source imagery, they aligned precisely with the corresponding scenes, providing an accurate basis for model training and evaluation.</p>
</sec>
<sec id="s2-3-2">
<label>2.3.2</label>
<title>Adding normalized bands</title>
<p>To enhance spectral separability between land and water surfaces, two normalized spectral indices were computed and appended to the original four-band imagery (RGB &#x2b; IR1): the Normalized Difference Water Index (NDWI) (<xref ref-type="bibr" rid="B20">McFeeters, 1996</xref>) and the Normalized Difference Vegetation Index (NDVI) (<xref ref-type="bibr" rid="B30">Rouse et al., 1974</xref>). NDWI highlights open-water regions, improving the detection of subtle water boundaries, while NDVI highlights vegetated terrain, aiding the distinction of land-water boundaries under varying illumination and surface conditions.</p>
</sec>
<sec id="s2-3-3">
<label>2.3.3</label>
<title>Image and label slicing</title>
<p>All experiments were carried out in PyTorch on a server equipped with two NVIDIA GeForce RTX 3090 GPUs (24&#xa0;GB each, total 48&#xa0;GB) running CUDA 12.4 and driver 550.90.07, enabling efficient parallel training. Because the WV2 scenes were too large to process in their entirety, both imagery and annotation masks were divided into smaller tiles of size 512 <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mo>&#xd7;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 512 pixels. Polygon annotations were rasterized into class-coded mask arrays, where pixel values corresponded to the following classes: 0 &#x3d; background, 1 &#x3d; water, and 2 &#x3d; land. The slicing procedure ensured perfect spatial alignment between each image tile and its corresponding label. To prevent overrepresentation of homogeneous regions, the dataset was curated so that approximately 90% of the image tiles contained both land and water pixels, capturing transition zones essential for learning land and water interfaces. The remaining 10% of tiles represented single-class regions (e.g., open water or inland tundra) to help the model distinguish pure land and water surfaces and improve generalization to spatially uniform areas. For image edges smaller than the target tile size, zero-padding was applied to both imagery and masks to maintain consistent dimensions across the dataset.</p>
</sec>
<sec id="s2-3-4">
<label>2.3.4</label>
<title>Normalization of image bands</title>
<p>Each image tile, including the RGB and infrared (IR) bands, was normalized independently using z-score normalization, where pixel values were scaled according to the mean and standard deviation of the training dataset. NoData pixels and padded regions were excluded from these calculations to avoid biasing the normalization statistics. The NDWI and NDVI indices were computed from the original (non-normalized) reflectance values and, being ratio-based measures, were already normalized by definition and required no further adjustment. For inference, the mean and standard deviation of the RGB and IR bands derived from the training data were applied to maintain consistency and mitigate lighting or sensor-specific variation across scenes.</p>
</sec>
</sec>
<sec id="s2-4">
<label>2.4</label>
<title>Data splitting and experimental setup</title>
<p>For all experiments, the dataset consisted of five WorldView-2 (WV2) satellite images acquired between 2010 and 2024. These images were divided into distinct subsets: three images were designated for training, one for validation, and one for testing. All tiles used for model evaluation were exclusively obtained from the held-out evaluation image, ensuring strict separation between training, validation, and test data. The tiling process resulted in 528 training tiles, 196 validation tiles, and 255 testing tiles, each measuring <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixels. While some tiles partially overlapped spatially across acquisitions, differences in capture year, illumination, and viewing geometry ensured that the evaluation imagery remained distinct from the training and validation datasets. This protocol was consistently applied to both the U-Net and DifFeat models, providing a robust basis for comparison and minimizing the risk of overfitting to specific acquisition conditions.</p>
</sec>
<sec id="s2-5">
<label>2.5</label>
<title>Cross-domain evaluation on large-scale WV2 and UAV imagery</title>
<p>To assess model generalization beyond training conditions, we evaluated both U-Net and DifFeat on large-scale Arctic imagery: (1) a large-scale WorldView-2 (WV2) satellite scene (<inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mn>18,000</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>20,000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixels) covering the Elson Lagoon region, and (2) a high-resolution unoccupied aerial vehicle (UAV) orthomosaic (<inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mn>20,000</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>20,000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixels at 2&#xa0;cm/pixel) from the same coastal area. These scenes were not part of the training or validation data, allowing us to evaluate cross-scale and cross-source generalization. The large-scale WV2 image was processed in the same manner as the training dataset, including band selection, normalization, and tiling into <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixel patches for model inference. The predictions of each tile were mosaicked to produce a seamless segmentation of the entire scene, with boundary extraction performed as described in <xref ref-type="sec" rid="s2-6-1">Sections 2.6.1</xref> and <xref ref-type="sec" rid="s2-7-1">2.7.1</xref>.</p>
<p>The UAV imagery was collected on 6 August 2024 using a Quantum Systems Trinity F90&#x2b; fixed-wing UAV equipped with a Micasense Altum-PT multispectral sensor. The flight was carried out 120&#xa0;m above ground level (AGL) over the Elson Lagoon coastline, following a corridor mapping path with 80% forward and side overlap. The resulting dataset was processed using a standard structure-from-motion (SfM) workflow, producing a 2&#xa0;cm/pixel ground sampling distance (GSD) orthomosaic. The images were downsampled by extracting <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:mn>5120</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5120</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> slices and resizing them to <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> to match the input size of the model, thus preserving large-scale spatial patterns while reducing fine spatial detail. The same preprocessing, inference, and boundary extraction steps as with the training data were applied to ensure methodological consistency. No model re-training or fine-tuning was performed for these tests. The results of the large-scale WV2 image and the UAV orthomosaic were used exclusively to assess the ability of the trained models to generalize to new spatial scales and imaging platforms.</p>
</sec>
<sec id="s2-6">
<label>2.6</label>
<title>U-Net</title>
<p>U-Net, a fully convolutional neural network originally developed for biomedical image segmentation (<xref ref-type="bibr" rid="B29">Ronneberger et al., 2015</xref>), features an encoder&#x2013;decoder structure with skip connections that allow simultaneous capture of global context and fine spatial detail. The encoder path extracts high-level contextual features through successive convolution and down-sampling operations, while the decoder path progressively reconstructs spatial details using up-sampling and convolution layers. Skip connections between corresponding encoder and decoder stages allow spatial information lost during down-sampling to be preserved and integrated during reconstruction, giving U-Net its distinctive U-shaped architecture.</p>
<p>In this study, U-Net was employed to perform semantic segmentation of land and water regions in high-resolution Arctic satellite imagery. Accurate delineation of these classes is essential for subsequent extraction of shoreline and bluff edge boundaries via post-processing. Because Arctic coastal transitions can be narrow, low contrast, and highly variable in appearance, the localization capability of U-Net is particularly well suited to capture fine-scale land-water interfaces that define these geomorphic boundaries. Training images and their corresponding label masks were prepared following the procedures described in <xref ref-type="sec" rid="s2-3">Section 2.3</xref>, including normalization, tiling, and class balance adjustments, emphasizing tiles containing both land and water pixels.</p>
<p>To enhance generalization, data augmentation was applied during training using random rotations and horizontal and vertical flips, exposing the model to varied orientations of coastal features. The network was trained using a hybrid loss function combining Dice loss, which mitigates class imbalance, and Boundary loss, which sharpens land&#x2013;water transitions by penalizing errors near class boundaries (<xref ref-type="bibr" rid="B1">Aryal et al., 2021</xref>). Optimization was performed with the Adam optimizer using a batch size of 8 and an initial learning rate of 0.001. A learning rate scheduler reduced the learning rate after 30 epochs to improve convergence stability. Model performance was evaluated using Intersection over Union (IoU), Precision, and Recall, computed for the land and water classes. These metrics provide a comprehensive assessment of segmentation accuracy and boundary fidelity, directly influencing the quality of the derived shoreline and bluff edge boundaries. Quantitative results are presented in the Results section.</p>
<sec id="s2-6-1">
<label>2.6.1</label>
<title>Inference</title>
<p>During inference, each large WV2 image was divided into smaller patches matching the model&#x2019;s input size. The trained U-Net model produced per-pixel class probabilities (softmax) for each patch, which were converted to discrete class labels (0 &#x3d; background, 1 &#x3d; water, 2 &#x3d; land) using an argmax operation. Both the predicted class maps and, optionally, the derived boundaries were written to pre-allocated larger arrays that preserved the original spatial dimensions and alignment across tiles. From each predicted patch, shoreline and bluff edge boundaries were extracted using the interface extraction method (<xref ref-type="sec" rid="s2-8">Section 2.8</xref>), which identifies the interfaces between land, water, and background regions. This boundary extraction step can be performed per patch or after mosaicking, depending on the desired output format. After processing all patches, the complete mosaicked raster was assembled to reconstruct the full scene.</p>
<p>The final outputs were then vectorized: the land and water segmentation was converted to polygon geometries, and when boundaries were extracted, the shoreline and bluff edge features were exported as polyline shapefiles. These products provide both region and line-based representations suitable for further spatial or ecological analysis. Unless explicitly stated elsewhere, no additional post-processing (e.g., smoothing or morphological filtering) was applied within the core U-Net pipeline. For cross-domain evaluation, U-Net models trained on satellite imagery were also applied to UAV-derived orthomosaics acquired for the same region in 2024 without additional fine-tuning to assess generalization performance.</p>
</sec>
</sec>
<sec id="s2-7">
<label>2.7</label>
<title>Differentiable feature clustering</title>
<p>Differentiable Feature Clustering (DifFeat) is an unsupervised learning technique that leverages feature representation learning for image segmentation (<xref ref-type="bibr" rid="B17">Kim et al., 2020</xref>). Its key advantage lies in its ability to perform segmentation without the need for labeled data, making it particularly valuable in Arctic settings where annotated datasets are scarce or expensive to obtain. DifFeat clusters pixels based on learned feature representations extracted from a convolutional feature map, where each pixel is represented as a multidimensional vector across channels. Clustering is performed by assigning each pixel to the channel (cluster) with the highest activation, effectively grouping pixels with similar spectral&#x2013;textural characteristics. In the implementation, the initial and minimum number of clusters are user-defined parameters. The model begins from the specified maximum and iteratively merges similar pixel groups during optimization until the minimum threshold is reached, controlling the final segmentation granularity. The optimization follows the loss formulation of <xref ref-type="bibr" rid="B17">Kim et al. (2020)</xref>, combining Cross-Entropy loss which encourages grouping of similar pixels and L1 loss which enforces spatial continuity and smoothness by penalizing abrupt feature changes between neighboring pixels. Together, these mechanisms produce spatially coherent clusters that correspond to distinct geomorphic surface types.</p>
<p>As an unsupervised approach, DifFeat generates multiple clusters from the input image slices (<xref ref-type="fig" rid="F2">Figure 2C</xref>). The model does not assign semantic classes to these clusters and the model&#x2019;s role is limited to grouping pixels with similar features. In our workflow, this is adapted into a minimally supervised method: a single manual step is used to visually inspect clusters from a reference slice and identify those corresponding to the two target classes (water and land). Once the target clusters are identified, their cluster IDs are fixed and used consistently across all slices and images, enabling fully automated extraction in subsequent predictions. If a selected cluster does not appear in a slice, it simply reflects that the class is absent from that portion of the scene. This fixed-ID mapping is central to reusing the model output without reintroducing manual work.</p>
<p>DifFeat&#x2019;s segmentation approach provides flexibility in cluster granularity, supporting both coarse-grained and fine-grained clustering depending on the desired level of feature separation. An important advantage of DifFeat in our research is its ability to operate effectively with minimal data, processing as little as a single image slice of size <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:mn>224</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>224</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> or <inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixels. This is particularly useful in Arctic coastal monitoring, where data are scarce due to remote site locations and environmental challenges such as persistent cloud cover. The input image slice is obtained from the slicing process explained in <xref ref-type="sec" rid="s2-3-3">Section 2.3.3</xref>, normalized to standardize the pixel intensities and improve the stability of the feature clustering. The model can process various combinations of spectral bands, including RGB, RGB&#x2b;IR, or RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI, providing flexibility in using additional information for clustering. The model is initialized with random weights and trained from scratch, ensuring that the learned features are specific to Arctic coastal imagery.</p>
<p>The resulting class-coded masks, which comprise land, water and background regions, were subsequently processed using the automated interface extraction workflow (<xref ref-type="sec" rid="s2-8">Section 2.8</xref>) to derive boundaries of the shoreline and the bluff edge. This unified post-processing step ensures methodological consistency between the supervised and unsupervised approaches, enabling direct comparison of segmentation accuracy, generalization, and data efficiency. <xref ref-type="fig" rid="F2">Figure 2</xref> illustrates the complete DifFeat workflow: (A) input WV2 image slice, (B) ground-truth mask displayed for comparison only (not used for model training), (C) clustering or segmentation map obtained from the model, (D) automatically extracted shoreline and bluff edge boundaries from land and water segments, and (E) the boundaries overlaid on the input image. Together, these panels demonstrate how DifFeat, despite operating without labeled data, produces coherent land and water delineations from which geomorphically meaningful shoreline and bluff edge boundaries can be objectively derived.</p>
<sec id="s2-7-1">
<label>2.7.1</label>
<title>Inference</title>
<p>During inference, the model processed each large WV2 image in overlapping/non-overlapping slices that matched the input size (<inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> pixels). Predictions were generated for each slice, and the fixed cluster ID mapping established during the reference selection step was applied to retain only the clusters corresponding to land and water. The slice-level outputs were then mosaicked into a seamless array matching the original scene dimensions. The resulting segmentation map was exported as a GeoPandas shapefile with polygon geometry representing land and water regions. Optionally, these polygons were passed through the automated interface extraction workflow (<xref ref-type="sec" rid="s2-8">Section 2.8</xref>) to obtain shoreline and bluff edge polylines. <xref ref-type="sec" rid="s11">Supplementary Figure S2</xref> illustrates this process: (A) shows the raw DifFeat prediction overlaid on the evaluation image, and (B) presents the corresponding derived boundaries after optional refinement.</p>
<p>By refinement, we refer to the limited manual post-processing performed for visualization purposes and to generate clean boundary shapefiles without artifacts at tile edges, as shown in <xref ref-type="fig" rid="F5">Figure 5</xref> and <xref ref-type="sec" rid="s11">Supplementary Figure S2B</xref>. This manual post-processing involved removing small noisy detections, refining edge continuity, and ensuring spatial coherence across tile boundaries. This cleaning step, though derived from automated output, was necessary because the model occasionally introduced artifacts in the large mosaic. The latter included misclassification of small inland water bodies or tundra features as bluff edges, and false shoreline detections along tile corners containing open water. These manual adjustments were applied only for the preparation of final visualizations and presentation materials, and were not used during automated boundary extraction or for quantitative evaluation of model performance. For cross-domain evaluation, DifFeat trained solely on one image slice of satellite imagery was also applied to a full satellite imagery scene and UAV orthomosaics without any fine-tuning, allowing for the assessment of generalization across sensors and spatial scales.</p>
</sec>
</sec>
<sec id="s2-8">
<label>2.8</label>
<title>Post-processing: interface extraction</title>
<p>Following segmentation, we derived geomorphic boundaries corresponding to the shoreline and bluff edge from the predicted land, water and background masks. Specifically, the model predicts two explicit classes, land and water, while unclassified regions (e.g., beach or bluff slopes) are treated as background. These boundaries are not directly predicted by the model but are instead derived from the spatial interfaces between the segmented regions. Boundaries were extracted by identifying adjacency relationships among these classes. For each class map, we first defined binary masks for land, water, and background. Using 8-connected morphological dilation with a 3 <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mo>&#xd7;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 3 structuring element, we delineated geomorphic boundaries by locating water pixels adjacent to land or background (shoreline) and land pixels adjacent to water or background (bluff edge). To remove artifacts, small closed loops were filtered out and boundary pixels along image edges were discarded. This approach ensures that both shoreline and bluff edge lines are objectively derived from the predicted land and water regions, without requiring manual delineation for each image or direct training of linear features. The same procedure can be applied to either ground-truth or predicted masks, enabling consistent boundary extraction for evaluation and analysis.</p>
</sec>
<sec id="s2-9">
<label>2.9</label>
<title>Evaluation metrics</title>
<p>Model performance was quantitatively assessed using four standard region-based metrics: Intersection over Union (IoU), Precision, Recall, and F1 score. All metrics were computed separately for the land and water classes using the manually annotated polygons as ground-truth reference. IoU was calculated as the ratio of the intersection to the union of predicted and ground-truth pixels for each class. Precision represents the proportion of predicted pixels that are correct, while Recall quantifies the proportion of ground-truth pixels that are correctly identified. The F1 score is the harmonic mean of Precision and Recall, summarizing segmentation accuracy for each class. These metrics were reported for both U-Net and DifFeat on the held-out evaluation set, allowing for direct comparison of segmentation performance across model types and spectral input combinations.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<label>3</label>
<title>Results</title>
<p>The following results summarize the performance of both segmentation approaches on land&#x2013;water classification tasks and derived boundary extraction in Arctic coastal imagery. The results are organized into three components: (1) quantitative evaluation of both models in different combinations of spectral inputs for land-water segmentation, (2) comparative analysis of segmentation quality, training efficiency, and boundary coherence between U-Net and DifFeat, and (3) evaluation of both models in large-scale and multisource imagery to evaluate cross-domain generalization.</p>
<sec id="s3-1">
<label>3.1</label>
<title>Quantitative evaluation of U-Net and DifFeat across spectral inputs</title>
<p>
<xref ref-type="table" rid="T1">Table 1</xref> summarizes the segmentation performance of U-Net and DifFeat in three spectral input combinations (RGB; RGB&#x2b;IR; RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI).</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Comparison of U-Net and DifFeat performance across input modalities using region-based metrics (<inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mi>&#x2191;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>: higher is better).</p>
</caption>
<table>
<thead valign="top">
<tr>
<th rowspan="2" align="left">Metric</th>
<th rowspan="2" align="left">Input image</th>
<th colspan="2" align="center">U-Net</th>
<th colspan="2" align="center">DifFeat</th>
</tr>
<tr>
<th align="center">Water</th>
<th align="center">Land</th>
<th align="center">Water</th>
<th align="center">Land</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="3" align="left">IoU <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mi>&#x2191;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">RGB</td>
<td align="center">0.16</td>
<td align="center">0.45</td>
<td align="center">0.92</td>
<td align="center">0.71</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR</td>
<td align="center">0.60</td>
<td align="center">0.43</td>
<td align="center">0.98</td>
<td align="center">0.87</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI</td>
<td align="center">0.58</td>
<td align="center">0.50</td>
<td align="center">0.95</td>
<td align="center">0.92</td>
</tr>
<tr>
<td rowspan="3" align="left">Precision <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:mi>&#x2191;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">RGB</td>
<td align="center">0.97</td>
<td align="center">0.45</td>
<td align="center">0.98</td>
<td align="center">0.96</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR</td>
<td align="center">0.72</td>
<td align="center">0.64</td>
<td align="center">0.99</td>
<td align="center">0.98</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI</td>
<td align="center">0.80</td>
<td align="center">0.65</td>
<td align="center">0.96</td>
<td align="center">0.97</td>
</tr>
<tr>
<td rowspan="3" align="left">Recall <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mi>&#x2191;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">RGB</td>
<td align="center">0.16</td>
<td align="center">0.99</td>
<td align="center">0.93</td>
<td align="center">0.73</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR</td>
<td align="center">0.77</td>
<td align="center">0.56</td>
<td align="center">0.98</td>
<td align="center">0.88</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI</td>
<td align="center">0.68</td>
<td align="center">0.69</td>
<td align="center">0.98</td>
<td align="center">0.94</td>
</tr>
<tr>
<td rowspan="3" align="left">F1 score <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>&#x2191;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="left">RGB</td>
<td align="center">0.28</td>
<td align="center">0.62</td>
<td align="center">0.95</td>
<td align="center">0.83</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR</td>
<td align="center">0.75</td>
<td align="center">0.60</td>
<td align="center">0.98</td>
<td align="center">0.93</td>
</tr>
<tr>
<td align="left">RGB&#x2b;IR&#x2b;NDWI&#x2b;NDVI</td>
<td align="center">0.74</td>
<td align="center">0.66</td>
<td align="center">0.97</td>
<td align="center">0.95</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s3-1-1">
<label>3.1.1</label>
<title>U-Net performance across spectral inputs</title>
<p>For the water class, U-Net performance improved markedly with additional spectral information. The IoU increased from 0.16 (RGB) to 0.60 (RGB&#x2b;IR), indicating that near-infrared bands enhanced discrimination of water from land. Using RGB inputs alone, the U-Net detected only the most distinct water regions, yielding very high precision (0.97) but low recall (0.16), indicating that many true water pixels were misclassified. Adding infrared bands increased sensitivity (recall &#x3d; 0.77) with moderately reduced precision (0.72), resulting in a much higher F1 score (0.28 <inline-formula id="inf17">
<mml:math id="m17">
<mml:mrow>
<mml:mo>&#x2192;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 0.75) and more balanced segmentation. Including NDWI and NDVI produced only minor or negative changes (IoU &#x3d; 0.58, F1 &#x3d; 0.74), suggesting redundancy among the derived indices and other spectral bands included in the model.</p>
<p>For the land class, U-Net achieved an IoU of 0.45 with RGB inputs and 0.43 with infrared band. Performance improved slightly with NDWI and NDVI (IoU &#x3d; 0.50, F1 &#x3d; 0.66), likely reflecting enhanced vegetation&#x2013;water contrast. Overall, U-Net relied strongly on spectral richness for accurate water segmentation, while land segmentation showed more modest gains.</p>
</sec>
<sec id="s3-1-2">
<label>3.1.2</label>
<title>DifFeat performance across spectral inputs</title>
<p>DifFeat, trained without labels, achieved consistently high segmentation accuracy across all input combinations. For the water class, performance was good with RGB inputs (IoU &#x3d; 0.92) and further improved with the addition of the infrared band (IoU &#x3d; 0.98), confirming that near-infrared information enhances separability between water and surrounding terrain. When NDWI and NDVI indices were included, performance decreased slightly (IoU &#x3d; 0.95, F1 &#x3d; 0.97) but remained well balanced, with precision and recall both exceeding 0.93.</p>
<p>For the land class, DifFeat maintained high and stable performance across all input combinations, with IoU increasing from 0.71 to 0.92 and F1 scores from 0.83 to 0.95, highlighting the benefit of incorporating additional spectral and index-based inputs. The model effectively utilized these complementary features, demonstrating robust generalization. Overall, DifFeat consistently outperformed U-Net under all spectral configurations, exhibiting superior adaptability to spectral variation and feature redundancy.</p>
</sec>
</sec>
<sec id="s3-2">
<label>3.2</label>
<title>Comparison of training efficiency and predictions of U-Net and DifFeat</title>
<p>The training times, parameters, and configurations for the U-Net and DifFeat clustering models are summarized in <xref ref-type="table" rid="T2">Table 2</xref>. U-Net contains approximately 1.943&#xa0;M learnable parameters, compared to 0.106&#xa0;M for DifFeat, reflecting the deeper encoder&#x2013;decoder structure of U-Net versus the shallower CNN in DifFeat. This makes DifFeat substantially lighter in terms of model complexity. The U-Net model required approximately 58&#xa0;min (3,538&#xa0;s) to complete 100 epochs using 528 labeled training images. In contrast, the DifFeat clustering model completed 100 epochs in just 4.61&#xa0;s on a single unlabeled image tile, a 99.87% reduction in training time. Both models were trained using <inline-formula id="inf18">
<mml:math id="m18">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>6</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> (six-band composite) input slices to ensure consistency in spectral inputs for comparison. These results highlight the substantial contrast in model size and training efficiency between the two approaches.</p>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Number of learnable parameters, total training time, and training settings for the evaluated models.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Model</th>
<th align="left">Parameters (M)</th>
<th align="left">Training time</th>
<th align="left">Settings</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">U-Net</td>
<td align="left">
<inline-formula id="inf19">
<mml:math id="m19">
<mml:mrow>
<mml:mo>&#x2248;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> 1.943</td>
<td align="left">58&#xa0;min/3,538&#xa0;s</td>
<td align="left">Epochs: 100; Images: 528</td>
</tr>
<tr>
<td align="left">DifFeat clustering</td>
<td align="left">
<inline-formula id="inf20">
<mml:math id="m20">
<mml:mrow>
<mml:mo>&#x2248;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> <bold>0.106</bold>
</td>
<td align="left">
<bold>4.61&#xa0;s</bold>
</td>
<td align="left">Epochs: 100; <bold>Images: 1</bold>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Training time indicates the total duration required to complete the full training process for the specified number of epochs and images. Epochs refers to one full pass through the training dataset, and images denotes to the number of training images used. Bold values highlight the most computationally and data-efficient values among the compared models.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>
<xref ref-type="fig" rid="F3">Figure 3</xref> presents qualitative comparisons between U-Net and DifFeat predictions across five representative Arctic coastal scenes. Each row (A&#x2013;E) shows a WV2 scene, the manually annotated land (green), water (blue), and background (cream) reference mask, and the predicted segmentation results from both models using different spectral inputs. Columns three and four show U-Net predictions trained on RGB and RGB &#x2b; IR &#x2b; NDWI &#x2b; NDVI inputs, respectively, while columns five and six display DifFeat predictions under the same input configurations. The labels &#x201c;S&#x201d; and &#x201c;B&#x201d; denote shoreline and bluff edge similarity scores, respectively, computed as the Intersection over Union (IoU) between each model&#x2019;s predicted boundaries and the ground-truth reference (shown for visualization only).</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Segmentation results from U-Net and DifFeat showing land (green), water (blue), and backgroud (cream) regions across five Arctic coastal scenes. Rows (<bold>A&#x2013;E</bold>) represent various test-set examples. Columns show false-color RGB input, ground truth reference masks, and predictions using RGB and RGB&#x002B;IR&#x002B;NDWI&#x002B;NDVI inputs. &#x201C;S&#x201D; and &#x201C;B&#x201D; indicate the IoU scores for the water and land classes, respectively, when comparing model predictions with manual annotations (ground truth).</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g003.tif">
<alt-text content-type="machine-generated">Segmentation results from U-Net and DifFeat showing land (green), water (blue), and backgroud (cream) regions across five Arctic coastal scenes. Rows (A&#x2013;E) represent various test-set examples. Columns show false-color RGB input, ground truth reference masks, and predictions using RGB and RGB&#x002B;IR&#x002B;NDWI&#x002B;NDVI inputs. &#x201c;S&#x201d; and &#x201c;B&#x201d; indicate the IoU scores for the water and land classes, respectively, when comparing model predictions with manual annotations (ground truth).</alt-text>
</graphic>
</fig>
<p>Across the examples, U-Net produced smooth, continuous land and water regions, closely following the manually annotated masks in most cases. Adding infrared and normalized indices improved delineation of subtle or low-contrast transitions, particularly where shallow water or moist tundra obscure spectral differences. In contrast, DifFeat generated more granular clusters with greater local variability, yet effectively captured fine-scale transition zones between land and water even without labeled training data. Multispectral inputs improved both models, yielding better spatial coherence and reduced false detections. To illustrate, Rows C and E of <xref ref-type="fig" rid="F3">Figure 3</xref> demonstrate the influence of surface texture and illumination, where both models exhibited small gaps or over-segmentation in darker regions when using RGB inputs. The addition of infrared and normalized indices mitigated these artifacts, improving consistency and reducing misclassification.</p>
<p>
<xref ref-type="fig" rid="F4">Figure 4</xref> presents the corresponding shoreline and bluff edge boundaries derived from the segmentation outputs shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. The first column shows the same WV2 input scenes for reference, while the subsequent panels display the extracted boundaries corresponding to each segmentation result in the same spatial arrangement. Blue lines represent the shoreline and orange lines represent the bluff edge, both automatically delineated from the predicted masks using the interface extraction algorithm described in <xref ref-type="sec" rid="s2-8">Section 2.8</xref>. The visual comparison highlights that U-Net boundaries were generally smoother and more continuous, whereas DifFeat boundaries were finer and often better aligned with subtle geomorphic transitions, though occasionally fragmented. When infrared and normalized indices were included, both models achieved improved alignment with the ground-truth boundaries with reduced false positives. Together, these results demonstrate that both models can generate accurate land/water segmentation, and corresponding shoreline/bluff edge boundaries, with U-Net favoring spatial smoothness and DifFeat excelling in fine-scale texture capture, providing complementary strengths for Arctic coastal monitoring.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Automatically extracted shoreline and bluff-edge boundaries derived from the land&#x2010;water segmentations. Rows <bold>(A&#x2013;E)</bold> represent various test-set examples. The first column shows the same WorldView-2 input scenes as in <xref ref-type="fig" rid="F3">Figure 3</xref> for reference, while the remaining columns display the corresponding predicted boundaries. Blue lines indicate shorelines, and orange lines indicate bluff edges.</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g004.tif">
<alt-text content-type="machine-generated">Automatically extracted shoreline and bluff-edge boundaries derived from the land-water segmentations. Rows (A&#x2013;E) represent various test-set examples. The first column shows the same WorldView-2 input scenes as in Figure 3 for reference, while the remaining columns display the corresponding predicted boundaries. Blue lines indicate shorelines, and orange lines indicate bluff edges.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-3">
<label>3.3</label>
<title>Evaluation on large-scale and multi-source imagery</title>
<sec id="s3-3-1">
<label>3.3.1</label>
<title>DifFeat predictions on WV2 satellite imagery</title>
<p>
<xref ref-type="fig" rid="F5">Figure 5A</xref> shows the full WV2 scene with shoreline and bluff edge boundaries automatically extracted from the DifFeat land and water segmentation results. A zoomed-in view in <xref ref-type="fig" rid="F5">Figure 5B</xref> (approximately 2&#xa0;km in length) demonstrates close spatial alignment between the extracted boundaries and the locations where these landforms would be expected, based on expert visual interpretation. However, some localized errors remain in the derived boundary positions. For example, <xref ref-type="fig" rid="F5">Figure 5C</xref> highlights a region where the extracted interfaces are offset by a few pixels, resulting in water sediments being misclassified as bluff edges and wave patterns interpreted as shoreline. In <xref ref-type="fig" rid="F5">Figure 5D</xref>, both the predicted bluff edge and the shoreline are spatially offset from the reference boundaries, and the shoreline shows a larger misalignment. Conversely, <xref ref-type="fig" rid="F5">Figure 5E</xref> shows a correctly delineated bluff edge, but with a minor misalignment of the shoreline.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>DifFeat model predictions on a WV2 image scene for two target classes: bluff edge (orange) and shoreline (blue). <bold>(A)</bold> The map and insets are shown after post-processing to improve visual and spatial coherence. Arrows indicate regions of interest highlighted in sub-figures <bold>(B&#x2013;E)</bold>. <bold>(B)</bold> Shows areas where predicted boundaries for both classes align with the reference. <bold>(C)</bold> Shows regions where predictions for both classes do not match the reference. <bold>(D)</bold> Highlights a region where the predicted bluff edge and shoreline are spatially offset from the reference, with a larger misalignment for the shoreline. <bold>(E)</bold> Illustrates a minor misaligned shoreline prediction. Line widths differ for each class to enhance clarity.</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g005.tif">
<alt-text content-type="machine-generated">DifFeat model predictions on a WV2 image scene for two target classes: bluff edge (orange) and shoreline (blue). (A) The map and insets are shown after post-processing to improve visual and spatial coherence. Arrows indicate regions of interest highlighted in sub-figures (B&#x2013;E). (B) Shows areas where predicted boundaries for both classes align with the reference. (C) Shows regions where predictions for both classes do not match the reference. (D) Highlights a region where the predicted bluff edge and shoreline are spatially offset from the reference, with a larger misalignment for the shoreline. (E) Illustrates a minor misaligned shoreline prediction. Line widths differ for each class to enhance clarity.</alt-text>
</graphic>
</fig>
<p>Overall, the derived boundaries correspond well to the expected shoreline and bluff edge positions. This alignment is illustrated in <xref ref-type="sec" rid="s11">Supplementary Figure S1</xref>, where the post-processed DifFeat predictions are overlaid on manually annotated features for visual comparison. It is important to note that manual annotations were digitized as land and water polygons, where one boundary edge represents the actual shoreline or bluff edge. As a result, model predictions align with the intended feature edge, rather than the polygon centroid or area, as illustrated in <xref ref-type="sec" rid="s11">Supplementary Figure S1</xref>. <xref ref-type="sec" rid="s11">Supplementary Figures S2&#x2013;S4</xref> provide additional examples of raw and cleaned predictions, spatial discontinuities, and clustering behavior. In tundra regions, the model occasionally misclassified polygonal ice-wedge features, surface cracks, and high-centered polygons as bluff edges (see <xref ref-type="sec" rid="s11">Supplementary Figure S3</xref> for examples of these misclassifications, panels D&#x2013;F). Boundary artifacts were also observed along the edges of water bodies and at tile borders, where portions of open water were sometimes interpreted as shoreline. These discontinuities occurred even when slice overlap was introduced during inference.</p>
</sec>
<sec id="s3-3-2">
<label>3.3.2</label>
<title>U-Net predictions on WV2 satellite imagery</title>
<p>U-Net land and water segmentation results for the large-scale WV2 scene are shown in <xref ref-type="sec" rid="s11">Supplementary Figure S5</xref> (top: extracted bluff edge boundaries; bottom: extracted shoreline boundaries), using models trained on different combinations of input bands.</p>
<p>For bluff edge boundaries (<xref ref-type="sec" rid="s11">Supplementary Figure S5</xref>, top), U-Net trained with RGB &#x2b; IR &#x2b; NDVI &#x2b; NDWI produced a continuous but noisy band that generally aligned with the actual bluff, though occasional extensions into the shoreline or water occurred. The RGB &#x2b; IR model, however, struggled with turbid water, highlighting the benefit of additional spectral indices. For shoreline boundaries (<xref ref-type="sec" rid="s11">Supplementary Figure S5</xref>, bottom), the RGB &#x2b; IR &#x2b; NDVI &#x2b; NDWI model produced sparse outputs with scattered noise over water, while the RGB &#x2b; IR model generated more continuous, though noisier, boundary segments that extended into the tundra. Both models showed limited improvement from the six-band composite.</p>
<p>This results contrasts with the prediction from smaller, mixed land-water patches shown in <xref ref-type="fig" rid="F3">Figure 3</xref>, where U-Net produces smoother segmentation, especially in land regions. However, the model&#x2019;s performance deteriorates in homogeneous patches, such as those containing only land or only water (<xref ref-type="sec" rid="s11">Supplementary Figure S7</xref>, rows 1&#x2013;3). In these regions, U-Net often generates spurious segments of the opposite class (e.g., isolated water pixels in land-only patches or land pixels in water-only patches). This issue is prominent in large-scale imagery, where many tiles are homogeneous. In such cases, the increased false positives accumulate, resulting in fragmented and noisy boundaries in the raw segmentation output. This behavior is mainly due to U-Net&#x2019;s challenge in accurately classifying pure, single-class regions, leading to spurious predictions that disrupt the overall segmentation.</p>
</sec>
<sec id="s3-3-3">
<label>3.3.3</label>
<title>DifFeat predictions on drone imagery</title>
<p>The land&#x2013;water segmentation outputs and extracted boundaries for the UAV scene are shown in <xref ref-type="fig" rid="F6">Figure 6A</xref>, with derived bluff edge boundaries highlighted in green. Across the image, the model consistently produced land&#x2013;water separations that, after interface extraction, yielded bluff edge lines in strong visual agreement with reference boundaries. <xref ref-type="fig" rid="F6">Figure 6B</xref> shows a zoomed-in region where the extracted bluff edge boundary aligns closely with the manually delineated reference. However, in <xref ref-type="fig" rid="F6">Figure 6C</xref>, tundra features are occasionally misclassified as bluff edges in the segmentation mask, reflecting generalization limitations. The model did not produce distinct shoreline boundaries from the UAV segmentation, likely due to weaker spectral contrast between water and adjacent tundra surfaces in the downsampled UAV data. This limitation may also relate to differences in image resolution, spectral response, or surface conditions unique to the UAV dataset.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Evaluation of the DifFeat model on a UAV-dervied drone orthomosaic. <bold>(A)</bold> Bluff edge predictions made by the DifFeat model trained on WV2 image slices, shown on UAV drone orthoimage of the same site. <bold>(B,C)</bold> highlight regions of correct and incorrect predictions, where the model&#x2019;s performance varies across different sections.</p>
</caption>
<graphic xlink:href="fenvs-13-1657984-g006.tif">
<alt-text content-type="machine-generated">Evaluation of the DifFeat model on a UAV-dervied drone orthomosaic. (A) Bluff edge predictions made by the DifFeat model trained on WV2 image slices, shown on UAV drone orthoimage of the same site. (B,C) highlight regions of correct and incorrect predictions, where the model&#x2019;s performance varies across different sections.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-3-4">
<label>3.3.4</label>
<title>U-Net predictions on drone imagery</title>
<p>The U-Net land&#x2013;water segmentation model, trained on annotated WV2 satellite imagery, was evaluated on UAV orthomosaics to test cross-domain generalization. <xref ref-type="sec" rid="s11">Supplementary Figure S6</xref> shows the derived boundaries extracted from the segmentation outputs, with blue representing the shoreline and orange representing the bluff edge. The shoreline boundaries were sparse and fragmented, appearing as short, discontinuous line segments separated by large unclassified regions. These extracted boundaries did not consistently follow the true shoreline and were often interrupted by gaps, with false positives scattered across tundra areas and over water surfaces exhibiting spectral variation. Bluff edge boundaries were largely absent; the few segments that appeared were fragmented and confined mostly to image margins or corners, without forming coherent or continuous lines.</p>
</sec>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>In this study, we performed land and water segmentation of Arctic coastal imagery using both a supervised deep learning model (U-Net) and an unsupervised clustering method, DifFeat. U-Net serves as a representative supervised baseline widely used in coastal studies, while DifFeat demonstrates the potential to reduce dependence on annotated data in data-sparse environments. From the resulting land and water masks, shoreline and bluff edge boundaries were automatically derived using an interface extraction algorithm. DifFeat was trained entirely without labeled data and minimal manual supervision was employed for cluster selection, where an expert identified the clusters corresponding to land and water. To our knowledge, this is the first comparison of supervised and unsupervised approaches for these geomorphic boundaries in an Arctic setting, providing a foundation for automated coastal change detection. Beyond shoreline mapping, our study also derives another coastal morphological feature, the bluff edge, which is often overlooked but serves as a critical indicator for coastal change analysis and land-ocean exchange.</p>
<p>U-Net established a reliable supervised baseline for land and water segmentation, performing consistently across spectral inputs when applied to WV-2 imagery. The inclusion of the infrared (IR) band substantially improved water segmentation accuracy by enhancing contrast between tundra and open water, while NDVI and NDWI contributed minor or inconsistent gains. Land segmentation remained comparatively stable and less sensitive to spectral augmentation, indicating that land surfaces were easier to classify than water. When applied to UAV imagery, U-Net produced sparse and fragmented segmentation masks, reflecting limited cross-domain generalization from satellite-trained models. While the U-Net baseline provided a useful comparative model for tile-wise training and evaluation, it did not generalize well to full-scene or cross-domain imagery, producing scattered and noisy predictions. This is due to U-Net&#x2019;s performance on mixed and homogeneous patches, which becomes critical when scaling to whole-scene analysis. Since a larger proportion of tiles in full-scene WV-2 imagery are homogeneous, the noise accumulates, becoming more prominent in large-scale outputs, leading to fragmented and unreliable boundaries.</p>
<p>DifFeat demonstrated strong and consistent performance for land and water segmentation, maintaining stable precision and recall across all spectral input combinations. This stability indicates that the model effectively learned discriminative spectral&#x2013;spatial features without supervision and remained robust to input variability. Water segmentation benefited from the inclusion of infrared information, while vegetation and water indices (NDVI, NDWI) provided smaller but consistent improvements for land classification. The outputs were spatially coherent and boundary-aligned, producing clean masks from which shoreline and bluff edge boundaries were automatically extracted using the interface extraction algorithm. Overall, DifFeat&#x2019;s consistent performance across sensors and spectral inputs highlights its potential as a label-efficient and generalizable alternative to supervised segmentation in data-limited Arctic environments. Its unsupervised clustering framework can be readily extended to other remote sensing applications, where grouping pixels with similar spatial-spectral characteristics enables efficient mapping of diverse surface features without the need for retraining or labeled data.</p>
<p>Beyond segmentation accuracy, DifFeat&#x2019;s efficiency is a notable strength. Training for the DifFeat model was completed in just a few seconds for a single <inline-formula id="inf21">
<mml:math id="m21">
<mml:mrow>
<mml:mn>512</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> image tile, and large-scene inference required minimal computational and manual effort. The training time is 99.87% shorter than that of U-Net. Despite this simplicity, the outputs were visually coherent and directly suitable for downstream refinement. This efficiency opens opportunities to use DifFeat outputs as sparse supervisory labels or enriched input features for supervised models like U-Net, reducing the need for extensive manual annotation while improving boundary precision (<xref ref-type="bibr" rid="B6">Clark et al., 2022</xref>). In data-scarce Arctic settings where fine-grained delineation is essential, such a hybrid strategy offers a scalable and label-efficient pathway for future coastal monitoring.</p>
<p>Since the training data contained only land and water labels, the shoreline and bluff edge were implicitly represented within a shared boundary. As a result, U-Net learned coarse class separations rather than the precise geomorphic interfaces needed for accurate boundary delineation. Although U-Net serves as a strong supervised baseline, recovering fine spatial details, especially at object boundaries, remains challenging (<xref ref-type="bibr" rid="B35">Zhou et al., 2023</xref>). Future work could explore boundary-aware or attention-enhanced U-Net variants to improve localization along geomorphic boundaries. More precise, line-based annotations could also benefit both training and evaluation, though such data remain difficult to obtain at large scales.</p>
<p>DifFeat predictions, meanwhile, require post-processing to mitigate tiling artifacts and occasional tundra misclassifications. These issues likely stem from local variability in cluster assignments and the absence of topographic context. Future work should therefore focus on automating cluster selection to improve spatial consistency across tiles and on integrating auxiliary datasets such as DEMs, which may enhance bluff edge detection by capturing subtle elevation transitions. Incorporating DEMs could likewise benefit supervised models such as U-Net by providing complementary elevation cues for boundary refinement. A further limitation is that both models were trained solely on cloud-free, open-water season imagery, limiting their generalization to imagery acquired under more challenging conditions which could include residual snow, or sea ice cover, or cloud shadows. Incorporating post-processing workflows for cloud/shadow removal or spectral inpainting could help address these limitations.</p>
<p>In summary, this study demonstrates supervised and unsupervised segmentation approaches as complementary tools for Arctic coastal monitoring. U-Net offers structured, annotation-driven learning, whereas DifFeat enables scalable, data-efficient mapping without manual labels. DifFeat&#x2019;s core strength lies in grouping pixels with similar spectral and spatial characteristics into semantically meaningful clusters at varying levels of granularity, which can then be interpreted and assigned semantic labels. Together, these approaches advance automated extraction of shoreline and bluff edge features, two key indicators of permafrost-driven land loss, establishing a foundation for scalable Arctic coastal monitoring and supporting future efforts in climate adaptation, risk assessment, and community resilience across rapidly changing Arctic permafrost landscapes.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="sec" rid="s11">Supplementary Material</xref>, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>HB: Writing &#x2013; review and editing, Formal Analysis, Writing &#x2013; original draft, Software, Validation, Conceptualization, Methodology, Visualization. SV: Conceptualization, Funding acquisition, Writing &#x2013; review and editing, Visualization, Validation, Data curation, Supervision, Methodology. OF: Supervision, Methodology, Project administration, Software, Formal Analysis, Resources, Validation, Writing &#x2013; review and editing. SP: Data curation, Validation, Writing &#x2013; review and editing. CT: Funding acquisition, Writing &#x2013; review and editing.</p>
</sec>
<ack>
<title>Acknowledgements</title>
<p>The authors would like to thank the Utqia&#x121;vik I&#xf1;upiat Corporation (UIC) Science and the I&#xf1;upiat people of the Utqia&#x121;vik community for their continued support. Additionally, we thank past and present members of the Systems Ecology Lab (SEL) for help with data collection. Geospatial support for this work was provided by the Polar Geospatial Center (PGC) under NSF-OPP awards 1043681, 1559691, and 2129685. The authors acknowledge the use of ChatGPT (version 4, OpenAI) for assistance in grammar checking, rephrasing, and improving the overall readability of the manuscript.</p>
</ack>
<sec sec-type="COI-statement" id="s8">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The authors declare that Generative AI was used in the creation of this manuscript. Generative AI, in particular, ChatGPT (version 4, OpenAI) was used for assistance in grammar checking, rephrasing, and improving the overall readability of the manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s11">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fenvs.2025.1657984/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fenvs.2025.1657984/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="Supplementaryfile1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<fn-group>
<fn fn-type="custom" custom-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/196803/overview">Peng Liu</ext-link>, Chinese Academy of Sciences (CAS), China</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1099896/overview">Dustin Whalen</ext-link>, Geological Survey of Canada, Canada</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3141937/overview">Giulio Passerotti</ext-link>, The University of Melbourne Department of Computing and Information Systems, Australia</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3146927/overview">Heidi Rodenhizer</ext-link>, Woods Hole Research Center, United States</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aryal</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Escarzaga</surname>
<given-names>S. M.</given-names>
</name>
<name>
<surname>Vargas Zesati</surname>
<given-names>S. A.</given-names>
</name>
<name>
<surname>Velez-Reyes</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Fuentes</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Tweedie</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Semi-automated semantic segmentation of arctic shorelines using very high-resolution airborne imagery, spectral indices and weakly supervised machine learning approaches</article-title>. <source>Remote Sens.</source> <volume>13</volume>, <fpage>4572</fpage>. <pub-id pub-id-type="doi">10.3390/rs13224572</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bartsch</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ley</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Nitze</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Pointner</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Vieira</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Feasibility study for the application of synthetic aperture radar for coastal erosion rate quantification across the arctic</article-title>. <source>Front. Environ. Sci.</source> <volume>8</volume>, <fpage>143</fpage>. <pub-id pub-id-type="doi">10.3389/fenvs.2020.00143</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Beamish</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Raynolds</surname>
<given-names>M. K.</given-names>
</name>
<name>
<surname>Epstein</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Frost</surname>
<given-names>G. V.</given-names>
</name>
<name>
<surname>Macander</surname>
<given-names>M. J.</given-names>
</name>
<name>
<surname>Bergstedt</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Recent trends and remaining challenges for optical remote sensing of arctic tundra vegetation: a review and outlook</article-title>. <source>Remote Sens. Environ.</source> <volume>246</volume>, <fpage>111872</fpage>. <pub-id pub-id-type="doi">10.1016/j.rse.2020.111872</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bengoufa</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Niculescu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mihoubi</surname>
<given-names>M. K.</given-names>
</name>
<name>
<surname>Belkessa</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Rami</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Rabehi</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Machine learning and shoreline monitoring using optical satellite images: case study of the Mostaganem shoreline, Algeria</article-title>. <source>J. Applied Remote Sensing</source> <volume>15</volume>, <fpage>026509</fpage>. <pub-id pub-id-type="doi">10.1117/1.jrs.15.026509</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cassidy</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wiseman</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lange</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Eilers</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Bradley</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Seasonal coastal erosion rates calculated from planetscope imagery in arctic Alaska</article-title>. <source>Remote Sens.</source> <volume>16</volume>, <fpage>2365</fpage>. <pub-id pub-id-type="doi">10.3390/rs16132365</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Clark</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Moorman</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Whalen</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Vieira</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Multiscale object-based classification and feature extraction along arctic coasts</article-title>. <source>Remote Sens.</source> <volume>14</volume>, <fpage>2982</fpage>. <pub-id pub-id-type="doi">10.3390/rs14132982</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Creel</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Guimond</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>B. M.</given-names>
</name>
<name>
<surname>Nielsen</surname>
<given-names>D. M.</given-names>
</name>
<name>
<surname>Bristol</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Tweedie</surname>
<given-names>C. E.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>Permafrost thaw subsidence, sea-level rise, and erosion are transforming alaska&#x2019;s arctic coastal zone</article-title>. <source>Proc. Natl. Acad. Sci.</source> <volume>121</volume>, <fpage>e2409411121</fpage>. <pub-id pub-id-type="doi">10.1073/pnas.2409411121</pub-id>
<pub-id pub-id-type="pmid">39625976</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dang</surname>
<given-names>K. B.</given-names>
</name>
<name>
<surname>Vu</surname>
<given-names>K. C.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>T. D. L.</given-names>
</name>
<name>
<surname>Pham</surname>
<given-names>T. P. N.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Application of deep learning models to detect coastlines and shorelines</article-title>. <source>J. Environ. Manag.</source> <volume>320</volume>, <fpage>115732</fpage>. <pub-id pub-id-type="doi">10.1016/j.jenvman.2022.115732</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Efimova</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bartsch</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pointner</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Arctic coastline mapping with sentinel-2 data</article-title>. <source>AGU Fall Meet. Abstr.</source> <volume>2020</volume>, <fpage>C003</fpage>&#x2013;<lpage>C008</lpage>.</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<collab>European Space Agency</collab> (<year>2025</year>). <article-title>Worldview-2 mission overview</article-title>.</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fritz</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Vonk</surname>
<given-names>J. E.</given-names>
</name>
<name>
<surname>Lantuit</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Collapsing arctic coastlines</article-title>. <source>Nat. Clim. Change</source> <volume>7</volume>, <fpage>6</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1038/nclimate3188</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gabarr&#xf3;</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Hughes</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Wilkinson</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Bertino</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Bracher</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Diehl</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>Improving satellite-based monitoring of the polar regions: identification of research and capacity gaps</article-title>. <source>Front. Remote Sens.</source> <volume>4</volume>, <fpage>952091</fpage>. <pub-id pub-id-type="doi">10.3389/frsen.2023.952091</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Heidler</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Mou</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Baumhoer</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Dietz</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>X. X.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Hed-unet: combined segmentation and edge detection for monitoring the antarctic coastline</article-title>. <source>IEEE Transactions Geoscience Remote Sensing</source> <volume>60</volume>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1109/tgrs.2021.3064606</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Irrgang</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Bendixen</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Farquharson</surname>
<given-names>L. M.</given-names>
</name>
<name>
<surname>Baranskaya</surname>
<given-names>A. V.</given-names>
</name>
<name>
<surname>Erikson</surname>
<given-names>L. H.</given-names>
</name>
<name>
<surname>Gibbs</surname>
<given-names>A. E.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Drivers, dynamics and impacts of changing arctic coasts</article-title>. <source>Nat. Rev. Earth Environ.</source> <volume>3</volume>, <fpage>39</fpage>&#x2013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1038/s43017-021-00232-1</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jones</surname>
<given-names>B. M.</given-names>
</name>
<name>
<surname>Arp</surname>
<given-names>C. D.</given-names>
</name>
<name>
<surname>Whitman</surname>
<given-names>M. S.</given-names>
</name>
<name>
<surname>Grosse</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Nitze</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>A decade of remotely sensed observations highlight complex processes linked to coastal permafrost bluff erosion in the arctic</article-title>. <source>Geophys. Res. Lett.</source> <volume>45</volume>, <fpage>4872</fpage>&#x2013;<lpage>4881</lpage>.</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Juma</surname>
<given-names>G. A.</given-names>
</name>
<name>
<surname>Meunier</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Herstoff</surname>
<given-names>E. M.</given-names>
</name>
<name>
<surname>Irrgang</surname>
<given-names>A. M.</given-names>
</name>
<name>
<surname>Fritz</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Weber</surname>
<given-names>C.</given-names>
</name>
<etal/>
</person-group> (<year>2025</year>). <article-title>Future arctic: how will increasing coastal erosion shape nearshore planktonic food webs?</article-title> <source>Limnol. Oceanogr. Lett.</source> <volume>10</volume>, <fpage>5</fpage>&#x2013;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.1002/lol2.10446</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kim</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Kanezaki</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Tanaka</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Unsupervised learning of image segmentation based on differentiable feature clustering</article-title>. <source>IEEE Trans. Image Process.</source> <volume>29</volume>, <fpage>8055</fpage>&#x2013;<lpage>8068</lpage>. <pub-id pub-id-type="doi">10.1109/tip.2020.3011269</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kirillov</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Mintun</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Ravi</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Mao</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Rolland</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Gustafson</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). &#x201c;<article-title>Segment anything</article-title>,&#x201d; in <source>Proceedings of the IEEE/CVF international conference on computer vision</source>, <fpage>4015</fpage>&#x2013;<lpage>4026</lpage>.</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>McAllister</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Payo</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Novellino</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Dolphin</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Medina-Lopez</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Multispectral satellite imagery and machine learning for the extraction of shoreline indicators</article-title>. <source>Coast. Eng.</source> <volume>174</volume>, <fpage>104102</fpage>. <pub-id pub-id-type="doi">10.1016/j.coastaleng.2022.104102</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>McFeeters</surname>
<given-names>S. K.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>The use of the normalized difference water index (ndwi) in the delineation of open water features</article-title>. <source>Int. J. Remote Sens.</source> <volume>17</volume>, <fpage>1425</fpage>&#x2013;<lpage>1432</lpage>. <pub-id pub-id-type="doi">10.1080/01431169608948714</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Miller</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Bonsell</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>McTigue</surname>
<given-names>N. D.</given-names>
</name>
<name>
<surname>Kelley</surname>
<given-names>A. L.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>The seasonal phases of an arctic lagoon reveal the discontinuities of ph variability and co<sub>2</sub> flux at the air&#x2013;sea interface</article-title>. <source>Biogeosciences</source> <volume>18</volume>, <fpage>1203</fpage>&#x2013;<lpage>1221</lpage>. <pub-id pub-id-type="doi">10.5194/bg-18-1203-2021</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<collab>National Snow and Ice Data Center</collab> (<year>2023</year>). <article-title>Frozen ground and permafrost</article-title>.</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nelson</surname>
<given-names>P. R.</given-names>
</name>
<name>
<surname>Maguire</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Pierrat</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Orcutt</surname>
<given-names>E. L.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Serbin</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Remote sensing of tundra ecosystems using high spectral resolution reflectance: opportunities and challenges</article-title>. <source>J. Geophys. Res. Biogeosciences</source> <volume>127</volume>, <fpage>e2021JG006697</fpage>. <pub-id pub-id-type="doi">10.1029/2021jg006697</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nielsen</surname>
<given-names>D. M.</given-names>
</name>
<name>
<surname>Pieper</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Barkhordarian</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Overduin</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Ilyina</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Brovkin</surname>
<given-names>V.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Increase in arctic coastal erosion and its sensitivity to warming in the twenty-first century</article-title>. <source>Nat. Clim. Change</source> <volume>12</volume>, <fpage>263</fpage>&#x2013;<lpage>270</lpage>. <pub-id pub-id-type="doi">10.1038/s41558-022-01281-0</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nielsen</surname>
<given-names>D. M.</given-names>
</name>
<name>
<surname>Chegini</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Maerz</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Brune</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Mathis</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Dobrynin</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>Reduced Arctic Ocean co2 uptake due to coastal permafrost erosion</article-title>. <source>Nat. Clim. Change</source> <volume>14</volume>, <fpage>968</fpage>&#x2013;<lpage>975</lpage>. <pub-id pub-id-type="doi">10.1038/s41558-024-02074-3</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Park</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Shoreline change analysis with deep learning semantic segmentation using remote sensing and gis data</article-title>. <source>KSCE J. Civ. Eng.</source> <volume>28</volume>, <fpage>928</fpage>&#x2013;<lpage>938</lpage>. <pub-id pub-id-type="doi">10.1007/s12205-023-1604-9</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Philipp</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Dietz</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ullmann</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kuenzer</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Automated extraction of annual erosion rates for arctic permafrost coasts using sentinel-1, deep learning, and change vector analysis</article-title>. <source>Remote Sens.</source> <volume>14</volume>, <fpage>3656</fpage>. <pub-id pub-id-type="doi">10.3390/rs14153656</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rantanen</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Karpechko</surname>
<given-names>A. Y.</given-names>
</name>
<name>
<surname>Lipponen</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Nordling</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Hyv&#xe4;rinen</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Ruosteenoja</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>The arctic has warmed nearly four times faster than the globe since 1979</article-title>. <source>Commun. Earth Environment</source> <volume>3</volume>, <fpage>168</fpage>. <pub-id pub-id-type="doi">10.1038/s43247-022-00498-3</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>U-net: convolutional networks for biomedical image segmentation</article-title>,&#x201d; in <source>Medical image computing and computer-assisted intervention (MICCAI)</source> (<publisher-name>Springer</publisher-name>), <fpage>234</fpage>&#x2013;<lpage>241</lpage>.</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Rouse</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Haas</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Schell</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Deering</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>1974</year>). &#x201c;<article-title>Monitoring vegetation systems in the Great Plains with erts</article-title>,&#x201d; in <source>Third Earth resources technology Satellite-1 symposium</source> (<publisher-name>Technical presentations NASA</publisher-name>), <volume>1</volume>, <fpage>309</fpage>&#x2013;<lpage>317</lpage>.</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scala</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Manno</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Ciraolo</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Semantic segmentation of coastal aerial/satellite images using deep learning techniques: an application to coastline detection</article-title>. <source>Comput. Geosciences</source> <volume>192</volume>, <fpage>105704</fpage>. <pub-id pub-id-type="doi">10.1016/j.cageo.2024.105704</pub-id>
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schuur</surname>
<given-names>E. A.</given-names>
</name>
<name>
<surname>McGuire</surname>
<given-names>A. D.</given-names>
</name>
<name>
<surname>Sch&#xe4;del</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Grosse</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Harden</surname>
<given-names>J. W.</given-names>
</name>
<name>
<surname>Hayes</surname>
<given-names>D. J.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>Climate change and the permafrost carbon feedback</article-title>. <source>Nature</source> <volume>520</volume>, <fpage>171</fpage>&#x2013;<lpage>179</lpage>. <pub-id pub-id-type="doi">10.1038/nature14338</pub-id>
<pub-id pub-id-type="pmid">25855454</pub-id>
</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wenzl</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Baumhoer</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Dietz</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Kuenzer</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Vegetation changes in the arctic: a review of earth observation applications</article-title>. <source>Remote Sens.</source> <volume>16</volume>, <fpage>4509</fpage>. <pub-id pub-id-type="doi">10.3390/rs16234509</pub-id>
</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hong</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Sea-land segmentation using deep learning techniques for landsat-8 oli imagery</article-title>. <source>Mar. Geod.</source> <volume>43</volume>, <fpage>105</fpage>&#x2013;<lpage>133</lpage>. <pub-id pub-id-type="doi">10.1080/01490419.2020.1713266</pub-id>
</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>An overview of coastline extraction from remote sensing data</article-title>. <source>Remote Sens.</source> <volume>15</volume>, <fpage>4865</fpage>. <pub-id pub-id-type="doi">10.3390/rs15194865</pub-id>
</mixed-citation>
</ref>
</ref-list>
</back>
</article>