<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Remote Sens.</journal-id>
<journal-title>Frontiers in Remote Sensing</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Remote Sens.</abbrev-journal-title>
<issn pub-type="epub">2673-6187</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1521446</article-id>
<article-id pub-id-type="doi">10.3389/frsen.2025.1521446</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Remote Sensing</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Automated depth correction of bathymetric LiDAR point clouds using PointCNN semantic segmentation</article-title>
<alt-title alt-title-type="left-running-head">Paul et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/frsen.2025.1521446">10.3389/frsen.2025.1521446</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Paul</surname>
<given-names>Ovi</given-names>
</name>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2879475/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Ekhtari</surname>
<given-names>Nima</given-names>
</name>
<uri xlink:href="https://loop.frontiersin.org/people/3157349/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Glennie</surname>
<given-names>Craig L.</given-names>
</name>
<uri xlink:href="https://loop.frontiersin.org/people/1051559/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
</contrib>
</contrib-group>
<aff>
<institution>Department of Civil and Environmental Engineering</institution>, <institution>National Center for Airborne Laser Mapping (NCALM)</institution>, <institution>University of Houston</institution>, <addr-line>Houston</addr-line>, <addr-line>TX</addr-line>, <country>United States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/149987/overview">Eduardo Landulfo</ext-link>, Instituto de Pesquisas Energ&#xe9;ticas e Nucleares (IPEN), Brazil</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2139450/overview">Yaohui Liu</ext-link>, Shandong Jianzhu University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2349795/overview">Xin Wu</ext-link>, Beijing University of Posts and Telecommunications (BUPT), China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Ovi Paul, <email>opaul@uh.edu</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>15</day>
<month>08</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>6</volume>
<elocation-id>1521446</elocation-id>
<history>
<date date-type="received">
<day>01</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>25</day>
<month>07</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Paul, Ekhtari and Glennie.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Paul, Ekhtari and Glennie</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>The study explores deep learning to perform direct semantic segmentation of bathymetric lidar points to improve bathymetry mapping. Focusing on river bathymetry, the goal is to accurately and simultaneously classify points on the benthic layer, water surface, and ground near riverbanks. These classifications are then used to apply depth correction to all points within the water column. The study aimed to classify the scene into four classes: river surface, riverbed, ground, and other (for points outside of those three classes), focusing on the river surface and riverbed classes. To achieve this, PointCNN, a convolutional neural network model adept at handling unorganized and unstructured data in 3D space was implemented. The model was trained with airborne bathymetric lidar data from the Swan River in Montana and the Eel River in California. The model was tested on the Snake River in Wyoming to evaluate its performance. These diverse bathymetric datasets across the United States provided a solid foundation for the model&#x2019;s robust testing. The results were strong for river surface classification, achieving an Intersection over Union of (0.89) and a Kappa coefficient of (0.92), indicating high reliability and minimal errors. The riverbed classification also showed an IoU of (0.7) and a slightly higher Kappa score of (0.76). Depth correction was then performed on riverbed points, proportional to the calculated depth from a surface model formed by Delaunay triangulation of ground and river surface points. The automated process performs significantly faster than traditional manual classification and depth correction processes, saving time and expense. Finally, corrected depths were quantitatively validated by comparing with independent Acoustic Doppler Current Profiler measurements from the Snake River, obtaining a mean depth error of 2&#xa0;cm and an Root mean square error of 16&#xa0;cm. These validation results show the reliability and accuracy of the proposed automated bathymetric depth correction workflow.</p>
</abstract>
<kwd-group>
<kwd>bathymetric lidar</kwd>
<kwd>deep learning</kwd>
<kwd>3D convolutional neural network</kwd>
<kwd>graph neural network</kwd>
<kwd>semantic segmentation</kwd>
<kwd>PointCNN</kwd>
<kwd>riverbed depth correction</kwd>
</kwd-group>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Lidar Sensing</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Lidar systems produce 3D point clouds, providing high-resolution data essential for geoscientists to analyze natural and artificial features. These clouds consist of millions of individual points, each with x, y, and z coordinates and possibly other features like intensity and RGB; collectively forming a comprehensive 3D model of the surveyed area. Aerial bathymetric lidar systems are designed to map underwater features using laser pulses that can penetrate water, normally operating at a wavelength of <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>532</mml:mn>
<mml:mtext>&#x2009;nm</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula> (green). Bathymetric lidar sensors provide high-resolution data of submerged environments, precisely documenting shallow underwater terrain (<xref ref-type="bibr" rid="B40">Pan et al., 2015a</xref>). An accurate digital model of the benthic layer is crucial for mapping habitats in fluvial and coastal areas (<xref ref-type="bibr" rid="B30">Letard et al., 2022</xref>), enhancing flood inundation models (<xref ref-type="bibr" rid="B52">Teng et al., 2015</xref>), tracking sediment transport (<xref ref-type="bibr" rid="B4">Anderson and Pitlick, 2014</xref>), and for nearshore bathymetry (<xref ref-type="bibr" rid="B3">Albright and Glennie, 2021</xref>). It is invaluable for coastal management to monitor erosion, identify submerged hazards (<xref ref-type="bibr" rid="B43">Pricope and Bashit, 2023</xref>), guide infrastructure planning, and conserve ecosystems like coral reefs and seagrasses (<xref ref-type="bibr" rid="B59">Wilson, 2011</xref>; <xref ref-type="bibr" rid="B39">Pan et al., 2014</xref>).</p>
<p>Processing bathymetric lidar data has significant challenges, primarily due to the refraction of the laser pulses in water. While shifts in the <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> coordinates are minor and dependent upon the scan angle, the <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> coordinate (elevation) shifts are more pronounced. This is because water, with a nominal refractive index of approximately 1.33 (<xref ref-type="bibr" rid="B61">Xiao et al., 2024</xref>), bends and slows the light, distorting true depths because the travel time of the laser pulses in water is unknown without precise knowledge of both the water surface and benthic layer locations (<xref ref-type="bibr" rid="B2">Agrafiotis et al., 2019</xref>; <xref ref-type="bibr" rid="B48">Schwarz et al., 2021</xref>). Fluvial environments face significant challenges due to fluctuating water elevations along a river&#x2019;s course, especially downstream, where levels can drop rapidly (<xref ref-type="bibr" rid="B35">Mandlburger et al., 2015</xref>). Dams exacerbate this, causing abrupt elevation changes that complicate water surface point detection and classification (<xref ref-type="bibr" rid="B31">Letard et al., 2024</xref>). Rivers often have complex hydrological structures with multiple braided channels that diverge and rejoin, connecting with abandoned channels or adjacent floodplain lakes, adding to the complexity. The Snake River in Wyoming, United States, as shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, exemplifies these challenges.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Snake River, Wyoming, in <bold>(a)</bold> a dam, and <bold>(b)</bold> shows a braided river channel, which contributes to the complexity of the river system due to varying flow dynamics, sediment transport, and multiple interconnecting channels.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g001.tif">
<alt-text content-type="machine-generated">Satellite image showing the Snake River flowing from Jackson Lake, marked by elevation from 1958 to 2128 meters. Insets detail the Jackson Lake Dam and braided river channels near Schwabacher Landing. A compass rose and scale bars are included.</alt-text>
</graphic>
</fig>
<p>The industry standard to correct unknown travel time in water involves manually classifying the water surface and water column points using a filtering method, creating two surface model using benthic and water surface points separately, calculating the depth of water column and waterbed points, and finally correcting their elevation proportionate to their depth based on the travel time in water and the resulting change in the speed of light. For the same Snake River dataset, previous studies such as <xref ref-type="bibr" rid="B29">Legleiter et al. (2016)</xref> applied an adaptive TIN filter (<xref ref-type="bibr" rid="B7">Axelsson, 2000</xref>) to identify candidate water surface points, followed by manual editing to refine surface and bottom separation. A channel centerline was then fitted to the surface points, and refraction correction was applied to bottom returns using the time difference between green laser surface and benthic echoes. Two separate surface models were generated for the water surface and riverbed, and morphological filtering was used to remove noise. <xref ref-type="bibr" rid="B40">Pan et al. (2015a)</xref> used a continuous wavelet transform to pick seed peaks in each full waveform, then manually classified those peaks into water surface, water-column, and benthic returns and corrected their ranges using the refractive index. The manual process of refraction and depth correction for bathymetric lidar data is inherently time-consuming and susceptible to human error due to the repetitive nature of the tasks. Therefore, developing an automated methodology to ensure precision, efficiency, and repeatability is needed. Automating these corrections can significantly reduce processing time and minimize the introduction of errors, thereby improving the reliability of bathymetric measurements.</p>
<p>Current machine learning research using bathymetric airborne laser scanning (ALS) focuses on detecting structures (<xref ref-type="bibr" rid="B54">Tsai et al., 2021</xref>), and identifying vegetation. <xref ref-type="bibr" rid="B12">Daniel and Dupont (2020)</xref> used a convolutional neural network (CNN) for seabed features, and <xref ref-type="bibr" rid="B15">Erena et al. (2019)</xref> highlight drones for high-resolution topo-bathymetric monitoring. While shallow water applications are well-documented (<xref ref-type="bibr" rid="B55">Wang and Philpot, 2007</xref>; <xref ref-type="bibr" rid="B23">Hyejin Kim et al., 2023</xref>), deep learning for water-level semantic segmentation is limited. Innovative methods like pseudo-waveform decomposition (<xref ref-type="bibr" rid="B23">Hyejin Kim et al., 2023</xref>) and multispectral imaging (<xref ref-type="bibr" rid="B36">Mandlburger et al., 2021</xref>) exist, but purely point cloud applications remain largely unexplored, indicating a need for further research. <xref ref-type="bibr" rid="B9">Bouhdaoui et al. (2014)</xref> studied the impact of complex water bottom geometries on peak time shifting in bathymetric lidar waveforms, using the Wa-LID waveform simulator to model different depths, slopes, and footprint sizes. <xref ref-type="bibr" rid="B65">Zhou et al. (2023)</xref> addressed water depth bias correction by subdividing the water area into sub-regions based on water depths and biases, using a subdivision algorithm and least-squares regression. <xref ref-type="bibr" rid="B2">Agrafiotis et al. (2019)</xref> applied a support vector regression (SVR) model to correct depth underestimation in point clouds from structure from motion (SfM) and multi-view stereo (MVS) techniques, using known depth data from bathymetric lidar to enhance accuracy and robustness by fusing the lidar and image-based point clouds. For aerial bathymetric lidar data alone, semantic segmentation is crucial yet underexplored for segmenting river surfaces using deep learning.</p>
<p>Recent advances in 3D point cloud segmentation have led to a wide range of deep learning models categorized by their core processing methods. Pointwise MLP-based models such as PointNet (<xref ref-type="bibr" rid="B44">Qi et al., 2016</xref>) and PointNet&#x2b;&#x2b; (<xref ref-type="bibr" rid="B45">Qi et al., 2017</xref>) learn per-point features using shared multilayer perceptrons and aggregate them with symmetric functions. RandLA-Net (<xref ref-type="bibr" rid="B22">Hu et al., 2020</xref>) and ShellNet (<xref ref-type="bibr" rid="B62">Zhang et al., 2019</xref>) improve efficiency and local structure learning, though MLP-based methods generally lack strong spatial context modeling. Volumetric methods, such as SEGCloud (<xref ref-type="bibr" rid="B51">Tchapmi et al., 2017</xref>) and SparseConvNet (<xref ref-type="bibr" rid="B19">Graham et al., 2018</xref>), convert point clouds into voxel grids before processing, which allows the use of 3D convolutions but sacrifices geometric precision and incurs high memory and computation due to voxelization. These methods do not operate directly on the raw point cloud and may lose fine-grained details. Spherical projection models such as SqueezeSeg (<xref ref-type="bibr" rid="B60">Wu et al., 2018</xref>) transform 3D data into 2D spherical range images to enable fast processing with 2D convolutions. However, this transformation alters the spatial structure and introduces distortions, limiting the model&#x2019;s ability to preserve the native geometry of the point cloud. Point convolution methods and graph-based methods represent the most reliable classes of architectures. PointCNN (<xref ref-type="bibr" rid="B33">Li et al., 2018</xref>) and DGCNN (<xref ref-type="bibr" rid="B57">Wang et al., 2019b</xref>) are widely recognized as the standard baselines in these categories due to their consistent performance. PointCNN transforms unordered inputs into canonical forms for convolution, while models like PCNN (<xref ref-type="bibr" rid="B6">Atzmon et al., 2018</xref>), ConvPoint (<xref ref-type="bibr" rid="B10">Boulch, 2020</xref>), and KPConv (<xref ref-type="bibr" rid="B53">Thomas et al., 2019</xref>) apply continuous or deformable filters. Graph-based models such as DGCNN and GACNet (<xref ref-type="bibr" rid="B56">Wang et al., 2019a</xref>) dynamically build neighborhood graphs and extract features using edge-based mechanisms. According to <xref ref-type="bibr" rid="B8">Bello et al. (2020)</xref>, many deep learning models for point cloud segmentation achieve comparable performance across standard benchmarks. Therefore, we focus on PointCNN and DGCNN, which are inherently different in design but have been widely used across diverse applications and reflect the two dominant paradigms of point cloud learning (<xref ref-type="bibr" rid="B34">Lumban-Gaol et al., 2021</xref>; <xref ref-type="bibr" rid="B25">Koguciuk et al., 2019</xref>).</p>
<p>3D semantic segmentation is widely applied in computer vision and remote sensing, providing point-wise segmentation of point clouds. In environmental applications, topographic and bathymetric lidar data are used for land cover mapping (<xref ref-type="bibr" rid="B5">Arief et al., 2018</xref>; <xref ref-type="bibr" rid="B14">Ekhtari et al., 2018</xref>; <xref ref-type="bibr" rid="B63">Zhang et al., 2022</xref>), distinguishing terrestrial and aquatic features to support hydrological modeling, flood risk analysis, and environmental monitoring (<xref ref-type="bibr" rid="B64">Zhao et al., 2016</xref>). Building on segmentation methods, PointCNN (<xref ref-type="bibr" rid="B33">Li et al., 2018</xref>) and DGCNN (<xref ref-type="bibr" rid="B57">Wang et al., 2019b</xref>) represent two leading architectures for 3D point cloud processing. PointCNN, adapted from PointNet&#x2b;&#x2b; (<xref ref-type="bibr" rid="B45">Qi et al., 2017</xref>), introduces hierarchical feature learning and has been applied in environmental mapping (<xref ref-type="bibr" rid="B17">Fareed et al., 2023</xref>), 3D object recognition (<xref ref-type="bibr" rid="B25">Koguciuk et al., 2019</xref>), and autonomous navigation (<xref ref-type="bibr" rid="B1">A Arief et al., 2019</xref>). DGCNN extends PointNet (<xref ref-type="bibr" rid="B44">Qi et al., 2016</xref>) with dynamic graph-based convolutions to better capture local and global shape information. These models are not only technically robust but also highly relevant in practical domains, establishing themselves as essential tools for modern 3D data understanding.</p>
<p>Achieving accurate segmentation of bathymetric lidar through deep learning models like PointCNN or DGCNN can eliminate the need for manual processes, which are time-consuming and prone to human error due to their qualitative and repetitive nature. Manual classification lacks repeatability, as two individuals performing the classification independently may get different results. Utilizing deep learning models can streamline the workflow by removing the necessity for additional steps such as delineating the river, incorporating external images for verification, and manually classifying the water surface. This new automated approach enhances efficiency, accuracy and repeatability in the classification of bathymetric lidar data by performing water column classification with high precision. Although PointCNN and DGCNN presented as candidate segmentation approaches in this study, the workflow remains architecture-agnostic; any point-cloud network (e.g., PointNet&#x2b;&#x2b;, KPConv, or a transformer-based model) capable of reliably separating water-surface from water-bed points can be substituted without altering the subsequent refraction-correction and depth-estimation steps. To our knowledge, this is the first model trained and tested for simultaneous semantic segmentation for both the river surface and riverbed. The segmentation of these two layers aids in the calculation of water depth, making it more automated and accurate for refraction and depth correction of the benthic layer.</p>
</sec>
<sec id="s2">
<title>2 Study area and methods</title>
<sec id="s2-1">
<title>2.1 Study area</title>
<p>The three datasets used in this study, shown in <xref ref-type="table" rid="T1">Table 1</xref>, are bathymetric lidar survey data for three rivers in the United States: the Swan River in Montana, the Snake River in Wyoming, and the Eel River in California. All three datasets were collected by the National Center for Airborne Laser Mapping (NCALM) (<ext-link ext-link-type="uri" xlink:href="http://ncalm.cive.uh.edu">ncalm.cive.uh.edu</ext-link>) and the data includes independent ground truth data. The ground truth was created using manual classification, ensuring high accuracy through multiple iterations. The process involved using Terrasolid&#x2019;s Terrascan software to guarantee precise data labeling, with verification using Google Earth images. The Swan River survey was conducted in 2023, the Snake River in 2012, and the Eel River in 2014. These surveys cover a range of geographical areas, with the Swan River encompassing 53 <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>km</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, the Snake River 77 <inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>km</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, and the Eel River 57 <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>km</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. However, only a narrow corridor along each river was selected for analysis to focus the study on only the river itself, significantly reducing the total point count and concentrating solely on river classes as shown in <xref ref-type="table" rid="T2">Table 2</xref>. The surveyed corridor length of these rivers is 27&#xa0;km for the Swan River, 44&#xa0;km for the Snake River, and 204&#xa0;km for the Eel River. An important aspect of these surveys is the point density, expressed as the number of lidar points per square meter. This metric is used for determining the survey data&#x2019;s resolution and level of detail. The Swan River survey has a high point density of 28 <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>pts/m</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, which suggests a very detailed and fine-scale survey. In contrast, the Snake River survey has a much lower point density of 3 <inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>pts/m</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, indicating a coarser resolution. The Eel River survey falls between these two, with a point density of 14 <inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>pts/m</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. The sensor technologies employed for these surveys also differ, reflecting technological advancements and specific sensor capabilities. The Snake River and Eel River surveys were conducted using the Optech AQUARIUS sensor (<xref ref-type="bibr" rid="B20">Guo et al., 2018</xref>) and the Swan River survey used the RIEGL VQ-840-G sensor (<xref ref-type="bibr" rid="B42">Pfennigbauer et al., 2022</xref>). The datasets used in our study are diverse in terms of river hydrogeomorphology, lidar sensor technologies, and point cloud densities. This diversity enhances the robustness and generalizability of any machine learning model, allowing it to effectively learn and predict river channel evolution in various geographic and environmental contexts.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Details of lidar datasets used in this research.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">River name</th>
<th align="left">State in USA</th>
<th align="left">Year collected</th>
<th align="center">River corridor length (km)</th>
<th align="center">Point density (points/m<sup>2</sup>)</th>
<th align="left">Sensor</th>
<th align="center">Total points (million)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Swan (<xref ref-type="bibr" rid="B38">Marshall, 2023</xref>)</td>
<td align="left">Montana</td>
<td align="left">2023</td>
<td align="center">27</td>
<td align="center">28</td>
<td align="left">RIEGL VQ-840-G</td>
<td align="center">1,469</td>
</tr>
<tr>
<td align="left">Eel (<xref ref-type="bibr" rid="B13">Dietrich, 2014</xref>)</td>
<td align="left">California</td>
<td align="left">2014</td>
<td align="center">204</td>
<td align="center">14</td>
<td align="left">Optech AQUARIUS</td>
<td align="center">822</td>
</tr>
<tr>
<td align="left">Snake (<xref ref-type="bibr" rid="B28">Legleiter, 2024</xref>)</td>
<td align="left">Wyoming</td>
<td align="left">2012</td>
<td align="center">44</td>
<td align="center">4</td>
<td align="left">Optech AQUARIUS</td>
<td align="center">213</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Total points after clipping to river center corridor and class percentage distribution for each river dataset.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">River name</th>
<th align="center">Total Points (million)</th>
<th align="center">Other (%)</th>
<th align="center">Ground (%)</th>
<th align="center">Riverbed (%)</th>
<th align="center">River surface (%)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Swan</td>
<td align="center">186</td>
<td align="center">71.9</td>
<td align="center">14.3</td>
<td align="center">11.8</td>
<td align="center">2.0</td>
</tr>
<tr>
<td align="left">Eel</td>
<td align="center">44</td>
<td align="center">50.7</td>
<td align="center">31.0</td>
<td align="center">12.2</td>
<td align="center">6.1</td>
</tr>
<tr>
<td align="left">Snake</td>
<td align="center">20</td>
<td align="center">18.7</td>
<td align="center">44.8</td>
<td align="center">26.7</td>
<td align="center">9.7</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The data for all three rivers were manually classified into four categories: river surface, riverbed, ground, and other. The river surface class represents the interface between the water and the atmosphere, capturing the mostly specular reflection of laser light on the water surface. The riverbed class encompasses the benthic layer at the bottom of the water, including both the sediment and any submerged vegetation. The riverbed class also accounts for volumetric scattering in the water column. Volumetric scattering occurs when light interacts with suspended particles and water molecules. These interactions cause the lidar signal to scatter in multiple directions, affecting its penetration depth and the peak of the backscatter signal received. The ground class includes terrestrial surfaces, excluding aquatic or vegetative components, clearly differentiated from the marine environment. Lastly, the other class includes all remaining data points that do not fall into the previous three categories and predominantly consist of vegetation. This categorization ensures comprehensive mapping and analysis of the riverine and adjacent environments. To focus on the riverine environment, only a subset of each dataset was used, including points along a 50-m corridor from the centerline of each river. The class distribution for this corridor is given in <xref ref-type="table" rid="T2">Table 2</xref>. Even after reducing to a narrow along the river, the ground and other classes still comprise the majority of points, whereas the river surface and riverbed classes account for only a small percentage.</p>
<p>To validate our depth corrections on the Snake River, we used the ground truth acquired by <xref ref-type="bibr" rid="B29">Legleiter et al. (2016)</xref>. They surveyed Rusty Bend (a specific bend in the Snake River) with a SonTek RiverSurveyor S5 Acoustic Doppler Current Profiler (ADCP) mounted on a kayak and tied to an RTK-GPS base station, getting depth soundings with 0.001&#xa0;m resolution and 1% accuracy over 0.2&#x2013;15&#xa0;m. Along the channel margins, they also measured water edge elevations with RTK-GPS to correct for instrument drift and kayak offset. All ADCP returns within 0.5&#xa0;m of a GPS control point were cross-calibrated to ensure vertical consistency. We used these rigorously corrected and georeferenced ADCP measurements as the bathymetric ground-truth along the Snake River for validating our fully automated depth correction model.</p>
</sec>
<sec id="s2-2">
<title>2.2 Methodology</title>
<p>The methodology starts with the data processing steps used to convert light detection and ranging files into uniformly sized blocks suitable for deep learning input. The process continues with data augmentation techniques aimed at increasing the dataset&#x2019;s size and generalizability for the machine learning model. The model is then discussed with an in-depth analysis of the PointCNN implementation using ESRI&#x2019;s deep learning module arcgis.learn (<xref ref-type="bibr" rid="B16">Esri, 2024</xref>). Different metrics and error rates are then used to evaluate the model&#x2019;s performance. Finally, the methodology addresses depth and refraction corrections to achieve the desired corrected final riverbed points, as shown in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Workflow diagram showing the steps involved in this research. The process starts with input data, followed by pre-processed data where data cleaning and transformation occur. Augmenting the dataset is the next step in enhancing the training dataset. After the training and testing phase, the model is evaluated against the ground truth. Finally, depth and refraction corrections are applied to the riverbed points classified by the model.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g002.tif">
<alt-text content-type="machine-generated">Flowchart depicting a data processing pipeline. It starts with &#x22;Input,&#x22; leading to &#x22;Pre-process Data,&#x22; then &#x22;Augmenting Data,&#x22; followed by &#x22;Model Training and Testing.&#x22; The path splits: one leads to &#x22;Evaluation&#x22; and then &#x22;Output,&#x22; while the other leads to &#x22;Depth &#x26; Refraction Correction&#x22; and back to &#x22;Augmenting Data.&#x22;</alt-text>
</graphic>
</fig>
<sec id="s2-2-1">
<title>2.2.1 Data preprocessing</title>
<p>Point cloud data cannot be directly fed into a deep learning model due to its unstructured and unorganized nature. To make it suitable for deep learning, the data must be preprocessed into a uniform dataset, as demonstrated in methods such as, Pointnet&#x2b;&#x2b; (<xref ref-type="bibr" rid="B45">Qi et al., 2017</xref>) and PointCNN (<xref ref-type="bibr" rid="B33">Li et al., 2018</xref>). The entire dataset is broken down into smaller blocks, each with a constant size of <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> x <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. Block size <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is selected based on the specific application requirements as determined by the user. Block sizes are based on the data characteristics, scene complexity, and point density.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Point cloud data (with ground truth) is divided into frames, as seen on the left side of the diagram. Points are extracted from each frame to create uniform-sized blocks <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> x <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mi mathvariant="normal">S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> (square blocks are required for the model) with a fixed number of points <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> using Farthest Point Sampling (FPS). <inline-formula id="inf17">
<mml:math id="m17">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> number of blocks are generated after a single frame is fully sampled.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g003.tif">
<alt-text content-type="machine-generated">Map with overlapping red squares representing blocks of constant size, overlaying a colored terrain. An arrow points to a diagram showing stacked blocks labeled as &#x22;Points Extracted to Make the Blocks.&#x22; Text explains the number of blocks and points, with a north arrow indicating orientation. Scale bar shows 0 to 300 meters.</alt-text>
</graphic>
</fig>
<p>The gridded squares shown in <xref ref-type="fig" rid="F3">Figure 3</xref> are the frames that contain the raw point cloud sets. Points are sampled from these frames to create the final product called a block. The sampling from each frame starts with an initial point selected randomly. This point serves as the starting reference for the sampling process. Because the initial point is chosen randomly, it introduces some variability in the sampling process; different initial points can lead to different sampled sets. From this initial point, Farthest Point Sampling (FPS) (<xref ref-type="bibr" rid="B45">Qi et al., 2017</xref>) calculates the Euclidean distance to every other point in the point cloud frame. The point farthest from this initial point, i.e., the point with the maximum distance, is then selected and added to the sampled set. In subsequent iterations, the algorithm updates the newly added point as the reference to find the next farthest point. This iterative process continues until the desired number of points are chosen and a block is created. Each block is represented as a set of points:<disp-formula id="equ1">
<mml:math id="m18">
<mml:mrow>
<mml:mfenced open="{" close="}">
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>where <inline-formula id="inf18">
<mml:math id="m19">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of points within the block. Similar to block size, the value of <inline-formula id="inf19">
<mml:math id="m20">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> can be decided based on point density, however, another critical factor is computational power. The larger the value of <inline-formula id="inf20">
<mml:math id="m21">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the more dedicated memory is required. Once the sample set is complete, the process repeats on the remaining points in the frame until all points are sampled to create blocks. FPS effectively spreads out the sampled points, ensuring a uniform and comprehensive coverage of the entire point cloud. This method preserves the structural diversity and spatial distribution of the original dataset.</p>
</sec>
<sec id="s2-2-2">
<title>2.2.2 Augmenting data</title>
<p>A rotation matrix about the z-axis is extensively used to augment point cloud data in various machine learning and computer vision applications (<xref ref-type="bibr" rid="B49">Shi et al., 2021</xref>; <xref ref-type="bibr" rid="B11">Choi et al., 2021</xref>). This matrix, given by:<disp-formula id="equ2">
<mml:math id="m22">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mtable class="matrix">
<mml:mtr>
<mml:mtd columnalign="center">
<mml:mi>cos</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mo>&#x2212;</mml:mo>
<mml:mi>sin</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="center">
<mml:mi>sin</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mi>cos</mml:mi>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>&#x3b8;</mml:mi>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mn>0</mml:mn>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd columnalign="center">
<mml:mn>0</mml:mn>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mn>0</mml:mn>
</mml:mtd>
<mml:mtd columnalign="center">
<mml:mn>1</mml:mn>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>allows for the rotation of 3D points around the z-axis by an angle <inline-formula id="inf21">
<mml:math id="m23">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. Augmenting point cloud data through such rotations helps artificially increase the dataset&#x2019;s size, introducing variations that can improve the robustness and generalization of machine learning models. Three data orientations are created by applying this transformation to all points in a point cloud by 90<inline-formula id="inf22">
<mml:math id="m24">
<mml:mrow>
<mml:mo>&#xb0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, 180<inline-formula id="inf23">
<mml:math id="m25">
<mml:mrow>
<mml:mo>&#xb0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and 270<inline-formula id="inf24">
<mml:math id="m26">
<mml:mrow>
<mml:mo>&#xb0;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. This augmentation technique ensures that the models are not overly dependent on specific orientations and can effectively handle diverse real-world scenarios.</p>
</sec>
<sec id="s2-2-3">
<title>2.2.3 Deep learning model</title>
<p>PointNet&#x2b;&#x2b; (<xref ref-type="bibr" rid="B45">Qi et al., 2017</xref>) extends the original PointNet (<xref ref-type="bibr" rid="B44">Qi et al., 2016</xref>) by capturing local geometric structures in point clouds through a hierarchical framework. It divides point clouds into local regions and recursively applies PointNet using set abstraction layers that perform sampling, grouping, and multi-scale feature aggregation. This enables effective learning of both local and global features, making PointNet&#x2b;&#x2b; suitable for classification and segmentation tasks. It has been successfully applied to semantic segmentation of airborne lidar in urban environments. For example, <xref ref-type="bibr" rid="B50">Shin et al. (2022)</xref> used it for building extraction by incorporating multiple lidar returns, while <xref ref-type="bibr" rid="B24">Jing et al. (2021)</xref> applied it to large-scale point cloud segmentation.</p>
<p>PointCNN (<xref ref-type="bibr" rid="B33">Li et al., 2018</xref>) further improves classification by learning spatial relationships through its X-Conv (X-transformation convolution) module, which reorders and weights local neighborhoods using a learned transformation matrix. This enhances local feature aggregation and pattern recognition in 3D space. PointCNN has shown strong performance across diverse applications, including powerline classification (<xref ref-type="bibr" rid="B27">Kumar et al., 2024</xref>), tree species identification (<xref ref-type="bibr" rid="B21">Hell et al., 2022</xref>), and UAS-LiDAR ground point classification in agriculture (<xref ref-type="bibr" rid="B17">Fareed et al., 2023</xref>), outperforming traditional filters like CSF and PMF. Due to its robustness and consistent performance across various ALS segmentation tasks, PointCNN was selected for bathymetric lidar data classification in this study.</p>
<p>The X-Conv operation in PointCNN (<xref ref-type="bibr" rid="B33">Li et al., 2018</xref>), as depicted in <xref ref-type="fig" rid="F4">Figure 4</xref>, is designed to extract higher-level features from unstructured point cloud data, enabling more precise and effective point cloud analysis.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>X-Convolution operation from PointCNN extracting higher features from unstructured point cloud data. <bold>(a)</bold> Initial point cloud with scattered points (orange circles). <bold>(b)</bold> Local regions are formed after the first X-Conv layer, and the relationships between neighboring points are established (indicated by the red arrows). The initial 10 points are reduced to 6 (cyan circles) based on feature and spatial relationships. <bold>(c)</bold> After the second X-Conv layer, more refined local regions are created (red circles), with stronger connections among neighboring points. This leads to a higher-level representation of the point cloud with only three remaining points.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g004.tif">
<alt-text content-type="machine-generated">Diagram showing a three-step process labeled (a), (b), and (c). In (a), scattered orange dots represent initial points. In (b), red arrows connecting blue dots show the X-Conv transformation creating a network. In (c), further transformations with fewer connections are depicted.</alt-text>
</graphic>
</fig>
<p>Step 1: For each selected center point (shown in <xref ref-type="fig" rid="F4">Figure 4a</xref>), the <inline-formula id="inf25">
<mml:math id="m27">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> nearest neighbors are collected, and their coordinates are stacked into a matrix:<disp-formula id="equ3">
<mml:math id="m28">
<mml:mrow>
<mml:mi>P</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x22a4;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>where <inline-formula id="inf26">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> are the coordinates (e.g., <inline-formula id="inf27">
<mml:math id="m30">
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf28">
<mml:math id="m31">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> positions), and <inline-formula id="inf29">
<mml:math id="m32">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of neighboring points. <xref ref-type="fig" rid="F4">Figure 4b</xref> illustrates these neighbors connected to the center.</p>
<p>Step 2: Each neighbor&#x2019;s coordinate <inline-formula id="inf30">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is passed through a point-wise multilayer perceptron (MLP), producing a higher-dimensional feature embedding:<disp-formula id="equ4">
<mml:math id="m34">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">MLP</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>where <inline-formula id="inf31">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the dimensionality of the extracted features. Simultaneously, a second MLP generates a transformation matrix:<disp-formula id="equ5">
<mml:math id="m36">
<mml:mrow>
<mml:mi mathvariant="script">X</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo movablelimits="false" form="prefix">MLP</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>which learns to both re-order and re-weight the <inline-formula id="inf32">
<mml:math id="m37">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> neighbors. The transformed coordinates and features are then combined:<disp-formula id="equ6">
<mml:math id="m38">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mi mathvariant="script">X</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>,</mml:mo>
<mml:mspace width="2em"/>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="script">X</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2295;</mml:mo>
<mml:mi mathvariant="script">X</mml:mi>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>where <inline-formula id="inf33">
<mml:math id="m39">
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represents any additional raw point attributes (such as intensity or color), and <inline-formula id="inf34">
<mml:math id="m40">
<mml:mrow>
<mml:mo>&#x2295;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> denotes feature concatenation along the feature dimension.</p>
<p>Step 3: Finally, the ordered and weighted feature block <inline-formula id="inf35">
<mml:math id="m41">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is passed through a linear layer with learnable weights <inline-formula id="inf36">
<mml:math id="m42">
<mml:mrow>
<mml:mi>W</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b4;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>:<disp-formula id="equ7">
<mml:math id="m43">
<mml:mrow>
<mml:mo movablelimits="false" form="prefix">Conv</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>W</mml:mi>
<mml:msup>
<mml:mrow>
<mml:mi>F</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2a;</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>producing the output features for the center point, where <inline-formula id="inf37">
<mml:math id="m44">
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the output feature dimension.</p>
<p>This operation combines the <inline-formula id="inf38">
<mml:math id="m45">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> neighbour features into a single vector of length <inline-formula id="inf39">
<mml:math id="m46">
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, producing one new point that carries richer information than any individual neighbour. Repeating the same procedure in deeper layers <xref ref-type="fig" rid="F4">Figure 4c</xref> reduces the number of points (for example, <inline-formula id="inf40">
<mml:math id="m47">
<mml:mrow>
<mml:mn>10</mml:mn>
<mml:mo>&#x2192;</mml:mo>
<mml:mn>6</mml:mn>
<mml:mo>&#x2192;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> in the <xref ref-type="fig" rid="F4">Figure 4</xref>) while increasing each point&#x2019;s feature dimension, thereby enlarging the receptive field and capturing progressively more contextual detail.</p>
<p>Because the transformation matrix <inline-formula id="inf41">
<mml:math id="m48">
<mml:mrow>
<mml:mi mathvariant="script">X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is learned directly from each local neighbourhood, X&#x2013;Conv simultaneously: (i) re-orders points into a data driven canonical sequence, ensuring permutation invariance, (ii) assigns geometry aware weights that emphasises informative neighbours and suppresses noise, giving it shape adaptivity. These properties suit the unordered and highly detailed nature of lidar data.</p>
<p>The architecture diagram in <xref ref-type="fig" rid="F5">Figure 5</xref> shows Esri&#x2019;s implementation of PointCNN (<xref ref-type="bibr" rid="B16">Esri, 2024</xref>), which is different from the original PointCNN primarily in the parameter settings for each X-Conv layer. In this configuration, the number of points <inline-formula id="inf42">
<mml:math id="m49">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> decreases from 8,192 to 1,024, while the number of channels <inline-formula id="inf43">
<mml:math id="m50">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> increases from 128 to 768, allowing deeper layers to extract more complex features. The neighborhood size <inline-formula id="inf44">
<mml:math id="m51">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>K</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> ranges from 96 to 128, representing the number of nearest neighbors, and dilation rates <inline-formula id="inf45">
<mml:math id="m52">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> of 8 and 16 enable the model to capture multi-scale spatial patterns. The original PointCNN architecture used different parameter values and was tuned for specific datasets and tasks, often involving additional configurations for feature extraction and layer-wise optimization.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>PointCNN architecture shows the first layer receiving inputs of point clouds, followed by the convolutional layers consisting of X-Conv layers. The final output layer gives the classified outputs of points designated by <inline-formula id="inf46">
<mml:math id="m53">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>O</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g005.tif">
<alt-text content-type="machine-generated">Diagram of a neural network with an input layer of 8192 nodes at the top and an output layer of 8192 nodes at the bottom. Between them are three X-Conv layers with varying configurations: X-Conv(N=8192, C=128, K=96, D=8), X-Conv(N=6144, C=256, K=96, D=8), and X-Conv(N=3072, C=512, K=128, D=16). Connections are depicted as lines between nodes across layers.</alt-text>
</graphic>
</fig>
<p>For comparison, a second model was also trained and tested, Dynamic Graph CNN (DGCNN), which enhances PointNet (<xref ref-type="bibr" rid="B44">Qi et al., 2016</xref>) for point cloud data with its EdgeConv operation, dynamically updating point connections based on learned features to capture local geometric structures. DGCNN applies CNN principles to dynamic graphs, adapting Graph Neural Networks (GNNs) for unstructured data, but updates its graph structure during training (<xref ref-type="bibr" rid="B46">Scarselli et al., 2009</xref>). This combination improves DGCNN&#x2019;s effectiveness in classification and segmentation tasks (<xref ref-type="bibr" rid="B57">Wang et al., 2019b</xref>). <xref ref-type="bibr" rid="B58">Widyaningrum et al. (2021)</xref> applied DGCNN to classify airborne laser scanning point clouds in urban areas, using datasets from Surabaya and the Netherlands. They explored various input features, block sizes, and loss functions and concluded that DGCNN is highly effective for urban ALS point cloud classification, achieving near-production quality results. Due to its native ability to compute unorganized 3D data, similar to how NLP (Natural Language Processing) models handle text, DGCNN was used to determine if a graph-based CNN could outperform a 3D CNN like PointCNN for bathymetric lidar data.</p>
</sec>
<sec id="s2-2-4">
<title>2.2.4 Refraction and depth correction</title>
<p>The depth of riverbed points can be calculated using a water surface model. Such a model is formed by Delaunay triangulation of water surface and ground points which results in a seamless Triangulated Irregular Network (TIN). An interpolation function <inline-formula id="inf47">
<mml:math id="m54">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>TIN</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> can then be used to obtain the equivalent water surface elevation for each riverbed point. Using this function, we calculated the interpolated surface height at the precise X and Y positions of each riverbed point with the function:<disp-formula id="equ8">
<mml:math id="m55">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>interpolated</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>TIN</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>The interpolated surface points are then used to compute the water depths by subtracting the interpolated surface points from the original Z-values for the riverbed points:<disp-formula id="equ9">
<mml:math id="m56">
<mml:mrow>
<mml:mtext>Water&#x2009;Depth</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>river</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>surface</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>The water depth is then adjusted for the refraction of light in water. This correction factor, given by the refractive index of water (approximately <inline-formula id="inf48">
<mml:math id="m57">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>water</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.33</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> for light refraction), is applied to obtain the final corrected riverbed depths:<disp-formula id="equ10">
<mml:math id="m58">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>corrected</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>Water&#x2009;Depth</mml:mtext>
<mml:mo>&#xd7;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>water</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>Finally, the refraction corrections for the riverbed points are applied to the horizontal coordinates. Using the methodology described in <xref ref-type="bibr" rid="B41">Pan et al. (2015b)</xref>, this process involves adjusting each riverbed point return for refraction at the air-water interface. Using Snell&#x2019;s law, the adjustment accounts for the change in refractive indices between air and water, requiring correction for the lidar pulse angles. This iterative process continues until the corrections for the angle are minimized, ensuring accurate horizontal adjustments.</p>
</sec>
</sec>
</sec>
<sec id="s3">
<title>3 Experiments and results</title>
<p>This section presents a comprehensive analysis of the experiments conducted, focused on comparing neural networks, the impact of both varying block sizes and the number of points used in a block on semantic segmentation performance, and the final model training process. For model evaluation, IoU, Accuracy, Precision, Recall, F1, Kappa, Commission Error (CE), and Omission Error (OE) metrics were used. The results underscore the effectiveness of PointCNN to deliver balanced precision and recall, its robustness across different segmentation tasks, and present the optimal configurations for achieving high performance in complex riverine environments.</p>
<sec id="s3-1">
<title>3.1 Hardware configuration and training parameters</title>
<p>All of the training was performed on an NVIDIA Quadro P6000 GPU (24&#xa0;GB VRAM) under Ubuntu 22.04.5 LTS. For both PointCNN and DGCNN, Adam optimizer was used with a learning rate of 0.004 and a dropout rate of 0.25. In PointCNN, we used point-wise cross-entropy loss with a batch size of 6. For DGCNN, cross-entropy loss was used with a batch size of 2. Both setups required approximately 22&#xa0;GB of GPU memory. Block size and per-block point-sampling configurations were explored in <xref ref-type="sec" rid="s3-2">Section 3.2</xref>, and then the optimal setting was selected for the final model.</p>
</sec>
<sec id="s3-2">
<title>3.2 Block size and number of points in a block</title>
<p>Analyzing semantic segmentation performance across various block sizes (S) and number of points (P) reveals crucial insights into how these parameters impact the model&#x2019;s effectiveness. This analysis focuses exclusively on classifying the primary classes, river surface, and riverbed, as those are needed to apply depth and refraction correction.</p>
<sec id="s3-2-1">
<title>3.2.1 Analysis for number of points in a block</title>
<p>Impact of varying number of block points (P &#x3d; 32,768, P &#x3d; 16,384, P &#x3d; 8,192): For this experiment, a constant block size of 50 &#xd7; 50 square meters is used with a varying number of points. The training and testing were done on the Swan River dataset only. In the top half of <xref ref-type="table" rid="T3">Table 3</xref>, a model trained with P &#x3d; 16,384 consistently delivers balanced performance across both classes. The performance metrics for the river surface class across different numbers of points highlight that the P=16384 configuration is optimal. This number of points in a block achieves the highest IoU at 0.94 and the highest Kappa coefficient at 0.96 for the river surface, indicating the most accurate model performance. The CE remains consistently low at 0.01 across all numbers of points, but the OE is notably lower at 0.05 for P=16384, compared to 0.08 for both P=32768 and P=8192. This indicates that the P=16384 configuration has fewer missed detections, making it the best performer among the tested block point counts. While the observed differences in metrics are not huge, they are significant in bathymetric lidar segmentation. In this specific application, even small improvements in accuracy can significantly impact the quality of underwater terrain mapping, where precision is crucial. The point density on this configuration, 6 points/m<sup>2</sup> appears to provide the optimal balance between feature quality and computational efficiency.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Performance Metrics for different experiment settings.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th colspan="5" align="center">Optimal number of block points, Swan river, MT</th>
</tr>
<tr>
<th align="left">Class</th>
<th align="left">Metric</th>
<th align="center">P &#x3d; 8,192 (3&#xa0;pts/m<sup>2</sup>)</th>
<th align="left">P &#x3d; 16,384 (6&#xa0;pts/m<sup>2</sup>)</th>
<th align="left">P &#x3d; 32,768 (13&#xa0;pts/m<sup>2</sup>)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="4" align="left">River Surface</td>
<td align="left">IoU</td>
<td align="center">0.92</td>
<td align="center">0.94</td>
<td align="center">0.92</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.94</td>
<td align="center">0.96</td>
<td align="center">0.94</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.01</td>
<td align="center">0.01</td>
<td align="center">0.01</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.08</td>
<td align="center">0.05</td>
<td align="center">0.08</td>
</tr>
<tr>
<td rowspan="4" align="left">Riverbed</td>
<td align="left">IoU</td>
<td align="center">0.81</td>
<td align="center">0.83</td>
<td align="center">0.78</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.86</td>
<td align="center">0.88</td>
<td align="center">0.83</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.09</td>
<td align="center">0.11</td>
<td align="center">0.14</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.12</td>
<td align="center">0.08</td>
<td align="center">0.11</td>
</tr>
</tbody>
</table>
<table>
<thead>
<tr>
<th colspan="5" align="center">Optimal Block Size, Snake River, WY</th>
</tr>
<tr>
<th align="left">Class</th>
<th align="left">Metric</th>
<th align="center">S &#x3d; 30&#xa0;m (18&#xa0;pts/m<sup>2</sup>)</th>
<th align="center">S &#x3d; 50&#xa0;m (6&#xa0;pts/m<sup>2</sup>)</th>
<th align="center">S &#x3d; 100&#xa0;m (1&#xa0;pts/m<sup>2</sup>)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="4" align="left">River Surface</td>
<td align="left">IoU</td>
<td align="center">0.34</td>
<td align="center">0.58</td>
<td align="center">0.55</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.42</td>
<td align="center">0.67</td>
<td align="center">0.64</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.15</td>
<td align="center">0.04</td>
<td align="center">0.06</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.64</td>
<td align="center">0.41</td>
<td align="center">0.43</td>
</tr>
<tr>
<td rowspan="4" align="left">Riverbed</td>
<td align="left">IoU</td>
<td align="center">0.41</td>
<td align="center">0.56</td>
<td align="center">0.49</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.44</td>
<td align="center">0.62</td>
<td align="center">0.58</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.40</td>
<td align="center">0.29</td>
<td align="center">0.17</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.44</td>
<td align="center">0.27</td>
<td align="center">0.45</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-2-2">
<title>3.2.2 Block size analysis</title>
<p>Comparison across block sizes, S &#x3d; 100&#xa0;m, S &#x3d; 50&#xa0;m, and S &#x3d; 30&#xa0;m: The training of these models was done using the Swan River datasets, but the testing was done on our benchmark Snake River dataset. The Swan River was initially used as test data to determine the block size, but the performance was similar for all blocks. However, the performance varies when it is used on unseen datasets. The performance on varying block sizes with the constant number of points (16,384) is shown in the bottom half of <xref ref-type="table" rid="T3">Table 3</xref>. The performance metrics for the river surface class across different block sizes show that the S &#x3d; 50&#xa0;m configuration performs best. It achieves the highest IoU (0.58) and Kappa (0.67), indicating better model accuracy. The CE is also lowest at 0.04, suggesting fewer false positives. Although the OE is slightly lower at 0.41 compared to 0.43 for S &#x3d; 100&#xa0;m, it is significantly better than 0.64 for S &#x3d; 30&#xa0;m. For the riverbed class, S &#x3d; 50&#xa0;m also shows the best performance with the highest IoU (0.56), Kappa (0.62), and the lowest OE (0.27), though it has a higher CE (0.29) than S &#x3d; 100&#xa0;m. Overall, the S &#x3d; 50&#xa0;m configuration provides the best balance of accuracy and error rates for both classes, which has a point density of 6 points/m<sup>2</sup> as well.</p>
</sec>
<sec id="s3-2-3">
<title>3.2.3 Optimal configuration for semantic segmentation</title>
<p>Considering the relationship between block size and the number of points on model performance, our experiments suggest that a block size of S &#x3d; 50&#xa0;m combined with P &#x3d; 16,384 provides the most effective configuration for bathymetric segmentation. This setting has a point density of 6 points/m<sup>2</sup>, balancing spatial context and local feature detail to achieve high segmentation accuracy across both river surface and riverbed classes. This point density ensures sufficient coverage to distinguish complex riverine geometry without overwhelming the model with redundant points or excessive computational burden. Validating the experiments on both the Swan River (training) and the Snake River (testing) demonstrated that this configuration generalized well across different river environments, supporting its applicability to diverse bathymetric mapping tasks.</p>
</sec>
</sec>
<sec id="s3-3">
<title>3.3 Neural network comparison</title>
<p>We first tested PointCNN&#x2019;s performance against DGCNN with optimized values for P and S (as defined in <xref ref-type="sec" rid="s3-1">Section 3.1</xref>). We set P to 8,192 and S to 50&#xa0;m. Both the training and testing were done using the Swan River data. In <xref ref-type="table" rid="T4">Table 4</xref>, PointCNN has more balanced results between Precision and Recall, leading to higher F1 scores than DGCNN. This balance suggests that PointCNN is better at identifying the correct points.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Performance Metrics for PointCNN vs. DGCNN Models, Swan River, MT.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Class</th>
<th align="left">Metric</th>
<th align="center">PointCNN</th>
<th align="center">DGCNN</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="4" align="left">River Surface</td>
<td align="left">IoU</td>
<td align="center">0.92</td>
<td align="center">0.63</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.94</td>
<td align="center">0.72</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.01</td>
<td align="center">0.01</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.08</td>
<td align="center">0.36</td>
</tr>
<tr>
<td rowspan="4" align="left">Riverbed</td>
<td align="left">IoU</td>
<td align="center">0.81</td>
<td align="center">0.59</td>
</tr>
<tr>
<td align="left">Kappa</td>
<td align="center">0.86</td>
<td align="center">0.68</td>
</tr>
<tr>
<td align="left">CE</td>
<td align="center">0.09</td>
<td align="center">0.09</td>
</tr>
<tr>
<td align="left">OE</td>
<td align="center">0.12</td>
<td align="center">0.37</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>DGCNN manages to achieve low CE for the river surface and riverbed classes but struggles with high OE. This indicates a tendency to misclassify points more than PointCNN, which is particularly problematic in classes requiring better scores, like riverbed and river surface for our application. PointCNN&#x2019;s consistently high performance across both classes reflects its robustness and adaptability to segmentation challenges. The model&#x2019;s ability to maintain low error rates while achieving high IoU and Kappa scores shows its effectiveness in semantic segmentation tasks.</p>
</sec>
<sec id="s3-4">
<title>3.4 Final model</title>
<p>For the final modeling, 80% of the data from the Swan and Eel Rivers was used to train the model, while the remaining 20% was set aside for testing on seen data. The results of these tests on the seen data are detailed in <xref ref-type="sec" rid="s14">Supplementary Appendix SA</xref>, demonstrating the model&#x2019;s performance across various metrics for the Swan and Eel Rivers. For the final evaluation and analysis, the focus is on the unseen data from the Snake River. This unseen data serves as a test of the model&#x2019;s ability to generalize and accurately classify river features in previously unobserved environments, providing insights into its overall performance. On average, the model training took 2&#xa0;h and 50&#xa0;min per epoch. Over 30 epochs, the total training time amounted to approximately 85&#xa0;h. For inference, the classification of the entire Snake River dataset was completed in approximately 12&#xa0;min.</p>
<sec id="s3-4-1">
<title>3.4.1 Training and validation loss</title>
<p>The loss graph for the training model in <xref ref-type="fig" rid="F6">Figure 6</xref> shows the changes in training loss and validation loss over 30 epochs. Initially, both losses decrease rapidly, indicating effective learning. By the fifth epoch losses drop significantly below 0.4. After the 13th epoch, the loss values hardly change, and the reason the training loss becomes slightly lower than the validation loss after epoch 13 is due to the onset of overfitting. Therefore, the model was selected at the 13th epoch to prevent overfitting and maintain good generalization performance for the validation dataset. This decision is crucial to balance underfitting and overfitting, optimizing the model&#x2019;s performance on unseen data.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Loss graph for training model over 30 epochs. Both the training loss and validation loss is measured to examine model fit.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g006.tif">
<alt-text content-type="machine-generated">Line graph titled &#x22;Loss Graph of the Training Model&#x22; shows loss on the y-axis and epochs on the x-axis. It features two lines: blue for training loss and orange for validation loss. Both lines decrease sharply initially, then plateau around 0.2, indicating model convergence.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-4-2">
<title>3.4.2 Confusion matrix</title>
<p>A normalized confusion matrix was constructed to show the inter-class relationships, providing insights into the patterns of misclassifications. The normalized confusion matrix demonstrates the classification performance of a model for four categories: river surface, riverbed, ground, and other, with a particular emphasis on the river-related classes. In <xref ref-type="fig" rid="F7">Figure 7</xref>, the riverbed class was correctly identified in 88% of instances. However, 8% of riverbed instances were misclassified as ground, 2% as other, and 2% as river surface. In the case of the river surface class, the model achieved its highest accuracy, correctly classifying 93% of instances. Despite this high accuracy, there were some misclassifications: 2% of river surface instances were incorrectly predicted as riverbed and ground, and 3% as other. These results highlight the model&#x2019;s strong ability to accurately identify river features, especially the river surface, while indicating some confusion, primarily with ground and other classes. This suggests that while the model effectively distinguishes river features, there is still room for improvement to reduce misclassification.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Normalized confusion matrix for the Snake river, Wyoming.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g007.tif">
<alt-text content-type="machine-generated">Normalized confusion matrix displaying prediction labels against true labels for four categories: River Surface, Riverbed, Ground, and Other. High accuracy is indicated on the diagonal with values: 0.93 for River Surface, 0.88 for Riverbed, 0.83 for Ground, and 0.73 for Other. Off-diagonal values indicate misclassifications. A color gradient highlights accuracy, with darker shades representing higher values.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-4-3">
<title>3.4.3 Performance metrics</title>
<p>The data in <xref ref-type="table" rid="T5">Table 5</xref> showcases the performance metrics of the classification model across four classes: river surface, riverbed, ground, and other. The model exhibits strong performance, with IoU scores above 0.67 for all classes. Notably, the river surface class achieved the highest scores across all metrics, with an IoU of 0.89 and a Kappa statistic of 0.92, indicating excellent agreement. Precision is consistently high, although slightly lower for the ground and riverbed classes. The recall is exceptionally high for the riverbed, suggesting good sensitivity. The F1 scores, which balance precision and recall, are robust across all classes. The highlighted results in <xref ref-type="table" rid="T5">Table 5</xref> for the river surface indicate the most important classifications where the model performed well. CE is lowest for the river surface, while OE was minimal for the riverbed and river surface, highlighting the model&#x2019;s accuracy in detecting these features. Our training data was focused on areas along river corridors, resulting in less training data from the ground and other classes. This focus aligns with our application to classify river surface and riverbed classes for depth correction. However, PointCNN also shows potential for classifying non-bathymetric classes with further training.</p>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Performance metrics for the Snake river, WY, S &#x3d; 50 and P &#x3d; 16,384.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Class</th>
<th align="center">IoU</th>
<th align="center">Accuracy</th>
<th align="center">Precision</th>
<th align="center">Recall</th>
<th align="center">F1</th>
<th align="center">Kappa</th>
<th align="center">CE</th>
<th align="center">OE</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">River Surface</td>
<td align="center">
<bold>0.89</bold>
</td>
<td align="center">0.97</td>
<td align="center">0.95</td>
<td align="center">0.93</td>
<td align="center">0.94</td>
<td align="center">
<bold>0.92</bold>
</td>
<td align="center">
<bold>0.05</bold>
</td>
<td align="center">
<bold>0.07</bold>
</td>
</tr>
<tr>
<td align="left">Riverbed</td>
<td align="center">0.70</td>
<td align="center">0.90</td>
<td align="center">0.77</td>
<td align="center">0.88</td>
<td align="center">0.82</td>
<td align="center">0.76</td>
<td align="center">0.23</td>
<td align="center">0.12</td>
</tr>
<tr>
<td align="left">Ground</td>
<td align="center">0.67</td>
<td align="center">0.90</td>
<td align="center">0.78</td>
<td align="center">0.83</td>
<td align="center">0.80</td>
<td align="center">0.74</td>
<td align="center">0.22</td>
<td align="center">0.17</td>
</tr>
<tr>
<td align="left">Other</td>
<td align="center">0.68</td>
<td align="center">0.91</td>
<td align="center">0.91</td>
<td align="center">0.73</td>
<td align="center">0.81</td>
<td align="center">0.76</td>
<td align="center">0.09</td>
<td align="center">0.27</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3-5">
<title>3.5 Depth adjustment</title>
<p>A detailed analysis of a river segmentation using both a top-down and a cross-section view derived from the lidar data is shown in <xref ref-type="fig" rid="F8">Figure 8</xref>. The top section of the image presents a classified map of the river and its surrounding areas, with different colors representing the classes. The cross-section view compares adjusted and unadjusted riverbed depths from the river surface. The raw data is biased due to the bending and slowing of light as it passes through the water. The adjusted depth from the river surface is measured at 0.86&#xa0;m, accounting for the refraction and distortion effects caused by the interaction of the laser pulses with the water column. By comparing these two measurements, the image highlights the significant impact of the change in speed of light on the accuracy of bathymetric data. Adjusting for these distortions is crucial for producing reliable maps of underwater features.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>The plan view at the top displays the spatial distribution of different classes. The profile below shows the unadjusted and adjusted depths of riverbed class, along with other classes in a vertical view.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g008.tif">
<alt-text content-type="machine-generated">3D visualization of a river section, with labeled regions: red for unadjusted riverbed, blue for adjusted riverbed, cyan for river surface, maroon for ground, and green for others. A vertical cross-section shows unadjusted riverbed depth at 1.14 meters and adjusted depth at 0.86 meters from the river surface. Arrow indicates north direction.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3-6">
<title>3.6 Depth evaluation with ADCP measurements</title>
<p>To evaluate our automated bathymetry lidar depth correction workflow, we validated our results using 1,595 independent high-accuracy ADCP depth measurements collected from the clear-water Snake River (<xref ref-type="bibr" rid="B29">Legleiter et al., 2016</xref>). Metrics used for the depth accuracy validation are based on <xref ref-type="bibr" rid="B40">Pan et al. (2015a)</xref> and <xref ref-type="bibr" rid="B29">Legleiter et al. (2016)</xref>. Mean error was used to quantify the systematic offset between predicted and true depths. The standard deviation of errors measures the variability of individual errors around the mean, reflecting the consistency of predictions. Root mean square error (RMSE) combines both bias and variability into a single metric, giving greater weight to larger errors. Mean absolute error (MAE) captures the typical magnitude of errors without considering their direction. Finally, the coefficient of determination (R<sup>2</sup>) assessed how well the predicted depths explained the variability in the true depth measurements, with values closer to 1 indicating stronger predictive performance.</p>
<p>Initially, our uncorrected bathymetric data showed a mean bias of 45&#xa0;cm, indicating systematic overestimation. After refraction correction, this bias was eliminated, reducing the mean bias to 2&#xa0;cm as shown in <xref ref-type="table" rid="T6">Table 6</xref>. This improvement highlights the effectiveness of the method in addressing systematic depth overestimation. Compared to previous studies using the same dataset, our bathymetry depth accuracy show improved mean bias. <xref ref-type="bibr" rid="B40">Pan et al. (2015a)</xref> and <xref ref-type="bibr" rid="B29">Legleiter et al. (2016)</xref> reported mean bias of approximately 6&#xa0;cm and 8&#xa0;cm, respectively, both higher than the 2&#xa0;cm error achieved by our automated method. The RMSE decreased from 0.47&#xa0;m to 0.16&#xa0;m after correction, and the MAE decreased from 0.45&#xa0;m to 0.12&#xa0;m. We observed a slight increase in the standard deviation of errors, from 0.12&#xa0;m before correction to 0.15&#xa0;m. This increase indicates greater variability among corrected measurements rather than persistent systematic errors. The model significantly improves bathymetric accuracy by effectively eliminating bias and achieving precision on par with previously published continuous wavelet transform (CWT) based approaches. These results confirm the reliability and consistency of our workflow for automated bathymetric depth estimation.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Comparison of depth accuracy for Snake River dataset.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Study</th>
<th align="center">Mean (m)</th>
<th align="center">Std.&#xa0;Dev. (m)</th>
<th align="center">
<inline-formula id="inf49">
<mml:math id="m59">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">RMSE (m)</th>
<th align="center">MAE (m)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Our Model (Before depth correction)</td>
<td align="center">0.45</td>
<td align="center">0.12</td>
<td align="center">0.95</td>
<td align="center">0.47</td>
<td align="center">0.45</td>
</tr>
<tr>
<td align="left">
<xref ref-type="bibr" rid="B40">Pan et al. (2015a)</xref>
</td>
<td align="center">0.06</td>
<td align="center">0.14</td>
<td align="center">0.93</td>
<td align="center">N/A</td>
<td align="center">N/A</td>
</tr>
<tr>
<td align="left">
<xref ref-type="bibr" rid="B29">Legleiter et al. (2016)</xref>
</td>
<td align="center">0.08</td>
<td align="center">0.16</td>
<td align="center">0.95</td>
<td align="center">N/A</td>
<td align="center">N/A</td>
</tr>
<tr>
<td align="left">Our Model (After depth correction)</td>
<td align="center">
<bold>0.02</bold>
</td>
<td align="center">0.15</td>
<td align="center">0.95</td>
<td align="center">0.16</td>
<td align="center">0.12</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-7">
<title>3.7 Error analysis</title>
<p>The misclassification of the riverbed class as ground can be attributed to several factors. The riverbed and the ground have similar characteristics, especially near the edges of the river where small changes in water level is the only factor governing the difference between ground and riverbed classes. This likely confused classification algorithms, leading to misclassification. Vegetation and debris on the riverbed further complicated classification, as plants, fallen branches, or other debris were mistaken for ground. These misclassifications often occurred around the river&#x2019;s edges, as shown in <xref ref-type="fig" rid="F9">Figure 9a</xref>. One thing to note is that this occurs in very shallow water regions, where the elevation of the ground and water is nearly the same, and the river surface class is absent because the river surface and benthic layer returns overlap in the lidar response (<xref ref-type="bibr" rid="B40">Pan et al., 2015a</xref>).</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>
<bold>(a)</bold> Plan and profile view of two representative areas along the Snake River. The highlighted areas demonstrate locations where the algorithm misclassified data, particularly around the river edges. The circled points are areas where the riverbed class is misclassified as ground and <italic>vice versa</italic>. Histograms showing the frequency distribution of riverbed points misclassified as <bold>(b)</bold> ground and <bold>(c)</bold> river surface, plotted against their depth values (distance from river surface to benthic layer) for the Snake River dataset. The red line in each subplot indicates the depth error that would result from the misclassification due to the improper application of the speed of light correction. Misclassifications are more frequent at shallower depths, with errors increasing with water depth, the overall impact on point cloud accuracy remains minimal.</p>
</caption>
<graphic xlink:href="frsen-06-1521446-g009.tif">
<alt-text content-type="machine-generated">Top image shows vertical cross-sections of terrain classifications with colors indicating misclassified areas, riverbed, river surface, ground, and others. Misclassifications are highlighted. Bottom left graph (b) shows frequency of riverbed points misclassified as river surface, with depth error line. Bottom right graph (c) shows frequency of riverbed points misclassified as ground, with depth error line. Total misclassified points are labeled in both graphs.</alt-text>
</graphic>
</fig>
<p>In <xref ref-type="fig" rid="F9">Figure 9b</xref>, the histogram (blue bars) shows that the majority of misclassifications occur at shallower depths, with the frequency decreasing as depth increases. A total of 431,225 points are misclassified. The red line represents the depth error, which rises linearly from 0.0 to 0.8&#xa0;m as the depth ranges from 0.0 to 2.5&#xa0;m. Misclassifications are more frequent at shallower depths, with the errors increasing with water depth. However, because a majority of the misclassifications are in shallow areas, the overall point cloud errors due to the improper application of refraction and speed of light corrections is minimal for this majority of misclassified points. The misclassification for river-related classes decreases with increasing water depth and shallow river regions are more susceptible to errors. As depicted in <xref ref-type="fig" rid="F9">Figure 9c</xref>, most misclassifications between the riverbed and river surface classes occur within the 0.4&#xa0;m depth range. The histogram (blue bars) reveals the majority of misclassifications occur at shallower depths, with the frequency decreasing as depth increases. A total of 127,162 points are misclassified. The red line illustrates the depth error, which rises linearly from 0.0 to 0.2&#xa0;m as the depth increases from 0.0 to 0.6&#xa0;m. The algorithm struggles to accurately differentiate between the benthic layer and the water surface in extremely shallow areas (below 20&#xa0;cm), due to the laser pulse width of the bathymetric lidar system (8&#xa0;ns for Aquarius (<xref ref-type="bibr" rid="B18">Fernandez-Diaz et al., 2014</xref>) and 1.5&#xa0;ns for the Riegl 840-G (<xref ref-type="bibr" rid="B37">Mandlburger et al., 2020</xref>). The shallow water lidar returns are normally a convolution of the water surface and benthic layer, and therefore distinguishing these two surfaces from convolved lidar returns is difficult, both for the lidar processing algorithms (e.g., <xref ref-type="bibr" rid="B47">Schwarz et al., 2019</xref>) and the machine learning results presented herein. Since most misclassifications occur at very shallow depths where the differences between the riverbed and surface are minimal, the impact due to incorrect application of depth corrections and therefore the accuracy of the final point cloud is minor. <xref ref-type="fig" rid="F9">Figures 9b,c</xref> show that although the algorithm struggles in shallow regions, the depth error caused by misclassification remains relatively small and localized for the majority of the misclassified points. As a result, the overall effect on point cloud accuracy is minimal.</p>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<title>4 Discussion</title>
<sec id="s4-1">
<title>4.1 Model generalization</title>
<p>The strong agreement between our corrected lidar depths and independent ADCP measurements on the Snake River demonstrates the robustness of the semantic segmentation results. The high classification accuracy achieved for the river surface and riverbed classes suggests that the model generalized well across different river environments, despite being trained on datasets from geographically distinct rivers. Because accurate river surface and riverbed identification is important for reliable depth estimation, the close match with the ADCP ground truth confirms that the semantic segmentation was precise and can support real-world depth correction applications. These findings indicate that our model effectively transfers across diverse bathymetric conditions, leading to accurate and consistent results even on an unseen river system. Because of the strong generalization performance, we found no need for additional data or further data augmentation techniques for training.</p>
</sec>
<sec id="s4-2">
<title>4.2 Computational efficiency and scalability</title>
<p>Deep learning models are often computationally intensive, particularly for large-scale 3D point cloud data. This was evident with DGCNN, which required nearly 7&#xa0;h per epoch with a limited batch size of 2 due to its dynamic graph recomputation at every layer. In contrast, PointCNN demonstrated strong computational efficiency by completing each training epoch in approximately 2&#xa0;h and 50&#xa0;min with a larger batch size of 6. This efficiency comes from the fixed neighborhood structure, where local point relationships are established once during initial sampling and reused throughout the network, avoiding the costly repeat of graph construction. During inference, PointCNN further demonstrated scalability by classifying the entire Snake River dataset (approximately 20 million points) in just 12&#xa0;min. These results show that by integrating an efficient deep learning model like PointCNN into our workflow, we achieve fast, scalable, and practical processing for large-scale bathymetric lidar applications.</p>
</sec>
<sec id="s4-3">
<title>4.3 Model agnostic</title>
<p>Although PointCNN yielded the strongest results in our experiments, the primary aim of this study was not to optimize a specific architecture but to establish a generalizable workflow for automating bathymetric lidar classification and depth correction. The segmentation stage was designed to be modular, allowing for the substitution of any deep learning model capable of distinguishing the river surface from the riverbed points. Numerous point-cloud segmentation architectures have been developed in recent years with varying trade-offs in accuracy, complexity, and computational efficiency. Our workflow demonstrated that such models can be integrated interchangeably without altering the downstream refraction-correction or depth-estimation components. This flexibility enables the method to benefit from future advancements in 3D point-cloud learning and provides a robust foundation for developing fully automated bathymetric processing pipelines.</p>
</sec>
<sec id="s4-4">
<title>4.4 Practical application</title>
<p>This study presents an automated workflow specifically aimed at supporting engineering surveys that require accurate bathymetric mapping. By validating our corrected lidar depths against high-accuracy ADCP measurements, we demonstrated that the method reliably captures riverbed topography with minimal bias. The ability to generate precise riverbed digital elevation models (DEMs) or estimate river volumes makes this approach highly suitable for bathymetric engineering surveys. By automating the classification process, the workflow enables large-scale surveys to reduce time, cost, and manual effort traditionally associated with riverine bathymetric surveys.</p>
</sec>
<sec id="s4-5">
<title>4.5 Misclassification in shallow water</title>
<p>In PointCNN, X-Conv operations transform local neighborhoods relative to a center point by applying a learned transformation matrix. The effectiveness of classification relies on meaningful spatial differences <inline-formula id="inf50">
<mml:math id="m60">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and feature differences among neighboring points. In shallow water regions, river surface and riverbed points have minimal vertical separation. Similarly, along riverbanks, the horizontal and vertical distances between riverbed and ground points are small, resulting in overlapping local geometries. When such local differences diminish, the learned transformation matrix <inline-formula id="inf51">
<mml:math id="m61">
<mml:mrow>
<mml:mi>X</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> cannot effectively distinguish between classes, leading to increased misclassification errors.</p>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>5 Conclusion</title>
<p>The high scores achieved for both water surface and underwater feature segmentation highlight that deep learning models, such as PointCNN, can differentiate and classify accurately. We used about 275&#xa0;km of river corridor bathymetric lidar data for training and testing. Specifically, 27&#xa0;km from the Swan River and 204&#xa0;km from the Eel River were used for training, while 44&#xa0;km from the Snake River were used for testing and evaluation. This comprehensive dataset provided a robust foundation for evaluating the effectiveness of our methods to identify and classify river surfaces and riverbeds. We only trained and tested the model on subsets of ALS data using narrow strips along the river corridor. Despite the variability in sensor types, data density, and locations, the performance of the model was not affected, demonstrating that it is weakly correlated with these three factors.</p>
<p>The model achieved high accuracy using only the basic x, y, and z point cloud coordinates as features, without the need for additional attributes such as intensity or number of returns. Only the coordinates were sufficient for classification, indicating that semantic segmentation can be effectively achieved by focusing solely on scene geometry. The structural (geometrical) characteristics of the four classes used in this study were distinct enough for the model to learn their differences.</p>
<p>The experiments highlighted the optimal configurations for block size and the number of points in a block. The 50 m block size and 16,384 points per block emerged as the most effective setup for achieving high precision and recall in segmenting narrow and shallow rivers. The 3D CNN(PointCNN) model&#x2019;s performance was also better than a graph-based model (DGCNN), providing better precision and recall in our segmentation tasks. Successfully using the semantic segmentation of PointCNN to automate riverbed data depth correction eliminates the need for manual classification and significantly enhances efficiency. The model enables precise depth corrections by achieving IoU scores of 0.89 for river surface and 0.7 for riverbed classifications, accounting for light refraction effects in water. Errors were primarily in shallow water, where the depth correction has a negligible effect on the final point cloud accuracy.</p>
<p>The independent validation using ADCP measurements highlights the robustness and accuracy of our depth correction workflow. Achieving a mean depth bias of only 2&#xa0;cm and an RMSE of 16&#xa0;cm confirms that the automated method produces results consistent with the field data. This close agreement with ADCP ground truth demonstrates that the model both generalizes well to unseen river environments and meets the accuracy standards required for real-world engineering applications. ADCP validation showed the effectiveness of our approach in delivering operationally reliable bathymetric data suitable for detailed bathymetry mapping.</p>
<p>In summary, the application of deep learning in our study significantly improves traditional manual processes. Conventional methods of analyzing point cloud data are time-consuming, labor-intensive, and prone to human error. Our automated deep learning approach reduces time and resource demands while ensuring consistent and precise analysis of point cloud data. This automation enhances the reliability and efficiency of bathymetry mapping, making it a valuable tool for researchers and professionals in the field.</p>
</sec>
<sec id="s6">
<title>6 Limitations</title>
<p>Experiments using the number of returns as an additional feature revealed that the x, y, and z coordinates alone provided sufficient geometric information for the model to accurately differentiate between the river surface and riverbed, rendering the number of returns unnecessary. We also considered using intensity as a feature, focusing on first returns only to avoid the potential misleading effects of partial returns caused by factors like water surface reflection and vegetation interference. However, this approach led to significant data imbalance, making it impractical to include intensity as a reliable feature in the model.</p>
</sec>
<sec id="s7">
<title>7 Future work</title>
<p>A potential improvement could be introduced by using attention mechanisms in the network architecture. In an attention-based framework, rather than treating all neighboring points equally, the model learns to assign different importance weights to each neighbor based on the relative spatial and feature differences from the center point. This dynamic weighting allows the network to focus on the most informative neighbors, improving its ability to distinguish between river surface, riverbed, and ground points, even under conditions of low geometric contrast. While similar attention-based concepts have been successfully applied to point cloud analysis in other domains, their potential for improving the semantic segmentation of bathymetric lidar remains largely unexplored. Future work could investigate adapting such approaches to further enhance classification accuracy and depth correction performance in challenging shallow water and riverbank environments.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s8">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="sec" rid="s14">Supplementary Material</xref>, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s9">
<title>Author contributions</title>
<p>OP: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Validation, Visualization, Writing &#x2013; original draft, Writing &#x2013; review and editing. NE: Conceptualization, Methodology, Supervision, Writing &#x2013; review and editing. CLG: Conceptualization, Supervision, Writing &#x2013; review and editing.</p>
</sec>
<sec sec-type="funding-information" id="s10">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. Partial funding for the authors was provided by a facility grant from the National Science Foundation (EAR/IF &#x23;1830734) and a grant from the U.S. Army Corps of Engineers Cold Regions Research and Engineering.</p>
</sec>
<ack>
<p>The authors would like to thank Abhinav Singhania with NCALM for his invaluable contribution in manually classifying bathymetric lidar datasets used to train and test our models in this research.</p>
</ack>
<sec sec-type="COI-statement" id="s11">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s12">
<title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="s13">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s14">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frsen.2025.1521446/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frsen.2025.1521446/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="Supplementaryfile1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>A Arief</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Arief</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bhat</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Indahl</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Tveite</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Density-adaptive sampling for heterogeneous point cloud object segmentation in autonomous vehicle applications</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops</conf-name>.</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Agrafiotis</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Skarlatos</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Georgopoulos</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Karantzalos</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Depthlearn: learning to correct the refraction on point clouds derived from aerial imagery for accurate dense shallow water bathymetry based on svms-fusion with lidar point clouds</article-title>. <source>Remote Sens.</source> <volume>11</volume>, <fpage>2225</fpage>. <pub-id pub-id-type="doi">10.3390/rs11192225</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Albright</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Nearshore bathymetry from fusion of sentinel-2 and icesat-2 observations</article-title>. <source>IEEE Geoscience Remote Sens. Lett.</source> <volume>18</volume>, <fpage>900</fpage>&#x2013;<lpage>904</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2020.2987778</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Anderson</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pitlick</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Using repeat lidar to estimate sediment transport in a steep stream</article-title>. <source>J. Geophys. Res. (Earth Surf.)</source> <volume>119</volume>, <fpage>621</fpage>&#x2013;<lpage>643</lpage>. <pub-id pub-id-type="doi">10.1002/2013JF002933</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arief</surname>
<given-names>H. A.</given-names>
</name>
<name>
<surname>Strand</surname>
<given-names>G.-H.</given-names>
</name>
<name>
<surname>Tveite</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Indahl</surname>
<given-names>U. G.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Land cover segmentation of airborne lidar data using stochastic atrous network</article-title>. <source>Remote Sens.</source> <volume>10</volume>, <fpage>973</fpage>. <pub-id pub-id-type="doi">10.3390/rs10060973</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Atzmon</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Maron</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lipman</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Point convolutional neural networks by extension operators</article-title>. <source>ACM TOG</source> <volume>37</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1145/3197517.3201301</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Axelsson</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2000</year>). &#x201c;<article-title>Dem generation from laser scanner data using adaptive tin models</article-title>,&#x201d; in <source>International archives of photogrammetry and remote sensing</source> (<publisher-loc>Amsterdam, Netherlands</publisher-loc>: <publisher-name>International Society for Photogrammetry and Remote Sensing</publisher-name>), <volume>33</volume>, <fpage>111</fpage>&#x2013;<lpage>118</lpage>.</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bello</surname>
<given-names>S. A.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Adam</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Review: deep learning on 3d point clouds</article-title>. <source>Remote Sens.</source> <volume>12</volume>, <fpage>1729</fpage>. <pub-id pub-id-type="doi">10.3390/rs12111729</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bouhdaoui</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bailly</surname>
<given-names>J.-S.</given-names>
</name>
<name>
<surname>Baghdadi</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Abady</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Modeling the water bottom geometry effect on peak time shifting in lidar bathymetric waveforms</article-title>. <source>IEEE Geoscience Remote Sens. Lett.</source> <volume>11</volume>, <fpage>1285</fpage>&#x2013;<lpage>1289</lpage>. <pub-id pub-id-type="doi">10.1109/lgrs.2013.2292814</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Boulch</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Convpoint: continuous convolutions for point cloud processing</article-title>. <source>Comput. Graph. Forum</source> <volume>88</volume>, <fpage>24</fpage>&#x2013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1016/j.cag.2020.02.005</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Choi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Kwak</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Part-aware data augmentation for 3d object detection in point cloud</article-title>. <source>Computer Vision and Pattern Recognition</source>, <fpage>3391</fpage>&#x2013;<lpage>397</lpage>. <pub-id pub-id-type="doi">10.1109/iros51168.2021.9635887</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Daniel</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Dupont</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Investigating fully convolutional network to semantic labelling of bathymetric point cloud</article-title>. <source>ISPRS Ann. Photogrammetry, Remote Sens. Spatial Inf. Sci.</source> <volume>2</volume>, <fpage>657</fpage>&#x2013;<lpage>663</lpage>. <pub-id pub-id-type="doi">10.5194/isprs-annals-V-2-2020-657-2020</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dietrich</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2014</year>). <source>Eel river critical zone observatory July 2014 lidar survey</source>. <publisher-name>OpenTopography</publisher-name>.</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ekhtari</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Fernandez-Diaz</surname>
<given-names>J. C.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Classification of airborne multispectral lidar point clouds for land cover mapping</article-title>. <source>IEEE J. Sel. Top. Appl. Earth Observations Remote Sens.</source> <volume>11</volume>, <fpage>2068</fpage>&#x2013;<lpage>2078</lpage>. <pub-id pub-id-type="doi">10.1109/JSTARS.2018.2835483</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Erena</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Atenza</surname>
<given-names>J. F.</given-names>
</name>
<name>
<surname>Garc&#xed;a-Galiano</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Dom&#xed;nguez</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Bernab&#xe9;</surname>
<given-names>J. M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Use of drones for the topo-bathymetric monitoring of the reservoirs of the segura river basin</article-title>. <source>Water</source> <volume>11</volume>, <fpage>445</fpage>. <pub-id pub-id-type="doi">10.3390/w11030445</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<collab>Esri</collab> (<year>2024</year>). <article-title>Arcgis, version 2.1.0.2</article-title>. <publisher-name>Esri Inc</publisher-name>. <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://pypi.org/project/arcgis/">https://pypi.org/project/arcgis/</ext-link>. (Accessed August 1, 2025).</comment>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fareed</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Flores</surname>
<given-names>J. P.</given-names>
</name>
<name>
<surname>Das</surname>
<given-names>A. K.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Analysis of uas-lidar ground points classification in agricultural fields using traditional algorithms and pointcnn</article-title>. <source>Remote Sens.</source> <volume>15</volume>, <fpage>483</fpage>. <pub-id pub-id-type="doi">10.3390/rs15020483</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fernandez-Diaz</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Carter</surname>
<given-names>W. E.</given-names>
</name>
<name>
<surname>Shrestha</surname>
<given-names>R. L.</given-names>
</name>
<name>
<surname>Sartori</surname>
<given-names>M. P.</given-names>
</name>
<name>
<surname>Singhania</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Early results of simultaneous terrain and shallow water bathymetry mapping using a single-wavelength airborne lidar sensor</article-title>. <source>IEEE J. Sel. Top. Appl. Earth Observations Remote Sens.</source> <volume>7</volume>, <fpage>623</fpage>&#x2013;<lpage>635</lpage>. <pub-id pub-id-type="doi">10.1109/JSTARS.2013.2265255</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Graham</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Engelcke</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>van der Maaten</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>3D semantic segmentation with submanifold sparse convolutional networks</article-title>. <source>CVPR</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1711.10275</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Guo</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Tian</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Gaussian half-wavelength progressive decomposition method for waveform processing of airborne laser bathymetry</article-title>. <source>Remote Sens.</source> <volume>10</volume>, <fpage>35</fpage>. <pub-id pub-id-type="doi">10.3390/rs10010035</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hell</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Brandmeier</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Briechle</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Krzystek</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Classification of tree species and standing dead trees with lidar point clouds using two deep neural networks: Pointcnn and 3dmfv-net</article-title>. <source>Open Geospatial Data, Softw. Stand.</source> <volume>7</volume>, <fpage>103</fpage>&#x2013;<lpage>121</lpage>. <pub-id pub-id-type="doi">10.1007/s41064-022-00200-4</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Rosa</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Markham</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Trigoni</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Randla-net: efficient semantic segmentation of large-scale point clouds</article-title>. <source>CVPR</source>. <pub-id pub-id-type="doi">10.48550/arXiv.1911.11236</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hyejin Kim</surname>
<given-names>J. L.</given-names>
</name>
<name>
<surname>Jaehoon</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wie</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Water bottom and surface classification algorithm for bathymetric lidar point clouds of very shallow waters</article-title>. <source>Can. J. Remote Sens.</source> <volume>49</volume>, <fpage>2172957</fpage>. <pub-id pub-id-type="doi">10.1080/07038992.2023.2172957</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jing</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Guan</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zang</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Multispectral lidar point cloud classification using se-pointnet&#x2b;&#x2b;</article-title>. <source>Remote Sens.</source> <volume>13</volume>, <fpage>2516</fpage>. <pub-id pub-id-type="doi">10.3390/rs13132516</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Koguciuk</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Chechlinski</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>El-Gaaly</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>3D object recognition with ensemble learning&#x2014;a study of point cloud-based deep learning models</article-title>, <source>Computer Vision and Pattern Recognition</source>, <fpage>100</fpage>&#x2013;<lpage>114</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-33723-0_9</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Nandy</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Soni</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Lohani</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Powerline extraction from aerial and mobile lidar data using deep learning</article-title>. <source>Earth Sci. Inf.</source> <volume>17</volume>, <fpage>2819</fpage>&#x2013;<lpage>2833</lpage>. <pub-id pub-id-type="doi">10.1007/s12145-024-01310-w</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Legleiter</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2024</year>). <source>Lidar survey of the snake river, WY 2012</source>. <publisher-name>OpenTopography</publisher-name>.</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Legleiter</surname>
<given-names>C. J.</given-names>
</name>
<name>
<surname>Overstreet</surname>
<given-names>B. T.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Fernandez-Diaz</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Singhania</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Evaluating the capabilities of the casi hyperspectral imaging system and aquarius bathymetric lidar for measuring channel morphology in two distinct river environments</article-title>. <source>Earth Surf. Process. Landforms</source> <volume>41</volume>, <fpage>344</fpage>&#x2013;<lpage>363</lpage>. <pub-id pub-id-type="doi">10.1002/esp.3794</pub-id>
</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Letard</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Collin</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Corpetti</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lague</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Pastol</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ekelund</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Classification of land-water continuum habitats using exclusively airborne topobathymetric lidar green waveforms and infrared intensity point clouds</article-title>. <source>Remote. Sens.</source> <volume>14</volume>, <fpage>341</fpage>. <pub-id pub-id-type="doi">10.3390/rs14020341</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Letard</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lague</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Le Guennec</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Lef&#xe8;vre</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Feldmann</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Leroy</surname>
<given-names>P.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>3dmasc: accessible, explainable 3d point clouds classification. application to bi-spectral topo-bathymetric lidar data</article-title>. <source>ISPRS J. Photogrammetry Remote Sens.</source> <volume>207</volume>, <fpage>175</fpage>&#x2013;<lpage>197</lpage>. <pub-id pub-id-type="doi">10.1016/j.isprsjprs.2023.11.022</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bu</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Di</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Pointcnn: convolution on x-transformed points</article-title>,&#x201d; in <source>Advances in neural information processing systems</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Bengio</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wallach</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Larochelle</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Grauman</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Cesa-Bianchi</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Garnett</surname>
<given-names>R.</given-names>
</name>
</person-group> (<publisher-name>Montr&#x00e9;al, Canada: Curran Associates, Inc.</publisher-name>), <volume>31</volume>.</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lumban-Gaol</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Amri</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Rachmadi</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Nugroho</surname>
<given-names>H. A.</given-names>
</name>
<name>
<surname>Arymurthy</surname>
<given-names>A. M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>A comparative study of point clouds semantic segmentation using three different neural networks on the railway station dataset</article-title>. <source>Int. Archives Photogrammetry, Remote Sens. Spatial Inf. Sci.</source> <volume>XLIII-B3-2021</volume>, <fpage>223</fpage>&#x2013;<lpage>229</lpage>. <pub-id pub-id-type="doi">10.5194/isprs-archives-XLIII-B3-2021-223-2021</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mandlburger</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Hauer</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wieser</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Pfeifer</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Topo-bathymetric lidar for monitoring river morphodynamics and instream habitats&#x2014;a case study at the pielach river</article-title>. <source>Remote Sens.</source> <volume>7</volume>, <fpage>6160</fpage>&#x2013;<lpage>6195</lpage>. <pub-id pub-id-type="doi">10.3390/rs70506160</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mandlburger</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>K&#xf6;lle</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>N&#xfc;bel</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Soergel</surname>
<given-names>U.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Bathynet: a deep neural network for water depth mapping from multispectral aerial images</article-title>. <source>PFG &#x2013; J. Photogrammetry, Remote Sens. Geoinformation Sci.</source> <volume>89</volume>, <fpage>71</fpage>&#x2013;<lpage>89</lpage>. <pub-id pub-id-type="doi">10.1007/s41064-021-00142-3</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mandlburger</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Pfennigbauer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Schwarz</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fl&#xf6;ry</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Nussbaumer</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Concept and performance evaluation of a novel uav-borne topo-bathymetric lidar sensor</article-title>. <source>Remote Sens.</source> <volume>12</volume>, <fpage>986</fpage>. <pub-id pub-id-type="doi">10.3390/rs12060986</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Marshall</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2023</year>). <source>Lidar survey to evaluate spatial heterogeneity and channel evolution, mt 2022</source>. <publisher-name>OpenTopography</publisher-name>.</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Fernandez-Diaz</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C. L.</given-names>
</name>
<name>
<surname>Starek</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Shallow water seagrass observed by high resolution full waveform bathymetric lidar</article-title>. <source>IEEE Geoscience Remote Sens. Symposium</source>, <fpage>1341</fpage>&#x2013;<lpage>1344</lpage>. <pub-id pub-id-type="doi">10.1109/IGARSS.2014.6946682</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Hartzell</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Fernandez-Diaz</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Legleiter</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Overstreet</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2015a</year>). <article-title>Performance assessment of high resolution airborne full waveform lidar for shallow river bathymetry</article-title>. <source>Remote Sens.</source> <volume>7</volume>, <fpage>5133</fpage>&#x2013;<lpage>5159</lpage>. <pub-id pub-id-type="doi">10.3390/rs70505133</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pan</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Glennie</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Legleiter</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Overstreet</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2015b</year>). <article-title>Estimation of water depths and turbidity from hyperspectral imagery using support vector regression</article-title>. <source>IEEE Geoscience Remote Sens. Lett.</source> <volume>12</volume>, <fpage>2165</fpage>&#x2013;<lpage>2169</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2015.2453636</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Pfennigbauer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Rieger</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Schwarz</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Ullrich</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2022</year>). &#x201c;<article-title>Impact of beam parameters on the performance of a topo-bathymetric lidar sensor</article-title>,&#x201d; in <source>Laser radar technology and applications XXVII</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Kamerman</surname>
<given-names>G. W.</given-names>
</name>
<name>
<surname>Magruder</surname>
<given-names>L. A.</given-names>
</name>
<name>
<surname>Turner</surname>
<given-names>M. D.</given-names>
</name>
</person-group> (<publisher-loc>SPIE</publisher-loc>: <publisher-name>International Society for Optics and Photonics</publisher-name>), <fpage>12110</fpage>. <pub-id pub-id-type="doi">10.1117/12.2618794</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pricope</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Bashit</surname>
<given-names>M. S.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Emerging trends in topobathymetric lidar technology and mapping</article-title>. <source>Int. J. Remote Sens.</source> <volume>44</volume>, <fpage>7706</fpage>&#x2013;<lpage>7731</lpage>. <pub-id pub-id-type="doi">10.1080/01431161.2023.2287564</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Qi</surname>
<given-names>C. R.</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Mo</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Guibas</surname>
<given-names>L. J.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Pointnet: deep learning on point sets for 3d classification and segmentation</article-title>. <source>CoRR</source>, <fpage>1612.00593</fpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1612.00593</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Qi</surname>
<given-names>C. R.</given-names>
</name>
<name>
<surname>Yi</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Su</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Guibas</surname>
<given-names>L. J.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Pointnet&#x2b;&#x2b;: deep hierarchical feature learning on point sets in a metric space</article-title>,&#x201d; in <source>Advances in neural information processing systems</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Guyon</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Luxburg</surname>
<given-names>U. V.</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Wallach</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Fergus</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Vishwanathan</surname>
<given-names>S.</given-names>
</name>
</person-group>, (<publisher-name>Long Beach, CA: Curran Associates, Inc.</publisher-name>), <fpage>30</fpage>.</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Scarselli</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Gori</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Tsoi</surname>
<given-names>A. C.</given-names>
</name>
<name>
<surname>Hagenbuchner</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Monfardini</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>The graph neural network model</article-title>. <source>IEEE Trans. Neural Netw.</source> <volume>20</volume>, <fpage>61</fpage>&#x2013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1109/TNN.2008.2005605</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schwarz</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Mandlburger</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Pfennigbauer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Pfeifer</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Design and evaluation of a full-wave surface and bottom-detection algorithm for lidar bathymetry of very shallow waters</article-title>. <source>ISPRS J. Photogrammetry Remote Sens.</source> <volume>150</volume>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1016/j.isprsjprs.2019.02.002</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schwarz</surname>
<given-names>R. K.</given-names>
</name>
<name>
<surname>Pfeifer</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Pfennigbauer</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Mandlburger</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Depth measurement bias in pulsed airborne laser hydrography induced by chromatic dispersion</article-title>. <source>IEEE Geoscience Remote Sens. Lett.</source> <volume>18</volume>, <fpage>1332</fpage>&#x2013;<lpage>1336</lpage>. <pub-id pub-id-type="doi">10.1109/LGRS.2020.3003088</pub-id>
</citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Shi</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Pv-rcnn: point-voxel feature set abstraction for 3d object detection</article-title>.</citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shin</surname>
<given-names>Y.-H.</given-names>
</name>
<name>
<surname>Son</surname>
<given-names>K.-W.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>D.-C.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Semantic segmentation and building extraction from airborne lidar data with multiple return using pointnet&#x2b;&#x2b;</article-title>. <source>Appl. Sci.</source> <volume>12</volume>, <fpage>1975</fpage>. <pub-id pub-id-type="doi">10.3390/app12041975</pub-id>
</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tchapmi</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Choy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Armeni</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Gwak</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Savarese</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Segcloud: semantic segmentation of 3d point clouds</article-title>. <source>3DV</source>, <fpage>537</fpage>&#x2013;<lpage>547</lpage>. <pub-id pub-id-type="doi">10.1109/3dv.2017.00067</pub-id>
</citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Teng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Vaze</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dutta</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Marvanek</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Rapid inundation modelling in large floodplains using lidar dem</article-title>. <source>Water Resour. Manag.</source> <volume>29</volume>, <fpage>2619</fpage>&#x2013;<lpage>2636</lpage>. <pub-id pub-id-type="doi">10.1007/s11269-015-0960-8</pub-id>
</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Thomas</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Deschaud</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Marcotegui</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Goulette</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Guibas</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Kpconv: flexible and deformable convolution for point clouds</article-title>. <source>ICCV</source>, <fpage>6410</fpage>&#x2013;<lpage>6419</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2019.00651</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tsai</surname>
<given-names>C.-M.</given-names>
</name>
<name>
<surname>Lai</surname>
<given-names>Y.-H.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>Y.-D.</given-names>
</name>
<name>
<surname>Chung</surname>
<given-names>Y.-J.</given-names>
</name>
<name>
<surname>Perng</surname>
<given-names>J.-W.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Multi-dimensional underwater point cloud detection based on deep learning</article-title>. <source>Sensors</source> <volume>21</volume>, <fpage>884</fpage>. <pub-id pub-id-type="doi">10.3390/s21030884</pub-id>
</citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>C.-K.</given-names>
</name>
<name>
<surname>Philpot</surname>
<given-names>W. D.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Using airborne bathymetric lidar to detect bottom type variation in shallow waters</article-title>. <source>Remote Sens. Environ.</source> <volume>106</volume>, <fpage>123</fpage>&#x2013;<lpage>135</lpage>. <pub-id pub-id-type="doi">10.1016/j.rse.2006.08.003</pub-id>
</citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Solomon</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2019a</year>). <article-title>Graph attention convolution for point cloud semantic segmentation</article-title>. <source>CVPR</source>, <fpage>10288</fpage>&#x2013;<lpage>10297</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2019.01054</pub-id>
</citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Sarma</surname>
<given-names>S. E.</given-names>
</name>
<name>
<surname>Bronstein</surname>
<given-names>M. M.</given-names>
</name>
<name>
<surname>Solomon</surname>
<given-names>J. M.</given-names>
</name>
</person-group> (<year>2019b</year>). <article-title>Dynamic graph cnn for learning on point clouds</article-title>. <source>ACM Trans. Graph.</source> <volume>38</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1145/3326362</pub-id>
</citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Widyaningrum</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Fajari</surname>
<given-names>M. K.</given-names>
</name>
<name>
<surname>Lindenbergh</surname>
<given-names>R. C.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Airborne laser scanning point cloud classification using the dgcnn deep learning method</article-title>. <source>Remote Sens.</source> <volume>13</volume>, <fpage>859</fpage>. <pub-id pub-id-type="doi">10.3390/rs13050859</pub-id>
</citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wilson</surname>
<given-names>J. C.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Mapping coral reef complexes using airborne lidar bathymetry technology</article-title>. <source>Oceans&#x2019;11 MTS/IEEE Kona</source> <volume>1&#x2013;5</volume>, <fpage>1</fpage>&#x2013;<lpage>5</lpage>. <pub-id pub-id-type="doi">10.23919/oceans.2011.6107165</pub-id>
</citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wan</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Yue</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Keutzer</surname>
<given-names>K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Squeezeseg: convolutional neural nets with recurrent crf for real-time road-object segmentation from 3d lidar point cloud</article-title>. <source>ICRA</source>, <fpage>1887</fpage>&#x2013;<lpage>1893</lpage>. <pub-id pub-id-type="doi">10.1109/icra.2018.8462926</pub-id>
</citation>
</ref>
<ref id="B61">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xiao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Guo</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>The influence of refractive index changes in water on airborne lidar bathymetric errors</article-title>. <source>J. Mar. Sci. Eng.</source> <volume>12</volume>, <fpage>435</fpage>. <pub-id pub-id-type="doi">10.3390/jmse12030435</pub-id>
</citation>
</ref>
<ref id="B62">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Hua</surname>
<given-names>B.-S.</given-names>
</name>
<name>
<surname>Yeung</surname>
<given-names>S.-K.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Shellnet: efficient point cloud convolutional neural networks using concentric shells statistics</article-title>. <source>ICCV</source>, <fpage>1607</fpage>&#x2013;<lpage>1616</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2019.00169</pub-id>
</citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lei</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Introducing improved transformer to land cover classification using multispectral lidar point clouds</article-title>. <source>Remote Sens.</source> <volume>14</volume>, <fpage>3808</fpage>. <pub-id pub-id-type="doi">10.3390/rs14153808</pub-id>
</citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Ljungholm</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Malmqvist</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bianco</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Hansson</surname>
<given-names>L.-A.</given-names>
</name>
<name>
<surname>Svanberg</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Inelastic hyperspectral lidar for profiling aquatic ecosystems</article-title>. <source>Laser and Photonics Rev.</source> <volume>10</volume>, <fpage>807</fpage>&#x2013;<lpage>813</lpage>. <pub-id pub-id-type="doi">10.1002/lpor.201600093</pub-id>
</citation>
</ref>
<ref id="B65">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2023</year>). <article-title>Adaptive model for the water depth bias correction of bathymetric lidar point cloud data</article-title>. <source>Int. J. Appl. Earth Observation Geoinformation</source> <volume>118</volume>, <fpage>103253</fpage>. <pub-id pub-id-type="doi">10.1016/j.jag.2023.103253</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>