<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2025.1524380</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artificial Intelligence</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>A mobile hybrid deep learning approach for classifying 3D-like representations of Amazonian lizards</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>da Silva</surname> <given-names>Arthur Gonsales</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2868901/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>de Oliveira</surname> <given-names>Roger Pinho</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2889901/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>de Oliveira Bastos</surname> <given-names>Caio</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2639444/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>de Carvalho</surname> <given-names>Elena Almeida</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Gomes</surname> <given-names>Bruno Duarte</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2639303/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Departamento de Ci&#x00EA;ncia de Dados, Instituto Tecnol&#x00F3;gico Vale</institution>, <addr-line>Bel&#x00E9;m</addr-line>, <country>Brazil</country></aff>
<aff id="aff2"><sup>2</sup><institution>Centro de Ci&#x00EA;ncias Biol&#x00F3;gicas e da Sa&#x00FA;de, Universidade da Amaz&#x00F4;nia</institution>, <addr-line>Bel&#x00E9;m</addr-line>, <country>Brazil</country></aff>
<aff id="aff3"><sup>3</sup><institution>Instituto de Ci&#x00EA;ncias Biol&#x00F3;gicas, Universidade Federal do Par&#x00E1;</institution>, <addr-line>Bel&#x00E9;m</addr-line>, <country>Brazil</country></aff>
<author-notes>
<fn fn-type="edited-by" id="fn0001">
<p>Edited by: Shanwen Sun, Northeast Forestry University, China</p>
</fn>
<fn fn-type="edited-by" id="fn0002">
<p>Reviewed by: Felipe Polivanov Ottoni, Federal University of Maranh&#x00E3;o, Brazil</p>
<p>Jung-Il Kim, Korea Institute of Ocean Science and Technology (KIOST), Republic of Korea</p>
<p>Hongmei Zhang, Northeast Forestry University, China</p>
</fn>
<corresp id="c001">&#x002A;Correspondence: Bruno Duarte Gomes, <email>brunodgomes@ufpa.br</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>12</day>
<month>08</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>8</volume>
<elocation-id>1524380</elocation-id>
<history>
<date date-type="received">
<day>06</day>
<month>01</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>07</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2025 da Silva, de Oliveira, de Oliveira Bastos, de Carvalho and Gomes.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>da Silva, de Oliveira, de Oliveira Bastos, de Carvalho and Gomes</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Image classification is a highly significant field in machine learning (ML), especially when applied to address longstanding and challenging issues in the biological sciences, such as specie recognition and biodiversity conservation. In this study, we present the development of a hybrid machine learning-based tool suitable for deployment on mobile devices. This tool is aimed at processing and classifying three-dimensional samples of endemic lizard species from the Amazon rainforest. The dataset used in our experiment was collected at the Museu Paraense Em&#x00ED;lio Goeldi (MPEG), Bel&#x00E9;m-PA, Brazil, and comprises three species: (a) <italic>Anolis fuscoauratus</italic>; (b) <italic>Hoplocercus spinosus</italic>; and (c) <italic>Polychrus marmoratus</italic>. We compared the effectiveness of four artificial neural networks (ANN) for feature extraction: (a) MobileNet; (b) MobileNetV2; (c) MobileNetV3-Small; and (d) MobileNetV3-Large. Additionally, we evaluated five classical ML models for classifying the extracted patterns: (a) Support Vector Machine (SVM); (b) GaussianNB (GNB); (c) AdaBoost (ADB); (d) K-Nearest Neighbors (KNN); and (e) Random Forest (RF). The performance metrics of all classifiers were very close, we used the McNemar&#x2019;s test on each model&#x2019;s confusion matrix to evaluate and compare their statistical significance. Our best model was a combination of a 2.9 million parameters MobileNetV3-Small as the feature extractor, with a linear kernel-based SVM as the classifier, which achieved accuracy of 0.955, precision of 0.948, recall of 0.948, and f1-score of 0.948. The results indicated that the use of a small deep learning (DL) model, in combination with a classical ML algorithm, emerges as a viable technique for classifying three-dimensional representations of lizard species samples. Such an approach facilitates taxonomic identification work for professionals in the field and provides a tool adaptable for integration into mobile data recording equipment, such as smartphones, and benefiting from more morphological features extracted from three-dimensional samples instead of two-dimensional images.</p>
</abstract>
<kwd-group>
<kwd>hybrid machine learning</kwd>
<kwd>3D representations</kwd>
<kwd>Amazonian lizards</kwd>
<kwd>MobileNet</kwd>
<kwd>species classification</kwd>
</kwd-group>
<counts>
<fig-count count="5"/>
<table-count count="4"/>
<equation-count count="5"/>
<ref-count count="46"/>
<page-count count="10"/>
<word-count count="7008"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Machine Learning and Artificial Intelligence</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>In the Squamata order, which comprises species with bodies covered by scales, among other characteristics, the classification of lizards is based on multiple morphological features (<xref ref-type="bibr" rid="ref30">Pyron et al., 2013</xref>). According to (<xref ref-type="bibr" rid="ref37">Stewart and Daniel, 1075</xref>), these morphological characteristics are referred to as microornamentations and are most prominent in the dorsal scales of the head, trunk, and tails of each individual. Modern biodiversity data collection equipment, such as sound recorders, camera traps, and other imaging methods, allow the measurement of many parameters, making it possible to extract vast amounts of information in a relatively inexpensive manner. This technology has become increasingly popular among scientists and helps to answer questions such as: (a) Which species occur in a given area?; (b) What are their activities/behavior?; and (c) How many individuals inhabit the region? (<xref ref-type="bibr" rid="ref12">Gomez Villa et al., 2017</xref>). The success in inventorying and monitoring forest lizard species relies on robust monitoring, recognition, and sampling, and currently represents one of the most complex tasks in the field of herpetological conservation (<xref ref-type="bibr" rid="ref2">Bell, 2009</xref>).</p>
<p>One of the most used data types in problems involving biodiversity conservation with specialized image models is camera trap images (<xref ref-type="bibr" rid="ref21">Miao et al., 2019</xref>). The aim of remote monitoring can range from species identification to inferring the abundance and distribution of important conservation animals, but these motivations typically share a common goal: to classify target species (<xref ref-type="bibr" rid="ref7">Chen et al., 2019</xref>). This interest in remote monitoring is accompanied by several challenges in large-scale identification (<xref ref-type="bibr" rid="ref7">Chen et al., 2019</xref>).</p>
<p>The most recent research in automated identification of animal species can be divided into two distinct types: laboratory-based investigation (LBI), and field-based investigation (FBI) (<xref ref-type="bibr" rid="ref19">Martineau et al., 2017</xref>). For LBI, a pre-established image acquisition protocol must be followed to standardize the sampling and use of specimens, which are typically handled by a specialized biologist. This contrasts significantly with FBI, where a mobile device or camera is usually employed for the image acquisition process of the individuals (<xref ref-type="bibr" rid="ref19">Martineau et al., 2017</xref>).</p>
<p>In studies of insect classification, for instance, LBI is the most commonly used method due to the highly manual handling of specimens (<xref ref-type="bibr" rid="ref19">Martineau et al., 2017</xref>). On the other hand, the identification of mammals and fish is typically accomplished using field-recorded images, while automated recognition of plant species can benefit from both the controlled environment of a laboratory and field conditions (<xref ref-type="bibr" rid="ref44">Weinstein, 2018</xref>). These studies focus on the use of Machine Learning (ML) with Convolutional Neural Networks (CNN), which are models specialized in image processing that extract high-level abstractions from data and are considered the state-of-the-art for tasks involving image classification (<xref ref-type="bibr" rid="ref42">W&#x00E4;ldchen and M&#x00E4;der, 2018</xref>).</p>
<p>The most common type of algorithm learning used for image classification is supervised learning, where input data (samples) are fed into the model along with their corresponding labels (class names), and the algorithms are trained to map the input information to the output label, such as the name of a species, for example (<xref ref-type="bibr" rid="ref26">Norouzzadeh et al., 2021</xref>).</p>
<p>Before the emergence of computer vision (CV) models and artificial intelligence (AI) algorithms in general, the process of identifying and conserving animal species was and still is, in some places, carried out manually with a high dependence on human activities, which imposes several limitations on the task (<xref ref-type="bibr" rid="ref39">Tuia et al., 2022</xref>). These limitations, mainly physical and cognitive, hinder the understanding of species distribution and diversity. For instance, the counting of colonies of seabirds and cave-dwelling bats conducted by humans tends to significantly underestimate the actual number of individuals (<xref ref-type="bibr" rid="ref39">Tuia et al., 2022</xref>). This scenario of limitations and uncertainties changed with the advent of large-scale AI-driven automation of these tasks.</p>
<p>With recent advances in automated image classification and information gathering, new approaches have become possible (<xref ref-type="bibr" rid="ref28">Pinho et al., 2023</xref>). Several existing examples demonstrate the applications of automatic classification based on deep learning (DL) using taxonomic data from different species (<xref ref-type="bibr" rid="ref42">W&#x00E4;ldchen and M&#x00E4;der, 2018</xref>). <xref ref-type="table" rid="tab1">Table 1</xref> summarizes recent studies where CV algorithms were employed to perform automated species identification on a diverse range of other taxonomic datasets (<xref ref-type="bibr" rid="ref44">Weinstein, 2018</xref>; <xref ref-type="bibr" rid="ref39">Tuia et al., 2022</xref>; <xref ref-type="bibr" rid="ref4">Bolon et al., 2022</xref>; <xref ref-type="bibr" rid="ref9">Durso et al., 2021</xref>; <xref ref-type="bibr" rid="ref3">Binta Islam et al., 2023</xref>).</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Recent studies employing computer vision algorithms for species classification across various taxonomic groups.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Species</th>
<th align="center" valign="top">Samples</th>
<th align="left" valign="top">Architecture</th>
<th align="center" valign="top">Accuracy</th>
<th align="left" valign="top">Study</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Reptiles</td>
<td align="center" valign="top">386,006</td>
<td align="left" valign="top">Vision Transformer (ViT)</td>
<td align="center" valign="top">0.962</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref4">Bolon et al. (2022)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Reptiles</td>
<td align="center" valign="top">82,601</td>
<td align="left" valign="top">EfficientNet</td>
<td align="center" valign="top">0.870</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref9">Durso et al. (2021)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Lizards &#x0026; Amphibians</td>
<td align="center" valign="top">6,045</td>
<td align="left" valign="top">MobileNetV2</td>
<td align="center" valign="top">0.820</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref11">Gill et al. (2024)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Lizards &#x0026; Amphibians</td>
<td align="center" valign="top">2,700</td>
<td align="left" valign="top">VGG16</td>
<td align="center" valign="top">0.870</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref3">Binta Islam et al. (2023)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Fishes</td>
<td align="center" valign="top">1,080</td>
<td align="left" valign="top">Image Processing + SVM</td>
<td align="center" valign="top">0.942</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref35">Sharmin et al. (2019)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Fishes</td>
<td align="center" valign="top">3,068</td>
<td align="left" valign="top">U-NET&#x202F;+&#x202F;CNN</td>
<td align="center" valign="top">0.979</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref32">Robillard et al. (2023)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Lizards &#x0026; Amphibians</td>
<td align="center" valign="top">828</td>
<td align="left" valign="top">CNN</td>
<td align="center" valign="top">0.600</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref16">Islam and Valles (2020)</xref>
</td>
</tr>
<tr>
<td align="left" valign="top">Mammals</td>
<td align="center" valign="top">326</td>
<td align="left" valign="top">Mask R-CNN&#x202F;+&#x202F;ResNet101</td>
<td align="center" valign="top">0.980</td>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref9001">Gray et al. (2019)</xref>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>A total of seven comparable studies where pre-trained state-of-the-art deep learning models were employed on taxonomic datasets of different groups were analyzed in order to provide more robustness to our research. The sample size and accuracy score are reported below for each study.</p>
</table-wrap-foot>
</table-wrap>
<p>As can be seen in <xref ref-type="table" rid="tab1">Table 1</xref>, most studies used pre-trained models. This is the case because when pre-trained networks are employed either as feature extractors or efficiently optimized for the new dataset, there exists a strong correlation between the high accuracy achieved by the model on its original pre-training phases with its score in the new training demand (<xref ref-type="bibr" rid="ref18">Kornblith et al., 2019</xref>). Thus, incremental or transfer learning only requires the pre-trained model to generalize an additional predictive pattern that might be present in the dataset while retaining its previous optimal weights, often gathered on ImageNet Large-Scale Visual Recognition Competition (ILSVRC) (<xref ref-type="bibr" rid="ref9">Durso et al., 2021</xref>).</p>
<p>Despite the widespread use of CNNs in taxonomic databases (<xref ref-type="bibr" rid="ref44">Weinstein, 2018</xref>; <xref ref-type="bibr" rid="ref39">Tuia et al., 2022</xref>; <xref ref-type="bibr" rid="ref4">Bolon et al., 2022</xref>; <xref ref-type="bibr" rid="ref9">Durso et al., 2021</xref>; <xref ref-type="bibr" rid="ref3">Binta Islam et al., 2023</xref>), our literature review revealed no applications of these models to three-dimensional representations of Amazonian lizards. In this study, we have developed an open-source system for the automatic classification of three-dimensional samples of Amazonian lizard species, adaptable for deployment on mobile equipment such as smartphones. We employed state-of-the-art DL and ML techniques for image processing and classification using the family of CNNs known as MobileNets (<xref ref-type="bibr" rid="ref15">Howard et al., 2017</xref>; <xref ref-type="bibr" rid="ref34">Sandler et al., 2018</xref>; <xref ref-type="bibr" rid="ref14">Howard et al., 2019</xref>), together with classical ML models, which demonstrated exceptional efficiency in similar tasks. Making use of 3D representations of specimens as samples turned our approach unique, and significantly benefited our models with relevant morphological information about each species when compared to typical 2D representations as employed by <xref ref-type="bibr" rid="ref16">Islam and Valles (2020)</xref>, which used the same family of pre-trained CNNs we used in this study, but achieved less accuracy.</p>
</sec>
<sec sec-type="materials|methods" id="sec2">
<label>2</label>
<title>Materials and methods</title>
<sec id="sec3">
<label>2.1</label>
<title>Data collection and sample processing</title>
<p>Data was collected at MPEG, located in Bel&#x00E9;m, Para, Brazil. MPEG is the second-oldest scientific research institution in Brazil, founded in 1866, and it houses a local herpetological collection with approximately 100,000 specimens of amphibians and reptiles (<xref ref-type="bibr" rid="ref8">Da Costa Prudente et al., 2019</xref>). Three species were selected for collection, namely: (a) <italic>Anolis fuscoauratus</italic>; (b) <italic>Hoplocercus spinosus</italic>; and (c) <italic>Polychrus marmoratus</italic>; all species found in the Amazon region (<xref ref-type="bibr" rid="ref41">Vitt et al., 2003</xref>; <xref ref-type="bibr" rid="ref38">Torres-Carvajal and De Queiroz, 2009</xref>; <xref ref-type="bibr" rid="ref23">Murphy et al., 2017</xref>). <xref ref-type="fig" rid="fig1">Figure 1</xref> below shows pictures of individuals from each species.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>The three species selected for this study. <bold>(A)</bold> <italic>Anolis fuscoauratus</italic>, <bold>(B)</bold> <italic>Hoplocercus spinosus</italic>, <bold>(C)</bold> <italic>Polychurs marmoratus</italic>. All the specimens were preserved in alcohol, and only individuals with good preservation conditions were selected.</p>
</caption>
<graphic xlink:href="frai-08-1524380-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Panel A shows a small lizard with partially shed skin on its tail, resting on a blue object. Panel B depicts a rough-scaled lizard with shedding skin on dark fabric. Panel C portrays a lizard with a long tail lying on black fabric, also shedding skin.</alt-text>
</graphic>
</fig>
<p>All specimens were preserved in alcohol, and the preservation conditions of each sample were a determining factor in selecting both the individuals and species chosen for this study. The selected individuals were then placed on a black cloth, and positioned on the collection bench to mitigate any visual noise that could interfere with identification. This simple strategy can be easily replicated in any environment, as in field data collection routines.</p>
<p>In recent studies using three-dimensional samples for species classification, the use of Light Detection and Ranging (LiDAR), and Spectral Imaging (SI) are extensive, particularly in studies using plants as specimens (<xref ref-type="bibr" rid="ref20">M&#x00E4;yr&#x00E4; et al., 2021</xref>; <xref ref-type="bibr" rid="ref25">Nezami et al., 2020</xref>; <xref ref-type="bibr" rid="ref29">Polonen et al., 2018</xref>). However, these technologies are costly and require highly specialized expertise, making them impractical for everyday use by experts in both laboratory and field settings. Furthermore, using impractical solutions such as LiDAR and SI makes it almost impossible to safely and easily reproduce the results, especially in areas where research funding is unstable.</p>
<p>As a solution, we adopted smartphone-based image capture from the dorsal, lateral, and ventral points of view to compose our samples. The use of smartphones offers a cost-effective alternative, enabling broader accessibility and usability for species classification. As can be seen in <xref ref-type="fig" rid="fig2">Figure 2</xref>, three photos of each individual were taken, where each will represent one channel of a final RGB-like three-dimensional sample.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>A sample of <italic>Anolis fuscoauratus</italic>, composed of three perspectives. <bold>(A)</bold> dorsal, <bold>(B)</bold> lateral, <bold>(C)</bold> ventral views. The images are converted to grayscale and then arranged into a matrix of dimensions 1 x 224 x 224 x 3, with each image occupying one color channel.</p>
</caption>
<graphic xlink:href="frai-08-1524380-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Three images labeled A, B, and C show a brown lizard on a dark surface. Image A displays a top view, image B shows a side view, and image C presents another top view with the tail bent in a loop. Each image highlights different angles and postures of the lizard.</alt-text>
</graphic>
</fig>
<p>It was necessary to remove some images due to poor quality; a total of 80 three-dimensional samples, totaling 240 unique images, remained. Among these, there were 49 samples of <italic>Anolis fuscoauratus</italic>, 22 samples of <italic>Hoplocercus spinosus</italic>, and 9 samples of <italic>Polychrus marmoratus</italic>.</p>
<p>Subsequently, all samples were resized to dimensions of 224 &#x00D7; 224 pixels and standardized to conform to the input layer requirements of our Convolutional Neural Network (CNN), which are standard for the MobileNet family of models. The dataset was then partitioned into training/validation and test sets, adhering to an 80&#x2013;20% split, respectively. This approach was chosen over the inclusion of an additional hold-out validation set, with a preference for employing cross-validation. The <xref ref-type="fig" rid="fig2">Figure 2</xref> below shows one sample composed of different perspectives.</p>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Data augmentation for addressing class imbalance</title>
<p>We used TensorFlow&#x2019;s (TF) image data generator module (<xref ref-type="bibr" rid="ref1">Abadi et al., 2016</xref>) for data augmentation, where random modifications such as Flip, Crop, and Translate, were applied to the samples without altering their fundamental characteristics, thus generating new synthetic observations in our dataset (<xref ref-type="bibr" rid="ref45">Xu et al., 2023</xref>). The outcome of data augmentation resulted in an increase from 80 initial three-dimensional samples to 3,900 in the training set, balanced between species. This increases the robustness of our model on handling different imaging conditions in different collection environments. The <xref ref-type="fig" rid="fig3">Figure 3</xref> illustrates the data augmentation process.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>The data augmentation process illustrated. The original image set was split into train and test sets, and then the augmented images were generated for the training set. No images from the test set were used to augment data in the training set.</p>
</caption>
<graphic xlink:href="frai-08-1524380-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Diagram illustrating a machine learning workflow. A "Training Set" comprised of lizard images is augmented and used for "Training with Cross Validation". A "Test Set", also of lizard images, is used for "Model Evaluation". Arrows indicate the workflow progression.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>Models selection and definition</title>
<p>We selected the class of MobileNet models for developing our species identification system. This class consists of highly efficient algorithms for mobile CV applications and embedded systems (<xref ref-type="bibr" rid="ref15">Howard et al., 2017</xref>). There are three main MobileNet models: (a) MobileNet; (b) MobileNetv2; and (c) MobileNetV3, with the latter having two variants, namely: Large and Small (<xref ref-type="bibr" rid="ref15">Howard et al., 2017</xref>; <xref ref-type="bibr" rid="ref34">Sandler et al., 2018</xref>; <xref ref-type="bibr" rid="ref14">Howard et al., 2019</xref>).</p>
<p>The first model (MobileNet) is based on depth wise separable convolutions, which are a form of factorized convolutions that transform a regular convolution operation into depth wise, which significantly reduces both computational cost and model size, having 4.3&#x202F;M adjustable parameters, with a lower memory footprint in comparison to other major CNNs (<xref ref-type="bibr" rid="ref15">Howard et al., 2017</xref>). The second model (MobileNetV2) introduces the new <italic>inverted residual with a linear bottleneck</italic> module (<xref ref-type="bibr" rid="ref34">Sandler et al., 2018</xref>), which expands to a higher dimension a compressed low-dimensional representation of the input data and then filters it using a lightweight depthwise convolution, having a slightly higher memory requirement than MobileNet, with a more robust architecture comprised of 3.5&#x202F;M adjustable parameters. The third model (MobileNetV3) features an efficient redesign of the network architecture, coupled with a segmentation decoder that optimizes resource consumption for both of its variants, the Large, for devices with greater availability of resources, having 5.4&#x202F;M adjustable parameters, and the Small, for scenarios with more limited processing power, having a total of 2.5&#x202F;M adjustable parameters (<xref ref-type="bibr" rid="ref14">Howard et al., 2019</xref>).</p>
<p>We used and compared the performance of all available MobileNet network variants as feature extractors only. We did not retrain the models, and we appended a Global Average Pooling 2D layer at the end of each model for dimensionality reduction, and then we replaced their classification layers with classical ML algorithms.</p>
<p>The selection of classical ML algorithms was based on the criteria that it has to be commonly applied in research with biological databases (<xref ref-type="bibr" rid="ref17">Jovel and Greiner, 2021</xref>), and pre-implemented in Scikit-learn (SKL) (<xref ref-type="bibr" rid="ref27">Pedregosa et al., 2008</xref>). The chosen models were: (a) Support Vector Machine (SVM) with linear, rbf, poly kernels; (b) K-Nearest Neighbors (KNN); (c) Random Forest (RF); (d) GaussianNB (GNB); and (e) AdaBoost (ADB).</p>
<p>We adopted this hybrid approach because there is enough evidence showing that using pre-trained models, such as MobileNets as feature extractors, can transfer their high accuracies acquired on ILSVRC to the new models they compose, without the need for computationally expensive retraining (<xref ref-type="bibr" rid="ref18">Kornblith et al., 2019</xref>; <xref ref-type="bibr" rid="ref36">Sowmya et al., 2023</xref>; <xref ref-type="bibr" rid="ref22">Michele et al., 2019</xref>). Moreover, the composition of a hybrid model with a classical algorithm serving as the final classifier drastically reduces the likelihood of the model presents overfitting (<xref ref-type="bibr" rid="ref22">Michele et al., 2019</xref>).</p>
</sec>
<sec id="sec6">
<label>2.4</label>
<title>Feature extraction and dimensionality reduction</title>
<p>From the original data, we generated four new datasets of features, each one extracted with a different variant of MobileNet (V1, V2, V3-Large, and V3-Small), we call these full-features datasets. To assess the complexity and the effectiveness of feature separation across our classes, we applied the t-distributed Stochastic Neighbor Embedding (t-SNE) to each of the full-feature datasets. t-SNE is a method that compresses high-dimensional data into a two-or three-dimensional map (<xref ref-type="bibr" rid="ref40">van der Maaten and Hinton, 2008</xref>), effectively transforming high cardinality information into a lower-dimensional compressed space.</p>
<p>Lastly, we used the RF algorithm to ascertain the relative importance of features within each full-features dataset (<xref ref-type="bibr" rid="ref13">Haq et al., 2019</xref>). Subsequently, a significance threshold of 0.01 was applied to retain only those features ranking highest in importance. This process yielded a subset of 20 columns constituting the top-ranked features for each respective full-features dataset.</p>
</sec>
<sec id="sec7">
<label>2.5</label>
<title>Model training and evaluation</title>
<p>For comparison, we trained our ML models on each full-features dataset, and also on each 20 top-ranked features dataset. All datasets were normalized with <italic>MinMaxScaler</italic> (<xref ref-type="bibr" rid="ref31">Raju et al., 2020</xref>). The training was cross-validated, with the <italic>k-fold</italic> and <italic>random state</italic> parameters set to 4, and 42, respectively. For models&#x2019; performance evaluation, we used a total of five different metrics, namely: (a) accuracy; (b) precision; (c) recall; (d) f1-score; and (e) confusion matrix.</p>
<sec id="sec8">
<label>2.5.1</label>
<title>Accuracy</title>
<p>Accuracy denotes the ratio of true positives (TP) and true negatives (TN), against the overall predictions, also comprised of false positives (FP) and false negatives (FN) (<xref ref-type="bibr" rid="ref24">Naser and Alavi, 2023</xref>). It is calculated as follows:<disp-formula id="E1">
<mml:math id="M1">
<mml:mtext mathvariant="italic">Accuracy</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">TN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula></p>
</sec>
<sec id="sec9">
<label>2.5.2</label>
<title>Precision</title>
<p>Precision denotes the ratio of correctly predicted true instances over the total number of positively predicted instances (<xref ref-type="bibr" rid="ref24">Naser and Alavi, 2023</xref>). It is calculated as follows:<disp-formula id="E2">
<mml:math id="M2">
<mml:mtext mathvariant="italic">Precision</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula></p>
</sec>
<sec id="sec10">
<label>2.5.3</label>
<title>Recall</title>
<p>Recall denotes the ratio of correctly predicted true instances over the total number of positive instances (<xref ref-type="bibr" rid="ref24">Naser and Alavi, 2023</xref>). It is calculated as follows:<disp-formula id="E3">
<mml:math id="M3">
<mml:mtext mathvariant="italic">Recall</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula></p>
</sec>
<sec id="sec11">
<label>2.5.4</label>
<title>F1-score</title>
<p>F1-Score is the harmonic mean of precision and recall (<xref ref-type="bibr" rid="ref24">Naser and Alavi, 2023</xref>). It is calculated as follows:<disp-formula id="E4">
<mml:math id="M4">
<mml:mi>F</mml:mi>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext mathvariant="italic">Score</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn>
<mml:mi>x</mml:mi>
<mml:mfrac>
<mml:mtext mathvariant="italic">PrecisionxRecall</mml:mtext>
<mml:mrow>
<mml:mtext mathvariant="italic">Precision</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="italic">Recall</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula></p>
</sec>
<sec id="sec12">
<label>2.5.5</label>
<title>Confusion matrix</title>
<p>A confusion matrix presents a summary of correctly and misclassified samples of a classification problem (<xref ref-type="bibr" rid="ref24">Naser and Alavi, 2023</xref>). The entries of a confusion matrix are all the positive and negative predictions described so far.</p>
</sec>
</sec>
<sec id="sec13">
<label>2.6</label>
<title>Bayesian optimization evaluation</title>
<p>We made an additional evaluation using Bayesian Optimization (BO) in an attempt to further improve the best ML model&#x2019;s hyperparameters. By using BO, a surrogate for the model&#x2019;s objective function is created, and a Gaussian Regressor quantifies the uncertainty for the surrogate (<xref ref-type="bibr" rid="ref10">Frazier, 2018</xref>). The formula below shows the acquisition function Expected Improvement (EI), adopted in this study.<disp-formula id="E5">
<mml:math id="M5">
<mml:msub>
<mml:mi mathvariant="italic">EI</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2254;</mml:mo>
<mml:msub>
<mml:mi>E</mml:mi>
<mml:mi>n</mml:mi>
</mml:msub>
</mml:math>
</disp-formula></p>
<p>The EI tells us how much we expect to improve our best result if we try a new set of optimizable parameters <italic>x</italic>, and it is popular due to its multi-modal nature and effective balance between exploration and exploitation of the search space for the best set of hyperparameters that will produce the lowest error on the model (<xref ref-type="bibr" rid="ref43">Wang et al., 2017</xref>). The metrics resulting from this attempt were compared to the model trained without the help of BO.</p>
</sec>
<sec id="sec14">
<label>2.7</label>
<title>Statistical analysis</title>
<p>The McNemar&#x2019;s test is a statistical test particularly suitable for comparing the performances of two classification models on the same dataset, assuming a null hypothesis (H0) of no statistical difference between the two proportions being compared. The test was used to evaluate both model performance and the effectiveness of BO throughout this study.</p>
<p>First, we assessed potential performance differences between the best-performing models trained with the full-feature dataset, both with and without BO. This process was then repeated for the best models trained on the reduced dataset (top 20 features), again comparing models with and without BO.</p>
<p>Finally, McNemar&#x2019;s test was used to determine if there were any significant differences between the best models trained with the full-feature and reduced datasets, regardless of the use of BO during training.</p>
</sec>
<sec id="sec15">
<label>2.8</label>
<title>Classification pipeline technologies</title>
<p>Our open-source pipeline was developed using Python (<xref ref-type="bibr" rid="ref33">Rossum, 1995</xref>), the TF DL framework (<xref ref-type="bibr" rid="ref1">Abadi et al., 2016</xref>), and the SKL ML framework (<xref ref-type="bibr" rid="ref27">Pedregosa et al., 2008</xref>). Images were captured using an HTC One M8 smartphone (4MP &#x00D7; 2688 &#x00D7; 1520 440 ppi camera). The classification pipeline comprises five main stages:<list list-type="bullet">
<list-item>
<p>Capture a dorsal photo of the specimen.</p>
</list-item>
<list-item>
<p>Capture a lateral photo of the specimen.</p>
</list-item>
<list-item>
<p>Capture a ventral photo of the specimen.</p>
</list-item>
<list-item>
<p>Compose a three-dimensional sample from the acquired images.</p>
</list-item>
<list-item>
<p>Classify the lizard species with our trained model.</p>
</list-item>
</list></p>
</sec>
</sec>
<sec sec-type="results" id="sec16">
<label>3</label>
<title>Results</title>
<sec id="sec17">
<label>3.1</label>
<title>Datasets complexity analysis</title>
<p>The complexity of each dataset significantly influenced the performance of classical ML algorithms. <xref ref-type="fig" rid="fig4">Figure 4</xref> illustrates the differences in clustering for each dataset, as revealed by t-SNE (<xref ref-type="bibr" rid="ref40">van der Maaten and Hinton, 2008</xref>).</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>The t-SNE plot for dataset complexity analysis. <bold>(A)</bold> MobileNet; <bold>(B)</bold> MobileNet V2; <bold>(C)</bold> MobileNet V3-Large; and <bold>(D)</bold> MobileNet V3-Small. The features extracted by MobileNet V3-Small demonstrate the most homogeneous and well-separated clusters of data points for all species. The X and Y axes represent compressed dimensions.</p>
</caption>
<graphic xlink:href="frai-08-1524380-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Scatter plots labeled A, B, C, and D show data points distributed across compressed dimensions. Each plot displays three clusters: blue, orange, and green. Variations in clustering patterns are visible across the plots. Axes are labeled "Compressed Dimension 1" and "Compressed Dimension 2."</alt-text>
</graphic>
</fig>
<p>Analyses (a) and (c) show good separation between clusters, but the samples within each cluster are more dispersed. In contrast, analysis (b) reveals greater class overlap, although clusters are relatively well-concentrated. Analysis (d), based on MobileNetV3-Small-extracted data, demonstrates the optimal balance between cluster separation and sample concentration, with minimal class overlap.</p>
</sec>
<sec id="sec18">
<label>3.2</label>
<title>Model performance analysis</title>
<p>The trained models demonstrated similar performance across all datasets, indicating that both full-feature and reduced datasets successfully captured essential morphological and structural patterns, such as microornamentations (<xref ref-type="bibr" rid="ref37">Stewart and Daniel, 1075</xref>). This facilitated model generalization despite variations in the number of extracted features. <xref ref-type="table" rid="tab2">Table 2</xref> presents the top-performing models trained with cross-validation using all features extracted by MobileNet variants.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Cross-validated average performance metrics of classic ML models on each full-features dataset.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Feature Extractor</th>
<th align="left" valign="top">Best model</th>
<th align="center" valign="top">Accuracy</th>
<th align="center" valign="top">Precision</th>
<th align="center" valign="top">Recall</th>
<th align="center" valign="top">F1-score</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">MobileNetV3-Small</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.974</td>
<td align="center" valign="top">0.985</td>
<td align="center" valign="top">0.965</td>
<td align="center" valign="top">0.973</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV1</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.970</td>
<td align="center" valign="top">0.981</td>
<td align="center" valign="top">0.949</td>
<td align="center" valign="top">0.961</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV2</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.951</td>
<td align="center" valign="top">0.964</td>
<td align="center" valign="top">0.924</td>
<td align="center" valign="top">0.937</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV3-Large</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.953</td>
<td align="center" valign="top">0.969</td>
<td align="center" valign="top">0.928</td>
<td align="center" valign="top">0.942</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Four models acting as feature extractors were tested together with five different classical machine learning algorithms acting as classifiers. The accuracy, precision, recall, and f1-score for all hybrid models, on the full-features dataset, are reported below.</p>
</table-wrap-foot>
</table-wrap>
<p>The MobileNet V3-Small + Linear SVM classifier consistently outperformed other models on full-feature datasets. This superior performance might be attributed to the relatively lower complexity and clearer class separation of the dataset generated with this MobileNet variant, as evidenced by <xref ref-type="fig" rid="fig4">Figure 4</xref>. Other datasets exhibited greater class overlap and less cluster concentration.</p>
<p>While the dataset with only the 20 top-ranked features exhibited reduced homogeneity and class separation, it remained representative of the underlying data. Notably, the MobileNet V3-Small + Linear SVM classifier again demonstrated comparable performance, leading the results on this reduced dataset as well as shown in <xref ref-type="table" rid="tab3">Table 3</xref>.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Cross-validated average performance metrics of classic ML models on each 20 top-ranked features dataset.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Feature Extractor</th>
<th align="left" valign="top">Best model</th>
<th align="center" valign="top">Accuracy</th>
<th align="center" valign="top">Precision</th>
<th align="center" valign="top">Recall</th>
<th align="center" valign="top">F1-score</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">MobileNetV3-Small</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.968</td>
<td align="center" valign="top">0.970</td>
<td align="center" valign="top">0.962</td>
<td align="center" valign="top">0.965</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV1</td>
<td align="left" valign="top">RBF SVM</td>
<td align="center" valign="top">0.958</td>
<td align="center" valign="top">0.955</td>
<td align="center" valign="top">0.936</td>
<td align="center" valign="top">0.944</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV3-Large</td>
<td align="left" valign="top">RFC</td>
<td align="center" valign="top">0.935</td>
<td align="center" valign="top">0.934</td>
<td align="center" valign="top">0.923</td>
<td align="center" valign="top">0.925</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV2</td>
<td align="left" valign="top">Linear SVM</td>
<td align="center" valign="top">0.898</td>
<td align="center" valign="top">0.905</td>
<td align="center" valign="top">0.859</td>
<td align="center" valign="top">0.874</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Four models acting as feature extractors were tested together with five different classical machine learning algorithms acting as classifiers. The accuracy, precision, recall, and f1-score for all hybrid models, on the 20 top-ranked features dataset, are reported below.</p>
</table-wrap-foot>
</table-wrap>
<p>The reduced dataset saw more complex classical ML algorithms among the top performers compared to the full-feature dataset (<xref ref-type="table" rid="tab2">Table 2</xref>). This suggests a need for increased model complexity to compensate for the information loss resulting from feature selection.</p>
</sec>
<sec id="sec19">
<label>3.3</label>
<title>Bayesian optimization effectiveness analysis and model skill evaluation</title>
<p>McNemar&#x2019;s test was used to assess statistical differences between hybrid models trained with and without Bayesian Optimization (BO), using both full-feature and reduced datasets.</p>
<p>For the full-feature dataset, the model trained without BO significantly outperformed the BO-trained model (&#x03C7;<sup>2</sup>&#x202F;=&#x202F;0.0, <italic>p</italic>&#x202F;=&#x202F;3.05e-5), achieving an accuracy of 0.991, precision of 0.987, recall of 0.992, and an F1-score of 0.990 on the test set.</p>
<p>In contrast, for the reduced dataset, the BO-trained model showed superior performance (&#x03C7;<sup>2</sup>&#x202F;=&#x202F;14.0, <italic>p</italic>&#x202F;=&#x202F;8.58e-11), with an accuracy of 0.955, precision of 0.948, recall of 0.948, and an F1-score of 0.948.</p>
<p>Finally, McNemar&#x2019;s test revealed no significant difference between the best-performing full-feature model and the best-performing reduced-feature model (&#x03C7;<sup>2</sup>&#x202F;=&#x202F;7.0, <italic>p</italic>&#x202F;=&#x202F;0.80361). Thus, the less complex model trained with the reduced dataset can be safely used. <xref ref-type="fig" rid="fig5">Figure 5</xref> shows the normalized confusion matrix for the MobileNetV3-Small + Linear SVM model on the test set.</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Confusion matrix of the MobileNetV3-Small + Linear SVM model trained with the reduced dataset. This confusion matrix corresponds to the best-performing MobileNetV3-Small + Linear SVM model trained on the dataset with the 20 top-ranked features. Despite the highest number of misclassified samples being from the Polychrus species among all classes, the overall performance, on a per-sample basis, was proven to be highly efficient.</p>
</caption>
<graphic xlink:href="frai-08-1524380-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Confusion matrix illustrating the model's classification performance for three labels: Anolis, Polyrus, and Hoplocercus. True label Anolis is predicted correctly 96% of the time, Polycrus 97%, and Hoplocercus 92%.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec20">
<label>3.4</label>
<title>Classification pipeline trainable parameters</title>
<p>The <xref ref-type="table" rid="tab4">Table 4</xref> summarizes the trainable parameters of the final classification pipeline, which includes a Min-Max scaler and a linear kernel SVM.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Trainable parameters of the assembled final classification pipeline.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Parameter</th>
<th align="left" valign="top">Value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">minmax_rescaler__clip</td>
<td align="left" valign="top">False</td>
</tr>
<tr>
<td align="left" valign="top">minmax_rescaler__copy</td>
<td align="left" valign="top">True</td>
</tr>
<tr>
<td align="left" valign="top">minmax_rescaler__feature_range</td>
<td align="left" valign="top">(0, 1)</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__C</td>
<td align="left" valign="top">0.10284379327993369</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__class_weight</td>
<td align="left" valign="top">None</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__dual</td>
<td align="left" valign="top">False</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__fit_intercept</td>
<td align="left" valign="top">True</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__intercept_scaling</td>
<td align="left" valign="top">1</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__loss</td>
<td align="left" valign="top">squared_hinge</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__max_iter</td>
<td align="left" valign="top">1,000</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__multi_class</td>
<td align="left" valign="top">&#x2018;ovr&#x2019;</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__penalty</td>
<td align="left" valign="top">&#x2018;l2&#x2019;</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__random_state</td>
<td align="left" valign="top">42</td>
</tr>
<tr>
<td align="left" valign="top">linear_svm_classifier__tol</td>
<td align="left" valign="top">0.0001</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>All the trainable model parameters for the classification algorithms are listed below to complete reproducibility of this study.</p>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec sec-type="discussion" id="sec21">
<label>4</label>
<title>Discussion</title>
<p>This research sought to evaluate the efficacy of classifying three-dimensional representations of Amazonian lizard species using cutting-edge deep learning algorithms. The aim was to create a mobile-ready classification pipeline that could be integrated into biodiversity monitoring equipment. The use of image triplets, each containing dorsal, lateral, and ventral views of the specimens, is a distinctive approach compared to all the most recent and comparable studies. Our findings demonstrate that this approach is not only feasible but also an efficient means of automated classification. Furthermore, our unique dataset, collected at MPEG, one of Brazil&#x2019;s oldest and most renowned research institutions, sets this study apart from a substantial portion of recent research.</p>
<sec id="sec22">
<label>4.1</label>
<title>Comparative analysis with existing research</title>
<p>Although several studies have applied deep learning to reptile images (<xref ref-type="bibr" rid="ref44">Weinstein, 2018</xref>; <xref ref-type="bibr" rid="ref39">Tuia et al., 2022</xref>; <xref ref-type="bibr" rid="ref4">Bolon et al., 2022</xref>; <xref ref-type="bibr" rid="ref9">Durso et al., 2021</xref>; <xref ref-type="bibr" rid="ref3">Binta Islam et al., 2023</xref>), most focus on a broader scope of reptiles and amphibians, not specifically lizards (<xref ref-type="bibr" rid="ref3">Binta Islam et al., 2023</xref>; <xref ref-type="bibr" rid="ref35">Sharmin et al., 2019</xref>; <xref ref-type="bibr" rid="ref11">Gill et al., 2024</xref>). A notable exception is (<xref ref-type="bibr" rid="ref11">Gill et al., 2024</xref>), which used MobileNetV2 to classify an open-access dataset of reptiles and amphibians, including lizards as one class. Unlike our approach, they treated each image independently, without aggregating triplets. Despite a larger dataset for fine-tuning, their accuracy of 0.820 was significantly lower than our best model. This discrepancy might be due to their higher number of classes, potentially increasing the model&#x2019;s learning difficulty. However, our dataset arguably presents higher complexity due to variations in dorsal, lateral, and ventral points-of-view, which may have forced our models to learn more detailed morphological patterns. Thus, our use of 3D representations and image triplets might be advantageous for capturing such details, ultimately leading to improved classification performance.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="sec23">
<label>5</label>
<title>Conclusion</title>
<p>Our study elucidates the potential for the classification of three-dimensional representations of lizard species through the utilization of mobile-ready deep learning models in the context of biodiversity monitoring. The deployment of three-dimensional representations of the specimens, generated from image triplets comprising dorsal, lateral, and ventral perspectives of the animals, has proven efficacious in capturing intricate morphological patterns. This approach facilitates robust feature extraction, distinct class separation, and enhanced classification accuracy. The capacity of the model to be readily deployed on mobile devices further augments its potential for field applications in biodiversity research and conservation endeavors.</p>
<p>Future research initiatives should focus on augmenting the number of supported species, as well as assessing the impact of this increment in our model of choice, potentially exploring new models and architectures, thereby contributing to the burgeoning field of deep learning-based lizard species classification. Currently, there is a paucity of extensive published studies in this domain for direct comparison. Additionally, efforts to incorporate a broader spectrum of preserved specimens would address the limitations imposed by the current dataset.</p>
<p>Another critical aspect warranting further evaluation is the usability of deep learning-based applications across diverse biodiversity datasets. Research by <xref ref-type="bibr" rid="ref5">Campos et al. (2024)</xref>, <xref ref-type="bibr" rid="ref6">Campos et al. (2023)</xref> has demonstrated that the efficacy of artificial intelligence algorithms in species identification can vary significantly depending on the animal dataset. This variability underscores the potential utility of applications, such as the one proposed in this study, as supportive technologies for field experts rather than as standalone solutions intended to replace human expertise.</p>
<p>Overall, this study underscores the necessity for further development and investigation of reliable models for biodiversity monitoring and research, with particular emphasis on endemic Amazonian lizards. The promising results presented herein pave the way for future advancements in this critical area of conservation science.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec24">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="ethics-statement" id="sec25">
<title>Ethics statement</title>
<p>The requirement of ethical approval was waived by Comit&#x00EA; de &#x00C9;tica do Museu Paraense Em&#x00ED;lio Goeldi for the studies involving animals because the research exclusively utilized preserved specimens from the herpetological collection of the Museu Paraense Em&#x00ED;lio Goeldi (MPEG). No living animals were handled, manipulated, or harmed during this study. The specimens used were already part of the museum&#x2019;s scientific collection, preserved in alcohol, and had been previously collected under appropriate permits for museum curation purposes. The research involved only photographic documentation of these preserved specimens for the development and validation of the artificial intelligence classification system. Therefore, no additional animal ethics committee approval was required for this type of archival specimen-based research. The studies were conducted in accordance with the local legislation and institutional requirements.</p>
</sec>
<sec sec-type="author-contributions" id="sec26">
<title>Author contributions</title>
<p>AS: Conceptualization, Data curation, Formal analysis, Investigation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. RO: Conceptualization, Investigation, Methodology, Validation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. CB: Data curation, Formal analysis, Validation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. EC: Conceptualization, Data curation, Investigation, Supervision, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. BG: Formal analysis, Validation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing, Data curation, Methodology, Supervision.</p>
</sec>
<sec sec-type="funding-information" id="sec27">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research and/or publication of this article.</p>
</sec>
<ack>
<p>We thank the Herpetology Laboratory from the Museu Paraense Emilio Goeldi for permitting us to collect the data used in this study. Our sincere appreciation also goes to the Instituto Tecnologico Vale for their support, providing full research scholarships for two of the authors, AGS and COB.</p>
</ack>
<sec sec-type="COI-statement" id="sec28">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec29">
<title>Generative AI statement</title>
<p>The author(s) declare that no Gen AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="sec30">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Abadi</surname> <given-names>M.</given-names></name> <name><surname>Agarwal</surname> <given-names>A.</given-names></name> <name><surname>Barham</surname> <given-names>P.</given-names></name> <name><surname>Brevdo</surname> <given-names>E.</given-names></name> <name><surname>Chen</surname> <given-names>Z.</given-names></name> <name><surname>Citro</surname> <given-names>C.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Tensor flow: large-scale machine learning on heterogeneous distributed systems</article-title>. <source>arXiv</source>.</citation></ref>
<ref id="ref2"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bell</surname> <given-names>T. P.</given-names></name></person-group> (<year>2009</year>). <article-title>A novel technique for monitoring highly cryptic lizard species in forests</article-title>. <source>Herpetol. Conserv. Biol.</source> <volume>4</volume>, <fpage>415</fpage>&#x2013;<lpage>425</lpage>.</citation></ref>
<ref id="ref3"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Binta Islam</surname> <given-names>S.</given-names></name> <name><surname>Valles</surname> <given-names>D.</given-names></name> <name><surname>Hibbitts</surname> <given-names>T. J.</given-names></name> <name><surname>Ryberg</surname> <given-names>W. A.</given-names></name> <name><surname>Walkup</surname> <given-names>D. K.</given-names></name> <name><surname>Forstner</surname> <given-names>M. R. J.</given-names></name></person-group> (<year>2023</year>). <article-title>Animal species recognition with deep convolutional neural networks from ecological camera trap images</article-title>. <source>Animals</source> <volume>13</volume>:<fpage>1526</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ani13091526</pub-id>, PMID: <pub-id pub-id-type="pmid">37174563</pub-id></citation></ref>
<ref id="ref4"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bolon</surname> <given-names>I.</given-names></name> <name><surname>Picek</surname> <given-names>L.</given-names></name> <name><surname>Durso</surname> <given-names>A. M.</given-names></name> <name><surname>Alcoba</surname> <given-names>G.</given-names></name> <name><surname>Chappuis</surname> <given-names>F.</given-names></name> <name><surname>Ruiz De Casta&#x00F1;eda</surname> <given-names>R.</given-names></name></person-group> (<year>2022</year>). <article-title>An artificial intelligence model to identify snakes from across the world: opportunities and challenges for Global Health and herpetology</article-title>. <source>SSRN Electron. J.</source> 16, e0010647. doi: <pub-id pub-id-type="doi">10.1371/journal.pntd.0010647</pub-id></citation></ref>
<ref id="ref5"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Campos</surname> <given-names>D. S.</given-names></name> <name><surname>de Oliveira</surname> <given-names>R. F.</given-names></name> <name><surname>Oliveira Vieira</surname> <given-names>L. de</given-names></name> <name><surname>Bragan&#x00E7;a</surname> <given-names>P. H. N.</given-names></name> <name><surname>Guimar&#x00E3;es</surname> <given-names>E. C.</given-names></name> <name><surname>Katz</surname> <given-names>A. M.</given-names></name> <etal/></person-group> (<year>2024</year>). <article-title>Well-intentioned initiatives hinder understanding biodiversity conservation: an essay on a recent deep-learning image classifier for Amazonian fishes</article-title> <source>Rev. Fish Biol. Fish.</source> <fpage>187</fpage>&#x2013;<lpage>200</lpage> doi: <pub-id pub-id-type="doi">10.1007/s11160-024-09901-y</pub-id> <volume>35</volume></citation></ref>
<ref id="ref6"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Campos</surname> <given-names>D. S.</given-names></name> <name><surname>Oliveira</surname> <given-names>R. F. D.</given-names></name> <name><surname>Vieira</surname> <given-names>L. D. O.</given-names></name> <name><surname>Bragan&#x00E7;a</surname> <given-names>P. H. N. D.</given-names></name> <name><surname>Nunes</surname> <given-names>J. L. S.</given-names></name> <name><surname>Guimar&#x00E3;es</surname> <given-names>E. C.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Revisiting the debate: documenting biodiversity in the age of digital and artificially generated images</article-title>. <source>Web Ecol.</source> <volume>23</volume>, <fpage>135</fpage>&#x2013;<lpage>144</lpage>. doi: <pub-id pub-id-type="doi">10.5194/we-23-135-2023</pub-id></citation></ref>
<ref id="ref7"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>R.</given-names></name> <name><surname>Little</surname> <given-names>R.</given-names></name> <name><surname>Mihaylova</surname> <given-names>L.</given-names></name> <name><surname>Delahay</surname> <given-names>R.</given-names></name> <name><surname>Cox</surname> <given-names>R.</given-names></name></person-group> (<year>2019</year>). <article-title>Wildlife surveillance using deep learning methods</article-title>. <source>Ecol. Evol.</source> <volume>9</volume>, <fpage>9453</fpage>&#x2013;<lpage>9466</lpage>. doi: <pub-id pub-id-type="doi">10.1002/ece3.5410</pub-id>, PMID: <pub-id pub-id-type="pmid">31534668</pub-id></citation></ref>
<ref id="ref8"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Da Costa Prudente</surname> <given-names>A. L.</given-names></name> <name><surname>Ramos</surname> <given-names>L.</given-names></name> <name><surname>Silva</surname> <given-names>T.</given-names></name> <name><surname>Sarmento</surname> <given-names>J.</given-names></name> <name><surname>Dourado</surname> <given-names>A.</given-names></name> <name><surname>Silva</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Dataset from the snakes (Serpentes, reptiles) collection of the Museu Paraense Em&#x00ED;lio Goeldi, Par&#x00E1;, Brazil</article-title>. <source>Biodivers Data J.</source> <volume>7</volume>:<fpage>e34013</fpage>. doi: <pub-id pub-id-type="doi">10.3897/BDJ.7.e34013</pub-id>, PMID: <pub-id pub-id-type="pmid">31086477</pub-id></citation></ref>
<ref id="ref9"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Durso</surname> <given-names>A. M.</given-names></name> <name><surname>Moorthy</surname> <given-names>G. K.</given-names></name> <name><surname>Mohanty</surname> <given-names>S. P.</given-names></name> <name><surname>Bolon</surname> <given-names>I.</given-names></name> <name><surname>Salath&#x00E9;</surname> <given-names>M.</given-names></name> <name><surname>Ruiz De Casta&#x00F1;eda</surname> <given-names>R.</given-names></name></person-group> (<year>2021</year>). <article-title>Supervised learning computer vision benchmark for Snake species identification from photographs: implications for herpetology and Global Health</article-title>. <source>Front Artif Intell.</source> <volume>4</volume>:<fpage>582110</fpage>. doi: <pub-id pub-id-type="doi">10.3389/frai.2021.582110</pub-id>, PMID: <pub-id pub-id-type="pmid">33959704</pub-id></citation></ref>
<ref id="ref10"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Frazier</surname> <given-names>P. I.</given-names></name></person-group> (<year>2018</year>). <article-title>A tutorial on Bayesian optimization</article-title>. <source>arXiv</source>.</citation></ref>
<ref id="ref11"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Gill</surname> <given-names>K. S.</given-names></name> <name><surname>Gupta</surname> <given-names>R</given-names></name> <name><surname>Malhotra</surname> <given-names>S</given-names></name> <name><surname>Devliyal</surname> <given-names>S</given-names></name> <name><surname>Sunil</surname> <given-names>G</given-names></name></person-group> (<year>2024</year>). Classification of reptiles and amphibians using transfer learning and deep convolutional neural networks. In: 2024 IEEE 9th International Conference for Convergence in Technology (I2CT) [Internet]. Pune, India: IEEE. pp. 1&#x2013;5. Available at: <ext-link xlink:href="https://ieeexplore.ieee.org/document/10544030/" ext-link-type="uri">https://ieeexplore.ieee.org/document/10544030/</ext-link> (Accessed November 2, 2024).</citation></ref>
<ref id="ref12"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gomez Villa</surname> <given-names>A.</given-names></name> <name><surname>Salazar</surname> <given-names>A.</given-names></name> <name><surname>Vargas</surname> <given-names>F.</given-names></name></person-group> (<year>2017</year>). <article-title>Towards automatic wild animal monitoring: identification of animal species in camera-trap images using very deep convolutional neural networks</article-title>. <source>Ecol. Inform.</source> <volume>41</volume>, <fpage>24</fpage>&#x2013;<lpage>32</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ecoinf.2017.07.004</pub-id></citation></ref>
<ref id="ref9001"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Gray</surname> <given-names>P. C.</given-names></name> <name><surname>Bierlich</surname> <given-names>K. C.</given-names></name> <name><surname>Mantell</surname> <given-names>S. A.</given-names></name> <name><surname>Friedlaender</surname> <given-names>A. S.</given-names></name> <name><surname>Goldbogen</surname> <given-names>J. A.</given-names></name> <name><surname>Johnston</surname> <given-names>D. W.</given-names></name></person-group> (<year>2019</year>). Drones and convolutional neural networks facilitate automated and accurate cetacean species identification and photogrammetry. <source>Methods Ecol Evol.</source> <volume>10</volume>, <fpage>1490</fpage>&#x2013;<lpage>1500</lpage>.</citation></ref>
<ref id="ref13"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Haq</surname> <given-names>A. U.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Peng</surname> <given-names>H.</given-names></name> <name><surname>Rahman</surname> <given-names>S. U.</given-names></name></person-group> (<year>2019</year>). <article-title>Combining multiple feature-ranking techniques and clustering of variables for feature selection</article-title>. <source>IEEE Access</source> <volume>7</volume>, <fpage>151482</fpage>&#x2013;<lpage>151492</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2019.2947701</pub-id></citation></ref>
<ref id="ref14"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Howard</surname> <given-names>A.</given-names></name> <name><surname>Sandler</surname> <given-names>M.</given-names></name> <name><surname>Chu</surname> <given-names>G.</given-names></name> <name><surname>Chen</surname> <given-names>L. C.</given-names></name> <name><surname>Chen</surname> <given-names>B.</given-names></name> <name><surname>Tan</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Searching for MobileNetV3</article-title>. <source>arXiv</source>.</citation></ref>
<ref id="ref15"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Howard</surname> <given-names>A. G.</given-names></name> <name><surname>Zhu</surname> <given-names>M.</given-names></name> <name><surname>Chen</surname> <given-names>B.</given-names></name> <name><surname>Kalenichenko</surname> <given-names>D.</given-names></name> <name><surname>Wang</surname> <given-names>W.</given-names></name> <name><surname>Weyand</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Mobile nets: efficient convolutional neural networks for mobile vision applications</article-title>. <source>arXiv</source>.</citation></ref>
<ref id="ref16"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Islam</surname> <given-names>SB</given-names></name> <name><surname>Valles</surname> <given-names>D</given-names></name></person-group> (<year>2020</year>). &#x201C;Identification of wild species in Texas from camera-trap images using deep neural network for conservation monitoring.&#x201D; In: <italic>2020 10th Annual Computing and Communication Workshop and Conference (CCWC)</italic>. Las Vegas, NV, USA: IEEE. pp. 0537&#x2013;0542.</citation></ref>
<ref id="ref17"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jovel</surname> <given-names>J.</given-names></name> <name><surname>Greiner</surname> <given-names>R.</given-names></name></person-group> (<year>2021</year>). <article-title>An introduction to machine learning approaches for biomedical research</article-title>. <source>Front. Med.</source> <volume>8</volume>:<fpage>771607</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmed.2021.771607</pub-id>, PMID: <pub-id pub-id-type="pmid">34977072</pub-id></citation></ref>
<ref id="ref18"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Kornblith</surname> <given-names>S</given-names></name> <name><surname>Shlens</surname> <given-names>J</given-names></name> <name><surname>Le</surname> <given-names>QV</given-names></name></person-group> (<year>2019</year>). &#x201C;Do Better ImageNet Models Transfer Better?&#x201D; in <italic>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) [internet]</italic>. Long Beach, CA, USA: IEEE. pp. 2656&#x2013;2666.</citation></ref>
<ref id="ref19"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Martineau</surname> <given-names>C.</given-names></name> <name><surname>Conte</surname> <given-names>D.</given-names></name> <name><surname>Raveaux</surname> <given-names>R.</given-names></name> <name><surname>Arnault</surname> <given-names>I.</given-names></name> <name><surname>Munier</surname> <given-names>D.</given-names></name> <name><surname>Venturini</surname> <given-names>G.</given-names></name></person-group> (<year>2017</year>). <article-title>A survey on image-based insect classification</article-title>. <source>Pattern Recogn.</source> <volume>65</volume>, <fpage>273</fpage>&#x2013;<lpage>284</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.patcog.2016.12.020</pub-id></citation></ref>
<ref id="ref20"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>M&#x00E4;yr&#x00E4;</surname> <given-names>J.</given-names></name> <name><surname>Keski-Saari</surname> <given-names>S.</given-names></name> <name><surname>Kivinen</surname> <given-names>S.</given-names></name> <name><surname>Tanhuanp&#x00E4;&#x00E4;</surname> <given-names>T.</given-names></name> <name><surname>Hurskainen</surname> <given-names>P.</given-names></name> <name><surname>Kullberg</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Tree species classification from airborne hyperspectral and LiDAR data using 3D convolutional neural networks</article-title>. <source>Remote Sens. Environ.</source> <volume>256</volume>:<fpage>112322</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.rse.2021.112322</pub-id></citation></ref>
<ref id="ref21"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Miao</surname> <given-names>Z.</given-names></name> <name><surname>Gaynor</surname> <given-names>K. M.</given-names></name> <name><surname>Wang</surname> <given-names>J.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name> <name><surname>Muellerklein</surname> <given-names>O.</given-names></name> <name><surname>Norouzzadeh</surname> <given-names>M. S.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Insights and approaches using deep learning to classify wildlife</article-title>. <source>Sci. Rep.</source> <volume>9</volume>:<fpage>8137</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-019-44565-w</pub-id>, PMID: <pub-id pub-id-type="pmid">31148564</pub-id></citation></ref>
<ref id="ref22"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Michele</surname> <given-names>A.</given-names></name> <name><surname>Colin</surname> <given-names>V.</given-names></name> <name><surname>Santika</surname> <given-names>D. D.</given-names></name></person-group> (<year>2019</year>). <article-title>Mobilenet convolutional neural networks and support vector machines for palmprint recognition</article-title>. <source>Procedia Comput. Sci.</source> <volume>157</volume>, <fpage>110</fpage>&#x2013;<lpage>117</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.procs.2019.08.147</pub-id></citation></ref>
<ref id="ref23"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Murphy</surname> <given-names>J. C.</given-names></name> <name><surname>Lehtinen</surname> <given-names>R. M.</given-names></name> <name><surname>Charles</surname> <given-names>S. P.</given-names></name> <name><surname>Wasserman</surname> <given-names>D.</given-names></name> <name><surname>Anton</surname> <given-names>T.</given-names></name> <name><surname>Brennan</surname> <given-names>P. J.</given-names></name></person-group> (<year>2017</year>). <article-title>Cryptic multicolored lizards in the <italic>Polychrus marmoratus</italic> group (Squamata: Sauria: Polychrotidae) and the status of <italic>Leiolepis auduboni</italic> Hallowell</article-title>. <source>Amphib. Reptile.</source> <volume>11</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>.</citation></ref>
<ref id="ref24"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Naser</surname> <given-names>M. Z.</given-names></name> <name><surname>Alavi</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Insights into performance fitness and error metrics for machine learning</article-title>. <source>Archit. Struct. Constr.</source> <volume>3</volume>, <fpage>499</fpage>&#x2013;<lpage>517</lpage>.</citation></ref>
<ref id="ref25"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nezami</surname> <given-names>S.</given-names></name> <name><surname>Khoramshahi</surname> <given-names>E.</given-names></name> <name><surname>Nevalainen</surname> <given-names>O.</given-names></name> <name><surname>P&#x00F6;l&#x00F6;nen</surname> <given-names>I.</given-names></name> <name><surname>Honkavaara</surname> <given-names>E.</given-names></name></person-group> (<year>2020</year>). <article-title>Tree species classification of drone hyperspectral and RGB imagery with deep learning convolutional neural networks</article-title>. <source>Remote Sens</source> <volume>12</volume>:<fpage>1070</fpage>. doi: <pub-id pub-id-type="doi">10.3390/rs12071070</pub-id></citation></ref>
<ref id="ref26"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Norouzzadeh</surname> <given-names>M. S.</given-names></name> <name><surname>Morris</surname> <given-names>D.</given-names></name> <name><surname>Beery</surname> <given-names>S.</given-names></name> <name><surname>Joshi</surname> <given-names>N.</given-names></name> <name><surname>Jojic</surname> <given-names>N.</given-names></name> <name><surname>Clune</surname> <given-names>J.</given-names></name></person-group> (<year>2021</year>). <article-title>A deep active learning system for species identification and counting in camera trap images</article-title> <comment>Schofield M, editor.</comment> <source>Methods Ecol. Evol.</source> <volume>12</volume>, <fpage>150</fpage>&#x2013;<lpage>161</lpage>. doi: <pub-id pub-id-type="doi">10.1111/2041-210X.13504</pub-id></citation></ref>
<ref id="ref27"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pedregosa</surname> <given-names>F.</given-names></name> <name><surname>Varoquaux</surname> <given-names>G.</given-names></name> <name><surname>Gramfort</surname> <given-names>A.</given-names></name> <name><surname>Michel</surname> <given-names>V.</given-names></name> <name><surname>Thirion</surname> <given-names>B.</given-names></name></person-group> (<year>2008</year>). <article-title>Scikit-learn: machine learning in python</article-title>. <source>J. Mach. Learn.</source> <volume>9</volume>, <fpage>2825</fpage>&#x2013;<lpage>2830</lpage>.</citation></ref>
<ref id="ref28"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pinho</surname> <given-names>C.</given-names></name> <name><surname>Kaliontzopoulou</surname> <given-names>A.</given-names></name> <name><surname>Ferreira</surname> <given-names>C. A.</given-names></name> <name><surname>Gama</surname> <given-names>J.</given-names></name></person-group> (<year>2023</year>). <article-title>Identification of morphologically cryptic species with computer vision models: wall lizards (Squamata: Lacertidae: <italic>Podarcis</italic>) as a case study</article-title>. <source>Zool. J. Linnean Soc.</source> <volume>198</volume>, <fpage>184</fpage>&#x2013;<lpage>201</lpage>. doi: <pub-id pub-id-type="doi">10.1093/zoolinnean/zlac087</pub-id></citation></ref>
<ref id="ref29"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Polonen</surname> <given-names>I</given-names></name> <name><surname>Annala</surname> <given-names>L</given-names></name> <name><surname>Rahkonen</surname> <given-names>S</given-names></name> <name><surname>Nevalainen</surname> <given-names>O</given-names></name> <name><surname>Honkavaara</surname> <given-names>E</given-names></name> <name><surname>Tuominen</surname> <given-names>S</given-names></name> <etal/></person-group> (<year>2018</year>). &#x201C;Tree species identification using 3D spectral data and 3D convolutional neural network.&#x201D; In: <italic>2018 9th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS)</italic>. Amsterdam, Netherlands: IEEE. pp. 1&#x2013;5.</citation></ref>
<ref id="ref30"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Pyron</surname> <given-names>R.</given-names></name> <name><surname>Burbrink</surname> <given-names>F. T.</given-names></name> <name><surname>Wiens</surname> <given-names>J. J.</given-names></name></person-group> (<year>2013</year>). <article-title>A phylogeny and revised classification of Squamata, including 4161 species of lizards and snakes</article-title>. <source>BMC Evol. Biol.</source> <volume>13</volume>:<fpage>93</fpage>. doi: <pub-id pub-id-type="doi">10.1186/1471-2148-13-93</pub-id>, PMID: <pub-id pub-id-type="pmid">23627680</pub-id></citation></ref>
<ref id="ref31"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Raju</surname> <given-names>VNG</given-names></name> <name><surname>Lakshmi</surname> <given-names>KP</given-names></name> <name><surname>Jain</surname> <given-names>VM</given-names></name> <name><surname>Kalidindi</surname> <given-names>A</given-names></name> <name><surname>Padma</surname> <given-names>V</given-names></name></person-group> (<year>2020</year>). &#x201C;Study the influence of normalization/transformation process on the accuracy of supervised classification.: In <italic>2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT)</italic>. Tirunelveli, India: IEEE. pp. 729&#x2013;735.</citation></ref>
<ref id="ref32"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Robillard</surname> <given-names>A. J.</given-names></name> <name><surname>Trizna</surname> <given-names>M. G.</given-names></name> <name><surname>Ruiz-Tafur</surname> <given-names>M.</given-names></name> <name><surname>D&#x00E1;vila Panduro</surname> <given-names>E. L.</given-names></name> <name><surname>Santana</surname> <given-names>C. D.</given-names></name> <name><surname>White</surname> <given-names>A. E.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Application of a deep learning image classifier for identification of Amazonian fishes</article-title>. <source>Ecol. Evol.</source> <volume>13</volume>:<fpage>e9987</fpage>. doi: <pub-id pub-id-type="doi">10.1002/ece3.9987</pub-id>, PMID: <pub-id pub-id-type="pmid">37143991</pub-id></citation></ref>
<ref id="ref33"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Rossum</surname> <given-names>GV</given-names></name></person-group> (<year>1995</year>). Python Tutorial Centrum voor Wiskunde en Informatica Amsterdam. The Netherlands.</citation></ref>
<ref id="ref34"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Sandler</surname> <given-names>M</given-names></name> <name><surname>Howard</surname> <given-names>A</given-names></name> <name><surname>Zhu</surname> <given-names>M</given-names></name> <name><surname>Zhmoginov</surname> <given-names>A</given-names></name> <name><surname>Chen</surname> <given-names>LC</given-names></name></person-group> (<year>2018</year>). &#x201C;MobileNetV2: inverted residuals and linear bottlenecks.&#x201D; In: <italic>2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition</italic>. Salt Lake City, UT: IEEE. p. 4510&#x2013;4520.</citation></ref>
<ref id="ref35"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Sharmin</surname> <given-names>I.</given-names></name> <name><surname>Islam</surname> <given-names>N. F.</given-names></name> <name><surname>Jahan</surname> <given-names>I.</given-names></name> <name><surname>Ahmed Joye</surname> <given-names>T.</given-names></name> <name><surname>Rahman</surname> <given-names>M. R.</given-names></name> <name><surname>Habib</surname> <given-names>M. T.</given-names></name></person-group> (<year>2019</year>). <article-title>Machine vision based local fish recognition</article-title>. <source>SN Appl. Sci.</source> <volume>1</volume>:<fpage>1529</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s42452-019-1568-z</pub-id></citation></ref>
<ref id="ref36"><citation citation-type="book"><person-group person-group-type="author"><name><surname>Sowmya</surname> <given-names>M.</given-names></name> <name><surname>Balasubramanian</surname> <given-names>M.</given-names></name> <name><surname>Vaidehi</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). &#x201C;<article-title>Classification of animals using MobileNet with SVM classifier</article-title>&#x201D; in <source>Computational methods and data engineering</source>. eds. <person-group person-group-type="editor"><name><surname>Asari</surname> <given-names>V. K.</given-names></name> <name><surname>Singh</surname> <given-names>V.</given-names></name> <name><surname>Rajasekaran</surname> <given-names>R.</given-names></name> <name><surname>Patel</surname> <given-names>R. B.</given-names></name></person-group>, <series>Lecture Notes on Data Engineering and Communications Technologies; vol. 139</series> (<publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer Nature Singapore</publisher-name>), <fpage>347</fpage>&#x2013;<lpage>358</lpage>.</citation></ref>
<ref id="ref37"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Stewart</surname> <given-names>G.</given-names></name> <name><surname>Daniel</surname> <given-names>R. S.</given-names></name></person-group> (<year>1075</year>). <article-title>Microornamentation of lizard scales: some variations and taxonomic correlations</article-title>. <source>Herpetologica</source> <volume>31</volume>, <fpage>117</fpage>&#x2013;<lpage>130</lpage>.</citation></ref>
<ref id="ref38"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Torres-Carvajal</surname> <given-names>O.</given-names></name> <name><surname>De Queiroz</surname> <given-names>K.</given-names></name></person-group> (<year>2009</year>). <article-title>Phylogeny of hoplocercine lizards (Squamata: Iguania) with estimates of relative divergence times</article-title>. <source>Mol. Phylogenet. Evol.</source> <volume>50</volume>, <fpage>31</fpage>&#x2013;<lpage>43</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ympev.2008.10.002</pub-id>, PMID: <pub-id pub-id-type="pmid">18952184</pub-id></citation></ref>
<ref id="ref39"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Tuia</surname> <given-names>D.</given-names></name> <name><surname>Kellenberger</surname> <given-names>B.</given-names></name> <name><surname>Beery</surname> <given-names>S.</given-names></name> <name><surname>Costelloe</surname> <given-names>B. R.</given-names></name> <name><surname>Zuffi</surname> <given-names>S.</given-names></name> <name><surname>Risse</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Perspectives in machine learning for wildlife conservation</article-title>. <source>Nat. Commun.</source> <volume>13</volume>:<fpage>792</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41467-022-27980-y</pub-id>, PMID: <pub-id pub-id-type="pmid">35140206</pub-id></citation></ref>
<ref id="ref40"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>van der Maaten</surname> <given-names>L.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2008</year>). <article-title>Visualizing data using t-SNE</article-title>. <source>J. Mach. Learn.</source> <volume>9</volume>, <fpage>2579</fpage>&#x2013;<lpage>2605</lpage>.</citation></ref>
<ref id="ref41"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Vitt</surname> <given-names>L. J.</given-names></name> <name><surname>Avila-Pires</surname> <given-names>T. C. S.</given-names></name> <name><surname>Zani</surname> <given-names>P. A.</given-names></name> <name><surname>Sartorius</surname> <given-names>S. S.</given-names></name> <name><surname>Esp&#x00F3;sito</surname> <given-names>M. C.</given-names></name></person-group> (<year>2003</year>). <article-title>Life above ground: ecology of <italic>Anolis fuscoauratus</italic> in the Amazon rain forest, and comparisons with its nearest relatives</article-title>. <source>Can. J. Zool.</source> <volume>81</volume>, <fpage>142</fpage>&#x2013;<lpage>156</lpage>. doi: <pub-id pub-id-type="doi">10.1139/z02-230</pub-id></citation></ref>
<ref id="ref42"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>W&#x00E4;ldchen</surname> <given-names>J.</given-names></name> <name><surname>M&#x00E4;der</surname> <given-names>P.</given-names></name></person-group> (<year>2018</year>). <article-title>Machine learning for image based species identification</article-title>. <source>Methods Ecol. Evol.</source> <volume>9</volume>, <fpage>2216</fpage>&#x2013;<lpage>2225</lpage>.</citation></ref>
<ref id="ref43"><citation citation-type="other"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Van Stein</surname> <given-names>B</given-names></name> <name><surname>Emmerich</surname> <given-names>M</given-names></name> <name><surname>Back</surname> <given-names>T</given-names></name></person-group> (<year>2017</year>). &#x201C;A new acquisition function for Bayesian optimization based on the moment-generating function.&#x201D; In: <italic>2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC)</italic>. Banff, AB: IEEE. pp. 507&#x2013;512.</citation></ref>
<ref id="ref44"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Weinstein</surname> <given-names>B. G.</given-names></name></person-group> (<year>2018</year>). <article-title>A computer vision for animal ecology. Prugh L, editor</article-title>. <source>J. Anim. Ecol.</source> <volume>87</volume>, <fpage>533</fpage>&#x2013;<lpage>545</lpage>. doi: <pub-id pub-id-type="doi">10.1111/1365-2656.12780</pub-id>, PMID: <pub-id pub-id-type="pmid">29111567</pub-id></citation></ref>
<ref id="ref45"><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>M.</given-names></name> <name><surname>Yoon</surname> <given-names>S.</given-names></name> <name><surname>Fuentes</surname> <given-names>A.</given-names></name> <name><surname>Park</surname> <given-names>D. S.</given-names></name></person-group> (<year>2023</year>). <article-title>A comprehensive survey of image augmentation techniques for deep learning</article-title>. <source>Pattern Recogn.</source> <volume>137</volume>:<fpage>109347</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.patcog.2023.109347</pub-id></citation></ref>
</ref-list>
</back>
</article>