<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="correction">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Computer Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-9898</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fcomp.2025.1734114</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Correction</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Correction: Optimizing architectural-feature tradeoffs in Arabic automatic short answer grading: comparative analysis of fine-tuned AraBERTv2 models</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<collab>Frontiers Production Office</collab>
<xref ref-type="aff" rid="aff1"/>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/20170"/>
</contrib>
</contrib-group>
<aff id="aff1"><institution>Frontiers Media SA</institution>, <city>Lausanne</city>, <country country="ch">Switzerland</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Frontiers Production Office, <email xlink:href="mailto:production.office@frontiersin.org">production.office@frontiersin.org</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-11-12">
<day>12</day>
<month>11</month>
<year>2025</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>7</volume>
<elocation-id>1734114</elocation-id>
<history>
<date date-type="received">
<day>28</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>28</day>
<month>10</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2025 Frontiers Production Office.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Frontiers Production Office</copyright-holder>
<license>
<ali:license_ref start_date="2025-11-12">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<related-article id="RA1" related-article-type="corrected-article" journal-id="Front. Comput. Sci." journal-id-type="nlm-ta" vol="7" page="1683272" xlink:href="10.3389/fcomp.2025.1683272" ext-link-type="doi">A Correction on <article-title>Optimizing architectural-feature tradeoffs in Arabic automatic short answer grading: comparative analysis of fine-tuned AraBERTv2 models</article-title> by Mahmood, S. A. (2025). <italic>Front. Comput. Sci</italic>. 7:1683272. doi: <object-id>10.3389/fcomp.2025.1683272</object-id></related-article>
<kwd-group>
<kwd>large language model (LLMs)</kwd>
<kwd>AraBERT</kwd>
<kwd>neural network</kwd>
<kwd>Arabic natural language processing</kwd>
<kwd>educational assessment</kwd>
<kwd>Automated Short Answer Grading (ASAG)</kwd>
</kwd-group>
<counts>
<fig-count count="8"/>
<table-count count="7"/>
<equation-count count="0"/>
<ref-count count="0"/>
<page-count count="6"/>
<word-count count="1086"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Digital Education</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<p>There was a mistake in the article as published. <xref ref-type="table" rid="T1">Tables 1</xref>&#x02013;<xref ref-type="table" rid="T7">7</xref> and <xref ref-type="fig" rid="F1">Figures 1</xref>&#x02013;<xref ref-type="fig" rid="F8">8</xref> were published as supplementary material when they should have been added to the main article. The corrected figures and tables appear below.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Distribution of answers by question type.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Question type</bold></th>
<th valign="top" align="center"><bold>Question type (In Arabic)</bold></th>
<th valign="top" align="center"><bold>Total questions</bold></th>
<th valign="top" align="center"><bold>Total answers</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Define the scientific term</td>
<td valign="top" align="center"><inline-graphic xlink:href="fcomp-07-1734114-i0001.tif"/></td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">291</td>
</tr>
<tr>
<td valign="top" align="left">Explain</td>
<td valign="top" align="center"><inline-graphic xlink:href="fcomp-07-1734114-i0002.tif"/></td>
<td valign="top" align="center">21</td>
<td valign="top" align="center">830</td>
</tr>
<tr>
<td valign="top" align="left">What are the consequences of</td>
<td valign="top" align="center"><inline-graphic xlink:href="fcomp-07-1734114-i0003.tif"/></td>
<td valign="top" align="center">6</td>
<td valign="top" align="center">282</td>
</tr>
<tr>
<td valign="top" align="left">Justify or give reasons for</td>
<td valign="top" align="center"><inline-graphic xlink:href="fcomp-07-1734114-i0004.tif"/></td>
<td valign="top" align="center">10</td>
<td valign="top" align="center">465</td>
</tr>
<tr>
<td valign="top" align="left">What is the difference between</td>
<td valign="top" align="center"><inline-graphic xlink:href="fcomp-07-1734114-i0005.tif"/></td>
<td valign="top" align="center">5</td>
<td valign="top" align="center">217</td>
</tr>
<tr>
<td valign="top" align="left">Total</td>
<td valign="top" align="center">5 types</td>
<td valign="top" align="center">48</td>
<td valign="top" align="center">2,085</td>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Detailed distribution of randomly sampled responses across selected questions.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Q&#x02013;No</bold>.</th>
<th valign="top" align="left"><bold>Question type</bold></th>
<th valign="top" align="center"><bold>Total answers</bold></th>
<th valign="top" align="center"><bold>Training answers</bold></th>
<th valign="top" align="center"><bold>Test answers</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1</td>
<td valign="top" align="left">Define the scientific term</td>
<td valign="top" align="center">46</td>
<td valign="top" align="center">36</td>
<td valign="top" align="center">10</td>
</tr>
<tr>
<td valign="top" align="left">26</td>
<td valign="top" align="left">Explain</td>
<td valign="top" align="center">47</td>
<td valign="top" align="center">37</td>
<td valign="top" align="center">10</td>
</tr>
<tr>
<td valign="top" align="left">28</td>
<td valign="top" align="left">What are the consequences of</td>
<td valign="top" align="center">48</td>
<td valign="top" align="center">38</td>
<td valign="top" align="center">10</td>
</tr>
<tr>
<td valign="top" align="left">35</td>
<td valign="top" align="left">Justify or give reasons for</td>
<td valign="top" align="center">51</td>
<td valign="top" align="center">40</td>
<td valign="top" align="center">11</td>
</tr>
<tr>
<td valign="top" align="left">45</td>
<td valign="top" align="left">What is the difference between</td>
<td valign="top" align="center">36</td>
<td valign="top" align="center">28</td>
<td valign="top" align="center">8</td>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Performance evaluation of AraBERTv2 with MLP model using different feature sets: training vs. testing results.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Stage</bold></th>
<th valign="top" align="left"><bold>No. of feature</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>Pearson correlation</bold></th>
<th valign="top" align="center"><bold>Spearman&#x00027;s correlation</bold></th>
<th valign="top" align="center"><bold>Epoch 1&#x02013;5</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">AraBERTv2 with MLP</td>
<td valign="top" align="left">Training</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.14</td>
<td valign="top" align="center">1.51</td>
<td valign="top" align="center">0.847</td>
<td valign="top" align="center">0.85</td>
<td valign="top" align="center">898 &#x02192; 533 &#x02192; 347 &#x02192; 250 &#x02192; 156</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.2</td>
<td valign="top" align="center">1.58</td>
<td valign="top" align="center">0.818</td>
<td valign="top" align="center">0.816</td>
<td valign="top" align="center">1,026 &#x02192; 614 &#x02192; 263 &#x02192; 185</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">0.18</td>
<td valign="top" align="center">0.2</td>
<td valign="top" align="center">0.999</td>
<td valign="top" align="center">0.998</td>
<td valign="top" align="center">713 &#x02192; 34 &#x02192; 13 &#x02192; 9 &#x02192; 7</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Testing</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.31</td>
<td valign="top" align="center">1.76</td>
<td valign="top" align="center">0.803</td>
<td valign="top" align="center">0.808</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.48</td>
<td valign="top" align="center">1.9</td>
<td valign="top" align="center">0.744</td>
<td valign="top" align="center">0.746</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">1.77</td>
<td valign="top" align="center">2.22</td>
<td valign="top" align="center">0.691</td>
<td valign="top" align="center">0.689</td>
<td/>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Performance evaluation of AraBERTv2 with CNN model using different feature sets: training vs. testing results.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Stage</bold></th>
<th valign="top" align="left"><bold>No. of features</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>Pearson correlation</bold></th>
<th valign="top" align="center"><bold>Spearman&#x00027;s correlation</bold></th>
<th valign="top" align="center"><bold>Epoch 1&#x02013;5</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">AraBERTv2 with CNN</td>
<td valign="top" align="left">Training</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.22</td>
<td valign="top" align="center">1.59</td>
<td valign="top" align="center">0.849</td>
<td valign="top" align="center">0.843</td>
<td valign="top" align="center">1,092 &#x02192; 610 &#x02192; 427 &#x02192; 306 &#x02192; 227</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.17</td>
<td valign="top" align="center">1.53</td>
<td valign="top" align="center">0.833</td>
<td valign="top" align="center">0.832</td>
<td valign="top" align="center">1,057 &#x02192; 567 &#x02192; 379 &#x02192; 280 &#x02192; 205</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">0.24</td>
<td valign="top" align="center">0.27</td>
<td valign="top" align="center">0.999</td>
<td valign="top" align="center">0.998</td>
<td valign="top" align="center">773 &#x02192; 28 &#x02192; 12 &#x02192; 8 &#x02192; 6</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Testing</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.45</td>
<td valign="top" align="center">1.93</td>
<td valign="top" align="center">0.784</td>
<td valign="top" align="center">0.788</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.6</td>
<td valign="top" align="center">2.02</td>
<td valign="top" align="center">0.746</td>
<td valign="top" align="center">0.75</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">2.63</td>
<td valign="top" align="center">3.07</td>
<td valign="top" align="center">0.607</td>
<td valign="top" align="center">0.613</td>
<td/>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Performance evaluation of AraBERTv2 with LSTM model using different feature sets: training vs. testing results.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Stage</bold></th>
<th valign="top" align="left"><bold>No. of features</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>Pearson correlation</bold></th>
<th valign="top" align="center"><bold>Spearman&#x00027;s correlation</bold></th>
<th valign="top" align="center"><bold>Epoch 1&#x02013;5</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">AraBERTv2 with LSTM</td>
<td valign="top" align="left">Training</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.26</td>
<td valign="top" align="center">1.62</td>
<td valign="top" align="center">0.821</td>
<td valign="top" align="center">0.825</td>
<td valign="top" align="center">1,147 &#x02192; 718 &#x02192; 524 &#x02192; 356 &#x02192; 262</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.27</td>
<td valign="top" align="center">1.66</td>
<td valign="top" align="center">0.811</td>
<td valign="top" align="center">0.818</td>
<td valign="top" align="center">1,141 &#x02192; 675 &#x02192; 456 &#x02192; 349 &#x02192; 267</td>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">0.14</td>
<td valign="top" align="center">0.19</td>
<td valign="top" align="center">0.998</td>
<td valign="top" align="center">0.998</td>
<td valign="top" align="center">728 &#x02192; 62 &#x02192; 31 &#x02192; 22 &#x02192; 19</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Testing</td>
<td valign="top" align="left">2-feature</td>
<td valign="top" align="center">1.48</td>
<td valign="top" align="center">1.86</td>
<td valign="top" align="center">0.757</td>
<td valign="top" align="center">0.759</td>
<td/>
</tr>
 <tr>
<td/>
<td/>
<td valign="top" align="left">3-feature</td>
<td valign="top" align="center">1.6</td>
<td valign="top" align="center">2.03</td>
<td valign="top" align="center">0.757</td>
<td valign="top" align="center">0.77</td>
<td/>
</tr>
<tr>
<td/>
<td/>
<td valign="top" align="left">4-feature</td>
<td valign="top" align="center">3.62</td>
<td valign="top" align="center">4.19</td>
<td valign="top" align="center">0.388</td>
<td valign="top" align="center">0.419</td>
<td/>
</tr></tbody>
</table>
</table-wrap>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Performance comparison of AraBERTv2 fine-tuned models with MLP, CNN, and LSTM architectures using different feature sets.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Fine-tuned models</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>Pearson correlation</bold></th>
<th valign="top" align="center"><bold>Spearman&#x00027;s correlation</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">2-features-AraBERTv2 with MLP</td>
<td valign="top" align="center"><bold>1.31</bold></td>
<td valign="top" align="center"><bold>1.76</bold></td>
<td valign="top" align="center"><bold>0.803</bold></td>
<td valign="top" align="center"><bold>0.808</bold></td>
</tr>
<tr>
<td valign="top" align="left">2-features-AraBERTv2 with CNN</td>
<td valign="top" align="center">1.45</td>
<td valign="top" align="center">1.93</td>
<td valign="top" align="center">0.784</td>
<td valign="top" align="center">0.788</td>
</tr>
<tr>
<td valign="top" align="left">2-features-AraBERTv2 with LSTM</td>
<td valign="top" align="center">1.48</td>
<td valign="top" align="center">1.86</td>
<td valign="top" align="center">0.757</td>
<td valign="top" align="center">0.759</td>
</tr>
<tr>
<td valign="top" align="left">3-features-AraBERTv2 with MLP</td>
<td valign="top" align="center">1.48</td>
<td valign="top" align="center">1.9</td>
<td valign="top" align="center">0.744</td>
<td valign="top" align="center">0.746</td>
</tr>
<tr>
<td valign="top" align="left">3-features-AraBERTv2 with CNN</td>
<td valign="top" align="center">1.6</td>
<td valign="top" align="center">2.02</td>
<td valign="top" align="center">0.746</td>
<td valign="top" align="center">0.75</td>
</tr>
<tr>
<td valign="top" align="left">3-features-AraBERTv2 with LSTM</td>
<td valign="top" align="center"><bold>1.6</bold></td>
<td valign="top" align="center"><bold>2.03</bold></td>
<td valign="top" align="center"><bold>0.757</bold></td>
<td valign="top" align="center"><bold>0.77</bold></td>
</tr>
<tr>
<td valign="top" align="left">4-features-AraBERTv2 with MLP</td>
<td valign="top" align="center"><bold>1.77</bold></td>
<td valign="top" align="center"><bold>2.22</bold></td>
<td valign="top" align="center"><bold>0.691</bold></td>
<td valign="top" align="center"><bold>0.689</bold></td>
</tr>
<tr>
<td valign="top" align="left">4-features-AraBERTv2 with CNN</td>
<td valign="top" align="center">2.63</td>
<td valign="top" align="center">3.07</td>
<td valign="top" align="center">0.607</td>
<td valign="top" align="center">0.613</td>
</tr>
<tr>
<td valign="top" align="left">4-features-AraBERTv2 with LSTM</td>
<td valign="top" align="center">3.62</td>
<td valign="top" align="center">4.19</td>
<td valign="top" align="center">0.388</td>
<td valign="top" align="center">0.419</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The bold values represent the optimal results obtained from our experimental analysis.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>Comparative performance evaluation of Arabic Automated Short Answer Grading (ASAG) systems.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr style="background-color:#919498;color:#ffffff">
<th valign="top" align="left"><bold>Criterion/study</bold></th>
<th valign="top" align="left"><bold>Methodology</bold></th>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="center"><bold>Best RMSE</bold></th>
<th valign="top" align="center"><bold>Best Pearson/Spearman</bold></th>
<th valign="top" align="center"><bold>Key strength</bold></th>
<th valign="top" align="left"><bold>Primary limitation</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Our study (AraBERTv2)</td>
<td valign="top" align="left">-Fine-tuned AraBERTv2 with MLP/CNN/LSTM -Tested 2/3/4 feature configurations</td>
<td valign="top" align="left">AS-ARSG (2,133 answers)</td>
<td valign="top" align="center">1.31</td>
<td valign="top" align="left">-Pearson: 0.803 -Spearman: 0.808</td>
<td valign="top" align="left">Optimal balance between generalizability and accuracy with limited data</td>
<td valign="top" align="left">Performance degradation in LSTM with added features</td>
</tr>
<tr>
<td valign="top" align="left">(4)</td>
<td valign="top" align="left">Latent Semantic Analysis (LSA) with local/hybrid weighting</td>
<td valign="top" align="left">AR-ASAG (2,133 answers)</td>
<td valign="top" align="center">N/A</td>
<td valign="top" align="left">N/A</td>
<td valign="top" align="left">Effective semantic weighting</td>
<td valign="top" align="left">Limited capacity for capturing complex contextual relationships</td>
</tr>
<tr>
<td valign="top" align="left">(19)</td>
<td valign="top" align="left">-BERT vs. Word2Vec/AWN comparison -Intensive text preprocessing</td>
<td valign="top" align="left">-AR-ASAG (2,133) -Jordanian History (550)</td>
<td valign="top" align="center">1.00308</td>
<td valign="top" align="left">Pearson: 0.841902</td>
<td valign="top" align="left">Demonstrated BERT&#x00027;s superiority over traditional approaches</td>
<td valign="top" align="left">Heavy dependency on text normalization and stemming</td>
</tr></tbody>
</table>
</table-wrap>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>General workflow of the proposed automated Arabic short-answer grading model using AraBERTv2.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0001.tif">
<alt-text content-type="machine-generated">Flowchart illustrating the process of training and testing with the AR-ASAG dataset. The sequence includes dataset loading, preprocessing, and splitting into 80% training and 20% testing. The training subset undergoes feature selection and AraBERT training, leading to finetuned AraBERT models. These models are evaluated and compared, followed by visualization to determine the best AraBERT model. Arrows indicate the workflow and connections among the steps.</alt-text>
</graphic>
</fig>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>The AraBERT_MLP training methodology.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0002.tif">
<alt-text content-type="machine-generated">Bar charts compare AraBERTv2 with LSTM across training and testing phases. The top charts show MAE and RMSE, and Pearson and Spearman correlations for different features in training. The bottom charts depict the same metrics for testing, highlighting variations in error values and correlation coefficients across two, three, and four features.</alt-text>
</graphic>
</fig>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>The AraBERT_CNN training methodology.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0003.tif">
<alt-text content-type="machine-generated">Bar charts showing AraBERTv2 with CNN performance during training and testing. Training error bars compare MAE and RMSE for 2, 3, and 4 features, with RMSE generally higher. Correlation values for Pearson and Spearman increase with more features. Testing error values increase with more features, while correlation values decrease slightly as features increase.</alt-text>
</graphic>
</fig>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>The AraBERT_LSTM training methodology.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0004.tif">
<alt-text content-type="machine-generated">Flowchart depicting the AraBERT with MLP training stage. Feature selection leads to three models: 2-features (red), 3-features (green), and 4-features (purple). Each model connects to AraBERT with CNN, then fine-tuning AraBERT stages, ending with evaluation and comparison.</alt-text>
</graphic>
</fig>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>Performance evaluation of AraBERTv2 with MLP model using different feature sets: training vs. testing results.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0005.tif">
<alt-text content-type="machine-generated">Bar charts comparing AraBERTv2 with MLP performance in training and testing phases. In training, 4-feature shows minimal MAE and RMSE, with high Pearson and Spearman correlations. In testing, 2-feature has lower error values than 3-feature and 4-feature, though 4-feature performs slightly better in correlation values.</alt-text>
</graphic>
</fig>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>Performance evaluation of AraBERTv2 with CNN model using different feature sets: training vs. testing results.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0006.tif">
<alt-text content-type="machine-generated">Diagram illustrating a machine learning process, titled &#x0201C;AraBERT with MLP Training Stage.&#x0201D; It starts with &#x0201C;Feature Selection,&#x0201D; leads to three models: &#x0201C;2-features model,&#x0201D; &#x0201C;3-features model,&#x0201D; and &#x0201C;4-features model.&#x0201D; Each model goes to &#x0201C;AraBERT with CNN,&#x0201D; followed by &#x0201C;Fine tuning AraBERT,&#x0201D; and ends with &#x0201C;Evaluation and comparison.&#x0201D; Arrows indicate the flow direction.</alt-text>
</graphic>
</fig>
<fig position="float" id="F7">
<label>Figure 7</label>
<caption><p>Performance evaluation of AraBERTv2 with LSTM model using different feature sets: training vs. testing results.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0007.tif">
<alt-text content-type="machine-generated">Scatter plot titled &#x0201C;Model Performance: Error vs Spearman Correlation&#x0201D; showing different models' performance using colored markers: AraBERTv2 with MLP, CNN, and LSTM. The x-axis represents MAE (mean absolute error), where lower is better, and the y-axis represents Spearman&#x02019;s rank correlation, where higher is better. The plot uses different shapes to indicate feature numbers. Most points cluster between 1.5 to 2.0 MAE and 0.65 to 0.80 correlation, with one outlier beyond 3.5 MAE and below 0.45 correlation.</alt-text>
</graphic>
</fig>
<fig position="float" id="F8">
<label>Figure 8</label>
<caption><p>Fine-tuned models performance: MAE vs. spearman correlation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1734114-g0008.tif">
<alt-text content-type="machine-generated">Diagram of a machine learning workflow featuring AraBERT with MLP. It starts with feature selection, separating into three models: a 2-feature model in red, a 3-feature model in green, and a 4-feature model in blue. These feed into the AraBERT with MLP stage, which then advances to fine-tuning AraBERT in individual boxes. An evaluation and comparison stage follows, indicated by arrows.</alt-text>
</graphic>
</fig>
<p>All in-text Supplementary Table and Supplementary Figure in-text citations have been changed to Table and Figure in-text citations.</p>
<p>The original version of this article has been updated.</p>
</body>
<back>
<sec sec-type="ai-statement" id="s1">
<title>Generative AI statement</title>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Approved by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/739598/overview">Frontiers Editorial Office</ext-link>, Frontiers Media SA, Switzerland</p>
</fn>
</fn-group>
</back>
 </article>

