<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Public Health</journal-id>
<journal-title>Frontiers in Public Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Public Health</abbrev-journal-title>
<issn pub-type="epub">2296-2565</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpubh.2025.1479095</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Public Health</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Explainable AI-based feature importance analysis for ovarian cancer classification with ensemble methods</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Kodipalli</surname> <given-names>Ashwini</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1505122/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Devi</surname> <given-names>V. Susheela</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Guruvare</surname> <given-names>Shyamala</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<xref ref-type="corresp" rid="c002"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1937368/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Ismail</surname> <given-names>Taha</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2982790/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
</contrib-group>
<aff id="aff1"><sup>1</sup><institution>Department of Computer Science and Automation, Indian Institute of Science</institution>, <addr-line>Bangalore, Karnataka</addr-line>, <country>India</country></aff>
<aff id="aff2"><sup>2</sup><institution>School of Computer Science and Engineering, RV University, Bangalore</institution>, <addr-line>Karnataka</addr-line>, <country>India</country></aff>
<aff id="aff3"><sup>3</sup><institution>Department of Obstetrics and Gynecology, Kasturba Medical College, Manipal, Manipal Academy of Higher Education, Manipal</institution>, <addr-line>Karnataka</addr-line>, <country>India</country></aff>
<aff id="aff4"><sup>4</sup><institution>Department of Radiology, Kanachur Institute of Medical Sciences</institution>, <addr-line>Mangaluru, Karnataka</addr-line>, <country>India</country></aff>
<author-notes>
<fn fn-type="edited-by" id="fn0001">
<p>Edited by: Gilles Guillot, CSL Behring AG, Switzerland</p>
</fn>
<fn fn-type="edited-by" id="fn0002">
<p>Reviewed by: Xianlong Zeng, Ohio University, United States</p>
<p>Raffaella Massafra, National Cancer Institute Foundation (IRCCS), Italy</p>
</fn>
<corresp id="c001">&#x002A;Correspondence: Ashwini Kodipalli, <email>kashwini@iisc.ac.in</email></corresp>
<corresp id="c002">Shyamala Guruvare, <email>shyamala.g@manipal.edu</email></corresp>
</author-notes>
<pub-date pub-type="epub">
<day>26</day>
<month>03</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>13</volume>
<elocation-id>1479095</elocation-id>
<history>
<date date-type="received">
<day>11</day>
<month>08</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>11</day>
<month>02</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2025 Kodipalli, Devi, Guruvare and Ismail.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Kodipalli, Devi, Guruvare and Ismail</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Ovarian Cancer (OC) is one of the leading causes of cancer deaths among women. Despite recent advances in the medical field, such as surgery, chemotherapy, and radiotherapy interventions, there are only marginal improvements in the diagnosis of OC using clinical parameters, as the symptoms are very non-specific at the early stage. Owing to advances in computational algorithms, such as ensemble machine learning, it is now possible to identify complex patterns in clinical parameters. However, these complex patterns do not provide deeper insights into prediction and diagnosis. Explainable artificial intelligence (XAI) models, such as LIME and SHAP Kernels, can provide insights into the decision-making process of ensemble models, thus increasing their applicability.</p>
</sec>
<sec>
<title>Methods</title>
<p>The main aim of this study is to design a computer-aided diagnostic system that accurately classifies and detects ovarian cancer. To achieve this objective, a three-stage ensemble model and a game-theoretic approach based on SHAP values were built to evaluate and visualize the results, thus analyzing the important features responsible for prediction.</p>
</sec>
<sec>
<title>Results and Discussion</title>
<p>The results demonstrate the efficacy of the proposed model with an accuracy of 98.66%. The proposed model&#x2019;s consistency and advantages are compared with single classifiers. The SHAP values of the proposed model are validated using conventional statistical methods such as the <italic>p</italic>-test and Cohen&#x2019;s <italic>d</italic>-test to highlight the efficacy of the proposed method. To further validate the ranking of the features, we compared the <italic>p</italic>-values and Cohen&#x2019;s <italic>d</italic>-values of the top five and bottom five features. The study proposed and validated an AI-based method for the detection, diagnosis, and prognosis of OC using multi-modal real-life data, which mimics the move of a clinician approach with a demonstration of high performance. The proposed strategy can lead to reliable, accurate, and consistent AI solutions for the detection and management of OC with higher patient experience and outcomes at low cost, low morbidity, and low mortality. This can be beneficial for millions of women living in resource-constrained and challenging economies.</p>
</sec>
</abstract>
<kwd-group>
<kwd>interpretable AI</kwd>
<kwd>ensemble models</kwd>
<kwd>bagging</kwd>
<kwd>boosting</kwd>
<kwd>machine learning</kwd>
<kwd><italic>p</italic>-value</kwd>
<kwd>Cohen&#x2019;s</kwd>
<kwd>SHAP</kwd>
</kwd-group>
<counts>
<fig-count count="17"/>
<table-count count="4"/>
<equation-count count="0"/>
<ref-count count="45"/>
<page-count count="17"/>
<word-count count="8854"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Digital Public Health</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>Ovarian cancer (OC) is one of the most fatal gynecological cancers in women. With delayed detection and complexity of disease progression being the main challenges in the treatment of OC, it continues to be a considerable health concern. In 2023, ovarian cancer stands as the most prominent cause of death among gynecological cancers in the United States and is the second most common type of gynecological cancer. Approximately 19,000 new ovarian cancer cases are expected to be diagnosed in the United States annually, with a projected 13,000 casualties (<xref ref-type="bibr" rid="ref1">1</xref>). Approximately 300,000 new cases are diagnosed annually worldwide (<xref ref-type="bibr" rid="ref2">2</xref>). Approximately 80% of patients face a recurrence of the disease although they would have undergone initial treatment (<xref ref-type="bibr" rid="ref3">3</xref>). These facts emphasize the imperative need for developments in early detection and treatment strategies, which is an ongoing research field.</p>
<p>Diagnosis in the initial stages and its precise classification is crucial for cancer diagnosis and the development of personalized treatment plans. Yet, traditional diagnostic methods often fall short of the required accuracy and transparency, restricting their efficacy in clinical practice.</p>
<p>The use of the latest developments in the fields of machine learning (ML) and artificial intelligence (AI) could result in promising improvements in the diagnostic process. These technologies can aid in creating sophisticated models that can analyze large datasets and identify the complex patterns of different cancer types. Ensemble models, which combine the predictive capabilities of multiple classifiers, have exhibited better performance in classification tasks than single classifiers. Regardless of their effectiveness, these models encounter challenges with transparency and interpretability when applied in clinical settings.</p>
<p>The current study intends to focus on these challenges by incorporating an explainable AI (XAI) approach to the interpretation and classification of OC using an ensemble of machine learning models. The key objectives are to utilize the advantages of ensemble learning while applying explainable AI techniques and verify that the model&#x2019;s decision-making process is transparent and comprehensible to clinicians.</p>
<p>The following are the main contributions to this study:</p>
<list list-type="bullet">
<list-item>
<p>A three-stage ensemble model is proposed for accurate classification.</p>
</list-item>
<list-item>
<p>Explainable AI methods (LIME and SHAP) are applied to gain a deeper understanding of the decision-making process of the proposed model.</p>
</list-item>
<list-item>
<p>The results of the proposed model are validated using the p-test and Cohen&#x2019;s <italic>d</italic>-test.</p>
</list-item>
<list-item>
<p>The feature importance provided by the proposed model is validated using the Wilcoxon signed-rank test (<xref ref-type="bibr" rid="ref4">4</xref>).</p>
</list-item>
</list>
<p>The structure of the paper is organized as follows: Section 2 reviews related studies involving single classifiers, ensemble models, and deep learning approaches. Section 3 details the methodology, including an in-depth description of the datasets and the architectural framework. Section 4 presents the results obtained from the study, and the discussion section provides a detailed comparative study with the existing related studies. Finally, Section 5 concludes with potential directions for future research.</p>
<p>By integrating explainable AI with ensemble learning, this study seeks to bridge the gap between advanced machine learning methodologies and their practical implementation in medical diagnostics. This approach aims to enhance the detection and treatment of ovarian cancer by combining the predictive accuracy of ensemble models with the interpretability provided by explainable AI techniques.</p>
</sec>
<sec id="sec2">
<label>2</label>
<title>Related studies</title>
<p>Machine learning (ML) and deep learning (DL) have revolutionized various fields, and their impact on medicine is particularly profound. These technologies leverage vast amounts of data and sophisticated algorithms to enhance diagnostic accuracy, personalize treatment, predict disease outbreaks, and streamline administrative processes. This section provides an overview of how the ML and DL methods have transformed medical practices and research.</p>
<sec id="sec3">
<label>2.1</label>
<title>Related studies using single classifiers</title>
<p>Nopour et al. (<xref ref-type="bibr" rid="ref5">5</xref>) used a logistic regression (LR) model and an eXtreme Gradient Boosting (XGBoost) model for OC prediction and reported that XGBoost exhibited better accuracy than the regression model. Li et al. (<xref ref-type="bibr" rid="ref6">6</xref>) adopted a logistic regression (LR)-based radiomics model out of other omics models (support vector machine (SVM), K-nearest neighbors (KNNs), Random Forest (RF), and XGBoost) and achieved an accuracy of 84.8%. Kori et al. (<xref ref-type="bibr" rid="ref7">7</xref>) built a model using principal component analysis and ML algorithms that showed 96.7% sensitivity and 100% specificity. Klein et al. (<xref ref-type="bibr" rid="ref8">8</xref>) built a model using MALDI imaging and ML algorithms such as linear discriminant analysis (LDA), SVM-lin, and SVM-rbf for epithelial ovarian cancer (EOC) histotype classification, yielding a mean accuracy of 80% for LDA, 80% for SVM-lin, and 74% for SVM-rbf. Chao et al. (<xref ref-type="bibr" rid="ref9">9</xref>) designed an LR model for OC prediction and obtained the area under the curve (AUC), sensitivity, and specificity values of 0.903, 89.2, and 82.3% in the training dataset and 0.891, 88.9, and 76.75% in the test dataset, respectively. Schilling et al. (<xref ref-type="bibr" rid="ref10">10</xref>) used ML models using K-means clustering, Naive Bayes, logistic regression, and SVM for disease prediction and revealed that the LR model performed the best with a 0.89 f1 score. Hong et al. (<xref ref-type="bibr" rid="ref11">11</xref>) built a radiomics model based on LASSO Cox regression and reported good discrimination in both training and validation sets with C-indexes of 0.694 (95% confidence interval [CI]: 0.613&#x2013;0.775) and 0.709 (95% CI: 0.517&#x2013;0.901), respectively. Bahado-Singh et al. (<xref ref-type="bibr" rid="ref12">12</xref>) worked with ML algorithms (SVM, LDA, prediction analysis for microarrays (PAMs), and generalized linear model (GLM)) and obtained optimal results of AUC value (95% CI) 1.00 (0.9000&#x2013;1.0) with SVM and AUC (95% CI) 0.99 (0.9000&#x2013;1.0) with GLM, PAMs, and LDA, along with achieving an AUC value (95% CI) 1.00 (0.9&#x2013;1.0) with SVM models for CpG analysis. Sheela Lavanya et al. (<xref ref-type="bibr" rid="ref13">13</xref>) applied ML techniques (KNNs, SVM, DTs) alongside explainable AI to aid the early detection of OC and demonstrated that SVM performed with 85% base model accuracy. Feng Zhan et al. (<xref ref-type="bibr" rid="ref14">14</xref>) designed the LR radiomic model to predict lymphocyte-specific protein tyrosine kinase (LCK) expression and overall survival in high-grade serous ovarian cancer (HGSOC) patients and achieved AUC values of 0.879 and 0.834, respectively. A.D Coles et al. (<xref ref-type="bibr" rid="ref15">15</xref>) performed a comparative analysis of the RF and LR models for retrospective detection of OC recurrences from chemotherapy data and revealed that RF achieved the highest F1 score among the two. Alexander Laios et al. (<xref ref-type="bibr" rid="ref16">16</xref>) performed ML-based risk prediction using the KNN and LDA techniques and QDA algorithms that achieved predictive accuracies of 0.80, 0.90, and 0.92, respectively.</p>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Related studies based on ensemble models</title>
<p>Subsequent studies used the following ensemble models for the prediction of OC intrinsically: Random Forest (RF), XGBoost, gradient-boosted trees (GBT), and Bagging. Gong et al. (<xref ref-type="bibr" rid="ref17">17</xref>) used nine supervised ML classifiers including RF and GBT, out of which RF performed with high stability and an accuracy of 0.60 including gut microbiota for chemoresistance to OC prediction using the RF model with an AUC value of 0.909. Ahamad et al. (<xref ref-type="bibr" rid="ref18">18</xref>) utilized a combination of XGBoost, GBT, and light gradient boosting machine (LGBM) to build classifier models, and the predictive analysis results revealed an accuracy of 91%. Hamidi et al. (<xref ref-type="bibr" rid="ref19">19</xref>) developed Boruta, a novel RF feature selection-based ML model, for identifying important biomarkers. Zhao et al. (<xref ref-type="bibr" rid="ref20">20</xref>) used LASSO regression, ridge regression, XGBoost, RF, and AdaBoost to build a LASSO Cox regression model. By performing decision curve analysis, the resulting nomogram indicated that the combined model for predicting 1-year and 3-year survival probabilities provided an optimal net benefit compared to using a single indicator. Wadapurkar et al. (<xref ref-type="bibr" rid="ref21">21</xref>) applied next-generation sequencing analysis to cancer driver genes and OC prediction using the XGBoost classification model and yielded an accuracy of 0.946. Abuzinadah et al. (<xref ref-type="bibr" rid="ref22">22</xref>) used the RF model, GBM, and also ensemble GBM&#x202F;+&#x202F;XGB models. GBM was reported to be the best with an accuracy of 87.14%, a recall of 87.53%, and a precision of 87.58%. Zeng et al. (<xref ref-type="bibr" rid="ref23">23</xref>) used a new multi-omics model including all the features and showed the best prediction performance with an AUC value of 0.911. Piedimonte et al. (<xref ref-type="bibr" rid="ref24">24</xref>) used a Random Forest model that was trained to predict the dichotomous outcome of optimal cytoreduction and obtained an AUC of 99.8%. Cheng et al. (<xref ref-type="bibr" rid="ref25">25</xref>) used a gradient-boosting decision tree algorithm in which the AUC value of the signature gene pair was 0.9658, whereas the AUC value of the individual signature gene-based prediction was only 0.6823. Seri Jeong et al. (<xref ref-type="bibr" rid="ref26">26</xref>) utilized the RF model for differentiating benign and malignant OC using combined cancer markers and achieved receiver operating characteristic (ROC) AUC values of 0.707, 0.680, 0.643, 0.657, and 0.624 for the biomarkers ROMA, HE4, CA125, LD, and NLR, respectively. Maria et al. (<xref ref-type="bibr" rid="ref27">27</xref>) used an XAI-based ensemble model for the accurate classification of ovarian cancer and achieved an accuracy of 83.2%. Additionally, the model proved to be highly effective in handling imbalanced datasets. Annarita et al. (<xref ref-type="bibr" rid="ref28">28</xref>) proposed a waterfall-based classification model for the classification of clinical data and ultrasound indicators and achieved an accuracy of 86.36%.</p>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>Related studies based on deep learning models</title>
<p>Wu et al. (<xref ref-type="bibr" rid="ref29">29</xref>) used an attention-based network by using a pre-trained model ResNet50 on ImageNet to obtain a C-index value of 0.5789 (0.5096&#x2013;0.6053) and a p-value of 0.00845, while the risk score indicated good prediction ability in the homologous recombination deficiency (HRD+) subgroup. Feng et al. (<xref ref-type="bibr" rid="ref30">30</xref>) used a backpropagation neural network model by using a pre-trained ResNet50 which outperformed the U-net-based network architecture by obtaining 89.11% average sensitivity and 96.37% specificity. Suganya et al. (<xref ref-type="bibr" rid="ref31">31</xref>) used MALDI imaging for the EOC classification. Out of several ML and DL models used in the research, neural network and convolutional neural network (CNN) were most suitable for EOC classification with an accuracy of 83% for NN, 85% for CNN, and a sensitivity of 69&#x2013;100% as well as specificity of 90&#x2013;99% for both the models, respectively. Rahul Mishra et al. (<xref ref-type="bibr" rid="ref32">32</xref>) used the Convolutional Neural Networks with Gray Wolf Optimization (CNNGWO) model and reported an improved diagnostic accuracy of OC to 98%, while Talib et al. (<xref ref-type="bibr" rid="ref33">33</xref>) trained a simple CNN-based DL model on 1798 images and achieved an accuracy of 81% and an AUC value of 0.89. Kasture et al. (<xref ref-type="bibr" rid="ref34">34</xref>) used outclassed traditional CNN algorithms, deep CNN, and pre-trained AlexNet models, improving the accuracy from 70 to 89.93%. Reilly et al. (<xref ref-type="bibr" rid="ref35">35</xref>) proposed an MIA3G model (deep feed-forward neural network) for OC risk assessment and achieved a sensitivity of 89.8%, specificity of 84.02%, positive predictive value (PPV) of 22.455, and negative predictive value (NPV) of 99.38%. David Joon Ho et al. (<xref ref-type="bibr" rid="ref36">36</xref>) worked with a deep multi-magnification network model (DMMN) for OC segmentation and obtained 86% recall and 84% precision. Ziyambe et al. (<xref ref-type="bibr" rid="ref37">37</xref>) built a CNN architecture model using the Xception network and obtained 94.43% accuracy, 95.02% sensitivity, and 93.16% specificity, making it a potential tool in assisting physicians for predicting OC. Dingdu Hu et al. (<xref ref-type="bibr" rid="ref38">38</xref>) conducted automatic segmentations of T2-weighted MRI using DL algorithms (CNN, U-Net, DeepLabv3, U-Net++, and PSPNet) along with transformers (TransUnet and Swin-Unet), and reported that U-Net++ performed the best with precision and recall rates of 83.8 and 88.2% in the internal test set and 82.5 and 72.5% in the external test set, respectively. Ravishankar et al. (<xref ref-type="bibr" rid="ref39">39</xref>) proposed a fuzzy CNN-based classifier with an initial accuracy of 98.37%, which later achieved 100% accuracy occasionally with an increase in several epochs. Wang et al. (<xref ref-type="bibr" rid="ref40">40</xref>) used two improved DL-based methods for predicting bevacizumab efficacy in EOC patients based on histopathological images and revealed that the proposed model using the MSH2 protein expression performed excellently with 100% accuracy, sensitivity, and specificity, respectively. Ju Young et al. (<xref ref-type="bibr" rid="ref41">41</xref>) developed a U-Net-based DL algorithm for the characterization of cancer-associated thrombosis of OC and achieved outstanding results with an AUC value of 0.99 along with high sensitivity and specificity. Jeya Sundari et al. (<xref ref-type="bibr" rid="ref42">42</xref>) developed an intelligent black widow optimization on image enhancement with a DL-based ovarian tumor diagnosis model, which achieved a maximum contrast of 0.97, a contrast-to-noise ratio (CNR) of 92.74%, a weighted peak signal-to-noise ratio (WPSNR) of 20.43, and a homogeneity of 0.94. Rui Yin et al. (<xref ref-type="bibr" rid="ref43">43</xref>) worked with a multitask deep learning model for high-grade serous ovarian carcinoma (HGSOC) prediction using CT images of OC, which resulted in AUC values of 0.87 (0.83&#x2013;0.89), 0.88 (0.86&#x2013;0.91), 0.86 (0.82&#x2013;0.89), and 0.79 (0.75&#x2013;0.82) in the training sets, validation sets, prospective sets, and external sets, respectively. Yang et al. (<xref ref-type="bibr" rid="ref44">44</xref>) developed the Ovarian Cancer Digital Pathology Index (OCDPI), a graph-based DL model, for prediction and treatment response in OC patients using histopathological images and demonstrated prognostic ability for overall survival prediction in the prostate, lung, colorectal, and ovarian cancer (PLCO; HR, 1.916; 95% CI, 1.380&#x2013;2.660; log-rank test, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001) and Harbin Medical University Cancer hospital (HMUCH) (HR, 2.796; 95% CI, 1.404&#x2013;5.568; log-rank test, <italic>p</italic>&#x202F;=&#x202F;0.0022) cohorts. It was observed that the performance of OCDPI was good in patients with low-grade tumors or fresh frozen slides. Laios et al. (<xref ref-type="bibr" rid="ref45">45</xref>) developed a predictive model using XGBoost and the deep neural network for the prediction of epithelial ovarian cancer and used a SHAP technique to discuss how intraoperative decisions during EOC cytoreduction can integrate human factors along with factual knowledge to optimize the selected trade-off in surgical effort. Huang et al. (<xref ref-type="bibr" rid="ref4">4</xref>) emphasized classification tasks and used a game-theoretic framework using Shap values for model development, evaluation, and result visualization. The case study findings highlighted the effectiveness, reliability, and benefits of the ML pipeline in comparison to traditional statistical methods.</p>
</sec>
</sec>
<sec sec-type="methods" id="sec6">
<label>3</label>
<title>Methodology</title>
<sec id="sec7">
<label>3.1</label>
<title>Data collection and description</title>
<p>The dataset comprises 1,500 cases of ovarian tumors, classified into borderline, malignant, and benign types. The authors meticulously transferred the data manually from medical records to an Excel file, following ethical medical practices under the vigilance of medical practitioners at the Kasturba Medical College, Karnataka, India. Each row represents a unique patient and the columns capture various clinical features pertinent to the presence or absence of ovarian cancer. The raw data from the records were structured into the following 28 columns in the Excel file: clinical parameters such as age, parity, family history, oral contraception, age at menarche, menopause, abdominal pain, menstrual abnormalities, dysmenorrhea, abdominal distention, loss of appetite, edema, pallor, P/V tenderness, and abdominal mass; ultrasound or CT parameters: bilateral, ascites, solid area, papillary projection, thick septa, vascularity, contrast enhancement, lymph nodes, and deposits; and tumor markers such as CA125, CA19-9, and CEA and the corresponding category or class label. There are three classes: 0 &#x2013; benign, 1 &#x2013; border, and 2 &#x2013; malignant.</p>
<p>Some of the feature descriptions are as follows:</p>
<list list-type="roman-lower">
<list-item>
<p>Parity: number of full-term pregnancies.</p>
</list-item>
<list-item>
<p>Oral contraception: history of using oral contraceptive pills.</p>
</list-item>
<list-item>
<p>Dysmenorrhea: painful menstruation.</p>
</list-item>
<list-item>
<p>Abdominal distention: swelling or enlargement of the abdomen.</p>
</list-item>
<list-item>
<p>Edema: swelling caused by excess fluid trapped in body tissues.</p>
</list-item>
<list-item>
<p>Nodes: presence of swollen lymph nodes.</p>
</list-item>
<list-item>
<p>Pallor: unusual paleness of the skin.</p>
</list-item>
<list-item>
<p>P/V tenderness: pain during a pelvic examination.</p>
</list-item>
<list-item>
<p>Abdominal mass: detectable lump in the abdomen.</p>
</list-item>
<list-item>
<p>Bilateral: side of the body where the tumor is located.</p>
</list-item>
<list-item>
<p>Ascites: accumulation of fluid in the peritoneal cavity.</p>
</list-item>
<list-item>
<p>Solid area: presence of solid regions within the tumor.</p>
</list-item>
<list-item>
<p>Papillary projection: finger-like extensions seen in tumors.</p>
</list-item>
<list-item>
<p>Thick septa: thickened partitions within the tumor.</p>
</list-item>
<list-item>
<p>Vascularity: blood supply to the tumor.</p>
</list-item>
<list-item>
<p>Contrast enhancement: increase in the visibility of structures using contrast medium in imaging.</p>
</list-item>
<list-item>
<p>Deposits: tumor spread or deposits in the abdomen.</p>
</list-item>
<list-item>
<p>CA125: blood marker often elevated in ovarian cancer.</p>
</list-item>
<list-item>
<p>CA19-9: tumor marker used in the diagnosis of ovarian cancer.</p>
</list-item>
<list-item>
<p>CEA: carcinoembryonic antigen, another tumor marker.</p>
</list-item>
</list>
<p>This dataset served as a vital tool for studying the clinical features and diagnostic markers of ovarian tumors, aiding in advancing medical research and treatment strategies.</p>
</sec>
<sec id="sec8">
<label>3.2</label>
<title>Architectural framework</title>
<p><xref ref-type="fig" rid="fig1">Figure 1</xref> depicts the process flow followed in the current study. The dataset consists of 1,500 samples, with benign, malignant, and border cases comprising 500 samples each. The data are stored in a CSV file. This is considered the Input Data file that holds the raw data. Further steps were performed on this data. Preprocessing steps such as feature scaling were performed using a standard min&#x2013;max scaler. A standard min&#x2013;max scaler was applied to rescale all features to a uniform range, typically [0, 1]. This ensured that features with different scales did not disproportionately influence the model. Feature Transformation using exponential transformation was used to address skewed data distributions; the model was enhanced to capture patterns in the data along with labeling instances into correct categories, and missing values were imputed using mean imputation, replacing missing entries with the mean of the respective features. This approach preserved the overall distribution of the data and maintained dataset integrity without introducing significant bias. The processed data were then provided as input for several machine learning classifier models. The ML classifiers K-nearest neighbors (KNNs), logistic regression (LR), support vector machine (SVM), decision trees (DTs), and Naive Bayes (NB) were used in the current study. For the ensemble classifiers, models such as Random Forest (RF), XGBoost, CatBoost, and GradientBoost were applied. All classifiers were evaluated based on their performance using evaluation metrics such as accuracy, AUC, and ROC. The best-performing models among all the models were chosen. Random Forest, XGBoost, and SVM with accuracies of 98.13, 97, and 97.06%, respectively, were the top three best-performing models and were used to design a three-stage ensemble architecture. These scores represent the performance of the models on the test data. Explainable AI methods were applied to the test data results and passed onto the final ensemble model. LIME and SHAP were used to visualize and explain the predictions made by the model, as they are widely used for explanation across domains.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Process diagram.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g001.tif"/>
</fig>
</sec>
</sec>
<sec sec-type="results" id="sec9">
<label>4</label>
<title>Results and discussion</title>
<sec id="sec10">
<label>4.1</label>
<title>Statistical analysis using machine learning models</title>
<p>The results have been computed by considering the complete dataset of clinical parameters, tumor markers, and ultrasound parameters. The accuracy rates of single classifiers and ensemble classifiers are depicted as bar charts in <xref ref-type="fig" rid="fig2">Figure 2</xref>, respectively. It can be observed that SVM and Random Forest exhibit the highest accuracy among the single classifiers and the ensemble models, respectively. Only SVM, Random Forest (RF), and XGBoost have an accuracy of 97% or above, with 97.06, 98.13, and 97%, respectively. Therefore, these models were used to build the three-stage ensemble classifier. In <xref ref-type="fig" rid="fig3">Figure 3</xref>, the accuracy rates of all the models used in the current study have been plotted along with the proposed three-stage ensemble classifier.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Accuracy of the single classifier (left) and ensemble classifiers (right).</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g002.tif"/>
</fig>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Comparison of single, ensemble, and three-stage ensemble models.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g003.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig4">Figure 4</xref> shows the ROC curves for SVM, Random Forest, and the proposed three-stage ensemble models. Their AUC&#x2013;ROC values are 0.92, 0.95, and 0.96, respectively. It is evident from the figure that the proposed model exhibits the best predictive behavior. In machine learning, feature importance is a technique that assigns a score to input features based on their importance for predicting the target variables. Since the proposed model is a combination of Random Forest, SVM, and XGBoost, it naturally provides the feature importance scores as they are the tree-based model. The RF model gives the feature importance based on the entropy score for the classification tree and aggregates the scores of each feature across all the trees in the forest. Each tree&#x2019;s feature importance is calculated based on the reduction in impurity (entropy value). Similar to RF, GradientBoost calculates the feature importance score using the total reduction in the log loss. The feature importance predicted on the overall dataset and only on the clinical parameters is shown in <xref ref-type="fig" rid="fig5">Figures 5A</xref>,<xref ref-type="fig" rid="fig5">B</xref>, respectively.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>ROC curves for the SVM (left), RF (middle), and three-stage ensemble architecture (right).</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g004.tif"/>
</fig>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Feature importance <bold>(A)</bold> on the overall dataset and <bold>(B)</bold> only on clinical parameters.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g005.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig5">Figure 5A</xref> shows the feature importance of each feature contributing to ovarian malignancy. It is evident that CEA, loss of appetite, CA19-9, CA125, abdominal distention, family history, age of menarche, and age are the highest contributors and more prominently influence the decision of which category the particular case will fall into. It can be seen from <xref ref-type="fig" rid="fig5">Figure 5B</xref> that the top features family history, loss of appetite, age, abdominal distention, oral contraception, and age of menarche are the highest contributors among clinical parameters and more prominently influence the decision of which category the particular case will fall into. In contrast, the features such as dysmenorrhea, menopause, menstrual abnormalities, and abdominal pain are the lowest contributors clinically, however, but are of significance when accounted for as an isolated event.</p>
</sec>
<sec id="sec11">
<label>4.2</label>
<title>Analysis using explainable AI</title>
<sec id="sec12">
<label>4.2.1</label>
<title>Lime</title>
<p>LIME helps in understanding how specific features of patients&#x2019; data influence the model prediction. As shown in <xref ref-type="fig" rid="fig6">Figure 6</xref>, the model predicted that the patients do not have a serious condition (benign). Out of the six features considered, &#x2018;Abdominal Distention&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Family History&#x2019;, and &#x2018;CA19-9&#x2032; indicate that there is no serious condition (negative contribution). On the other hand, &#x2018;CEA&#x2019; and &#x2018;CA125&#x2019; indicate slight risk (positive contribution) at the very initial stages. The actual values for these features are as follows: &#x2018;Abdominal Distention&#x2019; at 0.0, &#x2018;CEA&#x2019; at 1.59, &#x2018;CA19-9&#x2032; at 13.26, &#x2018;Loss of Appetite&#x2019; at 0.0, &#x2018;Family History&#x2019; at 0.0, and &#x2018;CA125&#x2019; at 25.94. Because &#x2018;Loss of Appetite&#x2019;, &#x2018;Abdominal Distention&#x2019;, &#x2018;Family History&#x2019;, and &#x2018;CA19-9&#x2032; fall within safe ranges, they contribute to the model prediction as benign. Although CEA and CA125 values are positively contributing, the values are very negligible. CA125 value is 25.94, which is outside the safe range, but the other factors are negatively contributing, and thus the case can be considered benign.</p>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>LIME instance for Class 0 (benign).</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g006.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig7">Figure 7</xref> shows a borderline case where the model predicted a possible risk (Class 1) of ovarian malignancy. Here, three features&#x2014;&#x2018;Abdominal Distention&#x2019;, &#x2018;Loss of Appetite&#x2019;, and &#x2018;Family History&#x2019;&#x2014;indicate a higher risk (positive contribution), while &#x2018;CEA&#x2019;, &#x2018;CA19-9&#x2019;, and &#x2018;CA125&#x2019; suggest a lower risk (negative contribution) as they are not de-ranged values. The values of these features are as follows: &#x2018;Abdominal Distention&#x2019; at 1.0, &#x2018;Loss of Appetite&#x2019; at 1.0, &#x2018;CA19-9&#x2019; at 10.0, &#x2018;CEA&#x2019; at 1.0, and &#x2018;CA125&#x2019; at 16.0. Because &#x2018;Abdominal Distention&#x2019;, &#x2018;Loss of Appetite&#x2019;, and &#x2018;Family History&#x2019; are within the concerning ranges, they contribute to the model prediction of possible risk, with the border values of CEA being yes and CA125 toward the border value.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>LIME instance for Class 1 (border case).</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g007.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig8">Figure 8</xref> shows that the model predicted a serious ovarian condition (malignant). The features&#x2014;&#x2018;Loss of Appetite&#x2019;, &#x2018;CEA&#x2019;, &#x2018;CA125&#x2019;, &#x2018;Abdominal Distention&#x2019;, and &#x2018;Age&#x2019;&#x2014;indicate a higher risk (positive contribution). The values of these features are as follows: &#x2018;Loss of Appetite&#x2019; at 1.00 indicates that the person is having a loss of appetite issue, &#x2018;Abdominal Distention&#x2019; at 1.0, &#x2018;CEA&#x2019; at 23.64 is also on the higher side, &#x2018;CA125&#x2019; at 143.38, &#x2018;Age&#x2019; at 71, and &#x2018;CA19-9&#x2019; at 1.89. Since all the features except CA19-9 contribute positively, the model prediction is revealed as malignant. The important factor to be considered in this case is the age which is 71; the values of other factors such as CA125, loss of appetite, and abdominal distention are high, leading to malignant prediction.</p>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>LIME instance for Class 2 (malignant case).</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g008.tif"/>
</fig>
</sec>
<sec id="sec13">
<label>4.2.2</label>
<title>SHAP</title>
<p>SHAP values help in understanding the importance of each feature in the model prediction for ovarian lesions. As shown in <xref ref-type="fig" rid="fig9">Figure 9</xref>, &#x2018;CEA&#x2019;, &#x2018;CA19-9&#x2019;, &#x2018;CA125&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Family History&#x2019;, and &#x2018;Abdominal Distention&#x2019; are the most important features in determining the model predictions; LIME also used the same features for predicting the cases. This indicates that these six features have the most significant impact on the model predicting a benign, borderline, or malignant condition.</p>
<fig position="float" id="fig9">
<label>Figure 9</label>
<caption>
<p>Importance of SHAP feature.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g009.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig10">Figure 10</xref> shows how the feature &#x2018;CEA (Carcinoembryonic Antigen)&#x2019; influences the prediction of a specific patient (observation number 9614). With a &#x2018;CEA&#x2019; value of 105.63, it contributes positively by 0.2168 points, indicating a higher risk.</p>
<fig position="float" id="fig10">
<label>Figure 10</label>
<caption>
<p>Contribution of CEA for observation number 614.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g010.tif"/>
</fig>
<p>In <xref ref-type="fig" rid="fig11">Figure 11</xref>, for another patient (observation number 892), the &#x2018;CEA&#x2019; value is 1, contributing negatively by &#x2212;0.1211 points, suggesting a lower risk. These examples show that the impact of each feature can be positive or negative depending on its value, helping doctors understand how different aspects of patient data influence the model prediction.</p>
<fig position="float" id="fig11">
<label>Figure 11</label>
<caption>
<p>Contribution of CEA for observation number 1129.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g011.tif"/>
</fig>
<sec id="sec14">
<label>4.2.2.1</label>
<title>Tree SHAP</title>
<p>Interpreting a stacked bar plot of Tree SHAP values with mean SHAP values involves an understanding of how each feature contributes to the model&#x2019;s predictions on average. This helps in identifying which features are the most influential and how they collectively affect the model output relative to the average prediction.</p>
<p><xref ref-type="fig" rid="fig12">Figure 12A</xref> shows that &#x2018;CEA&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;CA125&#x2019;, &#x2018;CA19-9&#x2019;, &#x2018;Abdominal Distention&#x2019;, &#x2018;Family History&#x2019;, &#x2018;Menarche&#x2019;, &#x2018;Age&#x2019;, and &#x2018;Menstrual Abnormalities&#x2019; are the top nine features attributing to model prediction in cases of ovarian lesions according to the Tree SHAP Explainer. The bar against &#x2018;CEA&#x2019; shows that this feature is highly influential in deciding &#x2018;Class 2&#x2019; when compared to &#x2018;Class 1&#x2019; and &#x2018;Class 0&#x2019;. The bar against &#x2018;Loss of Appetite&#x2019; shows that this feature is highly influential in deciding &#x2018;Class 0&#x2019; and &#x2018;Class 1&#x2019; when compared to &#x2018;Class 2&#x2019;. The bar against &#x2018;CA125&#x2019; shows that this feature is highly important for predicting &#x2018;Class 2&#x2019; when compared to the other two classes. The bar against &#x2018;Abdominal Distention&#x2019; shows that this feature is important for instances that belong to &#x2018;Class 0&#x2019;. <xref ref-type="fig" rid="fig12">Figure 12B</xref> shows that &#x2018;Oral Contraception&#x2019;, &#x2018;Family History&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Abdominal Distention&#x2019;, and &#x2018;Menarche&#x2019; are the top five features contributing clinically to the model prediction according to the Tree SHAP explainer. The bar against &#x2018;Family History&#x2019; shows that this feature is highly influential in deciding Classes 1 and 2 when compared to Class 0. The bar against &#x2018;Age&#x2019; shows that this feature is highly influential in deciding Class 2 and Class 1 when compared with Class 0. The bar against &#x2018;Oral Contraception&#x2019; is highly influential in deciding Class 2 when compared to Classes 1 and 0.</p>
<fig position="float" id="fig12">
<label>Figure 12</label>
<caption>
<p>Feature importance <bold>(A)</bold> on the overall dataset and <bold>(B)</bold> on clinical parameters.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g012.tif"/>
</fig>
<p>From the clinician&#x2019;s point of view, these observations indicate the weightage of these parameters as the conditions for prediction and suggest positive and negative predictive roles for each of the contributors. CEA has the highest positive predictive value for malignancy. CA19-9 has the highest predictive value for borderline tumors, whereas CA125, being a non-specific marker, is not as specifically predictive as any of the categories. For example, if SHAP predicts a tumor as malignant based on CEA and the same mass as benign based on CA125, it is more likely a malignant mass because CEA as an independent factor has the highest reliable predictivity. In the prediction model that included clinical and tumor marker parameters, abdominal distension had more of a negative predictive value than other parameters, indicating that the absence of abdominal distension rules out malignancy and the mass is likely to be benign. However, when only clinical parameters were considered in the prediction model and tumor markers were not included, abdominal distension had the predictive capacity to indicate malignancy by its presence.</p>
</sec>
<sec id="sec15">
<label>4.2.2.2</label>
<title>Sampling SHAP</title>
<p>Interpreting a stacked bar plot of Sampling SHAP values with mean SHAP values involves an understanding of how each feature contributes to the model predictions, considering the variability introduced by the sampling process. This approach helps in identifying influential features and their average impact on model predictions across different samples.</p>
<p><xref ref-type="fig" rid="fig13">Figure 13A</xref> shows that &#x2018;CEA&#x2019;, &#x2018;CA125&#x2019;, &#x2018;CA19-9&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Abdominal Distention&#x2019;, &#x2018;Family History&#x2019;, &#x2018;Menarche&#x2019;, &#x2018;Age&#x2019;, and &#x2018;Menstrual Abnormalities&#x2019; are the top nine features contributing to the model prediction of ovarian lesions according to the Sampling SHAP Explainer. The bar against &#x2018;CEA&#x2019; shows that this feature is highly influential in deciding &#x2018;Class 2&#x2019; when compared to &#x2018;Class 0&#x2019; and &#x2018;Class 1&#x2019;. The bar against &#x2018;CA125&#x2019; is highly influential in deciding Class 2 when compared to &#x2018;Class 0&#x2019; and &#x2018;Class 1&#x2019;. The bar against &#x2018;CA19-9&#x2019; shows that this feature is highly important for predicting &#x2018;Class 1&#x2019; when compared to the other two classes. The bar against &#x2018;Abdominal Distention&#x2019; shows that this feature is important for instances that belong to &#x2018;Class 0&#x2019;. Sampling SHAP was used when exact computation (as done in Tree SHAP) was impractical due to computational constraints. It involved approximating SHAP values through sampling subsets of features or instances. <xref ref-type="fig" rid="fig13">Figure 13B</xref> shows that &#x2018;Oral Contraception, &#x2018;Family History&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Abdominal Distention&#x2019;, and &#x2018;Menarche&#x2019; are the top five features contributing clinically to model prediction as per the Sampling SHAP Explainer. The bar against the features oral contraception, age, and family history shows that these features are highly influential in deciding &#x2018;Class 1 and 2&#x2032; when compared to &#x2018;Class 0&#x2032;. The bar against the features &#x2018;Loss of Appetite&#x2019; and &#x2018;Abdominal Distention&#x2019; is highly influential in deciding &#x2018;Class 0 and 1&#x2032; when compared to &#x2018;Class 2&#x2032;.</p>
<fig position="float" id="fig13">
<label>Figure 13</label>
<caption>
<p>Feature importance <bold>(A)</bold> on the overall dataset and <bold>(B)</bold> on clinical parameters.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g013.tif"/>
</fig>
</sec>
<sec id="sec16">
<label>4.2.2.3</label>
<title>Kernel SHAP</title>
<p>Interpreting a stacked bar plot of Kernel SHAP values with mean SHAP values involves an understanding of how each feature contributes to the model prediction, considering the kernel-based approach used for SHAP value computation. This approach helps in identifying influential features and their average impact on model predictions across different instances or data subsets.</p>
<p><xref ref-type="fig" rid="fig14">Figure 14A</xref> shows that &#x2018;CEA&#x2019;, &#x2018;CA125&#x2019;, &#x2018;Abdominal Distention&#x2019;, &#x2018;CA19-9&#x2019;, &#x2018;Loss of Appetite&#x2019;, &#x2018;Menarche&#x2019;, &#x2018;Family History&#x2019;, &#x2018;Age&#x2019;, and &#x2018;Dysmenorrhoea&#x2019; are the top nine features contributing to the model prediction of ovarian lesions according to the Kernel SHAP Explainer. The bar against &#x2018;CEA&#x2019; is highly influential in deciding between &#x2018;Class 2&#x2019; and &#x2018;Class 1&#x2019; when compared to &#x2018;Class 0&#x2019;. The bar against &#x2018;CA19-9&#x2019; shows that this is highly important for predicting &#x2018;Class 1&#x2019; when compared to the other two classes. The bar against &#x2018;CA125&#x2019; is highly influential in deciding &#x2018;Class 2&#x2019; when compared to &#x2018;Class 0 and Class 2&#x2019;. The bar against &#x2018;Abdominal Distention&#x2019; shows that this feature is important for instances that belong to &#x2018;Class 0&#x2019;. Sampling SHAP was used when exact computation (as conducted in Tree SHAP) was impractical due to computational constraints. This involved approximating SHAP values through sampling subsets of features or instances. <xref ref-type="fig" rid="fig14">Figure 14B</xref> shows the feature importance of clinical data as per the Kernel SHAP. The bars against the features &#x2018;Oral Contraception&#x2019; and &#x2018;Family History&#x2019; are highly influential in deciding &#x2018;Class 1 and 2&#x2019; when compared with &#x2018;Class 0&#x2019;.</p>
<fig position="float" id="fig14">
<label>Figure 14</label>
<caption>
<p>Feature importance <bold>(A)</bold> on the overall dataset and <bold>(B)</bold> on clinical parameters.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g014.tif"/>
</fig>
<p>From <xref ref-type="fig" rid="fig12">Figures 12&#x2013;14</xref>, it can be concluded that &#x2018;CEA&#x2019; and &#x2018;CA19-9&#x2019; are the top two features highly contributing to model prediction. Furthermore, it can be noted that Kernel SHAP performs the best in comparison with Tree and Sampling SHAP, showing consistency in explaining model prediction across all three classes.</p>
<p><xref ref-type="fig" rid="fig15">Figure 15</xref> shows the global view of the RF distribution, which corroborates the feature importance deduction that abdominal distention is one of the major contributing factors in classifying ovarian lesions. A mix of cases can be found in some instances that might not be a contributing factor but a few individual instances with relatively higher values could be studied further to understand the factors, if any, that set them apart. In conclusion, features with positive SHAP values for a particular instance can be indicative and helpful in stratifying a case. SHAP values will help in identifying the features that contribute to the model predictions. In <xref ref-type="fig" rid="fig16">Figure 16</xref>, the top nine features contributed to predicting particular training instances as benign cases. From the figure, it can be observed that while getting trained, the features have contributed to learning and the correct classification. It is also evident from the values that the features such as loss of appetite, abdominal distention, and family history are absent, the tumor markers such as CEA, CA19-19, and CA125 values are very much within the safe range, and the value of the most important feature Age is 18. This shows that SHAP has predicted the case as benign with a probability of 99.1%.</p>
<fig position="float" id="fig15">
<label>Figure 15</label>
<caption>
<p>SHAP values based on the proposed model.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g015.tif"/>
</fig>
<fig position="float" id="fig16">
<label>Figure 16</label>
<caption>
<p>Training sample of a benign case with a probability of 99.1%.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g016.tif"/>
</fig>
<p><xref ref-type="fig" rid="fig17">Figure 17</xref> shows that the particular instance is classified as not non-benign, with a probability of 0.1. It is evident from the value on the left-hand side that the values of tumor biomarkers CEA, CA125, and CA19-9 are very high. The values of the clinical parameters such as family history, loss of appetite, and abdominal distention are all positive, with age being 75. From <xref ref-type="fig" rid="fig16">Figures 16</xref>, <xref ref-type="fig" rid="fig17">17</xref>, it can be confirmed that the values of these features vary with huge differences.</p>
<fig position="float" id="fig17">
<label>Figure 17</label>
<caption>
<p>Training sample of a benign case with a probability of 0.1, which is a malignant case.</p>
</caption>
<graphic xlink:href="fpubh-13-1479095-g017.tif"/>
</fig>
<p><italic>p</italic>-value is the measure that determines the significance of the obtained results in the hypothesis tests. It represents the probability of obtaining the test results at least as extreme as the observed results. Cohen&#x2019;s <italic>d</italic> value is often used to evaluate the practical significance of a model&#x2019;s performance and the impact of the intervention. It can be observed from <xref ref-type="table" rid="tab1">Table 1</xref> that for CEA, loss of appetite, CA19-9, CA125, and abdominal distention, the p-value is extremely small (less than 0.5), indicating that the result is highly statistically significant. Cohen&#x2019;s <italic>d</italic> values of these variables suggest a large effect size, which means the impact of these variables on diagnosis is very significant.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p><italic>p</italic>-value and effect size (Cohen&#x2019;s <italic>d</italic>) for variables (influencing diagnosis and classification of ovarian mass lesions) ranked top 5 by SHAP.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Ranking</th>
<th align="left" valign="top">Variables</th>
<th align="center" valign="top"><italic>p</italic>-value</th>
<th align="center" valign="top">Cohen&#x2019;s <italic>d</italic> value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">CEA</td>
<td align="center" valign="top">3.7974 &#x00D7; 10<sup>&#x2212;10</sup></td>
<td align="center" valign="top">&#x2212;1.4000</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Loss of appetite</td>
<td align="center" valign="top">1.2299 &#x00D7; 10<sup>&#x2212;8</sup></td>
<td align="center" valign="top">3.2861</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">CA19-9</td>
<td align="center" valign="top">2.1076 &#x00D7; 10<sup>&#x2212;11</sup></td>
<td align="center" valign="top">&#x2212;1.3615</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">CA125</td>
<td align="center" valign="top">0.1115</td>
<td align="center" valign="top">0.0883</td>
</tr>
<tr>
<td align="left" valign="top">5</td>
<td align="left" valign="top">Abdominal distention</td>
<td align="center" valign="top">1.2299 &#x00D7; 10<sup>&#x2212;9</sup></td>
<td align="center" valign="top">3.2861</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From <xref ref-type="table" rid="tab2">Table 2</xref>, it can be observed that <italic>p</italic>-values are significantly larger and Cohen&#x2019;s d values are greater than 0.2, which indicates a much smaller effect size when compared with the most important variables. The negative Cohen&#x2019;s d values suggest a negative effect.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p><italic>p</italic>-value and effect size (Cohen&#x2019;s <italic>d</italic>) for variables (influencing diagnosis and classification of ovarian mass lesions) ranked bottom 5 by SHAP.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Ranking</th>
<th align="left" valign="top">Variables</th>
<th align="center" valign="top"><italic>p</italic>-value</th>
<th align="center" valign="top">Cohen&#x2019;s <italic>d</italic> value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">16</td>
<td align="left" valign="top">Menstrual abnormalities</td>
<td align="center" valign="top">4.1922</td>
<td align="center" valign="top">&#x2212;0.4490</td>
</tr>
<tr>
<td align="left" valign="top">17</td>
<td align="left" valign="top">Dysmenorrhea</td>
<td align="center" valign="top">4.1922</td>
<td align="center" valign="top">&#x2212;0.4490</td>
</tr>
<tr>
<td align="left" valign="top">18</td>
<td align="left" valign="top">Menopause</td>
<td align="center" valign="top">2.4639</td>
<td align="center" valign="top">0.4083</td>
</tr>
<tr>
<td align="left" valign="top">19</td>
<td align="left" valign="top">Parity (GPLA)</td>
<td align="center" valign="top">5.3066</td>
<td align="center" valign="top">&#x2212;0.3210</td>
</tr>
<tr>
<td align="left" valign="top">20</td>
<td align="left" valign="top">Abdominal pain</td>
<td align="center" valign="top">1.0</td>
<td align="center" valign="top">0.0</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="tab3">Table 3</xref> shows the difference in ranking based on SHAP values and p-values, which implies that while some features have more predictive capabilities, it might not be statistically as evident. Abdominal distention ranks higher in SHAP values than in p-values, indicating that it might be more important for prediction than its statistical significance might indicate.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Comparison of rankings generated by SHAP values and <italic>p</italic>-values.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Rank</th>
<th align="left" valign="top">Feature</th>
<th align="center" valign="top">SHAP value rank</th>
<th align="center" valign="top"><italic>p</italic>-value rank</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">CEA</td>
<td align="center" valign="top">1</td>
<td align="center" valign="top">4</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Loss of appetite</td>
<td align="center" valign="top">2</td>
<td align="center" valign="top">1</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">CA19-9</td>
<td align="center" valign="top">3</td>
<td align="center" valign="top">5</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">CA125</td>
<td align="center" valign="top">4</td>
<td align="center" valign="top">3</td>
</tr>
<tr>
<td align="left" valign="top">5</td>
<td align="left" valign="top">Abdominal distention</td>
<td align="center" valign="top">5</td>
<td align="center" valign="top">2</td>
</tr>
<tr>
<td align="left" valign="top">6</td>
<td align="left" valign="top">Family history</td>
<td align="center" valign="top">6</td>
<td align="center" valign="top">6</td>
</tr>
<tr>
<td align="left" valign="top">7</td>
<td align="left" valign="top">Menarche</td>
<td align="center" valign="top">7</td>
<td align="center" valign="top">8</td>
</tr>
<tr>
<td align="left" valign="top">8</td>
<td align="left" valign="top">Age</td>
<td align="center" valign="top">8</td>
<td align="center" valign="top">7</td>
</tr>
<tr>
<td align="left" valign="top">9</td>
<td align="left" valign="top">Menstrual abnormalities</td>
<td align="center" valign="top">9</td>
<td align="center" valign="top">9</td>
</tr>
<tr>
<td align="left" valign="top">10</td>
<td align="left" valign="top">Dysmenorrhea</td>
<td align="center" valign="top">10</td>
<td align="center" valign="top">10</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
</sec>
<sec sec-type="discussion" id="sec17">
<label>5</label>
<title>Discussion</title>
<p>In this section, we compare the results of the proposed model with those of the existing studies mentioned in Section 2.</p>
<p>After a rigorous survey of the available literature, from <xref ref-type="table" rid="tab4">Table 4</xref>, we found that there are not many studies in the literature that incorporate a combination of clinical, radiological, and serum-based parameters for ovarian cancer diagnosis and prognosis, which is an advancement in the state of the art beyond the existing literature&#x2014;this mapping of multi-modal data onto a common subspace benefit from the complementary information in each modality. Therefore, we attempted to fill this gap in the literature, specifically patient data available in clinical data and radiological and tumor markers. Learning these data proceeds by maximizing intra-class similarities and inter-class differences to obtain richer representations. The richer representations obtained from these data trained ensemble-based classifiers to diagnose cancer. The proposed approach also aligns with medical experts using information from different sources to corroborate and conclude their diagnosis.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Comparison with the existing relatable studies.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Literature</th>
<th align="left" valign="top">Dataset used</th>
<th align="left" valign="top">Type/modality of data</th>
<th align="center" valign="top">Number of samples considered</th>
<th align="center" valign="top">Number of features</th>
<th align="center" valign="top">No. of patients</th>
<th align="left" valign="top">Parameters used in the dataset</th>
<th align="left" valign="top">Methods used</th>
<th align="left" valign="top">Performance reported</th>
<th align="left" valign="top">XAI methods</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Nopour et al. (<xref ref-type="bibr" rid="ref5">5</xref>)</td>
<td align="left" valign="top">In-house</td>
<td align="left" valign="top">Clinical data</td>
<td align="center" valign="top">1,473</td>
<td align="center" valign="top">26</td>
<td align="center" valign="top">1,473</td>
<td align="left" valign="top">Age, Family history, family cancer syndrome, etc.</td>
<td align="left" valign="top">XGBoost, SVM, KNNs, ANN, RF</td>
<td align="left" valign="top">XGBoost: AUC&#x2013;ROC&#x202F;=&#x202F;0.93 (95% CI: [0.91&#x2013;0.95])</td>
<td align="left" valign="top">Not used</td>
</tr>
<tr>
<td align="left" valign="top">Lavanya J M et al. (<xref ref-type="bibr" rid="ref13">13</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Clinical data</td>
<td align="center" valign="top">349</td>
<td align="center" valign="top">49</td>
<td align="center" valign="top">349</td>
<td align="left" valign="top">22 general chemical tests, 19 blood routine tests, and 6 tumor markers, including age and menopause information</td>
<td align="left" valign="top">KNNs, SVM, Decision Trees, Max Voting, Boosting, Bagging, Stacking</td>
<td align="left" valign="top">Base accuracy&#x202F;=&#x202F;85%, Stacking: 89%</td>
<td align="left" valign="top">SHAP</td>
</tr>
<tr>
<td align="left" valign="top">Abuzinadah et al. (<xref ref-type="bibr" rid="ref22">22</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Clinical data</td>
<td align="center" valign="top">349</td>
<td align="center" valign="top">49</td>
<td align="center" valign="top">349</td>
<td align="left" valign="top">22 general chemical tests, 19 blood routine tests, and 6 tumor markers, including age and menopause information</td>
<td align="left" valign="top">Stacked Ensemble Model (Bagging + Boosting)</td>
<td align="left" valign="top">Accuracy&#x202F;=&#x202F;96.87%</td>
<td align="left" valign="top">SHAP</td>
</tr>
<tr>
<td align="left" valign="top">Schilling et al. (<xref ref-type="bibr" rid="ref10">10</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Gene expression</td>
<td align="center" valign="top">4,490</td>
<td align="center" valign="top">9,954</td>
<td align="center" valign="top">585</td>
<td align="left" valign="top">mRNA sequencing data</td>
<td align="left" valign="top">K-means, Naive Bayes, Logistic Regression, SVM</td>
<td align="left" valign="top">F1 score&#x202F;=&#x202F;0.89</td>
<td align="left" valign="top">SHAP</td>
</tr>
<tr>
<td align="left" valign="top">Chao et al. (<xref ref-type="bibr" rid="ref9">9</xref>)</td>
<td align="left" valign="top">In-house data</td>
<td align="left" valign="top">demographic and clinicopathologic features</td>
<td align="center" valign="top">6,809</td>
<td align="center" valign="top">67</td>
<td align="center" valign="top">6,809</td>
<td align="left" valign="top">Age, dysmenorrhea, Family history, CA125, Leiomyoma, etc.</td>
<td align="left" valign="top">Gradient-Boosting Decision Tree, Logistic Regression</td>
<td align="left" valign="top">LR: AUC&#x202F;=&#x202F;0.891, Sensitivity&#x202F;=&#x202F;88.9%, Specificity&#x202F;=&#x202F;76.7%</td>
<td align="left" valign="top">Not used</td>
</tr>
<tr>
<td align="left" valign="top">Ahamad et al. (<xref ref-type="bibr" rid="ref18">18</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Clinical data</td>
<td align="center" valign="top">349</td>
<td align="center" valign="top">49</td>
<td align="center" valign="top">349</td>
<td align="left" valign="top">22 general chemical tests, 19 blood routine tests, and 6 tumor markers, including age and menopause information</td>
<td align="left" valign="top">RF, SVM, DTs, XGBoost, LR, GBM, LGBM</td>
<td align="left" valign="top">Accuracy&#x202F;=&#x202F;91%</td>
<td align="left" valign="top">Not used</td>
</tr>
<tr>
<td align="left" valign="top">Chen et al. (<xref ref-type="bibr" rid="ref25">25</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Gene expression</td>
<td align="center" valign="top">4,490</td>
<td align="center" valign="top">9,954</td>
<td align="center" valign="top">585</td>
<td align="left" valign="top">mRNA sequencing data</td>
<td align="left" valign="top">Gradient Boosting Decision Tree</td>
<td align="left" valign="top">Accuracy&#x202F;=&#x202F;96.5%</td>
<td align="left" valign="top">Not used</td>
</tr>
<tr>
<td align="left" valign="top">Zeng et al. (<xref ref-type="bibr" rid="ref23">23</xref>)</td>
<td align="left" valign="top">Open access</td>
<td align="left" valign="top">Gene expression</td>
<td align="center" valign="top">4,490</td>
<td align="center" valign="top">9,954</td>
<td align="center" valign="top">585</td>
<td align="left" valign="top">mRNA sequencing data</td>
<td align="left" valign="top">Machine learning models (not specified explicitly)</td>
<td align="left" valign="top">AUC: 0.952 (BRCA1), 0.911 (multi-omics model)</td>
<td align="left" valign="top">Risk Score Interpretation for Survival Prediction</td>
</tr>
<tr>
<td align="left" valign="top">Proposed work</td>
<td align="left" valign="top">In-house</td>
<td align="left" valign="top">Clinical data, ultrasound parameters, and Tumor markers</td>
<td align="center" valign="top">1,500</td>
<td align="center" valign="top">20</td>
<td align="center" valign="top">1,500</td>
<td align="left" valign="top">CEA, CA19-9, CA125, Loss of Appetite, Abdominal Distention, Family History, Age of Menarche, Age, etc.</td>
<td align="left" valign="top">Three-stage ensemble model (SVM, RF, XGBoost)</td>
<td align="left" valign="top">Accuracy&#x202F;=&#x202F;98.66%</td>
<td align="left" valign="top">LIME and SHAP</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Hence, the innovative component and knowledge addition in this study is a unique combination of specific problem definition, usage of more relevant and near-match parameters that clinicians follow in their day-to-day practice, and translation of this cognition into an AI system, resulting in improved performance with reliability, consistency, and accuracy.</p>
</sec>
<sec id="sec18">
<label>6</label>
<title>Conclusion and future studies</title>
<p>This study demonstrates the development and application of an ensemble classifier along with explainability to accurately predict benign, borderline, and malignant ovarian cancers using clinical, ultrasound, and tumor marker data. It has been proved experimentally that the proposed three-stage ensemble architecture exhibited higher performance than the existing models in the literature, with an accuracy of 98.66%.</p>
<p>As a future direction, it is envisioned that if the subcategories of malignant ovarian cancer and histopathological images are considered in the classification process, performance may be affected, leaving lacunae for further development in the future.</p>
<p>The developed classification paradigm in this study has the potential to bring translational changes in the ovarian cancer management protocols in clinics, in particular, which can help in lowering the cost of treatment, lowering morbidity, and reducing the risk of mortality specifically for women in resource-constrained and economically challenged scenarios.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec19">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="ethics-statement" id="sec20">
<title>Ethics statement</title>
<p>The studies involving humans were approved by this study was carried out in collaboration with Kasturba Medical College, Manipal. The work was been approved by the ethical committee of the hospital. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants' legal guardians/next of kin in accordance with the national legislation and institutional requirements.</p>
</sec>
<sec sec-type="author-contributions" id="sec21">
<title>Author contributions</title>
<p>AK: Conceptualization, Data curation, Formal analysis, Methodology, Validation, Visualization, Writing &#x2013; original draft. VD: Supervision, Validation, Visualization, Writing &#x2013; review &#x0026; editing. SG: Resources, Supervision, Validation, Visualization, Writing &#x2013; review &#x0026; editing. TI: Validation, Visualization, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec sec-type="funding-information" id="sec22">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research and/or publication of this article.</p>
</sec>
<sec sec-type="COI-statement" id="sec23">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="sec24">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><citation citation-type="other"><person-group person-group-type="author"><collab id="coll1">American Cancer Society</collab></person-group>. (<year>2023</year>). Cancer Facts and Statistics. Available at: <ext-link xlink:href="https://www.cancer.org/research/cancer-facts-statistics/" ext-link-type="uri">https://www.cancer.org/research/cancer-facts-statistics/</ext-link></citation></ref>
<ref id="ref2"><label>2.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Siegel</surname> <given-names>RL</given-names></name> <name><surname>Miller</surname> <given-names>KD</given-names></name> <name><surname>Wagle</surname> <given-names>NS</given-names></name> <name><surname>Jemal</surname> <given-names>A</given-names></name></person-group>. <article-title>Cancer statistics, 2023</article-title>. <source>CA Cancer J Clin</source>. (<year>2023</year>) <volume>73</volume>:<fpage>17</fpage>&#x2013;<lpage>48</lpage>. doi: <pub-id pub-id-type="doi">10.3322/caac.21763</pub-id></citation></ref>
<ref id="ref3"><label>3.</label><citation citation-type="other"><person-group person-group-type="author"><collab id="coll2">Cancer Research Institute</collab></person-group>. (<year>2023</year>). Ovarian Cancer statistics and research. Available at: <ext-link xlink:href="https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/ovarian-cancer" ext-link-type="uri">https://www.cancerresearchuk.org/health-professional/cancer-statistics/statistics-by-cancer-type/ovarian-cancer</ext-link></citation></ref>
<ref id="ref4"><label>4.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>W</given-names></name> <name><surname>Suominen</surname> <given-names>H</given-names></name> <name><surname>Liu</surname> <given-names>T</given-names></name> <name><surname>Rice</surname> <given-names>G</given-names></name> <name><surname>Salomon</surname> <given-names>C</given-names></name> <name><surname>Barnard</surname> <given-names>AS</given-names></name></person-group>. <article-title>Explainable discovery of disease biomarkers: the case of ovarian cancer to illustrate the best practice in machine learning and Shapley analysis</article-title>. <source>J Biomed Inform</source>. (<year>2023</year>) <volume>141</volume>:<fpage>104365</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jbi.2023.104365</pub-id>, PMID: <pub-id pub-id-type="pmid">37062419</pub-id></citation></ref>
<ref id="ref5"><label>5.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Nopour</surname> <given-names>R</given-names></name></person-group>. <article-title>Screening ovarian cancer by using risk factors: machine learning assists</article-title>. <source>Biomed Eng Online</source>. (<year>2024</year>) <volume>23</volume>:<fpage>18</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12938-024-01219-x</pub-id>, PMID: <pub-id pub-id-type="pmid">38347611</pub-id></citation></ref>
<ref id="ref6"><label>6.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>J</given-names></name> <name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Ma</surname> <given-names>J</given-names></name> <name><surname>Wang</surname> <given-names>F</given-names></name> <name><surname>Cui</surname> <given-names>S</given-names></name> <name><surname>Ye</surname> <given-names>Z</given-names></name></person-group>. <article-title>Computed tomography-based radionics machine learning classifiers to differentiate type I and type II epithelial ovarian cancers</article-title>. <source>Eur Radiol</source>. (<year>2023</year>) <volume>33</volume>:<fpage>5193</fpage>&#x2013;<lpage>204</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00330-022-09318-w</pub-id>, PMID: <pub-id pub-id-type="pmid">36515713</pub-id></citation></ref>
<ref id="ref7"><label>7.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kori</surname> <given-names>M</given-names></name> <name><surname>Demirtas</surname> <given-names>TY</given-names></name> <name><surname>Comertpay</surname> <given-names>B</given-names></name> <name><surname>Arga</surname> <given-names>KY</given-names></name> <name><surname>Sinha</surname> <given-names>R</given-names></name> <name><surname>Gov</surname> <given-names>E</given-names></name></person-group>. <article-title>A 19-gene signature of serous ovarian Cancer identified by machine learning and systems biology: prospects for diagnostics and personalized medicine</article-title>. <source>OMICS</source>. (<year>2024</year>) <volume>28</volume>:<fpage>90</fpage>&#x2013;<lpage>101</lpage>. doi: <pub-id pub-id-type="doi">10.1089/omi.2023.0273</pub-id>, PMID: <pub-id pub-id-type="pmid">38320250</pub-id></citation></ref>
<ref id="ref8"><label>8.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Klein</surname> <given-names>O</given-names></name> <name><surname>Kanter</surname> <given-names>F</given-names></name> <name><surname>Kulbe</surname> <given-names>H</given-names></name> <name><surname>Jank</surname> <given-names>P</given-names></name> <name><surname>Denkert</surname> <given-names>C</given-names></name> <name><surname>Nebrich</surname> <given-names>G</given-names></name> <etal/></person-group>. <article-title>MALDI-imaging for classification of epithelial ovarian cancer histotypes from a tissue microarray using machine learning methods</article-title>. <source>Proteomics Clin Appl</source>. (<year>2019</year>) <volume>13</volume>:<fpage>e1700181</fpage>. doi: <pub-id pub-id-type="doi">10.1002/prca.201700181</pub-id>, PMID: <pub-id pub-id-type="pmid">30471200</pub-id></citation></ref>
<ref id="ref9"><label>9.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chao</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>S</given-names></name> <name><surname>Lang</surname> <given-names>J</given-names></name> <name><surname>Leng</surname> <given-names>J</given-names></name> <name><surname>Fan</surname> <given-names>Q</given-names></name></person-group>. <article-title>The application of risk models based on machine learning to predict endometriosis-associated ovarian cancer in patients with endometriosis</article-title>. <source>Acta Obstet Gynecol Scand</source>. (<year>2022</year>) <volume>101</volume>:<fpage>1440</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1111/aogs.14462</pub-id>, PMID: <pub-id pub-id-type="pmid">36210724</pub-id></citation></ref>
<ref id="ref10"><label>10.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Schilling</surname> <given-names>V</given-names></name> <name><surname>Beyerlein</surname> <given-names>P</given-names></name> <name><surname>Chien</surname> <given-names>J</given-names></name></person-group>. <article-title>A bioinformatics analysis of ovarian Cancer data using machine learning</article-title>. <source>Algorithms</source>. (<year>2023</year>) <volume>16</volume>:<fpage>330</fpage>. doi: <pub-id pub-id-type="doi">10.3390/a16070330</pub-id></citation></ref>
<ref id="ref11"><label>11.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hong</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Lin</surname> <given-names>D</given-names></name> <name><surname>Peng</surname> <given-names>J</given-names></name> <name><surname>Yuan</surname> <given-names>Q</given-names></name> <name><surname>Zeng</surname> <given-names>Y</given-names></name> <etal/></person-group>. <article-title>Development of a radiomic&#x2013;clinical nomogram for prediction of survival in patients with serous ovarian cancer</article-title>. <source>Clin Radiol</source>. (<year>2022</year>) <volume>77</volume>:<fpage>352</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.crad.2022.01.038</pub-id>, PMID: <pub-id pub-id-type="pmid">35264303</pub-id></citation></ref>
<ref id="ref12"><label>12.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Bahado-Singh</surname> <given-names>RO</given-names></name> <name><surname>Ibrahim</surname> <given-names>A</given-names></name> <name><surname>Al-Wahab</surname> <given-names>Z</given-names></name> <name><surname>Aydas</surname> <given-names>B</given-names></name> <name><surname>Radhakrishna</surname> <given-names>U</given-names></name> <name><surname>Yilmaz</surname> <given-names>A</given-names></name> <etal/></person-group>. <article-title>Precision gynecologic oncology: circulating cell-free DNA epigenomic analysis, artificial intelligence and the accurate detection of ovarian cancer</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>:<fpage>18625</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-022-23149-1</pub-id>, PMID: <pub-id pub-id-type="pmid">36329159</pub-id></citation></ref>
<ref id="ref13"><label>13.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>JM</surname> <given-names>SL</given-names></name> <name><surname>Subbulakshmi</surname> <given-names>P</given-names></name></person-group>. <article-title>Innovative approach towards early prediction of ovarian cancer: machine learning-enabled XAI techniques</article-title>. <source>Heliyon</source>. (<year>2024</year>) <volume>10</volume>:<fpage>e29197</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e29197</pub-id></citation></ref>
<ref id="ref14"><label>14.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhan</surname> <given-names>F</given-names></name> <name><surname>He</surname> <given-names>L</given-names></name> <name><surname>Yu</surname> <given-names>Y</given-names></name> <name><surname>Chen</surname> <given-names>Q</given-names></name> <name><surname>Guo</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name></person-group>. <article-title>A multimodal radiomic machine learning approach to predict the LCK expression and clinical prognosis in high-grade serous ovarian cancer</article-title>. <source>Sci Rep</source>. (<year>2023</year>) <volume>13</volume>:<fpage>16397</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-023-43543-7</pub-id>, PMID: <pub-id pub-id-type="pmid">37773310</pub-id></citation></ref>
<ref id="ref15"><label>15.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Coles</surname> <given-names>AD</given-names></name> <name><surname>McInerney</surname> <given-names>CD</given-names></name> <name><surname>Zucker</surname> <given-names>K</given-names></name> <name><surname>Cheeseman</surname> <given-names>S</given-names></name> <name><surname>Johnson</surname> <given-names>OA</given-names></name> <name><surname>Hall</surname> <given-names>G</given-names></name></person-group>. <article-title>Evaluation of machine learning methods for the retrospective detection of ovarian cancer recurrences from chemotherapy data</article-title>. <source>ESMO Real World Data Digit Oncol</source>. (<year>2024</year>) <volume>4</volume>:<fpage>100038</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.esmorw.2024.100038</pub-id></citation></ref>
<ref id="ref16"><label>16.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Laios</surname> <given-names>A</given-names></name> <name><surname>De Oliveira Silva</surname> <given-names>RV</given-names></name> <name><surname>Dantas De Freitas</surname> <given-names>DL</given-names></name> <name><surname>Tan</surname> <given-names>YS</given-names></name> <name><surname>Saalmink</surname> <given-names>G</given-names></name> <name><surname>Zubayraeva</surname> <given-names>A</given-names></name> <etal/></person-group>. <article-title>Machine learning-based risk prediction of critical care unit admission for advanced stage high grade serous ovarian cancer patients undergoing cytoreductive surgery: the Leeds-Natal score</article-title>. <source>J Clin Med</source>. (<year>2021</year>) <volume>11</volume>:<fpage>87</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jcm11010087</pub-id>, PMID: <pub-id pub-id-type="pmid">35011828</pub-id></citation></ref>
<ref id="ref17"><label>17.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Gong</surname> <given-names>TT</given-names></name> <name><surname>He</surname> <given-names>XH</given-names></name> <name><surname>Gao</surname> <given-names>S</given-names></name> <name><surname>Wu</surname> <given-names>QJ</given-names></name></person-group>. <article-title>Application of machine learning in the prediction of chemotherapy-resistant ovarian cancer based on gut microbiota</article-title>. <source>J Cancer</source>. (<year>2021</year>) <volume>12</volume>:<fpage>2877</fpage>&#x2013;<lpage>85</lpage>. doi: <pub-id pub-id-type="doi">10.7150/jca.46621</pub-id>, PMID: <pub-id pub-id-type="pmid">33854588</pub-id></citation></ref>
<ref id="ref18"><label>18.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahamad</surname> <given-names>MM</given-names></name> <name><surname>Aktar</surname> <given-names>S</given-names></name> <name><surname>Uddin</surname> <given-names>MJ</given-names></name> <name><surname>Rahman</surname> <given-names>T</given-names></name> <name><surname>Alyami</surname> <given-names>SA</given-names></name> <name><surname>Al-Ashhab</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Early-stage detection of ovarian cancer based on clinical data using machine learning approaches</article-title>. <source>J Pers Med</source>. (<year>2022</year>) <volume>12</volume>:<fpage>1211</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jpm12081211</pub-id>, PMID: <pub-id pub-id-type="pmid">35893305</pub-id></citation></ref>
<ref id="ref19"><label>19.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hamidi</surname> <given-names>F</given-names></name> <name><surname>Gilani</surname> <given-names>N</given-names></name> <name><surname>Arabi Belaghi</surname> <given-names>R</given-names></name> <name><surname>Yaghoobi</surname> <given-names>H</given-names></name> <name><surname>Babaei</surname> <given-names>E</given-names></name> <name><surname>Sarbakhsh</surname> <given-names>P</given-names></name> <etal/></person-group>. <article-title>Identifying potential circulating miRNA biomarkers for the diagnosis and prediction of ovarian cancer using the machine-learning approach: application of Boruta</article-title>. <source>Front Digit Health</source>. (<year>2023</year>) <volume>5</volume>:<fpage>1187578</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fdgth.2023.1187578</pub-id>, PMID: <pub-id pub-id-type="pmid">37621964</pub-id></citation></ref>
<ref id="ref20"><label>20.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zhao</surname> <given-names>C</given-names></name> <name><surname>Xiong</surname> <given-names>K</given-names></name> <name><surname>Zhao</surname> <given-names>F</given-names></name> <name><surname>Adam</surname> <given-names>A</given-names></name> <name><surname>Li</surname> <given-names>X</given-names></name></person-group>. <article-title>Glycosylation-related genes predict the prognosis and immune fraction of ovarian cancer patients based on weighted gene coexpression network analysis (WGCNA) and machine learning</article-title>. <source>Oxid Med Cell Longev</source>. (<year>2022</year>) <volume>2022</volume>:<fpage>3665617</fpage>. doi: <pub-id pub-id-type="doi">10.1155/2022/3665617</pub-id></citation></ref>
<ref id="ref21"><label>21.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wadapurkar</surname> <given-names>R</given-names></name> <name><surname>Bapat</surname> <given-names>S</given-names></name> <name><surname>Mahajan</surname> <given-names>R</given-names></name> <name><surname>Vyas</surname> <given-names>R</given-names></name></person-group>. <article-title>Machine learning approaches for prediction of ovarian cancer driver genes from mutational and network analysis</article-title>. <source>Data Technol Appl</source>. (<year>2024</year>) <volume>58</volume>:<fpage>62</fpage>&#x2013;<lpage>80</lpage>. doi: <pub-id pub-id-type="doi">10.1108/DTA-03-2022-0096</pub-id></citation></ref>
<ref id="ref22"><label>22.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Abuzinadah</surname> <given-names>N</given-names></name> <name><surname>Kumar Posa</surname> <given-names>S</given-names></name> <name><surname>Alarfaj</surname> <given-names>AA</given-names></name> <name><surname>Alabdulqader</surname> <given-names>EA</given-names></name> <name><surname>Umer</surname> <given-names>M</given-names></name> <name><surname>Kim</surname> <given-names>TH</given-names></name> <etal/></person-group>. <article-title>Improved prediction of ovarian Cancer using ensemble classifier and Shaply explainable AI</article-title>. <source>Cancers</source>. (<year>2023</year>) <volume>15</volume>:<fpage>5793</fpage>. doi: <pub-id pub-id-type="doi">10.3390/cancers15245793</pub-id>, PMID: <pub-id pub-id-type="pmid">38136346</pub-id></citation></ref>
<ref id="ref23"><label>23.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Zeng</surname> <given-names>H</given-names></name> <name><surname>Chen</surname> <given-names>L</given-names></name> <name><surname>Zhang</surname> <given-names>M</given-names></name> <name><surname>Luo</surname> <given-names>Y</given-names></name> <name><surname>Ma</surname> <given-names>X</given-names></name></person-group>. <article-title>Integration of histopathological images and multi-dimensional omics analyses predicts molecular features and prognosis in high-grade serous ovarian cancer</article-title>. <source>Gynecol Oncol</source>. (<year>2021</year>) <volume>163</volume>:<fpage>171</fpage>&#x2013;<lpage>80</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ygyno.2021.07.015</pub-id>, PMID: <pub-id pub-id-type="pmid">34275655</pub-id></citation></ref>
<ref id="ref24"><label>24.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Piedimonte</surname> <given-names>S</given-names></name> <name><surname>Erdman</surname> <given-names>L</given-names></name> <name><surname>So</surname> <given-names>D</given-names></name> <name><surname>Bernardini</surname> <given-names>MQ</given-names></name> <name><surname>Ferguson</surname> <given-names>SE</given-names></name> <name><surname>Laframboise</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Using a machine learning algorithm to predict the outcome of primary cytoreductive surgery in advanced ovarian cancer</article-title>. <source>J Surg Oncol</source>. (<year>2023</year>) <volume>127</volume>:<fpage>465</fpage>&#x2013;<lpage>72</lpage>. doi: <pub-id pub-id-type="doi">10.1002/jso.27137</pub-id>, PMID: <pub-id pub-id-type="pmid">36350138</pub-id></citation></ref>
<ref id="ref25"><label>25.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>K</given-names></name> <name><surname>Xu</surname> <given-names>H</given-names></name> <name><surname>Lei</surname> <given-names>Y</given-names></name> <name><surname>Lio</surname> <given-names>P</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Guo</surname> <given-names>H</given-names></name> <etal/></person-group>. <article-title>Integration and interplay of machine learning and bioinformatics approach to identify genetic interaction related to ovarian cancer chemoresistance</article-title>. <source>Brief Bioinform</source>. (<year>2021</year>) <volume>22</volume>:<fpage>bbab100</fpage>. doi: <pub-id pub-id-type="doi">10.1093/bib/bbab100</pub-id>, PMID: <pub-id pub-id-type="pmid">33971668</pub-id></citation></ref>
<ref id="ref26"><label>26.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jeong</surname> <given-names>S</given-names></name> <name><surname>Son</surname> <given-names>DS</given-names></name> <name><surname>Cho</surname> <given-names>M</given-names></name> <name><surname>Lee</surname> <given-names>N</given-names></name> <name><surname>Song</surname> <given-names>W</given-names></name> <name><surname>Shin</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Evaluation of combined cancer markers with lactate dehydrogenase and application of machine learning algorithms for differentiating benign disease from malignant ovarian cancer</article-title>. <source>Cancer Control</source>. (<year>2021</year>) <volume>28</volume>:<fpage>10732748211033401</fpage>. doi: <pub-id pub-id-type="doi">10.1177/10732748211033401</pub-id>, PMID: <pub-id pub-id-type="pmid">34923833</pub-id></citation></ref>
<ref id="ref27"><label>27.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Comes</surname> <given-names>MC</given-names></name> <name><surname>Arezzo</surname> <given-names>F</given-names></name> <name><surname>Cormio</surname> <given-names>G</given-names></name> <name><surname>Bove</surname> <given-names>S</given-names></name> <name><surname>Calabrese</surname> <given-names>A</given-names></name> <name><surname>Fanizzi</surname> <given-names>A</given-names></name> <etal/></person-group>. <article-title>An explainable machine learning ensemble model to predict the risk of ovarian cancer in BRCA-mutated patients undergoing risk-reducing salpingo-oophorectomy</article-title>. <source>Front Oncol</source>. (<year>2023</year>) <volume>13</volume>:<fpage>1181792</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fonc.2023.1181792</pub-id>, PMID: <pub-id pub-id-type="pmid">37519818</pub-id></citation></ref>
<ref id="ref28"><label>28.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Fanizzi</surname> <given-names>A</given-names></name> <name><surname>Arezzo</surname> <given-names>F</given-names></name> <name><surname>Cormio</surname> <given-names>G</given-names></name> <name><surname>Comes</surname> <given-names>MC</given-names></name> <name><surname>Cazzato</surname> <given-names>G</given-names></name> <name><surname>Boldrini</surname> <given-names>L</given-names></name> <etal/></person-group>. <article-title>An explainable machine learning model to solid adnexal masses diagnosis based on clinical data and qualitative ultrasound indicators</article-title>. <source>Cancer Med</source>. (<year>2024</year>) <volume>13</volume>:<fpage>e7425</fpage>. doi: <pub-id pub-id-type="doi">10.1002/cam4.7425</pub-id>, PMID: <pub-id pub-id-type="pmid">38923847</pub-id></citation></ref>
<ref id="ref29"><label>29.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>M</given-names></name> <name><surname>Zhu</surname> <given-names>C</given-names></name> <name><surname>Yang</surname> <given-names>J</given-names></name> <name><surname>Cheng</surname> <given-names>S</given-names></name> <name><surname>Yang</surname> <given-names>X</given-names></name> <name><surname>Gu</surname> <given-names>S</given-names></name> <etal/></person-group>. <article-title>Exploring prognostic indicators in the pathological images of ovarian cancer based on a deep survival network</article-title>. <source>Front Genet</source>. (<year>2023</year>) <volume>13</volume>:<fpage>1069673</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fgene.2022.1069673</pub-id>, PMID: <pub-id pub-id-type="pmid">36685892</pub-id></citation></ref>
<ref id="ref30"><label>30.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname> <given-names>Y</given-names></name></person-group>. <article-title>An integrated machine learning-based model for joint diagnosis of ovarian cancer with multiple test indicators</article-title>. <source>J Ovarian Res</source>. (<year>2024</year>) <volume>17</volume>:<fpage>45</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13048-024-01365-9</pub-id>, PMID: <pub-id pub-id-type="pmid">38378582</pub-id></citation></ref>
<ref id="ref31"><label>31.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Suganya</surname> <given-names>Y</given-names></name> <name><surname>Ganesan</surname> <given-names>S</given-names></name> <name><surname>Valarmathi</surname> <given-names>P</given-names></name> <name><surname>Suresh</surname> <given-names>T</given-names></name></person-group>. <article-title>A diagnosis of ovarian cyst using deep learning neural network with XGBoost algorithm</article-title>. <source>Int J Inf Technol</source>. (<year>2023</year>) <volume>15</volume>:<fpage>3499</fpage>&#x2013;<lpage>506</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s41870-023-01401-9</pub-id></citation></ref>
<ref id="ref32"><label>32.</label><citation citation-type="confproc"><person-group person-group-type="author"><name><surname>Mishra</surname> <given-names>R.</given-names></name> <name><surname>Kumar</surname> <given-names>M.</given-names></name> <name><surname>Brindha</surname> <given-names>S.</given-names></name> <name><surname>Sharma</surname> <given-names>A. K.</given-names></name> <name><surname>Raja</surname> <given-names>C.</given-names></name> <name><surname>Moharekar</surname> <given-names>T. T.</given-names></name></person-group> (<year>2023</year>) <article-title>An efficient deep learning model for intraoperative tissue classification in gynecological Cancer</article-title>. In <conf-name>2023 9th international conference on smart structures and systems (ICSSS)</conf-name> (pp. <fpage>1</fpage>&#x2013;<lpage>6</lpage>). <conf-loc>Chennai, India</conf-loc>: <publisher-name>IEEE</publisher-name>.</citation></ref>
<ref id="ref33"><label>33.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Talib</surname> <given-names>SM</given-names></name> <name><surname>Hasan</surname> <given-names>FK</given-names></name> <name><surname>Fakhir</surname> <given-names>AS</given-names></name> <name><surname>Abbas</surname> <given-names>AK</given-names></name> <name><surname>Karam</surname> <given-names>LS</given-names></name> <name><surname>Nour</surname> <given-names>AAM</given-names></name></person-group>. <article-title>Ovarian Cancer diagnosing by MRI: a preliminary study comparing deep learning and radiologist assessments</article-title>. <source>World Sci</source>. (<year>2023</year>) <volume>2</volume>:<fpage>117</fpage>&#x2013;<lpage>29</lpage>. Available at: <ext-link xlink:href="https://univerpubl.com/index.php/woscience/article/view/2397" ext-link-type="uri">https://univerpubl.com/index.php/woscience/article/view/2397</ext-link></citation></ref>
<ref id="ref34"><label>34.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Kasture</surname> <given-names>KR</given-names></name></person-group>. <article-title>A new deep learning method for automatic ovarian cancer prediction &#x0026; subtype classification</article-title>. <source>Turk J Comput Math Educ</source>. (<year>2021</year>) <volume>12</volume>:<fpage>1233</fpage>&#x2013;<lpage>42</lpage>. Available at: <ext-link xlink:href="https://www.proquest.com/docview/2622814176?sourcetype=Scholarly%20Journals" ext-link-type="uri">https://www.proquest.com/docview/2622814176?sourcetype=Scholarly%20Journals</ext-link></citation></ref>
<ref id="ref35"><label>35.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Reilly</surname> <given-names>G</given-names></name> <name><surname>Bullock</surname> <given-names>RG</given-names></name> <name><surname>Greenwood</surname> <given-names>J</given-names></name> <name><surname>Ure</surname> <given-names>DR</given-names></name> <name><surname>Stewart</surname> <given-names>E</given-names></name> <name><surname>Davidoff</surname> <given-names>P</given-names></name> <etal/></person-group>. <article-title>Analytical validation of a deep neural network algorithm for the detection of ovarian cancer</article-title>. <source>JCO Clin Cancer Inform</source>. (<year>2022</year>) <volume>6</volume>:<fpage>e2100192</fpage>. doi: <pub-id pub-id-type="doi">10.1200/CCI.21.00192</pub-id>, PMID: <pub-id pub-id-type="pmid">35671415</pub-id></citation></ref>
<ref id="ref36"><label>36.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ho</surname> <given-names>DJ</given-names></name> <name><surname>Chui</surname> <given-names>MH</given-names></name> <name><surname>Vanderbilt</surname> <given-names>CM</given-names></name> <name><surname>Jung</surname> <given-names>J</given-names></name> <name><surname>Robson</surname> <given-names>ME</given-names></name> <name><surname>Park</surname> <given-names>CS</given-names></name> <etal/></person-group>. <article-title>Deep interactive learning-based ovarian cancer segmentation of H&#x0026;E-stained whole slide images to study morphological patterns of BRCA mutation</article-title>. <source>J Pathol Inform</source>. (<year>2023</year>) <volume>14</volume>:<fpage>100160</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jpi.2022.100160</pub-id>, PMID: <pub-id pub-id-type="pmid">36536772</pub-id></citation></ref>
<ref id="ref37"><label>37.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ziyambe</surname> <given-names>B</given-names></name> <name><surname>Yahya</surname> <given-names>A</given-names></name> <name><surname>Mushiri</surname> <given-names>T</given-names></name> <name><surname>Tariq</surname> <given-names>MU</given-names></name> <name><surname>Abbas</surname> <given-names>Q</given-names></name> <name><surname>Babar</surname> <given-names>M</given-names></name> <etal/></person-group>. <article-title>A deep learning framework for the prediction and diagnosis of ovarian cancer in pre-and post-menopausal women</article-title>. <source>Diagnostics</source>. (<year>2023</year>) <volume>13</volume>:<fpage>1703</fpage>. doi: <pub-id pub-id-type="doi">10.3390/diagnostics13101703</pub-id>, PMID: <pub-id pub-id-type="pmid">37238188</pub-id></citation></ref>
<ref id="ref38"><label>38.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Hu</surname> <given-names>D</given-names></name> <name><surname>Jian</surname> <given-names>J</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Gao</surname> <given-names>X</given-names></name></person-group>. <article-title>Deep learning-based segmentation of epithelial ovarian cancer on T2-weighted magnetic resonance images</article-title>. <source>Quant Imaging Med Surg</source>. (<year>2023</year>) <volume>13</volume>:<fpage>1464</fpage>&#x2013;<lpage>77</lpage>. doi: <pub-id pub-id-type="doi">10.21037/qims-22-494</pub-id>, PMID: <pub-id pub-id-type="pmid">36915355</pub-id></citation></ref>
<ref id="ref39"><label>39.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ravishankar</surname> <given-names>TN</given-names></name> <name><surname>Jadhav</surname> <given-names>HM</given-names></name> <name><surname>Kumar</surname> <given-names>NS</given-names></name> <name><surname>Ambala</surname> <given-names>S</given-names></name></person-group>. <article-title>A deep learning approach for ovarian cysts detection and classification (OCD-FCNN) using fuzzy convolutional neural network</article-title>. <source>Meas Sens</source>. (<year>2023</year>) <volume>27</volume>:<fpage>100797</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.measen.2023.100797</pub-id></citation></ref>
<ref id="ref40"><label>40.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>CW</given-names></name> <name><surname>Lee</surname> <given-names>YC</given-names></name> <name><surname>Lin</surname> <given-names>YJ</given-names></name> <name><surname>Firdi</surname> <given-names>NP</given-names></name> <name><surname>Muzakky</surname> <given-names>H</given-names></name> <name><surname>Liu</surname> <given-names>TC</given-names></name> <etal/></person-group>. <article-title>Deep learning can predict bevacizumab therapeutic effect and microsatellite instability directly from histology in epithelial ovarian Cancer</article-title>. <source>Lab Investig</source>. (<year>2023</year>) <volume>103</volume>:<fpage>100247</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.labinv.2023.100247</pub-id>, PMID: <pub-id pub-id-type="pmid">37741509</pub-id></citation></ref>
<ref id="ref41"><label>41.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Ahn</surname> <given-names>JY</given-names></name> <name><surname>Carlos-Alcalde</surname> <given-names>W</given-names></name> <name><surname>Vasquez</surname> <given-names>M</given-names></name> <name><surname>Cho</surname> <given-names>MS</given-names></name> <name><surname>Afshar-Kharghan</surname> <given-names>V</given-names></name> <name><surname>Wong</surname> <given-names>ST</given-names></name></person-group>. <article-title>A deep learning-enabled digital pathology platform for characterization of cancer-associated thrombosis of ovarian cancer</article-title>. <source>Cancer Res</source>. (<year>2023</year>) <volume>83</volume>:<fpage>5406</fpage>&#x2013;<lpage>6</lpage>. doi: <pub-id pub-id-type="doi">10.1158/1538-7445.AM2023-5406</pub-id></citation></ref>
<ref id="ref42"><label>42.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Jeya Sundari</surname> <given-names>M</given-names></name> <name><surname>Brintha</surname> <given-names>NC</given-names></name></person-group>. <article-title>An intelligent black widow optimization on image enhancement with deep learning-based ovarian tumor diagnosis model</article-title>. <source>Comput Methods Biomech Biomed Eng Imaging Vis</source>. (<year>2023</year>) <volume>11</volume>:<fpage>598</fpage>&#x2013;<lpage>605</lpage>. doi: <pub-id pub-id-type="doi">10.1080/21681163.2022.2092036</pub-id>, PMID: <pub-id pub-id-type="pmid">39935898</pub-id></citation></ref>
<ref id="ref43"><label>43.</label><citation citation-type="other"><person-group person-group-type="author"><name><surname>Yin</surname> <given-names>R.</given-names></name> <name><surname>Guo</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Zhang</surname> <given-names>Q.</given-names></name> <name><surname>Dou</surname> <given-names>Z.</given-names></name> <name><surname>Wang</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2023</year>). Predicting NACT response and HGSOC from CT images in ovarian Cancer with multitask deep learning: a multicenter study. Available at: <ext-link xlink:href="https://ssrn.com/abstract=4359710" ext-link-type="uri">https://ssrn.com/abstract=4359710</ext-link>.</citation></ref>
<ref id="ref44"><label>44.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Z</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Zhuo</surname> <given-names>L</given-names></name> <name><surname>Sun</surname> <given-names>K</given-names></name> <name><surname>Meng</surname> <given-names>F</given-names></name> <name><surname>Zhou</surname> <given-names>M</given-names></name> <etal/></person-group>. <article-title>Prediction of prognosis and treatment response in ovarian cancer patients from histopathology images using graph deep learning: a multicenter retrospective study</article-title>. <source>Eur J Cancer</source>. (<year>2024</year>) <volume>199</volume>:<fpage>113532</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ejca.2024.113532</pub-id>, PMID: <pub-id pub-id-type="pmid">38241820</pub-id></citation></ref>
<ref id="ref45"><label>45.</label><citation citation-type="journal"><person-group person-group-type="author"><name><surname>Laios</surname> <given-names>A</given-names></name> <name><surname>Kalampokis</surname> <given-names>E</given-names></name> <name><surname>Johnson</surname> <given-names>R</given-names></name> <name><surname>Munot</surname> <given-names>S</given-names></name> <name><surname>Thangavelu</surname> <given-names>A</given-names></name> <name><surname>Hutson</surname> <given-names>R</given-names></name> <etal/></person-group>. <article-title>Factors predicting surgical effort using explainable artificial intelligence in advanced stage epithelial ovarian cancer</article-title>. <source>Cancers</source>. (<year>2022</year>) <volume>14</volume>:<fpage>3447</fpage>. doi: <pub-id pub-id-type="doi">10.3390/cancers14143447</pub-id>, PMID: <pub-id pub-id-type="pmid">35884506</pub-id></citation></ref>
</ref-list>
</back>
</article>