Next Article in Journal
Evaluation of Platelet Alloimmunization by Filtration Enzyme-Linked Immunosorbent Assay
Next Article in Special Issue
CRV-NET: Robust Intensity Recognition of Coronavirus in Lung Computerized Tomography Scan Images
Previous Article in Journal
Development and Implementation of an Internal Quality Control Sample to Standardize Oligomer-Based Diagnostics of Alzheimer’s Disease
Previous Article in Special Issue
An Ensembled Framework for Human Breast Cancer Survivability Prediction Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Framework for the Prediction and Diagnosis of Ovarian Cancer in Pre- and Post-Menopausal Women

1
Department of Electrical Engineering, Harare Polytechnic College, Causeway Harare P.O. Box CY407, Zimbabwe
2
Department of Electrical, Computer and Telecommunications Engineering, Botswana International University of Science and Technology, Palapye 10071, Botswana
3
Department of Industrial and Mechatronics Engineering, Faculty of Engineering & the Built Environment, University of Zimbabwe, Mt. Pleasant, 630 Churchill Avenue, Harare, Zimbabwe
4
Abu Dhabi University, Abu Dhabi 59911, United Arab Emirates
5
College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
6
Robotics and Internet of Things Laboratory, Prince Sultan University, Riyadh 12435, Saudi Arabia
7
EIAS Data Science Laboratory, Prince Sultan University, Riyadh 12435, Saudi Arabia
8
Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(10), 1703; https://doi.org/10.3390/diagnostics13101703
Submission received: 15 March 2023 / Revised: 17 April 2023 / Accepted: 25 April 2023 / Published: 11 May 2023
(This article belongs to the Special Issue Applications of Artificial Intelligence in Thoracic Imaging)

Abstract

:
Ovarian cancer ranks as the fifth leading cause of cancer-related mortality in women. Late-stage diagnosis (stages III and IV) is a major challenge due to the often vague and inconsistent initial symptoms. Current diagnostic methods, such as biomarkers, biopsy, and imaging tests, face limitations, including subjectivity, inter-observer variability, and extended testing times. This study proposes a novel convolutional neural network (CNN) algorithm for predicting and diagnosing ovarian cancer, addressing these limitations. In this paper, CNN was trained on a histopathological image dataset, divided into training and validation subsets and augmented before training. The model achieved a remarkable accuracy of 94%, with 95.12% of cancerous cases correctly identified and 93.02% of healthy cells accurately classified. The significance of this study lies in overcoming the challenges associated with the human expert examination, such as higher misclassification rates, inter-observer variability, and extended analysis times. This study presents a more accurate, efficient, and reliable approach to predicting and diagnosing ovarian cancer. Future research should explore recent advances in this field to enhance the effectiveness of the proposed method further.

1. Introduction

The yearly mortality for ovarian cancer is 151,900, making it the deadliest cancer globally [1]. According to Miller, it is women’s fifth highest cause of death. The most common kind of gynaecological carcinoma is ovarian cancer, which emanates from epithelial tissue, and 90% of the cases are due to this type. The five histologic carcinomas are Mucinous-Ovarian Cancer (M-OC), High-Grade Serous-Ovarian Cancer (H-GS-OC), Low-Grade Serous-Ovarian Cancer (L-GS-OC), Clear-Cell Ovarian Cancer (C-COC), and Endometrioid-Ovarian-Cancer (E-O-C), with poor prognosis at an advanced stage [2].
Early Identification boosts survival from 3% in Stage IV to about 90% in Stage I [3]. Van Haaften-Day et al. (2001) [4] found that Carbohydrate Antigen 125 (CA125) has been used for over four decades, but that its accuracy is not acceptable as it has not improved survivability. 50% of early-stage tumours, primarily type I ovarian cancers, and 92% of advanced-stage tumours, primarily type II ovarian cancers, have increased serum CA125 levels. Skates et al. [5] found that physiological variables influence normal CA125 serum concentrations and the presence of menopause impacts CA125 levels. Sopik et al. [6] found that benign illnesses also had increased CA125 levels, causing false positives. Only 20% of ovarian tumors express CA125. A screening’s anticipated sensitivity is around 80%. Moss et al. [7] confirmed that relying on the biomarker alone is misleading. Akinwunmi et al. [8] found that 1% of healthy people had increased serum levels, including 5% of patients with benign illnesses, e.g., endometriosis.
HE4, WFDC2, is overexpressed in Endometrioid ovarian cancer and less in epithelial tissues of the respiratory system and reproductive organs [9]. Yana-ranop et al. [10] found that the specificity of HE4 was 86% compared to that of CA125, and the AUC of it was 0.893 compared to that of CA125 0.865 [11]. HE4 levels vary in smokers (30%) and non-smokers (20%) [12]. Contraceptives impact HE4 levels. Ferraro et al. [13] found lower HE4 levels in oral contraceptive users (p = 0.008). Biopsy, imaging (US, CT, MRI, PET), and algorithms for learning (deep) Convolutional Neural networks (CNN) can predict and diagnose epithelial ovarian cancer (serous) accurately.
The challenge is that there is no effective screening method; hence, ovarian cancer is diagnosed when it has already advanced to Stage III or IV. Radiologists manually analyse and interpret medical images of a suspected cancer patient for cancer subtyping and staging. This results in the misclassification of the cancer subtypes, inter-observer variations, subjectivity, and time consumption. To address this, a deep CNN model was developed to predict and diagnose cancers.
Expert pathologists interpret Cellular morphology, defining OC categories and directing treatments [14]. Inter-observer variations cause inaccurate, suboptimal treatments, poor prognosis, and reduced life quality [15]. This shows the need for computational methods to accurately predict cancer class and diagnose cancer subtype.
Accurate prediction and diagnosis of the cell tumours are vital as they lead to proper prognosis and treatment, increasing survivability. Deep learning merits include:
  • Processing huge data and producing highly accurate predictions, reducing incorrect diagnoses.
  • Permitting early detection of ovarian carcinoma, increasing treatment success.
  • Permitting personalised treatment. Deep learning can predict how treatments affect different women, enabling personalised, efficient care.
Deep learning can improve patient outcomes and reduce mortality through early detection and personalised treatment. The algorithm could predict and diagnose the images in under 5 s with an F1-score of 0.94. The order of the paper is as follows: related work, the materials and methods used, followed by results and discussion, and, finally, the conclusion.

2. Related Work

Various deep-learning techniques have been studied to classify ovarian carcinoma based on the cell type. Personalized treatment plans for ovarian cancer patients depend on accurately identifying the type of ovarian cancer. In the past decade, various studies have aimed to enhance cancer screening outcomes in the preclinical stage, focusing on utilizing histopathological images and biomarkers, such as CA-125 and HE-4. However, biomarker-based detection can be slow and prone to missing detection [16,17]. CA-125 has limited accuracy and specificity in detecting early-stage ovarian cancer, and imaging methods such as CT, US, and MRI are commonly used for locating and detecting features of masses. However, the interpretation of images by expert radiographers can be time-consuming and prone to inter-observer variability [18].
Recent studies have focused on using machine learning algorithms for early prediction and diagnosis of ovarian cancer. Different methods have been proposed to extract features from ultrasound images and classify ovarian tumors. Some techniques include SVM, shallow neural network classifiers, and wavelet transform filters. The features extracted from the images include textural and pathological features, wavelet coefficients, homogeneousness, histogram, and grey difference multi-scaling. These features are then integrated with a support-vector-machine (SVM) classifier to classify the types of cysts fully [19,20,21,22,23,24]. El-Bendary and Belal [25] developed a combined approach for subtyping cancer stage and ovarian cancer classification using gene expression and clinical datasets. Boosting and Ensemble SVM achieved a classification accuracy of 80%, while other ML classifiers yielded 70.77%. The proposed method also had a high recall, specificity, precision, F measure, and AUC.
The authors in [26] developed a CAD method for diagnosing borderline ovarian cancer by analyzing S-HG images. They used a k-NN classifier, and an optimization tool called T-POT, which resulted in an average accuracy between 0.976 and 0.96. The T-POT classifier achieved AUROC values of 0.97, 1, 0.98, and 0.99 for normal, malignant, borderline, and benign cases, respectively.
Deep learning techniques offer the advantage of intricate feature learning directly from raw data, making it possible to define a system without manually creating features that are necessary for other machine learning methods. This property has drawn interest in examining the benefits of using deep learning in medical image processing [27,28]. Multiphoton Microscopy (MPM) imaging and wide-field fluorescence imaging (WFI) has been proposed for ovarian imaging to capture specific biomarkers that cannot be detected by MPM alone while providing high-resolution images of the entire ovary. Using the Linear Discriminant Analysis classification algorithm, Sawyer et al. [29] achieved an accuracy of 66.66%, 87.50%, and 62.5% for genotype, age, and treatment, respectively [30,31].
The article highlights the different methods researchers propose for detecting and classifying various types of cancer using deep learning algorithms. Lu et al. [32] proposed a Tumour Origin and Assessment method via Deep Learning (TOAD), yielding an AUC-ROC of 0.988. Booma et al. [33] introduced a method using max-pooling enhanced with ML algorithms, which achieved an accuracy of 89%. Wen et al. utilized a custom collection of 3D filters achieving accuracy with AU-ROC ranging from 83% to 90% [34] (Wen et al., 2016). Huttunen et al. [35] used deep neural networks to classify unstained tissue multiphoton microscopy images achieving 95% sensitivity and 97% specificity. Wang et al. [36] proposed a two-level deep transfer learning approach, achieving an accuracy of 87.54%. Lastly, Yu et al. [37] proposed a radiomics model based on a convolutional neural network (CNN), which achieved an AUC of 0.955 to 0.975. They suggested that a CNN-based radiomics method for predicting endometrial cancer and distinguishing histopathology slides with cancer cells, showing potential clinical value in detecting tumour cells and predicting chemotherapy response.
Sengupta et al. [38] used DHL with lamin protein distribution to diagnose ovarian cancer based on the morphology of the nucleus, showing potential as a useful marker in prognosis and histology. Zhang et al. [39] used radiomics and a CNN-based model to predict Endometrioid Cancer with promising clinical potential, although with a limited sample size. Liao et al. [40] proposed the MTDL method for addressing high-dimensional feature space issues and improving classification accuracy, with the potential for further application to additional dataset categories. Various studies have proposed promising methods for diagnosing and predicting gynecological cancers, including ovarian and endometrial cancers, using advanced techniques such as MPM, WFI, DHL, radiomics, and CNN-based models, which hold potential for improving clinical outcomes.
Guo et al. [41] used deep and machine-learning techniques for clustering and classification in ovarian cancer subtyping using gene expression features. Kasture et al. [42] proposed a DL approach for predicting and classifying ovarian cancer subtypes using histopathological images. The authors in [43] used F-R-CNN to classify ovarian images. Ghoniem et al. [44] proposed a hybrid DL using multimodal data, combining gene and histopathology images with ALO-optimized LSTM and CNN networks. The proposed study of Ghoniem et al. suggests a novel approach to diagnosing ovarian cancer using a hybrid evolutionary deep-learning model that integrates multiple types of data and various assessment indicators. By comparing the performance of the proposed model with nine other multi-modal fusion models, the study demonstrates that the proposed model achieves higher precision and accuracy in ovarian cancer diagnoses.
A visual example of early detection through conventional methods as shown in Figure 1.
Finally, Xao et al. [45] used multi-omics to identify biomarkers for early OC diagnosis. We requested readers to check out in details the figure [46], which shows women with genetic mutations, or a family cancer history are at high risk. Tissue and liquid biopsy were analyzed using omics technology, combined with ML technology for biomarker development, and combined with the machine learning algorithm to form omics [47]. In addition, the authors refer to figure [47], which relates to using multi-omics to identify biomarkers, which speeds up the process of discovering new biomarkers. The process is displayed in Figure 2.
In their seminal work, Arezzo et al. investigated the application of radiomics in the qualitative and quantitative analysis of images obtained from various imaging techniques, such as magnetic resonance imaging (MRI) and ultrasound [48]. The primary objective of the radiomics approach is to achieve personalized ovarian cancer medicine. However, they concluded that radiomics has not yet been widely adopted in clinical settings due to several challenges, including:
  • The absence of a benchmark protocol for feature extraction in each method involved.
  • Bias is introduced by differences in images acquired from various instruments.
  • A lack of prospective external validation of algorithmic models on large datasets, resulting in insufficient studies addressing the performance of radiomics [48]. Arezzo et al. emphasized that images cannot only show the extent of the disease but also aid in its diagnosis. They suggested that deep learning could potentially overcome some of the limitations of radiomics in image analysis.
In a subsequent study, Arezzo et al. [48] revealed that peritoneal tuberculosis (TBP) exhibits similar signs and symptoms to advanced ovarian cancer. TBP accounts for approximately 1% to 2% of tuberculosis cases. The similarity between TBP and ovarian carcinoma presents challenges in the differential diagnosis. Biopsy remains the definitive method for confirming the presence of the disease, while computed tomography (CT) and ultrasound can be employed for visualizing nodules [49]. The authors proposed that a deep learning framework could effectively enhance the accuracy of image analysis, providing more objective diagnoses when using these imaging techniques.
Reilly et al. [50] outline the development and validation process of MIA3G, a deep neural network algorithm that aims to detect ovarian cancer. The algorithm was trained on a dataset of 1067 specimens and validated on a separate set of 2000 specimens. The results show that MIA3G achieves a sensitivity of 89.8% and a specificity of 84.02% in detecting ovarian cancer.
Machine learning integrates and analyses high-throughput molecular experiments to create new biomarkers for understanding illness. Integrated data analysis through machine learning can reveal biologically meaningful biomarker candidates despite changes at each omics level. Artificial intelligence studies are predicted to improve precision medicine for OC significantly. Elyan et al. [51] reviewed advances in using deep learning to analyse images and compared ML and DL algorithms. Figure 3 compares machine learning techniques that require manual feature extraction and deep learning, which do the extraction automatically.
Machine learning (ML) [53] enables computers to learn from past data without creating algorithms to account for the unlimited number of possible feature combinations. DL techniques differ from traditional approaches that use manual labelling. DL techniques learn the features of incoming photos from start to end without feature extraction or engineering. Convolutional neural networks (C-NNs) can capture the underlying representation of images by using partially linked layers and weight sharing, as shown in Figure 4.
The studies highlighted the potential of deep learning and machine learning techniques in ovarian cancer subtyping and diagnosis using various data modalities, including gene expression, histopathological images, and ultrasound images. These methods have shown promising results in improving accuracy and identifying subtypes at the molecular level. However, the small sample sizes and the need for larger datasets remain a knowledge gap.
Table 1 compares machine learning methods used for diagnosing and subtyping ovarian cancer. The table lists various researchers’ studies, methodologies, and performance metrics, such as accuracy, sensitivity, specificity, and AUC-ROC.

3. Materials and Methods

This section discusses the methodology used in the research, including the dataset preparation and the proposed Convolutional Neural Network (CNN) architecture. The dataset used in this project consists of 200 images, with 100 images each of serous ovarian cancer and non-cancerous samples. The original dataset was obtained from The Cancer Genome Atlas TCGA repository and was augmented to 11,040 images for both classes to be effectively used in the deep learning architecture (Kasture et al., 2021) [42]. This dataset was split into 80% for the training set and 20% for validation. The proposed CNN architecture is shown in Figure 5.
The preprocessed dataset was then fed for convolution operation. The preprocessing conducted on the images was the elimination of the images that had errors, such as the ones which were encoded badly and did not have the jpeg extension. The data was cleaned to remove the images that were distorted. Convolution is a special line function used to extract a feature using some feature detectors, also called a kernel, which are applied to all inputs, and a series of numbers called a tensor. The product-wise element between each kernel element and the input tensor is calculated in each tensor area and summarized to find the output value in the corresponding outgoing tensor area called the feature map. With striding and padding included, the formula for determining the output image dimension was as follows:
output image dimension = n + 2 p f s + 1   ×   n + 2 p f s + 1
where
  • n = Number of input pixels
  • f = Number of pixels of filter
  • p = Padding
  • s = Stride
To mimic the real-world representation mathematically of the behavior of neurons, the result of convolution was passed on to the non-linear activation function, ReLU, defined mathematically as f x = m a x ( 0 , x ) The activation function choice includes the fact that it does not engage all of the neurons at the same time and does not stimulate all of the neurons at the same time. Hence during backpropagation, not all the neurons are activated. Next, the data was passed onto the pooling layer, which provided a standard downtime that reduced the size of the feature map element to introduce translation flexibility in tiny shifts and distortions, and that reduced the number of subsequent parameters that could be learned. The research used Maxpooling. The data was then fed into the flattening layer and converted into a one-dimensional array or (vector) numbers for inputting into the fully connected layer. After that, the data is fed to the fully connected layer of feedforward neural networks. Every node in the previous layer is linked to the next layer’s nodes by a learnable weight. The features from the pooling layers are mapped to the network’s outputs. Lastly, the result was fed to the output layer, which used SoftMax for classification. The Loss function used was cross entropy and was given by the following formula for binary classification:
l = y l o g p + 1 y l o g 1 p
where:
  • l = Loss function
  • p = The predicted probability
  • y = 0 or 1 in binary classification
In conclusion, the proposed CNN architecture and dataset preparation for ovarian cancer classification has been discussed. The algorithm used to train the augmented image dataset has been summarized, and Table 2 summarizes the hyperparameters used in the CNN architecture.

4. Results

The model was fed with 11,040 images––half healthy cells, and the other half infected with serous cancer subtype. The number of epochs was chosen for hyperparameter selection and incremented in tens. The training and validation accuracy were recorded.
Table 3 shows a steady increase in the training and validation accuracy as the number of epochs increased from 10 to 50, reaching a training accuracy of 99.52% and a validation accuracy of 99.91% after 50 epochs.
After the training process, the testing process was performed by uploading an image from the test dataset, and the algorithm would give the percentage of it being either serous or healthy cells. This result demonstrates the superiority of the Xception network, which achieves high training and validation accuracy with a small number of epochs. Compared to conventional generic convolutional neural networks, the model does not perform channel-wise convolution, reducing the number of connections and making the model lighter. As a result, excellent accuracy can be achieved with just 10 epochs (99.03% training accuracy and 94.43% validation accuracy).
The model accuracy shows that the percentage for training and validation accuracy rose as shown in Figure 6. This graph shows accuracy from 98.56% and 99.73% to 99.52% and 99.91% for 10 to 50 epochs, respectively.
The percentage loss (shown in Figure 7) for training and validation stood at 0.0445 and 0.0083 and 0.0147, and 0.0020 after 10 and 50 epochs, respectively.
Figure 8 shows the 93.02% of the images were correctly classified as health, and 95.12% were classified as serous cells. From the confusion matrix, valuable performance metrics can be derived, such as:
Sensitivity = T P T P + F N = 93.02 93.02 + 4.88 = 0.9502
The sensitivity of the model is 0.9502, which represents the model’s ability to predict and classify the health cell images correctly as health when the label class is health.
Specificity = T N T N + F P = 95.12 95.12 + 6.98 = 0.9316
The specificity represents the model’s ability to correctly predict and classify the infected cells as infected cells when the class label is serous.
Recall = T P T P + F N = 0.9502
Recall measures how well the model can detect positive events, as in the health images class detection.
Precision = T P T P + F P = 93.02 93.02 + 6.98 = 0.9302
Precision measures how well the model has assigned the positive events to the positive class.
The   F - measure = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n = 2 × 0.9502 × 0.9302 0.9502 + 0.9302 = 0.94
The F-score is the harmonic mean of the model’s precision and recall, and it measures the model’s accuracy on the dataset of health and serous.
Table 4 shows that the precision for health is higher than that of serous cells by 2 units, and the recall for both is equal at 0.93. The F1 score is 0.94 for all classes. The classification report summarizes the performance of the Xception network on the medical images for the entire dataset.
Table 5 shows the accuracy score for different classification models; deep hybrid learning, convolutional neural network, GoogleNet (v3), and linear discriminant analysis classification. The GoogleNet v3 achieved the highest accuracy of 0.925, and the linear discriminant achieved the lowest of 0.666. As a comparison, the accuracy of GoogleNet v3 can be attributed to the transfer learning process, whereby the features are input to a pre-trained architecture. The hyperparameters would have been tuned already. The method that was proposed outperformed the ones by other researchers. A classification accuracy of 0.94 can be attributed to the efficient use of the parameters. The network is built because it has batch normalization after every convolution, which helps fight the model’s overfitting.
Table 5 shows that a “Deep Hybrid Learning” model achieved an accuracy score of 0.907. This model combines deep learning and traditional machine learning techniques, such as decision trees or linear regression, to improve performance. The specific structure of the network used in the study is not provided. Still, it likely involves multiple layers of deep neural networks and incorporates techniques such as batch normalization to combat overfitting.

5. Discussion

In this study, we aimed to develop a deep-learning model to classify healthy and serous ovarian cancer cells accurately. Our model achieved a high accuracy of 94.43%, a sensitivity of 0.9502, and a specificity of 0.9316, demonstrating its potential to assist physicians in diagnosing and predicting ovarian cancer prognosis, ultimately improving patient outcomes.
Our findings build upon previous research in the field, with our model surpassing the performance of other models reported in the literature. The superior performance can be attributed to the depth-wise separable convolution, which reduces model complexity, and the absence of activation functions in intermediate layers and residual connections. These innovations enabled our model to achieve higher accuracy than the 80% reported by El-Nabawy et al. [25], the 66.66–87.5% by Sawyer et al. [29], and the 83.6% by Lu et al. [32].
However, the study has some limitations. A public TCGA dataset may not fully represent local patient populations, and the retrospective design does not account for disease progression. Additionally, the lack of clinical data, such as patient genetics, may affect model robustness. To address these issues, future research should involve partnerships with hospitals and research centers to validate the model on real-world data and incorporate clinical datasets.
Alternative explanations for the findings should also be considered. Model interpretation techniques could provide insights into the underlying factors driving the model’s predictions, and the model’s design can be further refined to improve performance.
The practical implications of our findings include the potential to support ovarian cancer diagnosis and monitoring in clinical settings. Our model also offers recommendations for future research, such as addressing limitations, improving model interpretability, and validating the model on diverse populations.

6. Conclusions

In summary, this study explored the potential of a deep learning model for predicting and diagnosing ovarian cancer. Our model achieved an accuracy of 94%, highlighting its promise for early prediction and diagnosis. The study contributes to the existing literature by demonstrating the potential of deep learning in preclinical cancer screening and personalized management of ovarian cancer.
The broader implications of our research include the potential impact on related fields, such as radiology, and real-world applications in ovarian cancer diagnosis and treatment optimization. Despite some limitations, our study offers suggestions for future research to build upon our findings, such as improving data quality and quantity, evaluating models on new independent datasets, and enhancing generalizability through diverse population studies.
Overall, this research emphasizes the potential of advanced computing techniques in the timely detection of ovarian cancer. By addressing current challenges and refining these models, we can develop more reliable and broadly applicable tools that can be adopted for practical medical purposes, ultimately improving cancer diagnosis, outcome prediction, and personalized medical care.

Author Contributions

Conceptualization, Q.A., M.A. (Mubarak Albathan) and A.H.; Data curation, A.Y., Q.A., M.B. and A.H.; Formal analysis, A.Y., T.M., M.U.T., M.B., M.A. (Muhammad Asim) and A.H.; Funding acquisition, M.A. (Mubarak Albathan); Investigation, B.Z., A.Y., T.M., M.U.T., M.A. (Mubarak Albathan), S.J. and A.H.; Methodology, B.Z., A.Y., T.M., M.U.T., S.J., M.B. and M.A. (Muhammad Asim); Project administration, M.A. (Mubarak Albathan); Resources, B.Z., T.M., M.U.T., S.J., M.B. and M.A. (Muhammad Asim); Software, B.Z., A.Y., T.M., M.U.T., Q.A., M.A. (Mubarak Albathan), M.B. and A.H.; Supervision, M.A. (Muhammad Asim); Validation, T.M., M.U.T., M.A. (Mubarak Albathan) and S.J.; Visualization, Q.A., M.A. (Mubarak Albathan) and M.A. (Muhammad Asim); Writing—original draft, B.Z., Q.A., S.J., M.B. and A.H.; Writing—review and editing, Q.A. and M.A. (Muhammad Asim). All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) for funding and supporting this work through Research Partnership Program no. RP-21-07-11.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset will be available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Torre, L.A.; Bray, F.; Siegel, R.L.; Ferlay, J.; Lortet-Tieulent, J.; Jemal, A. Global cancer statistics, 2012. CA Cancer J. Clin. 2015, 65, 87–108. [Google Scholar] [CrossRef] [PubMed]
  2. Reid, B.M.; Permuth, J.B.; Sellers, T.A. Epidemiology of ovarian cancer: A review. Cancer Biol. Med. 2017, 14, 9–32. [Google Scholar] [CrossRef] [PubMed]
  3. Vázquez, M.A.; Mariño, I.P.; Blyuss, O.; Ryan, A.; Gentry-Maharaj, A.; Kalsi, J.; Manchanda, R.; Jacobs, I.; Menon, U.; Zaikin, A. A quantitative performance study of two automatic methods for the diagnosis of ovarian cancer. Biomed. Signal Process. Control. 2018, 46, 86–93. [Google Scholar] [CrossRef] [PubMed]
  4. Van Haaften-Day, C.; Shen, Y.; Xu, F.; Yu, Y.; Berchuck, A.; Havrilesky, L.J.; De Bruijn HW, A.; Van Der Zee AG, J.; Bast, R.C.; Hacker, N.F. OVX1, Macrophage-Colony Stimulating Factor, and CA-125-II as Tumor Markers for Epithelial Ovarian Carcinoma A Critical Appraisal. Cancer Interdisciplin. Int. J. Am. Cancer Soc. 2001, 92, 2837–2844. [Google Scholar] [CrossRef]
  5. Skates, S.J.; Mai, P.; Horick, N.K.; Piedmonte, M.; Drescher, C.W.; Isaacs, C.; Armstrong, D.K.; Buys, S.S.; Rodriguez, G.C.; Horowitz, I.R.; et al. Large Prospective Study of Ovarian Cancer Screening in High-Risk Women: CA125 Cut-Point Defined by Menopausal Status. Cancer Prev. Res. 2011, 4, 1401–1408. [Google Scholar] [CrossRef]
  6. Sopik, V.; Rosen, B.; Giannakeas, V.; Narod, S.A. Why have ovarian cancer mortality rates declined? Part III. Prospects for the future. Gynecol. Oncol. 2015, 138, 757–761. [Google Scholar] [CrossRef]
  7. Moss, E.L.; Hollingworth, J.; Reynolds, T.M. The role of CA125 in clinical practice. J. Clin. Pathol. 2005, 58, 308–312. [Google Scholar] [CrossRef]
  8. Akinwunmi, B.O.; Babic, A.; Vitonis, A.F.; Cramer, D.W.; Titus, L.; Tworoger, S.S.; Terry, K.L. Chronic Medical Conditions and CA125 Levels among Women without Ovarian Cancer. Cancer Epidemiol. Biomark. Prev. 2018, 27, 1483–1490. [Google Scholar] [CrossRef]
  9. Drapkin, R.; von Horsten, H.H.; Lin, Y.; Mok, S.C.; Crum, C.P.; Welch, W.R.; Hecht, J.L. Human Epididymis Protein 4 (HE4) Is a Secreted Glycoprotein that Is Overexpressed by Serous and Endometrioid Ovarian Carcinomas. Cancer Res. 2005, 65, 2162–2169. [Google Scholar] [CrossRef]
  10. Yanaranop, M.; Anakrat, V.; Siricharoenthai, S.; Nakrangsee, S.; Thinkhamrop, B. Is the Risk of Ovarian Malignancy Algorithm Better Than Other Tests for Predicting Ovarian Malignancy in Women with Pelvic Masses? Gynecol. Obstet. Investig. 2017, 82, 47–53. [Google Scholar] [CrossRef]
  11. Wu, C.; Wang, Y.; Wang, F. Deep Learning for Ovarian Tumor Classification with Ultrasound Images. In Advances in Multimedia Information Processing–PCM 2018: 19th Pacific-Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 395–406. [Google Scholar] [CrossRef]
  12. Bolstad, N.; Øijordsbakken, M.; Nustad, K.; Bjerner, J. Human epididymis protein 4 reference limits and natural variation in a Nordic reference population. Tumor Biol. 2011, 33, 141–148. [Google Scholar] [CrossRef]
  13. Ferraro, S.; Schiumarini, D.; Panteghini, M. Human epididymis protein 4: Factors of variation. Clin. Chim. Acta 2015, 438, 171–177. [Google Scholar] [CrossRef] [PubMed]
  14. Jayson, G.C.; Kohn, E.C.; Kitchener, H.C.; Ledermann, J.A. Ovarian cancer. Lancet 2014, 384, 1376–1388. [Google Scholar] [CrossRef] [PubMed]
  15. Kommoss, S.; Pfisterer, J.; Reuss, A.; Diebold, J.; Hauptmann, S.; Schmidt, C.; Du Bois, A.; Schmidt, D.; Kommoss, F. Specialized Pathology Review in Patients with Ovarian Cancer. Int. J. Gynecol. Cancer 2013, 23, 1376–1382. [Google Scholar] [CrossRef]
  16. Yoshida-Court, K.; Karpinets, T.V.; Mitra, A.; Solley, T.N.; Dorta-Estremera, S.; Sims, T.T.; Delgado Medrano, A.Y.; El Alam, M.B.; Ahmed-Kaddar, M.; Lynn, E.J.; et al. Immune environment and antigen specificity of the T cell receptor repertoire of malignant ascites in ovarian cancer. PLoS ONE 2023, 18, e0279590. [Google Scholar] [CrossRef] [PubMed]
  17. de Leon, A.; Perera, R.; Nittayacharn, P.; Cooley, M.; Jung, O.; Exner, A.A. Ultrasound Contrast Agents and Delivery Systems in Cancer Detection and Therapy. Adv. Cancer Res. 2018, 139, 57–84. [Google Scholar] [CrossRef] [PubMed]
  18. Lusk, J.F.; Miranda, C.; Howell, M.; Chrest, M.; Eshima, J.; Smith, B.S. Photoacoustic Flow System for the Detection of Ovarian Circulating Tumor Cells Utilizing Copper Sulfide Nanoparticles. ACS Biomater. Sci. Eng. 2019, 5, 1553–1560. [Google Scholar] [CrossRef]
  19. Danaee, P.; Ghaeini, R.; Hendrix, D.A. A deep learning approach for cancer detection and relevant gene identification. In Proceedings of the 22nd Pacific Symposium on Biocomputing (PSB), Kohala Coast, HI, USA, 4–8 January 2017; pp. 219–229. [Google Scholar]
  20. Chen, S.-J.; Chang, C.-Y.; Chang, K.-Y.; Tzeng, J.-E.; Chen, Y.-T.; Lin, C.-W.; Hsu, W.-C.; Wei, C.-K. Classification of the Thyroid Nodules Based on Characteristic Sonographic Textural Feature and Correlated Histopathology Using Hierarchical Support Vector Machines. Ultrasound Med. Biol. 2010, 36, 2018–2026. [Google Scholar] [CrossRef]
  21. Acharya, U.R.; Sree, S.V.; Swapna, G.; Gupta, S.; Molinari, F.; Garberoglio, R.; Witkowska, A.; Suri, J.S. Effect of complex wavelet transform filter on thyroid tumor classification in three-dimensional ultrasound. Proc. Inst. Mech. Eng. Part H J. Eng. Med. 2013, 227, 284–292. [Google Scholar] [CrossRef]
  22. Katsigiannis, S.; Keramidas, E.G.; Maroulis, D. A Contourlet Transform Feature Extraction Scheme for Ultrasound Thyroid Texture Classification. Int. J. Eng. Intell. Syst. Electr. Eng. Commun. 2010, 18, 171. [Google Scholar]
  23. Chang, C.-Y.; Chen, S.-J.; Tsai, M.-F. Application of support-vector-machine-based method for feature selection and classification of thyroid nodules in ultrasound images. Pattern Recognit. 2010, 43, 3494–3506. [Google Scholar] [CrossRef]
  24. Chang, C.-Y.; Liu, H.-Y.; Tseng, C.-H.; Shih, S.-R. Computer-aided diagnosis for thyroid graves’ disease in ultrasound images. Biomed. Eng. Appl. Basis Commun. 2012, 22, 91–99. [Google Scholar] [CrossRef]
  25. El-Nabawy, A.; El-Bendary, N.; Belal, N.A. Epithelial Ovarian Cancer Stage Subtype Classification using Clinical and Gene Expression Integrative Approach. Procedia Comput. Sci. 2018, 131, 23–30. [Google Scholar] [CrossRef]
  26. Wang, G.; Sun, Y.; Jiang, S.; Wu, G.; Liao, W.; Chen, Y.; Lin, Z.; Liu, Z.; Zhuo, S. Machine learning-based rapid diagnosis of human borderline ovarian cancer on second-harmonic generation images. Biomed. Opt. Express 2021, 12, 5658–5669. [Google Scholar] [CrossRef] [PubMed]
  27. Ma, S.; Sigal, L.; Sclaroff, S. Learning Activity Progression in LSTMs for Activity Detection and Early Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1942–1950. [Google Scholar] [CrossRef]
  28. Aliamiri, A.; Shen, Y. Deep learning based atrial fibrillation detection using wearable photoplethysmography sensor. In Proceedings of the 2018 IEEE EMBS International Conference on Biomedical Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018; pp. 442–445. [Google Scholar]
  29. Sawyer, T.W.; Koevary, J.W.; Rice, F.P.S.; Howard, C.C.; Austin, O.J.; Connolly, D.C.; Cai, K.Q.; Barton, J.K. Quantification of multiphoton and fluorescence images of reproductive tissues from a mouse ovarian cancer model shows promise for early disease detection. J. Biomed. Opt. 2021, 24, 096010. [Google Scholar] [CrossRef]
  30. Liang, Q.; Wendelhag, I.; Wikstrand, J.; Gustavsson, T. A multiscale dynamic programming procedure for boundary detection in ultrasonic artery images. IEEE Trans. Med. Imaging 2000, 19, 127–142. [Google Scholar] [CrossRef]
  31. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  32. Lu, M.Y.; Chen, T.Y.; Williamson, D.F.K.; Zhao, M.; Shady, M.; Lipkova, J.; Mahmood, F. AI-based pathology predicts origins for cancers of unknown primary. Nature 2021, 594, 106–110. [Google Scholar] [CrossRef]
  33. Booma, P.M.; Thiruchelvam, V.; Ting, J.; Ho, S. Max pooling technique to detect and classify medical image for max pooling technique to detect and classify medical image for ovarian cancer diagnosis. Test Eng. Manag. J. 2020, 82, 8423–8442. [Google Scholar]
  34. Wen, B.; Campbell, K.R.; Tilbury, K.; Nadiarnykh, O.; Brewer, M.A.; Patankar, M.; Singh, V.; Eliceiri, K.W.; Campagnola, P.J. 3D texture analysis for classification of second harmonic generation images of human ovarian cancer. Sci. Rep. 2016, 6, 35734. [Google Scholar] [CrossRef]
  35. Huttunen, M.J.; Hassan, A.; McCloskey, C.W.; Fasih, S.; Upham, J.; Vanderhyden, B.C.; Boyd, R.W.; Murugkar, S. Automated classification of multiphoton microscopy images of ovarian tissue using deep learning. J. Biomed. Opt. 2018, 23, 66002. [Google Scholar] [CrossRef] [PubMed]
  36. Wang, C.-W.; Lee, Y.-C.; Chang, C.-C.; Lin, Y.-J.; Liou, Y.-A.; Hsu, P.-C.; Chang, C.-C.; Sai, A.-K.-O.; Wang, C.-H.; Chao, T.-K. A Weakly Supervised Deep Learning Method for Guiding Ovarian Cancer Treatment and Identifying an Effective Biomarker. Cancers 2022, 14, 1651. [Google Scholar] [CrossRef] [PubMed]
  37. Yu, K.-H.; Hu, V.; Wang, F.; Matulonis, U.A.; Mutter, G.L.; Golden, J.A.; Kohane, I.S. Deciphering serous ovarian carcinoma histopathology and platinum response by convolutional neural networks. BMC Med. 2020, 18, 236. [Google Scholar] [CrossRef] [PubMed]
  38. Sengupta, D.; Ali, S.N.; Bhattacharya, A.; Mustafi, J.; Mukhopadhyay, A.; Sengupta, K. Nuclear Morphology Optimized Deep Hybrid Learning (NUMODRIL): A novel architecture for accurate diagnosis/prognosis of Ovarian Cancer. bioRxiv 2020. [Google Scholar] [CrossRef]
  39. Zhang, Y.; Gong, C.; Zheng, L.; Li, X.; Yang, X. Deep Learning for Intelligent Recognition and Prediction of Endometrial Cancer. J. Health Eng. 2021, 2021, 1148309. [Google Scholar] [CrossRef] [PubMed]
  40. Liao, Q.; Ding, Y.; Jiang, Z.L.; Wang, X.; Zhang, C.; Zhang, Q. Multi-task deep convolutional neural network for cancer diagnosis. Neurocomputing 2019, 348, 66–73. [Google Scholar] [CrossRef]
  41. Guo, L.-Y.; Wu, A.-H.; Wang, Y.-X.; Zhang, L.-P.; Chai, H.; Liang, X.-F. Deep learning-based ovarian cancer subtypes identification using multi-omics data. BioData Min. 2020, 13, 10. [Google Scholar] [CrossRef]
  42. Kasture, K.R.; Shah, D.D.; Matte, P.N. Research Article A New Deep Learning method for Automatic Ovarian Cancer Prediction & Subtype classification. Turk. J. Comput. Math. Educ. Res. Artic. 2021, 12, 1233–1242. [Google Scholar]
  43. Kavitha, S.; Vidyaathulasiraman. Identification and classification of early stage Ovarian cancer using convolutional neural network. Ilkogr. Online-Elem. Educ. Online 2021, 20, 1908–1924. [Google Scholar] [CrossRef]
  44. Ghoniem, R.M.; Algarni, A.D.; Refky, B.; Ewees, A.A. Multi-Modal Evolutionary Deep Learning Model for Ovarian Cancer Diagnosis. Symmetry 2021, 13, 643. [Google Scholar] [CrossRef]
  45. Xiao, Y.; Bi, M.; Guo, H.; Li, M. Multi-omics approaches for biomarker discovery in early ovarian cancer diagnosis. Ebiomedicine 2022, 79, 104001. [Google Scholar] [CrossRef] [PubMed]
  46. Woman Ovarian Cancer Figure. Available online: https://ars.els-cdn.com/content/image/1-s2.0-S2352396422001852-gr1_lrg.jpg (accessed on 13 April 2023).
  47. Machine Learning Technology for Biomarker Development Figure. Available online: https://ars.els-cdn.com/content/image/1-s2.0-S2352396422001852-gr2.jpg (accessed on 13 April 2023).
  48. Arezzo, F.; Loizzi, V.; La Forgia, D.; Moschetta, M.; Tagliafico, A.S.; Cataldo, V.; Kawosha, A.A.; Venerito, V.; Cazzato, G.; Ingravallo, G.; et al. Radiomics Analysis in Ovarian Cancer: A Narrative Review. Appl. Sci. 2021, 11, 7833. [Google Scholar] [CrossRef]
  49. Arezzo, F.; Cormio, G.; La Forgia, D.; Kawosha, A.A.; Mongelli, M.; Putino, C.; Silvestris, E.; Oreste, D.; Lombardi, C.; Cazzato, G.; et al. The Application of Sonovaginography for Implementing Ultrasound Assessment of Endometriosis and Other Gynaecological Diseases. Diagnostics 2022, 12, 820. [Google Scholar] [CrossRef]
  50. Reilly, G.P.; Dunton, C.J.; Bullock, R.G.; Ure, D.R.; Fritsche, H.; Ghosh, S.; Pappas, T.C.; Phan, R.T. Validation of a deep neural network-based algorithm supporting clinical management of adnexal mass. Front. Med. 2023, 10, 1–11. [Google Scholar] [CrossRef] [PubMed]
  51. Elyan, E.; Vuttipittayamongkol, P.; Johnston, P.; Martin, K.; McPherson, K.; Moreno-García, C.F.; Jayne, C.; Sarker, M.K. Computer vision and machine learning for medical image analysis: Recent advances, challenges, and way forward. Artif. Intell. Surg. 2022, 2, 24–45. [Google Scholar] [CrossRef]
  52. Gumbs, A.A.; Frigerio, I.; Spolverato, G.; Croner, R.; Illanes, A.; Chouillard, E.; Elyan, E. Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery? Sensors 2021, 21, 5526. [Google Scholar] [CrossRef]
  53. Sarvamangala, D.R.; Kulkarni, R.V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 2022, 15, 1–22. [Google Scholar] [CrossRef]
Figure 1. A visual example of early detection through conventional methods.
Figure 1. A visual example of early detection through conventional methods.
Diagnostics 13 01703 g001
Figure 2. Tissue and liquid biopsy were analysed using omics technology, combined with ML technology for biomarker development, combined with the machine learning algorithm to form omics.
Figure 2. Tissue and liquid biopsy were analysed using omics technology, combined with ML technology for biomarker development, combined with the machine learning algorithm to form omics.
Diagnostics 13 01703 g002
Figure 3. The approach of Earlier C-V techniques (top) against DL (bottom). (Gumbs et al.) [52].
Figure 3. The approach of Earlier C-V techniques (top) against DL (bottom). (Gumbs et al.) [52].
Diagnostics 13 01703 g003
Figure 4. CNN model schematic (Sarvamangala & Kulkarni, 2022) [53].
Figure 4. CNN model schematic (Sarvamangala & Kulkarni, 2022) [53].
Diagnostics 13 01703 g004
Figure 5. The proposed architecture of the model.
Figure 5. The proposed architecture of the model.
Diagnostics 13 01703 g005
Figure 6. Model Accuracy for both training and validation after 50 epochs.
Figure 6. Model Accuracy for both training and validation after 50 epochs.
Diagnostics 13 01703 g006
Figure 7. Model loss for both training and validation after 50 epochs.
Figure 7. Model loss for both training and validation after 50 epochs.
Diagnostics 13 01703 g007
Figure 8. Confusion Matrix of the proposed model.
Figure 8. Confusion Matrix of the proposed model.
Diagnostics 13 01703 g008
Table 1. Comparative Analysis of Machine Learning Methods for Diagnosis and Subtyping of Ovarian Cancer.
Table 1. Comparative Analysis of Machine Learning Methods for Diagnosis and Subtyping of Ovarian Cancer.
StudyMethodologyAccuracy/AUC-ROC/Sensitivity/Specificity
El-Nabawy et al. (2018) [25]Boosting and Ensemble SVM for ovarian cancer classification using gene expression and clinical datasetsAccuracy: 80%
Wang et al. (2021) [26]CAD method using k-NN classifier and T-POT for diagnosing borderline ovarian cancer by analyzing S-HG imagesAverage Accuracy: 0.976–0.96
Sawyer et al. (2021) [29]MPM and WFI imaging techniques with Linear Discriminant Analysis classification algorithm for ovarian imagingAccuracy: 66.66%/87.50%/62.5%
Lu et al. (2021) [32]Tumour Origin and Assessment Method via Deep Learning (TOAD)AUC-ROC: 0.988
Booma et al. (2020) [33]Max-pooling enhanced with ML algorithmsAccuracy: 89%
Wen et al. (2016) [34]Custom collection of 3D filtersAU-ROC: 83–90%
Huttunen et al. (2018) [35]Deep neural networks for classifying unstained tissue multiphoton microscopy imagesSensitivity: 95%/Specificity: 97%
Wang et al. (2020) [36]Two-level deep transfer learning approachAccuracy: 87.54%
Yu et al. (2020) [37]Radiomics model based on a convolutional neural networkAUC: 0.955–0.975
Guo et al. (2020) [41]Deep and machine-learning techniques for clustering and classification in ovarian cancer subtyping using gene expression featuresN/A
Kasture et al. (2021) [42]DL approach for predicting and classifying ovarian cancer subtypes using histopathological imagesN/A
Wu et al. (2018) [11]DL approach to classifying OC tumors using ultrasound imagesN/A
Vazquez et al. (2018) [3]Bayesian change point for interpretation and RNN for classificationN/A
Kavitha et al. (2021) [43]F-R-CNN to classify ovarian imagesN/A
Ghoniem et al. (2021) [44]Hybrid DL using multimodal data, combining gene and histopathology images with ALO-optimized LSTM and CNN networksN/A
Xiao et al. (2022) [45]Multi-omics to identify biomarkers for early OC diagnosisN/A
Table 2. Hyperparameters and Configuration Settings of CNN for Image Classification.
Table 2. Hyperparameters and Configuration Settings of CNN for Image Classification.
HyperparameterValue
Number of convolutional layers2
Kernel size3 × 3
Number of filters32, 64
Pooling typeMax pooling
Pooling size2 × 2
Number of neurons in fully connected layer128
Activation functionReLU
Output activation functionSoftmax
Loss functionCross-entropy
OptimizerAdam
Learning rate0.001
Batch size32
Number of epochs50
Table 3. The number of epochs and the corresponding training and validation accuracies.
Table 3. The number of epochs and the corresponding training and validation accuracies.
EpochsTraining AccuracyValidation Accuracy
1099.0394.43
2099.4685.16
3099.5790.45
4099.8298.78
5099.5299.91
Table 4. Classification Report for Health and Serous Classifications.
Table 4. Classification Report for Health and Serous Classifications.
PrecisionRecallF1-ScoreSupport
Health0.950.930.945520
Serous0.930.930.945520
Accuracy 0.9411,040
Macro Avg0.940.940.9411,040
Weighted Avg0.940.940.9411,040
Table 5. Comparison of Accuracy scores for different models.
Table 5. Comparison of Accuracy scores for different models.
ModelAccuracy
Deep hybrid learning0.907
Convolutional Neural Network0.897
GoogleNet (V3)0.925
Linear Discriminant Analysis Classification0.666
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ziyambe, B.; Yahya, A.; Mushiri, T.; Tariq, M.U.; Abbas, Q.; Babar, M.; Albathan, M.; Asim, M.; Hussain, A.; Jabbar, S. A Deep Learning Framework for the Prediction and Diagnosis of Ovarian Cancer in Pre- and Post-Menopausal Women. Diagnostics 2023, 13, 1703. https://doi.org/10.3390/diagnostics13101703

AMA Style

Ziyambe B, Yahya A, Mushiri T, Tariq MU, Abbas Q, Babar M, Albathan M, Asim M, Hussain A, Jabbar S. A Deep Learning Framework for the Prediction and Diagnosis of Ovarian Cancer in Pre- and Post-Menopausal Women. Diagnostics. 2023; 13(10):1703. https://doi.org/10.3390/diagnostics13101703

Chicago/Turabian Style

Ziyambe, Blessed, Abid Yahya, Tawanda Mushiri, Muhammad Usman Tariq, Qaisar Abbas, Muhammad Babar, Mubarak Albathan, Muhammad Asim, Ayyaz Hussain, and Sohail Jabbar. 2023. "A Deep Learning Framework for the Prediction and Diagnosis of Ovarian Cancer in Pre- and Post-Menopausal Women" Diagnostics 13, no. 10: 1703. https://doi.org/10.3390/diagnostics13101703

APA Style

Ziyambe, B., Yahya, A., Mushiri, T., Tariq, M. U., Abbas, Q., Babar, M., Albathan, M., Asim, M., Hussain, A., & Jabbar, S. (2023). A Deep Learning Framework for the Prediction and Diagnosis of Ovarian Cancer in Pre- and Post-Menopausal Women. Diagnostics, 13(10), 1703. https://doi.org/10.3390/diagnostics13101703

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop