Next Article in Journal
Excessive Daytime Sleepiness in Hypertensive Patients: The Role of Major Depressive Disorder
Previous Article in Journal
Urine Protein to Creatinine Ratio for the Assessment of Bevacizumab-Associated Proteinuria in Patients with Gynecologic Cancers: A Diagnostic and Quality Improvement Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Immunohistochemistry Interpretation with a Machine Learning-Based Expert System

1
Faculty of Medicine and Pharmacy, Dunarea de Jos University of Galati, 35 AI Cuza St., 800010 Galati, Romania
2
Saint John Clinical Emergency Hospital for Children, 800487 Galati, Romania
3
Saint Apostle Andrew Emergency County Clinical Hospital, 177 Brailei St., 800578 Galati, Romania
*
Authors to whom correspondence should be addressed.
Diagnostics 2024, 14(17), 1853; https://doi.org/10.3390/diagnostics14171853
Submission received: 13 June 2024 / Revised: 26 July 2024 / Accepted: 22 August 2024 / Published: 24 August 2024
(This article belongs to the Section Pathology and Molecular Diagnostics)

Abstract

:
Background: In recent decades, machine-learning (ML) technologies have advanced the management of high-dimensional and complex cancer data by developing reliable and user-friendly automated diagnostic tools for clinical applications. Immunohistochemistry (IHC) is an essential staining method that enables the identification of cellular origins by analyzing the expression of specific antigens within tissue samples. The aim of this study was to identify a model that could predict histopathological diagnoses based on specific immunohistochemical markers. Methods: The XGBoost learning model was applied, where the input variable (target variable) was the histopathological diagnosis and the predictors (independent variables influencing the target variable) were the immunohistochemical markers. Results: Our study demonstrated a precision rate of 85.97% within the dataset, indicating a high level of performance and suggesting that the model is generally reliable in producing accurate predictions. Conclusions: This study demonstrated the feasibility and clinical efficacy of utilizing the probabilistic decision tree algorithm to differentiate tumor diagnoses according to immunohistochemistry profiles.

1. Introduction

Immunohistochemistry (IHC) is a critical staining technique that facilitates the determination of cellular origins through the analysis of specific antigen expressions within tissue samples, thereby aiding in both diagnostic and prognostic assessments [1]. In addition to determining the origin of tumors, IHC is used to subtype tumors, evaluate the effectiveness of therapies, predict patient prognosis through the assessment of prognostic markers, and differentiate precancerous lesions by examining molecular alterations [2]. However, the rapidly expanding body of knowledge regarding IHC positivity in certain neoplasms often leads to divergent interpretations in standard procedures, especially in certain complex cases [3].
The interpretation of IHC data can vary based on the experience and expertise of individual pathologists. Researchers now have access to thousands of new antibodies and IHC-staining data across a wide range of tumor types [4]. Consequently, it is impractical for pathologists to memorize every molecular signal identified by the constantly evolving repertoire of antibodies in tumors originating from various tissues [5]. This issue can be solved by the use of standardized IHC panels and algorithmic techniques for specific diagnoses, but in practice, every case has its unique characteristics, and in certain situations, the use of IHC panels might be both time-consuming and expensive [6]. Notably, discovering a method to differentiate between various tumor subtypes would be highly desirable, particularly concerning potential clinical applications [7]. To this end, it would be highly advantageous to develop a new and practical cancer diagnostic tool that ideally uses as few distinguishing markers as possible, which can be used as an efficient and reliable method to identify and classify various tumor types. In recent decades, machine-learning (ML) technologies [8] have facilitated the management of high-dimensional and complex cancer data by creating reliable and user-friendly automated diagnostic tools for clinical applications. This advancement has allowed us to uncover new insights into the relationships between these changes and human cancers [9].
The latest advancements in ML have significantly improved the efficiency and speed of processing raw medical data. By incorporating these advanced technologies into the diagnostic process, pathologists can acquire valuable knowledge, enabling them to analyze large volumes of medical data more effectively and efficiently [10]. Deep-learning techniques, recently developed and rapidly gaining popularity, excel due to their ability to perform automated feature extraction and representation learning. As part of the learning process, this approach can autonomously extract information from images, leading to significantly improved performance in classification tasks [11]. Numerous studies have demonstrated that deep-learning systems outperform cutting-edge techniques in various medical-imaging analytic tasks, including computer-aided diagnosis, segmentation, detection, and classification [12,13,14].
The aim of this study was to identify a model that could predict histopathological diagnoses based on specific immunohistochemical markers.

2. Materials and Methods

To provide a clear understanding of our methodology, we have included a flowchart (Figure 1) that outlines the key steps involved in developing the proposed machine-learning model for predicting histopathological diagnoses based on IHC markers.

2.1. Sample Data Collection and Preprocessing

In order to develop the database for the ML tool to process and analyze, we conducted a retrospective study between 2019 and 2023, collecting data from the Pathology Department of “Saint John” Hospital in Galati, Romania, after obtaining the approval of the Ethics Committee. For each case, informed consent was obtained from the patient, indicating agreement with the collection, storage, and use of biological samples for both diagnostic and scientific purposes. Any data related to patient identification, except the histopathological diagnosis and the IHC results, were blinded before data processing. The diagnoses following immunohistochemical evaluation were collected, and for each case, the specific IHC marker panels used, and their positive or negative results, were documented.
Cases with insufficient IHC profiles—those without any markers indicating the origin of the tumor, IHC profiles with fewer than three markers, or unclear findings—were eliminated.
After collecting the data, a total of 686 cases were included in the primary dataset. IHC antibodies, averaging 5.4 per case (range 3–11), were utilized in the diagnostic process, with a total of 58 IHC markers identified and used in various combinations in panels to support the histopathological diagnosis of each case. These markers were utilized based on their relevance to various tumor types and included common antibodies such as Cytokeratin AE1/AE3, Cytokeratin 7, Cytokeratin 20, p63, Cytokeratin 5/6, E-cadherin, Estrogen Receptor, Progesterone Receptor, p53, S100, Melan-A, Human Melanoma Black 45 (HMB45), Epithelial Membrane Antigen (EMA), CD45, CD20, CD3, CD10, CD34, Thyroid Transcription Factor 1 (TTF1), Vimentin, Actin, Calretinin, Alpha Methyl Acyl CoA Racemase (AMACR), Prostate-Specific Antigen (PSA), and Glial Fibrillary Acidic Protein (GFAP). A total of 32 histopathological diagnoses were identified, but due to the limited sample size for some diagnoses, only 20 were included in the computations performed by the ML model, with the following distribution: invasive breast carcinoma, ductal type (15.31%), squamous cell carcinoma (12.54%), prostatic acinar adenocarcinoma (9.33%), pulmonary adenocarcinoma, not otherwise specified (NOS) (8.75%), colorectal adenocarcinoma, NOS (7.43%), serous carcinoma (4.37%), endometrioid carcinoma (3.21%), melanoma (3.21%), invasive breast carcinoma, lobular type (3.06%), basal cell carcinoma (2.62%), urothelial carcinoma (2.62%), gastric adenocarcinoma, NOS (2.04%), small cell neuroendocrine carcinoma (1.89%), invasive breast cancer, other types (1.75%), colorectal mucinous adenocarcinoma (1.60%), glioblastoma (1.60%), hepatocellular carcinoma (1.46%), cholangiocarcinoma (1.46%), Kaposi sarcoma (1.46%), and undifferentiated pleomorphic sarcoma (1.46%). Diagnoses like malignant mesothelioma, mixed germ cell tumor, or non-Hodgkin lymphoma, with less than 10 cases, were categorized as “Other” by the program due to their limited numbers, representing 12.83% of the total cases.
To test the ML model, a secondary dataset comprising 21 cases from an alternative center was created to mitigate the bias associated with diagnoses conducted by the same pathologists. This new dataset adhered to the same inclusion and exclusion criteria as the primary dataset, encompassing the most frequent diagnoses, such as invasive ductal carcinoma of the breast or pulmonary adenocarcinoma, as well as diagnoses categorized as “Other”, such as pancreatic adenocarcinoma.
All the collected data were compiled into two Excel tables with 59 columns, with the first column containing the histopathological diagnosis and the remaining columns listing the IHC markers with their respective positive or negative results, or not applicable (N/A) if the marker was not used for that specific case.

2.2. Machine-Learning Algorithm Model and Model Development

The ML algorithm employed in our study leverages Bayes’ theorem, a fundamental concept in probability theory and statistics that connects conditional probabilities with marginal probabilities. It states that if the pre-event probability is known, the post-event probability can be determined. Conditional probabilities indicate the likelihood of one event occurring, given that another event has occurred. Additionally, the probabilities of detecting events separately are also considered [15].
IHC results are binary, indicating either positive or negative outcomes, and pathologists possess practical knowledge about the probability of specific IHC markers being positive or negative in each tumor, also based on well-documented information in textbooks and other publications. The probability of positivity represents the number of positive cases out of all the diagnoses. After obtaining the test results, the likelihood of each diagnosis can be computed by multiplying the prior probability by the chance of each test being positive or negative. This allows for the comparison of post-probability values to determine which diagnosis has the highest likelihood.
To generate a prediction model, the Exploratory Desktop software, version 6.12.5.2, developed by Exploratory, Inc. was used. Exploratory Desktop offers a straightforward and user-friendly interface for accessing several data sources, transforming and cleaning data, visualizing and analyzing data, and using a wide range of machine-learning algorithms to explore data and gain deeper insights quickly. XGBoost, or eXtreme Gradient Boosting, is a well-liked machine-learning algorithm that creates prediction models in the form of an ensemble of weak prediction models—usually Decision Trees—for regression and classification tasks. Similar to previous boosting techniques, it constructs the model step-by-step and extends it by permitting optimization of any differentiable loss function.

2.3. Model Training, Validation and Testing

The first step in training the model was to import our Excel dataset into the data frame from the Exploratory main interface. Once the data were loaded, the wrangling tools available were used to make some minor adjustments to eliminate any program errors, misinterpretation of data, or false results. In the second step, the XGBoost learning model was applied, where the input variable (target variable) was the histopathological diagnosis and the predictors (independent variables influencing the target variable) were the immunohistochemical markers. In the final step, the model training parameters were set; in the first instance, the default parameters provided by XGBoost were used and the model performance was checked followed by a grid search for hyperparameter tuning to find the best parameter combination.
To validate the model, the XGBoost advanced settings have an option that allows splitting the dataset into various ratios of training data and validation data, performing the training and validation at the same time. We chose a ratio of 7:3, and to avoid overfitting, we increased the number of iterations and the number of early stopping iterations. Finally, the negative log-likelihood was chosen for the multiclass classification metric. After the model was properly trained and validated, we saved it to apply the testing data set.
Testing the XGBoost model in Exploratory involved the following steps: importing the testing dataset in Exploratory, applying the trained model to the test data, and the program outputting the predicted values alongside the actual value in a new data frame. By following these detailed steps, we could effectively assess our model’s generalization ability and make informed decisions on any further tuning or adjustments.

2.4. Model Performance Evaluation

After the program computes the data, the micro- and macro-average precision scores, micro- and macro-average recall scores, accuracy rate, F1 score, and misclassification rate will be calculated. The most important metric is perhaps the F1 score, which is used to evaluate the performance of a classification model, such as XGBoost, by balancing precision and recall. It represents the harmonic mean of precision and recall, offering a single metric that encompasses both aspects of the model’s accuracy. When using models in the healthcare industry to make recommendations for diagnoses, the F1 score can be particularly helpful. A high F1 score suggests that the model is effective at identifying both positive and negative cases, which is crucial for reducing misdiagnosis and ensuring that patients receive the correct care [16].
Micro-average precision is the total of all the true positives divided by the total of all the true positives and false positives. In other words, it is the total number of accurate predictions divided by the total number of predictions. Macro-average precision is the arithmetic mean of the precision values for the different classes. The total of all the true positives for every class divided by the number of actual positives (as opposed to forecasted positives) yields the micro-average recall, and the macro-average recall is the arithmetic mean of all the recall scores for different classes [17].
One of the most straightforward performance metrics in machine learning is accuracy. It is a statistic that indicates how many of all the predictions were correct. This measure is quite simple to use in binary and multiclass classification problems, but it is important to understand its nuances and limitations. Accuracy is determined by dividing the total number of predictions by the sum of the true positives and true negatives. For example, a model would be 90% accurate if it made 100 predictions and 90 of them were correct (either true positives or true negatives). The misclassification rate is a machine-learning metric that represents the percentage of incorrect predictions made by a classification system [17].

3. Results

3.1. Model Training and Validation Results

Table 1 presents a comprehensive summary of the performance metrics for the XGBoost model across different histopathological diagnoses. The metrics include the F1 score, accuracy rate, misclassification rate, precision, and recall for both the training and validation datasets.
As demonstrated in Table 1, the XGBoost model exhibits high accuracy and recall, indicating its effectiveness in identifying colorectal adenocarcinoma NOS, despite its lower precision suggesting some false positives. Validation data confirm the model’s strong performance on unseen data, maintaining high precision and recall with a minor trade-off in recall.
For a better representation and understanding of the findings, we grouped the diagnoses according to the model performance as follows: Group 1: high performance and generalization; Group 2: strong training performance, minor validation drops; Group 3: moderate performance with room for improvement; and Group 4: poor performance and generalization issues.
Group 1: High Performance and Generalization
  • Prostatic acinar adenocarcinoma
  • Gastric adenocarcinoma, NOS
  • Pulmonary adenocarcinoma, NOS
  • Basal cell carcinoma
  • Invasive breast carcinoma, ductal type
These diagnoses show exceptional model performance, with high F1 scores, perfect or near-perfect recall, and high precision. The validation results indicate strong generalization to unseen data, suggesting robust and reliable model performance. For instance, the model’s perfect scores in training and very high validation scores for prostatic acinar adenocarcinoma and basal cell carcinoma highlight its robustness.
Group 2: Strong Training Performance, Minor Validation Drop
  • Hepatocellular carcinoma
  • Invasive breast carcinoma, lobular type
  • Small cell neuroendocrine carcinoma
  • Squamous cell carcinoma
  • Serous carcinoma
These diagnoses exhibit strong training performance, with high F1 scores, perfect or high recall, and high precision. The validation results show slight decreases in precision or recall but maintain high performance overall. For example, the model’s training recall of 1.0000 for invasive breast carcinoma, lobular type, and its validation precision of 0.9091 indicate strong performance with minor validation drops.
Group 3: Moderate Performance with Room for Improvement
  • Urothelial carcinoma
  • Cholangiocarcinoma
  • Glioblastoma
  • Melanoma
  • Kaposi sarcoma
  • Undifferentiated pleomorphic sarcoma
These diagnoses show moderate performance, with high precision in training but significant drops in the recall and F1 scores in terms of validation, indicating difficulties in generalizing to unseen data. For instance, the model’s perfect precision for glioblastoma in validation contrasts with its low recall, highlighting the need for improvement in detecting true positives. Similarly, the model’s improved recall in validation for Kaposi sarcoma suggests better performance on new data.
Group 4: Poor Performance and Generalization Issues
  • Colorectal mucinous adenocarcinoma
  • Endometrioid carcinoma
  • Invasive breast carcinoma, other types
These diagnoses exhibit poor performance, with low F1 scores and recall. The high precision in relation to training is misleading due to the very low recall, indicating many missed true cases. Validation metrics confirm this issue, showing significant decreases in precision and recall, suggesting difficulties in generalizing to unseen data. For example, the model’s low recall for colorectal mucinous adenocarcinoma and endometrioid carcinoma indicates it fails to identify many true cases.
Overall, the XGBoost model’s high recall and accuracy across multiple carcinoma types make it a valuable tool for early and accurate cancer detection. Its ability to generalize well to new data ensures its reliability in real-world clinical settings. However, improving precision for certain carcinoma types can enhance performance, reducing false positives and unnecessary follow-up procedures. By addressing these areas through targeted improvements and following best practices for model deployment, the XGBoost model can significantly enhance diagnostic accuracy, providing robust support for clinicians in cancer diagnosis and treatment planning.
Figure 2 illustrates the learning curves for an XGBoost model, depicting the negative log likelihood (NLL) against the number of iterations for both the training and validation datasets over 20 iterations. Both curves (training and validation) exhibit a general downward trend as the number of iterations increases. This indicates that the model is learning and improving its performance over time, with the NLL decreasing, which corresponds to better model predictions. The training NLL starts at around 1.6 and decreases steadily, flattening out to around 0.4 after 20 iterations. This consistent decrease signifies that the model is fitting the training data increasingly well. The validation NLL starts at around 1.8 and also decreases, but at a slower rate compared to the training curve. By the end of 20 iterations, it levels out around 0.6. The slower decrease compared to the training curve suggests that the model’s performance on unseen data is improving but not as rapidly as on the training data.
There is a noticeable gap between the training and validation NLL throughout the iterations, with the validation NLL consistently higher than the training NLL. This gap indicates that the model is fitting the training data better than the validation data, a common scenario that may suggest some degree of overfitting. However, since the validation NLL continues to decrease and does not show a marked increase or significant divergence, the overfitting is not severe. Both curves appear to be converging toward their respective minimums as the number of iterations increases, suggesting that additional iterations may not result in a substantial improvement in model performance.

3.2. Model Testing Results

After training and validating the model, a new set of 21 patients was used for testing. The following results were obtained after applying the model (Table 2). The model exhibits strong performance in predicting the following histopathologic diagnoses: squamous cell carcinoma, basal cell carcinoma, prostatic acinar adenocarcinoma, glioblastoma, invasive breast carcinoma ductal type, small cell neuroendocrine carcinoma, Kaposi sarcoma, and serous carcinoma. The precise alignment between the true labels and predicted labels for these diagnoses indicates the model’s high reliability and robustness, ensuring both high precision and recall. This consistency is crucial for clinical accuracy and minimizing the risk of misdiagnosis.
However, the model frequently misclassified pancreatic adenocarcinoma, gastric poorly cohesive carcinoma, pulmonary adenocarcinoma NOS, undifferentiated pleomorphic sarcoma, and endometrioid carcinoma into the “Other” category. This misclassification suggests that the model struggled to distinguish these specific cancer types, likely due to insufficient training data or overlapping features with other cancer types. For instance, pancreatic adenocarcinoma and undifferentiated pleomorphic sarcoma had few cases in the training set, resulting in inadequate learning by the model. Additionally, the misclassification of gastric poorly cohesive carcinoma as invasive breast carcinoma, lobular type indicates potential feature similarities or insufficient distinction in the IHC markers used.
The model’s high performance for several common cancer types indicates its potential value in clinical settings, aiding in early and accurate diagnosis, which is critical for effective treatment planning. However, the misclassifications and lower performance for certain rare or complex cancer types highlight the need for more comprehensive training data and the incorporation of additional diagnostic markers or advanced feature-engineering techniques to improve the model’s robustness.
Overall, the model shows promising results for many cancer types. Addressing its limitations through targeted improvements can enhance its diagnostic accuracy and reliability, thus providing better support for clinical decision-making.

3.3. Model Performance Evaluation

Table 3 presents a comprehensive performance summary of the XGBoost model, as evaluated on both training and validation datasets. The key metrics provided include the micro-average F score, macro-average F score, accuracy rate, and misclassification rate. These metrics collectively provide insights into the model’s predictive accuracy, generalizability, and reliability, which are critical for its application in clinical settings.
The micro-average F score aggregates the model’s performance across all instances to provide an overall measure of effectiveness. A score of 0.8597 for the training data indicates a high level of accuracy, reflecting the model’s ability to balance precision and recall. The validation score of 0.8137, though slightly lower, demonstrates strong performance and suggests that the model maintains a high level of predictive accuracy on unseen data. This minimal drop from training to validation suggests good generalization without significant overfitting.
The macro-average F score measures the model’s performance across all classes, treating each class equally. The training score of 0.8446 indicates consistent performance across different cancer types. The validation score of 0.8061, while slightly lower, still reflects robust performance on unseen data. This metric is particularly important for understanding the model’s ability to handle class imbalances and maintain performance across both common and rare cancer types.
The accuracy rate measures the proportion of correctly predicted instances. The high accuracy rate of 85.97% for the training data indicates that the model accurately classifies the majority of instances. The validation accuracy rate of 81.37%, though slightly lower, still reflects a high level of correctness on new data. The small decrease in accuracy from training to validation demonstrates the model’s reliability when applied to unseen data, which is critical for clinical applications.
The misclassification rate represents the proportion of incorrect predictions. A rate of 14.02% for the training data and 18.62% for the validation data highlights areas for improvement. The increased misclassification rate in the validation data suggests that while the model is effective, it encounters more challenges when predicting unseen cases.
Model Evaluation Summary:
  • Strengths:
    High precision and recall for several common cancer types, making it valuable for early and accurate diagnosis.
    Consistent performance across different cancer types, essential for handling class imbalances in clinical datasets.
  • Areas for Improvement:
    The increased misclassification rate in the validation data indicates the need for more comprehensive training data, especially for rare cancer types.
    Incorporating additional diagnostic markers or advanced feature-engineering techniques could improve the model’s ability to differentiate between similar cancer types, thereby reducing misclassification rates.

4. Discussion

4.1. State-of-the-Art Machine-Learning Algorithms for Histopathology

Traditionally, various machine-learning algorithms are applied to classification tasks, including Nearest Neighbor [18], Decision Trees [19], Artificial Neural Network [20], Support Vector Machine (SVM) [21], Ensemble Learning [22], and Convolutional Neural Network (CNN) [11]. These techniques aim to achieve high classification accuracy. However, they often struggle with challenges such as imbalanced misclassification costs across different classes. For example, these methods can have a higher likelihood of incorrectly identifying non-cancerous patients as cancerous or vice versa. This misclassification can lead to significant consequences, such as increased costs for further pathological examination.
Regarding cancer detection and tumoral grading, Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) are extensively used for histopathological image analysis. For example, DINOv2 has shown exceptional performance in classifying colorectal cancer tissue patches [23]. These models excel in handling large-scale image data and performing automated feature extraction and representation learning. Chen et al. [24] introduced a hierarchical image pyramid transformer architecture that combines representations from multiple fields of view to obtain a slide-level representation, which is effective for processing gigapixel images in histopathology.
Another area of application of machine-learning models is in tumor segmentation and biomarker detection. Hierarchical self-supervised learning frameworks are used for precise segmentation of tumor regions within histopathological images, analyzing images at multiple scales to enhance accuracy [25]. Echle et al. [26] have utilized machine learning to quantify biomarkers like HER2 in breast cancer, achieving high performance in real-time clinical analysis.
Several researchers have employed machine-learning models for feature extraction and survival prediction. For instance, Yao et al. [27] utilized attention-based multiple instance learning (MIL) networks for survival prediction by analyzing whole-slide images (WSIs). These networks focus attention on relevant patches, thereby enhancing the accuracy of survival predictions. Additionally, Hashimoto et al. [28] developed a multi-scale domain-adversarial convolutional neural network (CNN) for cancer subtype classification, utilizing unannotated histopathological images to mitigate performance variation across different sites. Other researchers have employed machine-learning models using IHC data to predict clinical outcomes in cancer patients. They compared the performance of these AI models with baseline features, establishing the superior efficacy of AI in predicting clinical outcomes [29,30].
XGBoost has been effectively combined with deep-learning models for tasks such as classifying breast tumors. For instance, combining DenseNet201 features with XGBoost has demonstrated high classification accuracy in breast cancer detection [25]. Also, XGBoost has been used to develop predictive models for high-risk conditions such as high-risk MASH (Maternal Alloimmune Hepatic Syndrome). The integration of SHAP (SHapley Additive exPlanations) enhances the interpretability of these models, which is crucial for clinical applications [31].
It can be noticed that the XGBoost model has not yet been explored for classifying a variety of histopathological diagnoses based on particular immunohistochemical markers.

4.2. Performance Metrics and Generalization

In the present study, we validated the estimated diagnostic probability of specific histopathological diagnoses based on particular immunohistochemical markers. This was achieved using a machine-learning system employing a probabilistic Decision Tree model, such as XGBoost. These models have been applied to identify lung cancer and colon cancer subtypes [32], predict lung metastases from thyroid cancer [33], and develop risk models for detecting lung cancer [34], all of which have demonstrated high performance.
Our study demonstrated a precision rate of 85.97% within the dataset, indicating a high level of performance and suggesting that the model is generally reliable in producing accurate predictions. The precision rate of the predicted diagnoses was generally higher compared to prior research on tumors of unknown origin, which reported a precision rate of 79.9% [35]. However, it was lower than a study using lymphoma cases, which achieved a precision rate of 95% [36].
A high precision rate was observed in diagnoses such as adenocarcinoma with various localizations and melanoma. The authors attributed these findings to the fact that the machine-learning system provides more accurate predictions when trained on a larger number of cases. Other studies have shown that the accuracy rate for each tumor type is closely related to the corresponding sample size. Moreover, an imbalance in the dataset complicates the achievement of high performance [37,38,39].
Trying to understand the issues that led to cases being classified under a different diagnosis than the correct one, we analyzed each case in the testing set individually. In the case of pancreatic adenocarcinoma classified as “Other”, the issue is that the training set included only six cases with this diagnosis, which was not sufficient for learning and creating a distinct class, as this diagnosis was not among the 20 subtypes mentioned in Table 1. A similar explanation can be applied to undifferentiated pleomorphic sarcoma, which had five cases in the training set. For poorly cohesive gastric carcinoma, the training set had only three cases, which did not allow for the creation of a class. However, due to the use of common markers for both diagnoses (CK7, HER2, E-cadherin), the model classified it as lobular breast carcinoma instead of labeling it as “Other”. The case of lung adenocarcinoma classified as “Other” is atypical; as there were sufficient data in the training set, the F1 score was quite high and the precision was 1.0000. The explanation for this could be that it was a poorly differentiated brain metastasis, requiring the use of a larger number of markers (specifically 11) to establish the diagnosis, which likely led to an inaccurate prediction. The situation where an endometrioid carcinoma had the predicted label “Other” is attributed to the fact that this class has an F1 score of 0.7778, indicating moderate performance, most likely due to the lack of specificity of the IHC markers for this tumor subtype. The challenge of imbalanced data is critical in histopathology, where rare cancer types are often underrepresented. This study’s results align with the broader literature emphasizing the need for comprehensive and balanced datasets to enhance model accuracy [25,31]. While the model showed high performance for common cancer types, its generalization to rare types was limited. This is consistent with the challenges faced by other state-of-the-art models, suggesting improvements in training data diversity and advanced feature-engineering are necessary to enhance generalization capabilities [23,31].
As aforementioned, the present study employed a binary tumor marker expression system (positive or negative) to forecast the histopathological diagnosis. Nevertheless, alternative investigations have focused on developing mathematical models capable of predicting cancer types using molecular characteristics, such as gene expression, DNA methylation profiles, or somatic alteration analysis [40,41,42,43]. Through the inclusion of driver gene mutations and complex structural variant-related features, Nguyen et al. [43] achieved the prediction of approximately 35 different cancer (sub)types with an accuracy of approximately 90%. Marquard et al. [44] successfully distinguished between 10 cancer types with 69% accuracy and 6 cancer types with 85% accuracy when considering solely somatic mutations or both somatic mutations and copy number alterations, respectively. Utilizing the presence of somatic point mutations and copy number alterations in 50 genes as predictive features, Soh et al. [38] attained an accuracy of approximately 77% across 28 distinct cancer types.
Another method for predicting histopathological diagnoses, documented in the scientific literature, involves employing machine-learning systems that analyze cellular properties like nuclear size and the homogeneity of cell populations. This approach holds clinical utility for distinguishing various cell groups. Moran et al. [45] utilized the cell detection module within QuPath, an open-source image analysis toolbox for digital pathology, to extract morphological features of Merkel cell carcinoma (MCC). Their findings indicated that the nuclear area and circularity were significant prognostic factors for MCC. Faridi et al. [46] employed region-growing methods to segment cells, extracting morphological features such as the nuclear size, solidity, and eccentricity to grade the pleomorphism scores of whole-slide images of breast cancer cases, and attained an accuracy of 86.6%.

4.3. Challenges, Limitations, and Recommendations for Future Research

The model frequently misclassified certain rare or complex cancer types. This misclassification indicates that the model struggled to differentiate these specific cancer types, likely due to insufficient training data or overlapping features with other cancer types. For example, pancreatic adenocarcinoma and poorly cohesive gastric carcinoma were often misclassified, highlighting the need for more comprehensive training data and advanced feature-engineering techniques to enhance the model robustness. This challenge aligns with the findings from other studies that emphasize the necessity for more extensive training data and sophisticated feature engineering to improve the model performance [23,35,36,37,38,39].
While the model demonstrated high performance for several common cancer types, its performance dropped when applied to rare or less common types. This indicates a limitation in the model’s ability to generalize across all cancer types. This study suggests that increasing the sample size for each diagnostic category and incorporating additional diagnostic markers could enhance the model’s generalization capabilities. This recommendation is consistent with other research that advocates for techniques like regularization to improve generalization [25,35,42].
This study notes that there was an imbalance in the dataset used for training the machine-learning model. Certain tumor types had significantly fewer cases, which affected the model’s ability to learn and accurately predict these diagnoses. For instance, the training set included only a small number of cases of pancreatic adenocarcinoma and undifferentiated pleomorphic sarcoma, making it difficult for the model to create distinct classes for these diagnoses. Increasing the sample size for each cancer type and incorporating additional diagnostic markers can improve the model’s robustness and accuracy [16,17,31].
To enhance the model’s learning and prediction capabilities, we could expand the dataset to include more cases for each cancer type, with a particular focus on rare ones. Additionally, incorporating a greater variety of diagnostic markers and employing advanced feature-engineering techniques could improve differentiation between similar cancer types. Exploring the use of more complex data systems, such as gene expression and DNA methylation profiles, would further capture the intricacies of tumor marker expressions.

5. Conclusions

This study demonstrated the feasibility and efficacy of utilizing the probabilistic Decision Tree algorithm to differentiate tumor diagnoses according to immunohistochemistry profiles.
The primary sources of error within this system include disease-specific markers, the presence of overlapping IHC profiles among diseases, a deficiency of site-specific markers, mixed or combined tumors, and atypical IHC profiles. Another source of error arises from the limited number of cases available in the training dataset for each diagnosis. Future research should address this by increasing the sample size for each diagnostic category, allowing for more robust analysis and validation of the findings. Nevertheless, this system can serve as a valuable aid to pathologists in rendering accurate diagnoses.

Author Contributions

Methodology, data curation, writing—original draft preparation, A.I.N. and M.N.; writing—review and editing, A.I.N., D.G.P., M.N., A.F., I.F. and A.N.; supervision, conceptualization and funding, A.N., A.I.N. and I.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the “Dunărea de Jos” University of Galati, VAT 3127522, and the APC was paid by the “Dunărea de Jos” University of Galati, VAT 3127522.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Saint John Clinical Emergency Hospital for Children, Galati, Romania (protocol code C676/20.11.2023).

Informed Consent Statement

Informed consent was obtained from all the subjects involved in the study.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. de Matos, L.L.; Trufelli, D.C.; de Matos, M.G.L.; da Silva Pinhal, M.A. Immunohistochemistry as an Important Tool in Biomarkers Detection and Clinical Practice. Biomark. Insights 2010, 5, 9–20. [Google Scholar] [CrossRef]
  2. Werner, B.; Campos, A.C.; Nadji, M.; Torres, L.F.B. Uso prático da imuno-histoquímica em patologia cirúrgica. J. Bras. Patol. Med. Lab. 2005, 41, 353–364. [Google Scholar] [CrossRef]
  3. DeYoung, B.R.; Wick, M.R. Immunohistologic Evaluation of Metastatic Carcinomas of Unknown Origin: An Algorithmic Approach. Semin. Diagn. Pathol. 2000, 17, 184–193. [Google Scholar]
  4. Dabbs, D. Diagnostic Immunohistochemistry, 3rd ed.; Elsevier Ltd.: Philadelphia, PA, USA, 2010; pp. 1–3. [Google Scholar]
  5. Kandalaft, P.L.; Gown, A.M. Practical Applications in Immunohistochemistry: Carcinomas of Unknown Primary Site. Arch. Pathol. Lab. Med. 2016, 140, 508–523. [Google Scholar] [CrossRef]
  6. Lin, F.; Prichard, J. (Eds.) Handbook of Practical Immunohistochemistry: Frequently Asked Questions, 2nd ed.; Springer: New York, NY, USA, 2015; pp. 17–30. [Google Scholar] [CrossRef]
  7. Ciriello, G.; Miller, M.L.; Aksoy, B.A.; Senbabaoglu, Y.; Schultz, N.; Sander, C. Emerging Landscape of Oncogenic Signatures across Human Cancers. Nat. Genet. 2013, 45, 1127–1133. [Google Scholar] [CrossRef]
  8. Greenman, C.; Stephens, P.; Smith, R.; Dalgliesh, G.L.; Hunter, C.; Bignell, G.; Davies, H.; Teague, J.; Butler, A.; Stevens, C.; et al. Patterns of Somatic Mutation in Human Cancer Genomes. Nature 2007, 446, 153–158. [Google Scholar] [CrossRef]
  9. Bertsimas, D.; Wiberg, H. Machine Learning in Oncology: Methods, Applications, and Challenges. JCO Clin. Cancer Inform. 2020, 4, 885–894. [Google Scholar] [CrossRef]
  10. Robertson, S.; Azizpour, H.; Smith, K.; Hartman, J. Digital Image Analysis in Breast Pathology—From Image Processing Techniques to Artificial Intelligence. Transl. Res. 2018, 194, 19–35. [Google Scholar] [CrossRef]
  11. Lecun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  12. Alirezazadeh, P.; Hejrati, B.; Monsef-Esfahani, A.; Fathi, A. Representation Learning-Based Unsupervised Domain Adaptation for Classification of Breast Cancer Histopathology Images. Biocybern. Biomed. Eng. 2018, 38, 671–683. [Google Scholar] [CrossRef]
  13. Bayramoglu, N.; Kannala, J.; Heikkila, J. Deep Learning for Magnification Independent Breast Cancer Histopathology Image Classification. In Proceedings of the International Conference on Pattern Recognition, Cancun, Mexico, 4–8 December 2016; pp. 2440–2445. [Google Scholar] [CrossRef]
  14. Gandomkar, Z.; Brennan, P.C.; Mello-Thoms, C. MuDeRN: Multi-Category Classification of Breast Histopathological Image Using Deep Residual Networks. Artif. Intell. Med. 2018, 88, 14–24. [Google Scholar] [CrossRef]
  15. Lesaffre, E.; Lawson, A.B. Bayesian Biostatistics, 1st ed.; John Wiley & Sons, Ltd.: Chichester, UK, 2012; pp. 15–18. [Google Scholar]
  16. Yacouby, R.; Axman, D. Probabilistic Extension of Precision, Recall, and F1 Score for More Thorough Evaluation of Classification Models. In Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems, Online, 12 November 2020; pp. 79–91. [Google Scholar] [CrossRef]
  17. Buckland, M.; Gey, F. The Relationship between Recall and Precision. J. Am. Soc. Inf. Sci. 1994, 45, 12–19. [Google Scholar] [CrossRef]
  18. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2016. [Google Scholar]
  19. Quinlan, J.R. Induction of Decision Trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  20. Amato, F.; López, A.; Peña Méndez, E.M.; Vaňhara, P.; Hampl, A.; Havel, J. Artificial Neural Networks in Medical Diagnosis. J. Appl. Biomed. 2013, 11, 47–58. [Google Scholar] [CrossRef]
  21. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  22. Ho, T.K. The Random Subspace Method for Constructing Decision Forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  23. Bahadir, C.D.; Omar, M.; Rosenthal, J.; Marchionni, L.; Liechty, B.; Pisapia, D.J.; Sabuncu, M.R. Artificial Intelligence Applications in Histopathology. Nat. Rev. Electr. Eng. 2024, 1, 93–108. [Google Scholar] [CrossRef]
  24. Chen, R.J.; Chen, C.; Li, Y.; Chen, T.Y.; Trister, A.D.; Krishnan, R.G.; Mahmood, F. Scaling Vision Transformers to Gigapixel Images via Hierarchical Self-Supervised Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  25. Springenberg, M.; Frommholz, A.; Wenzel, M.; Weicken, E.; Ma, J.; Strodthoff, N. From Modern CNNs to Vision Transformers: Assessing the Performance, Robustness, and Classification Strategies of Deep Learning Models in Histopathology. Med. Image Anal. 2023, 87, 102809. [Google Scholar] [CrossRef]
  26. Echle, A.; Rindtorff, N.T.; Brinker, T.J.; Luedde, T.; Pearson, A.T.; Kather, J.N. Deep Learning in Cancer Pathology: A New Generation of Clinical Biomarkers. Br. J. Cancer 2021, 124, 686–696. [Google Scholar] [CrossRef]
  27. Yao, J.; Zhu, X.; Jonnagaddala, J.; Hawkins, N.; Huang, J. Whole Slide Images Based Cancer Survival Prediction Using Attention Guided Deep Multiple Instance Learning Networks. Med. Image Anal. 2020, 65, 101789. [Google Scholar] [CrossRef]
  28. Hashimoto, N.; Fukushima, D.; Koga, R.; Takagi, Y.; Ko, K.; Kohno, K.; Nakaguro, M.; Nakamura, S.; Hontani, H.; Takeuchi, I. Multi-Scale Domain-Adversarial Multiple-Instance CNN for Cancer Subtype Classification with Unannotated Histopathological Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar] [CrossRef]
  29. Mi, H.; Bivalacqua, T.J.; Kates, M.; Seiler, R.; Black, P.C.; Popel, A.S.; Baras, A.S. Predictive Models of Response to Neoadjuvant Chemotherapy in Muscle-Invasive Bladder Cancer Using Nuclear Morphology and Tissue Architecture. Cell Rep. Med. 2021, 2, 100287. [Google Scholar] [CrossRef]
  30. Nakhli, R.; Moghadam, P.A.; Mi, H.; Farahani, H.; Baras, A.; Gilks, B.; Bashashati, A. Sparse Multi-Modal Graph Transformer with Shared-Context Processing for Representation Learning of Giga-Pixel Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17-24 June 2023; pp. 11547–11557. [Google Scholar]
  31. Unger, M.; Kather, J.N. A Systematic Analysis of Deep Learning in Genomics and Histopathology for Precision Oncology. BMC Med. Genom. 2024, 17, 48. [Google Scholar] [CrossRef]
  32. Hage Chehade, A.; Abdallah, N.; Marion, J.-M.; Oueidat, M.; Chauvet, P. Lung and Colon Cancer Classification Using Medical Imaging: A Feature Engineering Approach. Phys. Eng. Sci. Med. 2022, 45, 729–746. [Google Scholar] [CrossRef]
  33. Liu, W.; Wang, S.; Ye, Z.; Xu, P.; Xia, X.; Guo, M. Prediction of Lung Metastases in Thyroid Cancer Using Machine Learning Based on SEER Database. Cancer Med. 2022, 11, 2503–2515. [Google Scholar] [CrossRef]
  34. Li, Y.; Shen, S.; Wang, X.; Wang, M.; Zhang, X.; Zhong, B.; Zhan, Q.; Zhang, X.; Ren, H.; Jin, L. Prediction of Lung Cancer Risk in Chinese Population with Genetic-Environment Factor Using Extreme Gradient Boosting. Cancer Med. 2022, 11, 4469–4478. [Google Scholar] [CrossRef]
  35. Chong, Y.; Han, S.; Kim, H.; Noh, S.; Lee, E.J.; Kang, C.S.; Park, C.K. Diagnosis Prediction of Tumours of Unknown Origin Using ImmunoGenius, a Machine Learning-Based Expert System for Immunohistochemistry Profile Interpretation. Diagn. Pathol. 2021, 16, 19. [Google Scholar] [CrossRef]
  36. Chong, Y.; Kim, D.C.; Han, S.; Noh, S.; Kang, C.S.; Park, C.K. A Machine-Learning Expert-Supporting System for Diagnosis Prediction of Lymphoid Neoplasms Using a Probabilistic Decision-Tree Algorithm and Immunohistochemistry Profile Database. J. Pathol. Transl. Med. 2020, 54, 462–470. [Google Scholar] [CrossRef]
  37. Jiao, W.; Atwal, G.; Polak, P.; Karlic, R.; Cuppen, E.; Danyi, A.; de Ridder, J.; van Dyk, E.; Demichelis, F.; Spellman, P.; et al. A Deep Learning System Accurately Classifies Primary and Metastatic Cancers Using Passenger Mutation Patterns. Nat. Commun. 2020, 11, 728. [Google Scholar] [CrossRef]
  38. Soh, K.P.; Szczurek, E.; Sakoparnig, T.; Beerenwinkel, N. Predicting Cancer Type from Tumour DNA Signatures. Genome Med. 2017, 9, 48. [Google Scholar] [CrossRef]
  39. Lu, D.; Qiu, Y.; Hu, Z.; Peng, Y.; Wang, X.; Bao, H.; Xiao, Z.; Fan, X. Machine Learning Models to Predict Primary Sites of Metastatic Cervical Carcinoma from Unknown Primary. Front. Genet. 2020, 11, 614823. [Google Scholar] [CrossRef]
  40. Kang, S.; Li, Q.; Chen, Q.; Zhou, Y.; Park, S.; Lee, G. CancerLocator: Non-Invasive Cancer Diagnosis and Tissue-of-Origin Prediction Using Methylation Profiles of Cell-Free DNA. Genome Biol. 2017, 18, 53. [Google Scholar] [CrossRef] [PubMed]
  41. Hao, X.; Luo, H.; Krawczyk, M.; Wei, W.; Wang, W.; Wang, J.; Flagg, K.; Hou, J.; Zhang, H.; Yi, S.; et al. DNA Methylation Markers for Diagnosis and Prognosis of Common Cancers. Proc. Natl. Acad. Sci. USA 2017, 114, 7414–7419. [Google Scholar] [CrossRef]
  42. Ramaswamy, S.; Tamayo, P.; Rifkin, R.; Mukherjee, S.; Yeang, C.H.; Angelo, M.; Ladd, C.; Reich, M.; Latulippe, E.; Mesirov, J.P.; et al. Multiclass Cancer Diagnosis Using Tumor Gene Expression Signatures. Proc. Natl. Acad. Sci. USA 2001, 98, 15149–15154. [Google Scholar] [CrossRef] [PubMed]
  43. Nguyen, L.; Van Hoeck, A.; Cuppen, E. Machine Learning-Based Tissue of Origin Classification for Cancer of Unknown Primary Diagnostics Using Genome-Wide Mutation Features. Nat. Commun. 2022, 13, 4013. [Google Scholar] [CrossRef] [PubMed]
  44. Marquard, A.M.; Eklund, A.C.; Joshi, T.; Krzystanek, M.; Favero, F.; Wang, Z.; Jaiswal, A.; Trifonov, V.; Lindberg, J.; Szallasi, Z.; et al. TumorTracer: A Method to Identify the Tissue of Origin from the Somatic Mutations of a Tumor Specimen. BMC Med. Genom. 2015, 8, 58. [Google Scholar] [CrossRef]
  45. Moran, J.M.T.; Davids, M.S.; Rodriguez, V.; Girardi, M.; Krauthammer, M.; Mahmood, U. Large Nuclear Size Correlated with Better Overall Survival, Merkel Cell Polyomavirus Positivity, and Terminal Deoxynucleotidyl Transferase Expression in Merkel Cell Carcinoma. J. Am. Acad. Dermatol. 2021, 84, 550–552. [Google Scholar] [CrossRef]
  46. Faridi, P.; Danyali, H.; Helfroush, M.S.; Jahromi, M.A. An Automatic System for Cell Nuclei Pleomorphism Segmentation in Histopathological Images of Breast Cancer. In Proceedings of the 2016 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Philadelphia, PA, USA, 3 December 2016; pp. 1–5. [Google Scholar] [CrossRef]
Figure 1. Flowchart showing the study methodology.
Figure 1. Flowchart showing the study methodology.
Diagnostics 14 01853 g001
Figure 2. Learning curve of the XGBoost model: negative log likelihood vs. iterations.
Figure 2. Learning curve of the XGBoost model: negative log likelihood vs. iterations.
Diagnostics 14 01853 g002
Table 1. XGBoost learning model summary by class.
Table 1. XGBoost learning model summary by class.
Class F1 ScoreAccuracy RateMisclass. RatePrecisionRecall
Colorectal adenocarcinoma, NOSTraining0.83330.98190.01810.71431.0000
Validation0.82350.97050.02950.73680.9333
Colorectal adenocarcinoma, mucinousTraining0.33330.98190.01811.00000.2000
Validation0.21020.89210.10790.34780.5333
Prostatic acinar adenocarcinomaTraining1.00001.00000.00001.00001.0000
Validation0.92310.99500.00501.00000.8572
Gastric adenocarcinoma, NOSTraining0.90910.99100.00900.83331.0000
Validation0.88890.99580.00420.80001.0000
Pulmonary adenocarcinoma, NOSTraining0.88890.99100.00901.00000.8000
Validation0.90900.98520.01480.83331.0000
Basal cell carcinomaTraining0.95240.99550.00450.90911.0000
Validation1.00001.00000.00001.00001.0000
Endometrioid carcinomaTraining0.77780.98190.01810.87500.7000
Validation0.54540.97550.02450.60000.5000
Hepatocellular carcinomaTraining1.00001.00000.00001.00001.0000
Validation0.92300.99020.00981.00000.8571
Invasive breast carcinoma, other typesTraining0.57140.97290.02711.00000.4000
Validation0.54450.97920.02080.60000.5000
Invasive breast carcinoma, ductal typeTraining0.83330.98190.01810.71431.0000
Validation0.92170.96260.03740.86890.9815
Invasive breast carcinoma, lobular typeTraining0.95240.99550.00450.90911.0000
Validation0.93330.99510.00490.87501.0000
Small cell neuroendocrine carcinomaTraining0.80000.98190.01810.80000.8000
Validation0.77140.98760.01240.66670.3333
Squamous cell carcinomaTraining0.94740.99550.00451.00000.9000
Validation0.95230.98530.01470.90911.0000
Serous carcinomaTraining0.94740.99550.00451.00000.9000
Validation0.95230.98530.01470.90911.0000
Urothelial carcinomaTraining0.85710.98640.01360.81820.9000
Validation0.77780.98040.01960.70000.8750
CholangiocarcinomaTraining0.80000.99100.00901.00000.6667
Validation0.77140.98840.01160.66670.3333
GlioblastomaTraining0.86960.98640.01360.76921.0000
Validation0.66670.99510.00491.00000.5000
MelanomaTraining1.00001.00000.00001.00001.0000
Validation0.92310.99510.00491.00000.8571
Kaposi sarcomaTraining0.88890.99550.00451.00000.8000
Validation0.87880.98330.01670.87880.8788
Undifferentiated pleomorphic sarcomaTraining0.75000.99100.00901.00000.6000
Validation0.54550.97550.02450.60000.5000
OtherTraining0.82470.92310.07690.78430.8696
Validation0.82350.97060.02940.73680.9333
Table 2. Test dataset results.
Table 2. Test dataset results.
Histopathologic DiagnosisPredicted Label
Squamous cell carcinomaSquamous cell carcinoma
Basal cell carcinomaBasal cell carcinoma
Prostatic acinar adenocarcinomaProstatic acinar adenocarcinoma
GlioblastomaGlioblastoma
Invasive breast carcinoma, ductal typeInvasive breast carcinoma, ductal type
Squamous cell carcinomaSquamous cell carcinoma
Squamous cell carcinomaSquamous cell carcinoma
Pancreatic adenocarcinomaOther
Gastric poorly cohesive carcinomaInvasive breast carcinoma, lobular type
Pulmonary adenocarcinoma, NOSOther
Undifferentiated pleomorphic sarcomaOther
Small cell neuroendocrine carcinomaSmall cell neuroendocrine carcinoma
Invasive breast carcinoma, ductal typeInvasive breast carcinoma, ductal type
Invasive breast carcinoma, ductal typeInvasive breast carcinoma, ductal type
Kaposi sarcomaKaposi sarcoma
Small cell neuroendocrine carcinomaSmall cell neuroendocrine carcinoma
Pulmonary adenocarcinoma, NOSPulmonary adenocarcinoma, NOS
Prostatic acinar adenocarcinomaProstatic acinar adenocarcinoma
Small cell neuroendocrine carcinomaSmall cell neuroendocrine carcinoma
Serous carcinomaSerous carcinoma
Endometrioid carcinomaOther
Table 3. Model performance summary.
Table 3. Model performance summary.
Data TypeMicro-Average F ScoreMacro-Average F ScoreAccuracy RateMisclass. RateRows
Training 0.85970.84460.85970.1402481
Validation0.81370.80610.81370.1862204
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Neagu, A.I.; Poalelungi, D.G.; Fulga, A.; Neagu, M.; Fulga, I.; Nechita, A. Enhanced Immunohistochemistry Interpretation with a Machine Learning-Based Expert System. Diagnostics 2024, 14, 1853. https://doi.org/10.3390/diagnostics14171853

AMA Style

Neagu AI, Poalelungi DG, Fulga A, Neagu M, Fulga I, Nechita A. Enhanced Immunohistochemistry Interpretation with a Machine Learning-Based Expert System. Diagnostics. 2024; 14(17):1853. https://doi.org/10.3390/diagnostics14171853

Chicago/Turabian Style

Neagu, Anca Iulia, Diana Gina Poalelungi, Ana Fulga, Marius Neagu, Iuliu Fulga, and Aurel Nechita. 2024. "Enhanced Immunohistochemistry Interpretation with a Machine Learning-Based Expert System" Diagnostics 14, no. 17: 1853. https://doi.org/10.3390/diagnostics14171853

APA Style

Neagu, A. I., Poalelungi, D. G., Fulga, A., Neagu, M., Fulga, I., & Nechita, A. (2024). Enhanced Immunohistochemistry Interpretation with a Machine Learning-Based Expert System. Diagnostics, 14(17), 1853. https://doi.org/10.3390/diagnostics14171853

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop