Next Article in Journal
Psychological Impact of Pro-Anorexia and Pro-Eating Disorder Websites on Adolescent Females: A Systematic Review
Next Article in Special Issue
A Deep Learning Model for Classification of Endoscopic Gastroesophageal Reflux Disease
Previous Article in Journal
Effect of COVID-19 Pandemic on Orthopedic Surgery in Three Centers from Romania
Previous Article in Special Issue
Combinatorial K-Means Clustering as a Machine Learning Tool Applied to Diabetes Mellitus Type 2
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Feature Extraction Approach for Hematopoietic Cancer Subtype Classification

1
Database and Bioinformatics Laboratory, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
2
School of Medicine, Nankai University, Tianjin 300071, China
3
Department of Electrical Engineering, Faculty of Engineering, Chiang Mai University, Chiang Mai 50200, Thailand
4
Biomedical Engineering Institute, Chiang Mai University, Chiang Mai 50200, Thailand
5
Data Science Laboratory, Faculty of Information Technology, Ton Duc Thang University, Ho Chi Minh 700000, Vietnam
6
Department of Computer Science, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
*
Authors to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(4), 2197; https://doi.org/10.3390/ijerph18042197
Submission received: 30 December 2020 / Revised: 15 February 2021 / Accepted: 19 February 2021 / Published: 23 February 2021
(This article belongs to the Special Issue Disease Prediction, Machine Learning, and Healthcare)

Abstract

:
Hematopoietic cancer is a malignant transformation in immune system cells. Hematopoietic cancer is characterized by the cells that are expressed, so it is usually difficult to distinguish its heterogeneities in the hematopoiesis process. Traditional approaches for cancer subtyping use statistical techniques. Furthermore, due to the overfitting problem of small samples, in case of a minor cancer, it does not have enough sample material for building a classification model. Therefore, we propose not only to build a classification model for five major subtypes using two kinds of losses, namely reconstruction loss and classification loss, but also to extract suitable features using a deep autoencoder. Furthermore, for considering the data imbalance problem, we apply an oversampling algorithm, the synthetic minority oversampling technique (SMOTE). For validation of our proposed autoencoder-based feature extraction approach for hematopoietic cancer subtype classification, we compared other traditional feature selection algorithms (principal component analysis, non-negative matrix factorization) and classification algorithms with the SMOTE oversampling approach. Additionally, we used the Shapley Additive exPlanations (SHAP) interpretation technique in our model to explain the important gene/protein for hematopoietic cancer subtype classification. Furthermore, we compared five widely used classification algorithms, including logistic regression, random forest, k-nearest neighbor, artificial neural network and support vector machine. The results of autoencoder-based feature extraction approaches showed good performance, and the best result was the SMOTE oversampling-applied support vector machine algorithm consider both focal loss and reconstruction loss as the loss function for autoencoder (AE) feature selection approach, which produced 97.01% accuracy, 92.60% recall, 99.52% specificity, 93.54% F1-measure, 97.87% G-mean and 95.46% index of balanced accuracy as subtype classification performance measures.

1. Introduction

Lots of bioinformatics techniques have been developed for the disease detection and diagnosis of patients with incurable diseases such as cancer for several decades [1]. However, it still remains challenging to deal with cancer patients. Furthermore, the development of appropriate classification models using the gene which is extracted from patients is useful for early diagnosis of both patients and normal people. Cancer is a major disease which causes death, involving abnormal cell differentiation [2]. It is caused by many reasons, but the majority of cancers (90~95%) are due to genetic mutations from lifestyle factors such as smoking, obesity, alcohol and so on. The remaining 5~10% are caused due to inherited genes [3].
A hematopoietic malignancy is a neoplasm from hematopoietic cells in the bone marrow, lymph nodes, peripheral blood and lymphatic system which is related to organs of the hematopoietic system. Furthermore, it can be found in other organs such as the gastrointestinal system and the central nervous system [4]. Given that hematopoietic malignancies occur in the hematopoietic system, this malignancy is called a liquid tumor. While uncommon in other cancers, chromosomal translocations are a common cause of these diseases. Hematopoietic cancer accounts for 8~10% of all cancer diagnoses, and its mortality rate is also similar to this [5].
Historically, hematopoietic cancer is divided by whether the malignant location is in the blood or the lymph nodes. However, in 2001, the World Health Organization (WHO) introduced the WHO classification of tumors of hematopoietic and lymphoid tissue as a standard and updated it in 2008 and 2016 [6]. This WHO classification criterion focused on cell linkage rather than the location of the occurrence. According to the WHO classification, hematopoietic malignancies are mainly divided into leukemia, lymphoma and myeloma [7].
Leukemia is one type of hematopoietic cancer that results from genetic changes in hematopoietic cells in the blood or bone marrow. If an abnormality occurs in the bone marrow, the abnormally generated blood cells mix with blood in the body and spread widely into the body through the blood stream [8]. Most leukemia cases are diagnosed in adults aged over 65 years, but it is also commonly observed in children under the age of 15 [7]. The American Cancer Society (ACS) reported, in 2020, that the United States will see about 60,530 new cases and 23,100 deaths from leukemia [9]. Lymphoma is usually found in distinct stationary masses of lymphocytes, such as the lymph node, thymus or spleen. Like leukemia, lymphoma can also travel through the whole body by the blood stream. Commonly, lymphoma cases are divided into Hodgkin lymphoma, non-Hodgkin lymphoma, acquired immune deficiency syndrome (AIDS)-related lymphoma and primary central nervous system (CNS) lymphoma [8]. In 2020, the ACS reported that there will be approximately 85,720 new cases and 20,910 deaths from lymphoma [9]. Myeloma is a tumor that occurs in plasma cells which are differentiated from bone marrow, blood or other tissue. Plasma cells generate antibodies that protect against disease or infection, but when they develop abnormally, it interferes with antibody generation and causes confusion in the human immune system [8]. According to the estimation by the ACS, there would be 32,270 new cases and 12,830 deaths in 2020 [9].
There have been several studies conducted using genetic data-based cancer classification by machine learning and data mining approaches [10,11,12,13,14,15,16,17,18]. For instance, one study [10] utilized several machine learning approaches with naïve Bayes and k-nearest neighbor classification algorithm for breast cancer classification. Another study [11] applied classification algorithms for binary subtype cancer classification on acute myeloblastic leukemia microarray data. Other kinds of research also typically use machine learning algorithms such as support vector machine, decision tree, random forest, lasso and neural network [12,13,14,15,16,17,18].
Over the years, various approaches for data mining have been applied on many cancer research studies. Specifically, a deep learning method was applied in this area [19,20,21,22,23]. Ahmed M et al. [19] developed a breast cancer classification model using deep belief networks in an unsupervised part for learning input feature statistics. Additionally, in the supervised part, they adopted a conjugate gradient and Levenberg–Marquardt algorithm.
Furthermore, there are several studies on cancer subtype classification. A study [20] used a deep learning approach for kidney cancer subtype classification using miRNA data from The Cancer Genome Atlas (TCGA), which contained five different subtypes. They employed neighborhood component analysis for feature extraction and a long short-term memory (LSTM)-based classifier. DeepCC [22] architecture using the gene set enrichment analysis (GSEA) method and artificial neural network generated deep cancer subtype classification frameworks that made comparisons between machine learning algorithms such as support vector machine, logistic regression and gradient boost using a colon cancer dataset from TCGA which contains 14 subtype cancer labels, and a Deeptype [23] framework has been made for cancer subtype classification based on PAM50 [24]. They used a multi-layer neural network structure for adopting representation power to project on representation space.
Traditional approaches [10,11,12,13,14,15,16,17,18] have used statistical techniques for cancer classification. Sun et al. [17] used an entropy-based approach for feature extraction in a cancer classification model. However, this method has a disadvantage; that is, multiple classes cannot be applied at once because each cancer is classified one by one through binary classification for cancer classification. In the case of Deeptype [23], clustering is established using a specific gene set called PAM50, previously known for breast cancer.
However, since this PAM50 indicator is used as an already known indicator, subtypes can be classified through some valid information about breast cancer. Above all, for other cancers, including hematopoietic cancer, there is no gene set for subtype classification. In view of this, our work has a difference in feature extraction and subtype classification only from the gene expression data of hematologic cancer. In order to overcome the demerit of the multi-class classification task and the limitations due to the absence of a gene set, we applied a method of feature extraction using an autoencoder-based method among deep learning methods. In addition, we propose a subtype classification method in which reconstruction error is generated by the autoencoder. The classification error generated by the classification model and merged error are used as the loss function by referring to the loss function application methods of Deeptype [23].
The goal of this work is to develop an autoencoder-based feature extraction approach for hematopoietic cancer subtype classification. We not only focus on the five subtypes of hematopoietic cancer and conduct a study on classifying by applying deep learning techniques, but we also perform a feature extraction and calculation of two kinds of errors, which are reconstruction error and classification error. In the process, first a deep autoencoder approach is used for extracting suited features for building a classification model, and then the reconstruction error and classification error (cross-entropy loss and focal loss) are calculated for considering the data imbalance problem when building the classification model using the extracted features. To validate the deep autoencoder-based classification model, we compared other traditional feature selection algorithms and classification algorithms with an oversampling approach. We compared five widely used classification algorithms including logistic regression (LR), random forest (RF), k-nearest neighbor (KNN), artificial neural network (ANN) and support vector machine (SVM).
We compared our proposed method with traditional cancer classification and cancer subtype classification methods such as data mining and machine learning approaches which are not able to be used in the previous end-to-end approaches. Our end-to-end approach has multiple steps including feature engineering, data imbalance handling and a classification task. The objectives of this study are to extract features from a deep learning-based approach on the gene expression data for predicting hematopoietic cancer subtypes and develop an end-to-end deep learning-based classification model. The major contributions of this study are, briefly, as follows:
  • We propose an end-to-end approach without any manual engineering, which classifies hematopoietic cancer subtypes;
  • We adopt a non-linear transformation step by using a deep autoencoder to select deep features from gene expression data in hematopoietic cancer by adopting a deep learning architecture-based feature engineering task;
  • We implement a mixed loss function for the proposed deep learning model, considering both the compression of knowledge representation and the data imbalance problem.
The remainder of this paper is organized as follows: Section 2 introduces the hematopoietic cancer gene expression dataset from TCGA. Furthermore, the proposed deep autoencoder-based approach is explained in detail. In Section 3, the experimental results are provided. Finally, Section 4 discusses the experimental results with our conclusion.

2. Materials and Methods

2.1. Dataset

TCGA is a site that contains plenty of information and data related to human cancer. Currently, as of 2020, there are 47 types of cancer data, and each cancer’s data are provided various kinds of data such as gene expression, clinical data and methylation data from large numbers of patient with cancer [25]. Although some raw data, which include original sequence information, are treated by controlled data that have to be approved for use in experiments by TCGA, most data are freely accessible for researchers. We collected TCGA data from 2457 patients with hematopoietic cancer gene expression data. The collected gene expression data have five subtypes of hematopoietic cancer: lymphoid leukemia, myeloid leukemia, leukemia not otherwise specified (nos), mature B-cell leukemia and plasma cell neoplasm. The size of each hematopoietic subtype sample is 550 lymphoid leukemia cases, 818 myeloid leukemia, 104 leukemia nos, 113 mature B-cell leukemia and 860 plasma cell neoplasm. Furthermore, these data have 60,453 exons’ information with one gene expression profiling measurement. The level of gene expression is fragments per kilobase per million (FPKM) mapped measure [26]. This FPKM can be calculated by the following equation:
F P K M = T o t a l   f r a g m e n t s   o f   i n t e r e s t M a p p e d   r e a d s   ( M i l l i o n s ) × e x o n   l e n g t h   ( k b )
FPKM is a normalized estimation of gene expression based on RNA-seq data considering both the number of reads and the length of the exon, measured by kilobase unit. That is, a large FPKM means a large amount of expression per unit length, so the FPKM of a certain gene refers to a relative amount of gene expression.
The statistics of hematopoietic cancer are shown in Table 1. In a preprocessing step, we eliminated noisy and non-valued instances. These preprocessed data were used for the subtype classification in this experiment; they were divided into 80% for training and 20% for testing. However, as we introduced above, the dataset was considerably imbalanced.
Due to this data imbalance problem, we applied a cost function on the classification and feature extraction and oversampling method. We also used an autoencoder-based model for extracting the highly related gene expression data and compared this algorithm with other traditional dimension reduction algorithms.

2.2. Proposed Autoencoder-Based Approach

In the experiment, we propose a deep learning-based hematopoietic cancer subtype classification approach. Figure 1 shows the proposed approach which inputs the hematopoietic cancer gene expression data from TCGA and outputs the subtype classification result. This approach consists of an autoencoder feature extraction part and a machine learning-based cancer subtype classifier.
In the DAE structure, we employed the mean squared error (MSE) for measuring deep learning reconstruction loss when training the training set and adopted focal loss [27] as a measurement of the classification error in the classifier. Focal loss (FL) is an updated version of cross-entropy loss, which was used for class imbalance encountered during the model training. Therefore, our proposed autoencoder-based hematopoietic cancer subtype classification approach used the sum of both MSE as reconstruction loss and FL as classification loss as a cost function for this approach.
We performed this experiment on an Intel Xeon E3 1231 v3 processor with 32G memory and RTX 2060 (Gigabyte, New Taipei City, Taiwan). Additionally, we used Python 3.7 for parsing the data and analysis by implementing deep learning and machine learning libraries. The whole process of this experiment and the methodologies including the machine learning and deep learning approaches performed are explained in detail in the next section.

2.2.1. Feature Extraction using Deep Learning Approach on Gene Expression Data

In this research, we used a DAE-based feature selection approach. The autoencoder structure has a strong point in the non-linear feature selection and transformation. Additionally, we compared this DAE-based approach with traditional statistical-based feature selection approaches, which are Principal Component Analysis (PCA) [28] and Non-negative Matrix Factorization (NMF) [29]. PCA is one of most popular statistical techniques which relates factor analysis with multivariate analysis. This algorithm aims to represent the characteristics in a dataset as a small set of factors or a dataset that keeps important information. Furthermore, NMF is available for multivariate analysis. This algorithm based on linear algebra makes complex feature information into smaller non-negative matrices. Generally, PCA tends to group both positively and negatively correlated components; on the other hand, NMF divides factors into positive vectors. These kinds of statistical factor analyses access the linearity constraint, so we applied DAE techniques with non-linear calculations for obtaining better classification results.
We applied two DAE models. One was a normal autoencoder model (AE) [30,31] and the other was a variational autoencoder model (VAE) [32,33]. Both of the two autoencoder models were constructed using the Pytorch deep learning library in Python [34]. The structure of an AE is divided into two main parts: encoder and decoder. The encoder has an input layer with three fully connected hidden layers with 2000, 500 and 100 nodes. The decoder is comprised of two hidden layers, which are fully connected. The details of the autoencoder are explained in Appendix A.
Furthermore, similar with the AE structure, VAEs also have an encoder and a decoder. The VAE is normally used in semi-supervised learning models nowadays. Additionally, it can learn approximate inferences and be trained using the gradient descent method. The main approach of the VAE is using probability distribution to obtain samples which match the distribution. The details of the variational autoencoder are explained in Appendix B.

2.2.2. Hematopoietic Cancer Subtype Classification Model

For the subtype classification on hematopoietic cancer, we generated a feed-forward neural network (NN) which has one input layer, one hidden layer with 100 nodes and one output layer. The features we extracted from the DAE utilized the input of the NN. In here, we adopt R e L U and s i g m o i d for non-linear activation function. This NN has the two loss functions for measuring classification error, which is the error between the real value and the predicted value. The details of the NN are shown in Appendix C.
The two loss functions are cross-entropy loss (CE) and focal loss (FL). CE is the most widely used loss function in classification models. However, if there are plenty of class labels with the imbalance status, it incurs a loss with non-trivial magnitude. If there are a lot of easy examples to be classified, which means there is a large class imbalance on the dataset, the CE of major class labels is going to be larger than the minor classes.
For handling this class imbalance problem, a modulating factor is added to CE, defined as focal loss (FL). This modulating factor in focal loss is ( 1 p i ) γ . Here, γ is a focusing factor which can be a changeable parameter that ranges γ 0 . Due to this advantage, using FL can prevent the overfitting problem which can accompanied by class imbalance. The details of CE and FL are described in Appendix D.

2.2.3. Training the Models

The loss functions on autoencoder models are calculated by using the difference between the input and the output. The formulas of each loss function are shown below:
L D A E ( i n p u t ,   r e c o n s t r u c t i o n i n p u t ) = M e a n   S q u a r e d   E r r o r L N N ( h i d d e n e n c o d e 3 ,   o u t p u t ) = { C r o s s   E n t r o p y   l o s s F o c a l   l o s s L ( i n p u t ,   o u t p u t ) = L D A E + L N N
In this experiment, we adopted the Adam optimizer [35] for updating the weight and bias iterative based in the training data. This approach has some merits. One is that the step size does not affect the gradient rescaling. Another is that Adam can refer to previous gradients for updating the step size. We performed several trials for defining the learning rate and set it to 0.0001. The batch size of the training set was 128, with the maximum epoch size set to 3000 with an early stopping approach for checking the optimal epoch number.

3. Experimental Results

3.1. Results of Feature Selection

We extracted key features on hematopoietic cancer gene expression data by using DAE approaches. To compare this result with traditional feature selection algorithms, we used PCA and NMF methods. For this comparison, we coded PCA and NMF algorithms using a Scikit-learn [36] and DAE model using a Python deep learning library, Pytorch [34].
For pair comparison, we used t-Distributed Stochastic Neighbor Embedding (tSNE) [37] which is used in the conversion of high-dimensional data visualization into low-dimensional embedding. It converts high-dimensional Euclidian distance between data points into conditional probabilities for mapping low-dimension space and adopts KL-divergence [38] to minimize mismatch on the low-dimensional data representation. Using this technique, researchers can acquire more interpretable data visualization on high-dimensional data. In this experiment, we used the tSNE technique for mapping data into a two-dimensional plane (dimension X, Y) for DAE feature selection, and the visualization of the extracted features of each approach is shown in Figure 2.
Among the visualizations of the selected gene expression data using several approaches, we can see that the result of the DAE-based AE model is the most clearly well distinguished compared with other results.

3.2. Results of DAE Training Process

We extracted key features on hematopoietic cancer gene expression data by using DAE approaches which consist of AE and VAE. Each DAE approach was trained with 3000 epochs for each iteration. We also calculated the loss functions MSE, CE and FL. During the feature extraction process, we calculated the MSE from the AE model, and classification loss (CE and FL) was calculated on the model training class. Furthermore, the total loss was generated by merging MSE and classification loss. Figure 3 and Figure 4 show the loss function graphs for the cancer classification using MSE as the reconstruction error, CE and FL as the classification error and total error as the merged MSE and CE on AE and VAE, respectively. Figure 5 and Figure 6 show loss function graphs using MSE as the reconstruction error, FL as the classification error and total error as the calculated sum of MSE and FL on AE and VAE, respectively.

3.3. Hematopoietic Cancer Subtype Classification Model Interpretation

The Shapley Additive exPlanations (SHAP) method [39], which is based on game theory [40] and local explanations [41], is often used to describe a model’s output. The top 20 important bio-markers for hematopoietic cancer subtype classification, which were derived from the autoencoder, are shown in Table 2 and in Figure 7 as a SHAP summary plot.
As shown above in Figure 7, the most important bio-marker for hematopoietic cancer subtype classification is Ring Finger Protein 130 (RNF130). Including RNF130, the top 20 bio-markers are well known as oncogenes and some bio-markers are directly related to hematopoietic processes. For example, RNF130 is usually related to pathways on the innate immune system and Class I MHC (Major Histocompatibility Complex)-mediated antigen processing and presentation. This bio-marker is related to growth factor withdrawal-induced apoptosis of myeloid precursor cells [42]. Another example is Breast Cancer Anti-Estrogen Resistance Protein 1, Crk-Associated Substrate (BCAR1). Overexpression of BCAR1 is usually detected in many cancers such as breast cancer, lung cancer, anaplastic large cell lymphoma and chronic myelogenous leukemia [43].

3.4. Hematopoietic Cancer Subtype Classification Model Evaluation

For evaluating the hematopoietic cancer subtype classification, we used six classification performance metrics: accuracy (Acc), precision (Pre), recall (Rec), the harmonic mean of precision and recall, which is called F1-measure (F1), geometric mean (G-mean, GM) and index of balanced accuracy (IBA, α = 0.1 ). Furthermore, we generated a confusion matrix and a precision–recall curve (PR-curve) which were used for evaluating the imbalanced data classification performance. The equations below indicate the classification performance measurement.
A c c u r a c y   ( A c c ) = T P + T N T P + T N + F P + F N P r e c i s i o n   ( P r e ) = T P T P + F P R e c a l l   ( R e c ) = S e n s i t i v i t y = T N T N + F N S p e c i f i c i t y   ( S p e ) = T N T N + F P F 1 m e a s u r e   ( F 1 ) = 2 × R e c × P r e R e c + P r e = 2 × T P 2 × T P + F P + F N G e o m e t r i c   m e a n   ( G M ) = R e c × S p e I B A = { 1 + α × ( R e c S p e ) } × G M
where TP, TN, FP and FN are the acronyms of true positive, true negative, false positive and false negative, respectively. TP and TN are the numbers of subtypes correctly classified into positive class or negative class, respectively; FP represents the number of instances of incorrect classification into positive class. Similarly, FN is the number of instances of incorrect classification into negative class.
Furthermore, for verifying the DAE-based cancer subtype classification models, we compared these models with traditional statistics and machine learning-based classification algorithms such as logistic regression (LR) [44,45], random forest (RF) [46,47], k-nearest neighbor (KNN) [48], artificial neural network (ANN) [49] and support vector machine (SVM) [50]. For the class imbalance problem on the classification task, we used an oversampling algorithm named the synthetic minority oversampling technique (SMOTE) [51,52].
The overall designed flowchart of this experiment is shown in Figure 8.
Table 3, Table 4, Table 5, Table 6 and Table 7 represent the results of all combinations of the experiments, which include statistical approaches (PCA, NMF) and deep learning approaches (AE, VAE). Each results table includes SMOTE for handling class imbalance on the dataset. The CE, FL, RE, TOC and TOF keywords in the loss function column represent cross-entropy, focal loss, reconstruction error, cross-entropy + reconstruction error and focal loss + reconstruction error respectively.
The results of hematopoietic cancer subtype classification using the logistic regression classification algorithm are shown in Table 3. The result of AE using merged loss, which combined cross-entropy loss and reconstruction loss, with SMOTE shows the highest results on accuracy (96.33%), recall (91.73%) and F1-measure (91.99%). In addition, the result of AE using reconstruction error with SMOTE shows the highest result on G-mean (95.33%) and IBA (95.24%).
Table 4 shows the results of hematopoietic cancer subtype classification using the k-nearest neighborhood classification algorithm. The result of AE using merged loss, which combined cross-entropy loss and reconstruction loss, with SMOTE shows the highest results on accuracy (96.06%), recall (93.82%) and F1-measure (91.31%), and the result of AE using merged loss, which combined focal loss and reconstruction error, with SMOTE shows the highest results on specificity (99.12%), G-mean (96.59%) and IBA (92.84%).
The results of hematopoietic cancer subtype classification using the random forest classification algorithm are shown in Table 5. The result of AE using merged loss, which combined with cross-entropy loss and reconstruction loss, with SMOTE shows the highest results on accuracy (96.60%), and F1-measure (91.82%), and the result of AE using reconstruction error with SMOTE shows the highest results on G-mean (99.12%), and IBA (90.80%).
In Table 6, the results of hematopoietic cancer subtype classification using the support vector machine classification algorithm are shown. The result of AE using merged loss, which combined focal loss and reconstruction loss, with SMOTE shows the best results on the evaluation matrices of accuracy (97.01%), precision (92.68%), recall (94.60%), specificity (99.52%), F1-measure (93.54%), G-mean (97.87%) and IBA (95.46%).
In Table 7, the results of hematopoietic cancer subtype classification using the artificial neural network classification algorithm are shown. The result of AE using merged loss, which combined cross-entropy loss and reconstruction loss, with SMOTE shows the highest results on accuracy (96.74%), precision (93.62%) and F1-measure (92.47%), and the result of AE using reconstruction error without the oversampling method shows the highest results on G-mean (95.39%) and IBA (90.32%).
Combining all of above results, the top 10 results based on F1-measure are summarized in Table 8. As a summary of this experiment, we found that the result of autoencoder feature selection based on the support vector machine classification algorithm with the total loss, which combined focal loss and reconstruction loss, with SMOTE showed the best performance in accuracy (97.01%), recall (94.60%), specificity (99.52%), F1-measure (93.53%), G-mean (97.87%) and IBA (95.46%).
Figure 9, below, shows the top six PR-curves and Figure 10 shows the top six confusion matrices among the results shown in Table 8.

4. Discussion

In this paper, we suggested an autoencoder-based feature extraction approach for hematopoietic cancer subtype classification. The five major hematopoietic cancer subtypes were selected to create experimental data based on gene expression level. We also compared our approach with traditional feature extraction algorithms, PCA and NMF, which are widely used in cancer classification based on gene expression data. In addition, in consideration of the class imbalance problem occurring in multi-label classification, we applied the SMOTE oversampling algorithm.
In the experimental results, the traditional feature selection approaches, NMF and PCA, showed good performance, but our proposed DAE-based approach for subtype classification showed a better performance. For example, in the results of the SVM classifier using the SMOTE oversampling method, the PCA and NMF feature extraction approaches showed 90.63% and 90.22% accuracy, respectively, and the AE-based feature extraction approaches with cross-entropy error (CE), reconstruction error (RE) and merged error (CE + RE) showed 93.34%, 96.06% and 96.88% classification accuracy, respectively. This result was also the same when focal loss was applied instead of cross-entropy loss. The accuracy for each focal loss case that applied the AE-based feature extraction approach with focal loss (FL), reconstruction error (RE) and merged error (FL + RE) was 90.08%, 95.65% and 97.01%, respectively. Although SVM showed the best result with merged error (FL + RE) using focal loss, in other cases, we found that the merged error (CE + RE) using cross-entropy error showed the best performance in the other feature extraction approaches. Using those extracted results on classification algorithm, the result of subtype classification using the DAE-based feature selection approach showed better performance than traditional statistics and machine learning feature extraction approaches.
Furthermore, as shown in Table 8, when all of the results were summarized, we found that the AE-based feature extraction approach shows better performance than other feature extraction methods. In addition, when comparing the loss function, the results of applying both the classification error (FL/CE) and the reconstruction error (RE) together showed better performance rather than the single loss function, and the sampling method also showed better results when applying the SMOTE oversampling technique.

5. Conclusions

In this paper, we focused on the autoencoder-based feature extraction method to extract biological information from complex cancer data such as gene expression, clinical data and methylation data. We evaluated the proposed method on TCGA data samples from 2457 patients with hematopoietic cancer: lymphoid leukemia, myeloid leukemia, leukemia nos (not otherwise specified), mature B-cell leukemia and plasma cell neoplasm. To the best of our knowledge, there is no other research work on hematopoietic cancer using deep learning-based feature extraction techniques. We compared the proposed autoencoder-based method to the traditional state-of-the-art algorithms PCA and NMF, as well as another generative deep learning technique, VAE. We provided comprehensive experimental results that show the efficiency of our proposed method.
As shown in the experimental results, the proposed method shows higher performance than the other compared techniques in terms of various evaluation metrics. The proposed method with TOF loss achieved the highest accuracy (97.01%), precision (92.68%), recall (94.60%), specificity (99.52%), F1-measure (93.53%), G-mean (97.87%) and index imbalanced accuracy (95.46%) followed by the SVM classifier, which was trained on the sampled data by SMOTE. The learned representations contain rich, valuable information related to hematopoietic cancer which also can be used for other downstream tasks such as regression, classification, survival analysis, etc. We also applied the SHAP feature interpretation technique to our pre-trained model to explain the black box and show the importance of each bio-marker. By extracting bio-markers using deep learning structures, this study is expected to be helpful in enabling gene-specific treatment of patients. Furthermore, it is expected that this model will be helpful in the development of public healthcare through extensibility that can be applied not only to cancer but also to various diseases.
In conclusion, we found that our autoencoder-based feature extraction approach for hematopoietic cancer subtype classification algorithm showed good classification performance in multiclass classification, and our suggested approach showed better performance than PCA and NMF, which are widely used feature extraction methods for cancer classification. Furthermore, the problem of unbalanced data can be solved by applying the SMOTE method.

Author Contributions

K.H.P. conceived and performed the experiments; E.B. improved the experimental approach. Y.P., N.T.-U. and K.H.R. discussed and improved the contents of the manuscript. N.T.-U. provided critical insight and discussion. K.H.R. supervised and funded this work. All authors approved the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and Future Planning (No. 2019K2A9A2A06020672 and No. 2020R1A2B5B02001717).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The work of Kwang Ho Park was supported in part by the Biomedical Engineering Institute, Chiang Mai University.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Autoencoder

In this paper, we used an autoencoder (AE) and a variational autoencoder (VAE) as non-linear feature extraction methods. The structure of an AE is divided into two main parts: encoder and decoder. The encoder has an input layer with three fully connected hidden layers with 2000, 500 and 100 nodes. The decoder is comprised of two hidden layers, which are fully connected. These layers transpose the weights of the encoding and decoding using the equations below:
h i d d e n e n c o d e 1 = R e L U ( W 1 × i n p u t + b 1 ) h i d d e n e n c o d e 2 = R e L U ( W 2 × h i d d e n e n c o d e 1 + b 2 ) h i d d e n e n c o d e 3 = R e L U ( W 3 × h i d d e n e n c o d e 2 + b 3 ) h i d d e n d e c o d e 1 = R e L U ( W 3 × h i d d e n e n c o d e 3 + b 3 ) h i d d e n d e c o d e 2 = R e L U ( W 2 × h i d d e n d e c o d e 1 + b 2 ) r e c o n s t r u c t i o n i n p u t = s i g m o i d ( W 1 × h i d d e n d e c o d e 2 + b 1 )
where, W 1 , W 2 and W 3 are the weight vectors between each layer. Assume that the input size is N , the size of each W 1 , W 2 and W 3 is N × 2000 , 2000 × 500 and 500 × 100 , respectively. The b 1 , b 2 and b 3 are the bias information for each layer. Rectified Linear Unit ( R e L U ) and s i g m o i d function were applied for non-linear activation functions. The h i d d e n d e c o d e 1 was used as the extracted features of AE. For measuring reconstruction loss, we used MSE as a loss function between the original data ( x i ) and reconstructed data ( x ^ i ). The equation of MSE is shown below.
M S E = 1 N   ( x i x ^ i ) 2
This sequence of autoencoder processes is shown in Figure A1.
Figure A1. The procedure of autoencoder (AE) feature selection method. The encoder consists of three fully connected hidden layers which contain 2000, 500 and 100 nodes and the decoder consists of two fully connected hidden layers which contains 500 and 2000 nodes. This autoencoder structure is evaluated by reconstruction error using MSE measurement.
Figure A1. The procedure of autoencoder (AE) feature selection method. The encoder consists of three fully connected hidden layers which contain 2000, 500 and 100 nodes and the decoder consists of two fully connected hidden layers which contains 500 and 2000 nodes. This autoencoder structure is evaluated by reconstruction error using MSE measurement.
Ijerph 18 02197 g0a1

Appendix B. Variational Autoencoder

Similar to the AE structure, the VAE also has an encoder and a decoder. This VAE is normally used in semi-supervised learning models nowadays. Furthermore, it can learn approximate inferences and be trained using the gradient descent method. The main approach of VAEs is using probability distribution to obtain samples which match the distribution. This VAE calculates objective function using reconstruction loss and Kullback–Leibler divergence (KLdivergence) [38].
( X ) = E [ log P ( X | Z ) ] D K L [ Q ( Z | X ) | | P ( Z ) ]
where is a VAE objective function, which is known as a variational lower bound; X is input data and Z is a latent variable. The left term of VAE is the reconstruction loss—it makes the decoder reconstruct input. The right term is referred to as KL-divergence for minimizing the difference between the encoder’s distribution Q ( Z | X ) and prior distribution P ( Z ) . For calculating distribution, it generates mean and standard deviation. The goal of this objective function is to maximize the variational lower bound using the maximization of data generation and minimization the KL-divergence. In this paper, the VAE shared the construction with AE; however, due to the VAE considering the distribution of the input, the VAE applies the distribution calculating part before generating the latent space. Almost the same as the AE, the VAE transposes the weights of the encoding and decoding using the below equations:
h i d d e n e n c o d e 1 = R e L U ( W 1 × i n p u t + b 1 ) h i d d e n e n c o d e 2 = R e L U ( W 2 × h i d d e n e n c o d e 1 + b 2 ) h i d d e n e n c o d e 3 = E [ log P ( X | Z ) ] D K L [ Q ( Z | X ) | | P ( Z ) ] h i d d e n d e c o d e 1 = R e L U ( W 3 × h i d d e n e n c o d e 3 + b 3 ) h i d d e n d e c o d e 2 = R e L U ( W 2 × h i d d e n d e c o d e 1 + b 2 ) r e c o n s t r u c t i o n i n p u t = s i g m o i d ( W 1 × h i d d e n d e c o d e 2 + b 1 )
where, W 1 , W 2 and W 3 are the weight vectors between each layer. Assume that the input size is N , the size of W 1 and W 2 is N × 2000 and 2000 × 500 respectively. In the third layer on the encoder, the distribution of input is computed and the mean ( μ ) and standard deviation ( σ ) of input are generated for calculating latent space ( Z ). The b 1 , b 2 and b 3 are the bias information for each layer. In VAE, R e L U and s i g m o i d functions are also used as non-linear activation functions. The result h i d d e n d e c o d e 1 , called latent space ( Z ), is used as the extracted feature of the VAE. For measuring reconstruction loss, MSE is used as a loss function between the original data and reconstructed data, same as the AE.
This sequence of variational autoencoder processes is shown in Figure A2.
Figure A2. The procedure of the variational autoencoder (VAE) feature selection method. The encoder consists of three fully connected hidden layers which contain 2000, 500 and 100 nodes and the decoder consists two fully connected hidden layers which contain 500 and 2000 nodes. Unlike the autoencoder, this method uses Kullback–Leibler divergence (KL-divergence) for calculating reconstruction loss for variational autoencoder.
Figure A2. The procedure of the variational autoencoder (VAE) feature selection method. The encoder consists of three fully connected hidden layers which contain 2000, 500 and 100 nodes and the decoder consists two fully connected hidden layers which contain 500 and 2000 nodes. Unlike the autoencoder, this method uses Kullback–Leibler divergence (KL-divergence) for calculating reconstruction loss for variational autoencoder.
Ijerph 18 02197 g0a2

Appendix C. Subtype Classification Neural Network

For the NN for the subtype classification on hematopoietic cancer, we generated a feed-forward neural network (NN), which has one input layer, one hidden layer with 100 nodes and one output layer. The features we extracted from DAE utilized the input of NN.
h i d d e n l a y e r = R e L U ( W 4 × h i d d e n e n c o d e 3 + b 4 ) o u t p u t = s i g m o i d ( W 5 × h i d d e n l a y e r + b 5 )
where W 4 and W 5 indicate the weight vector between the layers between 100 × 100 and 100 × C ; C indicates the number of class labels—in this case, it is 5—and b 4 and b 5 are the bias of each NN layer’s node.

Appendix D. Cross-Entropy Loss and Focal Loss

For calculating classification loss, we used two loss functions: cross-entropy loss (CE) and focal loss (FL). CE is the most widely used loss function in classification models. The equation of CE is shown in the below equation:
C E ( p , y )   = i = 1 N y i log ( P ( i | S ) )
where y is the class label of the instance; N is the number of classification labels on the dataset; P ( i | S ) is the predicted probability. If the model predicts an instance correctly as y i , P ( i | S ) heads to the true distribution of y i . As a result, the loss function will be decreased. For simplicity, if p i is defined in the estimated probability of the model when class label y i = 1, it can be written by the below equation:
C E ( p , y ) = C E ( p i ) = log ( p i )
However, if there are plenty of class labels with the imbalance status, it incurs a loss with the non-trivial magnitude. If there are a lot of easy examples to be classified, which means there is a large class imbalance on the dataset, the CE of major class labels is going to be larger than the minor classes.
For handling this class imbalance problem, a modulating factor is added to CE, defined as focal loss (FL). This modulating factor in focal loss is ( 1 p i ) γ . Here, γ is a focusing factor which can be changeable parameter that ranges γ 0 . Therefore, the FL can be formulated as the below equation:
F L ( p , y ) = F L ( p i ) = ( 1 p i ) γ log ( p i )
For instance, if the focusing factor is set as γ = 2 and a sample probability p t = 0.9 , the loss contribution is 100 times lower than CE. On the other hand, the modulating factor makes the loss contribution of the example lower and makes it easier to classify and compare than when using traditional CE. Due to this advantage, using FL can prevent the overfitting problem which can be accompanied by class imbalance.

References

  1. Roizen, M.F. Hallmarks of cancer: The next generation. Yearb. Anesthesiol. Pain Manag. 2012, 2012, 13. [Google Scholar] [CrossRef]
  2. World Health Organization. “Cancer”. Available online: https://www.who.int/cancer/PRGlobocanFinal.pdf (accessed on 12 September 2018).
  3. Anand, P.; Kunnumakara, A.B.; Sundaram, C.; Harikumar, K.B.; Tharakan, S.T.; Lai, O.S.; Sung, B.; Aggarwal, B.B. Cancer is a preventable disease that requires major lifestyle changes. Pharm. Res. 2008, 25, 2097–2116. [Google Scholar] [CrossRef] [PubMed]
  4. Vardiman, J.W.; Thiele, J.; Arber, D.A.; Brunning, R.D.; Borowitz, M.J.; Porwit, A.; Harris, N.L.; Le Beau, M.M.; Hellström-Lindberg, E.; Tefferi, A.; et al. The 2008 revision of the World Health Organization (WHO) classification of myeloid neoplasms and acute leukemia: Rationale and important changes. Blood 2009, 114, 937–951. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Howlader, N.; Noone, A.M.; Krapcho, M.; Miller, D.; Brest, A.; Yu, M.; Ruhl, J.; Tatalovich, Z.; Mariotto, A.; Lewis, D.R.; et al. (Eds.) SEER Cancer Statistics Review, 1975–2017; National Cancer Institute: Bethesda, MD, USA, 2020. [Google Scholar]
  6. Bruneau, J.; Molina, T.J. WHO Classification of Tumors of Hematopoietic and Lymphoid Tissues; Springer: Cham, Switzerland, 2019. [Google Scholar]
  7. Hossfeld, D.K. World Health Organization classification of tumours: Pathology and genetics of tumours of haematopoietic and lymphoid tissues. Ann. Oncol. 2002, 13, 490. [Google Scholar] [CrossRef]
  8. Mak, T.W.; Saunders, M.E.; Jett, B.D. Primer to the Immune Response; Academic Cell: Amsterdam, The Netherlands, 2013. [Google Scholar]
  9. American Cancer Society. Cancer Facts & Figures 2020. Atlanta: American Cancer Society. 2020. Available online: https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/annual-cancer-facts-and-figures/2020/cancer-facts-and-figures-2020.pdf (accessed on 24 February 2020).
  10. Amrane, M.; Oukid, S.; Gagaoua, I.; Ensarİ, T. Breast cancer classification using machine learning. In Proceedings of the IEEE 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; pp. 1–4. [Google Scholar]
  11. Dwivedi, A.K. Artificial neural network model for effective cancer classification using microarray gene expression data. Neural Comput. Appl. 2016, 29, 1545–1554. [Google Scholar] [CrossRef]
  12. Park, K.H.; Ishag, M.I.M.; Ryu, K.S.; Li, M.; Ryu, K.H. Efficient ensemble methods for classification on clear cell renal cell carcinoma clinical dataset. In Proceedings of the Asian Conference on Intelligent Information and Database Systems, Dong Hoi City, Vietnam, 19–21 March 2018; Springer: Cham, Switzerland, 2018. [Google Scholar]
  13. Piao, Y.; Piao, M.; Ryu, K.H. Multiclass cancer classification using a feature subset-based ensemble from microRNA expression profiles. Comput. Biol. Med. 2017, 80, 39–44. [Google Scholar] [CrossRef]
  14. Piao, Y.; Park, K.H.; Ryu, K.H.; Xiang, R. Identification of DNA Methylation Signatures for Diagnosis of Lung Adenocarcinoma. In Proceedings of the 2019 IEEE 10th International Conference on Awareness Science and Technology (iCAST), Morioka, Japan, 23–25 October 2019; pp. 1–5. [Google Scholar]
  15. Cai, Z.; Xu, D.; Zhang, Q.; Zhang, J.; Ngai, S.M.; Shao, J. Classification of lung cancer using ensemble-based feature selection and machine learning methods. Mol. BioSyst. 2015, 11, 791–800. [Google Scholar] [CrossRef] [PubMed]
  16. Ferroni, P.; Zanzotto, F.M.; Riondino, S.; Scarpato, N.; Guadagni, F.; Roselli, M. Breast cancer prognosis using a machine learning approach. Cancers 2019, 11, 328. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Sun, L.; Zhang, X.; Qian, Y.; Xu, J.; Zhang, S. Feature selection using neighborhood entropy-based uncertainty measures for gene expression data classification. Inf. Sci. 2019, 502, 18–41. [Google Scholar] [CrossRef]
  18. Lu, H.; Gao, H.; Ye, M.; Wang, X. A hybrid ensemble algorithm combining AdaBoost and genetic algorithm for cancer classification with gene expression data. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019. [Google Scholar] [CrossRef]
  19. Abdel-Zaher, A.M.; Eldeib, A.M. Breast cancer classification using deep belief networks. Expert Syst. Appl. 2016, 46, 139–144. [Google Scholar] [CrossRef]
  20. Muhamed Ali, A.; Zhuang, H.; Ibrahim, A.; Rehman, O.; Huang, M.; Wu, A. A machine learning approach for the classification of kidney cancer subtypes using mirna genome data. Appl. Sci. 2018, 8, 2422. [Google Scholar] [CrossRef] [Green Version]
  21. Batbaatar, E.; Park, K.H.; Amarbayasgalan, T.; Davagdorj, K.; Munkhdalai, L.; Pham, V.H.; Ryu, K.H. Class-Incremental Learning with Deep Generative Feature Replay for DNA Methylation-based Cancer Classification. IEEE Access 2020, 8, 210800–210815. [Google Scholar] [CrossRef]
  22. Gao, F.; Wang, W.; Tan, M.; Zhu, L.; Zhang, Y.; Fessler, E.; Vermeulen, L.; Wang, X. DeepCC: A novel deep learning-based framework for cancer molecular subtype classification. Oncogenesis 2019, 8, 1–12. [Google Scholar] [CrossRef]
  23. Chen, R.; Yang, L.; Goodison, S.; Sun, Y. Deep-learning approach to identifying cancer subtypes using high-dimensional genomic data. Bioinformatics 2020, 36, 1476–1483. [Google Scholar] [CrossRef] [PubMed]
  24. Parker, J.S.; Mullins, M.; Cheang, M.C.; Leung, S.; Voduc, D.; Vickery, T.; Davies, S.; Fauron, C.; He, X.; Hu, Z.; et al. Supervised risk predictor of breast cancer based on intrinsic subtypes. J. Clin. Oncol. 2009, 27, 1160. [Google Scholar] [CrossRef]
  25. Genomic Data Commons Data Portal. Available online: https://portal.gdc.cancer.gov (accessed on 15 October 2020).
  26. Trapnell, C.; Williams, B.A.; Pertea, G.; Mortazavi, A.; Kwan, G.; van Baren, M.J.; Salzberg, S.L.; Wold, B.J.; Pachter, L. Transcript assembly and abundance estimation from RNA-Seq reveals thousands of new transcripts and switching among isoforms. Nat. Biotechnol. 2010, 28, 511. [Google Scholar] [CrossRef] [Green Version]
  27. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  28. Geladi, P.; Isaksson, H.; Lindqvist, L.; Wold, S.; Esbensen, K. Principal component analysis of multivariate images. Chemom. Intell. Lab. Syst. 1989, 5, 209–220. [Google Scholar] [CrossRef]
  29. Lee, D.D.; Seung, H.S. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems; Neural Information Processing Systems (NIPS): Denver, CO, USA, 2001; pp. 556–562. [Google Scholar]
  30. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Charte, D.; Charte, F.; García, S.; del Jesus, M.J.; Herrera, F. A practical tutorial on autoencoders for nonlinear feature fusion: Taxonomy, models, software and guidelines. Inf. Fusion 2018, 44, 78–96. [Google Scholar] [CrossRef]
  32. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  33. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  34. Ketkar, N. Introduction to pytorch. In Deep Learning with Python; Apress: Berkeley, CA, USA, 2017; pp. 195–208. [Google Scholar]
  35. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  36. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  37. Hinton, G.E.; Roweis, S. Stochastic neighbor embedding. Adv. Neural Inf. Process. Syst. 2002, 15, 857–864. [Google Scholar]
  38. Johnson, D.; Sinanovic, S. Symmetrizing the kullback-leibler distance. Available online: https://scholarship.rice.edu/handle/1911/19969 (accessed on 20 March 2001).
  39. Lundberg, S.; Lee, S.I. A unified approach to interpreting model predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar]
  40. Štrumbelj, E.; Kononenko, I. Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst. 2014, 41, 647–665. [Google Scholar] [CrossRef]
  41. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; Association for Computing Machinery: New York, NY, USA, 2016; pp. 1135–1144. [Google Scholar]
  42. Li, Z.; Wong, K.Y.; Calin, G.A.; Chng, W.J.; Chan, G.C.F.; Chim, C.S. Epigenetic silencing of miR-340-5p in multiple myeloma: Mechanisms and prognostic impact. Clin. Epigenetics 2019, 11, 1–13. [Google Scholar] [CrossRef]
  43. Barrett, A.; Pellet-Many, C.; Zachary, I.C.; Evans, I.M.; Frankel, P. p130Cas: A key signalling node in health and disease. Cell. Signal. 2013, 25, 766–777. [Google Scholar] [CrossRef]
  44. Walker, S.H.; Duncan, D.B. Estimation of the probability of an event as a function of several independent variables. Biometrika 1967, 54, 167–179. [Google Scholar] [CrossRef] [PubMed]
  45. Kleinbaum, D.G.; Dietz, K.; Gail, M.; Klein, M.; Klein, M. Logistic Regression; Springer: New York, NY, USA, 2002. [Google Scholar]
  46. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  47. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  48. Agrawal, R. K-nearest neighbor for uncertain data. Int. J. Comput. Appl. 2014, 105, 13–16. [Google Scholar]
  49. Dreiseitl, S.; Ohno-Machado, L. Logistic regression and artificial neural network classification models: A methodology review. J. Biomed. Inform. 2002, 35, 352–359. [Google Scholar] [CrossRef] [Green Version]
  50. Tang, Y. Deep learning using linear support vector machines. arXiv 2013, arXiv:1306.0239. [Google Scholar]
  51. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  52. Davagdorj, K.; Pham, V.H.; Theera-Umpon, N.; Ryu, K.H. XGBoost-based framework for smoking-induced noncommunicable disease prediction. Int. J. Environ. Res. Public Health 2020, 17, 6513. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overview of proposed autoencoder-based classification approach. We used hematopoietic cancer gene expression data from The Cancer Genome Atlas (TCGA). The deep autoencoder (DAE) model was used to extract deep features from these gene expression data as a lower-dimensional vector. In this study, we use an autoencoder (AE) and a variational autoencoder (VAE) as DAEs. The classifier is used to classify hematopoietic cancer subtypes. We summed the reconstruction loss on DAE and classification loss in the cost function. LR: logistic regression; RF: random forest; SVM: support vector machine; RBF: radial-based function; KNN: k-nearest neighbor; ANN: artificial neural network.
Figure 1. Overview of proposed autoencoder-based classification approach. We used hematopoietic cancer gene expression data from The Cancer Genome Atlas (TCGA). The deep autoencoder (DAE) model was used to extract deep features from these gene expression data as a lower-dimensional vector. In this study, we use an autoencoder (AE) and a variational autoencoder (VAE) as DAEs. The classifier is used to classify hematopoietic cancer subtypes. We summed the reconstruction loss on DAE and classification loss in the cost function. LR: logistic regression; RF: random forest; SVM: support vector machine; RBF: radial-based function; KNN: k-nearest neighbor; ANN: artificial neural network.
Ijerph 18 02197 g001
Figure 2. The extracted feature visualization from hematopoietic cancer gene expression data. (a,b) Visualization of training and test data extracted from principal component analysis (PCA), respectively. (c,d) Visualization of training and test data extracted from non-negative matrix factorization (NMF), respectively. (e,f) Visualization of training and test data extracted from the autoencoder (AE) using t-distributed stochastic neighbor embedding (tSNE), respectively. (g,h) Visualization of training and test data extracted from variational autoencoder (VAE) using tSNE, respectively. LL: lymphoid leukemia; PC: plasma cell neoplasm; NO: leukemia not otherwise specified (nos); ML: myeloid leukemia; MB: mature B-cell leukemia.
Figure 2. The extracted feature visualization from hematopoietic cancer gene expression data. (a,b) Visualization of training and test data extracted from principal component analysis (PCA), respectively. (c,d) Visualization of training and test data extracted from non-negative matrix factorization (NMF), respectively. (e,f) Visualization of training and test data extracted from the autoencoder (AE) using t-distributed stochastic neighbor embedding (tSNE), respectively. (g,h) Visualization of training and test data extracted from variational autoencoder (VAE) using tSNE, respectively. LL: lymphoid leukemia; PC: plasma cell neoplasm; NO: leukemia not otherwise specified (nos); ML: myeloid leukemia; MB: mature B-cell leukemia.
Ijerph 18 02197 g002aIjerph 18 02197 g002b
Figure 3. The training loss graphs on AE-based feature extraction. The mean squared error (MSE) was calculated as reconstruction loss and cross-entropy loss (CE) was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) CE-based classification loss graph for each epoch (blue). The CE value is decreased until 5.32 × 10−13; (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.0001. (c) Total loss (CE + MSE) graph for each epoch (green). The value of total loss is decreased until 0.0001.
Figure 3. The training loss graphs on AE-based feature extraction. The mean squared error (MSE) was calculated as reconstruction loss and cross-entropy loss (CE) was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) CE-based classification loss graph for each epoch (blue). The CE value is decreased until 5.32 × 10−13; (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.0001. (c) Total loss (CE + MSE) graph for each epoch (green). The value of total loss is decreased until 0.0001.
Ijerph 18 02197 g003
Figure 4. The training loss graphs on VAE-based feature extraction. The MSE was calculated as reconstruction loss, and CE was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) CE-based classification loss graph for each epoch (blue). The CE value is decreased until 5.43 × 10−13. (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.00013. (c) Total loss (CE + MSE) graph for each epoch (green). The value of total loss is decreased until 0.00014.
Figure 4. The training loss graphs on VAE-based feature extraction. The MSE was calculated as reconstruction loss, and CE was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) CE-based classification loss graph for each epoch (blue). The CE value is decreased until 5.43 × 10−13. (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.00013. (c) Total loss (CE + MSE) graph for each epoch (green). The value of total loss is decreased until 0.00014.
Ijerph 18 02197 g004
Figure 5. The training loss graphs on AE-based feature extraction. The MSE was calculated as reconstruction loss, and focal loss (FL) was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) FL-based classification loss graph for each epoch (purple). The FL value is decreased until 5.97 × 10−12. (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.001. (c) Total loss (FL + MSE) graph for each epoch (green). The value of total loss is decreased until 0.0001.
Figure 5. The training loss graphs on AE-based feature extraction. The MSE was calculated as reconstruction loss, and focal loss (FL) was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) FL-based classification loss graph for each epoch (purple). The FL value is decreased until 5.97 × 10−12. (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.001. (c) Total loss (FL + MSE) graph for each epoch (green). The value of total loss is decreased until 0.0001.
Ijerph 18 02197 g005
Figure 6. The training loss graphs on VAE-based feature extraction. The MSE was calculated as reconstruction loss, and FL was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) FL-based classification loss graph for each epoch (purple). The FL value is decreased until 5.43 × 10−13; (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.00014. (c) Total loss (FL + MSE) graph for each epoch (green). The value of total loss is decreased until 0.00014.
Figure 6. The training loss graphs on VAE-based feature extraction. The MSE was calculated as reconstruction loss, and FL was calculated by classification loss. The x-axis represents the number of epochs and y-axis represents the loss value. (a) FL-based classification loss graph for each epoch (purple). The FL value is decreased until 5.43 × 10−13; (b) MSE-based reconstruction loss graph for each epoch (red). The MSE value is decreased until 0.00014. (c) Total loss (FL + MSE) graph for each epoch (green). The value of total loss is decreased until 0.00014.
Ijerph 18 02197 g006
Figure 7. The SHAP summary plot for the top 20 important bio-markers for hematopoietic cancer subtype classification.
Figure 7. The SHAP summary plot for the top 20 important bio-markers for hematopoietic cancer subtype classification.
Ijerph 18 02197 g007
Figure 8. The experimental flowchart of hematopoietic cancer subtype classification.
Figure 8. The experimental flowchart of hematopoietic cancer subtype classification.
Ijerph 18 02197 g008
Figure 9. Precision–recall curves (PR-curves) for the top six hematopoietic subtype cancer classification results by F1-measure. (a) PR-curve for the autoencoder using FL + RE with synthetic minority oversampling technique (SMOTE) on support vector machine (SVM) classifier; (b) PR-curve for the autoencoder using CL + RE with SMOTE on SVM classifier; (c) PR-curve for the autoencoder using CL + RE with SMOTE on artificial neural network (ANN) classifier; (d) PR-curve for the autoencoder using FL + RE with SMOTE on ANN classifier; (e) PR-curve for the autoencoder using CL + RE with SMOTE on LR classifier; (f) PR-curve for the autoencoder using CL + RE with SMOTE on RF classifier.
Figure 9. Precision–recall curves (PR-curves) for the top six hematopoietic subtype cancer classification results by F1-measure. (a) PR-curve for the autoencoder using FL + RE with synthetic minority oversampling technique (SMOTE) on support vector machine (SVM) classifier; (b) PR-curve for the autoencoder using CL + RE with SMOTE on SVM classifier; (c) PR-curve for the autoencoder using CL + RE with SMOTE on artificial neural network (ANN) classifier; (d) PR-curve for the autoencoder using FL + RE with SMOTE on ANN classifier; (e) PR-curve for the autoencoder using CL + RE with SMOTE on LR classifier; (f) PR-curve for the autoencoder using CL + RE with SMOTE on RF classifier.
Ijerph 18 02197 g009
Figure 10. Confusion matrix for the top six hematopoietic subtype cancer classification results by F1-measure. (a) represents the confusion matrix for the autoencoder using FL + RE with SMOTE on SVM classifier; (b) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on SVM classifier; (c) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on ANN classifier; (d) represents the confusion matrix for the autoencoder using FL + RE with SMOTE on ANN classifier; (e) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on LR classifier; (f) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on RF classifier.
Figure 10. Confusion matrix for the top six hematopoietic subtype cancer classification results by F1-measure. (a) represents the confusion matrix for the autoencoder using FL + RE with SMOTE on SVM classifier; (b) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on SVM classifier; (c) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on ANN classifier; (d) represents the confusion matrix for the autoencoder using FL + RE with SMOTE on ANN classifier; (e) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on LR classifier; (f) represents the confusion matrix for the autoencoder using CL + RE with SMOTE on RF classifier.
Ijerph 18 02197 g010
Table 1. Summary of the hematopoietic cancer dataset.
Table 1. Summary of the hematopoietic cancer dataset.
Subtype# SampleTrainTest
Lymphoid leukemias (LL)550440110
Myeloid leukemias (ML)818654164
Leukemia nos (NO)1048321
Mature B-cell leukemias (MB)1139023
Plasma cell neoplasms (PC)860688172
Table 2. Top 20 gene/protein bio-markers produced by AE and Shapley Additive exPlanations (SHAP).
Table 2. Top 20 gene/protein bio-markers produced by AE and Shapley Additive exPlanations (SHAP).
NoCodeGene/Protein
1ENSG00000113269.12RNF130
2ENSG00000050820.15BCAR1
3ENSG00000180535.3BHLHA15
4ENSG00000167861.14HID1
5ENSG00000126264.8HCST
6ENSG00000183508.4TENT5C
7ENSG00000173715.14C11orf80
8ENSG00000149054.13ZNF215
9ENSG00000104635.12SLC39A14
10ENSG00000182511.10FES
11ENSG00000134285.9FKBP11
12ENSG00000171067.9C11orf24
13ENSG00000258227.5CLEC5A
14ENSG00000258013.2RPL3P13
15ENSG00000076554.14TPD52
16ENSG00000110665.10C11orf21
17ENSG00000110777.10POU2AF1
18ENSG00000197879.13MYO1C
19ENSG00000177045.7SIX5
20ENSG00000151729.9SLC25A4
Table 3. Evaluation matrices for subtype classification using logistic regression (LR).
Table 3. Evaluation matrices for subtype classification using logistic regression (LR).
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
PCANoNo0.93480.88430.87330.98330.87720.92670.8492
SMOTE0.93610.88740.88930.98670.88520.94930.8946
NMFNoNo0.94160.87290.84530.98500.85150.91250.8210
SMOTE0.94160.87610.90080.98540.88760.94170.8792
AECENo0.65630.38690.39720.85410.33980.51240.2482
SMOTE0.82070.77990.83580.93170.73760.85520.7207
RENo0.94570.92840.83880.98900.85040.93190.8589
SMOTE0.95240.89160.90540.98880.89700.95330.9024
TOCNo0.95520.90650.87380.98610.88580.92150.8385
SMOTE0.96330.92290.91730.98820.91990.95150.8988
FLNo0.48780.39140.29840.85180.24840.50970.2456
SMOTE0.68080.63610.72630.94350.60490.85800.7242
RENo0.93480.89460.82480.97870.83520.91760.8320
SMOTE0.94290.88610.86860.97700.87470.92720.8513
TOFNo0.96330.91030.90790.98860.90880.93040.8559
SMOTE0.95920.89910.92790.98990.91140.95160.8987
VAECENo0.31250.18140.19790.79980.18520.39970.1502
SMOTE0.19840.19540.15600.79740.16180.39510.1467
RENo0.32340.21180.20780.79620.19680.38360.1382
SMOTE0.22150.20670.21490.79980.18930.40160.1517
TOCNo0.35050.26490.22990.79230.22330.39560.1472
SMOTE0.24180.24130.23170.79850.20920.41050.1586
FLNo0.31110.21200.26390.81300.19610.43600.1791
SMOTE0.21060.19710.18760.80000.17520.38160.1366
RENo0.32880.22300.21100.80210.20390.40260.1524
SMOTE0.21740.21740.22090.79410.19240.37590.1326
TOFNo0.31520.19160.20060.80040.18840.40030.1506
SMOTE0.22690.20960.20820.80180.18940.39300.1451
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Table 4. Evaluation matrices for subtype classification using k-nearest neighbor (KNN).
Table 4. Evaluation matrices for subtype classification using k-nearest neighbor (KNN).
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
PCANoNo0.91710.82930.80470.97870.81360.88740.7738
SMOTE0.91710.82250.83170.97940.82480.90250.8025
NMFNoNo0.91170.81270.80490.97720.80400.88690.7730
SMOTE0.91850.82940.86020.97980.84280.91800.8327
AECENo0.93890.88880.82350.98700.83410.91950.8345
SMOTE0.94020.86850.90740.98500.88300.95250.9015
RENo0.93070.89080.82420.98660.83660.91550.8267
SMOTE0.91170.81230.86510.98420.83410.93760.8712
TOCNo0.95380.93060.86370.98730.87880.91440.8243
SMOTE0.96060.89340.93820.98600.91310.94010.8760
FLNo0.91980.85760.79900.97830.78790.87110.7435
SMOTE0.89270.80600.86050.97770.82090.91070.8187
RENo0.94020.88750.84410.98020.84940.89710.7920
SMOTE0.91850.80590.86470.97700.82900.92420.8453
TOFNo0.95240.89370.87580.99000.87730.92700.8488
SMOTE0.94570.85540.93650.99120.88550.96590.9284
VAECENo0.32070.20940.21080.79590.20560.38640.1402
SMOTE0.09180.32920.18430.79910.09090.39460.1463
RENo0.33560.34590.22650.79710.22580.38840.1417
SMOTE0.10870.13450.22580.79660.09970.37010.1284
TOCNo0.31660.26150.21660.79960.21570.39650.1478
SMOTE0.11680.13970.24240.80110.10590.40750.1562
FLNo0.28260.17070.19000.97870.17600.88740.7738
SMOTE0.09510.14750.17970.97940.08550.90250.8025
RENo0.30570.18520.20110.97720.19230.88690.7730
SMOTE0.08830.14730.17100.97980.08130.91800.8327
TOFNo0.29880.16910.18500.98700.17620.91950.8345
SMOTE0.11010.20220.22650.98500.10120.95250.9015
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Table 5. Evaluation matrices for subtype classification using random forest (RF).
Table 5. Evaluation matrices for subtype classification using random forest (RF).
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
PCANoNo0.94160.88360.84440.98690.85840.92320.8417
SMOTE0.95110.89940.88860.98780.89360.93690.8691
NMFNoNo0.94020.90530.82470.98450.83810.90190.8005
SMOTE0.95110.91650.88020.98750.89510.93510.8655
AECENo0.92120.94700.76160.96600.76680.82310.6595
SMOTE0.93890.88570.84390.96860.86200.86370.7313
RENo0.94290.91180.80910.98740.82280.90810.8120
SMOTE0.95920.91140.91070.99010.91090.95610.9080
TOCNo0.95110.93070.83350.98670.85190.90600.8082
SMOTE0.96600.92890.91020.98880.91820.93910.8734
FLNo0.86820.71970.74230.97590.72570.86100.7253
SMOTE0.88860.85690.79900.97740.80550.88740.7739
RENo0.94970.93590.82720.97800.83840.88030.7605
SMOTE0.96200.91230.90210.98530.90670.93360.8628
TOFNo0.94700.90440.83930.98820.85130.91440.8242
SMOTE0.96060.91310.91570.99110.91420.94870.8926
VAECENo0.36960.22680.22580.79360.19380.38640.1403
SMOTE0.30160.16770.18790.79740.17620.39270.1449
RENo0.32610.17640.19720.80230.16180.40630.1552
SMOTE0.31660.18240.20170.79690.19050.39070.1434
TOCNo0.36820.18220.22290.79430.18220.38680.1405
SMOTE0.28940.22160.19130.79780.18770.39530.1468
FLNo0.34650.21010.21040.80490.17650.41080.1587
SMOTE0.28400.20720.18950.79890.18440.39630.1476
RENo0.31110.17350.18970.80670.15880.41460.1617
SMOTE0.32070.38770.21090.80310.20540.40550.1546
TOFNo0.33290.17980.20260.80040.17000.40120.1513
SMOTE0.27450.22270.17770.80770.17270.41710.1637
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Table 6. Evaluation matrices for subtype classification using support vector machine (SVM).
Table 6. Evaluation matrices for subtype classification using support vector machine (SVM).
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
PCANoNo0.86140.87090.61190.96130.62020.76700.5677
SMOTE0.90630.79630.86200.97780.8160.91810.8332
NMFNoNo0.86960.77250.61410.96270.62150.76890.5705
SMOTE0.90220.78060.84530.97780.79940.91040.8181
AECENo0.92530.74920.77220.97460.75890.86450.7318
SMOTE0.93340.86900.89020.97730.87640.92450.8459
RENo0.94570.89940.83490.98870.85270.93780.8708
SMOTE0.96060.91580.91640.98970.91600.96470.9261
TOCNo0.95790.91730.87590.98940.88860.93320.8614
SMOTE0.96880.92560.93090.99110.92810.96010.9162
FLNo0.91710.74450.77200.97670.75620.86630.7349
SMOTE0.90080.83530.86330.97870.84040.91770.8322
RENo0.95380.91210.84850.98140.86300.90210.8013
SMOTE0.95650.89810.89360.98460.89520.93710.8700
TOFNo0.95920.91160.88650.99290.89680.95540.9061
SMOTE0.97010.92680.94600.99520.93540.97870.9546
VAECENo0.32610.16740.19810.79410.16000.38670.1405
SMOTE0.30030.16720.18740.79510.17010.38980.1428
RENo0.32340.13160.19440.79830.15500.39620.1475
SMOTE0.31650.18770.19670.79670.17610.39160.1441
TOCNo0.38720.20270.23520.79610.19260.39070.1434
SMOTE0.3750.20800.23280.79950.20960.39820.1490
FLNo0.31790.22730.19310.80880.15530.41920.1654
SMOTE0.30980.17400.19400.81350.17440.43020.1743
RENo0.34380.33870.20750.80220.16500.40460.1539
SMOTE0.32200.18810.20150.80470.18240.41170.1595
TOFNo0.34380.23720.20800.80490.16900.41120.1590
SMOTE0.33560.19710.20900.80640.19040.41490.1619
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Table 7. Evaluation matrices for subtype classification using artificial neural network (ANN).
Table 7. Evaluation matrices for subtype classification using artificial neural network (ANN).
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
PCANoNo0.94160.86770.84310.98200.84950.91790.8320
SMOTE0.92800.86600.84280.96650.84340.87990.7615
NMFNoNo0.95240.90280.86680.98670.87830.92250.8404
SMOTE0.94430.87280.88990.98510.88070.93440.8645
AECENo0.91300.74230.71410.96770.71840.84210.6926
SMOTE0.94020.87980.90860.96820.88880.91730.8331
RENo0.95920.93670.88650.99140.90280.95390.9032
SMOTE0.95920.93180.89100.98930.90540.94720.8898
TOCNo0.96200.92500.87410.98960.88840.93860.8721
SMOTE0.96740.93620.91510.98800.92470.93530.8658
FLNo0.90220.70050.75890.96750.72750.82670.6656
SMOTE0.87770.80470.84900.97630.81240.91260.8225
RENo0.96200.91800.90580.98460.91110.92840.8525
SMOTE0.95650.90860.89700.98260.90150.92430.8448
TOFNo0.96330.92350.89870.99230.90930.95150.8981
SMOTE0.96330.91650.92680.99060.92100.93850.8719
VAECENo0.31110.19860.20630.79260.20030.37540.1322
SMOTE0.27310.19140.19200.79760.19090.39400.1459
RENo0.30300.18240.19440.79650.18740.38530.1394
SMOTE0.30300.20650.20310.79690.20060.39670.1479
TOCNo0.30300.20980.20050.80540.19790.40720.1559
SMOTE0.33150.23190.24550.80710.24730.41740.1639
FLNo0.29210.19100.19390.80680.18930.41620.1630
SMOTE0.31520.22560.22100.80010.22080.39880.1495
RENo0.26770.17370.17720.80290.17320.40620.1551
SMOTE0.30430.22200.22130.80310.22050.41450.1617
TOFNo0.31930.21270.20910.80640.20520.41140.1592
SMOTE0.29350.19340.19560.80050.19290.42030.1664
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Table 8. Top 10 performances for hematopoietic cancer subtype classification ordered by F1-measure.
Table 8. Top 10 performances for hematopoietic cancer subtype classification ordered by F1-measure.
Feature ExtractionLossSamplingAccPreRecSpeF1GMIBA
SVMAETOFSMOTE0.97010.92680.94600.99520.93540.97870.9546
SVMAETOCSMOTE0.96880.92560.93090.99110.92810.96010.9162
ANNAETOCSMOTE0.96740.93620.91510.9880.92470.93530.8658
ANNAETOFSMOTE0.96330.91650.92680.99060.9210.93850.8719
LRAETOCSMOTE0.96330.92290.91730.98820.91990.95150.8988
RFAETOCSMOTE0.9660.92890.91020.98880.91820.93910.8734
SVMAERESMOTE0.96060.91580.91640.98970.9160.96470.9261
RFAETOFSMOTE0.96060.91310.91570.99110.91420.94870.8926
KNNAETOCSMOTE0.96060.89340.93820.9860.91310.94010.876
LRAETOFSMOTE0.95920.89910.92790.98990.91140.95160.8987
CE = Cross entropy loss, FL = Focal loss, RE = Reconstruction error, TOC = CE + RE, TOF = FL + RE, Bold = The highest score.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, K.H.; Batbaatar, E.; Piao, Y.; Theera-Umpon, N.; Ryu, K.H. Deep Learning Feature Extraction Approach for Hematopoietic Cancer Subtype Classification. Int. J. Environ. Res. Public Health 2021, 18, 2197. https://doi.org/10.3390/ijerph18042197

AMA Style

Park KH, Batbaatar E, Piao Y, Theera-Umpon N, Ryu KH. Deep Learning Feature Extraction Approach for Hematopoietic Cancer Subtype Classification. International Journal of Environmental Research and Public Health. 2021; 18(4):2197. https://doi.org/10.3390/ijerph18042197

Chicago/Turabian Style

Park, Kwang Ho, Erdenebileg Batbaatar, Yongjun Piao, Nipon Theera-Umpon, and Keun Ho Ryu. 2021. "Deep Learning Feature Extraction Approach for Hematopoietic Cancer Subtype Classification" International Journal of Environmental Research and Public Health 18, no. 4: 2197. https://doi.org/10.3390/ijerph18042197

APA Style

Park, K. H., Batbaatar, E., Piao, Y., Theera-Umpon, N., & Ryu, K. H. (2021). Deep Learning Feature Extraction Approach for Hematopoietic Cancer Subtype Classification. International Journal of Environmental Research and Public Health, 18(4), 2197. https://doi.org/10.3390/ijerph18042197

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop