Next Article in Journal
Aerodynamic Characteristics Analysis of Rectifier Drum of High-Speed Train Environmental Monitoring Devices
Next Article in Special Issue
Learning Spatial–Spectral-Dimensional-Transformation-Based Features for Hyperspectral Image Classification
Previous Article in Journal
Capture of CO2 Using Mixed Amines and Solvent Regeneration in a Lab-Scale Continuous Bubble-Column Scrubber
Previous Article in Special Issue
The Fast Detection of Crop Disease Leaves Based on Single-Channel Gravitational Kernel Density Clustering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sample-Pair Envelope Diamond Autoencoder Ensemble Algorithm for Chronic Disease Recognition

1
Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu 610041, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Academy of Chips Technology, China Electronics Technology Group, Chongqing 401332, China
4
School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2023, 13(12), 7322; https://doi.org/10.3390/app13127322
Submission received: 27 April 2023 / Revised: 13 June 2023 / Accepted: 15 June 2023 / Published: 20 June 2023
(This article belongs to the Special Issue Application of Artificial Intelligence in Visual Signal Processing)

Abstract

:
Chronic diseases are severe and life-threatening, and their accurate early diagnosis is difficult. Machine-learning-based processes of data collected from the human body using wearable sensors are a valid method currently usable for diagnosis. However, it is difficult for wearable sensor systems to obtain high-quality and large amounts of data to meet the demands of diagnostic accuracy. Furthermore, existing feature-learning methods do not deal with this problem well. To address the above issues, a sample-pair envelope diamond autoencoder ensemble algorithm (SP_DFsaeLA) is proposed. The proposed algorithm has four main components. Firstly, sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM) is designed to find pairs of samples that are close to each other in a manifold neighborhood. Secondly, the feature-embedding stacked sparse autoencoder (FESSAE) is designed to extend features. Thirdly, a staged feature reduction mechanism is designed to reduce redundancy in the extended features. Fourthly, the sample-pair-based model and single-sample-based model are combined by weighted fusion. The proposed algorithm was experimentally validated on nine datasets and compared with the latest algorithm. The experimental results show that the algorithm is significantly better than existing representative algorithms and it achieves the highest improvement of 22.77%, 21.03%, 24.5%, 27.89%, and 10.65% on five criteria over the state-of-the-art methods.

1. Introduction

Chronic diseases have been the greatest threat to human life in recent decades and it is vital to diagnose and predict them before reducing the rate of fatality [1]. As chronic diseases are long-lasting and have slow progression, the lack of wearable sensor monitoring system is currently limiting disease diagnosis, thus affecting timely treatment of early patients. At present, wearable sensors can be used to continuously collect physiological signals from patients fast and then the obtained data can be processed by machine learning algorithms [2]. The structure of a machine-learning-based chronic disease recognition system is shown in Figure 1. Firstly, physiological data from the human body are collected by the wearable sensors. Then, they are stored in a chronic disease database by transmission. After that, machine learning algorithms are used to process and analyze data to provide diagnosis results back to the patient or hospital. For example, a continuous dataset of vital signs is obtained using optical and pressure sensors, followed by feature extraction and classification to obtain predictions for chronic diseases [3]. Data collected by a set of sensors are fused and then fed into a classifier model for early prediction of heart disease [4]. Currently, many wearable medical sensor systems enable data collection but do not provide high-quality data and timely medical diagnosis. It is well known that feature learning in machine learning has a large impact on recognition results. Therefore, the improvement of feature learning in machine learning algorithms for chronic diseases is important. It is a major motivation of this paper.
Machine learning technology has enabled new tools for recognition of chronic diseases because of its efficiency in data processing. Research on machine learning methods for chronic disease recognition is focused on the following two areas: feature learning and classifier design.
Feature learning is particularly important. Feature learning algorithms mainly include feature selection methods and feature extraction methods. Feature selection is the procedure of selecting the best features from all the features available to distinguish between classes. Feature extraction converts features from a higher dimensional to a lower dimensional space. Some techniques used for the diagnosis of chronic diseases include linear discriminant analysis (LDA), generalized discriminant analysis (GDA), principal component analysis (PCA), and so on.
Classifiers can discover hidden patterns in existing human physiological data. Common classifiers [5,6,7,8,9,10] include the radial basis function network (RBF Network), decision tree (DT), naive Bayesian (NB), logistic regression (LR), functional tree (FT), logistic model tree (LMT), support vector machine (SVM), k-nearest neighbor (KNN), multi-layer perceptron (MLP), random forest (RF), and recurrent neural network (RNN). Ekanayake et al. [11] selected high-quality features and used 11 machine learning methods for model training: KNN, support vector classifier (SVC) with a linear kernel, LR, decision tree classifier, SVC with radial basis function (RBF) kernel, Gaussian NB, RF, a classical neural network, extra trees classifier, Adaboost classifier, and extreme gradient boosting (XGB) classifier.
Motivation: These techniques developed by these studies improve the performance for chronic disease detection. In particular, feature learning can provide high-quality features, thus reducing classifier complexity and improving recognition performance. However, the existing feature learning algorithms did not take into account the characteristics of datasets on chronic diseases well, which typically are (1) small or medium-sized features and samples, usually with no more than 35 features and no more than 1500 samples according to the relevant literature; (2) complex correlation between features and class label (disease status); and (3) a requirement for high recognition accuracy. Deep learning has powerful feature extraction capabilities but cannot obtain sufficiently high quality of features in the condition of small samples. Existing non-parameter or low-parameter feature reduction algorithms, such as PCA and LDA, can considerably reduce the number of features in small samples. However, they use the original features and do not work well when the original features have low quality. Therefore, it is important and challenging to obtain a high-quality feature set in the case of small or medium-sized features and samples.
To address the above issues, this paper proposes a solution–sample-pair envelope diamond autoencoder ensemble algorithm (SP_DFsaeLA). Initially, the sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM) is designed to capture the intrinsic structure of manifold neighboring samples. Then, the feature embedding stacked sparse autoencoder (FESSAE) is designed to extend features. To eliminate redundancy between the expanded features, a nonparametric staged feature reduction mechanism is designed. The first stage uses the L1 regularized feature reduction algorithm and the second stage uses an improved manifold dimensionality reduction algorithm to further reduce the number of features. This mechanism combining SP_EMNCM, feature expansion, and feature reduction is called the sample-pair diamond-like feature learning mechanism. Figure 2a illustrates the flow of non-parameter or low-parameter feature reduction methods, such as PCA, LDA, LPP, and Relief. Figure 2b illustrates the flow of the deep feature learning method. The proposed sample-pair diamond-like feature learning mechanism is illustrated in Figure 2c.
The main contributions of this article are as follows:
  • A diamond-like feature learning mechanism is proposed for the first time in this paper, realizing feature expansion and reduction with a diamond topology. The traditional feature learning algorithm has low adaptivity, and the deep learning algorithm is dependent on a large number of samples. Both cannot meet the requirement for high accuracy from chronic disease recognition with low samples. The proposed diamond-like feature learning mechanism includes feature expansion mechanism and feature reduction mechanism. It first expands features to enhance the representation capability, then reduces features to enhance the generalization capability, thereby reducing the requirement for a large number of samples. In short, this mechanism combines deep learning and traditional feature reduction methods’ advantages together and can better adapt to the data characteristics of chronic diseases and the requirement for high recognition accuracy than the deep learning and traditional feature reduction methods.
  • Existing feature learning methods only focus on the original sample itself and ignore the neighborhood relationship of the sample. This leads to the samples’ features being easily affected by noise because the features are reduced and the relative relationship (e.g., manifold structure) is possibly being broken easily. The SP_EMNCM proposed in this paper is used to mine the sample nearest manifold neighbor structure information to form an envelope-like structure, thus enriching the feature information and improving its representational capacity.
  • Existing deep stack autoencoders learn poor-quality features at small sample sizes and have limited feature complementary fusion performance due to insufficient complementarity. The FESSAE is designed as a lightweight deep network for feature extraction of chronic sensing data. The FESSAE network introduces original features into the training process and structure of the network to improve the complementary nature between the higher quality feature and original feature, thus achieving high-quality deep features in small and medium sample sizes.
  • Existing feature expansion methods do not consider both between-sample structural information and between-feature complementary information in the expansion process, which leads to the introduction of a large number of ineffective features. The proposed feature expansion mechanism includes SP_EMNCM and FESSAE. It considers both the manifold neighbor structure information between samples and the feature complementarity, thereby improving original features richer and better.
  • Existing feature reduction methods do not adequately select and extract features for chronic disease recognition. The proposed staged feature reduction mechanism in this paper first selects the expanded features based on L1 regularization, then reduces the dimensionality of the most important features based on manifold learning, thus making the features more compact without losing useful information.
The following is the organization of the paper: Section 2 discuss some existing machine learning methods for chronic disease recognition. The proposed method is mainly described in Section 3. The experimental results are analyzed and reported in Section 4. Finally, in Section 5, the main contributions and possible limitations are discussed.

2. Related Work

In this section, we first briefly discuss some existing machine learning methods for chronic disease recognition and then discuss their advantages and disadvantages.
Feature learning refers to the process of automatically discovering and extracting meaningful representations or features from original data. Ahmed H et al. [12] used Relief and univariate feature selection to select great features from the dataset. Shrivas et al. [13] introduced a union-based feature selection technique for predicting chronic kidney disease. Chormunge et al. [14] realized a new relevance and cluster-based feature selection method in order to reduce the dimensionality issue in data mining tasks. Sawhney et al. [15] used penalization functions combined with the existing fitness function of the binary firefly method to reduce the feature set and improve cancer classification accuracy. Jayaraman et al. [16] have combined particle swarm optimization and gravitational cuckoo search algorithms for managing the features that exist in heart disease classification systems. Paul A K et al. [17] used weighted least squares to select effective attributes, and Rasitha [18] used LDA to classify hypothyroid disease. Mohamed et al. [19] used PCA to reduce the dimensions of medical data of type 2 diabetes. Shahbazi et al. [20] used GDA to minimize the number of features in the feature space and the overlap of samples. Lu H et al. [21] developed a patient network and machine learning method that combines the attributes of the patient network with sample features. Taghizadeh E et al. [22] chose Analysis of Variance, mutual information, additional tree classifiers, logistic regression for feature selection, and then used PCA again for feature reduction. Khan A et al. [23] understood the pattern of complications of type 2 diabetes and then analyzed codes and their relationships were then analyzed to construct a comorbidity network.
For classifier design, Ge et al. [24] researched a new multi-label neural network to predict chronic diseases. El-baz et al. [25] proposed a combinative classifier based on the KNN classifier in the prediction module of a hybrid intelligent system for breast cancer tumor recognition. Polat [26] used the SVM, KNN, RF, and LDA methods to classify medical databases after attribute weighting. Cheruku et al. [27] introduced a new hybrid decision support to a bat optimization method and a system based on rough set theory. In this hybrid system, the redundant feature is effectively reduced by generating fuzzy rules. Maniruzzaman et al. [28] proposed a Gaussian-based classification model for diabetes and investigated the performance of a Gaussian process using three kernels. Alhassan and Zainon [29] presented a deep-belief network for heart disease diagnosis. Abdollahi J et al. [30] used 10 machine learning algorithms as the basic algorithms in a stack generalization algorithm to predict chronic diseases and implemented a hybrid meta-algorithm for prediction. Fatan M [31] used advanced fusion techniques, deep learning segmentation methods, and survival analysis to automatically segment tumor and predict survival outcome in head-and-neck squamous cell carcinoma cancer. Rezaeijo S M [32] used hierarchical clustering to improve the validity of mpMR image-based for prostate tumor classifications.
In Table 1, the advantages and disadvantages of the method proposed in this paper and previous methods are listed.
In general, these researches have improved the performance of chronic disease detection. In particular, feature learning can provide high-quality features to improve recognition performance. However, existing feature learning algorithms do not take into account the characteristics of chronic disease datasets well. So, this paper proposed a diamond-like feature learning mechanism, realizing feature expansion and reduction with a diamond topology. It first expands features to enhance the representation capability, then reduces features to enhance the generalization capability, thereby reducing the requirement for a large number of samples.

3. Materials and Methods

3.1. Problem Formulation

Suppose a chronic disease dataset X = { ( x i , y i ) } i = 1 N , where x i is ith sample and y i is its corresponding label. The sample vector can be expressed as x i = { x i ( 1 ) , x i ( 2 ) , , x i ( M ) } , where M is the number of features. Through feature learning operator (method) ϕ ( ) , x i is transformed to x ^ i = ϕ ( x i ) by feature learning. Traditional feature learning algorithms reduce the dimensionality to obtain x ˜ i = ϕ t r ( x i ) and then train the classifier C to obtain prediction results. The traditional non-parameter or low-parameter feature learning algorithms are poorly adaptive and their feature extraction capability is unsatisfactory when the original features are complex. Deep learning method ϕ d ( ) is conducted to obtain reduced dimensional features x i = ϕ d ( x i ) . Deep learning has powerful feature extraction capabilities but suffers from a small-sample-size problem. To solve this problem above, the SP_DFsaeLA algorithm is proposed to find the optimal ϕ ( ) , thereby improving subsequent classification accuracy.

3.2. Proposed Algorithm’s Framework

In this section, we introduce the proposed algorithm. Sample-pair diamond stacked sparse autoencoder ensemble learning algorithm (SP_DFsaeLA) is developed to better recognize chronic diseases according to the characteristics of chronic disease data. This algorithm consists of four parts, as illustrated in Figure 3. First, the sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM) is designed by searching the manifold nearest samples in manifold neighborhood and generating sample pairs. Second, the feature embedding stacked sparse autoencoder (FESSAE) is designed to extend features. Third, a staged feature reduction mechanism is designed to remove the feature redundancy. It includes L1 regularization and weighted locality preserving discriminant projection (L1_wLPPD). This mechanism for achieving feature expansion followed by reduction is called the diamond-like feature learning mechanism (DFLM), as show in Figure 4. Fourth, a sample-pair-based model is constructed for SP_EMNCM and DFLM and a single-sample-based model is constructed for single sample and DELM. The sample-pair-based model and single-sample-based model are combined by weighted fusion.
The main symbols used in this paper are listed with their meanings in Table 2.

3.3. Sample-Pair Envelope Manifold Neighborhood Concatenation Mechanism (SP_EMNCM)

SP_EMNCM is designed to catch the inherent structure of similar manifold neighboring samples in the sample envelope space. It is known that geodesic distances can reveal similar relationships between samples located on nonlinear manifolds. Therefore, we introduce geodesic distance as the shortest distance into sample envelope space. The geodesic distance between the target sample and all samples is calculated, where the sample with the smallest geodesic distance is called the manifold nearest sample. Then, the original samples are concatenated with the manifold nearest samples to generate sample pairs.
Given a multi-dimensional sample envelope space R N M , containing N sample vectors { x 1 , x 2 , . . . , x i , . . . , x n } , each sample vector can be expressed as x i = { x i ( 1 ) , x i ( 2 ) , . . . , , . . . , x i ( m ) } . y ( x i ) is the class label of x i . d e ( x i , x j ) denotes the Euclidean distance between x i and x j . The process of searching for the manifold nearest sample is as follows:
  • Divide the sample by class label. If y ( x i ) = y ( x j ) , x i and x j are divided into the same sample set.
  • Determine the neighborhood relationships. Compute distance matrix D l from neighborhoods about each sample using Euclidean distance in the same class sample set.
  • Compute shortest distances. If the nearest neighbor graph has edge ( x i , x j ) , the shortest distance is d e ( x i , x j ) . Otherwise, d g x i , x j = + . The shortest distance is defined as:
    d g x i , x j = d e ( x i , x j ) min { d g x i , x j , d g x i , x k + d g ( x k , x j ) } I f   x i   a n d   x j   a r e   n e i g h o r s o t h e r s
Then, obtaining the shortest distance matrix D g = [ d g ( x i , x j ) ] N × N . The sample with the smallest distance value in each row of D g is retained as the manifold nearest sample.
After the manifold nearest samples are searched, original samples X i and the manifold nearest samples X j are concatenated into sample pairs X p a i r = X i , X j . The sample pair generation process is shown in the following pseudo code of Algorithm 1 and Figure 5.
Algorithm 1: Sample-pair envelope manifold neighborhood concatenation mechanism
Input: Dataset D = ( X , Y )
1: Data set D classified by category according to sample labels
2: For  i = 1 : N
3:     Divide the sample by class label
4:     Calculate shortest distances between this sample and all other samples of its class
5:     Sort all shortest distances thus obtained
6:     Select the manifold nearest sample x j to x i
7:     Concatenate samples x i and x j to generate sample pair x p a i r
8: End For
Output:  D p a i r = ( X p a i r , Y p a i r )
SP_EMNCM is different from existing methods in that the latter are based on original samples and do not consider structural information between samples. SP_EMNCM combines the closest samples and extracts information between two samples by expanding the features. Figure 6 illustrates the differences between the single-sample-based model and the proposed sample-pair-based model. Seen from Figure 6, the sample-pair-based model not only considers the original samples, but also the manifold neighborhood relationship between the samples.

3.4. Diamond-like Feature Learning Mechanism (DFLM)

3.4.1. FESSAE-Based Feature-Expansion Mechanism

In order to extend the features, an improved FESSAE was designed. FESSAE is a lightweight deep network.
FESSAE is improved compared to a stacked sparse autoencoder (SSAE). The FESSAE model is shown in Figure 7.
The key element of the FESSAE is the embedding combination between two adjacent autoencoders. Let H ( k ) = [ h 1 k , h 2 k , , h N k ] denote the k-th hidden layer’s output matrix and X ( O ) R N M the sample input to the first layer (original input) of the FESSAE. Firstly, the output of the previous hidden layer H ( k 1 ) is added to the original input X ( o ) to obtain combined feature as follows:
E ( k ) = [ ( X ( o ) ) T ; H ( k 1 ) ]
Then, the combined feature E ( k ) is transformed as follows:
L E k = G T E ( k )
where G is the appropriate sparse transformation matrix. The objective function of the feature-embedded unit is defined as follows to filter partially redundant features:
m a x G t r ( G T E ( k ) ( E ( k ) ) T G ) s . t . G i j = d
where d is the number of high-quality features retained in feature extraction. After processing by the feature-embedded unit, the k-th autoencoder’s input data are L ( E ( k ) ) .
The output data of the hidden layer are divided into two groups, and the ratio of the two groups of features is kept consistent with the ratio of the two types of data of the encoder for that layer. The hidden-layer output features are expressed as H = [ H Γ 1 , H Γ 2 ] . The addition group sparsity constraint is represented as follows:
ψ H = g = 1 2 | | H Γ g ( k ) | | 1
After introducing embedding elements and sparse constraints in the structure during training, the objective function of the k-th autoencoder (k > 1) of the FESSAE is expressed as follows:
a r g θ min 1 N i = 1 N | | L ( E ( k ) ) L ( E ( k ) ) | | 2 + λ ( | | W k 1 | | 2 + | | W k 2 | | 2 ) + β ( j = 1 d ( k ) K L ( ρ | | ρ j ) + g = 1 2 | | H Γ g ( k ) | | 1 )
where L ( E ( k ) ) is the k-th autoencoder’s output. λ and β denote the regularization coefficient and sparsity coefficient, respectively. ρ denotes the sparse parameter. The numerical value increases monotonously as the difference between ρ and ρ j increases.
The proposed FESSAE is outlined in Algorithm 2.
Algorithm 2: FESSAE-based feature expansion algorithm
Input: Samples X ( o )
1: Set parameters: λ , β , ρ , d ( k ) , number of iterations.
2: Pretraining:
3:    Train the first layer of FESSAE and extract the hidden layer’s output H(1) 
4:    For k = 2:K
5:       Calculate transformation matrix
6:       Calculate the output of the feature-embedded unit L ( E ( k ) ) by Equations (1)–(2)
7:       Train k-th layer with objective function in Equation (5)
8:       Extract output H(k) of k-th hidden layer
9:    End For
10: End Pretraining
11: Stack hidden layers and softmax layer
12: Fine-tune entire network
13: Get the output of the last hidden layer as a deep features
Output: Deep features

3.4.2. L1_wLPPD-Based Feature Reduction Mechanism

The feature-reduction mechanism consists of two stages: L1 regularization and weighted locality preserving discriminant projection.
The first stage used is L1 regularization, a common feature selection method, which obtains a sparse feature vector. The optimization objective function is as follows:
arg min θ i = 1 N ( y i m = 1 M ^ θ m x i m ) 2 + α m = 1 M ^ | θ m |
where the first term is original error term and the second is L1 regularization term. α denotes regularization factor. θ m is m-th feature’s regression coefficient. Equation (6) is optimized using the proximal gradient-descent method, and each gradient-descent iteration is:
θ ( k + 1 ) = arg m i n θ C 2 | | θ z | | 2 2 + α | | θ | | 1
where z = θ ( k ) C 1 f ( θ ( k ) ) and f θ k = i = 1 N ( y i ( θ ( k ) ) T x i ) 2 . C is a constant greater than zero, and the components of θ = [ θ 1 , θ 2 , . . , θ m , . . , θ ( M ^ ) ] are independent of each other. The solution of the L1 regular expression obtained using the iterative soft-threshold function can be expressed as follows:
θ ( k + 1 ) = s o f t α C 1 ( θ ( k ) C 1 f ( θ ( k ) ) ) = s i g n ( θ ( k ) ) ( | θ ( k ) | α C )
where s i g n ( ) is a symbolic function. The corresponding non-zero component features in θ ( k + 1 ) are selected as the feature subset, that is, features selected by L1 regularization.
The second-stage feature reduction is based on a weighted locality-preserving discriminant projection (wLPPD). X ~ denotes the samples after the first stage of reduction, p is the total number of classes. After sampling, the total number of samples is N s and the number of samples for the p-th class is N s p . The inter-class variance of the N s nearest-neighbor samples of the local sample center of the sample set X ~ is S B , defined as:
S B ϕ = p = 1 P B p B p T
where B p is the difference between the local sample center of the p-th class and the local sample center, that is, B p = ( 1 / N s p ) i = 1 N s p x ~ i ( p ) ( 1 / N s ) i = 1 N s x ~ i . The intra-class variance matrix of the N s p nearest-neighbor samples of the class center of the p-th class sample is S W , defined as:
S W ϕ = p = 1 P i = 1 N s p i p ( i p ) T
where i p is the difference between the i-th sample of class p and the local sample class center of class p, that is, i p = x ~ i ( p ) ( 1 / N s p ) i = 1 N s p x ~ i ( p ) .
The regularization term is as follows:
J ( A ) = i = 1 N j = 1 N A T x ~ i A T x ~ j 2 W i j = T r ( i = 1 N j = 1 N ( 2 A T x ~ i x ~ i T A 2 A T x ~ i x ~ j T A ) W i j ) = T r ( i = 1 N A T x ~ i D i i x ~ i T A i = 1 N j = 1 N W i j A T x ~ i x ~ j T A )
where D i i = j = 1 N W i j is the diagonal matrix. W is the affinity matrix. Equation (11) also can be written as T r ( A T X ~ ( D W ) X ~ T A ) . Setting L = D W , the locality preservation regularization term can be expressed as:
J A = T r ( A T X ~ L X ~ T A )
Using Equations (9)–(12), the proposed wLPPD can be expressed as:
min A T r A T S W ϕ A s . t . T r A T S B ϕ A γ J A = κ I
where κ is a constant and γ represents the Lagrange penalty factor. Through adding the Lagrange multiplier η , Equation (13) is rewritten as:
L A , η = T r A T S W ϕ A η ( T r A T S B ϕ A γ J A κ I )
Taking the partial derivative of Equation (14) and setting L ( A , η ) A = 0 , the result is:
S W ϕ A ( S B ϕ γ X ~ L X ~ T ) = η A
After obtaining the projection matrix A , we take the top l eigenvectors corresponding to the biggest eigenvalues of A to obtain projection matrix A l . Using the LPPD in each subspace separately, we can obtain T projection matrices: A l 1 , A l 2 …, A l T . The final projection matrix A l E is obtained through the weighted integration of A l t in each subspace projection transformation.
In addition, an ensemble learning method is adopted for the fusion mechanism. In particular, the sample sampling rate is set to δ S and the sampling rate of the features to δ F , and the mixed-feature dataset is sampled according to the bagging strategy q times, forming q subsets. The proposed L1_wLPPD staged feature-reduction mechanism is then summarized in Algorithm 3.
Algorithm 3: L1_wLPPD-based feature-reduction algorithm
Input: Data X ^ after feature expansion
1: First-stage feature reduction
2:    Hybrid feature selection for X ^ based on Equations (6)–(8)
3: End first-stage feature reduction
4: Second-stage feature reduction
5:     δ S = 0.7 , δ F = 0.5 .
6:    Sample q times to form q subsets.
7:    Train q-th SVM:
8:    For i = 1:T
9:      Choose ns train samples randomly
10:      Calculate scatter matrixes S B and S W using Equations (9)–(10)
11:      Calculate diagonal matrix D and the Laplacian matrix L
12:      Solve mapping matrix A using Equation (15)
13:    End For
14:    Obtain final mapping matrix A l E
15:    Map q-th subset to train SVM.
16: End second-stage feature reduction
17: Obtain the ultimate class label
Output: Predicted result

4. Experimental Results and Analysis

Three sets of experiments are carried out to verify the effectiveness of the algorithm in this paper. The first experiment is based on the ablation method to verify the effectiveness of the SP_EMNCM and DFLM in the algorithm and to verify the innovation points of the algorithm. The second experiment compares the proposed algorithm with existing representative feature learning algorithms, representative autoencoders, and the state-of-the-art algorithms to verify the proposed methods. The third experiment analyzes the effects of some important parameters including coefficient for the L2 weight regularization, sparsity regularization coefficient and sparse parameter, the type of classifier, and the number of classifiers.

4.1. Experimental Conditions

The proposed algorithm’s performance is tested on several relevant datasets. Two publicly available chronic disease datasets (Pima Indians Diabetes and Statlog Heart Data Set) are selected for the experiments; they include data on cardiovascular diseases and diabetes, which are the two major chronic diseases [33,34]. The basic information on the datasets is presented in Table 3. The Statlog Heart Data Set and Pima Indians Diabetes are representative datasets for diabetes and heart disease, respectively. In addition, Table 3 includes some other chronic diseases, including Parkinson’s and Alzheimer’s.
We set three AEs in the proposed FESSAE. The number of hidden units (neurons in hidden layer) is selected by considering dimensional range of the feature vectors for different datasets, and a grid search is used to find the best values. The adjustable parameters in the proposed FESSAE include coefficient for regularization and sparsity parameters. The relevant parameters are listed in Table 4, including three parameters of tuned FESSAE objective function and the number of FESSAE hidden layer’s neural units for each dataset.
In the experiments, the K-fold cross-validation technique was employed to evaluate the performance of the SP_DFsaeLA. The programing tool is R2018b MATLAB.

4.2. Evaluation Criteria

Accuracy (Acc), sensitivity (Sens), precision (Prec), F1_score, and specificity (Spec), which are calculated from the confusion matrix, are chosen as metrics to assess the effectiveness of the model. The confusion matrix or error matrix is a useful tool for visualizing the classifier’s whole performance. These datasets used in this study are multi-classified and binary, with a 2 × 2 confusion matrix for binary and n × n confusion matrix for multi-classification. The Acc, Sen, Pre, Spec, and F1_score of each category can be expressed as follows:
A c c = T P + T N T P + F P + F N + T N
P r e c = T P T P + F P
S e n s = T P T P + F N
S p e c = T N T N + F P
F 1 s c o r e = 2 ( P r e c × S e n s ) P r e c + S e n s

4.3. Ablation Experiments

4.3.1. Effectiveness Analysis of SP_EMNCM

In order to validate the efficacy of SP_EMNCM, we conducted a comparative analysis between single-sample-based model and the sample-pair-based model. The two models are illustrated in Figure 6. To eliminate the influence of other factors, such as different feature learning methods, the two methods are compared directly on the SVM and RF classifiers. Within Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10, the designation “SS” refers to a single-sample-based model, while “PS” refers to sample-pair-based model. The results are compared in Table 5 and Table 6. Table 7 and Table 8 compare the results obtained with the LDA. Results with the FESSAE are compared in Table 9 and Table 10.
As seen from Table 5 and Table 6, all the sample-pair-based methods performed better than the single-sample-based method for these datasets on the SVM and RF classifiers. This indicates that the SP_EMNCM is more effective than the single-sample-based methods.
As seen from Table 7 and Table 8, most of the sample-pair-based methods are better than the single-sample-based results in these databases with the LDA on the SVM and RF classifiers. Thus, in general, SP_EMNCM is effective for manifold feature reduction.
As seen from Table 9 and Table 10, most of the sample pair results in the dataset are better than single sample results on the SVM and RF classifiers. Thus, in general, SP_EMNCM is effective under deep learning.

4.3.2. Effectiveness Analysis of DFLM

Verification of the performance of the different steps in DFLM is conducted by ablation experiments. First, the sample pairs are input to the classification model as original features (OFs) input. Second, the samples are processed with the FESSAE and the OFs converted into deep features (DFs). Third, OFs and DFs are combined to construct new features (CF). The CF is processed by the L1_wLPPD to become the reduced features (CF and DFLM). Different feature sets are used with the same classifier and dataset. A comparison of the results is shown in Table 11.
Table 11 shows that in some cases the CF and DFLM has better results. This result indicates that the proposed diamond-like feature learning mechanism (DFLM) is effective. The accuracy of DFs also performs well. The results show that FESSAE is effective, and it can learn as many high-quality features as possible and expand features well. The fact that the CF and DFLM performs better than the CF means that the proposed feature reduction mechanism is effective. The accuracy of the CF does not improve as much as that of the OFs. We hypothesize that the reason is that the simple combination of the OFs and DFs leads to high redundancy. Therefore, it needs to consider feature reduction. Different datasets have different results, due to characteristics of data and disease. For example, most metrics of PID, Wisconsin and Heart datasets have higher results in CF and DFLM but, for the AD, PD, lung cancer, and Maxlettle datasets, the metrics have higher results in DFs in most cases. We hypothesize that this is due to the former’s sample size being medium-sized with dimensionality of 200 to 768, but the number of features is small. While the latter’s features are medium-scale with dimension of 20 to 35, the number of samples is large or small. The difference between the number of samples and features possibly is one of the reasons for the different performance of the metrics in different datasets and may also be related to the internal structure of the data.
To visualize the DFLM, the features extracted at the different stages are shown in Figure 8. Here, four datasets (PID, Heart, WDBC, and Wisconsin) are used.
As shown in Figure 8, the change in the number of features tends to increase first and then decrease as the stages of the proposed method progress. The distribution of number of features in the stages appear as a diamond topology. The number of features increased after combining the OFs and DFs. After the first L1_wLPPD stage, the expanded features are reduced. The second L1_wLPPD stage further reduces the number of features. Finally, high-quality features are obtained for small or medium size of samples and features.

4.4. Algorithm Comparison

4.4.1. Comparison with Typical Feature Learning Algorithms

In order to evaluate the performance of the SP_DFsaeLA, its performance is compared to those of feature learning algorithms that are representative. These included both feature selection methods, such as least absolute shrinkage and selection operator (LASSO) [44] and Relief [45], as well as feature extraction algorithms, such as PCA [46], LDA [47], and locality preserving projection (LPP) [48]. The SVM is used as the base classifier, since it is a commonly used classifier. The results are listed in Table 12.
The results in Table 12 show that the SP_DFsaeLA compared with the other feature-learning algorithms has superior performance, can obtain better quality features than other feature learning algorithms, and can considerably improve the classification accuracy. The first possible reason is that the compared methods are based on the original features, but the proposed algorithm expands the original features, thereby obtaining higher-quality features efficiently. The second possible reason is the high effectiveness of the two-stage feature-reduction mechanism.

4.4.2. Comparison with Representative Stacked Autoencoders

To verify the accuracy of FESSAE, we experimentally compared its performance to that of some representative stacked autoencoders that are representative, including the stacked autoencoder (SAE) [49], stacked sparse autoencoder (SSAE) [50], stacked denoising autoencoder (SDAE) [51], stacked pruning sparse autoencoder (SPSAE) [52], and SSAE combined with LASSO (SSAE and LASSO) [53]. The performance of FESSAE without sparse term is verified, and the method is referred to as ESAE. There are some main reasons why the autoencoders are considered as a deep learning method: (1) the proposed algorithm includes an autoencoder, so representative autoencoders are compared for fairness; (2) the autoencoder is a kind of deep network, which is more suitable for datasets with a small number of samples and features compared with other kinds of deep learning methods.
Since the number of the samples and features is small or medium, the number of hidden layers and neurons cannot be large. For a fair comparison, the SP_EMNCM is not considered and all parameters are set to the same parameters for all the methods. The observed accuracy is presented in Table 13. Three datasets (PD, AD, and Vehicle) are used.
As seen from Table 13, the FESSAE achieved the highest accuracy, showing that the proposed FESSAE is effective. Table 13 shows that the sparse constraints can substantially enhance the classifier accuracy. For all datasets, the proposed FESSAE achieved the best classification accuracy.

4.4.3. Comparison with Recent Chronic Disease Detection Algorithms

To further verify the accuracy of SP_DFsaeLA algorithm, its performance is compared to those of state-of-the-art algorithms proposed in recent years for classifying chronic diseases, including those by Hasan et al. [54], Wang et al. [55], Guia et al. [56], Hasan et al. [57], Lu H et al. [21], and Taghizadeh E et al. [22], Abdollahi J et al. [30]. For a more visual representation of the comparative algorithm information, we describe the state-of-the-art (SOTA) studies with their pros and cons in Table 14. Table 15 and Table 16 show that all the models perform better in either positive or negative chronic disease prediction; the two cross-validation methods (CV, including k-fold and Holdout) are used for fair comparison. As seen from the table, the SP_DFsaeLA is better than the others in precision or F1_score or both. The cases are similar under fivefold and Hold out CV.

4.5. Parameter Analysis

4.5.1. Analysis of FESSAE Parameters

The FESSAE model is a critical part of the SP_DFsaeLA. Therefore, it is critical to assess the impact of different FESSAE parameter settings on overall performance. First, the effect of the FESSAE’s sparsity parameter is discussed. The sparsity parameter, ranging from 0.02 to 0.1, is chosen based on previous research.
As seen from Figure 9, different sparse parameters give different results. The sparse parameter had an apparent impact for the FESSAE algorithm’s accuracy. The optimal sparse parameter is different for different datasets, and there is no fixed selection criterion. The datasets used have different numbers of samples, feature dimensions, and numbers of categories, and the optimal sparse parameter may be related to these factors.
The effects of the λ in the range of 10−5 to 10−3 and β in the range of 1 to 6 on the performance of the proposed FESSAE are analyzed together. Figure 10 shows that, when the λ is fixed, the β had relatively little effect on the network. When the β is fixed, the closer the λ is to 0, the more robust is the model.

4.5.2. Analysis of Classifier Type

The impact of different classifiers (extreme learning machines (ELM), SVM, and RF) on the proposed algorithm’s performance is experimentally studied. The results are listed in Table 17.
Table 17 shows that, for five of the eight datasets, the highest accuracy, sensitivity, and specificity are achieved with the SVM. For the AD dataset, classification accuracy obtained using SVM is 71.11%, which is 10 and 3.33% higher than those with RF and ELM, respectively, and the sensitivity is 71.11%, which is 10 and 3.33% higher than those with RF and ELM. In addition, the specificity is 85.56%, which is 5% and 1.67% better than those with RF and ELM, and it had better stability, since the SVM is most accurate in most cases. Overall, the SVM outperformed the RF and ELM.

4.5.3. Analysis of the Number of Classifiers

To verify whether subclassification has an effect on the performance, we designed experiments with different numbers of sub-classifiers and the results are shown in Figure 11. From Figure 11 it is clear that the accuracy of the different numbers of sub-classifiers did not differ significantly.

5. Discussion and Conclusions

In recent years, chronic diseases have become a serious threat to human health. Currently, a practical approach is to first use wearable sensors to collect data from the human body and then process the data using machine learning. Machine learning algorithm is an effective tool for the analysis of sensing data of chronic disease. The machine learning method mainly includes two parts: feature learning and classifier, where feature learning is very important. Therefore, it is important and challenging to study highly efficient feature learning methods for sensor monitoring of chronic disease. Traditional feature learning methods are restricted from original features and cannot construct high-quality features, whereas deep feature learning methods suffer from the small-sample problems.
To overcome these limitations, this paper proposed a solution—sample-pair envelope diamond stacked sparse autoencoder ensemble learning algorithm (SP_DFsaeLA). First, the sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM) is designed by searching the manifold nearest samples in manifold neighborhood and generating sample pairs. Second, the feature embedding stacked sparse autoencoder (FESSAE) is designed to extend features. Third, a staged feature reduction mechanism is designed to reduce extended feature redundancy. The first stage uses the L1 regularized feature-reduction algorithm and the second stage uses an improved manifold dimensionality reduction algorithm to further reduce features. Fourth, the sample-pair-based model and single-sample-based model are combined by weighted fusion.
In the experimental section, three sets of experiments are organized to verify the proposed method’s effectiveness to validate the effectiveness of the innovation points and to research the algorithm’s parameters. The results show that the SP_EMNCM is valid apparently. Compared with feature extraction methods, such as LDA, the accuracy of SP_DFsaeLA’s classification recognition is higher because the chronic disease samples are often random. However, LDA is not sensitive to non-Gaussian-distributed samples. Compared with feature selection algorithms, including Relief, the advantages of SP_DFsaeLA are its greater learnability and lower tolerance to faults. Relief gives higher weights to all features that are correlated with the class, so the limitation of Relief is in that it does not effectively remove redundant features. Compared with deep learning methods (SAEs), the accuracy of SP_DFsaeLA is improved by up to 20.5%. However, SAE only considers the feature extraction without considering the intrinsic structural information between samples.
As mentioned previously, effective feature learning of sensor data for chronic diseases is important and challenging, and there are some contributions for solving the problem. Firstly, a diamond-like feature learning mechanism is proposed in this paper. This mechanism combines deep learning and traditional feature-reduction methods’ advantages and can better adapt to the data characteristics of chronic diseases and the requirement for high recognition accuracy than the deep learning and traditional feature-reduction methods. Secondly, this paper proposes a sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM), which has the advantage of enriching the feature information. Thirdly, this paper designs lightweight FESSAE, which has the advantage of improving the complementarity of original features with deep features and achieving high-quality deep features in small or medium sample sizes. Fourth, a feature expansion mechanism is proposed by combining SP_EMNCM and FESSAE, which has the advantage of improving the original features richer and better by considering the sample structure relationship and feature complementary relationship. Fifth, a two-stage feature reduction mechanism for L1_wLPPD is proposed, and the advantages of this method are making the features more compact without losing high-quality features.
As far as we know, the proposed method is designed for feature learning of chronic diseases’ sensor data, and no similar published reports have been found. In addition, the SP_DFsaeLA is a framework approach, which is more inclusive of concrete algorithms. The proposed method can produce different variations by using different feature learning methods and different classifiers. Therefore, the SP_DFsaeLA has good generalization. There are limited types of the feature learning algorithms, but they are representative and some of them had been used for sensing of chronic diseases. Compared with other deep learning algorithms, the autoencoder has lightweight structure and parameters. Therefore, it is discussed in this paper as a representative deep neural network. That is a major reason why only autoencoder rather than other deep learning methods is involved.
Although the proposed method is effective, it has limitations for datasets with a large number of samples and features. In future work, other types of deep neural networks can be considered for further verification and improvement. In addition, the proposed algorithm can be validated on more datasets and embedded in portable systems for practical diagnosis of chronic diseases.

Author Contributions

Conceptualization, Y.L.; methodology, X.Q. and Y.L.; supervision, X.Q. and Z.Z.; software, formal analysis, writing—original draft preparation, Y.Z.; data acquisition, Z.Z. and J.M.; investigation and visualization, J.M. writing—reviewing and editing, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

No new data were created.

Acknowledgments

We thank the support of NSFC (No. U21A20448, 61771080); this work was supported by Sichuan Science and Technology Program (No. 2019ZDZX0006) and the Talents by Sichuan provincial Party Committee Organization Department, and Science and Technology Service Network Initiative (KFJ-STS-QYZD-2021-21-001).

Conflicts of Interest

The authors declare no conflict of interest pertaining to this work.

References

  1. Alhassan, A.M.; Zainon, W.M.N.W. Review of feature selection, dimensionality reduction and classification for chronic disease diagnosis. IEEE Access 2021, 9, 87310–87317. [Google Scholar] [CrossRef]
  2. Yin, H.; Jha, N.K. A health decision support system for disease diagnosis based on wearable medical sensors and machine learning ensembles. IEEE Trans. Multi Scale Comput. Syst. 2017, 3, 228–241. [Google Scholar] [CrossRef]
  3. Wu, J.; Chang, L.; Yu, G. Effective data decision-making and transmission system based on mobile health for chronic disease management in the elderly. IEEE Syst. J. 2020, 15, 5537–5548. [Google Scholar] [CrossRef]
  4. Muzammal, M.; Talat, R.; Sodhro, A.H.; Pirbhulal, S. A multi-sensor data fusion enabled ensemble approach for medical data from body sensor networks. Inf. Fusion 2020, 53, 155–164. [Google Scholar] [CrossRef]
  5. Abreu, P.H.; Santos, M.S.; Abreu, M.H.; Andrade, B.; Silva, D.C. Predicting breast cancer recurrence using machine learning techniques: A systematic review. ACM Comput. Surv. 2016, 49, 1–40. [Google Scholar] [CrossRef]
  6. Gunarathne, W.; Perera, K.; Kahandawaarachchi, K. Performance evaluation on machine learning classification techniques for disease classification and forecasting through data analytics for chronic kidney disease (CKD). In Proceedings of the 2017 IEEE 17th International Conference on Bioinformatics and Bioengineering, Washington, DC, USA, 23–25 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 291–296. [Google Scholar]
  7. Yildirim, P. Chronic kidney disease prediction on imbalanced data by multilayer perceptron: Chronic kidney disease prediction. In Proceedings of the 41st IEEE Computer Software and Applications Conference, Turin, Italy, 4–8 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 193–198. [Google Scholar]
  8. Zou, Q.; Qu, K.; Luo, Y.; Yin, D.; Ju, Y.; Tang, H. Predicting diabetes mellitus with machine learning techniques. Front. Genet. 2018, 9, 515. [Google Scholar] [CrossRef]
  9. Rubini, L.J.; Eswaran, D.P. Generating comparative analysis of early stage prediction of chronic kidney disease. Int. J. Mod. Eng. Res. 2015, 5, 49–55. [Google Scholar]
  10. Sinha, P.; Sinha, P. Comparative study of chronic kidney disease prediction using KNN and SVM. Int. J. Eng. Res. Technol. 2015, 4, 608–612. [Google Scholar]
  11. Ekanayake, I.U.; Herath, D. Chronic kidney disease prediction using machine learning methods. In Proceedings of the 2020 Moratuwa Engineering Research Conference, Moratuwa, Sri Lanka, 28–30 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 260–265. [Google Scholar]
  12. Ahmed, H.; Younis, E.M.; Hendawi, A.; Ali, A.A. Heart disease identification from patients’ social posts, machine learning solution on spark. Future Gener. Comput. Syst. 2020, 111, 714–722. [Google Scholar] [CrossRef]
  13. Shrivas, A.K.; Sahu, S.K.; Hota, H.S. Classification of chronic kidney disease with proposed union based feature selection technique. Soc. Sci. Res. Netw. Electron. J. 2018, 4, 26–27. [Google Scholar] [CrossRef]
  14. Chormunge, S.; Jena, S. Correlation based feature selection with clustering for high dimensional data. J. Electr. Syst. Inf. Technol. 2018, 5, 542–549. [Google Scholar] [CrossRef]
  15. Sawhney, R.; Mathur, P.; Shankar, R. A firefly algorithm based wrapper-penalty feature selection method for cancer diagnosis. In Proceedings of the 18th International Conference on Computational Science and Its Applications, Melbourne, VIC, Australia, 2–5 July 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 438–449. [Google Scholar]
  16. Jayaraman, V.; Sultana, H.P. Artificial gravitational cuckoo search algorithm along with particle bee optimized associative memory neural network for feature selection in heart disease classification. J. Ambient. Intell. Humaniz. Comput. 2019, 1–10. [Google Scholar] [CrossRef]
  17. Paul, A.K.; Shill, P.C.; Rabin, M.R.I.; Murase, K. Adaptive weighted fuzzy rule-based system for the risk level assessment of heart disease. Appl. Intell. 2018, 48, 1739–1756. [Google Scholar] [CrossRef]
  18. Rasitha, G. Predicting thyroid disease using linear discriminant analysis (LDA) data mining technique. Int. J. Mod. Trends Eng. Res. 2016, 4, 4–6. [Google Scholar] [CrossRef]
  19. Mohamed, E.I.; Linder, R.; Perriello, G.; Di Daniele, N.; Pöppl, S.J.; De Lorenzo, A. Predicting type 2 diabetes using an electronic nose-based artificial neural network analysis. Diabetes Nutr. Metab. 2002, 15, 215–221. [Google Scholar] [PubMed]
  20. Shahbazi, F.; Asl, B.M. Generalized discriminant analysis for congestive heart failure risk assessment based on long-term heart rate variability. Comput. Methods Programs Biomed. 2015, 122, 191–198. [Google Scholar] [CrossRef]
  21. Lu, H.; Uddin, S.; Hajati, F.; Moni, M.A.; Khushi, M. A patient network-based machine learning model for disease prediction: The case of type 2 diabetes mellitus. Appl. Intell. 2022, 52, 2411–2422. [Google Scholar] [CrossRef]
  22. Taghizadeh, E.; Heydarheydari, S.; Saberi, A.; JafarpoorNesheli, S.; Rezaeijo, S.M. Breast cancer prediction with transcriptome profiling using feature selection and machine learning methods. BMC Bioinform. 2022, 23, 410. [Google Scholar] [CrossRef]
  23. Khan, A.; Uddin, S.; Srinivasan, U. Comorbidity network for chronic disease: A novel approach to understand type 2 diabetes progression. Int. J. Med. Inform. 2018, 115, 1–9. [Google Scholar] [CrossRef]
  24. Ge, R.; Zhang, R.; Wu, Q.; Wang, P. Prediction of chronic diseases with multi-label neural network. IEEE Access 2020, 127, 24–25. [Google Scholar] [CrossRef]
  25. El-Baz, A.H. Hybrid intelligent system-based rough set and ensemble classifier for breast cancer diagnosis. Neural Comput. Appl. 2015, 26, 437–446. [Google Scholar] [CrossRef]
  26. Polat, K. Similarity-based attribute weighting methods via clustering algorithms in the classification of imbalanced medical datasets. Neural Comput. Appl. 2018, 30, 987–1013. [Google Scholar] [CrossRef]
  27. Cheruku, R.; Edla, D.R.; Kuppili, V.; Dharavath, R. RST-BatMiner: A fuzzy rule miner integrating rough set feature selection and bat optimization for detection of diabetes disease. Appl. Soft Comput. 2017, 67, 764–780. [Google Scholar] [CrossRef]
  28. Maniruzzaman; Kumar, N.; Abedin, M.; Islam, S.; Suri, H.S.; El-Baz, A.S.; Suri, J.S. Comparative approaches for classification of diabetes mellitus data: Machine learning paradigm. Comput. Methods Programs Biomed. 2017, 152, 23–34. [Google Scholar] [CrossRef]
  29. Alhassan, A.M.; Wan Zainon, W.M.N. Taylor bird swarm algorithm based on deep belief network for heart disease diagnosis. Appl. Sci. 2020, 10, 6626. [Google Scholar] [CrossRef]
  30. Abdollahi, J.; Nouri-Moghaddam, B.; Ghazanfari, M. Deep Neural Network Based Ensemble learning Algorithms for the healthcare system (diagnosis of chronic diseases). arXiv 2021, arXiv:2103.08182. [Google Scholar]
  31. Fatan, M.; Hosseinzadeh, M.; Askari, D.; Sheikhi, H.; Rezaeijo, S.M.; Salmanpour, M.R. Fusion-based head and neck tumor segmentation and survival prediction using robust deep learning techniques and advanced hybrid machine learning systems. In Head and Neck Tumor Segmentation and Outcome Prediction: Second Challenge, Proceedings of the HECKTOR 2021, Held in Conjunction with MICCAI 2021, Strasbourg, France, 27 September 2021; Springer International Publishing: Cham, Switzerland, 2022; pp. 211–223. [Google Scholar]
  32. Rezaeijo, S.M.; Hashemi, B.; Mofid, B.; Bakhshandeh, M.; Mahdavi, A.; Hashemi, M.S. The feasibility of a dose painting procedure to treat prostate cancer based on mpMR images and hierarchical clustering. Radiat. Oncol. 2021, 16, 182. [Google Scholar] [CrossRef] [PubMed]
  33. Hegde, S.; Mundada, M.R. Early prediction of chronic disease using an efficient machine learning algorithm through adaptive probabilistic divergence based feature selection approach. Int. J. Pervasive Comput. Commun. 2020, 17, 20–36. [Google Scholar] [CrossRef]
  34. Simon, D.S.; Fraser Paul, J. Kidney disease in the global burden of disease study 2017. Nat. Rev. Nephrol. 2019, 15, 193. [Google Scholar]
  35. Lichman, M. UCI Machine Learning Repository. 2013. Available online: http://archive.ics.uci.edu/ml (accessed on 1 January 2022).
  36. Smith, J.W.; Everhart, J.E.; Dickson, W.C.; Knowler, W.C.; Johannes, R.S. Using the ADAP learning algorithm to forecast the onset of diabetes mellitus. In Proceedings of the Symposium on Computer Applications in Medical Care, Washington, DC, USA, 6–9 November 1988; Volume 10, pp. 261–265. [Google Scholar]
  37. Sakar, B.E.; Isenkul, M.E.; Sakar, C.O.; Sertbas, A.; Gurgen, F.; Delil, S.; Apaydin, H.; Kursun, O. Collection and analysis of a Parkinson speech dataset with multiple types of sound recordings. IEEE J. Biomed. Health Inform. 2013, 17, 828–834. [Google Scholar] [CrossRef] [PubMed]
  38. Tan, X.; Liu, Y.; Li, Y.; Wang, P.; Zeng, X.; Yan, F.; Li, X. Localized instance fusion of MRI data of Alzheimer’s disease for classification based on instance transfer ensemble learning. Biomed. Eng. Online 2018, 17, 49. [Google Scholar] [CrossRef] [Green Version]
  39. Mangasarian, O.L.; Wolberg, S. Breast cancer diagnosis and prognosis via linear programming. Oper. Res. 1995, 43, 570–577. [Google Scholar] [CrossRef] [Green Version]
  40. Little, M.; McSharry, P.E.; Hunter, E.J.; Spielman, J.; Ramig, L.O. Suitability of dysphonia measurements for telemonitoring of Parkinson’s disease. IEEE Trans. Bio-Med. Eng. 2009, 56, 1015. [Google Scholar] [CrossRef] [Green Version]
  41. Asuncion, A. UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences. 2007. Available online: http://www.ics.uci.edu/~mlearn/MLRepository.html (accessed on 1 January 2022).
  42. Merz, C.J. UCI Repository of Machine Learning Databases. 1998. Available online: http://archive.ics.uci.edu/ (accessed on 5 January 2022).
  43. Beer, D.G.; Kardia, S.L.; Huang, C.C.; Giordano, T.J.; Levin, A.M.; Misek, D.E.; Lin, L.; Chen, G.; Gharib, T.G.; Thomas, D.G.; et al. Gene-expression profiles predict survival of patients with lung adenocarcinoma. Nat. Med. 2002, 8, 816–824. [Google Scholar] [CrossRef]
  44. Yamada, M.; Jitkrittum, W.; Sigal, L.; Xing, E.P.; Sugiyama, M. High-dimensional feature selection by feature-wise kernelized lasso. Neural Comput. 2014, 26, 185–207. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Sun, Y.; Lou, X.; Bao, B. A novel relief feature selection algorithm based on mean-variance model. J. Inf. Comput. Sci. 2011, 8, 3921–3929. [Google Scholar]
  46. Wold, S.; Esbensen, K.; Geladi, P. Principal component analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  47. Li, C.H.; Kuo, B.C.; Lin, C.T. LDA-based clustering algorithm and its application to an unsupervised feature extraction. IEEE Trans. Fuzzy Syst. 2010, 19, 152–163. [Google Scholar] [CrossRef]
  48. He, X.; Niyogi, P. Locality preserving projections. Adv. Neural Inf. Process. Syst. 2003, 16, 153–160. [Google Scholar]
  49. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Goswami, G.; Vatsa, M.; Singh, R. Face verification via learned representation on feature-rich video frames. IEEE Trans. Inf. Secur. 2017, 12, 1686–1698. [Google Scholar] [CrossRef]
  51. Görgel, P.; Simsek, A. Face recognition via deep stacked denoising sparse autoencoders (DSDSA). Appl. Math. Comput. 2019, 355, 325–342. [Google Scholar] [CrossRef]
  52. Kampffmeyer, M.; Løkse, S.; Bianchi, F.M.; Jenssen, R.; Livi, L. The deep kernelized autoencoder. Appl. Soft Comput. 2018, 71, 816–825. [Google Scholar] [CrossRef] [Green Version]
  53. Zhu, H.; Cheng, J.; Zhang, C.; Wu, J.; Shao, X. Stacked pruning sparse denoising autoencoder based intelligent fault diagnosis of rolling bearings. Appl. Soft Comput. 2020, 88, 106060. [Google Scholar] [CrossRef]
  54. Hasan, M.K.; Alam, M.A.; Das, D.; Hossain, E.; Hasan, M. Diabetes prediction using ensembling of different machine learning classifiers. IEEE Access 2020, 8, 76516–76531. [Google Scholar] [CrossRef]
  55. Wang, Q.; Cao, W.; Guo, J.; Ren, J.; Cheng, Y.; Davis, D.N. DMP_MI: An effective diabetes mellitus classification algorithm on imbalanced data with missing values. IEEE Access 2019, 7, 102232–102238. [Google Scholar] [CrossRef]
  56. De Guia, J.D.; Concepcion, R.S.; Bandala, A.A.; Dadios, E.P. Performance comparison of classification algorithms for diagnosing chronic kidney disease. In Proceedings of the 2019 IEEE 11th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment, and Management, Laoag, Philippines, 29 November–1 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–7. [Google Scholar]
  57. Hasan, K.A.; Hasan, M.A.M. Prediction of clinical risk factors of diabetes using multiple machine learning techniques resolving class imbalance. In Proceedings of the 2020 23rd International Conference on Computer and Information Technology, Dhaka, Bangladesh, 19–21 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
Figure 1. The structure of chronic disease recognition system based on machine learning.
Figure 1. The structure of chronic disease recognition system based on machine learning.
Applsci 13 07322 g001
Figure 2. Proposed method and traditional feature-based learning methods: (a) non-parameter or low-parameter feature reduction; (b) deep feature learning; (c) sample-pair diamond-like feature learning mechanism.
Figure 2. Proposed method and traditional feature-based learning methods: (a) non-parameter or low-parameter feature reduction; (b) deep feature learning; (c) sample-pair diamond-like feature learning mechanism.
Applsci 13 07322 g002
Figure 3. Proposed algorithm’s framework.
Figure 3. Proposed algorithm’s framework.
Applsci 13 07322 g003
Figure 4. Diamond-like feature learning mechanism (DFLM).
Figure 4. Diamond-like feature learning mechanism (DFLM).
Applsci 13 07322 g004
Figure 5. Sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM).
Figure 5. Sample-pair envelope manifold neighborhood concatenation mechanism (SP_EMNCM).
Applsci 13 07322 g005
Figure 6. Sample-pair-based model and single-sample-based model.
Figure 6. Sample-pair-based model and single-sample-based model.
Applsci 13 07322 g006
Figure 7. FESSAE model.
Figure 7. FESSAE model.
Applsci 13 07322 g007
Figure 8. Variation in the number of features in DFLM: (a) PID, (b) Heart, (c) WDBC, and (d) Wisconsin databases.
Figure 8. Variation in the number of features in DFLM: (a) PID, (b) Heart, (c) WDBC, and (d) Wisconsin databases.
Applsci 13 07322 g008
Figure 9. Effect of sparse proportion on FESSAE with (a) PID; (b) Heart; (c) WDBC; and (d) Wisconsin datasets.
Figure 9. Effect of sparse proportion on FESSAE with (a) PID; (b) Heart; (c) WDBC; and (d) Wisconsin datasets.
Applsci 13 07322 g009
Figure 10. Effect of penalty term coefficients on FESSAE performance: (a) single sample of PID; (b) sample pair of PID; (c) single sample of Heart; (d) sample pair of Heart.
Figure 10. Effect of penalty term coefficients on FESSAE performance: (a) single sample of PID; (b) sample pair of PID; (c) single sample of Heart; (d) sample pair of Heart.
Applsci 13 07322 g010
Figure 11. Effect of the number of classifiers on accuracy.
Figure 11. Effect of the number of classifiers on accuracy.
Applsci 13 07322 g011
Table 1. Strengths and weaknesses of the proposed method and previous methods.
Table 1. Strengths and weaknesses of the proposed method and previous methods.
MethodStrengthsWeaknesses
Ahmed H et al. [12]Predicting heart disease with real-time streaming data.Low adaptivity.
Shrivas et al. [13]High robustness and computational efficiency.Only focus on the original sample itself and ignore the neighborhood relationship of the sample.
Chormunge et al. [14]Reduce the dimensionality issue in data mining tasks.Sensitive to large changes in data distribution.
Sawhney et al. [15]Effectively reducing the number of features and saving computational overhead.Limited improvement in classification accuracy.
Rasitha [18]Effective in reducing feature dimensionality.Not sensitive to non-Gaussian distributed samples.
Mohamed et al. [19]Relatively simple and computationally efficient.Only original features are operated on. When the quality of the original feature is not high, it does not work well.
El-baz et al. [25]High classification accuracy and interpretability.Increasing computational complexity.
Maniruzzaman et al. [28]Robust to noise and outliers.In the case of small sample sizes, it is not possible to obtain sufficiently high feature quality.
Alhassan and Zainon [29]Utilizes advanced deep learning architecture for feature learning.High computational and training resource requirements.
SP_DFsaeLA
(proposed)
Both the manifold neighbor structure information between samples and the feature complementarity are considered, leading to a richer features and better improvement on accuracy. More adapted to the data characteristics of wearable chronic diseases recognition.Requires careful tuning of hyperparameters.
Table 2. Symbol description.
Table 2. Symbol description.
SymbolMeaning
X R N M Input data. N and M are sample size and feature size.
Y Sample label.
X p a i r R N 2 M Generated sample pairs. N and 2M are the sample size and feature size, respectively.
W 1 Encoder’s weight.
W 2 Decoder’s weight.
b 1 , b 2 Bias vector of the autoencoder.
dNumber of hidden units.
H R N d Hidden feature.
d ( k ) kth autoencoder number of hidden layer units.
β Coefficient for the sparsity regularization term.
λ Coefficient for the L2 weight regularization term.
ρ Sparse parameter.
ρ ^ j All training samples’ average activation value on the j-th hidden neuron
X ^ R N M ^ Data after feature expansion.
S B ϕ Inter-class variance matrix.
S W ϕ Intra-class variance matrix.
Table 3. Basic information of datasets used in study.
Table 3. Basic information of datasets used in study.
DatasetInstancesAttributesClassRelevant Paper
Statlog Heart Data Set (Heart)270132Reference [35]
Pima Indians Diabetes Data Set (PID)76882Reference [36]
Parkinson Speech Dataset (PD)1040262Reference [37]
Alzheimer’s disease (AD)90323Reference [38]
Breast Cancer Wisconsin Original (Wisconsin)68392Reference [39]
Maxlettle Parkinson Dataset (Maxlettle)195222Reference [40]
Statlog Vehicle Silhouettes (Vehicle)846184Reference [41]
Breast Cancer Wisconsin Diagnostic(WDBC)569302Reference [42]
lung cancer32563Reference [43]
Table 4. Parameter information.
Table 4. Parameter information.
ParameterParameter MeaningParameter Value
λ Coefficient for the L2 weight regularization term.1 × 10−4, 1 × 10−3, 1 × 10−2
β Coefficient for the sparsity regularization term.1, 2, 3, 4, 5, 6
ρ Sparse parameter.[0.02, 0.1]
PID hidden unitsNumber of FESSAE hidden-layer units in PID dataset120–40–16
Maxlettle hidden unitsNumber of FESSAE hidden-layer units in Maxlettle dataset160–80–42
Heart hidden unitsNumber of FESSAE hidden-layer units in Heart dataset100–60–24
PD hidden unitsNumber of FESSAE hidden-layer units in PD dataset200–100–48
AD hidden unitsNumber of FESSAE hidden-layer units in AD dataset240–120–60
Wisconsin hidden unitsNumber of FESSAE hidden-layer units in Wisconsin dataset120–40–16
Vehicle hidden unitsNumber of FESSAE hidden-layer units in Vehicle dataset120–60–32
WDBC hidden unitsNumber of FESSAE hidden-layer units in WDBC dataset200–110–60
lung cancer hidden unitsNumber of FESSAE hidden-layer units in lung cancer dataset424–210–106
Table 5. Comparison of sample-pair-based and single-sample-based methods on SVM.
Table 5. Comparison of sample-pair-based and single-sample-based methods on SVM.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS74.6177.0970.5 78.7273.44
PS74.2979.3079.4879.1179.32
MaxlettleSS80.2583.5676.8983.5679.56
PS81.2484.3779.3383.3381.02
HeartSS82.9287.2477.5088.3381.94
PS91.6794.1989.1794.1791.54
PDSS65.5864.9967.3163.8466.02
PS68.9469.8467.3170.5868.29
ADSS46.6748.5246.6773.3347.55
PS58.8965.7858.8979.4462.07
WisconsinSS96.8695.5498.3295.4096.91
PS97.7098.3197.0798.3297.67
VehicleSS82.1682.2382.1594.0482.19
PS84.6584.7684.6594.9084.71
WDBCSS96.9498.6195.3298.5996.87
PS99.0610098.1110099.04
lung cancerSS52.3841.6733.3363.3337.04
PS91.7587.7891.1193.6789.33
Table 6. Comparison of sample-pair-based and single-sample-based methods on RF.
Table 6. Comparison of sample-pair-based and single-sample-based methods on RF.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS79.4780.5277.9880.9879.09
PS84.8883.9086.5683.2285.14
MaxlettleSS87.6888.0189.3386.0088.08
PS92.6896.3689.1196.0091.87
HeartSS87.5090.5784.1790.8387.17
PS95.4297.5393.3397.5095.31
PDSS81.3581.2291.7380.9681.41
PS87.6987.0289.0486.3587.83
ADSS50.0052.0450.0075.0050.87
PS74.4475.5074.4474.9674.96
WisconsinSS98.5397.9599.1597.9198.54
PS99.1698.7799.5898.7499.17
VehicleSS86.6886.4986.6895.5786.59
PS90.8390.7990.8296.9590.80
WDBCSS96.7096.8496.7196.6796.76
PS98.3598.6798.1298.5798.38
lung cancerSS66.6750.0038.8975.0043.75
PS91.7487.7791.1193.6789.33
Table 7. Comparison of sample pair and single samples after LDA on SVM.
Table 7. Comparison of sample pair and single samples after LDA on SVM.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS74.9877.8770.1679.8373.67
PS80.2180.0880.5879.8580.30
MaxlettleSS86.4692.7879.1193.7885.32
PS97.0098.1896.0098.0096.83
HeartSS86.6790.7881.6791.6785.94
PS94.1795.1293.3395.0094.13
PDSS67.2166.6368.8465.5867.65
PS69.8170.7868.8871.7369.09
ADSS62.2264.6862.2281.1163.23
PS74.4475.4974.4487.2274.95
WisconsinSS97.2896.3498.3296.2297.31
PS98.5497.9699.1697.9198.54
VehicleSS83.5483.5083.5394.5183.51
PS88.1888.5488.1796.0688.36
WDBCSS98.5699.5297.6599.5298.58
PS99.2910098.5810099.28
lung cancerSS57.1461.1161.1178.3361.11
PS93.8194.4495.0097.0094.66
Table 8. Comparison of sample pair and single samples after LDA on RF.
Table 8. Comparison of sample pair and single samples after LDA on RF.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS81.5282.5480.2282.8482.24.
PS84.3282.6887.3081.3384.85
MaxlettleSS89.5791.5789.1189.7889.44
PS89.6794.1885.3394.0089.15
HeartSS89.1792.9685.0093.3388.71
PS93.7595.7591.6795.8393.61
PDSS80.1981.4278.2682.1279.80
PS85.0985.2185.1985.0185.08
ADSS62.2266.2462.2281.1164.12
PS55.5658.6055.5677.7856.94
WisconsinSS98.3297.5699.1697.4898.34
PS98.9597.9810097.9198.97
VehicleSS89.1989.4689.2096.3989.33
PS92.5792.7092.5797.5392.63
WDBCSS98.5999.5397.6499.5298.57
PS98.5999.5297.6599.5298.57
lung cancerSS83.3383.3388.8991.6786.02
PS94.2893.3396.6796.6794.92
Table 9. Comparison of sample pairs and single samples after FESSAE on SVM.
Table 9. Comparison of sample pairs and single samples after FESSAE on SVM.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS77.4477.3079.1575.7777.81
PS83.8088.6079.1488.4883.03
MaxlettleSS99.0010098.0010098.95
PS98.8910097.7810098.82
HeartSS91.2591.6490.8391.6791.16
PS97.0897.5996.6797.5097.07
PDSS75.0073.2679.3370.6776.07
PS81.4980.9282.6980.2981.69
ADSS71.1174.9671.1185.5672.95
PS75.5678.7075.5687.7877.07
WisconsinSS99.1698.3810098.3299.18
PS99.5899.1810099.1699.58
VehicleSS80.7782.3480.7893.5781.53
PS84.5385.5684.5497.8285.05
WDBCSS97.1797.8096.7197.6297.18
PS99.7699.5310099.5399.76
lung cancerSS85.7183.3391.6791.6787.30
PS97.1496.6798.3398.3397.46
Table 10. Comparison of sample pairs and single samples after FESSAE on RF.
Table 10. Comparison of sample pairs and single samples after FESSAE on RF.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDSS77.4477.3079.1575.7777.81
PS83.8088.6079.1488.4883.03
MaxlettleSS92.6197.5087.5697.7891.89
PS96.7810093.3310096.32
HeartSS91.2591.6490.8391.6791.16
PS97.0897.5996.6797.5097.07
PDSS75.0073.2679.3370.6776.07
PS81.4980.9282.6980.2981.69
ADSS58.8960.8258.8979.4459.78
PS62.2264.0862.2281.1163.04
WisconsinSS99.1698.3810098.3299.18
PS99.5899.1710099.1699.58
VehicleSS80.7782.3480.7893.5781.54
PS84.5385.5784.5494.8285.05
WDBCSS97.4197.7997.1897.6297.42
PS99.7699.5310099.5399.76
lung cancerSS50.0050.0038.8975.0043.75
PS87.6287.7891.1193.6789.33
Table 11. Result of principal stages of the proposed method.
Table 11. Result of principal stages of the proposed method.
DatasetMethodAcc (%)Prec (%)Sens (%)Spec (%)F1_Score (%)
PIDOFs 79.2979.3079.4879.1179.32
DFs 83.8088.6079.1488.4883.03
CF82.3186.9777.6686.9781.38
CF and DFLM 84.5488.5080.9988.1184.18
HeartOFs 91.6794.1989.1794.1791.54
DFs 97.0897.5996.6797.5097.07
CF96.2597.5195.0097.5096.13
CF and DFLM96.6797.5995.8397.5096.65
PDOFs 68.9469.8467.3170.5868.29
DFs 81.4980.9282.6980.2981.69
CF78.7381.3675.9681.4977.53
CF and DFLM79.9380.0579.8180.0579.60
ADOFs 58.8965.7858.8979.4462.07
DFs 75.5678.7075.5687.7877.07
CF68.8969.9768.8984.4469.41
CF and DFLM 72.2274.5572.2286.1173.36
WisconsinOFs 97.7098.3197.0798.3297.67
DFs 99.5899.1810099.1699.58
CF98.9598.7699.1698.7498.95
CF and DFLM 99.3798.7710098.7499.58
VehicleOFs 84.6584.7684.6594.9084.71
DFs 84.5385.5684.5497.8285.05
CF85.5486.3785.5595.1885.95
CF and DFLM 87.8088.7387.8295.9388.27
WDBCOFs 99.0610098.1110099.04
DFs 99.7699.5310099.5399.76
CF98.8299.0998.5899.0598.82
CF and DFLM99.5310099.5310099.53
MaxlettleOFs81.2484.3779.3383.3381.02
DFs98.8910097.7810098.82
CF97.8410095.5610097.65
CF and DFLM 97.8910095.5610097.50
lung cancerOFs91.7587.7891.1193.6789.33
DFs97.1496.6798.3398.3397.46
CF87.6287.7891.1193.6789.40
CF and DFLM 95.2491.6796.6795.8393.84
Table 12. Comparison of typical feature learning algorithms.
Table 12. Comparison of typical feature learning algorithms.
DatasetPerformance IndicesOFs
(%)
PCA (%)LDA (%)LPP (%)Relief (%)LASSO (%)SP_DFsaeLA
(Proposed) (%)
PIDAcc79.2980.0380.2180.0378.1777.5984.54
Prec79.3079.8280.0979.3878.3077.7688.50
Sens79.4880.5980.5881.3378.3477.5980.99
Spec79.1179.4879.8578.7477.9977.6088.11
F1_score79.3280.1380.3080.2578.2477.6784.18
MaxlettleAcc81.2491.7397.0095.8490.6284.5297.89
Prec84.3791.3394.8596.3788.3291.86100
Sens79.3393.7810095.7893.7877.3395.56
Spec83.3389.7894.0096.0087.5691.78100
F1_score81.0291.9197.2395.8790.9282.6597.50
HeartAcc91.6795.4294.1794.5893.3388.7596.67
Prec94.1995.9395.1294.3694.8589.5497.59
Sens89.1795.0093.3395.0091.6788.3395.83
Spec94.1795.8395.0094.1795.0089.1797.50
F1_score91.5495.3994.1294.6193.1988.7696.65
WDBCAcc99.0699.2999.2999.2999.2998.1199.53
Prec10010010010099.5399.05100
Sens98.1198.5898.5898.5799.0597.1799.53
Spec10010010010099.5299.06100
F1_score99.0499.2899.2899.2799.2898.0999.53
PDAcc68.9471.2569.9070.5871.2567.2179.93
Prec69.8471.8970.8071.3571.7266.6780.05
Sens67.3170.1968.2769.4270.3868.6579.81
Spec70.5872.3171.5471.7372.1265.7780.05
F1_score68.2970.8869.2670.1270.8667.4979.60
ADAcc58.8971.1172.2266.6768.8962.2272.22
Prec65.7874.0180.0066.8373.7066.3074.54
Sens58.8971.1172.2266.6768.8962.2272.22
Spec79.4485.5686.1183.3384.4481.1186.11
F1_score62.0772.0075.9166.7571.1964.0773.36
VehicleAcc84.6587.5587.1587.4486.7982.7787.80
Prec84.7687.6187.5687.4687.2083.2088.73
Sens84.6587.5587.1787.4486.7882.7787.82
Spec94.8995.8695.0595.1495.6094.2595.93
F1_score84.7187.5887.3687.4586.9982.9988.27
WisconsinAcc97.7098.5398.5498.5498.1197.7099.37
Prec98.3197.5598.3497.9497.5897.5298.77
Sens97.0699.5798.7499.1698.7497.91100
Spec98.3297.4998.3297.9197.4997.4998.74
F1_score97.6798.5598.5398.5498.1397.7099.38
lung can-cerAcc91.7591.7594.2884.6591.4288.5795.24
Prec87.7887.7894.4484.7692.2290.0091.67
Sens91.1191.1196.1184.6592.2290.0096.67
Spec93.6793.6797.0094.8995.6794.3395.83
F1_score89.3389.3395.2484.7192.2290.0093.84
Table 13. Classification accuracy of different deep autoencoder classifiers.
Table 13. Classification accuracy of different deep autoencoder classifiers.
DatasetSAE [49]
(%)
SSAE [50]
(%)
SDAE [51] (%)SPSAE [52] (%)SSAE and LASSO [53]
(%)
ESAE
(%)
Proposed FESSAE
(%)
PD64.1566.4866.4866.2265.8760.2975.67
AD57.6761.6759.5861.7857.6660.9471.11
Vehicle67.3070.0072.0074.7680.0674.8980.77
Table 14. SOTA studies with their pros and cons.
Table 14. SOTA studies with their pros and cons.
LiteratureMethodProsCons
Hasan M K et al. [54]A framework for diabetes prediction is proposed in which data standardization, feature selection, and different Machine Learning classifiers.High robustness.Requires significant time to search for the best combination.
Wang Q et al. [55]An effective prediction algorithm for diabetes is proposed, which include Naïve Bayes method, adaptive synthetic sampling method and random forest.Relatively simple and computationally efficient.Low universality.
Guia J et al. [56]Six classification algorithms were used to predict chronic kidney disease, including support vector classifier, decision trees, random forest, Gaussian Naive Bayes, Multilayer Perceptron, and K-Nearest neighbors.Saving computational overhead.Classification performance is dependent on tuning parameters.
Hasan K A et al. [57]Using logistic regression, analysis of variance and to identify by using multiple supervised machine learning algorithms.Ability to select the most significant risk factors associated with diabetes.Limited improvement in classification accuracy.
Lu H et al. [21]A patient network and machine learning approach is developed to combine the attributes of the patient network with the sample features and eight machine learning models are used to predict disease.Discover the potential characteristics of the patient.Classification performance is too dependent on the dataset.
Taghizadeh E et al. [22]Three groups of machine learning algorithms were employed: four feature selection procedures are employed and compared to select the most valuable feature; Principal Component Analysis, 13 classification algorithms accompanied with automated hyperparameter tuning.Effectively reducing the number of features.The sample size was relatively large.
Abdollahi J et al. [30]Selection of 10 machine learning algorithms as the basic algorithms in a stack generalization algorithm to predict chronic diseases and implementation of a hybrid meta-algorithm for prediction.High classification accuracy and interpretability.Only original features are operated on. When the quality of the original feature is not high, it does not work well.
Table 15. Performance comparison of proposed method with state-of-the-art chronic disease detection algorithms.
Table 15. Performance comparison of proposed method with state-of-the-art chronic disease detection algorithms.
DatasetPerformance IndicesLiterature [54]
(5-Fold)
Literature [55]
(5-Fold)
Literature [56]
(Holdout)
Literature [57]
(5-Fold)
Proposed Method
(5-Fold)
Proposed Method
(Holdout)
PIDAcc (%)88.887.180.0475.7184.5483.18
Sens (%)78.985.473.51-80.9974.07
Spec (%)93.4---88.1192.45
Prec (%)84.280.684.93-88.5090.91
F1_score (%)-82.978.80-84.1881.63
WDBCAcc (%)97.395.697.6595.6199.7698.82
Sens (%)83.095.195.35-99.53100
Spec (%)95.7---10098.82
Prec (%)87.896.0100-10097.67
F1_score (%)-95.597.62-99.7698.82
WisconsinAcc (%)98.197.196.8496.7999.3798.95
Sens (%)53.397.197.87-100100
Spec (%)99.5---98.7497.87
Prec (%)86.797.295.83-98.7797.96
F1_score (%)-97.196.84-99.3898.97
HeartAcc (%)73.986.785.4282.9696.6795.83
Sens (%)74.881.775.00-95.83100
Spec (%)73.0---97.5091.67
Prec (%)69.790.894.74-97.5992.31
F1_score (%)-86.083.72-96.6596.00
Table 16. Comparison of proposed method with state-of-the-art chronic disease detection algorithms.
Table 16. Comparison of proposed method with state-of-the-art chronic disease detection algorithms.
DatasetPerformance IndicesLiterature [21]
(5-Fold)
Literature [30]
(5-Fold)
Literature [22]
(5-Fold)
Proposed Method
(5-Fold)
PIDAcc (%)79.3881.1380.1984.54
Sens (%)87.5083.0279.2580.99
Spec (%)71.2579.2581.1388.11
Prec (%)75.2780.0080.7788.50
F1_score (%)80.9281.4880.0084.18
WDBCAcc (%)96.8598.8295.2999.76
Sens (%)96.8310092.8699.53
Spec (%)96.8897.6797.67100
Prec (%)96.8397.6797.50100
F1_score (%)96.8398.8295.1299.76
WisconsinAcc (%)94.4197.8995.7999.37
Sens (%)94.3710095.74100
Spec (%)94.4495.7495.8398.74
Prec (%)94.3796.0095.7498.77
F1_score (%)94.3797.8696.0099.38
HeartAcc (%)83.3395.8385.4296.67
Sens (%)91.6710083.3395.83
Spec (%)75.0091.6787.5097.50
Prec (%)78.5792.3186.9697.59
F1_score (%)84.6296.0085.0096.65
Table 17. Classification accuracy of proposed algorithm with different classifiers.
Table 17. Classification accuracy of proposed algorithm with different classifiers.
DatasetSVM (%)RF (%)ELM (%)
AccSpecSensAccSpecSensAccSpecSens
PID84.5488.1180.9983.6185.8581.3880.8392.5469.14
Maxlettle10010010098.8910097.7899.0098.00100
Heart96.6797.5095.8396.6797.5095.8396.6797.5095.83
PD81.7381.3582.1281.4478.2784.6280.6780.0081.35
AD71.1185.5671.1161.1180.5661.1167.7883.8967.78
Wisconsin99.3798.7410099.1698.3210099.1699.1599.15
Vehicle87.8095.9387.8287.0695.6787.0687.5595.1687.55
WDBC99.7610099.5397.4197.6397.1698.8298.5899.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Ma, J.; Qin, X.; Li, Y.; Zhang, Z. Sample-Pair Envelope Diamond Autoencoder Ensemble Algorithm for Chronic Disease Recognition. Appl. Sci. 2023, 13, 7322. https://doi.org/10.3390/app13127322

AMA Style

Zhang Y, Ma J, Qin X, Li Y, Zhang Z. Sample-Pair Envelope Diamond Autoencoder Ensemble Algorithm for Chronic Disease Recognition. Applied Sciences. 2023; 13(12):7322. https://doi.org/10.3390/app13127322

Chicago/Turabian Style

Zhang, Yi, Jie Ma, Xiaolin Qin, Yongming Li, and Zuwei Zhang. 2023. "Sample-Pair Envelope Diamond Autoencoder Ensemble Algorithm for Chronic Disease Recognition" Applied Sciences 13, no. 12: 7322. https://doi.org/10.3390/app13127322

APA Style

Zhang, Y., Ma, J., Qin, X., Li, Y., & Zhang, Z. (2023). Sample-Pair Envelope Diamond Autoencoder Ensemble Algorithm for Chronic Disease Recognition. Applied Sciences, 13(12), 7322. https://doi.org/10.3390/app13127322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop