Next Article in Journal
An Image Encryption Transmission Scheme Based on a Polynomial Chaotic Map
Previous Article in Journal
A New Thermodynamic Model to Approximate Properties of Subcritical Liquids
Previous Article in Special Issue
Protein Is an Intelligent Micelle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Weight-Based Feature Selection (WBFS) Algorithm Classifies Lung Cancer Subtypes Using Proteomic Data

1
School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710129, China
2
Xi’an Key Laboratory of Stem Cell and Regenerative Medicine, Institute of Medical Research, Northwestern Polytechnical University, Xi’an 710072, China
*
Authors to whom correspondence should be addressed.
Entropy 2023, 25(7), 1003; https://doi.org/10.3390/e25071003
Submission received: 3 April 2023 / Revised: 27 June 2023 / Accepted: 28 June 2023 / Published: 29 June 2023
(This article belongs to the Special Issue Information Theory in Computational Biology)

Abstract

:
Feature selection plays an important role in improving the performance of classification or reducing the dimensionality of high-dimensional datasets, such as high-throughput genomics/proteomics data in bioinformatics. As a popular approach with computational efficiency and scalability, information theory has been widely incorporated into feature selection. In this study, we propose a unique weight-based feature selection (WBFS) algorithm that assesses selected features and candidate features to identify the key protein biomarkers for classifying lung cancer subtypes from The Cancer Proteome Atlas (TCPA) database and we further explored the survival analysis between selected biomarkers and subtypes of lung cancer. Results show good performance of the combination of our WBFS method and Bayesian network for mining potential biomarkers. These candidate signatures have valuable biological significance in tumor classification and patient survival analysis. Taken together, this study proposes the WBFS method that helps to explore candidate biomarkers from biomedical datasets and provides useful information for tumor diagnosis or therapy strategies.

1. Introduction

Lung cancer is one of the deadliest cancers in the world, and lung adenocarcinoma (LUAD) and lung squamous cell carcinoma (LUSC) are the most common subtypes, which have drastically different biological signatures [1]. The precise biomarkers to be implemented in clinical settings remain ambiguous, and accurate classification of lung cancers into clinically significant subtypes is of utmost importance in making therapeutic decisions; this gap demands immediate attention.
The Cancer Proteome Atlas (TCPA) database [2] is a large-scale proteomic database that contains molecular and clinical data from over 11,000 patient samples across 32 different cancer types. The Cancer Genome Atlas (TCGA) database [3,4] is a comprehensive genomic database that also contains the clinical information of patients. Both TCPA and TCGA databases are publicly available and have been widely used by researchers around the world. They provide valuable resources for cancer research, enabling researchers to investigate the molecular mechanisms underlying cancer and identify new targets for therapeutic intervention. However, discovering key biomarkers for lung cancer from the extensive datasets in these databases is a difficult task in biomedical research [5].
Feature selection is an important task in machine learning and data analysis, where the goal is to identify a subset of relevant features from a large set of potential predictors [6]. One approach to feature selection is based on information theory; such an approach involves quantifying the amount of information that a feature provides about the outcome variable of interest. This approach is especially useful in situations where the number of potential predictors is large and the relationship between the predictors and the outcome variable is complex. Information-theoretic feature selection methods are based on measures such as entropy, mutual information, and conditional mutual information. These measures are used to assess the relevance of a feature to the outcome variable, as well as the redundancy between features. The goal is to select a subset of features that maximizes the amount of relevant information while minimizing the amount of redundant information.
One application of information-theoretic feature selection is in the field of biomarker discovery, where the goal is to identify a set of molecular features (such as genes or proteins) that are associated with a particular disease or condition [7]. By using information-theoretic methods to select key biomarkers, researchers can gain insights into the underlying biological mechanisms of the disease, as well as develop diagnostic and therapeutic tools. In the past two decades, many studies on filter feature selection algorithms based on information theory have been reported due to their robustness and computational efficiency [8]. The pioneering work of Battiti employed the mutual information (MI) criterion to evaluate a set of candidate features, and his research proved that mutual information is suitable for assessing arbitrary dependencies between random variables [8,9]. To date, most feature selection methods based on information theory obtain the optimal feature subset by maximizing correlation and minimizing redundancy, such as Mutual Information Maximisation (MIM) [10], Mutual Information Based Feature Selection (MIFS) [9], Mutual Information Feature Selector under Uniform Information Distribution (MIFS-U) [11], Max-Relevance and Min-Redundancy (mRMR) [12], Conditional Informative Feature Extraction (CIFE) [13], and Conditional Redundancy (CONDRED) [14]. Most existing feature selection strategies aim to maximize the classification performance only by considering the relevance between features and classes and redundancy between pairs of features without assessing interactions and complementarity; consequently, some features with discriminative ability may be mistakenly removed. The concept of interaction has been widely explored in different scenarios [15]. For example, exploring the interaction information of microarray gene expression data is beneficial for discovering cancer biomarkers and identifying cancer subtype classifications [16].
This study involved an extensive examination of a lung cancer protein expression dataset sourced from the TCPA database, in conjunction with phenotype and survival information from the TCGA database. Firstly, it aimed to provide an overview of the functional proteomic heterogeneity of LUAD and LUSC tumor samples. Secondly, a unique filter weight-based feature selection (WBFS) method that assesses selected features and candidate features was used to identify candidate protein biomarkers that exhibited superior performance in classifying the two major lung cancer subtypes. Additionally, Bayesian etworks (BNs) were utilized to identify the direct impact factors that had causal relationships with the classification of the two subtypes. Finally, the potential clinical implications of the candidate protein biomarkers for prognosis were evaluated.

2. Materials and Methods

2.1. Data Acquisition and Preprocessing

We extracted the lung cancer subset from the original dataset (TCGA-PANCAN32-L4.zip belongs to the level 4 data set, and the batch effect has been processed in the past) in TCPA (https://tcpaportal.org/tcpa, accessed on 11 January 2023), resulting in 687 tumor cases comprising 362 LUAD samples and 325 LUSC samples, with 258 proteins. The proteins that contained missing values (“NA”) in over 50% of the samples were removed, and the “NA” values for six proteins were replaced by their average values. This led to the creation of a proteome profiling dataset consisting of 687 tumor samples and 217 proteins. We also downloaded relevant clinical information, including survival information and phenotype information, for these 687 tumor samples from the TCGA data (11 January 2023).

2.2. Proteome Profiling Analysis

T-Distributed Stochastic Neighbor Embedding (t-SNE) is a non-linear technique that is particularly useful for visualizing high-dimensional data in low-dimensional space [17]. Principal Component Analysis (PCA) is a linear method that aims to reduce the number of features in a dataset while retaining the maximum amount of information. It can be used to capture the most important patterns in the data and visualization of the classification performance. Unlike PCA, t-SNE is not a linear method and instead tries to preserve the local structure of the data in the low-dimensional space.
We utilized t-SNE and PCA techniques based on the “Rtsne” [18] and “FactoMineR” [19] packages to visualize the differences between the original proteins and the selected proteomic biomarkers.

2.3. Using WBFS to Obtain Candidate Protein Biomarkers

High-dimensional datasets have stimulated the development of feature selection techniques. In this section, some basics of information theory are outlined, and then a new feature selection algorithm based on information theory is proposed.
Mutual information is the overlapping information of two entropies and is commonly utilized to analyze the correlation and statistical dependency between two random variables. It has been employed in various feature selection algorithms for obtaining the redundancy between the input features X and the classifications Y as well as the correlation between any pair of features.
I ( X ; Y ) = x X y Y p ( x , y ) log p ( x , y ) p ( x ) p ( y )
where p ( x ) is the probability when X = x , and p ( x , y ) is the joint probability when X = x , Y = y .
Conditional mutual information is the mutual information of X and Y given the third variable Z and can be computed by:
I ( X ; Y | Z ) = z Z p ( z k ) x X y Y p ( x , y | z ) log p ( x , y | z ) p ( x | z ) p ( y | z )
where p ( x | z ) is the conditional probability of X = x when Z = z is known, and p ( x , y | z ) is the conditional probability of X = x , Y = y when Z = z is known.
As an extension of mutual information, 3-way interaction information denotes the information shared by three random variables X , Y , Z and can be determined using conditional information and mutual information:
I ( X ; Y ; Z ) = I ( X , Y ; Z ) I ( X ; Z ) I ( Y ; Z )
Interaction information can be positive, zero, or negative; it is positive when the two variables of X , Y together provide information that cannot be provided by any one of them individually. Therefore, I ( X ; Z ) < I ( X ; Z | Y ) is always true when X , Y have an interaction relationship.
Table 1 summarizes all symbols used for the WBFS formula derivation.
For a dataset D that has m features F = { f 1 , f 2 , , f m } and n samples, the aim of a feature selection method is to find an optimal feature subset F = { f 1 , f 2 , , f k } that can maximize the objective function J ( f k ) and yield the same or a better classification accuracy than the original feature set as much as possible, where k m and F F . The concepts of relevance, redundancy, complementarity, and interaction are critical for a feature selection approach. The relevance between feature and class can be expressed using the mutual information of I ( f i ; C ) ; f i is called a stronger correlated feature than f j when I ( f i ; C ) > I ( f j ; C ) . In addition, the mutual information I ( f i ; f j ) between pairs of features is called redundancy. Considering correlation alone is definitely not sufficient because redundant features may reduce the classification performance and increase computational complexity; therefore, the redundancy should be eliminated while a new candidate feature f k is selected into S from the candidate feature set F / S . However, there is still a flaw in the objective function J ( f k ) based only on relevance and redundancy due to the absence of supplementary information when a new candidate feature f k is introduced into S .
We used the following equation as the objective function to screen the candidate features individually:
J ( f k ) = arg max f k F S , f j S ω I ( f k ; C | f j )
Specifically, instead of using equal weight to define the influence of selected features on candidate features, a dynamic parameter ω is considered to evaluate the differences of the selected features and updated when a new feature f k is introduced into S .
For two interaction features, f k and f j , I ( f j ; C ) < I ( f j ; C | f k ) will be true if their interaction information is positive. In other words, the addition of feature f k will produce a positive influence in predicting C for f j . Hence, we can use I ( f k ; C | f j ) I ( f k ; C ) to measure the weight ω , which can be expressed as follows:
ω = 1 + 2 ( I ( f j ; C | f k ) I ( f j ; C ) ) H ( f j ) + H ( C )
Theorem 1.
0 ω < 2 .
Proof of Theorem 1.
For feature f k , f j and class C , we have:
0 I ( f j ; C ) H ( f j ) 0 I ( f j ; C ) H ( C ) 0 2 I ( f j ; C ) H ( f j ) + H ( C ) 0 I ( f j ; C | f k ) H ( f j ) 0 I ( f j ; C | f k ) H ( C ) 0 2 I ( f j ; C | f k ) H ( f j ) + H ( C )
Then,
0 2 I ( f j ; C ) H ( f j ) + H ( C ) 1 0 2 I ( f j ; C | f k ) H ( f j ) + H ( C ) 1 1 2 ( I ( f j ; C | f k ) I ( f j ; C ) ) H ( f j ) + H ( C ) 1
Hence,
0 ω = 1 + 2 ( I ( f j ; C | f k ) I ( f j ; C ) ) H ( f j ) + H ( C ) 2
Based on Equations (4) and (5), the final objective function of WBFS is shown as:
J W B F S ( f k ) = arg max f k F S , f j S ( 1 + 2 ( I ( f j ; C | f k ) I ( f j ; C ) ) H ( f j ) + H ( C ) ) I ( f k ; C | f j )
Our proposed method of WBFS consists of two components: ω and I ( f k ; C | f j ) . The first component considers not only the importance of features in S but also the influence of candidate feature f k F / S on the relevance between S and C . The strong positive interaction of features f k F / S and f j S , the larger this value. The second component, I ( f k ; C | f j ) , is conditional mutual information, which fully covers the relevance between candidate feature f k and class C based on the selected feature f j S . Algorithm 1 presents the pseudocode of the WBFS algorithm:
Algorithm 1 Weight-based feature selection (WBFS)
I n p u t :   F ,   S ,   C ,   K   K     F
O u t p u t :   F
I n i t i a l i z a t i o n :   S   ,   F   ;
f o r   i = 0 ; i < F ; i + + d o
   M I i   I f i ; C ;
e n d   f o r
S f     w h i c h   I f ; C = max M I ;
S 1 ;
w h i l e   S < K     d o
   f o r   k = 0 ; k < F \ S ; k + + d o
     J W B F S f k = a r g m a x f k F S , f j S 1 + 2 I ( f j ; C | f k ) I f j ; C H f j + H C I ( f k ; C | f j ) ;
     S f k ;
   e n d   f o r
e n d   w h i l e
F S ;
R e t u r n   F ;
In Algorithm 1, H ( f j ) and H ( C ) represent the entropy of the feature f j and the categorical variable C , respectively.
When we need to select the top K features from the original feature set F , the steps of the WBFS algorithm are as follows:
First, we obtain the mutual information between each feature in F and class C . Second, the feature f , which has the largest relevance among the features, is added into S . Third, for the candidate feature set F / S , Equation (6) is utilized to obtain the feature f k and add it to the selected feature set S . When the number of features in S is equal to K , S is the optimal feature set F .
The time consumption largely depends on the computation times of mutual information or conditional mutual information for information theory-based feature selection methods. For a dataset D with m samples and n features, the time complexity is O ( m ) when mutual information or conditional mutual information is calculated. When a new feature is introduced into the selected feature set, mutual information or conditional mutual information among all features must be calculated; the time complexity is O ( m n ) for one iteration. Therefore, for WBFS, the final time complexity is O ( k m n ) if the top k features are chosen from the original features. The MIFS [9], mRMR [12], CIFE [13] and CONDRED [14] methods have the same time complexity as WBFS; and the time complexity of MIM [10] is O ( m n ) because only the relevance between features and classes is considered.

2.4. Using Bayesian Networks to Discover Causalities

Causal discovery aims to find causal relations by analyzing observational data [20]. As a widely used graphical model in bioinformatics and employed for related aspects of classification and prediction, the Bayesian network (BN) provides a convenient and coherent way to represent uncertainty from a perspective of probability [21]. BN models can estimate the joint probability distribution P over a vector of random variables X = X 1 , , X n , and the joint probability distribution factorized as a product of several conditional distributions denotes the dependency or independency structure by a directed acyclic graph (DAG):
P ( X 1 , , X n ) = i = 1 n P ( X i | P a ( X i G ) )
where P ( X 1 , , X n ) is the joint probability, P a ( X i G ) are the parent nodes of X i and P ( X i | P a ( X i G ) ) is the conditional probability of X i when the parent nodes of X i are known. Tsamardinos et al. were the first to establish the connection between local causal discovery and feature selection [22]. Yu et al. developed a unified view of causal and non-causal feature selection methods based on the BN perspective, and their experimental results show that non-causal feature selection is more suitable for high-dimensional data, especially for gene sequencing data [23]. Therefore, the combination of non-causal feature selection and BN local causal discovery is a promising strategy for identifying key signatures of a target in a high-dimensional dataset. The regulatory network composed of the selected biomarkers defines the regulatory interactions among regulators and their potential targets and can be used for inference and prediction. We constructed a regulatory network using BN to explore the causality interaction between any two of the selected proteins using WBFS based on the Causal Learner toolbox, which is available on GitHub (https://github.com/z-dragonl/Causal-Learner, accessed on 11 January 2023) [24].

2.5. Receiver Operating Characteristic (ROC) and Survival Analysis

The candidate protein biomarkers obtained with WBFS were analyzed for ROC and survival curves in order to evaluate the potential applications of these candidate signatures as possible markers for lung cancer diagnosis and prognosis. Medcalc statistical software [25] was used to analyze the candidate proteins for distinguishing LUAD and LUSC samples, and the sensitivity, specificity, and area under the curve (AUC) values were obtained. A multiple regression model to evaluate the combination classification performance of specific set biomarkers was also assessed, and it can be expressed as follows:
m u l t i p l e   regression = 1 1 + e ( c o n s tan t + i = 1 n c o e f f i c i e n t i exp r e s s i o n i )
In the above Equation (8), the c o n s t a n t represents the value that would be predicted for the dependent variable if all the independent variables were simultaneously equal to zero, the c o e f f i c i e n t i is the slope of the linear relationship of the i - th protein, and the e x p r e s s i o n i is the expression value of the i - th protein.
For survival analysis, Kaplan–Meier survival curves were explored and visualized (using α = 0.05 as the test level) using the Survminer toolkit [26] in R based on age, gender, and the tumor, node, metastasis (TNM) stage (clinical information was collected in the TCGA dataset), as well as the expression pattern of candidate proteins in LUAD and LUSC subjects.

3. Results

3.1. The Distributions Overview of the LUAD and LUSC Tumor Samples

The t-stochastic neighbor embedding (t-SNE) algorithm and principal component analysis (PCA), using the “Rtsne” and “FactoMineR” packages, respectively, were employed to display the difference between the original proteins and the 10 proteomic biomarkers. Overviews of the original lung cancer dataset based on t-SNE and PCA are shown in Figure 1A,C, respectively. t-SNE and PCA were also performed based on the subset containing 10 signatures. As shown in Figure 1B, the samples using 10 protein signatures clustered significantly better than the 258 original proteins. From Figure 1C,D, we also observed that the PCA scores for Dim1 significantly increased from 14.4% to 36.4%.

3.2. Using WBFS Method Identifying Protein Signatures for Classifying the Two Cancer Subtypes

3.2.1. Evaluate WBFS Classification Performance Based on UCI Datasets

To evaluate the classification performance of the WBFS algorithm and the advantage in terms of prediction accuracy compared with other feature selection algorithms, 12 high-dimensional datasets with different numbers of features, samples, and classification labels from the UCI repository (https://archive.ics.uci.edu/mL/datasets.php, accessed on 11 January 2023) and additional studies were employed (https://github.com/jundongl/scikit-feature, accessed on 11 January 2023). We discretize continuous datasets using an equal-width strategy into five bins. Additionally, we introduced the ratio concept to evaluate the classification difficulty for each dataset, with a smaller value indicating a more challenging feature selection. For dataset D , N is the number of samples, m is the median arity of the features, and c is the number of classes. The ratio can be calculated by N / ( m c ) . All 12 datasets are described in Table 2.
The experiment was performed using MATLAB 2020 based on Windows 10 with a 1.6 GHz Intel Core CPU and 8 GB memory. Three types of classifiers, k-nearest neighbors (KNN) [27], naïve Bayes classifier (NBC) [28], and the Library for Support Vector Machines (LibSVM) [29], were used to test the prediction performance using selected features in the experiment. Ten-fold cross-validation was employed, in which each of these datasets was randomly partitioned into a training dataset (90%) and a testing dataset (10%) and fed to every feature selection algorithm 10 times. Five well-known feature selection algorithms were selected to be compared with the proposed WBFS method, including CONDRED [14], MIM [10], mRMR [12], DISR [30], and CIFE [13]. A t-test was used to determine the significant differences between two groups of classification accuracies, and a Win/Tie/Loss (W/T/L) paradigm was adopted for describing performance: Win means WBFS shows better than other baseline methods, Tie means that there is no statistically significant difference with other methods and other cases are classified as Loss. Specifically, the ‘*’ symbols and ‘v’ identify statistically significant (at the 0.1 level) wins or losses over the proposed WBFS method. The average accuracies, standard deviation, and W/T/L information of 12 datasets based on three classifiers are shown in Table 3, Table 4 and Table 5. The number of selected features for all methods was fixed at 15 and k = 5 for the KNN classifier.
In Table 3, Table 4 and Table 5 bold values indicate the largest value among the six feature selection methods. WBFS obtains the best average classification accuracies of 84.48%, 73.15%, and 83.01% and occupies first place on five datasets (Lung, Krvskp, Madelon, Musk, and Waveform), six datasets (Breast, Colon, Ionosphere, Krvskp, Sonar, Waveform), and six datasets (Breast, Krvskp, Madelon, Musk, Sonar, Waveform) based on the three classifiers KNN, NB and LibSVM, respectively. For the other methods of CONDRED [14], MIM [10], mRMR [12], DISR [30], and CIFE [13], the numbers of cases for which they achieve the highest classification accuracy are 1, 2, 2, 1, 1 with KNN; 1, 1, 0, 1, 3 with NBC; and 1, 1, 2, 1, 1 with LibSVM, respectively.
From the perspective of W/T/L, WBFS obtains the best result in most cases. According to Table 5, WBFS achieves significantly higher classification accuracy than the CONDRED, MIM, mRMR, DISR, and CIFE methods with case numbers of 7, 6, 3, 6, and 8. For Table 3 and Table 4, the numbers are 6, 5, 3, 5, 7 and 7, 4, 3, 4, 4.
Considering that the above six feature selection algorithms are all based on a feature ranking strategy, an effective method for further comparing their classification accuracies is to individually add features for learning. We employed this strategy using a range of 1–30 selected features. Figure 2 shows the comparative results based on six representative datasets (Lung, Madelon, Sonar, Musk, Krvskp and Waveform). For each subgraph, the x-axis refers to the first k selected features, and the y-axis represents the average classification accuracy of three classifiers for the top k selected features. In Figure 2, it is obvious that the proposed WBFS method shows better average accuracies than the others in most cases based on the Lung, Madelon, Sonar, and Musk datasets. For instance, WBFS obtains the highest average classification accuracy of 76.39% with only seven features, while accuracies of 68.50%, 67.44%, 70.40%, 74.31%, and 64.84% are achieved by MIM, mRMR, DISR, CIFE and CONDRED, respectively, with the same number of selected features. For Krvskp and Waveform, the accuracies of each method are comparable because their ratio value is sufficiently large; WBFS still obtains the best accuracy in most cases.
In addition, we also compared the classification accuracy of feature selection subsets obtained by the WBFS algorithm and ensemble learning algorithm Random Forest (RF) based on the Isolet, Sonar, Waveform, Madelon, and Splice datasets, as shown in Figure S1. The results suggest that both methods have similar classification accuracy on low-dimensional datasets such as Waveform, Sonar, and Splice. However, WBFS demonstrates a clear advantage in classification accuracy on high-dimensional datasets such as Madelon and Isolet.
All of the above experiments demonstrate that the WBFS technique performs better than the other methods in terms of classification accuracy, which lays the foundation for its application in bioinformatics data.

3.2.2. Using WBFS to Obtain the Top 10 Candidate Biomarkers

We discretized the expression data into three bins by utilizing an equal-width strategy for WBFS according to the three expression states of protein (high expression, normal expression, and low expression). The top 10 proteins (MIG6, GAPDH, NDRG1_pT346, BRD4, CD26, TFRC, INPP4B, GSK3ALPHABETA, IGFBP2, and DUSP4) were screened by applying the WBFS algorithm, and they can be regarded as the candidate biomarkers that are most closely associated with the classification of lung cancer subtypes.

3.3. Using Bayesian Networks to Discover Causalities

In Figure 3, we can observe that the direct determinants for differentiating the subtypes of lung cancer include the protein markers MIG6, NDRG1_pT346, BRD4, CD26, INPP4B, and DUSP4, while the others can be considered as having an indirect impact on the classification outcome. Therefore, the Bayesian network (BN) provides significant factors for mechanistic studies at the molecular level.

3.4. ROC and Survival Analysis of the Candidate Protein Signatures

From Figure 4 and Table 6, it is possible to accurately distinguish LUAD and LUSC by each of the 10 protein biomarkers, as indicated by the AUC, which had a value up to 0.72 (p < 0.001). Multiple regression results of the 10 biomarkers indicated that eight protein markers, MIG6, GAPDH, NDRG1_pT346, BRD4, CD26, TFRC, INPP4B, and DUSP4, had significant differentiation abilities in disease classification between the LUSC and LUAD groups, as shown in Table S1. The detailed results of multiple regression analysis are shown in Table 6. The combination of the eight features achieved the best classification performance (AUC = 0.960, p < 0.001).
Figure 5 shows the Kaplan–Meier survival curves of the above factors, which demonstrates that MIG6, TFRC, INPP4B and IGFBP2 factors have a significant impact on the overall survival rate of LUAD patients (p < 0.05), and the MIG6, TFRC, and INPP4B factors have a significant negative prognosis correlation (p < 0.05). For LUSC patients, the factors of MIG6, GAPDH, CD26, and TFRC are all significantly correlated with the overall survival rate of patients (p < 0.05), and TFRC was the only factor that has a significant positive prognosis correlation (p < 0.05).

4. Discussion

Lung cancer is one of the most common malignant tumors worldwide. Accurate classification of lung cancers into clinically significant subtypes is of utmost importance, making it critical for clinical and scientific researchers to discover new molecular markers and potential diagnostic and therapeutic targets [6,31]. Additionally, it is worth noting that the TCPA platform has already standardized protein expression information, while the TCGA database catalogs and explores cancer-causing genomic alterations, establishing a comprehensive “atlas” of cancer genomic signatures. By integrating the information from TCPA and TCGA, we could identify potential diagnostic and prognostic biomarkers for lung tumor subtypes to help understand the underlying mechanisms of tumorigenesis and improve approaches or standards for cancer diagnosis and therapy.
Feature selection is a fundamental technique for achieving efficient data reduction in high-dimensional datasets, such as those used in biological and text data mining [8,32,33]. This strategy is crucial for identifying accurate models and key factors in classification or clustering tasks [34,35]. Many studies have focused on developing feature selection strategies that return an optimal feature subset. Information theory-based approaches, such as mutual information and its extensions, can efficiently evaluate the relevance, redundancy, or interaction information [36,37,38]. However, most proposed approaches concentrate primarily on relevance or redundancy without considering interaction information. In this paper, we present a new weight-based feature selection (WBFS) method that considers the weight between selected features and candidate features based on information theory. Our proposed algorithm outperforms other methods in terms of classification performance, as demonstrated through experiments on 12 datasets and three classifiers.
Feature selection methods have an interesting application in biomarker discovery from high-throughput genomics data, which can guide clinical diagnosis by identifying the most discriminative biomarkers for a particular problem. In this study, the WBFS method was applied to real-world lung cancer datasets to obtain key protein biomarkers for the classification of lung tumor subtypes. The causalities among the selected biomarkers were also explored using the Bayesian network (BN) strategy. The study found six promising protein biomarkers that were directly correlated with the classification of lung cancer subtypes, including MIG6, NDRG1_pT346, BRD4, CD26, INPP4B, and DUSP4. ROC analysis and survival curves based on selected proteins were also implemented. The ROC analysis has shown that 10 protein biomarkers were able to accurately distinguish between LUAD and LUSC with an AUC value of up to 0.72 and that eight of the protein markers had significant utility in disease classification. When combined, the eight features produced an AUC of 0.960, indicating strong classification performance. Survival curves showed that MIG6, TFRC, INPP4B, and IGFBP2 had a significant impact on the overall survival rate of LUAD patients, while MIG6, GAPDH, CD26, and TFRC were all significantly correlated with the overall survival rate of LUSC patients. A study [39] demonstrated that MIG6 is essential for controlling the regulation of cell signaling in lung cancer cells.
This work involves some limitations that are worth noting. Firstly, our WBFS method is based on the feature ranking strategy, where the selection of the number of features k is critical for achieving optimal classification performance and accuracy. Unfortunately, there is currently no effective strategy for determining the optimal value of k . Moreover, in the era of big data, it is more important than ever to identify the causal relationships among large datasets. However, most of the proposed feature selection methods are non-causal, and thus, it is necessary to establish a connection between non-causal feature selection and causality inference. In our future work, we plan to focus on two main aspects:
(1)
Developing a more efficient feature selection algorithm that can automatically determine the optimal number of selected features based on the intrinsic dimension of a high-dimensional dataset.
(2)
Investigating the relationship between causal and non-causal feature selection methods and applying non-causal feature selection methods to the Bayesian network. This will facilitate Bayesian network structure learning based on high-dimensional data.

5. Conclusions

In this paper, we developed a new weight-based feature selection (WBFS) method to identify protein biomarkers to distinguish between two different subtypes (LUAD and LUSC) of lung cancer. Additionally, we further explored the direct influencing factors for lung cancer subtype classification using Bayesian networks. Six promising protein biomarkers (MIG6, NDRG1_pT346, BRD4, CD26, INPP4B, and DUSP4) specifically contributed to the subtype classification. These biomarkers, along with others, were able to accurately distinguish between LUAD and LUSC with a high AUC value. The identification and characterization of these biomarkers hold enormous potential for improving the diagnosis, treatment, and outcomes of lung cancer patients.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/e25071003/s1, Table S1: Multiple regression results of the selected features in disease classification, Figure S1: BWFS vs. RF classification accuracy comparison based on five datasets: (A) Waveform, (B) Splice, (C) Sonar, (D) Madelon, (E) Isolet. The X axis shows the number of selected features, and the Y axis shows the average classification accuracy of three classifiers (KNN, NBC and LibSVM).

Author Contributions

Conceptualization, Y.W. and J.W.; methodology, Y.W. and X.R.; validation, X.R. and P.S.; writing, original draft preparation, Y.W.; writing, review and editing, X.G. and J.W.; supervision, J.W. and X.G.; funding acquisition, X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 81900134.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available at https://github.com/jihanwang/WBFS (accessed on 11 January 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Relli, V.; Trerotola, M.; Guerra, E.; Alberti, S. Abandoning the notion of non-small cell lung cancer. Trends Mol. Med. 2019, 25, 585–594. [Google Scholar] [CrossRef]
  2. Li, J.; Lu, Y.; Akbani, R.; Ju, Z.; Roebuck, P.L.; Liu, W.; Yang, J.-Y.; Broom, B.M.; Verhaak, R.G.W.; Kane, D.W.; et al. TCPA: A resource for cancer functional proteomics data. Nat. Methods 2013, 10, 1046–1047. [Google Scholar] [CrossRef] [Green Version]
  3. Lv, J.; Zhu, Y.; Ji, A.; Zhang, Q.; Liao, G. Mining TCGA database for tumor mutation burden and their clinical significance in bladder cancer. Biosci. Rep. 2020, 40, BSR20194337. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Yan, H.; Qu, J.; Cao, W.; Liu, Y.; Zheng, G.; Zhang, E.; Cai, Z. Identification of prognostic genes in the acute myeloid leukemia immune microenvironment based on TCGA data analysis. Cancer Immunol. Immunother. 2019, 68, 1971–1978. [Google Scholar] [CrossRef]
  5. Song, X.-F.; Zhang, Y.; Guo, Y.-N.; Sun, X.-Y.; Wang, Y.-L. Variable-size cooperative coevolutionary particle swarm optimization for feature selection on high-dimensional data. IEEE Trans. Evol. Comput. 2020, 24, 882–895. [Google Scholar] [CrossRef]
  6. Kumar, S.; Patnaik, S.; Dixit, A. Predictive models for stage and risk classification in head and neck squamous cell carcinoma (HNSCC). PeerJ 2020, 8, e9656. [Google Scholar] [CrossRef]
  7. Torres, R.; Judson-Torres, R.L. Research techniques made simple: Feature selection for biomarker discovery. J. Investig. Dermatol. 2019, 139, 2068–2074.e1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Vergara, J.R.; Estévez, P.A. A review of feature selection methods based on mutual information. Neural Comput. Appl. 2014, 24, 175–186. [Google Scholar] [CrossRef]
  9. Battiti, R. Using mutual information for selecting features in supervised neural net learning. IEEE Trans. Neural Netw. 1994, 5, 537–550. [Google Scholar] [CrossRef] [Green Version]
  10. Lewis, D.D. Feature Selection and Feature Extraction for Text Categorization. In Proceedings of the Speech and Natural Language: Proceedings of a Workshop Held at Harriman, Harriman, NY, USA, 23–26 February 1992. [Google Scholar]
  11. Kwak, N.; Choi, C.-H. Input feature selection for classification problems. IEEE Trans. Neural Netw. 2002, 13, 143–159. [Google Scholar] [CrossRef]
  12. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef] [PubMed]
  13. Lin, D.; Tang, X. Conditional Infomax Learning: An Integrated Framework for Feature Extraction and Fusion. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2006; pp. 68–82. [Google Scholar]
  14. Brown, G.; Pocock, A.; Zhao, M.-J.; Luján, M. Conditional likelihood maximisation: A unifying framework for information theoretic feature selection. J. Mach. Learn. Res. 2012, 13, 27–66. [Google Scholar]
  15. Wan, J.; Chen, H.; Li, T.; Yang, X.; Sang, B. Dynamic interaction feature selection based on fuzzy rough set. Inf. Sci. 2021, 581, 891–911. [Google Scholar] [CrossRef]
  16. Nakariyakul, S. A hybrid gene selection algorithm based on interaction information for microarray-based cancer classification. PLoS ONE 2019, 14, e0212333. [Google Scholar] [CrossRef]
  17. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  18. Krijthe, J.; van der Maaten, L.; Krijthe, M.J.; Package ‘Rtsne’. R Package Version 0.13 2017URL. 2018. Available online: https://github.com/jkrijthe/Rtsne (accessed on 11 January 2023).
  19. Lê, S.; Josse, J.; Husson, F. FactoMineR: An R package for multivariate analysis. J. Stat. Softw. 2008, 25, 1–18. [Google Scholar] [CrossRef] [Green Version]
  20. Glymour, C.; Zhang, K.; Spirtes, P. Review of causal discovery methods based on graphical models. Front. Genet. 2019, 10, 524. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Chen, S.H.; Pollino, C.A. Good practice in Bayesian network modelling. Environ. Model. Softw. 2012, 37, 134–135. [Google Scholar] [CrossRef]
  22. Tsamardinos, I.; Aliferis, C.F. Towards Principled Feature Selection: Relevancy, Filters and Wrappers. In International Workshop on Artificial Intelligence and Statistics; Christopher, M.B., Brendan, J.F., Eds.; PMLR, Proceedings of Machine Learning Research: Philadelphia, PA, USA, 2003; Volume R4, pp. 300–307. [Google Scholar]
  23. Yu, K.; Guo, X.; Liu, L.; Li, J.; Wang, H.; Ling, Z.; Wu, X. Causality-based feature selection: Methods and evaluations. ACM Comput. Surv. 2020, 53, 1–36. [Google Scholar] [CrossRef]
  24. Ling, Z.; Yu, K.; Zhang, Y.; Liu, L.; Li, J. Causal learner: A toolbox for causal structure and markov blanket learning. Pattern Recognit. Lett. 2022, 163, 92–95. [Google Scholar] [CrossRef]
  25. Schoonjans, F.; Zalata, A.; Depuydt, C.; Comhaire, F. MedCalc: A new computer program for medical statistics. Comput. Methods Programs Biomed. 1995, 48, 257–262. [Google Scholar] [CrossRef] [PubMed]
  26. Kassambara, A.; Kosinski, M.; Biecek, P.; Fabian, S. Survminer: Drawing Survival Curves Using ‘ggplot2′, R Package version 0.3; R Core Team: Vienna, Austria, 2017; p. 1. [Google Scholar]
  27. Kramer, O.; Kramer, O. K-nearest neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors; Springer: Berlin/Heidelberg, Germany, 2013; pp. 13–23. [Google Scholar]
  28. Leung, K.M. Naive bayesian classifier. Polytech. Univ. Dep. Comput. Sci./Financ. Risk Eng. 2007, 2007, 123–156. [Google Scholar]
  29. Chang, C.-C.; Lin, C.-J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  30. Meyer, P.E.; Bontempi, G. On the use of variable complementarity for feature selection in cancer classification. In Proceedings of the Applications of Evolutionary Computing: EvoWorkshops 2006: EvoBIO, EvoCOMNET, EvoHOT, EvoIASP, EvoINTERACTION, EvoMUSART, and EvoSTOC, Budapest, Hungary, 10–12 April 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 91–102. [Google Scholar]
  31. Kumari, P.; Kumar, S.; Sethy, M.; Bhue, S.; Mohanta, B.K.; Dixit, A. Identification of therapeutically potential targets and their ligands for the treatment of OSCC. Front. Oncol. 2022, 12, 910494. [Google Scholar] [CrossRef] [PubMed]
  32. Wang, J.; Wang, Y.; Zeng, Y.; Huang, D. Feature selection approaches identify potential plasma metabolites in postmenopausal osteoporosis patients. Metabolomics 2022, 18, 86. [Google Scholar] [CrossRef]
  33. Wang, Y.; Gao, X.; Ru, X.; Sun, P.; Wang, J. A hybrid feature selection algorithm and its application in bioinformatics. PeerJ Comput. Sci. 2022, 8, e933. [Google Scholar] [CrossRef]
  34. Gnana, D.A.A.; Balamurugan, S.A.A.; Leavline, E.J. Literature review on feature selection methods for high-dimensional data. Int. J. Comput. Appl. 2016, 136, 9–17. [Google Scholar]
  35. Llamedo, M.; Martínez, J.P. Heartbeat Classification Using Feature Selection Driven by Database Generalization Criteria. IEEE Trans. Biomed. Eng. 2011, 58, 616–625. [Google Scholar] [CrossRef]
  36. Koller, D.; Sahami, M. Toward Optimal Feature Selection; Stanford InfoLab: Stanford, CA, USA, 1996. [Google Scholar]
  37. Guo, B.; Nixon, M.S. Gait feature subset selection by mutual information. IEEE Trans. Syst. MAN Cybern.-Part A Syst. Hum. 2008, 39, 36–46. [Google Scholar]
  38. Ircio, J.; Lojo, A.; Mori, U.; Lozano, J.A. Mutual information based feature subset selection in multivariate time series classification. Pattern Recognit. 2020, 108, 107525. [Google Scholar] [CrossRef]
  39. Walsh, A.M.; Lazzara, M.J. Regulation of EGFR trafficking and cell signaling by Sprouty2 and MIG6 in lung cancer cells. J. Cell Sci. 2013, 126, 4339–4348. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Overview of the proteome profiling across lung adenocarcinoma (LUAD) and lung squamous cell carcinoma (LUSC) samples. The subgraphs (A,B) are the results performing t-SNE; the subgraphs (C,D) are the results performing PCA; (A) t-SNE and (C) PCA showed the clusters of samples based on whole proteome profiling; (B) t-SNE and (D) PCA showed the clusters of samples based on the top 10 selected protein profiling.
Figure 1. Overview of the proteome profiling across lung adenocarcinoma (LUAD) and lung squamous cell carcinoma (LUSC) samples. The subgraphs (A,B) are the results performing t-SNE; the subgraphs (C,D) are the results performing PCA; (A) t-SNE and (C) PCA showed the clusters of samples based on whole proteome profiling; (B) t-SNE and (D) PCA showed the clusters of samples based on the top 10 selected protein profiling.
Entropy 25 01003 g001
Figure 2. Average classification accuracy based on three classifiers vs. different number of selected features for six datasets: (A) Lung, (B) Madelon, (C) Sonar, (D) Musk, (E) Krvskp, (F) Waveform.
Figure 2. Average classification accuracy based on three classifiers vs. different number of selected features for six datasets: (A) Lung, (B) Madelon, (C) Sonar, (D) Musk, (E) Krvskp, (F) Waveform.
Entropy 25 01003 g002
Figure 3. The Bayesian network (BN) for the top 10 selected signatures and class label based on lung tumor dataset. These directed arrows represent the causal relationship between every two nodes. The cyan nodes are the key factors that are critical for subtype classification, and the yellow node represents the lung subtype classification.
Figure 3. The Bayesian network (BN) for the top 10 selected signatures and class label based on lung tumor dataset. These directed arrows represent the causal relationship between every two nodes. The cyan nodes are the key factors that are critical for subtype classification, and the yellow node represents the lung subtype classification.
Entropy 25 01003 g003
Figure 4. ROC curves of the 10 selected protein features and the combination test of 8 biomarkers. (A) shows the ROC curves of the 6 protein features of BRD4, CD26, DUSP4, GAPDH, GSK3ALPHABETA and IGFBP2; (B) shows the ROC curves of the 4 protein features of INPP4B, MIG6, NDRG1_pT346, TFRC and the combination test curve of 8 biomarkers.
Figure 4. ROC curves of the 10 selected protein features and the combination test of 8 biomarkers. (A) shows the ROC curves of the 6 protein features of BRD4, CD26, DUSP4, GAPDH, GSK3ALPHABETA and IGFBP2; (B) shows the ROC curves of the 4 protein features of INPP4B, MIG6, NDRG1_pT346, TFRC and the combination test curve of 8 biomarkers.
Entropy 25 01003 g004
Figure 5. The Kaplan–Meier curves for lung adenocarcinoma (LUAD) (AD) and lung squamous cell carcinoma (LUSC) (EH). The values of all protein expressions were divided into high and low groups by cut-off values. The horizontal axis represents the survival time (day), and the vertical axis represents the overall survival rate.
Figure 5. The Kaplan–Meier curves for lung adenocarcinoma (LUAD) (AD) and lung squamous cell carcinoma (LUSC) (EH). The values of all protein expressions were divided into high and low groups by cut-off values. The horizontal axis represents the survival time (day), and the vertical axis represents the overall survival rate.
Entropy 25 01003 g005
Table 1. The symbols used in feature selection method.
Table 1. The symbols used in feature selection method.
SymbolsDescriptionSymbolsDescription
D Dataset n The sample size of D
F Original feature set m The feature size of F
C Class labels | S | The feature size of S
S The selected feature subset K The size of F
F Optimal feature subset f j Feature number j
f i Feature number i F / S The candidate feature set
f k The candidate feature J ( f k ) The objective function
Table 2. Description of 12 datasets for experiment.
Table 2. Description of 12 datasets for experiment.
Dataset Name.Dataset Source# Instances# Features# ClassesRatio
Lunghttps://github.com/jundongl/scikit-feature (accessed on 11 January 2023)7332574
Colonhttps://github.com/jundongl/scikit-feature (accessed on 11 January 2023)622000210
Isolethttps://github.com/jundongl/scikit-feature (accessed on 11 January 2023)15606172612
Sonarhttps://archive.ics.uci.edu/dataset/151/connectionist+bench+sonar+mines+vs+rocks (accessed on 11 January 2023)20860220
Ionospherehttps://archive.ics.uci.edu/dataset/52/ionosphere (accessed on 11 January 2023)35134235
Breasthttps://archive.ics.uci.edu/dataset/17/breast+cancer+wisconsin+diagnostic (accessed on 11 January 2023)56930257
Landsathttps://archive.ics.uci.edu/dataset/146/statlog+landsat+satellite (accessed on 11 January 2023)6435366215
Madelonhttps://github.com/jundongl/scikit-feature (accessed on 11 January 2023)26005002260
Splicehttps://archive.ics.uci.edu/dataset/69/molecular+biology+splice+junction+gene+sequences (accessed on 11 January 2023)3175603265
Waveformhttps://archive.ics.uci.edu/dataset/108/waveform+database+generator+version+2 (accessed on 11 January 2023)5000403333
Krvskphttps://archive.ics.uci.edu/mL/datasets/Chess+(King-Rook+vs.+King-Pawn) (accessed on 11 January 2023)3196362533
Muskhttps://archive.ics.uci.edu/dataset/75/musk+version+2 (accessed on 11 January 2023)65981662660
Table 3. Accuracy (%) of selected features using the k-nearest neighbors (KNN) algorithm.
Table 3. Accuracy (%) of selected features using the k-nearest neighbors (KNN) algorithm.
No.DatasetCONDREDMIMmRMRDISRCIFEWBFS
1Lung46.43 ± 16.94 *77.68 ± 18.08 *85.71 ± 15.0686.07 ± 14.7261.25 ± 16.79 *88.93 ± 8.94
2Breast93.32 ± 3.07 *95.61 ± 2.7795.08 ± 1.6193.33 ± 3.94 *92.79 ± 2.67 *95.43 ± 2.06
3Colon83.81 ± 13.6587.14 ± 13.0590.24 ± 11.53 v83.81 ± 13.6587.14 ± 6.8585.48 ± 12.27
4Ionosphere85.46 ± 5.1486.32 ± 4.2485.44 ± 4.9786.32 ± 5.02 v85.17 ± 6.1784.60 ± 5.11
5Isolet20.06 ± 2.74 *29.74 ± 3.50 *65.06 ± 4.29 v57.69 ± 3.32 *50.90 ± 4.00 *59.94 ± 3.45
6Krvskp88.95 ± 1.69 *96.25 ± 0.9494.15 ± 1.26 *93.37 ± 1.04 *94.4 ± 1.27 *96.25 ± 0.94
7Landsat84.51 ± 1.5483.03 ± 1.12 *84.79 ± 1.1084.16 ± 1.6485.30 ± 1.1684.48 ± 1.76
8Madelon76.38 ± 2.43 *76.85 ± 2.18 *57.5 ± 3.82 *81.42 ± 1.78 *81.65 ± 3.0782.92 ± 1.84
9Musk50.85 ± 2.18 *87.06 ± 1.00 *83.13 ± 0.95 *69.25 ± 1.82 *76.66 ± 1.82 *91.66 ± 1.37
10Sonar75.45 ± 9.1983.17 ± 10.5178.33 ± 8.5983.64 ± 9.1280.24 ± 8.9981.17 ± 10.29
11Splice82.87 ± 1.7882.43 ± 2.0282.43 ± 2.0282.43 ± 2.0279.97 ± 2.47 *82.43 ± 2.02
12Waveform80.52 ± 1.4480.52 ± 1.4480.52 ± 1.4480.52 ± 1.4469.94 ± 2.02 *80.52 ± 1.44
Average72.3880.4881.8781.8378.7884.48
W/T/L6/6/05/7/03/7/25/6/17/5/0
The ‘*’ symbols and ‘v’ identify statistically significant (at the 0.1 level) wins or losses over the proposed WBFS method.
Table 4. Accuracy (%) of selected features using the naïve Bayes classifier (NBC) algorithm.
Table 4. Accuracy (%) of selected features using the naïve Bayes classifier (NBC) algorithm.
No.DatasetCONDREDMIMmRMRDISRCIFEWBFS
1Lung51.96 ± 24.4358.75 ± 24.6961.61 ± 23.3561.61 ± 23.3560.18 ± 22.5960.18 ± 22.11
2Breast89.99 ± 3.60 *92.79 ± 3.1493.32 ± 2.8462.74 ± 6.29 *92.79 ± 2.2693.32 ± 2.96
3Colon80.48 ± 10.75 *88.57 ± 13.6588.57 ± 13.6588.57 ± 13.6569.05 ± 18.58 *88.81 ± 11.06
4Ionosphere72.66 ± 9.70 *64.10 ± 6.04 *64.10 ± 6.04 *64.10 ± 6.04 *80.35 ± 10.45 *84.62 ± 7.75
5Isolet22.24 ± 3.96 *24.17 ± 2.99 *26.92 ± 4.46 *16.79 ± 2.80 *55.83 ± 3.29 v43.08 ± 4.35
6Krvskp52.22 ± 1.8852.22 ± 1.8852.22 ± 1.8852.22 ± 1.8852.22 ± 1.8852.22 ± 1.88
7Landsat75.46 ± 1.71 *72.82 ± 1.76 *76.53 ± 1.8576.64 ± 1.5277.62 ± 2.16 v76.38 ± 1.94
8Madelon59.58 ± 3.1059.62 ± 2.36 v58.96 ± 2.3159.19 ± 2.5259.15 ± 3.6559.12 ± 2.15
9Musk73.78 ± 1.49 *77.39 ± 1.79 *61.56 ± 1.60 *82.69 ± 0.96 v88.88 ± 0.93 v78.87 ± 1.35
10Sonar64.79 ± 13.19 *70.12 ± 15.7466.74 ± 16.0668.67 ± 15.47 *71.57 ± 9.6173.48 ± 16.34
11Splice88.69 ± 8.3888.62 ± 8.3788.62 ± 8.3788.62 ± 8.3787.74 ± 7.96 *88.62 ± 8.37
12Waveform79.08 ± 1.2679.08 ± 1.2679.08 ± 1.2679.08 ± 1.2677.34 ± 0.91 *79.08 ± 1.26
Average67.58 69.0268.1966.7472.7373.15
W/T/L7/5/04/7/13/9/04/7/14/5/3
The ‘*’ symbols and ‘v’ identify statistically significant (at the 0.1 level) wins or losses over the proposed WBFS method.
Table 5. Accuracy (%) of selected features using the Library for Support Vector Machines (LibSVM) algorithm.
Table 5. Accuracy (%) of selected features using the Library for Support Vector Machines (LibSVM) algorithm.
No.DatasetCONDREDMIMmRMRDISRCIFEWBFS
1Lung43.75 ± 24.1246.25 ± 24.69 *41.25 ± 27.3646.25 ± 24.69 v28.93 ± 23.28 *42.5 ± 24.09
2Breast93.32 ± 2.60 *94.38 ± 2.7295.25 ± 1.9092.97 ± 2.35 *92.79 ± 3.25 *95.42 ± 2.40
3Colon62.86 ± 22.1469.29 ± 19.97 v64.52 ± 23.0764.52 ± 23.0764.52 ± 23.0764.52 ± 23.07
4Ionosphere83.47 ± 6.99 *88.60 ± 4.4889.16 ± 4.24 v88.90 ± 4.72 *87.46 ± 4.7187.18 ± 5.25
5Isolet28.33 ± 2.36 *33.78 ± 4.34 *68.21 ± 2.92 v61.15 ± 1.87 *61.79 ± 2.59 *64.68 ± 3.46
6Krvskp75.28 ± 2.55 *77.63 ± 2.1072.37 ± 2.39 *72.03 ± 2.38 *72.59 ± 2.03 *77.63 ± 2.10
7Landsat82.77 ± 1.4080.57 ± 1.56 *82.78 ± 1.4082.81 ± 1.6283.37 ± 1.35 v82.69 ± 1.50
8Madelon65.85 ± 1.41 *65.62 ± 1.28 *60.35 ± 1.57 *67.77 ± 1.94 *67.81 ± 2.56 *69.50 ± 2.39
9Musk84.59 ± 1.27 *84.60 ± 1.23 *84.59 ± 1.27 *84.59 ± 1.27 *90.21 ± 0.93 *91.12 ± 1.10
10Sonar64.83 ± 8.47 *74.00 ± 12.71 *76.4 ± 13.5674.95 ± 12.3575.40 ± 10.3176.88 ± 11.64
11Splice89.39 ± 2.2389.17 ± 2.0789.17 ± 2.0789.17 ± 2.0787.62 ± 1.78 *89.17 ± 2.07
12Waveform83.62 ± 1.9783.62 ± 1.9783.62 ± 1.9783.62 ± 1.9777.32 ± 1.10 *83.62 ± 1.97
Average72.38 80.48 81.87 81.83 78.78 83.01
W/T/L7/5/06/5/13/7/26/5/18/3/1
The ‘*’ symbols and ‘v’ identify statistically significant (at the 0.1 level) wins or losses over the proposed WBFS method.
Table 6. The Receiver Operating Characteristic (ROC) analysis of 10 selected protein biomarkers and a combination test of 8 biomarkers.
Table 6. The Receiver Operating Characteristic (ROC) analysis of 10 selected protein biomarkers and a combination test of 8 biomarkers.
No.Protein BiomarkerAUCSensitivitySpecificityp-Value
1BRD40.74959.3882.60<0.0001
2CD260.82475.7075.40<0.0001
3DUSP40.73367.4067.40<0.0001
4GAPDH0.77573.5067.70<0.0001
5GSK3ALPHABETA0.76566.5075.40<0.0001
6IGFBP20.72265.2071.80<0.0001
7INPP4B0.76774.2066.90<0.0001
8MIG60.83380.0072.70<0.0001
9NDRG1_pT3460.72962.2076.50<0.0001
10TFRC0.87681.2080.90<0.0001
11The combination of eight biomarkers0.96090.1592.54<0.0001
The detailed information of the eight biomarkers in combination test are shown in Table S1.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Gao, X.; Ru, X.; Sun, P.; Wang, J. The Weight-Based Feature Selection (WBFS) Algorithm Classifies Lung Cancer Subtypes Using Proteomic Data. Entropy 2023, 25, 1003. https://doi.org/10.3390/e25071003

AMA Style

Wang Y, Gao X, Ru X, Sun P, Wang J. The Weight-Based Feature Selection (WBFS) Algorithm Classifies Lung Cancer Subtypes Using Proteomic Data. Entropy. 2023; 25(7):1003. https://doi.org/10.3390/e25071003

Chicago/Turabian Style

Wang, Yangyang, Xiaoguang Gao, Xinxin Ru, Pengzhan Sun, and Jihan Wang. 2023. "The Weight-Based Feature Selection (WBFS) Algorithm Classifies Lung Cancer Subtypes Using Proteomic Data" Entropy 25, no. 7: 1003. https://doi.org/10.3390/e25071003

APA Style

Wang, Y., Gao, X., Ru, X., Sun, P., & Wang, J. (2023). The Weight-Based Feature Selection (WBFS) Algorithm Classifies Lung Cancer Subtypes Using Proteomic Data. Entropy, 25(7), 1003. https://doi.org/10.3390/e25071003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop