Next Article in Journal
The GHSZ Argument: A Gedankenexperiment Requiring More Denken
Previous Article in Journal
Towards the Prediction of Rearrest during Out-of-Hospital Cardiac Arrest
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency Criteria

1
School of Computer Science, Wuhan University, Wuhan 430072, China
2
Department of Information System, Faculty of Computers and Informatics, Suez Canal University, Ismailia 41522, Egypt
3
Department of Computer Science and Information Technology, La Trobe University, Melbourne 3086, Australia
*
Authors to whom correspondence should be addressed.
Entropy 2020, 22(7), 757; https://doi.org/10.3390/e22070757
Submission received: 7 May 2020 / Revised: 13 June 2020 / Accepted: 28 June 2020 / Published: 9 July 2020

Abstract

:
The main challenge of classification systems is the processing of undesirable data. Filter-based feature selection is an effective solution to improve the performance of classification systems by selecting the significant features and discarding the undesirable ones. The success of this solution depends on the extracted information from data characteristics. For this reason, many research theories have been introduced to extract different feature relations. Unfortunately, traditional feature selection methods estimate the feature significance based on either individually or dependency discriminative ability. This paper introduces a new ensemble feature selection, called fuzzy feature selection based on relevancy, redundancy, and dependency (FFS-RRD). The proposed method considers both individually and dependency discriminative ability to extract all possible feature relations. To evaluate the proposed method, experimental comparisons are conducted with eight state-of-the-art and conventional feature selection methods. Based on 13 benchmark datasets, the experimental results over four well-known classifiers show the outperformance of our proposed method in terms of classification performance and stability.

1. Introduction

Nowadays, classification systems have a lot of contributions in different domains such as bioinformatics, medical analysis, text categorization, pattern recognition, and intrusion detection [1]. The main challenge of these systems is to deal with high dimensionality data, which may include redundant or irrelevant features [2]. These features have a negative effect on classification systems which can lead to (1) reducing the classification accuracy, (2) reducing the classification speed, (3) increasing the classification complexity. To overcome these limitations, features selection introduces an effective solution to reduce the dimensionality of data by selecting the significant features and discarding the undesirable ones [3].
Feature selection methods are divided into three categories: filter [4], embedded [5], and wrapper [6]. These methods can be also classified into two groups according to the role of classifiers in the feature selection process: classification-independent (filter method), and classification dependent group (embedded, and wrapper method) [3]. The former depends only on the data characteristics without considering classifiers in the selection process, while the latter depends on classifiers to assess the significance of features in the selection process. Although the classification-dependent group can return the best feature selection subset, it requires more computational cost as a result of the classification process. Moreover, the selected features related only to the used classifier in the feature selection process. For this reason, classification-independent is more practical for high dimensionality data [7]. In this study, filter feature selection is our interest rather than embedded and wrapper due to its benefits such as simplicity, practicality, scalability, efficiency, and generality [8].
The success of filter methods depends on the amount of extracted information from data characteristics [9]. Motivated by this hypothesis, many theories have been introduced to find the best filter feature selection method such as information theory [10], and rough set theory [11]. Information theory measures can rank the features not only according to their relevancy to class but also with respect to the redundancy of features [12]. Moreover, These measures outperform other measures as correlation due to its ability to deal with linear and non-linear relations [3]. Rough set theory can select a subset of features according to their dependency to class [13]. The main advantages of rough set measures are simplicity, and no user-defined parameter is required. However, the traditional measures of these theories share common limitation, they can not deal directly with continuous features. To overcome this limitation, many research studies have been extended by integrating the previous theories with fuzzy set theory [14,15,16]. Feature selection based fuzzy sets is not only suitable for any kind of data but also extracts more information from classes compared with the traditional feature selection methods [14]. In addition to its ability to deal with noise data [17].
Traditional methods based on previous theories estimate the feature significance based on either individually or dependency discriminative ability. Consequently, there is no general feature selection method, which returns the best feature subset with all datasets [18]. The traditional solution is to understand the data characteristics before the feature selection process. This solution is not efficient because of the high computational cost of expert analysis. To overcome this limitation, a new research direction, called an ensemble feature selection, is introduced, which combines more than one feature selection to cover all situations [2].
In this study, we propose a new ensemble feature selection method (fuzzy feature selection based on relevancy, redundancy, and dependency (FFS-RRD)) to utilize the previous theories. Firstly, we proposed a new method, called fuzzy weighted relevancy-based FS (FWRFS) to estimate the individually discriminative ability. Then, we combined it with fuzzy lower approximation-based FS (L-FRFS) to estimate the dependency discriminative ability [16]. The former method extracts two relations: relevancy and redundancy, while the latter extracts the dependency relation. The aim is to investigate these relations and produce a unique and effective feature selection method to improve classification methods.
The paper is organized as follows: Section 2 presents the main criteria of feature selection: relevancy, redundancy, and dependency. Then, the related work is presented in Section 3. Section 4 introduces the proposed method: fuzzy feature selection based on relevancy, redundancy, and dependency (FFS-RRD). After that, the experiment setup is showed in Section 5. Section 6 analyzes the experimental results. Finally, the conclusion is reported in Section 7.

2. Relevancy, Redundancy, and Dependency Measures

Filter-based FS methods try to find the best feature subset based on data characteristics without depending on classification models [4]. For this reason, they depend on the characteristics of data to find the most significant features. Consequently, filter-based feature selection methods study different data relations such as the relation between features and class, and the relation among features. There are three well-known feature relations: relevancy, redundancy, and dependency.
Firstly, relevancy relation measures the amount of shared information between features and the class [15]. However, some features may have the same relevancy relation and do not add new information to discriminate the classes. These features are considered redundant and no need to be selected. Redundancy relation measures the amount of shared information among features [15]. Another important feature relation is dependency [16]. Dependency relation measures the membership degree of feature subset to class. In the following, we present the definitions of these relations based on the fuzzy set theory [15,16].
Given a dataset D = ( U , F C ) , where U = { u 1 , u 2 , , u m } is a finite set of m instances, F = { f 1 , f 2 , , f n } is a finite set of n features, and C = { c 1 , c 2 , , c l } is a finite set of l classes. Let f : U V f , where V f is the feature value on U. Every feature f F can be represented by fuzzy equivalence relation E f on U and defined by the following fuzzy relation matrix M ( E f ) .
M ( E f ) = e 11 e 12 e 1 m e 21 e 2 m e m 1 e m 2 e m m
where e i j = E ( x i , x j ) is the fuzzy equivalence relation that defines the similarity degree between x i and x j , where x i , x j U .
Fuzzy equivalence class [ x i ] E f of x i is defined by the following fuzzy set on U:
[ x i ] E f = e i 1 x 1 + e i 2 x 2 + + e i m x m
Fuzzy entropy of feature f based on E f is defined as
H ( f ) = 1 m i = 1 m log m | [ x i ] E f |
where, the cardinal value of [ x i ] E f is defined as | [ x i ] E f | = j = 1 m e i j .
Indiscernibility relation I N D ( F ¯ ) defines a set of objects that have the same equivalence class, where F ¯ F . The fuzzy partition of U on I N D ( F ¯ ) is defined by U / I N D ( F ¯ ) = { X 1 , X 2 , , X F ¯ } , where X k = { x j U | [ x i ] E = [ x j ] E } , and x i X x , k = { 1 , 2 , , F ¯ } .
The fuzzy lower approximation of a single fuzzy equivalence class X is defined as
μ E F ¯ X ̲ ( x i ) = inf x j U I ( μ E F ¯ ( x i , x j ) , μ X ( x j ) )
where μ E F ¯ ( x i , x j ) = f F ¯ μ E f ( x i , x j ) , and I = m i n ( 1 , 1 x i + x j ) is the fuzzy Łukasiewicz implicator.
The fuzzy positive region determines all the objects on U that discriminate the classes of U / I N D ( C ) based on a set of features F ¯ . The fuzzy positive region is defined as
μ P O S E F ¯ ( C ) ( x i ) = sup X U / C μ E F ¯ ̲ X ( x i )

2.1. Relevancy

Let E f and E C are two fuzzy relations of feature f and class C on U, respectively. Then, the fuzzy mutual information between f and C is defined as
I ( f ; C ) = 1 m i = 1 m log m | [ x i ] E f [ x i ] E C | | [ x i ] E f | . | [ x i ] E C |

2.2. Redundancy

Let E f 1 and E f 2 be two fuzzy relations of features f 1 and f 2 on U, respectively. Then, the fuzzy mutual information between f 1 and f 2 is defined as
I ( f 1 ; f 2 ) = 1 m i = 1 m log m | [ x i ] E f 1 [ x i ] E f 2 | | [ x i ] E f 1 | . | [ x i ] E f 2 |

2.3. Dependency

Let F ¯ is a set of features, the dependency degree of F ¯ is defined as
F ¯ ( C ) = x i U μ P O S E F ¯ ( C ) ( x i ) | [ x i ] E f |

2.4. Example

To illustrate the computations of previous relations, a small example is presented in Table 1. Firstly, we estimate the relation matrix of each feature based on the following similarity equation [15]:
E f ( x i , x j ) = exp ( x i x j )
As C contains discrete values, we estimate the relation matrix according to the crisp way [19].
The relation matrix of f 1 is:
M ( E f 1 ) = 1.00 0.55 0.82 0.67 1.00 0.55 1.00 0.67 0.82 0.55 0.82 0.67 1.00 0.82 0.82 0.67 0.82 0.82 1.00 0.67 1.00 0.55 0.82 0.67 1.00
The relation matrix of f 2 is:
M ( E f 2 ) = 1.00 0.67 0.82 0.74 1.00 0.67 1.00 0.82 0.90 0.67 0.82 0.82 1.00 0.90 0.82 0.74 0.90 0.90 1.00 0.74 1.00 0.67 0.82 0.74 1.00
The relation matrix of C is:
M ( E C ) = 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0 0.0 1.0
The fuzzy entropy of f 1 is:
H ( f 1 ) = 1 5 ( log 5 4.04 + log 5 3.59 + log 5 4.13 + log 5 3.98 + log 5 4.04 ) = 0.34
The fuzzy entropy of C is:
H ( C ) = 1 5 ( log 5 3 + log 5 2 + log 5 3 + log 5 2 + log 5 3 ) = 0.97
The relevancy between f 1 and C is:
I ( f 1 , C ) = 1 5 ( log 5 2.82 4.04 3 + log 5 1.82 3.59 2 + log 5 2.64 4.13 3 + log 5 1.82 3.98 2 + log 5 2.82 4.04 3 ) = 0.21
The redundancy between f 1 and f 1 is:
I ( f 1 , f 2 ) = 1 5 ( log 5 4.04 4.04 4.23 + log 5 3.59 3.59 4.06 + log 5 4.13 4.13 4.36 + log 5 3.98 3.98 4.29 + log 5 4.04 4.04 4.23 ) = 0.24
For the first object x 1 of f 1 , the fuzzy lower approximation of a single fuzzy equivalence class X = 1 is:
μ E f 1 X = 1 ̲ ( x 1 ) = inf x j U I ( μ E F ¯ ( x 1 , x j ) , μ X = 1 ( x j ) ) = inf { I ( 1.00 , 1.0 ) , I ( 0.55 , 0.0 ) , I ( 0.82 , 1.0 ) , I ( 0.67 , 0.0 ) , I ( 1.00 , 1.0 ) } = 0
For the first object x 1 of f 1 , The fuzzy lower approximation of a single fuzzy equivalence class X = 0 is:
μ E f 1 X = 0 ̲ ( x 1 ) = inf x j U I ( μ E F ¯ ( x 1 , x j ) , μ X = 0 ( x j ) ) = inf { I ( 1.00 , 0.0 ) , I ( 0.55 , 1.0 ) , I ( 0.82 , 0.0 ) , I ( 0.67 , 1.0 ) , I ( 1.00 , 0.0 ) } = 0.33
Similarly, for the remaining objects of f 1 :
μ E f 1 X = 1 ̲ ( x 2 ) = 0.33 , μ E f 1 X = 0 ̲ ( x 2 ) = 0 μ E f 1 X = 1 ̲ ( x 3 ) = 0.0 , μ E f 1 X = 1 ̲ ( x 3 ) = 0.18 μ E f 1 X = 1 ̲ ( x 4 ) = 0.18 , μ E f 1 X = 1 ̲ ( x 4 ) = 0.0 μ E f 1 X = 1 ̲ ( x 5 ) = 0.0 , μ E f 1 X = 1 ̲ ( x 5 ) = 0.33
The fuzzy positive region for the first object x = 1 is:
μ P O S E f 1 ( C ) ( x 1 ) = sup X 1 , 0 μ E f 1 ̲ X ( x 1 ) = sup X 1 , 0 ( 0 , 1 ) = 0.33
For the remaining objects, the fuzzy positive region are:
μ P O S E f 1 ( C ) ( x 2 ) = 0.33 μ P O S E f 1 ( C ) ( x 3 ) = 0.18 μ P O S E f 1 ( C ) ( x 4 ) = 0.18 μ P O S E f 1 ( C ) ( x 5 ) = 0.33
The dependency degree of f 1 is:
f 1 ( C ) = x i U μ P O S E f 1 ( C ) ( x i ) | U | = 0.33 + 0.33 + 0.18 + 0.18 + 0.33 5 = 0.27

3. Related Works

Filter approach evaluates the feature significant based on the characteristics of data only with full independence of classification models [1]. Although the filter approach has many benefits over embedded and wrapper approaches, it may fail to find the best feature subset [20]. For this reason, a great research effort has been introduced to study the feature characteristics with the aim to find the significant features that improve classification models.
Among a variety of evaluation measures, mutual information (MI) has a popularity solution in feature selection based information theory due to its ability to define different relation of features such as relevancy, and redundancy. The main advantages of MI are [3]: (1) ability to deal with deal linear and non-linear relations among features; (2) ability to deal with both categorical and numerical features. In the past decades, MI has been used in many feature selection methods. Mutual information maximization (MIM) [21] defines the significance of features based on the relevancy relation. It suffers from the redundant features. After that, mutual information based feature selection (MIFS) [22] has been introduced and improved in MIFS-U [23] to define the significance of features based on both relevancy and redundancy relation. However, both methods require a predefined parameter to balance between the relevancy and redundancy relations. In [24], minimum redundancy maximum relevance (mRMR) proposes automatic value to estimate the predefined parameter of MIFS, and MIFS-U. In the literature, several feature selection methods have been proposed to find the best estimation of the relevancy and redundancy relations such as joint mutual information (JMI) [25], conditional mutual information maximization (CMIM) [26], joint mutual information maximization (JMIM) [27], and max-relevance and max-independence (MRI) [28]. However, previous studies of feature selection based mutual information do not consider the balance of selected/candidate feature relevancy relation. To avoid this limitation, Zhang et al. [29] has introduced a new method to keep the balance between the feature relevancy relations, called feature selection based on weighted relevancy (WRFS).
Another important solution in the filter approach is rough set which used to measure the dependency relation of features. Feature selection based rough set tries to find the minimal feature subset that maximizes the informative structure of all features (termed a reduct) [30]. The main advantages of the rough set are (1) analyzing only the hidden facts in data, (2) extracting the hidden knowledge of data without additional user-defined information, and (3) returning a minimal knowledge structure of data [19]. Many studies on feature selection based on rough set have been done. Rough Set Attribute Reduction (RSAR) defines the significance of a subset of features based on the dependency relation [31]. However, there is no guarantee to return the minimum feature subset. Han et al. [32] proposes an alternative dependency relation to reduce the computational cost of the feature selection process. Zhong et al. [33] defines the significance of the feature subset based on the discernibility matrix. However, it is impractical for high dimensionality data. In Entropy Based Reduction (EBR) [34], the significance of the feature subset is defined based on entropy which returns the maximum amount of information. In the literature of rough sets, further feature selection methods have been introduced such as Variable precision rough sets (VPRS) [35], and parameterized average support heuristic (PASH) [36].
However, both MI and rough set share common limitations when dealing with features of continuous values [19,37]. There are two traditional solutions have been proposed to overcome this limitation: parzen window [38], and discretization process [39]. The former has some limitations: firstly it requires a predefined parameter to compute the window function [40]. Secondly, it does not work efficiently with high dimensional data of spare samples [15]. The latter may lead to loss of feature information [41]. To overcome these limitations, FS based information theory and FS based rough set have been extended by fuzzy set theory to deal with continuous features directly [14,19]. However, most of FS methods based information theory focus on relevancy and redundancy relation, while FS methods based on rough set focus on dependency relation. The former depends on individually discriminative ability, while the latter depends on dependency discriminative ability. As a result, the traditional methods do not take the benefits of all types of discriminative ability.

4. Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency (FFS-RRD)

In this section, we present our proposed method, called FFS-RRD, as a filter feature selection method. The effectiveness of filter methods depends on the amount of extracted information from the data characteristics. To promote our proposed method, we used both individually and dependency discriminative ability based on three criteria: relevancy, redundancy, and dependency. FFS-RRD aims to maximize both relevancy and dependency relations and minimize the redundancy ones. To design our proposed method, firstly, we modified WRFS to overcome their limitations: (1) it can not deal with continuous features without the discretization process which may lead to loss of feature information. (2) WRFS does not consider dependency relation in the feature selection process. To overcome these limitations, we estimated WRFS based on the fuzzy concept instead of the probability concept. The extended method, called FWRFS, can deal with any numerical data without the discretization process. Then, we combined FWRFS with fuzzy-rough lower approximations (L-FRFS) [16] to extract the dependency relation. Consequently, we proposed a unique FS method, called FFSRRD, which maximizes both relevancy and dependency, and minimizes the redundancy relation. The three relations can extract more information from the dataset to promote the discriminate ability of feature selection. Figure 1 shows the process of the proposed method FFS-RRD. Both FWRFS and L-FRFS are applied on the same dataset. FWRFS selects the most relevant features and removes the redundancy ones, while L-FRFS selects the most dependency feature subset. The results of each method are combined to return the final feature selection subset. In our study, we used one of the popular combination methods called MIN [2]. MIN method assigns the minimum position of each feature among different results of feature selection methods to be ranked position in the final result.
The algorithm of the proposed method is presented in Algorithm 1. FFS-RRD depends on a combination of two methods: For the first method, FWRFS is used to return the ranked feature set that maximizes the relevancy and minimizes the redundancy. In the first step (Lines 1–3), The main parameters are initialized: ranked feature set ( R 1 ), candidate feature ( c a n d i d a t e ), and the current selected feature set ( s e l e c t e d ). Then, the feature of maximum relevancy with class is selected to be the first ranked feature in R 1 , and removed from the feature set F (Lines 4–8). After that, the feature of maximum relevancy with class and minimum redundancy with selected features is added to R 1 , and removed from F. This process is repeated until all features of F are ranked in R 1 (Lines 9–14). For the second method, L-FRFS is used to return the subset of features that maximizes the dependency relation. In the first step (Lines 15–17), the main parameters are initialized: selected feature subset ( R 2 ), temporary feature (T), maximum dependency degree ( s e l e c t ), and the last maximum dependency degree ( l a s t ). Then, the feature of maximum dependency is added to R 2 . This process is repeated until the maximum possible dependency degree of features be produced (Lines 18–25). Finally, the result of both methods is combined by M I N ( R 1 , R 2 ) to select the final feature subset (Line 26–27).
Algorithm 1: FFS-RRD: fuzzy feature selection based relevancy, redundancy, and dependency.
Entropy 22 00757 i001

5. Experiment Setup

The main goal of the feature selection process is to improve the classification performance with the minimum feature selection subset. To validate our proposed method, we used four classifiers to compare the proposed method with eight feature selection methods based on benchmark datasets. Figure 2 shows the framework of our experiment. In the following, we present more details about the experiment setup.

5.1. Dataset

Our experiment was conducted based on 13 benchmark datasets from machine learning repository (UCI) [42]. The datasets support different classification problems of binary and multi-class data. Table 2 presents a brief description of the experimental datasets.

5.2. Compared Feature Selection Methods

Table 3 shows the compared FS methods and their discriminative ability. The compared methods can be divided into two groups: probability-based, and fuzzy-based. Firstly, probability-based group uses the probability concept to estimate information measures. The probability-based group consists of CIFE [43], JMI [25], JMIM [27], WRFS [29], CMIM3 [44], JMI3 [44], and MIGM [45]. This group depends on the discretization process before implementation of feature selection methods. In our experiment, the discretization process transforms the continuous features into discrete features with ten equal intervals [46].
Unlike a probability-based group which requires discretization preprocess, fuzzy-based group uses the fuzzy concept to estimate information measures. The fuzzy-based group includes L-FRFS [16], and the proposed method FFS-RRD. This group depends on similarity relation which transforms each feature into a fuzzy equivalence relation. In our experiment, we used the following similarity relation [15].
E f ( x i , x j ) = exp ( x i x j )

5.3. Evaluation Metrics

The main factors characterize the quality of feature selection methods are its classification performance and stability [47]. The evaluation of our experiment is divided into two parts: classification performance and stability evaluation. Classification performance requires classification models to evaluate the effect of feature selection methods on improving the classification performance, while stability measures the robustness of feature selection methods.

5.3.1. Classification Performance

To evaluate the classification performance, we used three metrics: classification accuracy, F-measure ( β = 1 ), AUC. The experiment depends on four classifiers: Naive bayes (NB), support vector machine (SVM), K-nearest neighbors (KNN, K = 3), and decision tree (DT). To find reliable results, we used 10-fold cross-validation where the dataset is divided into ten equal parts, nine for the training phase and one for the test phase [48]. This process is repeated ten times. Then, we calculate the average results to compute the score of accuracy, F-measure, and AUC.
In this experiment, we used a threshold to cut the ranked features and return a subset of selected features. The threshold is the median position of the ranked features (or the nearest integer position if the number of ranked features is even). For L-FRFS, we used the same threshold if the size of the returned subset is more than the median of all features.

5.3.2. Stability Evaluation

The confidence of feature selection method is not only about the improvement of classification performance but also related to the robustness of the method [49]. The robustness of feature selection method against any small change of data, as a noise, is called feature selection stability [50]. In the stability experiment, we injected the data by 10% of noise which is generated based on standard deviation and the gaussian distribution of each feature [51]. Then, we run the feature selection method to return the sequence of features. This process is repeated for ten times with a new returned sequence each time. After that, we measure the stability for each feature selection method based on Kuncheva stability measure which is defined as [52]:
K u n s t a b = 2 p ( p 1 ) i = 1 p 1 j = i + 1 p K u n i n d e x ( R i , R j )
where p is the number of feature selection sequences, and K u n i n d e x ( R i , R j ) is the Kuncheva stability index between two feature selection sequences R i , and R j which is defined as:
K u n i n d e x ( R i , R j ) = w n r 2 r ( n r ) ,
where w = | R i R j | , r = | R i | = | R j | , and n is the total number of features.

6. Results Analysis

6.1. Classification Performance

6.1.1. Accuracy

Based on NB classifier, it is obvious that FFS-RRD achieved the maximum average accuracy with score 83.4%, as shown in Table 4. The proposed method was more accurate than compared methods by the range from 0.4% to 1.8%. The order of methods ranked after FFS-RRD was JMIM, followed by JMI, both CMIM3 and JMI3, MIGM, WRFS, L-FRFS, and CIFE.
According to SVM classifier, FFS-RRD achieved the maximum average accuracy of all datasets by 86.4%, while L-FRFS achieved the minimum average accuracy by 84.1%, as shown in Table 5. The proposed method outperformed other methods in the range from 0.5% to 2.3%. The second-best feature selection method was JMI, followed by CMIM3, both JMIM and JMI3, WRFS, MIGM, and CIFE.
In the case of KNN classifier, FFS-RRD also was the best feature selection method in the term of average accuracy by 85.4%, while L-FRFS was the worst method by 82.5%, as shown in Table 6. After that, MIGM achieved the second-best method, followed by both JMI and JMIM, JMI3, WRFS, CMIM3, CIFE. The proposed method achieved better accuracy in the range from 0.5% to 2.9%.
Similarly, FFS-RRD kept the best average accuracy of DT classifier by 84.5%, as shown in Table 7. The proposed method outperformed other methods in the range from 0.4% to 1.4%. In contrast, both CIFE and L-FRFS achieved the worst results by 83.1%. The second-best feature selection method was JMI, followed by JMIM, both WRFS and JMI3, MIGM, and CMIM3.

6.1.2. F-Measure

Figure 3 shows the F-measure of the compared methods based on the four used classifiers. In NB classifier, FSS-RRD achieved the maximum average F-measure by 88.5%, while MIGM achieved the minimum score by 82.4%. The proposed method outperformed other methods in the range from 1.5% to 6.1%. Similarly, FSS-RRD achieved the maximum average F-measure using SVM by 88.9%, while MIGM achieved the minimum score by 83.5%. The proposed method outperformed other methods in the range from 0.5% to 5.4%. According to KNN classifier, WRFS achieved the maximum average F-measure by 87.7%, while CMIM3 achieved the minimum score by 81.6%. The proposed method achieved the fourth-best position in this case. In DT classifier, FSS-RRD achieved the maximum average F-measure by 87.5%, while MIGM achieved the minimum score by 82.3%. The proposed method outperformed other methods in the range from 0.7% to 5.2%.

6.1.3. AUC

It is obvious that the proposed method achieved the highest AUC compared with other methods using all classifiers (Figure 4). According to NB, FSS-RDD achieved the maximum AUC by 87.8%, while CIFE achieved the minimum AUC by 86.2%. The proposed method outperformed other methods in the range from 0.2% to 1.6%. In SVM classifier, FSS-RDD also achieved the maximum AUC by 81.1%, while L-FRFS achieved the minimum score by 78.2%. The proposed method outperformed other methods in the range from 0.9% to 2.9%. Similarly, FSS-RDD also achieved the maximum AUC using KNN by 85.1%, while L-FRFS achieved the minimum score by 83.3%. The proposed method outperformed other methods in the range from 0.4% to 1.8%. Using DT classifier, FSS-RDD kept the best method by 82.2%, L-FRFS kept the worst method by 79.9%. The proposed method outperformed other methods in the range from 0.4% to 2.3%.
Figure 5 shows the average score of the four classifiers in terms of accuracy, F-measure, and AUC. For accuracy term, FFS-RRD achieved the highest accuracy for all classifiers by 84.9%, followed by JMI, JMIM, JMI3, CMIM3, MIGM, WRFS, CIFE, and L-FRFS by 84.4%, 84.3%, 84.2%, 84.1%, 83.9%, 83.8%, 83.3%, and 82.8%, respectively. The proposed method outperformed other methods in the range from 0.5% to 2.1%. Similarly, FFS-RRD achieved the highest average of F-measure by 87.6%. Then, JMIM achieved the second-best method, followed by JMI, WRFS, L-FRFS, CIFE, JMI3, MIGM, CMIM3 with score 87.4%, 87.3%, 86.8%, 85.7%, 85.7%, 85.3%, 83.4%, 83.9%, 82.8%, respectively. The proposed method outperformed other methods in the range from 0.2% to 4.7%. According to AUC, FFS-RRD achieved the highest AUC with a score 84.0%. JMI3 achieved the second-best method, followed by JMI, CMIM3, WRFS, MIGM, JMIM, CIFE, and L-FRFS by 83.6%, 83.5%, 83.4%, 83.3%, 83.2%, 83.0%, 82.4%, and 81.9%, respectively. The proposed method outperformed other methods in the range from 0.4% to 2.1%.

6.2. Stability

Figure 6 shows the average stability across the first half thresholds on all datasets. FFS-RRS achieved the maximum average of stability by 84.3%, while MIGM achieved the minimum score by 67.9%. After that, L-FRFS achieved the second-best method by 78.6%, followed by JMI3, CIFE, JMI, WRFS, CMIM3, and JMIM with an average score 76.0%. 75.8%, 73.5%, 73.0%, 72.3%, and 71.9%, respectively. The proposed method outperformed other methods in the range from 5.7% to 16.4%. Figure 7 shows a box-plot of average stability for all compared methods on the median threshold. In box-plot, the black circle represents the stability median, while the box represents both lower and upper quartiles. As shown in the box-plot, the stability result of the proposed method is better and more consistent than compared methods.
By considering the previous results, it is obvious that FFS-RRD achieved the best experimental results in the term of classification performance and feature stability. This is expected where the proposed method considers the individually and dependency discriminative ability of features. On the other hand, it is obvious that fuzzy-based methods are more stable than probability-based methods. The reason returns to using fuzzy sets to estimate the feature significance without information loss. Consequently, it helps fuzzy-based methods to be more stable against the noise.

7. Conclusions

In this paper, we have proposed an ensemble feature selection method, fuzzy feature selection based on relevancy, redundancy, and dependency criteria (FFS-RRD). Unlike the traditional methods, FFS-RRD depends on both individually and dependency discriminative ability. FFS-RRD aims to extract the significant relations from data characteristics to find the best feature subset that improves the performance of classification models. The proposed method consist of combination of two methods: FWRFS, and L-FRFS. FWRFS maximizes the relevancy and minimizes the redundancy relation, while L-FRFS maximizes the dependency relation.
Compared with eight state-of-the-art and conventional FS methods, experiments on 13 benchmark datasets indicate the outperformance of the proposed method in classification performance and stability. Classification performance includes three measures accuracy, F-measure, and AUC. The proposed method FFS-RRD achieved the highest average score of accuracy, and AUC on all datasets, while it achieved the highest average of F-measure on most of the classifiers except KNN classifier. On the other hand, the proposed method achieved the highest average of stability compared with other feature selection methods. In future work, we will extend the proposed method to explore their effect on multi-label classification models.

Author Contributions

Methodology, O.A.M.S., F.L., and Y.-P.P.C.; software, O.A.M.S.; investigation, F.L., X.C., and Y.-P.P.C.; writing—original draft preparation, O.A.M.S., and F.L.; writing—review and editing, F.L., O.A.M.S., X.C., and Y.-P.P.C.; supervision, F.L., and X.C.; funding acquisition, F.L. All authors read and approved the final version of the manuscript.

Acknowledgments

This research has been supported by the National Natural Science Foundation (61572368).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Macedo, F.; Oliveira, M.R.; Pacheco, A.; Valadas, R. Theoretical foundations of forward feature selection methods based on mutual information. Neurocomputing 2019, 325, 67–89. [Google Scholar] [CrossRef] [Green Version]
  2. Bolón-Canedo, V.; Alonso-Betanzos, A. Ensembles for feature selection: A review and future trends. Inf. Fusion 2019, 52, 1–12. [Google Scholar] [CrossRef]
  3. Lee, S.; Park, Y.T.; d’Auriol, B.J. A novel feature selection method based on normalized mutual information. Appl. Intell. 2012, 37, 100–120. [Google Scholar]
  4. Lazar, C.; Taminau, J.; Meganck, S.; Steenhoff, D.; Coletta, A.; Molter, C.; de Schaetzen, V.; Duque, R.; Bersini, H.; Nowe, A. A survey on filter techniques for feature selection in gene expression microarray analysis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2012, 9, 1106–1119. [Google Scholar] [CrossRef]
  5. Imani, M.B.; Keyvanpour, M.R.; Azmi, R. A novel embedded feature selection method: A comparative study in the application of text categorization. Appl. Artif. Intell. 2013, 27, 408–427. [Google Scholar] [CrossRef]
  6. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  7. Hu, L.; Gao, W.; Zhao, K.; Zhang, P.; Wang, F. Feature selection considering two types of feature relevancy and feature interdependency. Expert Syst. Appl. 2018, 93, 423–434. [Google Scholar] [CrossRef]
  8. Saeys, Y.; Inza, I.; Larrañaga, P. A review of feature selection techniques in bioinformatics. Bioinformatics 2007, 23, 2507–2517. [Google Scholar] [CrossRef] [Green Version]
  9. Oreski, D.; Oreski, S.; Klicek, B. Effects of dataset characteristics on the performance of feature selection techniques. Appl. Soft Comput. 2017, 52, 109–119. [Google Scholar] [CrossRef]
  10. Bonev, B. Feature Selection Based on Information Theory; Universidad de Alicante: Alicante, Spain, 2010. [Google Scholar]
  11. Caballero, Y.; Alvarez, D.; Bello, R.; Garcia, M.M. Feature selection algorithms using rough set theory. In Proceedings of the IEEE Seventh International Conference on Intelligent Systems Design and Applications (ISDA 2007), Rio de Janeiro, Brazilm, 20–24 October 2007; pp. 407–411. [Google Scholar]
  12. Che, J.; Yang, Y.; Li, L.; Bai, X.; Zhang, S.; Deng, C. Maximum relevance minimum common redundancy feature selection for nonlinear data. Inf. Sci. 2017, 409, 68–86. [Google Scholar] [CrossRef]
  13. Pawlak, Z. Rough Sets: Theoretical Aspects of Reasoning about Data; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1991. [Google Scholar]
  14. Hu, Q.; Yu, D.; Xie, Z. Information-preserving hybrid data reduction based on fuzzy-rough techniques. Pattern Recognit. Lett. 2006, 27, 414–423. [Google Scholar] [CrossRef]
  15. Yu, D.; An, S.; Hu, Q. Fuzzy mutual information based min-redundancy and max-relevance heterogeneous feature selection. Int. J. Comput. Intell. Syst. 2011, 4, 619–633. [Google Scholar] [CrossRef]
  16. Jensen, R.; Shen, Q. New approaches to fuzzy-rough feature selection. IEEE Trans. Fuzzy Syst. 2008, 17, 824–838. [Google Scholar] [CrossRef] [Green Version]
  17. Hüllermeier, E. Fuzzy sets in machine learning and data mining. Appl. Soft Comput. 2011, 11, 1493–1505. [Google Scholar] [CrossRef] [Green Version]
  18. Freeman, C.; Kulić, D.; Basir, O. An evaluation of classifier-specific filter measure performance for feature selection. Pattern Recognit. 2015, 48, 1812–1826. [Google Scholar] [CrossRef]
  19. Jensen, R.; Shen, Q. Fuzzy-rough sets for descriptive dimensionality reduction. In Proceedings of the 2002 IEEE World Congress on Computational Intelligence. 2002 IEEE International Conference on Fuzzy Systems, FUZZ-IEEE’02, Honolulu, HI, USA, 12–17 May 2002; Proceedings (Cat. No. 02CH37291). Volume 1, pp. 29–34. [Google Scholar]
  20. Vergara, J.R.; Estévez, P.A. A review of feature selection methods based on mutual information. Neural Comput. Appl. 2014, 24, 175–186. [Google Scholar] [CrossRef]
  21. Lewis, D.D. Feature selection and feature extraction for text categorization. In Proceedings of the Workshop on Speech and Natural Language; Association for Computational Linguistics: Stroudsburg, PA, USA, 1992; pp. 212–217. [Google Scholar]
  22. Battiti, R. Using mutual information for selecting features in supervised neural net learning. IEEE Trans. Neural Netw. 1994, 5, 537–550. [Google Scholar] [CrossRef] [Green Version]
  23. Kwak, N.; Choi, C.H. Input feature selection for classification problems. IEEE Trans. Neural Netw. 2002, 13, 143–159. [Google Scholar] [CrossRef]
  24. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Onpattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  25. Yang, H.; Moody, J. Feature selection based on joint mutual information. In Proceedings of the International ICSC Symposium on Advances in Intelligent Data Analysis, Genova, Italy, 1–4 June 1999; pp. 22–25. [Google Scholar]
  26. Fleuret, F. Fast binary feature selection with conditional mutual information. J. Mach. Learn. Res. 2004, 5, 1531–1555. [Google Scholar]
  27. Bennasar, M.; Hicks, Y.; Setchi, R. Feature selection using joint mutual information maximisation. Expert Syst. Appl. 2015, 42, 8520–8532. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, J.; Wei, J.M.; Yang, Z.; Wang, S.Q. Feature selection by maximizing independent classification information. IEEE Trans. Knowl. Data Eng. 2017, 29, 828–841. [Google Scholar] [CrossRef]
  29. Zhang, P.; Gao, W.; Liu, G. Feature selection considering weighted relevancy. Appl. Intell. 2018, 48, 4615–4625. [Google Scholar] [CrossRef]
  30. Hassanien, A.E.; Suraj, Z.; Slezak, D.; Lingras, P. Rough Computing: Theories, Technologies and Applications; IGI Global Hershey: Hershey, PA, USA, 2008. [Google Scholar]
  31. Chouchoulas, A.; Shen, Q. Rough set-aided keyword reduction for text categorization. Appl. Artif. Intell. 2001, 15, 843–873. [Google Scholar] [CrossRef]
  32. Han, J.; Hu, X.; Lin, T.Y. Feature subset selection based on relative dependency between attributes. In International Conference on Rough Sets and Current Trends in Computing; Springer: Berlin, Germany, 2004; pp. 176–185. [Google Scholar]
  33. Zhong, N.; Dong, J.; Ohsuga, S. Using rough sets with heuristics for feature selection. J. Intell. Inf. Syst. 2001, 16, 199–214. [Google Scholar] [CrossRef]
  34. Jensen, R.; Shen, Q. Fuzzy–rough attribute reduction with application to web categorization. Fuzzy Sets Syst. 2004, 141, 469–485. [Google Scholar] [CrossRef] [Green Version]
  35. Ziarko, W. Variable precision rough set model. J. Comput. Syst. Sci. 1993, 46, 39–59. [Google Scholar] [CrossRef] [Green Version]
  36. Zhang, M.; Yao, J. A rough sets based approach to feature selection. In Proceedings of the IEEE Annual Meeting of the Fuzzy Information, Banff, AB, Canada, 27–30 June 2004; Processing NAFIPS’04. Volume 1, pp. 434–439. [Google Scholar]
  37. Ching, J.Y.; Wong, A.K.; Chan, K.C.C. Class-dependent discretization for inductive learning from continuous and mixed-mode data. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 17, 641–651. [Google Scholar] [CrossRef]
  38. Kwak, N.; Choi, C.H. Input feature selection by mutual information based on Parzen window. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 1667–1671. [Google Scholar] [CrossRef] [Green Version]
  39. Garcia, S.; Luengo, J.; Sáez, J.A.; Lopez, V.; Herrera, F. A survey of discretization techniques: Taxonomy and empirical analysis in supervised learning. IEEE Trans. Knowl. Data Eng. 2012, 25, 734–750. [Google Scholar] [CrossRef]
  40. Herman, G.; Zhang, B.; Wang, Y.; Ye, G.; Chen, F. Mutual information-based method for selecting informative feature sets. Pattern Recognit. 2013, 46, 3315–3327. [Google Scholar] [CrossRef]
  41. Shen, Q.; Jensen, R. Selecting informative features with fuzzy-rough sets and its application for complex systems monitoring. Pattern Recognit. 2004, 37, 1351–1363. [Google Scholar] [CrossRef]
  42. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2017. [Google Scholar]
  43. Lin, D.; Tang, X. Conditional infomax learning: An integrated framework for feature extraction and fusion. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2006; pp. 68–82. [Google Scholar]
  44. Sechidis, K.; Azzimonti, L.; Pocock, A.; Corani, G.; Weatherall, J.; Brown, G. Efficient feature selection using shrinkage estimators. Mach. Learn. 2019, 108, 1261–1286. [Google Scholar] [CrossRef] [Green Version]
  45. Wang, X.; Guo, B.; Shen, Y.; Zhou, C.; Duan, X. Input Feature Selection Method Based on Feature Set Equivalence and Mutual Information Gain Maximization. IEEE Access 2019, 7, 151525–151538. [Google Scholar] [CrossRef]
  46. Dougherty, J.; Kohavi, R.; Sahami, M. Supervised and unsupervised discretization of continuous features. In Machine Learning Proceedings 1995; Elsevier: Amsterdam, The Netherlands, 1995; pp. 194–202. [Google Scholar]
  47. Li, Y.; Si, J.; Zhou, G.; Huang, S.; Chen, S. FREL: A stable feature selection algorithm. IEEE Trans. Neural Networks Learn. Syst. 2014, 26, 1388–1402. [Google Scholar] [CrossRef]
  48. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. Ijcai 1995, 14, 1137–1145. [Google Scholar]
  49. Pes, B.; Dessì, N.; Angioni, M. Exploiting the ensemble paradigm for stable feature selection: A case study on high-dimensional genomic data. Inf. Fusion 2017, 35, 132–147. [Google Scholar] [CrossRef]
  50. Nogueira, S.; Brown, G. Measuring the stability of feature selection. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2016; pp. 442–457. [Google Scholar]
  51. Tsai, Y.S.; Yang, U.C.; Chung, I.F.; Huang, C.D. A comparison of mutual and fuzzy-mutual information-based feature selection strategies. In Proceedings of the 2013 IEEE International Conference on Fuzzy Systems (FUZZ), Hyderabad, India, 7–10 July 2013; pp. 1–6. [Google Scholar]
  52. Kuncheva, L.I. A stability index for feature selection. In Proceedings of the 25th IASTED International Multi-Conference Artificial Intelligence and Applications, Innsbruck, Austria, 12–14 February 2007; pp. 421–427. [Google Scholar]
Figure 1. The process of our proposed method fuzzy feature selection based relevancy, redundancy, and dependency (FFS-RRD): firstly, the fuzzy relation matrix is generated for each feature in the dataset. Then, fuzzy mutual information maximizes the relevancy and minimizes the redundancy, while fuzzy rough set maximizes the dependency. Finally, the results are combined to find the selected features.
Figure 1. The process of our proposed method fuzzy feature selection based relevancy, redundancy, and dependency (FFS-RRD): firstly, the fuzzy relation matrix is generated for each feature in the dataset. Then, fuzzy mutual information maximizes the relevancy and minimizes the redundancy, while fuzzy rough set maximizes the dependency. Finally, the results are combined to find the selected features.
Entropy 22 00757 g001
Figure 2. An experimental framework of the proposed method fuzzy feature selection based relevancy, redundancy, and dependency (FFS-RRD): firstly, a discretization process is applied before probability-based methods. Then, the compared methods are evaluated in terms of classification performance, and stability.
Figure 2. An experimental framework of the proposed method fuzzy feature selection based relevancy, redundancy, and dependency (FFS-RRD): firstly, a discretization process is applied before probability-based methods. Then, the compared methods are evaluated in terms of classification performance, and stability.
Entropy 22 00757 g002
Figure 3. Average F-measure on the four used classifiers. Our proposed method (FFS-RRD) achieved the best result in all cases except KNN.
Figure 3. Average F-measure on the four used classifiers. Our proposed method (FFS-RRD) achieved the best result in all cases except KNN.
Entropy 22 00757 g003
Figure 4. Average AUC on the four used classifiers. Our proposed method (FFS-RRD) achieved the best result in all cases.
Figure 4. Average AUC on the four used classifiers. Our proposed method (FFS-RRD) achieved the best result in all cases.
Entropy 22 00757 g004
Figure 5. Average score of all classifiers in terms of accuracy, F-measure, and AUC. Our proposed method (FFS-RRD) achieved the best result.
Figure 5. Average score of all classifiers in terms of accuracy, F-measure, and AUC. Our proposed method (FFS-RRD) achieved the best result.
Entropy 22 00757 g005
Figure 6. Average stability across the first half thresholds on all datasets. Our proposed method (FFS-RRD) achieved the best result.
Figure 6. Average stability across the first half thresholds on all datasets. Our proposed method (FFS-RRD) achieved the best result.
Entropy 22 00757 g006
Figure 7. Average stability for all compared methods on the median threshold. Our proposed method (FFS-RRD) achieved the best result.
Figure 7. Average stability for all compared methods on the median threshold. Our proposed method (FFS-RRD) achieved the best result.
Entropy 22 00757 g007
Table 1. An example of a small dataset, contains two features ( f 1 , f 2 ), and class C.
Table 1. An example of a small dataset, contains two features ( f 1 , f 2 ), and class C.
f 1 f 2 C
0.2
0.8
0.4
0.6
0.2
0.1
0.5
0.3
0.4
0.1
1
0
1
0
1
Table 2. Description of datasets used in the experiments.
Table 2. Description of datasets used in the experiments.
DatasetBrief# Instances# Features# Classes
Breast Cancer Wisconsin (Prognostic)BCW Prognostic198332
Breast Cancer Wisconsin (Diagnostic)BCW Diagnostic569312
Climate Model Simulation CrashesCMSC540182
Credit ApprovalCredit Approval690152
DermatologyDermatology336346
Diabetic Retinopathy DebrecenDRD1151192
FertilityFertility10092
Statlog (Heart)Heart270132
IonosphereIonosphere351342
IrisIris15043
Libras MovementLibras Movement3609015
QSAR biodegradationQSAR1055412
ZooZoo101167
Table 3. The extracted feature relations of compared feature selection methods.
Table 3. The extracted feature relations of compared feature selection methods.
Ref.FS GroupFS MethodDiscriminative Ability
IndividuallyDependency
[43]Probability-basedCIFE
[25]JMI
[27]JMIM
[29]WRFS
[44]CMIM3
[44]JMI3
[45]MIGM
[16]Fuzzy-basedL-FRFS
ProposedFFS-RRD
Table 4. Average classification accuracy on Naive bayes (NB) classifier: our proposed method achieved the best result.
Table 4. Average classification accuracy on Naive bayes (NB) classifier: our proposed method achieved the best result.
DatasetCIFEJMIJMIMWRFSCMIM3JMI3MIGML-FRFSFFS-RRD
BCW Prognostic73.969.573.865.468.668.867.771.369.8
BCW Diagnostic92.093.493.493.693.292.492.985.993.6
CMSC91.993.891.994.193.692.893.893.793.8
Credit Approval85.683.583.786.982.483.784.876.885.4
Dermatology96.193.995.393.198.094.695.396.296.1
DRD57.660.357.657.557.760.660.357.657.5
Fertility87.987.988.088.088.088.087.988.088.0
Heart80.183.081.084.183.083.980.175.181.5
Ionosphere77.082.686.280.884.884.877.489.889.0
Iris94.694.694.694.693.593.593.595.095.0
Libras Movement51.461.760.850.559.059.059.461.060.0
QSAR78.278.777.178.578.778.177.677.780.0
Zoo94.295.196.096.096.196.096.993.594.9
Average81.682.983.081.882.882.882.181.783.4
Table 5. Average classification accuracy on support vector machine (SVM) classifier: our proposed method achieved the best result.
Table 5. Average classification accuracy on support vector machine (SVM) classifier: our proposed method achieved the best result.
DatasetCIFEJMIJMIMWRFSCMIM3JMI3MIGML-FRFSFFS-RRD
BCW Prognostic76.376.877.976.376.477.776.676.377.5
BCW Diagnostic96.397.397.397.596.995.095.688.496.3
CMSC91.592.091.593.291.991.691.992.191.9
Credit Approval85.585.585.585.585.585.585.573.785.5
Dermatology95.895.595.894.298.296.895.796.896.5
DRD68.067.268.067.767.567.067.268.067.7
Fertility88.088.088.088.088.088.088.088.088.0
Heart79.784.384.384.281.383.779.776.382.5
Ionosphere76.679.078.581.281.782.978.086.787.6
Iris95.795.795.795.993.993.993.994.595.0
Libras Movement72.276.174.767.675.574.974.576.575.3
QSAR84.284.384.582.384.283.784.483.884.1
Zoo92.895.388.793.393.989.594.192.095.3
Average84.885.985.485.185.885.485.084.186.4
Table 6. Average classification accuracy on K-nearest neighbors (KNN) classifier: our proposed method achieved the best result.
Table 6. Average classification accuracy on K-nearest neighbors (KNN) classifier: our proposed method achieved the best result.
DatasetCIFEJMIJMIMWRFSCMIM3JMI3MIGML-FRFSFFS-RRD
BCW Prognostic75.275.077.176.673.574.176.974.077.0
BCW Diagnostic93.393.597.195.495.194.695.287.594.8
CMSC90.192.390.192.090.992.492.693.493.0
Credit Approval84.783.985.185.584.485.385.176.484.6
Dermatology95.095.995.693.797.796.996.196.196.3
DRD64.363.764.364.162.363.763.664.764.1
Fertility88.788.787.286.784.090.588.785.089.4
Heart77.779.179.181.779.377.777.671.778.2
Ionosphere81.880.980.383.082.383.383.582.784.2
Iris93.993.993.993.991.891.891.892.792.7
Libras Movement70.776.877.970.477.174.778.375.878.3
QSAR84.084.783.381.583.884.483.682.583.0
Zoo90.194.191.693.192.192.191.289.794.3
Average83.884.884.884.484.284.784.982.585.4
Table 7. Average classification accuracy on decision tree (DT) classifier: our proposed method achieved the best result.
Table 7. Average classification accuracy on decision tree (DT) classifier: our proposed method achieved the best result.
DatasetCIFEJMIJMIMWRFSCMIM3JMI3MIGML-FRFSFFS-RRD
BCW Prognostic73.873.773.172.572.873.073.076.673.2
BCW Diagnostic93.794.594.194.294.093.593.588.694.6
CMSC89.691.289.690.391.491.291.391.391.6
Credit Approval85.585.585.786.384.885.785.375.386.3
Dermatology93.692.392.291.393.594.192.594.394.9
DRD68.065.868.067.765.066.865.967.767.6
Fertility87.187.187.187.686.686.687.187.587.5
Heart73.980.381.380.676.377.674.074.375.5
Ionosphere88.488.488.489.788.888.488.088.689.0
Iris90.890.890.890.891.991.991.991.791.7
Libras Movement60.466.065.661.364.764.367.566.066.4
QSAR83.283.782.583.582.983.582.382.382.9
Zoo91.994.092.392.993.492.194.196.097.1
Average83.184.183.983.783.583.783.683.184.5

Share and Cite

MDPI and ACS Style

Salem, O.A.M.; Liu, F.; Chen, Y.-P.P.; Chen, X. Ensemble Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency Criteria. Entropy 2020, 22, 757. https://doi.org/10.3390/e22070757

AMA Style

Salem OAM, Liu F, Chen Y-PP, Chen X. Ensemble Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency Criteria. Entropy. 2020; 22(7):757. https://doi.org/10.3390/e22070757

Chicago/Turabian Style

Salem, Omar A. M., Feng Liu, Yi-Ping Phoebe Chen, and Xi Chen. 2020. "Ensemble Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency Criteria" Entropy 22, no. 7: 757. https://doi.org/10.3390/e22070757

APA Style

Salem, O. A. M., Liu, F., Chen, Y. -P. P., & Chen, X. (2020). Ensemble Fuzzy Feature Selection Based on Relevancy, Redundancy, and Dependency Criteria. Entropy, 22(7), 757. https://doi.org/10.3390/e22070757

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop