Next Article in Journal
A Novel Mine Cage Safety Monitoring Algorithm Utilizing Visible Light
Previous Article in Journal
SlimDeblurGAN-Based Motion Deblurring and Marker Detection for Autonomous Drone Landing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Method for Objective Selection of Information Sources Using Multi-Kernel SVM and Local Scaling

by
Henry Jhoán Areiza-Laverde
1,
Andrés Eduardo Castro-Ospina
1,
María Liliana Hernández
2 and
Gloria M. Díaz
1,*
1
MIRP Lab–Parque i, Instituto Tecnológico Metropolitano (ITM), Medellín 050013, Colombia
2
Grupo de Investigación del Instituto de Alta Tecnología Médica (IATM), Ayudas Diagnósticas Sura, Medellín 050026, Colombia
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(14), 3919; https://doi.org/10.3390/s20143919
Submission received: 17 May 2020 / Revised: 30 May 2020 / Accepted: 3 June 2020 / Published: 14 July 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Advancement on computer and sensing technologies has generated exponential growth in the data available for the development of systems that support decision-making in fields such as health, entertainment, manufacturing, among others. This fact has made that the fusion of data from multiple and heterogeneous sources became one of the most promising research fields in machine learning. However, in real-world applications, to reduce the number of sources while maintaining optimal system performance is an important task due to the availability of data and implementation costs related to processing, implementation, and development times. In this work, a novel method for the objective selection of relevant information sources in a multimodality system is proposed. This approach takes advantage of the ability of multiple kernel learning (MKL) and the support vector machines (SVM) classifier to perform an optimal fusion of data by assigning weights according to their discriminative value in the classification task; when a kernel is designed for representing each data source, these weights can be used as a measure of their relevance. Moreover, three algorithms for tuning the Gaussian kernel bandwidth in the classifier prediction stage are introduced to reduce the computational cost of searching for an optimal solution; these algorithms are an adaptation of a common technique in unsupervised learning named local scaling. Two real application tasks were used to evaluate the proposed method: the selection of electrodes for a classification task in Brain–Computer Interface (BCI) systems and the selection of relevant Magnetic Resonance Imaging (MRI) sequences for detection of breast cancer. The obtained results show that the proposed method allows the selection of a small number of information sources.

1. Introduction

In machine learning, multimodality refers to the simultaneous use of different information sources to solve a specific problem [1]. It is applied to improve some aspects of algorithms, such as the feature generation process or separation between classes, referring specifically to the machine learning area. The use of multimodal sources offers some advantages because it provides additional information about the problem being solved [2]. However, having multiple information sources can also become a problem as the implementation cost can be substantially increased due to the procedural and financial parts of the solution. Currently, different research fields conduct studies that involve the use of multimodal information sources to improve the performance of their works [3,4,5]; however, at the same time, this trend creates the need for studies into the control and optimization of the use of these sources so that the tasks to be carried out are much more efficient in terms of cost, development, and processing times [6,7,8].
Taking into account that in recent times machine learning algorithms have proven to be extremely useful for processing large amounts of data, different studies have been conducted to automatically reduce the amount of data. The most similar approach to the automatic selection of relevant information sources is feature selection through machine learning methods [9]. Such an approach is implemented mainly because an information source can be considered a group of features that share the same nature. The selection of relevant features has been widely addressed in the machine learning field, and it is possible to adapt classical feature selection methods to simultaneously select or delete groups of features, which is known as group feature selection [10,11]. The studies that apply group feature selection usually identify and distinguish the information sources in a specific research area, such as electrodes in a Brain–Computer Interface (BCI) system [12] or different frequency bands in multispectral and hyperspectral imaging [13].
The most basic group feature selection methods use a defined classification threshold; they take a reference value that must be reached to decide whether or not eliminating a feature group [11,14]. Other methods use more advanced procedures to determine the relevance of a set of features based on a penalty imposed during the learning stage of the algorithm [15]; aiming the self method eliminates or reduces the effect of the least relevant feature groups by decreasing the values of the weights assigned to them [14,16]. There is another pair of group feature selection methods known as backward elimination and forward addition [17]; they are usually presented together and consist of removing or adding features to the classification task, and analyzing the performance curve generated during training, seeking to identify the points where the performance of the algorithm is maximum.
Although there is a wide variety of strategies that can be implemented to select feature groups, there is a common challenge all these methods share: the fact that the implementation of algorithms based on single feature selection strategies does not guarantee that the whole information source can be seen as a complete and independent block. That is, these methods usually retain an individual notion of the features to determine an apparent relevance of the information source, thus losing the overall properties of the source and causing breaks between some important relationships between the features that compose the source.
One method that has grown in importance in recent years in the machine learning field, regarding the use of multiple information sources, is Multiple Kernel Learning (MKL) [18,19]. MKL allows the implementation of a similarity measure (kernel) associated with each information source in an independent way before each of those sources is integrated into the learning task [20,21], thus taking advantage of the information coming from each source. Besides, it enables users to obtain easy-to-interpret results in relation to the analysis of each information source [22].
MKL has brought a lot of advantages for different tasks in which it has been implemented, especially when used with the Support Vector Machine (SVM) method [23] as well as with other types of machine learning algorithms [24]. In addition, MKL is well documented in the state of the art and has proven to be useful for the identification and selection of relevant information sources [25,26]. This is because it allows the user to have a similarity measure of data for each information source without losing the possible internal relationships of the features that compose the source, thus enabling objective studies into relevance analyses that are very easy to interpret.
In the state of the art, the use of MKL has been reported when weights are assigned to each kernel associated with the information sources [27], which is an important feature of this methodology since these weights can be used to determine the relevance that each source represents for the implemented solution [28,29,30]. It has been effectively demonstrated that the selection of relevant information sources by penalizing kernel weights is very useful and provides information that is easy to interpret [31], in addition to being a method that can be taken to different research areas.
This work proposes a novel method to objectively select the most relevant information sources in a classification task; this method uses the local scaling technique to tune the parameters of Gaussian kernels associated with each information source by using the MKL. Instead of computing a unique kernel bandwidth for all data, the local scaling technique computes a kernel bandwidth for each sample; thus, it exploits local statistics of sample neighborhood, capturing structure in data [32]. This technique has been used in combination with MKL to perform adaptive unsupervised clustering [33]. In this paper, three different algorithms were also proposed to adapt the local scaling technique to be used during the prediction stage in a supervised classification task, allowing to reduce the computational complexity of the tuning process of kernel parameters considerably. The proposed method is evaluated over two real application tasks: the selection of electrodes for a classification task in Brain–Computer Interface (BCI) systems and the selection of relevant Magnetic Resonance Imaging (MRI) sequences for detection of breast cancer. The obtained results show that the proposed method is stable regarding the sources which are selected as relevant when any of the three proposed algorithms are applied.

2. Methodology

This paper proposes a novel method to address the problem of identifying and selecting relevant available information sources by solving a binary classification task in an objective way. The proposed method is based on the use of techniques that have been well studied in the machine learning area, but without the joint implementation reported in the literature. This section contains a detailed explanation of the techniques involved in the proposed method, beginning with the theoretical framework of each technique and finishing with the pseudocode of the algorithms designed to apply such a method.

2.1. Support Vector Machines (SVMs)

As mentioned before, the proposed method is only applicable to binary classification problems (for now). This restriction is due to the use of an SVM classifier as base learner in the machine learning task. The SVM is a well-known classifier designed adopting the structural risk minimization theory in order to produce a successful generalization of the prediction using unknown data [34]. The SVM classifier finds the boundary line that better discriminates the training samples contained in a database represented by x i , y i i = 1 N . As a result of this property, SVMs are known as large margin classifiers [35]. Each x i in the database is an input vector of dimension D, and y { 1 , + 1 } is the label vector which has N elements.
The classification function f ( x ) = w , x + b defines the distance to the hyperplane which can be seen as a membership degree assigned by the SVM to a test sample, where w is the weight associated with each sample x i , b is the hyperplane bias term, and the · , · operator refers to the dot product between vectors. The primal optimization problem of the SVM is presented in Equation (1), and w and b are computed when this problem is solved.
w * = min w 1 2 w 2 2 + C i = 1 N ξ i s . t . y i ( w , x + b ) 1 ξ i
where C is a regularization parameter defined by the user, and ξ is the vector of slack variables.
Lagrange Multipliers are used to solve the quadratic optimization problem in Equation (1) [34], changing the classification function to f ( x ) = i = 1 N α i y i K ( x i , x ) + b and creating the final objective function presented in Equation (2), namely, the dual function of the SVM.
max 0 α C i = 1 N α i 1 2 i = 1 N j = 1 N α i α j y i y j K ( x i , x j ) s . t . i = 1 N α i y i = 0
where α is the dual variables vector originated by applying Lagrange Multipliers. The term K ( x i , x j ) in Equation (2) is known as the kernel function; it also appears in the classification function. It is essential for the performance of the SVM since it generalizes the SVM classifier to solve non-linear problems by allowing the calculation of non-linear decision boundaries. A kernel represents a similarity measure (dot product) between two samples and is commonly expressed as K ( x i , x j ) , where K : R D × R D R . Such dot products are made implicitly in a high-dimensional Hilbert Space by using the kernel trick, without the explicit knowledge or use of a mapping function to a high-dimensional space [36].
There are different types of representations of a kernel function, which to be valid and represent dot products on Hilbert spaces must fulfill Mercer’s condition [36]. The simplest one is the dot product ( x i , x j ), namely, linear kernel, and one of the most commonly used is the Gaussian kernel, also known as radial basis function kernel, which is defined by K ( x i , x j ) = exp x i x j 2 2 σ 2 , where σ is the kernel bandwidth and must be a positive real number.
The algorithm most commonly used to handle the dual function of the SVM is the Sequential Minimal Optimization (SMO) algorithm [37]. There are a lot of programming libraries designed to successfully apply the SMO algorithm, e.g., the LIBSVM library is one the most widely implemented in the literature [38] because it has a cutting-edge repository that can be used over different programming languages and is well supported by the scientific community.

2.2. Kernel Bandwidth Tuning with Local Scaling

A recurrent challenge found in the literature is the correct tuning of the σ parameter when a Gaussian kernel is used in the SVM [39,40,41]. This problem has been addressed with different strategies, mainly metaheuristic optimization techniques that consume significant time and computational resources, which is why it remains an open issue in the machine learning area.
Zelnik-Manor and Perona [32] proposed a clever and intuitive method for tuning the σ parameter without using metaheuristic methods. Furthermore, their strategy allowed them to define a different σ i value related to each sample in the database instead of a global σ ; this strategy is known as local scaling. To compute the local scaling parameter for each x i in the database, it is necessary to analyze the local statistics of its neighborhood because the value of σ i is defined by Equation (3).
σ i = d x i , x K
where x K is the K-th nearest neighbor of sample x i and d ( · , · ) is some distance function used to evaluate the local statistics of the data. This work uses the Euclidean distance following the proposal of Zelnik-Manor and Perona [32]. The K value determines the neighborhood size and relies on the scale or density of the samples space, i.e., a larger K value represents a larger similarity among samples, while a small K value focuses on local similarities. Therefore, a new parameter should be tuned when local scaling is used. This parameter is the correct neighbor K to compute the distance that defines each σ i value.
If any optimization algorithm is used for tuning the σ value in the general radial basis function kernel, it should deal with an infinite searching space of real values. Otherwise, when the local scaling technique is used, even if K remains as a free parameter, the optimization problem is reduced to a limited space of integer values considering that K Z : 1 K < N . Besides, the neighborhood of x i will be defined for a small number of samples because using a large K value would cause the loss of the local scaling property. When the local scaling strategy is applied, the Gaussian kernel is computed using Equation (4).
K ( x i , x j ) = exp x i x j 2 2 σ i σ j

2.3. Multiple Kernel Learning (MKL)

In recent years, MKL has attracted interest in different research areas, especially because it allows users to take advantage of multiple information sources to solve machine learning problems, even when each information source has a different nature [25,42]. The MKL methodology establishes that multiple linear or non-linear combinations of kernels can be used instead of one single kernel. One of the most commonly used MKL functions is a weighted sum of kernels as shown in Equation (5). This function enables the use of the individual information provided by each data source, in addition to keeping the internal relationships between the features that compose the whole source.
K η ( x i , x j ) = m = 1 P η m K m ( x i m , x j m )
where P is the number of information sources that compose the database and η m represents the weight assigned to each kernel function K m .
An important detail of the MKL function presented in Equation (5) is the possibility of applying a penalization to the weights η in order to identify the most relevant information sources, i.e., the possibility of selecting the sources that provide the most useful information to solve the classification task that is being studied [43]. One of the types of penalization most commonly applied to the η weights is the 1 -norm penalization. The latter provides a straightforward interpretation of results because it fulfills the characteristic of being a sparse penalization type, making some weights equal zero and thus eliminating the least useful information sources [44]. The 1 -norm penalization satisfies the condition η Δ , where Δ is the domain of η and is defined by Equation (6), which corresponds to a convex sum, namely, the simplex condition.
Δ = η R + P : m = 1 P η m = 1 , η m 0
Equation (6) clearly defines how the η weights can be interpreted as relevance measures because each one of them is associated with an individual kernel, which, in turn, is associated with an independent information source. Additionally, all the information sources could be different in terms of nature and number of features without affecting the implementation of the strategy. Kloft et al. [45] and Xu et al. [42] conducted two different studies to find the correct way to train the η weights using a generalized penalization of the p -norms with p 1 . Although the analysis in the two studies was performed in a different way, the result was the same. They found an iterative optimization strategy that solves an SVM and updates the kernel weights with Equation (7) until a tolerance measure is satisfied.
η m = w m 2 2 p + 1 h = 1 P w h 2 2 p p + 1 1 p
where w m 2 2 = η m 2 i = 1 N j = 1 N α i α j y i y j K m x i m , x j m , which is computed from the dual function of the SVM.

2.4. Description of the Proposed Method

The proposed method uses MKL to select relevant sources from multimodal data, i.e., the multiple sources available for the classification task. One matrix represents each information source with N columns corresponding to the samples and an independent number of rows corresponding to the features. The local scaling technique is used to compute the σ i values representing the local statistics of the data, what is necessary to compute the P kernels associated with the information sources. Then, the global kernel K η obtained after applying the MKL function based on the weighted sum of kernels is fed into an SVM classifier to complete a training process and solve the task. Thus, obtaining a trained predictive model with the definitive η values associated with the relevance measures of the information sources. Figure 1 shows a rough representation of the proposed method as a flowchart.
Because the local scaling strategy was proposed to be applied over unsupervised learning tasks [32], there is no defined process to apply this technique in a supervised learning approach, where labeled data is used to create a predictive model and avoid using all the training information at the prediction stage. Explicitly using the SVM classifier, it is necessary to define a way to compute or estimate each σ value associated with the test samples, contemplating that an SVM only requires the information of some samples (support vectors) to classify unknown data, whereby there is missing information about the original local statistics in the prediction stage. This work proposes three different algorithms to expand local scaling applications to supervised learning tasks; these algorithms are described below.

2.4.1. Prediction without Data Reduction

The first idea that comes to mind in order to avoid this problem is to compute the distance between the test and training samples and assign the σ values to the test samples using the distance of the K-th neighbor. A clear disadvantage of this solution is that all the training samples will be necessary in the prediction stage in order to correctly compute the σ values of the test samples. Algorithm 1 describes how this solution can be applied.
Algorithm 1: Algorithm without data reduction
Sensors 20 03919 i001

2.4.2. Prediction with Support Vectors Only

Although the first algorithm represents the most accurate approach to compute the σ values of the test samples, it wastes one of the most important characteristics of the SVM: its ability to implement only a few training samples, known as support vectors, in the prediction stage. The second algorithm was designed with the aim of avoiding this issue, using only the σ values of the support vectors. In this algorithm, the distance between the test samples and support vectors is computed, assigning to each test sample the corresponding σ of its nearest support vector. Algorithm 2 describes how this strategy can be applied.
Algorithm 2: Algorithm with data reduction by support vectors only
Sensors 20 03919 i002

2.4.3. Prediction with Mean σ Values of Support Vectors

The second algorithm presents the best solution with respect to data processing and memory consumption, but it appears to ignore information that belongs to the training samples that were not selected as support vectors and could be important to determine the σ values of the test data. The third algorithm was designed to avoid this issue. In this algorithm, a new step is added to the training stage. It consists in identifying all the training samples that had not been selected as support vectors, assigning to each one of them the nearest support vector, and, finally, computing, for each support vector, a new σ value composed of the mean of the σ values belonging to its nearest training samples. In the prediction stage, the distance between the test samples and support vectors is computed, assigning to each test sample the σ mean associated with its nearest support vector. This process is described in Algorithm 3.
Algorithm 3: Data reduction by the mean of σ values related to the samples nearest to the support vectors
Sensors 20 03919 i003

3. Experimental Setup

The proposed method can be applied to any well-structured classification task with a multimodal designation. For this reason, we decided to test it in two different tasks and thus demonstrate its usefulness. The first application scenario was a well-known state-of-the-art problem: the selection of relevant electroencephalography (EEG) channels in a Brain–Computer Interface (BCI) classification task. The second scenario was a problem that has not been addressed by the machine learning field: the objective selection of relevant information sources, more specifically, the selection of relevant Magnetic Resonance Imaging (MRI) sequences for breast cancer detection. This section describes the configuration of the datasets and the structure of all the experiments conducted here to evaluate the proposed method.

3.1. BCI Dataset

This dataset was published for the BCI competition IV dataset 2a, held in 2008 by the Graz University of Technology in Austria [46]. The dataset is composed of EEG signals taken from 9 healthy subjects. The EEG signals can be categorized into 4 classes of motor imagery: left hand, right hand, feet, and tongue. Each subject completed two sessions on different days; each session consisted of 6 runs separated by short time lapses. A total of 48 trials were recorded in each run (12 per class), thus conducting 288 trials per session. In this study, only the motor imagery of the left and right hands was used since the proposed method is restricted to binary classification problems only. Besides, only the trials of one session were provided with their corresponding labels. Finally, 72 trials were carried out for imaging the movement of the left hand and 72 for the right hand of each subject. In order for the method (which is based on an SVM) to carry out an adequate classification task, the left hand signals were assigned to class 1 and the right hand signals were assigned to class 1 .
The signals were acquired with a sampling frequency of 250 Hz using 22 electrodes with an inter-electrode distance of 3 . 5 cm based on the international 10–20 system (see Figure 2). In this case, each electrode was an independent information source that provided descriptive data to be used in the problem solution. Hence, this dataset was composed of 22 information sources in relation to the notation presented in this document. In addition to the 22 electrodes, three electrooculography (EOG) channels were added to the setup during data acquisition. The data provided by these EOG channels were not used as information sources for the classification task; they were only used for applying a preprocessing step to remove artifacts from the EEG signals. Besides, two filters were applied to all the signals (including EOG) as another preprocessing step: (1) a bandpass between 0 . 5 Hz and 100 Hz and (2) a 50 Hz notch filter.
A feature generation process was applied to the preprocessed signals. Such process allowed us to extract descriptive measures and information to represent the signals in a compressed way. The mother wavelet transformation Daubechies of order 2 (db2) has proven to be efficient in describing EEG signals for classification tasks [47]; thus, it was selected to generate the features of the BCI dataset in this study. Different decomposition levels were considered, taking into account that this dataset is composed of motor imagery signals. As a bandpass filter between 0.5 Hz and 100 Hz was applied to the signals, the detail coefficients of the third (D3) and fourth (D4) decomposition levels are associated with the mu (6.25–12.50 Hz) and beta (12.50–25.00 Hz) rhythms, respectively. These rhythms are commonly used to process motor imagery EEG signals in the state of the art [48,49,50]; hence, they were computed in this dataset. Additionally, the frequency ranges 0.5–3.13 Hz and 3.13–6.25 Hz are also highlighted in the literature. They are associated with approximation (A5) and detail (D5) coefficients of the fifth decomposition level, respectively [50]. Thus, they were also used in this study.
After computing the wavelet coefficients D3, D4, D5, and A5, six statistical measures of each coefficient and the original signal were calculated: mean, median, mode, variance, kurtosis, and standard deviation [51]. Additionally, the energy of each coefficient and the original signal was also computed. As a result, we obtained a total of 35 features per information source.

3.2. Breast Cancer Dataset

The MRI breast cancer dataset was developed by Instituto Tecnológico Metropolitano (ITM) in collaboration with Instituto de Alta Tecnología Médica (IATM) in Medellín, Colombia. It is composed of 87 studies, and each study corresponds to one subject and presents at least one Region of Interest (ROI). In total, 146 ROIs were extracted from the entire dataset. Each study in the dataset is represented by nine information sources, and each source corresponds to an MRI sequence. The sequences included in the dataset are the relaxations T1 and T2, Diffusion-Weighted Imaging (DWI), Apparent Diffusion Coefficient (ADC), and subtractions 1 to 5 obtained from Dynamic Contrast Enhanced (DCE) sequences [52]. Each MRI sequence highlights different three-dimensional information from the tissue under analysis so that it can be ensured that each sequence corresponds to an independent information source. Figure 3 shows an example of the visualization of these sequences.
The complete dataset was evaluated by two expert radiologists from IATM, who classified each finding in the sequences using the Breast Imaging Reporting and Data System (BI-RADS) [53]. This system enabled them to assign a probability level of a finding to be cancer, where 5 is the highest level: highly suggestive of malignancy. In order to adapt the dataset to the classification method proposed in this paper, the ROIs were separated into two classes: class 1 for all the ROIs with a BI-RADS less than 3 and class 1 for the ROIs with a BI-RADS equal to or greater than 3. A total of 61 ROIs were associated with class 1 and 85 with class 1 applying this configuration.
Two different set of features were implemented to describe the information of the findings in each sequence, i.e., perceptual and radiomic features. Perceptual features refer to the properties of an image that can be captured by human perception [54], while radiomic features represent a quantitative approach to the analysis of medical images aimed at explaining the morphological and functional properties of a lesion [55].
A total of 10 perceptual features and 86 radiomic features were computed for each image sequence. Perceptual features were computed as the first five statistical moments from both the salient maps generated by the Graph-Based Visual Saliency method [56,57] and the original images; while radiomic features were extracted using the Pyradiomics toolbox of Python [58]. As described below, the experimental stage was performed by considering both each set of features as an independent information source and the combination of the two sets of features as one unique information source. Thus, a total of 18 information sources per instance were obtained, i.e., nine sources that correspond to the set of perceptual features for each sequence (P_T1, P_T2, P_ADC, P_DWI, P_SUB1, P_SUB2, P_SUB3, P_SUB4, and nine other sources that correspond to each set of radiomic features (R_T1, R_T2, R_ADC, R_DWI, R_SUB1, R_SUB2, R_SUB3, R_SUB4 and R_SUB5).
The main objective of integrating perceptual and radiomic features in this study is to obtain relevant results for both machines and radiologists.

3.3. Test Settings

The tests conducted here to evaluate the performance of the proposed method take into account different variations that it may present. They are mainly associated with the three algorithms proposed to adapt the local scaling method to the supervised learning task, the types of features generated from the MRI sequences, the number of subjects in the BCI dataset, and the two main types of penalties that can be applied to the η weights associated with the kernels (which correspond to the 1 -norm and the 2 -norm). Therefore, it is necessary to explain how the tests were configured for each dataset.

3.3.1. BCI Dataset Configuration

The tests performed using the BCI dataset were divided into two large groups determined by the penalty types ( 1 -norm and 2 -norm). Then, each of these groups was divided into three subgroups determined by the three algorithms proposed here to adapt the local scaling method. Finally, a test was conducted on each subgroup using each one of the nine subjects that compose the dataset. This resulted in a total of 54 different tests taken into account to evaluate the performance of the proposed method with the BCI dataset.

3.3.2. MRI Dataset Configuration

In the same way, as with the BCI dataset, the tests with the MRI dataset were divided into two large groups determined by the penalty types ( 1 -norm and 2 -norm). Then, each test group was divided into three subgroups determined by the available feature types. The first subgroup included only perceptual features; the second, only radiomic features; and the third, both feature types. Finally, a test was applied to each subgroup using each one of the three proposed algorithms. This resulted in a total of 18 different tests taken into account to evaluate the performance of the proposed method with the MRI dataset.

3.3.3. General Configuration of the Tests

Although the purpose of the proposed method is the objective selection of relevant information sources, it is necessary to evaluate the effectiveness of the method based on its performance in terms of the classification task. Therefore, we decided to implement 4 well-known performance measures for classification tasks [59]: accuracy, sensitivity, specificity, and geometric mean (Geo-Mean). They were computed using Equations (8) to (11), respectively.
Accuracy = Correctly   predicted   samples Total   number   of   samples
Sensitivity = True   positives True   positives   +   False   negatives
Specificity = True   negatives True   negatives   +   False   positives
Geo Mean = Sensitivity   ·   Specificity
Each dataset was divided maintaining a 20 % of the samples outside the training and validation stage to use them as test samples, while the remaining 80 % was used to train the model. For the training and validation process, the cross-validation method K-Fold was implemented with a total of 10 folds in order to be able to train the method in such a way that it could correctly generalize the predictions about test data.
Although the method proposed by Zelnik-Manor and Perona [32] set the K value to 7 because it produced good general results, in this work, K was defined as a parameter to be optimized, in addition to the regularization parameter C, and the parameter for the SMO algorithm τ . The Particle Swarm Optimization (PSO) algorithm was implemented to find the optimal values of these parameters. The PSO is one of the most commonly used optimization algorithms in the state of the art [60,61,62]. It uses cooperative and stochastic methods to find the optimal parameters of the function to be optimized; in this case, the performance of the classifier. In all the tests, we used 40 search particles and individual and social learning coefficients equal to 1 . 1931 [63], and the inertia value was dynamically established between 0 . 1 and 1 . 1 with a maximum value of 60 iterations. The cost function used for the optimization process was the geometric mean because it was the performance measure that showed the best stability and quick convergence regarding the optimal values that are found.
Once the optimization process was completed using the PSO and cross-validation in each test, the optimal parameters found were used to train a global model using the total training data. This model was finally employed to make the prediction on the test data that had been initially separated, thus obtaining the trained values of the η weights that determine the relevance of the information sources and the predictions on the test data that would allow us to measure the performance of the method as a classifier.
Furthermore, Equation (12) was implemented in order to measure the effectiveness of the method in terms of the reduction rate of the number of information sources, taking into account that the ideal situation is to obtain a high reduction rate because it means the use of a minimal number of sources to solve the classification task.
Reduction   rate = 1 Number   of   sources   selected Total   number   of   sources

4. Results and Discussion

This section reports the results obtained with the BCI and the MRI datasets. Such results are analyzed in detail and discussed taking into account some relevant studies in the literature that present reference points to assess the performance of the proposed method.

4.1. Results Obtained with the BCI Dataset

A total of 54 tests were conducted with this dataset, as described in the previous section. After completing the training stage and obtaining the optimal values for the free parameters, a prediction process was performed with the test data while a sequential reduction as applied to the number of information sources. This was possible because the global model obtained after the training stage contained the η values of each information source (each electrode), and these values are directly associated with the relevance measure obtained for each source. As a result, the most relevant information sources can be identified and sequentially eliminated from least to most relevant when the η values are sorted from highest to lowest.
Figure 4 shows the curves obtained with the process of reduction of information sources. Taking into account that there is a lot of information, the results were condensed in three different figures that represent the application of the three proposed algorithms. Such figures show the average value and standard deviation of the accuracy obtained for the nine subjects, separated by the application of the 1 -norm and 2 -norm.
Figure 4a shows the performance curves of Algorithm 1. In it, the highest mean performance with the lowest number of sources is obtained using the 1 -norm; nevertheless, the 2 -norm shows a very similar behavior, reaching a high performance when only 2 information sources are used. Figure 4b shows the performance curves of Algorithm 2. In this case, the 2 -norm achieves the highest performance with the lowest number of sources, more specifically, between 4 and 6 sources. Finally, Figure 4c shows the performance curves of Algorithm 3. Again, both the 1 -norm and the 2 -norm exhibit similar behavior, although the 2 -norm slightly outperforms the 1 -norm when only 5 information sources are used. Based on these results, we decided to include only the detailed performance specifications of the method using the 2 -norm because the performance of both penalization types was very similar, and including all the detailed performance results could seem redundant.
Table 1 details the classification performance results obtained with the BCI dataset when Algorithm 1 was applied with the 2 -norm. Said table compares the results obtained when the 22 information sources were used and the best result obtained per subject with the lowest number of sources that could be used. These sources are sorted by relevance from high to low. Evidently, the performance measures always improve when the number of sources is reduced. It is also important to highlight that each subject requires a different number of information sources and that the selected electrodes are different for each subject (with some coincidences in several cases). Table 1 also shows the time required by the computer to train the global model and make predictions with the test data. All these tests were conducted on a work station with an Intel processor of 16 cores at 3 . 00 GHz and 16 GB of RAM. The last column in the table shows the number of samples required by the algorithm to compute the prediction with the test data. In this case, all the training samples were required by the algorithm because that is precisely the main feature of Algorithm 1 and also its worst disadvantage. The last row shows the average computed using all the information obtained from all the subjects with its corresponding standard deviation. The relevant electrodes were selected by analyzing which of them were the most voted among all the subjects.
Table 2 details the results obtained with the BCI dataset when Algorithm 2 was used with the 2 -norm. These results are very similar to those obtained in Table 1. The most important aspect to highlight in this table is the fact that all the most relevant electrodes selected in the average results were also selected by Algorithm 1, which means high stability in the method among the algorithms. Furthermore, the results show again that the proposed method can improve the performance (measures) of the classification task and, at the same time, produce a reduction in the required number of samples (nearly 72 % ) because this algorithm only uses the support vectors to make predictions.
Table 3 presents the results of the method applied to the BCI dataset when Algorithm 3 and the 2 -norm are used. The electrodes selected as relevant are the same as in Table 1 and Table 2, thus confirming, once again, the stability of the method in terms of the selection of relevant information sources even when the type of algorithm is different. Table 3 also shows that Algorithm 3 reduces the number of samples required for the prediction by nearly 74 % . We should remember that this algorithm takes the information of the samples that the SVM does not select as support vectors for computing mean sigma values and takes advantage of all the available data.
Taking into account that the stability and reliability of the relevant information sources selected by the method are the most critical factors to analyze, we created Table 4 with the average relevance assigned to each electrode among all the subjects. The accuracy obtained in the classification of each subject was used to weight this relevance, considering that the electrodes selected as relevant by the method are more reliable when the obtained accuracy is higher. In Table 4, it can be seen that the method selected electrodes 8, 13, 14, 18, and 22 as relevant in all the configurations among all the algorithm and penalization types. This table was organized based on the information in Table 1, Table 2 and Table 3, by taking the first eight selected electrodes as relevant. Furthermore, if the penalization type is used to analyze the results presented in Table 4 separately, it can also be seen that electrode 7 was classified as relevant with the 1 -norm; and electrode 1, with the 2 -norm.

4.2. Results Obtained with the MRI Dataset

A total of 18 tests were conducted with the MRI dataset. As in the tests applied to the BCI dataset, a prediction process was performed on the test data while making a sequential reduction in the number of information sources after completing the training stage and obtaining the optimal values for the free parameters. Figure 5 shows the curves obtained with the process of reduction of information sources on the MRI dataset. Three different figures that represent the use of the perceptual features, radiomic features and a combination of both summarize the results. The figures show the average value and standard deviation of the accuracy obtained with the three algorithms, separated by the application of the 1 -norm and 2 -norm.
Figure 5a shows the performance curves obtained using only the perceptual features. In this figure, the 1 -norm clearly outperforms the 2 -norm, reaching the highest mean accuracy with only 5 information sources, which slowly decreases with fewer sources. Figure 5b shows the performance curves obtained using only the radiomic features. In this case, both the 1 -norm and the 2 -norm present a similar behavior, although the 2 -norm slightly outperforms the 1 -norm when 5 information sources are used. Finally, Figure 5c shows the performance curves using perceptual and radiomic features simultaneously. Again, both the 1 -norm and the 2 -norm exhibit a similar performance, and the 1 -norm slightly outperforms its counterpart when only 6 information sources are used. Based on these results, we decided to include only the detailed performance results of the method using the 1 -norm, although the results were remarkably similar to the two penalization types.
Table 5 details the results of the method using only the perceptual features and applying the 1 -norm. This table presents different performance measures obtained when all the information sources and only the most relevant information sources for each algorithm are implemented. In this case, the table specifies the sources selected as relevant to represent the MRI sequences and also indicates the corresponding reduction rate. Additionally, Table 5 shows the time needed by the algorithm to complete the training stage of the global model and the prediction stage with the training and test data. The last column shows the number of training samples that were needed to build the global model (support vectors).
It is evident in Table 5 that identifying which information sources are relevant for the classification task can significantly improve its performance. Algorithm 2 also presents better performance than the other two algorithms, obtaining the best results in all the performance measures when source reduction was applied. Note the coincidence in the information sources selected as relevant by Algorithms 2 and 3, which show a very similar performance that is also better than that of Algorithm 1.
Table 6 details the results obtained using the method with the 1 -norm and only the information sources obtained from the generation of radiomic features on the MRI sequences. In this case, the reduction rate was considerably lower with all the algorithms compared to the previous table, but again the algorithms showed appropriate stability regarding the sources selected as relevant. In addition, Algorithms 2 and 3 present an advantage, especially because they considerably reduce the amount of data that they require for the construction of the final model.
Finally, the results in Table 7 take into account a higher number of information sources because they combine the sources obtained from the generation of perceptual and radiomic features. In this case, Table 7 shows very similar results with the three algorithms, a reduction rate of more than half of the information sources with all the algorithms, and an almost perfect match of the information sources selected as relevant. It is possible to identify highly relevant measures in the MRI sequences represented by subtractions in most tests, with both the perceptual and radiomic features.

4.3. Comparison against Feature Selection Methods

The closest approaches to the automatic selection of sources using machine learning are the feature selection techniques developed in said research area [64]. It is possible to associate the selection of information sources with the selection of features, considering that it is reasonable to select individual features hoping to eliminate all the features of one specific information source. Hence, the feature selection technique could eliminate the whole information source.
A new group of tests was performed based on the above and aims to evaluate the contribution of the proposed method to the objective selection of information sources provided for a binary classification task (compared to classical feature selection techniques). These tests consisted of two different experiments: (1) a feature selection process applied to all the information sources in order to automatically eliminate some of these sources if the feature selection process eliminated all the features that composed them; and (2) the application of the proposed method to reduce the number of information sources and the subsequent implementation of the feature selection technique for reducing the number of features that composed the relevant information sources.
The technique used to perform the feature selection process was the Fisher Score [65]. This method is a supervised feature selection technique that seeks to preserve features that have a more uniform distribution in one class and a more dispersed one in the others. The Fisher Score method is represented by Equation (13).
F S f i = j = 1 l n j μ i , j μ i 2 j = 1 l n j s i , j 2
where l is the number of classes; μ i , the mean of the feature f i ; n j , the number of samples in the j-th class; and μ i , j and s i , j , the mean and the variance of f i in the j-th class, respectively [66].
Equation (13) is used to obtain a rank vector ordering the features from the most relevant to the least relevant. Once the features are ordered, they are eliminated one by one starting by the least relevant till identifying the number of features that generate the highest classification performance.
In the first experiment, the whole features extracted from all information sources are concatenated in a unique feature vector before applying the feature selection and source elimination process. Thus, the relevant information sources for the classification task will be those that retain at least one of its features in the set of features that reported the best performance. In the second experiment, the proposed source selection method is applied to find the most relevant information sources. Then, features extracted from these sources are concatenated in a unique feature vector, which is used as input to the feature selection process, in order to select the most discriminative features that improve the classification task. In this case, the relevant sources could also be reduced, by applying the same process described to the first experiment; however, all sources selected initially by the proposed approach were kept in the experiments performed in this work.
Table 8 shows the results obtained with the BCI dataset. The left side of the table groups the results of the first experiment; the right side, those of the second experiment. It can be seen that the feature selection technique improves the performance of the classifier and manages to eliminate some information sources for some subjects by itself in some cases. However, in most cases, it cannot reduce the number of information sources, even with a considerable reduction in the number of features. On the other hand, the right side of the table shows that the proposed method manages to reduce the number of information sources for all the subjects with the three proposed algorithms and, at the same time, maintains high performance in the classification task. These results also indicate that the proposed method can be used to reduce the number of information sources and, after that, a feature selection technique can be applied to reduce the number of features; the objective is to achieve the best results in terms of the highest performance with minimum data quantity.
Table 9 shows the results obtained with the MRI dataset. The results of the first experiment of tests are on the left side; and those of the second experiment, on the right side. It is essential to highlight that the tests achieved the highest performance regarding the classification task when only the feature selection technique was applied, but in this case the task required the use of all the information sources, missing the main objective of the tests. Conversely, the results on the right side of the table show that the proposed algorithm maintained a high performance during all the tests while it eliminated some information sources (more than half in most tests), successfully achieving the proposed objective.
The sources selected as relevant during this group of tests were the same as those presented in the results in Section 4.1 and Section 4.2, which again highlights the stability of the proposed method regarding the objective selection of relevant information sources. Moreover, additional tests were conducted to consider the application of the feature selection technique before reducing the information sources following the structure of tests carried out in some studies in the literature. However, the results of those tests are not reported in this paper because they exhibited the same behavior regarding the information sources selected as relevant and the performance measures they achieved, confirming the stability of the proposed algorithm remains in all the evaluation environments considered here.

5. Conclusions

This work proposes an automatic method that implements different types of machine learning algorithms to objectively select the most relevant information sources in a classification task. The proposed method computes a Gaussian kernel for each information source, performing the tuning of the kernel parameters by using the local scaling technique. This technique was adapted through three different algorithms to be used during the prediction stage in the supervised classification task implemented over an SVM. Algorithm 1 uses all the training samples to compute the kernel bandwidth parameter ( σ ) corresponding to each prediction sample. Algorithm 2 uses only the SVM support vectors to determine the σ , thus reducing the amount of information required by the trained model. In contrast, Algorithm 3 collects the neighborhood information of the support vectors to estimate the σ value of each prediction sample, attempting to keep the information of training samples that are not selected as support vectors while reducing the information required by the trained model.
Two real application tasks were used to evaluate the proposed method: the selection of electrodes for a classification task in Brain–Computer Interface (BCI) systems and the selection of relevant Magnetic Resonance Imaging (MRI) sequences for detection of breast cancer. The obtained results show that the proposed method is stable regarding the sources which are selected as relevant, even when the results are analyzed among the three proposed algorithms. Although the three algorithms presented similar performances and fulfilled the expected function of source selection, it is possible to find a relatively superior performance of Algorithm 2 when using the 1 -norm and likewise, a leading behavior of Algorithm 3 when using the 2 -norm. These two algorithms have the particular advantage of reducing the number of samples required to build the trained model since they only require the information of the samples selected by the SVM as support vectors.

Author Contributions

The work was performed with the substantial contribution of all the authors. H.J.A.-L. carried out the research study, implemented the algorithms, performed the experimentation and wrote the first draft-manuscript; G.M.D., M.L.H., and A.E.C.-O. conceived the project; H.J.A.-L., A.E.C.-O., and G.M.D. conceived and designed the experiments; M.L.H. has contributed to image analysis and interpretation. All authors have contributed to the preparation of the paper, the discussing of the results and the reviewing of the content of this article. All authors have read and agreed to the published version of the manuscript

Funding

This research was funded by MinCiencias (Colombia), Instituto Tecnológico Metropolitano, and Ayudas Diagnósticas Sura. Project RC740-2017.

Acknowledgments

We would like to acknowledge to the staff of the MIRP laboratory and the IATM Research Group for its technical support in the MRI image database generation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Culache, O.; Obadă, D.R. Multimodality as a Premise for Inducing Online Flow on a Brand Website: A Social Semiotic Approach. Procedia-Soc. Behav. Sci. 2014, 149, 261–268. [Google Scholar] [CrossRef] [Green Version]
  2. Markonis, D.; Schaer, R.; Müller, H. Evaluating multimodal relevance feedback techniques for medical image retrieval. Inf. Retr. J. 2016, 19, 100–112. [Google Scholar] [CrossRef]
  3. Adali, T.; Levin-Schwartz, Y.; Calhoun, V.D. Multimodal data fusion using source separation: Application to medical imaging. Proc. IEEE 2015, 103, 1494–1506. [Google Scholar] [CrossRef]
  4. Correa, A.G.; Orosco, L.; Laciar, E. Automatic detection of drowsiness in EEG records based on multimodal analysis. Med. Eng. Phys. 2014, 36, 244–249. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, Z.; Yin, H.; Chai, Y.; Yang, S.X. A novel approach for multimodal medical image fusion. Expert Syst. Appl. 2014, 41, 7425–7435. [Google Scholar] [CrossRef]
  6. Barachant, A.; Bonnet, S. Channel selection procedure using Riemannian distance for BCI applications. In Proceedings of the 2011 5th International IEEE/EMBS Conference on Neural Engineering, Cancun, Mexico, 27 April–1 May 2011; pp. 348–351. [Google Scholar]
  7. Eliseyev, A.; Moro, C.; Faber, J.; Wyss, A.; Torres, N.; Mestais, C.; Benabid, A.L.; Aksenova, T. L1-penalized N-way PLS for subset of electrodes selection in BCI experiments. J. Neural Eng. 2012, 9, 045010. [Google Scholar] [CrossRef]
  8. Meyer, J.S.; Siegel, M.J.; Farooqui, S.O.; Jaramillo, D.; Fletcher, B.D.; Hoffer, F.A. Which MRI sequence of the spine best reveals bone-marrow metastases of neuroblastoma? Pediatr. Radiol. 2005, 35, 778–785. [Google Scholar] [CrossRef]
  9. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 94. [Google Scholar] [CrossRef] [Green Version]
  10. Gan, G.; Ng, M.K.P. Subspace clustering with automatic feature grouping. Pattern Recognit. 2015, 48, 3703–3713. [Google Scholar] [CrossRef]
  11. Pir, D.; Brown, T. Acoustic Group Feature Selection Using Wrapper Method for Automatic Eating Condition Recognition. In Proceedings of the Sixteenth Annual Conference of the International Speech Communication Association, Dresden, Germany, 6–10 September 2015. [Google Scholar]
  12. Lal, T.N.; Schroder, M.; Hinterberger, T.; Weston, J.; Bogdan, M.; Birbaumer, N.; Scholkopf, B. Support vector channel selection in BCI. IEEE Trans. Biomed. Eng. 2004, 51, 1003–1010. [Google Scholar] [CrossRef] [Green Version]
  13. Sotoca, J.M.; Pla, F.; Sanchez, J.S. Band selection in multispectral images by minimization of dependent information. IEEE Trans. Syst. Man Cybern. Part C 2007, 37, 258–267. [Google Scholar] [CrossRef]
  14. Xiang, S.; Yang, T.; Ye, J. Simultaneous feature and feature group selection through hard thresholding. In Proceedings of the 20th ACM SIGKDD International Conference On Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 532–541. [Google Scholar]
  15. Schmidt, M. Least squares optimization with L1-norm regularization. CS542B Proj. Rep. 2005, 504, 195–221. [Google Scholar]
  16. Subrahmanya, N.; Shin, Y.C. Automated sensor selection and fusion for monitoring and diagnostics of plunge grinding. J. Manuf. Sci. Eng. 2008, 130, 031014. [Google Scholar] [CrossRef]
  17. Raza, H.; Cecotti, H.; Prasad, G. Optimising frequency band selection with forward-addition and backward-elimination algorithms in EEG-based brain-computer interfaces. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–17 July 2015; pp. 1–7. [Google Scholar]
  18. Li, Y.; Wu, F.X.; Ngom, A. A review on machine learning principles for multi-view biological data integration. Briefings Bioinform. 2016, 19, 325–340. [Google Scholar] [CrossRef] [PubMed]
  19. Ren, Y.; Zhang, L.; Suganthan, P.N. Ensemble classification and regression-recent developments, applications and future directions. IEEE Comput. Intell. Mag. 2016, 11, 41–53. [Google Scholar] [CrossRef]
  20. Gu, Y.; Liu, T.; Jia, X.; Benediktsson, J.A.; Chanussot, J. Nonlinear multiple kernel learning with multiple-structure-element extended morphological profiles for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3235–3247. [Google Scholar] [CrossRef]
  21. Althloothi, S.; Mahoor, M.H.; Zhang, X.; Voyles, R.M. Human activity recognition using multi-features and multiple kernel learning. Pattern Recognit. 2014, 47, 1800–1812. [Google Scholar] [CrossRef]
  22. Xu, C.; Tao, D.; Xu, C. A survey on multi-view learning. arXiv 2013, arXiv:1304.5634. [Google Scholar]
  23. Qiu, S.; Lane, T. A framework for multiple kernel support vector regression and its applications to siRNA efficacy prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. (TCBB) 2009, 6, 190–199. [Google Scholar]
  24. Gönen, M.; Margolin, A.A. Localized data fusion for kernel k-means clustering with application to cancer biology. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 1305–1313. [Google Scholar]
  25. Gönen, M.; Alpaydın, E. Multiple kernel learning algorithms. J. Mach. Learn. Res. 2011, 12, 2211–2268. [Google Scholar]
  26. Lanckriet, G.R.; Deng, M.; Cristianini, N.; Jordan, M.I.; Noble, W.S. Kernel-based data fusion and its application to protein function prediction in yeast. In Biocomputing 2004; World Scientific: Singapore, 2003; pp. 300–311. [Google Scholar]
  27. Lewis, D.P.; Jebara, T.; Noble, W.S. Support vector machine learning from heterogeneous data: An empirical analysis using protein sequence and structure. Bioinformatics 2006, 22, 2753–2760. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Foresti, L.; Tuia, D.; Timonin, V.; Kanevski, M.F. Time series input selection using multiple kernel learning. In Proceedings of the 18th European Symposium on Artificial Neural Networks, ESANN, Bruges, Belgium, 28–30 April 2010. [Google Scholar]
  29. Tuia, D.; Camps-Valls, G.; Matasci, G.; Kanevski, M. Learning relevant image features with multiple-kernel classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3780–3791. [Google Scholar] [CrossRef]
  30. Subrahmanya, N.; Shin, Y.C. Sparse multiple kernel learning for signal processing applications. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 788. [Google Scholar] [CrossRef] [PubMed]
  31. Gönen, M. Bayesian efficient multiple kernel learning. arXiv 2012, arXiv:1206.6465. [Google Scholar]
  32. Zelnik-Manor, L.; Perona, P. Self-tuning spectral clustering. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 5–8 December 2005; pp. 1601–1608. [Google Scholar]
  33. Zhang, L.; Hu, X. Locally adaptive multiple kernel clustering. Neurocomputing 2014, 137, 192–197. [Google Scholar] [CrossRef]
  34. Vapnik, V. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 1999. [Google Scholar]
  35. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  36. Schölkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2002. [Google Scholar]
  37. Fan, R.E.; Chen, P.H.; Lin, C.J. Working set selection using second order information for training support vector machines. J. Mach. Learn. Res. 2005, 6, 1889–1918. [Google Scholar]
  38. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
  39. Lessmann, S.; Stahlbock, R.; Crone, S.F. Genetic algorithms for support vector machine model selection. In Proceedings of the 2006 IEEE International Joint Conference on Neural Network Proceedings, Vancouver, BC, Canada, 16–21 July 2006; pp. 3063–3069. [Google Scholar]
  40. Gomes, T.A.; Prudêncio, R.B.; Soares, C.; Rossi, A.L.; Carvalho, A. Combining meta-learning and search techniques to select parameters for support vector machines. Neurocomputing 2012, 75, 3–13. [Google Scholar] [CrossRef] [Green Version]
  41. Liu, J.; Zio, E. SVM hyperparameters tuning for recursive multi-step-ahead prediction. Neural Comput. Appl. 2017, 28, 3749–3763. [Google Scholar] [CrossRef]
  42. Xu, Z.; Jin, R.; Yang, H.; King, I.; Lyu, M.R. Simple and efficient multiple kernel learning by group lasso. In Proceedings of the 27th international conference on machine learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 1175–1182. [Google Scholar]
  43. Areiza-Laverde, H.J.; Díaz, G.M.; Castro-Ospina, A.E. Feature Group Selection Using MKL Penalized with 1-norm and SVM as Base Learner. In International Workshop on Experimental and Efficient Algorithms; Springer International Publishing: Cham, Switzerland, 2018; pp. 136–147. [Google Scholar]
  44. Gönen, G.B.; Gönen, M.; Gürgen, F. Probabilistic and discriminative group-wise feature selection methods for credit risk analysis. Expert Syst. Appl. 2012, 39, 11709–11717. [Google Scholar] [CrossRef]
  45. Kloft, M.; Brefeld, U.; Sonnenburg, S.; Zien, A. Non-sparse regularization and efficient training with multiple kernels. arXiv 2010, arXiv:1003.0079. [Google Scholar]
  46. Brunner, C.; Leeb, R.; Müller-Putz, G.; Schlögl, A.; Pfurtscheller, G. BCI Competition 2008–Graz Data Set A; Institute for Knowledge Discovery (Laboratory of Brain-Computer Interfaces), Graz University of Technology: Graz, Austria, 2008; Volume 16. [Google Scholar]
  47. Subasi, A. Automatic recognition of alertness level from EEG by using neural network and wavelet coefficients. Expert Syst. Appl. 2005, 28, 701–711. [Google Scholar] [CrossRef]
  48. Li, Y.; Long, J.; Yu, T.; Yu, Z.; Wang, C.; Zhang, H.; Guan, C. An EEG-based BCI system for 2-D cursor control by combining Mu/Beta rhythm and P300 potential. IEEE Trans. Biomed. Eng. 2010, 57, 2495–2505. [Google Scholar] [CrossRef] [PubMed]
  49. Marın-Castrillón, D.; Restrepo-Agudelo, S.; Areiza-Laverde, H.; Castro-Ospina, A.; Duque-Munoz, L. Exploratory Analysis of Motor Imagery local database for BCI systems. In Proceedings of the I Congreso Internacional de Ciencias Básicas e Ingeniería—CICI 2016, Meta, Colombia, 19–21 October 2016. [Google Scholar]
  50. Amin, H.U.; Malik, A.S.; Ahmad, R.F.; Badruddin, N.; Kamel, N.; Hussain, M.; Chooi, W.T. Feature extraction and classification for EEG signals using wavelet transform and machine learning techniques. Australas. Phys. Eng. Sci. Med. 2015, 38, 139–149. [Google Scholar] [CrossRef] [PubMed]
  51. Ghaemi, A.; Rashedi, E.; Pourrahimi, A.M.; Kamandar, M.; Rahdari, F. Automatic channel selection in EEG signals for classification of left or right hand movement in Brain Computer Interfaces using improved binary gravitation search algorithm. Biomed. Signal Process. Control. 2017, 33, 109–118. [Google Scholar] [CrossRef]
  52. Haacke, E.M.; Brown, R.W.; Thompson, M.R.; Venkatesan, R. Magnetic Resonance Imaging: Physical Principles and Sequence Design; Wiley-Liss New York: New York, NY, USA, 1999; Volume 82. [Google Scholar]
  53. Liberman, L.; Menell, J.H. Breast imaging reporting and data system (BI-RADS). Radiol. Clin. 2002, 40, 409–430. [Google Scholar] [CrossRef]
  54. Borji, A.; Sihite, D.N.; Itti, L. Quantitative analysis of human-model agreement in visual saliency modeling: A comparative study. IEEE Trans. Image Process. 2013, 22, 55–69. [Google Scholar] [CrossRef] [Green Version]
  55. Shaikh, F.A.; Kolowitz, B.J.; Awan, O.; Aerts, H.J.; von Reden, A.; Halabi, S.; Mohiuddin, S.A.; Malik, S.; Shrestha, R.B.; Deible, C. Technical challenges in the clinical application of radiomics. JCO Clin. Cancer Inform. 2017, 1, 1–8. [Google Scholar] [CrossRef]
  56. Harel, J.; Koch, C.; Perona, P. Graph-based visual saliency. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; pp. 545–552. [Google Scholar]
  57. Areiza-Laverde, H.J.; Duarte-Salazar, C.A.; Hernández, L.; Castro-Ospina, A.E.; Díaz, G.M. Breast Lesion Discrimination Using Saliency Features from MRI Sequences and MKL-Based Classification. In Proceedings of the Iberoamerican Congress on Pattern Recognition, Havana, Cuba, 28–31 October 2019; pp. 294–305. [Google Scholar]
  58. Van Griethuysen, J.J.; Fedorov, A.; Parmar, C.; Hosny, A.; Aucoin, N.; Narayan, V.; Beets-Tan, R.G.; Fillion-Robin, J.C.; Pieper, S.; Aerts, H.J. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017, 77, e104–e107. [Google Scholar] [CrossRef] [Green Version]
  59. Fernández, A.; García, S.; Galar, M.; Prati, R.C.; Krawczyk, B.; Herrera, F. Performance measures. In Learning from Imbalanced Data Sets; Springer: Cham, Switzerland, 2018; pp. 47–61. [Google Scholar]
  60. Areiza-Laverde, H.J.; Castro-Ospina, A.E.; Peluffo-Ordóñez, D.H. Voice Pathology Detection Using Artificial Neural Networks and Support Vector Machines Powered by a Multicriteria Optimization Algorithm. In Proceedings of the Workshop on Engineering Applications, Medellín, Colombia, 17–19 October 2018; pp. 148–159. [Google Scholar]
  61. Dávila-Guzmán, M.A.; Alfonso-Morales, W.; Caicedo-Bravo, E.F. Heterogeneous architecture to process swarm optimization algorithms. TecnoLógicas 2014, 17, 11–20. [Google Scholar] [CrossRef] [Green Version]
  62. González-Pérez, J.E.; García-Gómez, D.F. Electric field relaxing electrodes design using particle swarm optimization and finite elements method. TecnoLógicas 2017, 20, 27–39. [Google Scholar] [CrossRef] [Green Version]
  63. Clerc, M. Beyond standard particle swarm optimisation. In Innovations and Developments of Swarm Intelligence Applications; IGI Global: Hershey, PA, USA, 2012; pp. 1–19. [Google Scholar]
  64. Rincón, J.S.; Castro-Ospina, A.E.; Narváez, F.R.; Díaz, G.M. Machine Learning Methods for Classifying Mammographic Regions Using the Wavelet Transform and Radiomic Texture Features. In Proceedings of the International Conference on Technology Trends, Babahoyo, Ecuador, 29–31 August 2018; pp. 617–629. [Google Scholar]
  65. Gu, Q.; Li, Z.; Han, J. Generalized fisher score for feature selection. arXiv 2012, arXiv:1202.3725. [Google Scholar]
  66. Vora, S.; Yang, H. A comprehensive study of eleven feature selection algorithms and their impact on text classification. In Proceedings of the 2017 Computing Conference, London, UK, 18–20 July 2017; pp. 440–449. [Google Scholar]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Sensors 20 03919 g001
Figure 2. Distribution of electrodes for EEG signals acquisition.
Figure 2. Distribution of electrodes for EEG signals acquisition.
Sensors 20 03919 g002
Figure 3. Visualization of the MRI sequences in the breast cancer dataset.
Figure 3. Visualization of the MRI sequences in the breast cancer dataset.
Sensors 20 03919 g003
Figure 4. Performance curves obtained with the BCI test dataset when the number of electrodes was reduced. (a) Performance curves of Algorithm 1; (b) Performance curves of Algorithm 2; (c) Performance curves of Algorithm 3.
Figure 4. Performance curves obtained with the BCI test dataset when the number of electrodes was reduced. (a) Performance curves of Algorithm 1; (b) Performance curves of Algorithm 2; (c) Performance curves of Algorithm 3.
Sensors 20 03919 g004
Figure 5. Performance curves obtained with the test dataset when the number of MRI sequences was reduced. (a) Performance curves using only the perceptual features; (b) Performance curves using only the radiomic features; (c) Performance curves using both perceptual and radiomic features.
Figure 5. Performance curves obtained with the test dataset when the number of MRI sequences was reduced. (a) Performance curves using only the perceptual features; (b) Performance curves using only the radiomic features; (c) Performance curves using both perceptual and radiomic features.
Sensors 20 03919 g005
Table 1. Results of selection of BCI channels applying the Algorithm 1 and 2 -norm.
Table 1. Results of selection of BCI channels applying the Algorithm 1 and 2 -norm.
SubjectNumber
of Sources
Performance Measures (%)Reduction
Rate
Relevant
Electrodes
Required
Time (s)
Support
Vectors
AccGeo-MSensSpec
12264.2963.8971.4357.140.1414, 13, 18, 1,
22, 8, 7, 19, 9,
6, 4, 21, 5, 10,
17, 12, 2, 16, 15
0.139116
(100.0 %)
1975.0074.9178.5771.43
22253.3351.6440.0066.670.5913, 1, 12, 6,
20, 7, 3, 8, 17
0.090116
(100.0 %)
976.6776.0166.6786.67
32260.7159.7671.4350.000.918, 130.016116
(100.0 %)
289.2988.6478.57100.00
42257.1456.6964.2950.000.9121, 220.015116
(100.0 %)
275.0072.8492.8657.14
52260.0058.5073.3346.670.823, 22, 1, 130.030116
(100.0 %)
463.3362.5473.3353.33
62263.3363.2566.6760.000.4118, 7, 1, 8, 12,
10, 17, 3, 11,
9, 6, 14, 13
0.089116
(100.0 %)
1370.0069.9266.6773.33
72246.4346.2950.0042.860.9122, 30.016116
(100.0 %)
267.8667.7664.2971.43
82275.0074.9178.5771.430.8218, 14, 6, 130.029116
(100.0 %)
485.7185.4292.8678.57
92271.4371.4371.4371.430.2718, 13, 12, 7,
2, 1, 6, 3, 22,
21, 20, 19, 4,
8, 15, 14
0.115116
(100.0 %)
1678.5778.2585.7171.43
Average22 61 . 30
± 8 . 71
60 . 71
± 8 . 98
65 . 24
± 12 . 40
57 . 36
± 10 . 74
0 . 64
± 0 . 30
13, 8, 6, 1,
22, 3, 14, 18
0 . 060
± 0 . 048
116
(100.0 %)
8
± 7
75 . 71
± 8 . 23
75 . 14
± 8 . 25
77 . 73
± 11 . 00
73 . 70
± 14 . 09
Table 2. Results of selection of BCI channels applying the Algorithm 2 and 2 -norm.
Table 2. Results of selection of BCI channels applying the Algorithm 2 and 2 -norm.
SubjectNumber
of Sources
Performance Measures (%)Reduction
Rate
Relevant
Electrodes
Required
Time (s)
Support
Vectors
AccGeo-MSensSpec
12267.8667.7671.4364.290.5014, 13, 18,
1, 8, 22, 19,
21, 9, 10, 4
0.06583
(71.6 %)
1175.0074.9178.5771.43
22266.6765.3253.3380.000.3612, 8, 13, 1,
17, 7, 16, 20,
21, 2, 6, 14,
9, 10
0.096110
(94.8 %)
1466.6765.3253.3380.00
32271.4371.0764.2978.570.828, 5, 13, 140.02475
(64.7 %)
475.0074.9171.4378.57
42246.4346.2950.0042.860.9122, 210.01772
(62.1 %)
278.5778.5778.5778.57
52256.6754.1673.3340.000.777, 1, 13,
3, 22
0.03078
(67.2 %)
566.6765.3280.0053.33
62270.0069.9266.6773.330.5518, 8, 1, 7,
12, 17, 3,
9, 10, 14
0.06699
(85.3 %)
1076.6776.5980.0073.33
72246.4346.2950.0042.860.9122, 30.01383
(71.6 %)
267.8667.7664.2971.43
82267.8667.7671.4364.290.8218, 14, 1, 130.02886
(74.1 %)
482.1482.0778.5785.71
92271.4371.4371.4371.430.7313, 2, 12,
22, 1, 18
0.03564
(55.2 %)
678.5778.2585.7171.43
Average22 62 . 75
± 10 . 25
62 . 22
± 10 . 40
63 . 55
± 9 . 77
61 . 96
± 15 . 98
0 . 71
± 0 . 19
1, 13, 14,
22, 18, 8
0 . 042
± 0 . 028
83
± 14
( 71 . 8 %
± 12 . 1 %)
6
± 4
74 . 13
± 5 . 73
73 . 74
± 6 . 14
74 . 50
± 9 . 99
73 . 76
± 9 . 10
Table 3. Results of selection of BCI channels applying the Algorithm 3 and 2 -norm.
Table 3. Results of selection of BCI channels applying the Algorithm 3 and 2 -norm.
SubjectNumber
of Sources
Performance Measures (%)Reduction
Rate
Relevant
Electrodes
Required
Time (s)
Support
Vectors
AccGeo-MSensSpec
12267.8667.7671.4364.290.7714, 13, 18,
9, 1
0.04178
(67.2 %)
578.5778.2585.7171.43
22263.3362.5453.3373.330.057, 1, 13, 6, 8, 9,
12, 2, 3, 14, 18,
17, 20, 10, 5, 21,
19, 15, 16, 11, 22
0.171105
(90.5 %)
2163.3362.5453.3373.33
32264.2964.2964.2964.290.918, 130.01778
(67.2 %)
282.1482.0785.7178.57
42250.0050.0050.0050.000.9122, 210.01899
(85.3 %)
271.4371.4371.4371.43
52256.6754.1673.3340.000.777, 1, 13,
3, 22
0.03978
(67.2 %)
566.6765.3280.0053.33
62266.6766.3373.3360.000.5018, 7, 8, 12,
1, 9, 17, 10,
3, 11, 14
0.090100
(86.2 %)
1173.3373.3373.3373.33
72260.7160.6164.2957.140.9122, 30.01783
(71.6 %)
264.2964.2964.2964.29
82267.8667.7664.2971.430.7718, 13, 14,
9, 1
0.04483
(71.6 %)
582.1482.0785.7178.57
92271.4371.4371.4371.430.7313, 2, 12,
22, 1, 18
0.04564
(55.2 %)
671.4371.4371.4371.43
Average22 63 . 20
± 6 . 59
62 . 76
± 6 . 91
65 . 08
± 8 . 55
61 . 32
± 11 . 00
0 . 70
± 0 . 28
13, 1, 18,
22, 9, 14, 3
0 . 054
± 0 . 049
85
± 13
( 73 . 6 %
± 11 . 5 %)
7
± 6
72 . 59
± 7 . 17
72 . 30
± 7 . 39
74 . 55
± 11 . 07
70 . 63
± 7 . 76
Table 4. Results of selection of BCI channels applying the Algorithms 1 to 3 with 1 -norm and 2 -norm.
Table 4. Results of selection of BCI channels applying the Algorithms 1 to 3 with 1 -norm and 2 -norm.
Using the 1 -NormUsing the 2 -Norm
Algorithm 1Algorithm 2Algorithm 3Algorithm 1Algorithm 2Algorithm 3
ElectrodeRelevElectrodeRelevElectrodeRelevElectrodeRelevElectrodeRelevElectrodeRelev
188.951410.22138.001310.08110.481310.74
147.46189.23147.8787.291310.45110.52
137.31137.17186.9767.22148.83188.91
86.09227.0886.8816.81228.63228.15
126.0916.4776.61226.73187.3597.19
75.98216.36126.6136.6786.90147.19
225.9276.27225.38145.79125.2236.47
105.8785.7765.32185.79215.1885.29
215.8165.5915.2575.6295.14125.03
114.40125.28195.23125.62105.1474.91
174.4094.87104.13214.2834.97103.30
64.3123.40173.99174.1574.94113.30
34.27153.40113.97202.9123.42173.30
24.16193.40153.8822.87173.3723.26
14.16113.1133.6842.8741.76213.26
52.69203.1122.68152.8751.7651.53
92.6941.7552.48192.87191.7661.53
192.6931.56212.4892.7161.57151.53
202.69101.5692.46102.71161.57161.53
151.37161.56162.4651.40201.57191.53
161.37171.56202.46161.40110.00201.53
41.3251.3241.22111.31150.0040.00
Table 5. Results of selection of MRI sequences with perceptual features and 1 -norm.
Table 5. Results of selection of MRI sequences with perceptual features and 1 -norm.
Performance
Measures
Using All
Sources
(%)
Applying
Sources
Reduction (%)
Reduction
Rate
Relevant
Sequences
Required
Time
(s)
Support
Vectors
Algorithm 1Accuracy73.3380.000.78P_ADC, P_SUB40.013117
(100.0 %)
Geo-Mean65.9179.59
Sensitivity94.1282.35
Specificity46.1576.92
Algorithm 2Accuracy80.0093.330.56P_SUB2, P_SUB4,
P_SUB3, P_SUB5
0.02197
(82.9 %)
Geo-Mean76.1093.21
Sensitivity94.1294.12
Specificity61.5492.31
Algorithm 3Accuracy80.0086.670.67P_SUB2, P_SUB5
P_SUB3
0.02097
(82.9 %)
Geo-Mean76.1085.09
Sensitivity94.1294.12
Specificity61.5476.92
Table 6. Results of selection of MRI sequences with radiomic features and 1 -norm.
Table 6. Results of selection of MRI sequences with radiomic features and 1 -norm.
Performance
Measures
Using All
Sources
(%)
Applying
Sources
Reduction (%)
Reduction
Rate
Relevant
Sequences
Required
Time
(s)
Support
Vectors
Algorithm 1Accuracy82.7686.210.33R_T2, R_T1
R_DWI, R_SUB3
R_SUB5, R_SUB4
0.073117
(100.0 %)
Geo-Mean84.0287.45
Sensitivity70.5976.47
Specificity100.00100.00
Algorithm 2Accuracy79.3189.660.33R_T1, R_DWI
R_SUB3, R_T2
R_SUB4, R_SUB5
0.06381
(69.2 %)
Geo-Mean80.4490.75
Sensitivity64.7182.35
Specificity100.00100.00
Algorithm 3Accuracy86.2189.660.22R_T1, R_DWI
R_SUB3, R_T2
R_SUB5, R_SUB4
R_SUB2
0.11392
(78.6 %)
Geo-Mean87.4590.75
Sensitivity76.4782.35
Specificity100.00100.00
Table 7. Results of selection of MRI sequences with perceptual and radiomic features with 1 -norm.
Table 7. Results of selection of MRI sequences with perceptual and radiomic features with 1 -norm.
Performance
Measures
Using All
Sources
(%)
Applying
Sources
Reduction (%)
Reduction
Rate
Relevant
Sequences
Required
Time
(s)
Support
Vectors
Algorithm 1Accuracy76.6786.670.72P_SUB2, P_SUB5
P_SUB4, P_SUB3
R_SUB3
0.033117
(100.0 %)
Geo-Mean75.5185.09
Sensitivity82.3594.12
Specificity69.2376.92
Algorithm 2Accuracy76.6790.000.67P_SUB2, P_SUB5
P_SUB4, P_SUB3
R_SUB3, R_SUB5
0.03997
(82.9 %)
Geo-Mean75.5189.24
Sensitivity82.3594.12
Specificity69.2384.62
Algorithm 3Accuracy76.6790.000.67P_SUB2, P_SUB5
P_SUB4, P_SUB3
R_SUB3, R_SUB5
0.05197
(82.9 %)
Geo-Mean75.5189.24
Sensitivity82.3594.12
Specificity69.2384.62
Table 8. Comparison of results applying feature selection by Fisher Score technique over the BCI data before and after the source reduction process.
Table 8. Comparison of results applying feature selection by Fisher Score technique over the BCI data before and after the source reduction process.
SubjectFeature Selection over All Information SourcesFeature Selection After Reduce Information Sources
AccuracyMinimal
Required
Features
Reduction
Rate
(by Sources)
AccuracyMinimal
Required
Features
Reduction
Rate
(by Sources)
With
All Features
Reducing
Features
With
All Features
Reducing
Features
Algorithm 1
164.2985.712800.0075.0085.713620.14
253.3376.676670.0076.6773.33440.59
360.7192.86200.6889.2992.86220.91
457.1475.002070.0075.0089.29230.91
560.0070.001300.0063.3376.67690.82
663.3380.0030.9570.0080.0030.95
746.4367.86250.5067.8664.2910.95
875.0082.14520.5085.7185.711190.82
971.4378.577670.0078.5775.00830.27
Average 61 . 30
± 8 . 71
78 . 76
± 7 . 70
239 . 00
± 287 . 44
0 . 29
± 0 . 37
75 . 71
± 8 . 23
80 . 32
± 8 . 97
80 . 67
± 112 . 50
0 . 71
± 0 . 31
Algorithm 2
167.8685.712430.0075.0085.713000.50
266.6773.331660.0066.6770.003890.36
371.4396.4320.9175.0096.4320.91
446.4375.002010.0078.5785.71200.91
556.6770.002810.0066.6773.331630.77
670.0083.334200.0076.6780.00880.55
746.4367.86350.4167.8664.29320.91
867.8682.14620.3682.1485.71250.86
971.4378.574450.0078.5775.00260.73
Average 62 . 75
± 10 . 25
79 . 15
± 8 . 88
206 . 11
± 159 . 35
0 . 19
± 0 . 32
74 . 13
± 5 . 73
79 . 58
± 9 . 89
116 . 11
± 140 . 10
0 . 72
± 0 . 21
Algorithm 3
167.8685.712440.0078.5778.57970.77
263.3370.001670.0063.3370.001550.05
364.2989.2920.9182.1496.43580.91
450.0067.8620.9171.4382.1470.91
556.6773.332810.0066.6773.331630.77
666.6783.334200.0073.3380.00100.82
760.7171.436550.0064.2964.29320.91
867.8682.1450.9582.1482.1450.95
971.4378.574450.0071.4378.571840.73
Average 63 . 20
± 6 . 59
77 . 96
± 7 . 62
246 . 78
± 229 . 61
0 . 31
± 0 . 46
72 . 59
± 7 . 17
78 . 39
± 9 . 02
79 . 00
± 72 . 67
0 . 76
± 0 . 28
Table 9. Comparison of results applying feature selection by Fisher Score technique over the MRI data before and after the source reduction process.
Table 9. Comparison of results applying feature selection by Fisher Score technique over the MRI data before and after the source reduction process.
AlgorithmFeature Selection Over All Information SourcesFeature Selection after Reduce Information Sources
AccuracyMinimal
Required
Features
Reduction
Rate
(by Sources)
AccuracyMinimal
Required
Features
Reduction
Rate
(by Sources)
With
All Features
Reducing
Features
With
All Features
Reducing
Features
Perceptual Features
173.3383.33260.1180.0080.0070.78
280.0090.00480.0093.3393.33260.56
380.0086.67230.2286.6790.00150.67
Radiomic Features
182.7696.551790.0086.2193.101710.33
279.3196.551730.0089.6693.101740.33
386.2196.551790.0089.6693.102690.22
Perceptual and Radiomic Features
176.6790.003020.0086.6790.00430.72
276.6790.005260.0090.0090.001240.67
376.6790.005260.0090.0090.001250.67

Share and Cite

MDPI and ACS Style

Areiza-Laverde, H.J.; Castro-Ospina, A.E.; Hernández, M.L.; Díaz, G.M. A Novel Method for Objective Selection of Information Sources Using Multi-Kernel SVM and Local Scaling. Sensors 2020, 20, 3919. https://doi.org/10.3390/s20143919

AMA Style

Areiza-Laverde HJ, Castro-Ospina AE, Hernández ML, Díaz GM. A Novel Method for Objective Selection of Information Sources Using Multi-Kernel SVM and Local Scaling. Sensors. 2020; 20(14):3919. https://doi.org/10.3390/s20143919

Chicago/Turabian Style

Areiza-Laverde, Henry Jhoán, Andrés Eduardo Castro-Ospina, María Liliana Hernández, and Gloria M. Díaz. 2020. "A Novel Method for Objective Selection of Information Sources Using Multi-Kernel SVM and Local Scaling" Sensors 20, no. 14: 3919. https://doi.org/10.3390/s20143919

APA Style

Areiza-Laverde, H. J., Castro-Ospina, A. E., Hernández, M. L., & Díaz, G. M. (2020). A Novel Method for Objective Selection of Information Sources Using Multi-Kernel SVM and Local Scaling. Sensors, 20(14), 3919. https://doi.org/10.3390/s20143919

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop