Next Article in Journal
Characterization of Multipath Effects in Indoor Positioning Systems by AoA and PoA Based on Optical Signals
Previous Article in Journal
A Sensor Array Realized by a Single Flexible TiO2/POMs Film to Contactless Detection of Triacetone Triperoxide
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Extraction and Classification of Citrus Juice by Using an Enhanced L-KSVD on Data Obtained from Electronic Nose

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
College of Electronic and Information Engineering, Southwest University, Chongqing 400715, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(4), 916; https://doi.org/10.3390/s19040916
Submission received: 23 January 2019 / Revised: 16 February 2019 / Accepted: 18 February 2019 / Published: 21 February 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Aroma plays a significant role in the quality of citrus fruits and processed products. The detection and analysis of citrus volatiles can be measured by an electronic nose (E-nose); in this paper, an E-nose is employed to classify the juice which is stored for different days. Feature extraction and classification are two important requirements for an E-nose. During the training process, a classifier can optimize its own parameters to achieve a better classification accuracy but cannot decide its input data which is treated by feature extraction methods, so the classification result is not always ideal. Label consistent KSVD (L-KSVD) is a novel technique which can extract the feature and classify the data at the same time, and such an operation can improve the classification accuracy. We propose an enhanced L-KSVD called E-LCKSVD for E-nose in this paper. During E-LCKSVD, we introduce a kernel function to the traditional L-KSVD and present a new initialization technique of its dictionary; finally, the weighted coefficients of different parts of its object function is studied, and enhanced quantum-behaved particle swarm optimization (EQPSO) is employed to optimize these coefficients. During the experimental section, we firstly find the classification accuracy of KSVD, and L-KSVD is improved with the help of the kernel function; this can prove that their ability of dealing nonlinear data is improved. Then, we compare the results of different dictionary initialization techniques and prove our proposed method is better. Finally, we find the optimal value of the weighted coefficients of the object function of E-LCKSVD that can make E-nose reach a better performance.

1. Introduction

Citrus juice is famous for its rich nutrition, delicious taste and aroma-make, and aroma is a significant factor that affects the quality of citrus fruits and processed products. It is important for us to study the aroma of citrus processed products due to the fact that characteristics of aroma can be a standard for testing in the field of food. When tasting a citrus, we can use multiple senses, such as aroma and taste, to feel the stimulus. Also, flavor is a crucial factor in determining what citrus quality and nutrition value is. At present, the main method of testing the citrus quality is sensory analysis and precision instrumental.
Sensory analysis is used in traditional processing extensively. Although this test method is extremely consumer-friendly, people are too emotional to have a similar standard of the same juice. At the same time, various degrees of fatigue problems will affect experimenters’ sensory analysis, which may cause the experimenters to come to inaccurate or even wrong conclusions and make the results subjective and unscientific.
Methods for precision instrumental analysis include gas chromatography (GC), gas chromatography-olfactometry (GC-O), gas chromatography-mass spectrometry (GC-MS), etc. [1,2]. The analysis of precision instruments has become the main analytical test method in the scientific investigation because of its high repeatability and the results of analysis being objective and scientific. However, there are still some drawbacks using an analysis of precision instruments method. First, it cannot be applied well to the production and living, which is restrictive to a full research platform with low cost. Secondly, the preprocess steps of a precision instrument analysis are complicated, and the instrument is inconvenient to operate, which means it is difficult for operators to meet strict skill requirements. Last but not least, consumers pay more attention to the overall feeling rather than to the specific ingredients of the juice when they smell the juice.
The evaluation of the odor for juice is important, so an efficient and low-threshold technique of the actual production is in demand, which can focus especially on the overall feeling of the product and analyze the attributes of the product scientifically and objectively.
An electronic nose (E-nose) is a device which is designed like the biological olfactory device to analyze the gas/odor [3,4]. It combines an array of gas sensors and series-related intelligent algorithms. Through the gas sensor array, an E-nose can gain response curves of the sensors to the target gas/odor, and then, intelligent algorithms are used to deal with the response to identify the gas/odor, which includes a single or complex component(s) [5]. Comparing the sensory analysis with the precision instrumental analysis, an E-nose is considerable, low cost and fast detection. It has been applied to various fields such as wound detection [6], disease detection [7], food engineering [8,9,10], environmental control [11], explosive detection [12], agriculture [13,14] and so on.
Reference [15] validated an E-nose to identify “La France” pears of different maturity and compared them by using chemical analysis methods (GC and GC-MS). The result shows that the judgment result of an E-nose is very consistent. An E-nose was used to study the quality of citrus and apple by Corrado Di Natale et al., and they proved that an E-nose has enough sensitivity and resolution to distinguish among the various classes and to correctly predict the number of defects (for apples) and storage days (for oranges). Moreover, References [16,17,18] proved that E-nose can be used to identify the maturity and quality of various fruits.
Feature extraction is the first step of the sensor signal processing and affects the accuracy of the pattern recognition a lot [19]. To get the utmost out of the experiment data, classifiers are used for learning a classification function or constructing a classification model based on the original data. Feature extraction and pattern recognition play important roles in helping an E-nose get a high recognition rate, and they are two independent steps traditionally.
Extracting more features from an E-nose response can help improve the classification accuracy. In the traditional way, we would like to get the maximum value of the steady-state response of sensors to construct the original feature matrix and use this matrix to classify, but its accuracy is not ideal. To get a better performance, some classical algorithms are used for reprocessing the original feature matrix. Principal component analysis (PCA) is a method to find several comprehensive indicators to represent many original features, so that these comprehensive indicators reflect the original variables as much as possible, and they are not related to each other [20]. Independent component analysis (ICA) [21] is a statistical method used to convert observed multidimensional vectors into statistically independent components for eliminating the redundancy of the original data. These algorithms help the classifier improve accuracy. However, PCA and ICA are both linear methods of which the performance is not very ideal when used to deal with the nonlinear data. The kernel PCA (KPCA) is a nonlinear method which finds a computationally tractable solution through a simple kernel function that intrinsically constructs a nonlinear mapping from the input space to the high-dimensional space and then performs a nonlinear PCA in the high-dimensional space [22].
When the feature matrix is obtained, it will be put into the classifier. The most common classifiers used in an E-nose is the linear discriminant analysis (LDA) [23], radial basis function neural network (RBFNN) [24] and support vector machine (SVM) [25]. LDA is a linear classify, and its target during the training process is to shorten the distance between the samples from the same class and largen the distance of samples from the different classes. With the help of the kernel function, LDA can classify the nonlinear data. RBFNN using a radial basis function as its nonlinear mapping function is not easy to fall into the local optimum, but the time it spends during the training process is a little long. SVM is a widely used classifier in an E-nose. Its training target is not only the highest classification accuracy of the training data set but also the highest classification accuracy of the test data set, and it has a good generalization ability.
We cannot help but think that if we unify the process of feature extraction and the classifier, as a joint indicator, can we get a better recognition rate. Inspired by this, we find the label consistent KSVD (L-KSVD), which combines the feature extraction with the classification. KSVD is a dictionary learning algorithm for creating a dictionary for spare representations via a singular value decomposition approach. KSVD is a generalization of the k-means clustering method, and it works by iteratively alternating between sparse coding the input data based on the current dictionary and updating the atoms in the dictionary to better fit the data [26]. The innovation of this paper is pointed out as follows:
(a)
The traditional L-KSVD cannot handle the problem of nonlinear data very well, and the kernel function is adopted in this paper to help L-KSVD deal with the nonlinear data obtained by the E-nose.
(b)
Choosing a proper dictionary is the first and most important step of L-KSVD, and a novel dictionary initialization method is proposed according to the data characteristics of the E-nose. With the help of the Enhance Quantum-behaved Particle Swarm Optimization (EQPSO), this method generates random numbers in binary and uses the recognition rate as a fitness function to decide which sensor response will be used to initialize the dictionary.
(c)
The weighted coefficients of the objective function of L-KSVD have a bigger impact on the classification accuracy, so these coefficients are standardized and then optimized with the help of EQPSO in this paper.
In the rest of this paper, we present the experiment of this paper in Section 2 and introduce the proposed L-KSVD in Section 3. The results and discussion will be presented in Section 4. Finally, the conclusions are drawn in Section 5.

2. Materials and Methods

In this project, we use the cold-pressed technology to process the Valencia oranges with the same maturity to obtain orange juice: 50 kg oranges are sued in this project, and the shape and size of each orange is almost same. These oranges come from 10 trees of similar age and growth in the same orchard and are picked at the same position of each tree (about 5 kg oranges from one tree). During the process of obtaining the experimental juice, filtration, sterilization and canning are implemented in turn. Then, the juice is stored in a storage tank (its volume is 30 L), and nitrogen is filled in the top of the storage tank to be isolated from air, and the air pressure of the tank is 202.6 KPa. On the bottom of the tank, there is a tap which is used for the sampling of the orange juice. Inside the tank, there is a blender, and every time before the sampling, we turn it on for 1 min to stir the orange juice thoroughly in case the orange juice might delaminate or precipitate, and the blender spins on its axis once every 1 s. The sampling is proceeded every 15 days, and more information of the above description can be found in Figure 1. After each sampling, the orange juice is put into 4 sterilized and identical glass vessels, and each of them contains 500 mL of orange juice.
The analysis of the orange juice aroma components is shown in Table 1, and the schematic diagram of the experimental E-nose system is shown in Figure 2.
During each sampling experiment of the E-nose, we set the temperature and humidity of the chamber of the E-nose at 25 °C and 40%, respectively. The rules used when we design and construct the sensor array are (a) the sensor array can respond to all classes of the juice odor; (b) each sensor has its own interesting odor and can also respond to other odors of juice; and (c) a high-performance cost ratio and an easy purchase are requirements. Table 2 lists the sensors selected by us and their sensitive characteristics. The gas sensor array was located in the Teflon chamber with a volume of 0.24 L, and the flow rate can be controlled by a flowmeter and a micro pump with a set value of 0.08 L/min. The practical E-nose system designed by us is shown in Figure 3.
In the rubber stopper, there are two holes with two thin Teflon tubes inserted: one Teflon tube fixed as close as possible to Valencia orange juice. The output gas from the tube containing VOCs of the Valencia orange juice flows out of the bottle and then flows into the chamber through a Teflon tube. As for each sampling experiment, the following three stages should be performed:
Step (a)
expose all sensors to clean air for 5 min to obtain the baseline.
Step (b)
introduce the target gas into the chamber for 7 min.
Step (c)
exposed the sensor array to clean air for 5 min again to clean the sensors and restore the baseline.
In each experiment, each cup of orange juice is tested by an E-nose 6 times so that 24 sample data are recorded in total. After 4 sampling experiments, the data set of this project, which contains 96 sample data, is obtained.
The sensors used by the E-nose of this paper are all metal oxide sensor (MOS). With a MOS, when its resistance wire contacts different gases, its resistivity will change, which will lead to the change of its resistance value. The E-nose hardware circuit designed in this paper can transform the change of resistance value to the change of voltage value, and then, we record this voltage value. Therefore, the voltage value is taken as the response value of the sensor in this paper. For this voltage value, we first use analog circuits to filter and amplify it, and then, the response of the sensor array is sampled by a 14-bit data acquisition system (DAS) which is bought from the market, and the output signal of DAS is sent to the computer by a USB data line. Figure 4 illustrates the response of the sensors when Valencia orange juice odor is introduced into the chamber. We can see that each response curve rises obviously from the fifth minute when the target gas begins to pass over the sensor array and recovers to the baseline after the twelfth minute when clean air is conveyed to wash the sensors.
Then the maximum value of the steady-state response of sensors is extracted to create the feature matrix of the E-nose. There are 96 samples in this matrix, and the dimension of each sample is 15. We randomly select 2/3 samples of each gas to establish the training data set, and the rest are used as the test data set. This feature matrix is called the original feature matrix, and it is the input of the proposed KSVD algorithm.

3. Methodology

3.1. KSVD and L-KSVD

KSVD is a generalization of the k-means clustering method, and it works by iteratively alternating between sparse coding the input data based on the current dictionary and updating the atoms in the dictionary to better fit the data. KSVD can be found widely used in applications such as image processing, audio processing, biology and document analysis. KSVD learns a shared dictionary, which optimizes the following objective function:
arg min D , Z Y D X F 2 s . t . i ,   x i 0 T
where Y = [ y 1 , y 2 , , y N ] R n × N are N input signals and each is in the n dimension; D = [ d 1 , d 2 , d K ] R n × K ( K >> n , making D over-complete) is a dictionary with atoms; X = [ x 1 , x 2 , , x N ] R K × N are N sparse codes of input signals Y ; and T is a constant, which controls the number of nonzero elements in x i less than T .
Solving the minimization of Equation (1) by a two-step iterative algorithm: Firstly, the dictionary is fixed, and the sparse coefficients X can be found. This is the problem of sparse coding, which can be solved by an orthogonal matching pursuit (OMP) [27,28]. Secondly, while fixing all other atoms in D , the sparse coefficient matrix X is fixed and dictionary D is updated one atom at the same time.
For each atom, d k and the corresponding k t h row of coefficient matrix X denoted by x T k define the group of samples that use this atom as ω k = { i | 1 i N , x T k ( i ) 0 } . The error matrix computes E k = Y i k d i × x T i , restricts E k by choosing only the columns in ω k and obtains E k R . Then, the following problem is solved:
arg min d k , z T k E k R d k × x T k F 2
where a singular value decomposition (SVD) is performed and E k R = U Δ V T .
A label consistent KSVD (L-KSVD) algorithm to learn a discriminative dictionary for sparse coding was presented by Zhuojin Jiang et al. This algorithm learns a single over-complete dictionary and an optimal linear classifier jointly. It yields dictionaries so that feature points with the same class labels will have similar sparse codes.
In order to use class labels of training data, the associating label information with each dictionary item (columns of the dictionary matrix) enforces discriminability in sparse codes during the dictionary learning process [29].
The aim is to leverage the supervised information of input signals to learn a reconstructive, discriminative dictionary and to include the classification error as a term in the objective function for dictionary learning; each dictionary item is chosen to represent a subset of the training signals ideally from a single class: in that case, each dictionary item d k can be associated with a particular label. Thus, there is an explicit correspondence between the dictionary items and the labels. Then, L-KSVD focuses on the effects of adding a label-consistent regularization term, subsequently, to learn both more balanced reconstructive and discriminative power, making the objective function with a joint classification error and label-consistent regularization term.
The performance of the linear classifier must be based on the discriminability of x (the input sparse codes). To accept the discriminative sparse codes that have the learned D , an objective function of dictionary construction is defined as follows:
< D , W , A , X > = arg min D , W , A , X Y D X 2 2 + α Q A X 2 2 + β H W X 2 2   s . t . i , x i 0 T
where α dominates the reconstruction and label-consistent regularization, H W X 2 2 represents the classification error, and W is the classifier parameters to make the classification dictionary optimal. L-KSVD makes the linear predictive classifier f ( x ; W ) = W x , with Q = [ q i 1 q i K ] R K × N as the input signal Y of the discriminative sparse codes of the classification, denote the discriminative sparse code q i = [ q i 1 q i K ] t = [ 0 1 , 1 , 0 ] t R K consistent with y i (input signal), when q i are nonzero values on these indices where y i and d k (dictionary item) use a communion label. For instance, It can be assumed D = [ d 1 d 6 ] and Y = [ y 1 y 6 ] , where y 1 , y 2 , d 1 and d 2 are from class 1; y 3 , y 4 , d 3 and d 4 are from class 2; and y 5 , y 6 , d 5 and d 6 are from class 3. Q could be defined as
Q [ 110000 110000 001100 000011 000011 ]
A is a linear transformation matrix, and L-KSVD identifies g = ( x ; A ) = A x ; in the sparse feature space R K , the original sparse code x is transformed to the most discriminative one. Q A X 2 2 , on behalf of the discriminative sparse-code error, performs X approximate discriminant sparse codes. It forces the signals from the same class to have very similar sparse representations and uses simple linear classifiers to achieve a good classification performance. H = [ h 1 h N ] R m × N is a class label of Y. h i = [ 0 , 0 1 0 , 0 ] t R m is a label vector the goes to the input signal y i , and the nonzero position indicates the location of the classes. α and β are scalars for controlling the relative contribution of the corresponding items.
Dictionaries learned with the method will adapt to the structure of the training data (resulting in a good representation of the strict sparse constraints on each member of the set) and will make discriminatory sparse codes X disregard the dictionary size. These sparse codes can be directly used by classifiers, such as in Reference [30]. The discriminating characteristics of sparse codes x are very important for the performance of linear classifiers.
The brief introduction of KSVD and L-KSVD are shown above. When we apply L-KSVD to solve the problem of data processing of an E-nose, we find that several problems need to be solved:
The distribution of data gained by E-nose data is nonlinear, so the effect of KSVD/L-KSVD is not ideal when they are used to process the data directly; since the sensor of an E-nose is cross-responsive, the data is redundant. When the dictionary of KSVD/L-KSVD is initialized, different sensor responses will be selected and the corresponding recognition rate will be different. The weighted coefficients of three parts of the L-KSVD objective function will determine the influence of each part on the final result.
During Section 3.2, Section 3.3 and Section 3.4, we will solve the above three problems, respectively. The L-KSVD enhanced by techniques from Section 3.2, Section 3.3 and Section 3.4 is called E-LCKSVD.

3.2. Kernel Function

L-KSVD for sparse coding has contributed a lot, which lies in explicitly integrating the discriminative sparse codes and a single predictive linear classifier into the objective function for dictionary learning. However, this algorithm cannot handle the problems of the nonlinear data very well. As nonlinear dynamical systems which are difficult to solve in mathematics and science, based on pattern recognition theory, the low-dimensional space linear inseparable model through nonlinear mapping to a high-dimensional feature space may be linearly separable, but if we use this technique in a high-dimensional space classification or regression directly, there is a big obstacle named dimension disaster which will exist in high-dimensional feature space operations.
It has been proved that the kernel function can be used to solve this problem effectively. A kernel is a nonnegative real-valued integrable function; it is desirable to define the function for most applications to satisfy two additional requirements: normalization and symmetry. As we know, several types of kernel functions are commonly used in many fields, especially in a support vector machine (SVM). SVM maps the sample space to a feature space of high or even infinite dimensions through nonlinear mapping so that the nonlinear separable problems in the original sample space are transformed into linearly separable problems in the feature space. As for the problems of classification and regression, it is likely that the sample set cannot be processed linearly in the lowdimensional sample space, but the linear partition (or regression) can be achieved through a linear hyper plane in the high-dimensional feature space. As the linear learning machine is established in the high-dimensional feature space, it does not increase the complexity of the calculations and avoids the dimensional disaster to some extent compared with the linear model. All of this is due to the kernel function expansion and computational theory.
The kernel function can be combined with different algorithms to form a variety of different methods based on kernel function technology, and the design of these two parts can be carried out separately. The combination of L-KSVD and the kernel function can solve the nonlinear problem of the E-nose.
In this paper, firstly, we use the RBF kernel to map the data of an E-nose and then run KSVD/L-KSVD, namely, in the high-dimensional space. The expression of RBF kernel is
k ( x i , x j ) = exp ( x i x j 2 σ 2 )
where σ is the scale factor which determines the distribution of the data mapped to the high-dimensional space, so it is a very important parameter. In this paper, an optimization algorithm named EQPSO is used to set its value.

3.3. Dictionary

Choosing a proper dictionary is the first and most important step of the sparse representation based on classifications with encouraging results [31]. Especially, dictionaries learned from training data obtain researchers’ attentions because the learned dictionaries usually lead to a better representation and achieve much success in classification.
The goal of dictionary learning is to learn an over-complete dictionary matrix D R n × K in which K contains signal-atoms (in this notation, columns of D ). A signal vector y R n can be represented, sparsely, as a linear combination of these atoms; to represent y , the representation vector x should satisfy the exact condition y = D x , or the approximate condition y D x , made precise by requiring that y D x p ε for some small value ε and some L p norm. The vector x R K contains the representation coefficients of the signal y . Typically, the norm p is selected as L 1 , L 2 or L .
If n < K and D is a full-rank matrix, there are an infinite number of solutions for the representation problem. Then, constraints should be set on the solution. Also, to ensure sparsity, the solution with the fewest number of nonzero coefficients is preferred, which means the sparsity representation is the solution of either ( P 0 ) min x x 0 subject to y = D x or ( P 0 , ε ) min x x 0 subject to y D x 2 ε , where the x 0 counts the nonzero entries in the vector x .
Minimizing the reconstruction error and satisfying the sparsity constraints to achieve the construction of D : Although the random dictionary initialization has an outstanding effect on images of compressing and restorations, it is not good enough to identify gas in an E-nose due to the fact that the cross-responsiveness of the sensor array, that is, each sensor, will respond to the same gas, which means that the response of the sensor is overlapping and redundant.
To select the best combination of sensors, we use random binary number to filter proper dictionary initialized atoms: 1 represents this sensor is selected, and 0 represents not. Remove the information which is redundant, and filter the characteristic representative data in this way. The binary parameter is provided by EQPSO (shown in Section 3.5), and the way to generate the random binary number can be found in Section 3.5.

3.4. Weighted Coefficients

At first, we use a linear predictive classifier f ( x ; W )   =   W x . An objective function for learning a dictionary D having both reconstructive and discriminative power can be defined as follows:
< D , W , A , X > =   arg   min D , W , A , X Y D X 2 2 + α Q A X 2 2   + β H W X 2 2   s . t . i ,   x i o T ,
In this paper, Equation (6) is changed by us to the following:
< D , W , A , X > =   arg   min D , W , A , X α Y D X 2 2 + β Q A X 2 2   + γ H W X 2 2   s . t . i ,   x i o T
where α + β + γ = 1 and α, β, γ ε (0,1). We define this process as normalization processing. In Equation (7), the value of α, β and γ changes the proportion of each part in the whole expression; it is a little bit more intuitive to see which part is going to have a bigger impact on the outcome result of the recognition rate. We can see the classification accuracy of E-LCKSVD is very different when the values of α and β are different from Table 3, so the values of α, β and γ have great influences on the classification accuracy. Some optimization technique must be employed to set their value.

3.5. EQPSO and the Optimization Problem of the Proposed E-LCKSVD

3.5.1. PSO, QPSO and EQPSO

Particle swarm optimization (PSO) is an evolutionary computation algorithm based on the swarm intelligence theory. This swarm intelligence algorithm for continuous searching space problems is widely used for its simple programming and fast convergence speed and is used in Reference [32]. Among them,
P i d ( t ) = c 1 r 1 d ( t ) P i d ( t ) + c 2 r 2 d ( t ) P g d ( t ) c 1 r 1 d ( t ) + c 2 r 2 d ( t ) ,   1 d D ,
and
ϕ d ( t ) = c 1 r 1 d c 1 r 1 d ( t ) + c 2 r 2 d ( t ) ,
where t is the current iteration number of the algorithm, r 1 d ( t ) and r 2 d ( t ) are random numbers in [0,1]; Pi is the current optimal position of the particle; and Pg is the global optimal position of the group. However, in a classic PSO, the particle search process is realized in the form of orbit and the particle’s flight speed is limited. Therefore, in the search process, the particle search space is limited to a limited search space, which cannot cover the whole feasible search space. A general PSO cannot guarantee convergence to the global optimal solution with probability 1, which is the deficiency of PSO.
In order to solve this shortcoming of PSO, quantum-behaved PSO (QPSO) has been proposed [33]. QPSO combine PSO with quantum mechanics. QPSO has great advantages in terms of search ability, convergence speed, accuracy and solving robustness. Compared with the other algorithm, one of the biggest characteristics of QPSO is the simple calculation and few control parameters. It is superior to the general PSO not only in search ability but also in its accuracy, and QPSO can guarantee the global convergence with probability 1.
The wave function ϕ ( X , t ) is used to determine the state of each particle and the definition of the average optimal position which is the center of the optimal position of all particles. We can get an updated Equation (10) for the position of the particle:
x i j ( t + 1 ) = p i j ( t ) ± α | M i j ( t ) x i j ( t ) | ln [ 1 u i j ( t ) ] u i j ( t ) ~ U ( 0 , 1 )
where α is called the compression expansion factor to regulate the rate of convergence of particles. However, when the number of iterations is not infinite, QPSO cannot guarantee to find the global optimal value. In practice, the number of iterations is always limited.
Enhanced QPSO (EQPSO) [34,35] is an improved QPSO algorithm, which can ensure to find the value closest to the optimal value in the case of limited iteration times. It has a better ability of making particles multifarious at the early iteration’s stage and performing in local searching ability at the later stage of iteration. Meanwhile, the convergence speed of EQPSO is faster than other considered PSOs, such as PSO and QPSO. Therefore, EQPSO is selected as the optimization algorithm in this paper.

3.5.2. Optimization Problem of the Proposed E-LCKSVD

The first step is to generate α and β ( γ = 1 α β ); these two numbers are real, and the value range is 0 to 1. Then, from 15 binary numbers which are used to determine the response of the sensor which will be used to initialize the dictionary, one of the responses of the sensor will be selected, only when the corresponding binary number is 1. The classification accuracy of the E-nose is the fitness function; a set of parameters can be chosen only if its corresponding fitness function reaches the maximum. The detailed information of the parameters that needed optimization is shown in Table 4.

4. Results and Discussion

As mentioned before, there are 96 samples of 4 classes (24 samples in each class), and we randomly select 16 samples from each class to build the training data set (16 × 4 = 64 samples), and the rest of the samples are used to build the test data set (8 samples in each class). Figure 5 gives the work flow of E-LCKSVD. In order to verify that the kernel function can help L-KSVD deal with the data of the E-nose more effectively, we have entered several sets of comparative experiments. To determine whether to add a kernel function to the map data to a variable controlled by a high-dimensional space, we compare the processed and unprocessed data into KSVD and L-KSVD respectively. Since KSVD does not have the ability of pattern recognition by itself, we input the sparse matrix obtained by it into the extreme learning machine (ELM) [36] for pattern recognition. Each set of procedures is repeated 10 times, and the one with the highest recognition rate was taken as the final result, which is shown in Table 5.
It can be seen obviously that the addition of the kernel function improves the discriminating precision significantly. As shown in Figure 6, KSVD combined with ELM is more effective in processing data, as does L-KSVD.
Next, we compare the dictionary initialization method proposed in this paper with the previous dictionary initialization methods. We found that the initialization of the dictionary can affect the classification accuracy.
As shown in Figure 7, a too small dictionary leads to the loss of information, resulting in a low recognition rate. Meanwhile, the process of training the dictionary will be complex when the dictionary is too large, and a large dictionary does not mean a high classification accuracy because the data of the sensor array is redundant. In addition, we found that even though the dimensions were the same, the result is different when different sensors are selected to initialize the dictionary, just like shown in Table 6. The classification accuracy of different dictionary initialization methods are shown in Table 7. Finally, the size of the dictionary is 14. Compared with the random initialization method, it is significant to select the initialization dictionary purposefully. The training set recognition rate of E-LCKSVD using the optimized dictionary initialization method is as high as 98.4%, and the test set recognition rate reaches 96.9%.
Figure 8 is the comparison of the recognition rate using several different weighted coefficients and the corresponding weight when the highest recognition rate is reached. It can be seen visually that the weights have a great influence on the effect of the whole system. Therefore, finding the appropriate weight value is an important step. Different weights combinations have the problem of the local optimal solution; we use EQPSO to avoid this, and then, we find the global optimization point: α = 0.7661 , β = 0.0351 .
Then, we compare the proposed method with the results of the existing common feature extraction and classifier in an E-nose. We use the maximum steady-state response of the original response curve, PCA (as the representative of the linear feature extraction algorithm) and KPCA (as the representative of the nonlinear feature extraction algorithm) to construct the feature matrix, and select SVM, RBFNN and K-LDA (the kernel LDA) as classifiers. The data processing results are shown in Table 8.
From Table 8, we can find that the best result is obtained when KPCA is used to extract the feature and SVM is the classifier, but this result (96.6/90.6%) is still a little worse than E-LCKSVD (98.4/96.9%). The advantages and drawbacks of all techniques used in this paper is shown in Table 9.

5. Conclusions

In this paper, an E-nose is used to classify the orange juice of different storage days. An intelligent algorithm is important for the E-nose. Feature extraction and classification are two key steps, and the existing techniques generally study the two links separately, but the classifier can only adjust itself and cannot train the feature extraction in the process of training, so the result is not satisfactory. In this paper, an enhanced L-KSVD technique, E-LCKSVD, is proposed. This technique can combine feature extraction with pattern recognition and adjust and optimize the feature extraction and classification according to the training results. Aiming at the traditional L-KSVD algorithm, combined with the characteristics of the data in this paper, we have made some improvements to the algorithm: Firstly, the nonlinear response of the E-nose is mapped to linear data by using the kernel function. Then, a dictionary initialization method with the help of EQPSO is proposed to initialize the dictionary of KSVD/L-KSVD, and a KSVD/L-KSVD dictionary initialization method suitable for the characteristics of E-nose data is obtained. Finally, the weighted coefficients of different parts of the objective function of L-KSVD are standardized, the influence of weighted coefficients on the recognition rate is studied, and an optimized setting method of weighted coefficients based on EQPSO is proposed. The experiments of using E-LCKSVD to get higher recognition accuracy results are very satisfactory. In short, we conclude from a series of results that E-LCKSVD is an ideal solution for distinguishing gas data from an E-nose.

Author Contributions

W.C. proposed the algorithm. C.L. was responsible for data analysis and the discussion of the results. P.J. was in charge of the project management and involved in discussions and the experimental analysis.

Funding

This research was funded by National Natural Science Foundation of China (Grant No. 61601376), Fundamental Science and Advanced Technology Research Foundation of Chongqing (Grant No. cstc2018jcyjA0867) and Chongqing Postdoctoral Science Foundation Special Funded (Grant No. Xm2017039).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Högnadóttir, A.; Rouseff, R. Identification of aroma active compounds in orange essence oil using gas chromatography-olfactometry and gas chromatography-mass spectrometry. J. Chromatogr. A 2003, 998, 201–211. [Google Scholar] [CrossRef]
  2. Biniecka, M.; Caroli, S. Analytical methods for the quantification of volatile aromatic compounds. Trends Anal. Chem. 2011, 30, 1756–1770. [Google Scholar] [CrossRef]
  3. Pearce, T. Computational parallels between the biological olfactory pathway and its analogue ‘The Electronic Nose’: Part II. Sensor-based machine olfaction. Biosystems 1997, 41, 69–90. [Google Scholar] [CrossRef]
  4. Bahraminejad, B.; Basri, S.; Isa, M.; Hambali, Z. Application of a sensor array based on capillary-attached conductive gas sensors for odor identification. Meas. Sci. Technol. 2010, 21, 085204. [Google Scholar] [CrossRef]
  5. Jia, P.; Tian, F.; He, Q.; Fan, S.; Liu, J.; Yang, S.X. Feature extraction of wound infection data for electronic nose based on a novel weighted KPCA. Sens. Actuators B Chem. 2014, 201, 555–566. [Google Scholar] [CrossRef]
  6. Markom, M.A.; MdShakaff, A.Y.; Adom, A.H.; Ahmad, M.N.; Hidayat, W.; Abdullah, A.H.; Fikri, N.A. Intelligent electronic nose system for basal stem rot disease detection. Comput. Electron. Agric. 2009, 66, 140–146. [Google Scholar] [CrossRef]
  7. Gobbi, E.; Falasconi, M.; Torelli, E.; Sberveglier, G. Electronic nose predicts high and low fumonisin contamination in maize culture. Food Res. Int. 2011, 44, 992–999. [Google Scholar] [CrossRef]
  8. Loutfi, A.; Coradeschi, S.; Mani, G.; Shankar, P.; Rayappan, J. Electronic noses for food quality: A review. J. Food Eng. 2015, 144, 103–111. [Google Scholar] [CrossRef]
  9. Peris, M.; Escudergilabert, L. A 21st century technique for food control: Electronic noses. Anal. Chim. Acta 2009, 638, 1–15. [Google Scholar] [CrossRef] [PubMed]
  10. Licen, S.; Barbieri, G.; Fabbris, A.; Briguglio, S.C.; Pillon, A.; Stel, F.; Barbieri, P. Odor Control Map: Self Organizing Map built from electronic nose signals and integrated by different instrumental and sensorial data to obtain an assessment tool for real environmental scenarios. Sens. Actuators B Chem. 2018, 263, 476–485. [Google Scholar] [CrossRef]
  11. Yang, L.; Guo, C.; Zhao, J.; Pan, Y. Research advance in explosive detection by electronic nose. J. Transducer Technol. 2005, 24, 1–3. [Google Scholar]
  12. Wilson, A. Diverse Applications of Electronic-Nose Technologies in Agriculture and Forestry. Sensors 2013, 13, 2295–2348. [Google Scholar] [CrossRef] [PubMed]
  13. Chen, J.; Sun, Y.; Shen, L. Reach and Application of Electronic Nose on Testing the Quality of Agriculture Products. Appl. Mech. Mater. 2015, 738–739, 116–124. [Google Scholar] [CrossRef]
  14. Tobitsuka, K. Aroma components of La France and comparison of aroma patterns of different pears. Nippon Nōgeikagaku Kaishi 2003, 77, 762–767. [Google Scholar] [CrossRef]
  15. Saevels, S.; Lammertyn, J.; Berna, A.Z.; Veraverbeke, E.A.; Natale, C.D.; Nicolai, B.M. Electronic nose as a non-destructive tool to evaluate the optimal harvest date of apples. Postharvest Biol. Technol. 2004, 31, 9–19. [Google Scholar] [CrossRef]
  16. Natale, C.; Macagnano, A.; Davide, F.; Amico, A.D.; Paolesse, R.; Boschi, T.; Faccio, M.; Ferric, G. An electronic nose for food analysis. Sens. Actuators B Chem. 1997, 44, 521–526. [Google Scholar] [CrossRef]
  17. Mastello, R.B.; Capobiango, M.; Chin, S.; Monteriro, M.; Marriott, P.J. Identification of odour-active compounds of pasteurised orange juice using multidimensional gas chromatography techniques. Food Res. Int. 2015, 75, 281–288. [Google Scholar] [CrossRef] [PubMed]
  18. Leem, C.; Dreyfus, S. A Neural Network Approach to Feature Extraction in Sensor Signal Processing for Manufacturing Automation. J. Intell. Mater. Syst. Struct. 1994, 5, 247–257. [Google Scholar] [CrossRef]
  19. Eldar, Y.; Kuppinger, P.; Bölcskei, H. Compressed Sensing of Block-Sparse Signals: Uncertainty Relations and Efficient Recovery. Mathematics 2009, 58, 3042–3054. [Google Scholar]
  20. Jolliffe, I.T. Principal Component Analysis. J. Mark. Res. 2002, 87, 513. [Google Scholar]
  21. Hyvärinen, A. Fast and robust fixed-point algorithms for independent component analysis. IEEE Trans. Neural Netw. 1999, 10, 626–634. [Google Scholar] [CrossRef] [PubMed]
  22. Romdhani, S.; Gong, S.; Psarrou, A. A multi-view nonlinear active shape model using kernel PCA. Proc. BMVC 1999, 10, 483–492. [Google Scholar]
  23. Martin, Y.G.; Pavon, J.L.P.; Cordero, B.M.; Pinto, C.G. Classification of vegetable oils by linear discriminant analysis of electronic nose data. Anal. Chim. Acta 1999, 384, 83–94. [Google Scholar] [CrossRef]
  24. Saraoğlu, H.M.; Selvi, A.O.; Ebeoğlu, M.A.; Tasaltin, C. Electronic nose system based on quartz crystal microbalance sensor for blood glucose and hbA1c levels from exhaled breath odor. IEEE Sens. J. 2013, 13, 4229–4235. [Google Scholar] [CrossRef]
  25. Brudzewski, K.; Osowski, S.; Markiewicz, T. Classification of milk by means of an electronic nose and SVM neural network. Sens. Actuators B Chem. 2004, 98, 291–298. [Google Scholar] [CrossRef]
  26. Ahron, M.; Elad, M.; Bruckstein, A. K-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation. IEEE Trans. Image Process. 2006, 15, 4311–4322. [Google Scholar] [CrossRef]
  27. Marsousi, M.; Abhari, K.; Babyn, P.; Alirezaie, J. MULTI-STAGE OMP sparse coding using local matching pursuit atoms selection. IEEE Int. Conf. Acoust. 2013, 32, 1783–1787. [Google Scholar]
  28. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G. Online dictionary learning for sparse coding. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 689–696. [Google Scholar]
  29. Yang, M.; Dai, D.; Shen, L.; Gool, L. Latent Dictionary Learning for Sparse Representation Based Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4138–4145. [Google Scholar]
  30. Thiagarajan, J.; Ramamurthy, K.; Spanias, A. Learning Stable Multilevel Dictionaries for Sparse Representation of Images. IEEE Trans. Neural Netw. Learn. Syst. 2013, 26, 1913–1926. [Google Scholar] [CrossRef] [PubMed]
  31. Fagin, R.; Halpern, J.; Moses, Y.; Vardi, M. Reasoning about Knowledge; MIT Press: Cambridge, MA, USA, 2003; Volume 67, pp. 503–525. [Google Scholar]
  32. Kennedy, J.; Eberhart, R. Particle swarm optimization. Icnn95-Int. Conf. Neural Netw. 2002, 4, 1942–1948. [Google Scholar]
  33. Zhou, H.; Wang, G. A new theory consistency index based on deduction theorems in several logic systems. Fuzzy Sets Syst. 2006, 157, 427–443. [Google Scholar] [CrossRef]
  34. Jia, P.; Duan, S.; Yan, J. An Enhanced Quantum-Behaved Particle Swarm Optimization Based on a Novel Computing Way of Local Attractor. Information 2015, 6, 633–649. [Google Scholar] [CrossRef]
  35. Jia, P.; Tian, F.; Fan, S.; He, Q.; Feng, J.; Yang, S.X. A novel sensor array and classifier optimization method of electronic nose based on enhanced quantum-behaved particle swarm optimization. Sens. Rev. 2014, 34, 304–311. [Google Scholar] [CrossRef]
  36. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. IEEE Int. Jt. Conf. Neural Netw. 2004, 2, 985–990. [Google Scholar]
Figure 1. The obtainment of experimental juice and the timeline of an E-nose sampling experiment.
Figure 1. The obtainment of experimental juice and the timeline of an E-nose sampling experiment.
Sensors 19 00916 g001
Figure 2. A schematic diagram of the experimental system.
Figure 2. A schematic diagram of the experimental system.
Sensors 19 00916 g002
Figure 3. A practical E-nose system.
Figure 3. A practical E-nose system.
Sensors 19 00916 g003
Figure 4. The response of the sensor array. Note: Strictly speaking, the response of the metal oxide sensor (MOS) should be the resistance value, and the response of the vertical coordinate in this diagram is the voltage value, which is the voltage value that we use in the circuit (designed by ourselves) to transform the change of resistance value to the change of voltage value. The resistance of the sensor is different when the gas environment is different, and the voltage value is also different, so we take the voltage value as the response value of sensors in this paper.
Figure 4. The response of the sensor array. Note: Strictly speaking, the response of the metal oxide sensor (MOS) should be the resistance value, and the response of the vertical coordinate in this diagram is the voltage value, which is the voltage value that we use in the circuit (designed by ourselves) to transform the change of resistance value to the change of voltage value. The resistance of the sensor is different when the gas environment is different, and the voltage value is also different, so we take the voltage value as the response value of sensors in this paper.
Sensors 19 00916 g004
Figure 5. The flow chart of the proposed enhanced label-consistent KSVD (E-LCKSVD).
Figure 5. The flow chart of the proposed enhanced label-consistent KSVD (E-LCKSVD).
Sensors 19 00916 g005
Figure 6. The classification accuracy of different KSVD techniques.
Figure 6. The classification accuracy of different KSVD techniques.
Sensors 19 00916 g006
Figure 7. The recognition accuracy of different dictionary sizes.
Figure 7. The recognition accuracy of different dictionary sizes.
Sensors 19 00916 g007
Figure 8. The classification accuracy of different values of α and β.
Figure 8. The classification accuracy of different values of α and β.
Sensors 19 00916 g008
Table 1. The analysis of orange juice aroma components.
Table 1. The analysis of orange juice aroma components.
No.CompoundNo.Compound
1Ethanal (C2H4O)26α-copaene (C15H24)
2Ethyl acetate (C4H8O2)27Decanal (C10H20O)
3Methyl butanoate (C5H10O2)28Linalool (C10H18O)
4α-pinene (C10H16)29Germacrene D (C15H24)
5α-thujene (C10H16)30Caryophyllene (C15H24)
6Ethyl butanoate (C6H12O2)31P-menth-1-en-4-ol (C10H18O)
7Butanoic acid,2-methyl-,ethyl ester (C7H14O2)32Citronellol acetate (C12H22O2)
8Hexanal (C6H12O)33β-farnesene (C15H24)
9β-pinene (C10H16)34p-menth-1-en-8-ol (C12H20O2)
10β-thujene (C10H16)35Valencene (C15H24)
113-carene (C10H16)36Nerol acetate (C12H20O2)
12α-phellandrene (C10H16)37(S)-carvone (C10H14O)
13β-myrcene (C10H16)381,6-octadiene,3-(1-ethoxyethoxy)-3,
7-dimethyl (C14H26O2)
14α-terpinene (C10H16)39Cadinene (C15H24)
15Limonene (C10H16)40Lavandulol acetate (C12H20O2)
16β-phellandrene (C10H16)411-cyclohexene-1-methanol,4-(1-methylethenyl)-,acetate (C12H18O2)
17Ethyl caproate (C8H16O2)421-octanol (C8H18O)
18γ-terpinene (C10H16)43Cyclopentanecarboxylic acid,2-ethylcyclohexyl ester(C14H24O2)
19β-ocimene (C10H16)442,2,3,5,6-pentamethyl-3-heptene (C12H24)
20β-cymene (C10H14)45α-caryophyllene (C15H24)
21Terpinolene (C10H16)461-undecanol (C11H24O)
22Cyclopentanone,2-methyl (C6H10O)471-propene 3-(2 cyclopentenyl)-2-methyl-1 (C21H22)
23Octanal (C8H16O)48Dimethyl phthalate (C10H10O4)
24Nonanal (C9H18O)492,4-diphenyl-4-methyl-1-pentene (C18H20)
25Octyl acetate(C10H20O2)502,4-diphenyl-4-methyl-2-pentene (C18H20)
Table 2. The main sensitive characteristics of gas sensors.
Table 2. The main sensitive characteristics of gas sensors.
SensorsSensitive Characteristics
TGS813Methane, Propane, Ethanol, Isobutane, Hydrogen, Carbon monoxide
TGS816Combustible gases, Methane, Propane, Butane, Carbon monoxide, Hydrogen, Ethanol, Isobutane
TGS822Organic solvent vapors, Methane, Carbon monoxide, Isobutane, n-Hexane, Benzene, Ethanol, Acetone
TGS2600Gaseous air contaminants, Methane, Carbon monoxide, Isobutane, Ethanol, Hydrogen
TGS2602VOCs, Odorous gases, Ammonia, Hydrogen sulfide, Toluene, Ethanol
TGS2610CEthanol, Methane, Propane, Combustible gases, Isobutane
TGS2611EMethane, Propane, Isobutane
TGS2620Vapors of organic solvents, Combustible gases, Methane, Carbon monoxide, Isobutane, Hydrogen, Ethanol
MQ135AHydrogen, Smoke, Carbon monoxide, Ethanol
MQ135Ammonia, Benzene series material, Acetone, Carbon monoxide, Ethanol, Smoke
MQ136Hydrogen sulfide, Sulfur dioxide
MQ137Ammonia, Trimethylamine, Ethanolamine
MS1100Formaldehyde, Benzene, Toluene, Xylene, Aromatic compound
MP4Methane, Combustible gases, Biogas, Natural gas
MP503Smoke, Isobutane, Formaldehyde, Ethanol
Note: the response of these sensors is nonspecific. Besides their main sensitive gas in Table 2, they are also sensitive to other gases.
Table 3. The classification accuracy of different weighted coefficients.
Table 3. The classification accuracy of different weighted coefficients.
No.αβγAcc_Train (%)Acc_Test (%)
10.60.30.192.287.5
20.30.60.189.184.4
30.10.30.684.475.0
40.330.330.3390.684.4
Table 4. Details of the parameters needing optimization.
Table 4. Details of the parameters needing optimization.
1. Weighted Coefficient2. Binary Number for Dictionary Initialization
α , β   and   γ
α + β + γ = 1, and α ,   β ,   γ ( 0 , 1 ) .
b 1 , b 2 , , b 15 ,
The value of b n , n = 1 , 2 , , 15 is 0 or 1.
Table 5. The classification accuracy of different KSVD techniques (%).
Table 5. The classification accuracy of different KSVD techniques (%).
Original DataKernel Data
KSVD+ELML-KSVDK-KSVD+ELME-LCKSVD
Acc_train76.984.488.993.8
Acc_test76.081.381.387.5
Table 6. The classification accuracy when different sensor responses are chosen to initialize the dictionary.
Table 6. The classification accuracy when different sensor responses are chosen to initialize the dictionary.
SensorsBinary Number
1TGS813100111110
2TGS816110111101
3TGS822101110110
4TGS2600111001011
5TGS26021011 11011
6TGS2610C111100111
7TGS2611E101011011
8TGS2620010101001
9MQ135A100011111
10MQ135110011111
11MQ136011011110
12MQ137011101100
13MS1100011110000
14MP4111100101
15MP503011110111
Acc_train (%)90.692.289.098.490.681.393.898.485.9
Acc_test (%)81.384.487.593.871.975.087.596.981.3
Table 7. The classification accuracy of different dictionary initialization methods (%).
Table 7. The classification accuracy of different dictionary initialization methods (%).
Normal Dictionary InitializationOptimized Dictionary Initialization
K-KSVD+ELME-LCKSVDK-KSVD+ELME-LCKSVD
Acc_train88.993.892.698.4
Acc_test81.387.585.496.9
Table 8. The results of the existing common feature extraction techniques and classifiers (%).
Table 8. The results of the existing common feature extraction techniques and classifiers (%).
No-DealingPCAKPCA
SVMAcc_train92.894.896.6
Acc_test81.393.890.6
RBFNNAcc_train89.591.094.8
Acc_test86.588.589.6
K-LDAAcc_train93.393.796.0
Acc_test88.586.588.5
Note: No-dealing means the feature matrix containing the maximum value of the steady-state response of sensors is put into the classifiers (SVM, RBFNN and K-LDA) directly.
Table 9. The advantages and drawbacks of all the techniques used in this paper.
Table 9. The advantages and drawbacks of all the techniques used in this paper.
Advantages and Drawbacks
PCAThis linear feature extraction algorithm obtains the new feature according to the variance contribution rate, but the effect is not satisfactory when dealing with the nonlinear data.
KPCAWith the help of the kernel function, the data can be mapped to a high-dimension space and then analyzed by PCA, which has the ability of processing the nonlinear data, but the high-dimension mapping increases the computational complexity.
K-LDALDA is a kind of supervised linear classifier. With the help of the kernel function, it has the ability to classify the nonlinear data to some extent, but the improvement of the recognition rate of the kernel function is limited.
RBFNNAn artificial neural network used in an E-nose earlier: using a radial basis function as the nonlinear mapping function, the recognition rate is better than K-LDA, but still lower than SVM.
SVMFor a long time, SVM is considered as an optimal classifier. With the help of the kernel function, SVM has an excellent ability to process data, but the recognition rate is affected by the quality of the input data.
E-LCKSVDThe feature extraction and classifier are integrated into one, considering the influence of the dictionary initialization, kernel function and weight coefficient of the objective function on the recognition rate; if the idea of semi-supervised learning can be added, it would be more valuable to use unlabeled data which is cheap and easily available.

Share and Cite

MDPI and ACS Style

Cao, W.; Liu, C.; Jia, P. Feature Extraction and Classification of Citrus Juice by Using an Enhanced L-KSVD on Data Obtained from Electronic Nose. Sensors 2019, 19, 916. https://doi.org/10.3390/s19040916

AMA Style

Cao W, Liu C, Jia P. Feature Extraction and Classification of Citrus Juice by Using an Enhanced L-KSVD on Data Obtained from Electronic Nose. Sensors. 2019; 19(4):916. https://doi.org/10.3390/s19040916

Chicago/Turabian Style

Cao, Wen, Chunmei Liu, and Pengfei Jia. 2019. "Feature Extraction and Classification of Citrus Juice by Using an Enhanced L-KSVD on Data Obtained from Electronic Nose" Sensors 19, no. 4: 916. https://doi.org/10.3390/s19040916

APA Style

Cao, W., Liu, C., & Jia, P. (2019). Feature Extraction and Classification of Citrus Juice by Using an Enhanced L-KSVD on Data Obtained from Electronic Nose. Sensors, 19(4), 916. https://doi.org/10.3390/s19040916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop