Next Article in Journal
Hybridized Hierarchical Watermarking and Selective Encryption for Social Image Security
Previous Article in Journal
Productivity vs. Evenness in the U.S. Financial Market: A Business Ecosystem Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Kernel-Free Quadratic Surface Support Vector Regression with Non-Negative Constraints

1
College of Mathematics and Systems Science, Xinjiang University, Urumuqi 830046, China
2
Institute of Mathematics and Physics, Xinjiang University, Urumuqi 830046, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(7), 1030; https://doi.org/10.3390/e25071030
Submission received: 8 June 2023 / Revised: 4 July 2023 / Accepted: 5 July 2023 / Published: 7 July 2023
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
In this paper, a kernel-free quadratic surface support vector regression with non-negative constraints (NQSSVR) is proposed for the regression problem. The task of the NQSSVR is to find a quadratic function as a regression function. By utilizing the quadratic surface kernel-free technique, the model avoids the difficulty of choosing the kernel function and corresponding parameters, and has interpretability to a certain extent. In fact, data may have a priori information that the value of the response variable will increase as the explanatory variable grows in a non-negative interval. Moreover, in order to ensure that the regression function is monotonically increasing on the non-negative interval, the non-negative constraints with respect to the regression coefficients are introduced to construct the optimization problem of NQSSVR. And the regression function obtained by NQSSVR matches this a priori information, which has been proven in the theoretical analysis. In addition, the existence and uniqueness of the solution to the primal problem and dual problem of NQSSVR, and the relationship between them are addressed. Experimental results on two artificial datasets and seven benchmark datasets validate the feasibility and effectiveness of our approach. Finally, the effectiveness of our method is verified by real examples in air quality.

1. Introduction

For regression problems, sometimes there is a priori information, such as the response variable increasing as the explanatory variable increases. It is more natural to expect that the the air quality will decrease when the pollution gas concentration increases. However, the model sometimes obtains regression coefficients that do not match this a priori information, which can reduce the credibility and prediction accuracy of the model. Therefore, to solve this problem, we restrict the range of values of the regression coefficients to ensure the soundness of the model.
At present, several types of constraints have been utilized, including non-negative constraints [1,2,3,4], monotonicity constraints [5,6,7,8], smoothing constraints [9,10,11], etc. Powell et al. [12] proposed a Bayesian hierarchical model for estimating constraints conditional random fields to analyze the relationship between air pollution and health. Moreover, non-negative constraints have been applied to various problems. The non-negative least squares problem (NNLS) was introduced by Lawson [13]. Chen et al. [14] presented non-negative distributed regression as an effective method specifically designed for analyzing data in wireless sensor networks. Shekkizhar et al. [15,16] proposed non-negative kernel regression to handle graph construction from data and dictionary learning. Additionally, Chapel et al. [17] proposed non-negative penalized linear regression to address the challenge of unbalanced optimal transport.
Due to its excellent generalization ability, support vector regression (SVR) [18] has been widely used in various fields, such as the financial industry [19,20] and construction industry [21,22]. However, the selection of appropriate kernel functions and their corresponding parameters can be time-consuming during experiments, prompting researchers to explore kernel-free regression models. Su proposed the non-negative constraints SVR (NNSVR) for analyzing air quality data, which is only suitable for regression problems since it cannot use kernel functions. Based on the idea of quadratic kernel-free support vector machine (QSSVC) [23], Gao et al. [24] proposed a kernel-free fuzzy reduced quadratic surface ν -support vector machine for Alzheimer’s disease classification. Zhou et al. [25] proposed a kernel-free QSSVC. For the regression problem, Ye et al. [26,27] proposed two kernel-free nonlinear regression models, quadratic surface kernel-free least squares SVR (QLSSVR) and ϵ -kernel-free soft QSSVR ( ϵ -SQSSVR), respectively. Zhai et al. [28] proposed a linear twin quadratic surface support vector regression. Zheng et al. [29] developed a hybrid QSSVR and applied it to stock indices and price forecasting. These advancements have attracted significant attention from the research community seeking more efficient approaches in regression analysis without relying on kernel functions.
In this paper, a kernel-free quadratic surface support vector regression with non-negative constraints (NQSSVR) is proposed, which incorporates the idea of a kernel-free technique and non-negative constraints. The main contributions are summarized as follows.
  • NQSSVR is proposed by utilizing the kernel-free technique, which avoids the complexity of choosing the kernel functions and their parameters, and has interpretability to some extent. In fact, the task of NQSSVR is to find a quadratic regression function to fit the data, so it can achieve better generalization ability than other linear regression methods.
  • The non-negative constraints with respect to the regression coefficients are added to construct the optimization problem of NQSSVR, which can obtain a monotonically increasing regression function with explanatory variables on a non-negative interval. In some cases, the value of the response variable grows as the explanatory variable grows. For example, when exploring the air quality examples, the air quality index will increase as the concentration of gases in the air increases.
  • Both the primal and dual problems can be solved, since our method does not involve kernel functions. In the theoretical analysis, the existence and uniqueness of solutions to the primal and dual problems, as well as their interconnections, are analyzed. In addition, the properties of regression function on the domain of definition are given.
  • Numerical experiments on artificial datasets demonstrate the visualization results of the regression function obtained by our NQSSVR. The results on benchmark datasets show that the comprehensive performance of the method is relatively better than that of linear-SVR and NNSVR. In addition, more importantly, by exploring the practical application of air quality, it can be shown that our method is more applicable than QLSSVR and ϵ -SQSSVR.
The paper is structured as follows. Section 2 introduces a brief introduction to the ϵ -SQSSVR model, and some definitions and notations. In Section 3, we construct the primal and dual problems for NQSSVR and analyze the corresponding properties. Section 4 presents the results of numerical experiments conducted on datasets. Finally, Section 5 provides conclusions from this study.

2. Background

In this section, we give the related definitions and notations, and review the ϵ -SQSSVR model.

2.1. Definition and Notations

The following mathematical notations are utilized in this paper. Lowercase bold and uppercase bold represent vectors and matrices, respectively. I is the identity matrix of any size, S m is the set of m-dimensional symmetric matrices, R n × m is the set of n × m dimensional matrices. Next, define the operators as follows.
Definition 1.
For any symmetry matrix U = ( u k l ) m × m S m , its half-vectorization operator can be defined as follows:
hvec ( U ) = ( u 11 , , u 1 m , u 22 , , u 2 m , , u m m ) T R m ( m + 1 ) 2 .
Definition 2.
For any vector u = ( u 1 , u 2 , , u m ) T R m , define the following quadratic operator:
lvec ( u ) = ( 1 2 u 1 u 1 , , u 1 u m , 1 2 u 2 u 2 , , u 2 u m , , 1 2 u m u m ) T R m ( m + 1 ) 2 .
Definition 3.
For a vector u = ( u 1 , u 2 , , u m ) T R m , define the vector-to-matrix operator as follows:
mat ( u ) u 1 u 2 u m 0 0 0 0 u 1 0 u 2 u m 0 0 0 0 0 u 1 0 u 2 u m R m × m ( m + 1 ) 2 .

2.2. ϵ -SQSSVR

Given the training set
T = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } ,
where x i R m , y i R , i = 1 , 2 , , n . The task of ϵ -SQSSVR is to seek the quadratic regression function
g ( x ) = 1 2 x T W x + b T x + c ,
where W S m , b R m , and c R . To obtain the regression function (5), the optimization problem is established as follows:
min W , b , c , ξ ( * ) 1 2 i = 1 n Wx i + b 2 + C i = 1 n ( ξ i + ξ i * ) ,
s . t . ( 1 2 x i T W x i + b T x i + c ) y i ϵ + ξ i , i = 1 , , n ,
y i ( 1 2 x i T W x i + b T x i + c ) ϵ + ξ i * , i = 1 , , n ,
ξ i 0 , ξ i * 0 , i = 1 , 2 , , n ,
where C > 0 is a penalty parameter, and ξ ( * ) = ( ξ 1 , ξ 1 * , , ξ n , ξ n * ) T is the slack vector. The optimization problem (6)–(9) is a quadratic programming problem, so it can be solved directly. In addition, this model uses the quadratic surface kernel-free technique, which avoids the difficulty of choosing the kernel function and the corresponding parameters.

3. Kernel-Free QSSVR with Non-Negative Constraints (NQSSVR)

In this section, we establish the primal and dual problems of the kernel-free QSSVR with the non-negative constraints (NQSSVR). The properties of primal and dual problems are discussed, and the properties of the regression function with non-negative constraints are proved.

3.1. Primal Problem

Given the training set T (4), to find the regression function (5), the following optimization problem is formulated
min W , b , c , ξ ( * ) 1 2 i = 1 n Wx i + b 2 + C i = 1 n ( ξ i + ξ i * ) ,
s . t . ( 1 2 x i T W x i + b T x i + c ) y i ϵ + ξ i , i = 1 , 2 , , n ,
y i ( 1 2 x i T W x i + b T x i + c ) ϵ + ξ i * , i = 1 , 2 , , n ,
w k l 0 , b k 0 , k , l = 1 , , m ,
ξ i 0 , ξ i * 0 , i = 1 , n ,
where W = ( w k l ) m × m S m , b = ( b 1 , , b m ) R m , c R . w k l 0 , b k 0 , k , l = 1 , , m mean that each component of W and b is greater than or equal to zero. C > 0 is the penalty parameter, and ξ ( * ) = ( ξ 1 , ξ 2 * , , ξ n , ξ n * ) T is a slack vector.
In the above optimization problem (10)–(14), we impose constraints on the regression coefficients, namely w k l 0 , b k 0 , k , l = 1 , , m . Restricting the range of values of regression coefficients can help us to obtain regression functions that are more consistent with a priori information. In addition, the optimization problem does not involve kernel functions, which can avoid the complicated process of kernel functions, and its parameters selection further reduced computation time.
According to Definitions 1–3, the primal optimization problem (10)–(14) is simplified to the following form:
min z , c , ξ ( * ) 1 2 z T Gz + C i = 1 n ( ξ i + ξ i * ) ,
s . t . ( z T s i + c ) y i ϵ + ξ i , i = 1 , n ,
y i ( z T s i + c ) ϵ + ξ i * , i = 1 , n ,
z 0 ,
ξ i 0 , ξ i * 0 , i = 1 , n ,
where z = [ w , b ] R m ( m + 3 ) 2 , w hvec ( W ) , G = i = 1 n ( V i ) T V i , V i [ mat ( x i ) , I ] R m × ( m ( m + 3 ) 2 ) , c R , I S m , s i lvec ( x i ) . z 0 means that each component of z is greater than or equal to to zero. The matrix G is positive semidefinite matrix, since G = i = 1 n ( V i ) T V i 0 .

3.2. Dual Model

Next, the Lagrange multiplier vectors α ( * ) = ( α 1 , α 1 * , , α n , α n * ) T , η ( * ) = ( η 1 , η 1 * , , η n , η n * ) T , and β R m ( m + 3 ) 2 are introduced to the optimization problem (15)–(19). The Lagrange function can be expressed as follows:
L ( z , c , ξ ( * ) , α ( * ) , η ( * ) , β ) = 1 2 z T Gz + C i = 1 n ( ξ i + ξ i * ) β T z i = 1 n α i ( y i + ϵ + ξ i ( z T s i + c ) ) i = 1 n η i ξ i i = 1 n α i * ( ϵ + ξ i + ( z T s i + c ) y i ) i = 1 n η i * ξ i * .
The Karush–Kuhn–Tuchker (KKT) conditions for problems (15)–(19) are given as follows:
z L = Gz i = 1 n ( α i α i * ) s i β = 0 ,
c L = i = 1 n ( α i α i * ) = 0 ,
ξ i L = C α i η i = 0 , i = 1 , n ,
ξ i * L = C α i * η i * = 0 , i = 1 , n ,
ϵ + ξ i ( z T s i + c ) + y i 0 , ξ i 0 , i = 1 , , n ,
ϵ + ξ i * y i + ( z T s i + c ) 0 , ξ i * 0 , i = 1 , , n ,
α i ( y i + ϵ + ξ i s i T z c ) = 0 , α i 0 , i = 1 , , n ,
α i * ( y i + ϵ + ξ i + s i T z + c ) = 0 , α i * 0 , i = 1 , , n ,
η i * ξ i * = 0 , η i ξ i = 0 , η i * 0 , η i 0 , i = 1 , n ,
β z = 0 , β 0 .
According to the Equation (21),
z = G 1 ( i = 1 n ( α i α i * ) s i + β ) .
By substituting (21)–(30) into (20), the dual problem of the optimization problem (15)–(19) is formulated as
min α ( * ) , β 1 2 ( i = 1 n ( α i α i * ) s i + β ) T G 1 ( i = 1 n ( α i α i * ) s i + β )
+ i = 1 n ( α i + α i * ) ϵ i = 1 n ( α i α i * ) y i ,
s . t . i = 1 n ( α i α i * ) = 0 ,
0 α i , α i * C , i = 1 , n ,
β j 0 , j = 1 , , m ( m + 3 ) 2 ,
where α ( * ) = ( α 1 , α 1 * , , α n , α n * ) T and β = ( β 1 , , β m ( m + 3 ) 2 ) T are vectors, and C > 0 is the penalty parameter. Since the proposed model does not involve kernel functions, both the primal and dual problems can be solved directly. Solving the primal problem saves the cost of computing the inverse matrix. In particular, when the input dimension is large, solving the dual problem is more convenient.

3.3. Some Theoretical Analysis

In this subsection, the theoretical properties of the primal and dual problems, as well as the regression function after adding the non-negative constraints, are analyzed into properties.
Theorem 1.
Given the training set T (4) and C > 0 , if G is a positive definite matrix and ( z * , c * , ξ * , ξ * * ) is the optimal solution to the primal problem (15)–(19), then z * is unique.
Proof. 
Suppose ( z ¯ , c ¯ , ξ ¯ , x i ¯ * ) and ( z ^ , c ^ , ξ ^ , ξ ^ * ) are the two optimal solutions of the primal problem. There exists μ ( 0 , 1 ) such that the following equation holds:
( z , c , ξ , ξ * ) = μ ( z ¯ , c ¯ , ξ ¯ , ξ ¯ * ) + ( 1 μ ) ( z ^ , c ^ , ξ ^ , ξ ^ * ) .
Clearly, ( z , c , ξ , ξ * ) is also an optimal solution to the optimization problem (15)–(19), so the following inequalities hold:
1 2 z T Gz + C i = 1 n ( ξ i + ξ i * ) 1 2 z ¯ T G z ¯ + C i = 1 n ( ξ i ¯ + ξ i ¯ * ) ,
1 2 z T Gz + C i = 1 n ( ξ i + ξ i * ) 1 2 z ^ T G z ^ + C i = 1 n ( ξ i ^ + ξ i ^ * ) .
Multiplying inequality (37) by μ and inequality (38) by ( 1 μ ) , then summing them up yields
1 2 z T Gz + C i = 1 n ( ξ i + ξ i * ) μ 2 z ^ T G z ^ + 1 μ 2 z ¯ T G z ¯ + C i = 1 n ( μ ξ i ^ + ( 1 μ ) ξ i ¯ + μ ξ i ¯ * + ( 1 μ ) ξ i ^ * ) .
By simplifying the Formula (39), we obtain
1 2 ( μ z ¯ + ( 1 μ ) z ^ ) T G ( μ z ¯ + ( 1 μ ) z ^ ) μ 2 z ^ T G z ^ + 1 μ 2 z ¯ T G z ¯ .
Then, we have
μ ( 1 μ ) ( z ¯ z ^ ) T G ( z ¯ z ^ ) 0 .
Since the matrix G is a positive definite matrix and μ ( 0 , 1 ) , inequality (41) holds if and only if z ¯ = z ^ . □
Theorem 2.
For the training set T (4) and C > 0 , if the matrix G is positive definite, the optimal solution α ( * ) = ( α 1 , α 1 * , , α n , α n * ) T of the dual problem (32)–(35) exists and is unique, and the optimal solution of the primal problem (15)–(19) can be expressed as
z = G 1 ( i = 1 n ( α i α i * ) s i + β )
c = y j + ϵ ( i = 1 n ( α i α i * ) s i + β ) G 1 s j , f o r s o m e α j ( 0 , C ) ,
or
c = y k ϵ ( i = 1 n ( α i α i * ) s i + β ) G 1 s k , f o r s o m e α k * ( 0 , C ) .
Proof. 
By the (21) equation in the KKT condition, we have
z = G 1 ( i = 1 n ( α i α i * ) s i + β ) .
If α ( * ) there exists components, α j and α k * . such that α j ( 0 , C ) or α j * = 0 , α k = 0 or α k * ( 0 , C ) , by the complementary slackness conditions, ξ j = ξ k * = 0 , we have
c = y j + ϵ ( i = 1 n ( α i α i * ) s i + β ) G 1 s j , α j ( 0 , C ) ,
or
c = y k ϵ ( i = 1 n ( α i α i * ) s i + β ) G 1 s k , α k * ( 0 , C ) .
Next, the properties of the regression function (5) after adding non-negative constraints with respect to the regression coefficients are analyzed. The domain D is defined as follows:
D = { x = ( [ x ] 1 , , [ x ] m ) T [ x ] 1 0 , , [ x ] m 0 } .
Theorem 3.
On the domain D (48), the regression function g ( x ) (5) is monotonically non-decreasing respect to each component of x if and only if the regression coefficients have the following restrictions:
w k l 0 , , k , l = 1 , , m , b k 0 , , k = 1 , , m ,
where w k l 0 , and b k 0 , k , l = 1 , , m mean that each component of W and b is greater than or equal to zero.
Proof. 
The function g ( x ) can be written as
g ( x ) = 1 2 k , l = 1 m [ x ] k w k l [ x ] l + k = 1 m b k [ x ] k + c .
It only remains to justification that this holds for the k-th component of x , the quadratic function containing [ x ] k can be expressed as
g ( [ x ] k ) = 1 2 [ x ] k w k k [ x ] k + [ x ] k l k m w k l [ x ] l + b k [ x ] k + c .
Taking the derivative of the above equation yields
g ( [ x ] k ) [ x ] k = l = 1 m w k l [ x ] l + b k .
On the domain D , the function g ( [ x ] k ) monotonically non-decreasing is equivalent to being non-negative at the right end of the above equation, so it is a necessary and sufficient condition to prove that the latter holds as follows:
w k l = w l k 0 , l = 1 , , m , b k 0 , k = 1 , , m .
Sufficiency is obvious, and we only need to prove necessity. Supposing b k < 0 and taking x = ( [ x ] 1 , , [ x ] m ) T = 0 , then we obtain the following:
g ( [ x ] k ) [ x ] k = l = 1 m w k l [ x ] l + b k = b k < 0 .
The above relation is contrary to the known conditions. Supposing the existence of w k l < 0 with all other components being zero. Obviously, when the [ x ] l is sufficiently large, the following formula holds:
g ( [ x ] k ) [ x ] k = l = 1 m w k l [ x ] l + b k = w k l [ x ] l + b k < 0 .
This is a contradict. Similarly, it can be shown that Theorem 3 holds. □

4. Numerical Experiments

To verify the validity of our proposed NQSSVR model, we compare it with other methods, including linear SVR (lin-SVR), SVR with Gaussian kernel (rbf-SVR), and polynomial kernel (poly-SVR), and linear SVR with non-negative constraints (NNSVR), as well as QLSSVR and ϵ -SQSSVR. The primal and dual problems of the NQSSVR method are denoted as NQSSVR(p) and NQSSVR(d), respectively. The above experiments are tested on 2 artificial datasets, 7 UCI [30] datasets, and AQCI datasets. All numerical experiments in this section are conducted on a computer equipped with a 2.50GHz (i7-9700) CPU and 8G RAM using MatlabR2016(a).
To validate the fitting performances of various methods, the following four evaluation criteria are introduced as shown in Table 1. Without loss of generality, let y i ^ and y ¯ i be the predicted and mean values, respectively. The penalty parameters C and ϵ -insensitive parameter, as well as the Gaussian kernel parameter σ , are selected from { 2 i i = 6 , 5 , , 5 , 6 } , while the polynomial kernel parameter p is selected from { 1 , 2 } . All methods are selected through 5-fold cross-validation to obtain the optimal parameters.

4.1. Artificial Datasets

The 2 artificial datasets are conducted to validate the performance of the NQSSVR model.
Example 1.
y i = x i + 0.3 + ζ , x i [ 0 , 1 ] , ζ N ( 0 , 0 . 1 2 ) , i = 1 , , 100 .
The data points are indicated by magenta “o”, the red solid line indicates the target function, and the blue dashed line indicates the regression function obtained by the kernel function method. The green dashed line and the black solid line indicate NQSSVR(p), and NQSSVR(d), respectively. The regression functions obtained by NQSSVR and different kernel functions are shown in Figure 1. From Figure 1, it can be seen that NQSSVR yields an approximately linear regression function as well as the other three methods. The regression coefficients obtained by solving the primal and dual problems of our proposed method are w = 0.0001 , b = 1.0050 , c = 0.2897 and w = 0.0001 , b = 0.9946 , c = 0.3097 , respectively. So, NQSSVR can handle linear regression problem.
Example 2.
z i = x i 2 + y i 2 + 0.1 + ζ , x i , y i [ 0 , 1 ] , ζ N ( 0 , 0 . 2 2 ) , i = 1 , , 200 .
The fitting results of an artificial dataset are shown in Figure 2. The data points are denoted by “·”. Figure 2a,b present the regression surfaces obtained from the primal and dual problems, respectively. On this dataset, the regression coefficients obtained by our method are W = [ 0.9982 , 0.0000 ; 0.0000 , 0.9956 ] , b = [ 0.0000 , 0.000 ] T , c = 0.1005 , and W = [ 0.9896 , 0.0000 ; 0.0000 , 0.9967 ] , b = [ 0.0000 , 0.0000 ] T , c = 0.1034 . From Figure 2 and the regression coefficients, the quadratic surface fitted by our method can be matched to the actual distribution of the data.
Table 2 shows the results of the our proposed model and SVR with kernel functions on the two datasets mentioned above. When considering the linear regression problems, it is evident that the five models yield n similar outcomes. However, when addressing the non-linear data, our method demonstrates superior performance compared to the other three methods, as evidenced by the smaller average values of RMSE and MAE. Moreover, the difference in R 2 values between our method and the optimal result is minimal. Notably, the T2 values reveal that our method exhibits faster computation times than SVR with kernel function. This advantage comes from the fact that it does not contain a kernel function, thus eliminating the need for kernel parameter selection.
For example 2, the influence of the parameter on the accuracy of our proposed method is analyzed. As can be seen by Figure 3, the penalty parameters C and insensitive loss parameter ϵ have a greater impact on the accuracy of the NQSSVR model. So, reasonable parameters can improve the accuracy of the model. In the next experiments, we choose the optimal parameters for the model within the defined parameter range.
Next, the average test time is compared for NQSSVR(p), NQSSVR(d), rbf-SVR, and poly-SVR. Since lin-SVR is only applicable to linear regression, no comparison is made here. The CPU running time of the above four methods in different dimensions and data points is shown in Table 3. Where the input dimensions m of data point are 2, 4, 8, and 16 and the number n of data points is 200, 400, 600, 800, and 1000, respectively. It is noteworthy that the time variation of NQSSVR(p) remains small as the number of data points increases for the same input dimension, outperforming both the rbf-SVR and poly-SVR methods. Moreover, when the number of data points is consistent, NQSSVR(p) exhibits shorter average test time costs compared to rbf-SVR and poly-SVR. Furthermore, as the input dimension of data points increases, the average test time for the dual problem is found to be shorter than that for the primal problem.

4.2. Benchmark Datasets

In this section, to further validate the reliability of the proposed method, the NQSSVR model is compared with the lin-SVR, poly-SVR, rbf-SVR, NNSVR, QLSSVR, and ϵ -SQSSVR models on seven benchmark datasets. Details of all datasets are listed in Table 4. All datasets are normalized before conducting data experiments, and are divided into training datasets, test datasets and a validation datasets in a ratio of 3:1:1. All methods are compared on evaluation criteria: MAE, RMSE, T1, T2. The top two relatively better results are highlighted in bold. All results were repeated 5 times and their mean values are calculated.
Table 5 lists the regression results of the eight methods on the seven datasets. In terms of the evaluation criteria RMSE and MAE, it can be seen that NQSSVR is significantly better than lin-SVR and NNSVR. For most of the datasets, the NQSSVR model outperforms QLSSVR and ε -SQSSVR, and is not significantly different from rbf-SVR and poly-SVR. In terms of time, our method is second only to QLSSVR and outperforms other methods.
To compare the performances of our proposed method and other six methods, the Friedman test and post hoc test are employed. Initially, the Friedman test is conducted with the null hypothesis states that all methods have the same performances. Furthermore, we can calculate the Friedman statistics for each evaluation criterion using the following formula.
τ χ 2 = 12 N K ( K + 1 ) ( i = 1 K R i 2 K ( K + 1 ) 2 4 ) ,
τ F = ( N 1 ) τ χ 2 N ( K 1 ) τ χ 2 ,
where N and K are, respectively, the numbers of datasets and methods, R i is the average rank of the i-th method.
According to the Formula (59), the Friedman statistics corresponding to the three criteria are 12.2124, 13.8361 and 35.1600, respectively. Next, for α = 0.05, the critical value of Friedman statistic is calculated to be F α = 2.2371 . Since the Friedman statistic on each regression criteria is greater than F α , so we reject the null hypothesis. That is, these 8 methods have significantly different performances on the 3 evaluation criteria. To further compare the difference of each method, we proceed with a post hoc test. Specifically, if the difference of average ranks for two methods is larger than the critical difference ( C D ), then their performances are considered to be significantly different. Where the C D value can be calculated by the Formula (60)
C D = q α K ( K + 1 ) 6 N .
For α = 0.05 , we know q α = 3.0308 . Thus, we obtain C D = 3.9685 by the Formula (60).
Figure 4 visually displays the results of Friedman test and Nemenyi post hoc test on three regression evaluation criteria, respectively. Where the average ranks of each method for three criteria are marked along an axis. The axis is turned so that the lowest (best) ranks is to the right of each criterion. Groups of methods that are not significantly different are linked by a red line. Statistically, the performance of NQSSVR (p) is not significantly different from rbf-SVR and poly-SVR in terms of RMSE, MAE. And our method ranks better than both the kernel-free quadratic surface and lin-SVR models on RMSE, MAE. In terms of time, our model ranks third and fourth, outperforming the SVR with kernel functions and ϵ -SQSSVR. In general, the comprehensive performance of our method is similar to rbf-SVR and poly-SVR, and completely superior to lin-SVR and NNSVR.

4.3. Air Quality Composite Index Dataset (AQCI)

This section uses two AQCI datasets, the monthly AQCI dataset and the daily AQCI dataset. These two datasets containing 18 data points and 841 data points, respectively. Each data point has six input features including nitrogen dioxide ( NO 2 ), sulfur dioxide ( SO 2 ), PM2.5, ozone ( O 3 ), carbon monoxide (CO), PM10, respectively. And the output response is AQCI. Our method is compared with QLSSVR, ϵ -SQSSVR and NNSVR.
In Figure 5, the value of AQCI has a tendency to increase as the values of the remaining five input features increase, except for the O 3 feature. Therefore, the regression function we obtain should be monotonically increasing on the monthly AQCI dataset.
Table 6 shows the experimental results of our NQSSVR and the other three methods on these two datasets. The accuracy of our model is better than that of QLSSVR and ϵ -SQSSVR on the datasets because our NQSSVR imposes non-negative constraints with respect to the regression coefficients. In addition, our model has greater accuracy than the NNSVR model. Because NNSVR can only obtain a linear regression function, but NQSSVR can obtain a quadratic regression function.
To investigate the effect of adding non-negative constraints on the accuracy of the regression function, we compare the regression coefficients W and b obtained by NQSSVR(p) with those obtained using the other three methods. Since NNSVR is a linear model, it has only linear term coefficients b and does not involve nonlinear term coefficients W . The regression coefficients obtained by the four methods are small, and for comparison purposes, we enlarge the W and b by a factor of 100 and 10, respectively, before drawing the figure. When the regression coefficient is negative, the color of the color block is closer to blue.
We want to obtain a regression function that is monotonically increasing, by Theorem 3 it is equivalent to the non-negative constraints with respect to regression coefficients. From the Figure 6 and Figure 7, we can see that ϵ -SQSSVR and QLSSVR obtained W and b contain negative numbers, so the regression functions obtained do not match the a priori information. However, our model yields regression coefficients that all match the a priori information. Therefore, adding non-negative constraints can improve the accuracy and reasonableness of the model. Since our method can obtain a quadratic regression function, it is therefore more accurate than the linear regression function obtained by NNSVR.
Figure 8 show the effect of the parameters on the performance of our model. Here we have chosen three evaluation criteria, including RMSE, MAE, and R 2 . Figure 8a–c displays the experimental results obtained for the dual problem, while Figure 8d–f presents the experimental results for the primal problem. It can be observed that the effect of the parameters on our proposed model remains significant. It is worth noting that our model receives ϵ -insensitive parameter influence more and is not very sensitive to the change of penalty parameter C.

5. Conclusions

For the regression problem, a novel kernel-free quadratic surface support vector regression with non-negative constraints (NQSSVR) is proposed by utilizing the kernel-free technique and introducing the non-negative constraints with respect to regression coefficients. Specifically, by using a quadratic surface to fit the data, the regression function is nonlinear and does not involve kernel functions, so the model is unnecessary to select kernel functions and corresponding parameters, and the obtained regression function has better interpretability. Moreover, adding non-negative constraints with respect to regression coefficients to the model ensures that the obtained regression function conforms to the monotonic non-decreasing characteristics on non-negative interval. In fact, when exploring air quality examples, there is a prior information that air quality indicators will increase with the increase in all gas concentrations in the atmosphere. Fortunately, we have proven that the quadratic regression function obtained by NQSSVR is monotonically non-decreasing on the non-negative interval if and only if the non-negative constraints with respect to the regression coefficients hold true. The results of numerical experiments on the two artificial datasets, seven benchmark datasets and air quality datasets demonstrate that our method is feasible and effective.
In this paper, we impose a non-negative restriction on the regression coefficients based on prior information. In the subsequent optimization problems, we can add different restrictions to the values of the regression coefficients based on the prior information. For example, part of the regression coefficients are restricted to non-negative intervals and part of the regression coefficients are unrestricted.

Author Contributions

Methodology, D.W. and Z.Y.; Software, D.W.; Writing—original draft, D.W.; Writing—review & editing, Z.Y. and J.Y.; Supervision, Z.Y., J.Y. and X.Y.; Funding acquisition, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. 12061071).

Data Availability Statement

All of the benchmark datasets used in our numerical experiments are from the UCI Machine Learning Repository, which are available at https://archive.ics.uci.edu/ml/351index.php (the above datasets accessed on 18 August 2021).

Acknowledgments

The authors would like to thank the editor and referees for their valuable comments and suggestions, which helped us to improve the results of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, L.; Harrach, B.; Seo, J.K. Monotonicity-based electrical impedance tomography for lung imaging. Inverse Probl. 2018, 34, 045005. [Google Scholar] [CrossRef] [Green Version]
  2. Chatterjee, S.; Guntuboyina, A.; Sen, B. On matrix estimation under monotonicity constraints. Bernoulli 2018, 24, 1072–1100. [Google Scholar] [CrossRef] [Green Version]
  3. Wang, J.; Qian, Y.; Li, F.; Liang, J.; Ding, W. Fusing fuzzy monotonic Decision Trees. IEEE Trans. Fuzzy Syst. 2019, 28, 887–900. [Google Scholar] [CrossRef]
  4. Henderson, N.C.; Varadhan, R. Damped anderson acceleration with restarts and monotonicity control for accelerating em and em-like algorithms. J. Comput. Graph. Stat. 2019, 28, 834–846. [Google Scholar] [CrossRef]
  5. Bro, R.; Sidiropoulos, N.D. Least squares algorithms under unimodality and non-negativity constraints. J. Chemom. J. Chemom. Soc. 1998, 12, 223–247. [Google Scholar] [CrossRef]
  6. Luo, X.; Zhou, M.C.; Li, S.; Hu, L.; Shang, M. Non-negativity constrained missing data estimation for high-dimensional and sparse matrices from industrial applications. IEEE Trans. Cybern. 2019, 50, 1844–1855. [Google Scholar] [CrossRef]
  7. Theodosiadou, O.; Tsaklidis, G. State space modeling with non-negativity constraints using quadratic forms. Mathematics 2019, 9, 1905. [Google Scholar] [CrossRef]
  8. Haase, V.; Hahn, K.; Schndube, H.; Stierstorfer, K.; Maier, A.; Noo, F. Impact of the non-negativity constraint in model-based iterative reconstruction from CT data. Med. Phys. 2019, 46, 835–854. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Yamashita, S.; Yagi, Y.; Okuwaki, R.; Shimizu, K.; Agata, R.; Fukahata, Y. Potency density tensor inversion of complex body waveforms with time-adaptive smoothing constraint. Geophys. J. Int. 2022, 231, 91–107. [Google Scholar] [CrossRef]
  10. Wang, J.; Deng, Y.; Wang, R.; Ma, P.; Lin, H. A small-baseline InSAR inversion algorithm combining a smoothing constraint and L1-norm minimization. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1061–1065. [Google Scholar] [CrossRef]
  11. Mammen, E.; Marron, J.S.; Turlach, B.A.; Wand, M.P. A general projection framework for constraints smoothing. Stat. Sci. 2001, 16, 232–248. [Google Scholar] [CrossRef]
  12. Powell, H.; Lee, D.; Bowman, A. Estimating constraints concentration–response functions between air pollution and health. Environmetrics 2012, 23, 228–237. [Google Scholar] [CrossRef]
  13. Lawson, C.L.; Hanson, R.J. Solving Least Squares Problems; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1995. [Google Scholar]
  14. Chen, J.; Richard, C.; Honeine, P.; Bermudez, J.C.M. Non-negative distributed regression for data inference in wireless sensor networks. In Proceedings of the 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 7–10 November 2010; pp. 451–455. [Google Scholar]
  15. Shekkizhar, S.; Ortega, A. Graph construction from data by non-negative kernel regression. In Proceedings of the ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3892–3896. [Google Scholar]
  16. Shekkizhar, S.; Ortega, A. NNK-Means: Dictionary learning using non-negative kernel regression. arXiv 2021, arXiv:2110.08212. [Google Scholar]
  17. Chapel, L.; Flamary, R.; Wu, H.; Févotte, C.; Gasso, G. Unbalanced optimal transport through non-negative penalized linear regression. Adv. Neural Inf. Process. Syst. 2021, 34, 23270–23282. [Google Scholar]
  18. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. Adv. Neural Inf. Process. Syst. 1996, 9, 779–784. [Google Scholar]
  19. Fauzi, A. Stock price prediction using support vector machine in the second Wave of COVID-19 pandemic. Insearch Inf. Syst. Res. J. 2021, 1, 58–62. [Google Scholar] [CrossRef]
  20. Huang, S.; Cai, N.; Pacheco, P.P.; Narrandes, S.; Wang, Y.; Xu, W. Applications of support vector machine (SVM) learning in cancer genomics. Cancer Genom. Proteom. 2018, 15, 41–51. [Google Scholar]
  21. Zhong, H.; Wang, J.; Jia, H.; Mu, Y.; Lv, S. Vector field-based support vector regression for building energy consumption prediction. Appl. Energy 2019, 242, 403–414. [Google Scholar] [CrossRef]
  22. Guo, H.; Nguyen, H.; Bui, X.N.; Armaghani, D.J. A new technique to predict fly-rock in bench blasting based on an ensemble of support vector regression and GLMNET. Eng. Comput. 2021, 37, 421–435. [Google Scholar] [CrossRef]
  23. Dagher, I. Quadratic kernel-free non-linear support vector machine. J. Glob. Optim. 2008, 41, 15–30. [Google Scholar] [CrossRef]
  24. Gao, Z.; Wang, Y.; Huang, M.; Luo, J.; Tang, S. A kernel-free fuzzy reduced quadratic surface ν-support vector machine with applications. Appl. Soft Comput. 2022, 127, 109390. [Google Scholar] [CrossRef]
  25. Zhou, J.; Tian, Y.; Luo, J.; Zhai, Q. Novel non-Kernel quadratic surface support vector machines based on optimal margin distribution. Soft Comput. 2022, 26, 9215–9227. [Google Scholar] [CrossRef]
  26. Ye, J.Y.; Yang, Z.X.; Li, Z.L. Quadratic hyper-surface kernel-free least squares support vector regression. Intell. Data Anal. 2021, 25, 265–281. [Google Scholar] [CrossRef]
  27. Ye, J.Y.; Yang, Z.X.; Ma, M.P.; Wang, Y.L.; Yang, X.M. ϵ-Kernel-free soft quadratic surface support vector regression. Inf. Sci. 2022, 594, 177–199. [Google Scholar] [CrossRef]
  28. Zhai, Q.; Tian, Y.; Zhou, J. Linear twin quadratic surface support vector regression. Math. Probl. Eng. 2020, 2020, 3238129. [Google Scholar] [CrossRef]
  29. Zheng, L.J.; Tian, Y.; Luo, J.; Hong, T. A novel hybrid method based on kernel-free support vector regression for stock indices and price forecasting. J. Oper. Res. Soc. 2023, 74, 690–702. [Google Scholar] [CrossRef]
  30. Dua, D.; Graff, C. UCI Machine Learning Repository. 2017. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 18 August 2021).
Figure 1. The visualization results on Example 1.
Figure 1. The visualization results on Example 1.
Entropy 25 01030 g001
Figure 2. The visualization results on Example 2.
Figure 2. The visualization results on Example 2.
Entropy 25 01030 g002
Figure 3. Effect of parameters on method performance NQSSVR.
Figure 3. Effect of parameters on method performance NQSSVR.
Entropy 25 01030 g003
Figure 4. The results of Friedman test and Nemenyi post hoc test.
Figure 4. The results of Friedman test and Nemenyi post hoc test.
Entropy 25 01030 g004
Figure 5. The relationship between six features and AQCI on monthly AQCI dataset.
Figure 5. The relationship between six features and AQCI on monthly AQCI dataset.
Entropy 25 01030 g005
Figure 6. W -matrices.
Figure 6. W -matrices.
Entropy 25 01030 g006
Figure 7. b -vectors.
Figure 7. b -vectors.
Entropy 25 01030 g007
Figure 8. Effect of parameters on method performance NQSSVR.
Figure 8. Effect of parameters on method performance NQSSVR.
Entropy 25 01030 g008
Table 1. Evaluation criteria.
Table 1. Evaluation criteria.
Evaluation CriteriaFormulas
R 2 R 2 = S S R S S T = i = 1 n ( y ^ i y ¯ i ) 2 i = 1 n ( y i y ¯ i ) 2
Mean Absolute Error (MAE)MAE = 1 n i = 1 n y i y ^ i
Root Mean Squared Error (RMSE)RMSE = 1 n i = 1 n ( y i y ^ i ) 2
T1Average test time
T2Time to select parameters
Table 2. Experimental results on artificial datasets.
Table 2. Experimental results on artificial datasets.
DatasetsAlgorithmsRMSEMAE R 2 T1T2
lin-SVR0.0901 ± 0.00100.0716 ± 0.00110.9159 ± 0.01180.2085 ± 0.018831.2781
poly-SVR0.0894 ± 0.00150.0716 ± 0.00120.9502 ± 0.01340.1412 ± 0.009636.9959
Example 1rbf-SVR0.0865 ± 0.00100.0658 ± 0.00140.9680 ± 0.01530.1554 ± 0.005562.9346
NQSSVR(p)0.0885 ± 0.00120.0714 ± 0.00180.9066 ± 0.02600.0612 ± 0.009416.2336
NQSSVR(d)0.0889 ± 0.00070.0716 ± 0.00040.8829 ± 0.00220.1278 ± 0.001924.2926
lin-SVR1.4762 ± 0.00851.2420 ± 0.00740.5141 ± 0.00210.4143 ± 0.0204133.6888
poly-SVR0.1983 ± 0.00150.1567 ± 0.00180.9941 ± 0.00170.3464 ± 0.0413130.0461
Example 2rbf-SVR0.2057 ± 0.00660.1615 ± 0.00470.9899 ± 0.00140.4600 ± 0.0282573.8611
NQSSVR(p)0.1969 ± 0.00160.1544 ± 0.00140.9930 ± 0.00140.1850 ± 0.013928.3872
NQSSVR(d)0.1982 ± 0.00170.1552 ± 0.00170.9896 ± 0.00120.2056 ± 0.023432.7823
Table 3. Experimental results on the benchmark datasets.
Table 3. Experimental results on the benchmark datasets.
Data Points∖DimensionsMethodsm = 2m = 4m = 8m = 16
n = 200poly-SVR3.9670 ± 0.11833.3426 ± 0.08133.6770 ± 0.05963.8512 ± 0.0430
rbf-SVR1.1682 ± 0.03181.0756 ± 0.05751.0694 ± 0.04371.1634 ± 0.0173
NQSSVR(p)0.1592 ± 0.01460.2542 ± 0.00660.4346 ± 0.01931.0058 ± 0.0608
NQSSVR(d)0.7322 ± 0.05570.8390 ± 0.02190.9502 ± 0.04491.1458 ± 0.0404
n = 400poly-SVR22.1766 ± 0.656521.0828 ± 0.683520.7110 ± 0.440723.2322 ± 07712
rbf-SVR4.1252 ± 0.17254.2152 ± 0.15464.2804 ± 0.17354.4358 ± 0.1441
NQSSVR0.3942 ± 0.01990.5334 ± 0.02190.9442 ± 0.02813.2328 ± 0.1464
NQSSVR(d)3.2784 ± 0.13393.8628 ± 0.09544.2040 ± 0.33104.1320 ± 0.1463
n = 600poly-SVR64.8220 ± 1.018564.9902 ± 1.643163.5246 ± 1.303169.9104 ± 2.0957
rbf-SVR9.6822 ± 0.42479.5922 ± 0.401510.0280 ± 0.643610.7400 ± 0.5887
NQSSVR0.6522 ± 0.01140.8818 ± 0.00641.4464 ± 0.00164.4658 ± 0.0995
NQSSVR(d)8.8504 ± 0.45679.4208 ± 0.331010.9936 ± 0.458812.3712 ± 1.3552
n = 800poly-SVR157.6060 ± 5.8607139.2794 ± 7.3092161.7490 ± 2.2436163.0944 ± 4.0222
rbf-SVR19.5872 ± 1.237717.9632 ± 1.259718.5810 ± 0.351419.8404 ± 1.4977
NQSSVR0.9408 ± 0.04071.2254 ± 0.05371.9072 ± 0.08326.2810 ± 0.2554
NQSSVR(d)16.5370 ± 0.535019.1804 ± 0.850823.1990 ± 1.212826.5350 ± 0.9734
n = 1000poly-SVR284.1451 ± 14.6313288.7770 ± 11.1508272.6956 ± 10.7412141.2644 ± 8.5463
rbf-SVR30.3622 ± 2.070432.5590 ± 2.496029.5694 ± 2.439530.9340 ± 9.9135
NQSSVR(p)1.4494 ± 0.19541.7290 ± 0.03102.2676 ± 0.04525.3260 ± 0.01937
NQSSVR(d)24.7908 ± 1.261428.3166 ± 1.489233.8146 ± 2.272244.1248 ± 1.5144
Table 4. Details of the benchmark datasets.
Table 4. Details of the benchmark datasets.
DatasetsAbbreviationsSample PointsAttributes
Concrete Slump TestConcrete1037
Computer HardwareComputer2099
Yacht HydrodynamicsYacht3087
Forest FiresForest51713
Energy efficiency (Heating)Energy(H)7688
Energy efficiency (Cooling)Energy(C)7688
Air qualityAir10676
Table 5. Experimental results on the datasets.
Table 5. Experimental results on the datasets.
DatasetsAlgorithmsRMSEMAET1T2
lin-SVR0.0703 ± 0.00320.0499 ± 0.00270.1618 ± 0.012224.2926
poly-SVR0.0476 ± 0.00160.0301 ± 0.00110.1708 ± 0.013049.2486
rbf-SVR0.0406 ± 0.00230.0275 ± 0.00150.1544 ± 0.0116170.5862
NNSVR0.0752 ± 0.00210.0597 ± 0.00200.0572 ± 0.00577.6194
ConcreteQLSSVR0.0571 ± 0.00220.0453 ± 0.00170.0300 ± 0.00205.0718
ϵ -SQSSVR0.0378 ± 0.00110.0242 ± 0.00140.3366 ± 0.018444.0096
NQSSVR(p)0.0381 ± 0.00160.0296 ± 0.00110.1608 ± 0.005124.0741
NQSSVR(d)0.0317 ± 0.03000.0267 ± 0.00160.1432 ± 0.002018.5869
lin-SVR0.0393 ± 0.00140.0199 ± 0.00030.5438 ± 0.030364.1097
poly-SVR0.0216 ± 0.00060.0084 ± 0.00020.5014 ± 0.0272161.7245
rbf-SVR0.0179 ± 0.00110.0093 ± 0.00050.4554 ± 0.0193507.4994
NNSVR0.0376 ± 0.00260.0217 ± 0.00090.0874 ± 0.006311.1911
ComputerQLSSVR0.0194 ± 0.00150.0099 ± 0.00040.0324 ± 0.00065.2495
ϵ -SQSSVR0.0139 ± 0.00070.0118 ± 0.00030.6674 ± 0.023074.2415
NQSSVR(p)0.0119 ± 0.00170.0077 ± 0.00070.2132 ± 0.011541.8217
NQSSVR(d)0.0098 ± 0.00120.0063 ± 0.00050.3018 ± 0.013126.7490
lin-SVR0.1428 ± 0.00100.1129 ± 0.00073.2864 ± 0.0893360.9378
poly-SVR0.0676 ± 0.02040.0559 ± 0.01053.4052 ± 0.0965798.5638
rbf-SVR0.0287 ± 0.00360.0204 ± 0.00111.6488 ± 0.0835307.3457
NNSVR0.1547 ± 0.00080.1012 ± 0.00030.1876 ± 0.018227.4442
YachtQLSSVR0.1056 ± 0.00050.0817 ± 0.00060.0620 ± 0.01108.7560
ϵ -SQSSVR0.0711 ± 0.00090.0551 ± 0.00071.8485 ± 0.105956.9315
NQSSVR(p)0.0965 ± 0.00060.0741 ± 0.00020.5432 ± 0.010567.1763
NQSSVR(d)0.0708 ± 0.00130.0533 ± 0.00101.8038 ± 0.0704167.1142
lin-SVR0.0528 ± 0.00100.0196 ± 0.000117.3862 ± 0.71611561.1308
poly-SVR0.0486 ± 0.00140.0179 ± 0.00016.8746 ± 0.06341603.4728
rbf-SVR0.0465 ± 0.00130.0170 ± 0.00003.2826 ± 0.1706812.7531
NNSVR0.0498 ± 0.00100.0198 ± 0.00010.2688 ± 0.015145.1113
ForesQLSSVR0.0500 ± 0.00090.0186 ± 0.00010.1756 ± 0.017029.1384
ϵ -SQSSVR0.0484 ± 0.00110.0196 ± 0.00010.7936 ± 0.1830696.7686
NQSSVR(p)0.0475 ± 0.00310.0183 ± 0.00010.7360 ± 0.0168120.5714
NQSSVR(d)0.0470 ± 0.00240.0175 ± 0.00019.1456 ± 0.4374540.1738
lin-SVR0.0801 ± 0.00010.0556 ± 0.000115.0628 ± 0.64601559.3228
poly-SVR0.0298 ± 0.00010.0218 ± 0.000112.9536 ± 0.31133408.8491
rbf-SVR0.0237 ± 0.00040.0208 ± 0.00035.7626 ± 0.333510,338.7745
NNSVR0.0826 ± 0.00060.0572 ± 0.00010.3374 ± 0.011552.2095
Energy(H)QLSSVR0.0704 ± 0.00030.0499 ± 0.00020.1482 ± 0.008124.4898
ϵ -SQSSVR0.0685 ± 0.00100.0464 ± 0.00049.8748 ± 0.15961065.6020
NQSSVR(p)0.0516 ± 0.00020.0405 ± 0.00020.8624 ± 0.0274105.5643
NQSSVR(d)0.0587 ± 0.00030.0423 ± 0.000116.8058 ± 0.6380679.5034
lin-SVR0.0855 ± 0.00090.0592 ± 0.001421.4692 ± 0.51192708.5711
poly-SVR0.0623 ± 0.00010.0408 ± 0.000228.3624 ± 0.89866579.0492
rbf-SVR0.0398 ± 0.00180.0276 ± 0.00089.2944 ± 0.066913,515.7641
NNSVR0.0922 ± 0.00060.0583 ± 0.00040.5539 ± 0.025988.6845
Energy(C)QLSSVR0.0820 ± 0.00040.0572 ± 0.00030.2338 ± 0.040939.7269
ϵ -SQSSVR0.0795 ± 0.00080.0512 ± 0.000514.7644 ± 0.12541782.9128
NQSSVR(p)0.0681 ± 0.00020.0476 ± 0.00011.4326 ± 0.0303175.3763
NQSSVR(d)0.0702 ± 0.00050.0463 ± 0.000210.3206 ± 0.27441161.6354
lin-SVR0.1965 ± 0.00040.1637 ± 0.000218.5452 ± 0.96603563.6484
poly-SVR0.1197 ± 0.00180.0884 ± 0.000132.7722 ± 0.70517896.3575
rbf-SVR0.1246 ± 0.00030.0727 ± 0.000212.2458 ± 0.391116,768.0504
NNSVR0.1963 ± 0.00020.1038 ± 0.00020.4592 ± 0.004177.6678
AirQLSSVR0.1346 ± 0.00020.0958 ± 0.00010.1078 ± 0.007717.8609
ϵ -SQSSVR0.1265 ± 0.00030.1033 ± 0.000320.6978 ± 0.15962066.3358
NQSSVR(p)0.1385 ± 0.00010.0936 ± 0.00020.6869 ± 0.0241122.4225
NQSSVR(d)0.1458 ± 0.00070.0965 ± 0.000616.4676 ± 0.7046677.5034
Table 6. Results on the AQCI datasets.
Table 6. Results on the AQCI datasets.
DatasetsAlgorithmsRMSEMAE R 2 T1T2
NNSVR0.0274 ± 0.00010.0244 ± 0.00121.0058 ± 0.06350.0432 ± 0.00847.1174
QLSSVR0.0783 ± 0.00110.0668 ± 0.00080.1027 ± 0.10520.0094 ± 0.00150.1103
monthly ϵ -SQSSVR0.1202 ± 0.00550.0898 ± 0.00520.9865 ± 0.10710.1356 ± 0.05341.8532
NQSSVR(p)0.0140 ± 0.00120.0115 ± 0.00111.0130 ± 0.04570.1600 ± 0.01562.3799
NQSSVR(d)0.0185 ± 0.00150.0156 ± 0.00111.0241 ± 0.06720.0482 ± 0.00720.4796
NNSVR0.0072 ± 0.00030.0060 ± 0.00011.0000 ± 0.00032.7365 ± 0.129237.1562
QLSSVR0.0071 ± 0.00010.0056 ± 0.00041.0000 ± 0.00020.1714 ± 0.01882.1832
daily ϵ -SQSSVR0.0071 ± 0.00020.0057 ± 0.00011.0001 ± 0.000132.4848 ± 0.5036405.5820
NQSSVR(p)0.0062 ± 0.00010.0054 ± 0.00011.0002 ± 0.00012.9516 ± 0.038943.7415
NQSSVR(d)0.0067 ± 0.00010.0058 ± 0.00011.0000 ± 0.00012.3086 ± 0.289728.2437
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, D.; Yang, Z.; Ye, J.; Yang, X. Kernel-Free Quadratic Surface Support Vector Regression with Non-Negative Constraints. Entropy 2023, 25, 1030. https://doi.org/10.3390/e25071030

AMA Style

Wei D, Yang Z, Ye J, Yang X. Kernel-Free Quadratic Surface Support Vector Regression with Non-Negative Constraints. Entropy. 2023; 25(7):1030. https://doi.org/10.3390/e25071030

Chicago/Turabian Style

Wei, Dong, Zhixia Yang, Junyou Ye, and Xue Yang. 2023. "Kernel-Free Quadratic Surface Support Vector Regression with Non-Negative Constraints" Entropy 25, no. 7: 1030. https://doi.org/10.3390/e25071030

APA Style

Wei, D., Yang, Z., Ye, J., & Yang, X. (2023). Kernel-Free Quadratic Surface Support Vector Regression with Non-Negative Constraints. Entropy, 25(7), 1030. https://doi.org/10.3390/e25071030

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop