Next Article in Journal
Joint Optimization of Energy Efficiency and User Outage Using Multi-Agent Reinforcement Learning in Ultra-Dense Small Cell Networks
Previous Article in Journal
Bytecode Similarity Detection of Smart Contract across Optimization Options and Compiler Versions Based on Triplet Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on Resistant Hierarchical Fuzzy Neural Networks

1
Department of Electrical Engineering, I-Shou University, Kaohsiung 84001, Taiwan
2
School of Electronic and Mechanical Engineering, Fujian Polytechnic Normal University, Fuqing 350300, China
3
Department of Technology Management, Open University of Kaohsiung, Kaohsiung 81208, Taiwan
4
Department of Information Engineering, I-Shou University, Kaohsiung 84001, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(4), 598; https://doi.org/10.3390/electronics11040598
Submission received: 30 January 2022 / Revised: 8 February 2022 / Accepted: 13 February 2022 / Published: 15 February 2022
(This article belongs to the Topic Data Science and Knowledge Discovery)

Abstract

:
Novel resistant hierarchical fuzzy neural networks are proposed in this study and their deep learning problems are investigated. These fuzzy neural networks can be used to model complex controlled plants and can also be used as fuzzy controllers. In general, real-world data are usually contaminated by outliers. These outliers may have undesirable or unpredictable influences on the final learning machines. The correlations between the target and each of the predictors are utilized to partition input variables into groups so that each group becomes the input variables of a fuzzy system in each level of the hierarchical fuzzy neural network. In order to enhance the resistance of the learning machines, we use the least trimmed squared error as the cost function. To test the resistance of learning machines to adverse effects of outliers, we add at the output node some noise from three different types of distributions, namely, normal, Laplace, and uniform distributions. Real-world datasets are used to compare the performances of the proposed resistant hierarchical fuzzy neural networks, resistant densely connected artificial neural networks, and densely connected artificial neural networks without noise.

1. Introduction

No matter how the data are collected, the data at hand usually contain outliers. These data points are well separated from the bulk of data points or deviate from the general pattern of the data in some fashion. The outliers may have adverse or unpredictable influence on the final discriminant or predictive functions. In the past, many methods were proposed in statistical regression to address the problems with the outliers [1,2,3,4,5,6]. Regression is one of the major tasks in machine learning [7,8,9,10], and it is extensively studied using various models. Regression is applied in science education, agriculture, and signal processing [11,12,13,14]. The purpose of regression is to find the relationship between input variables and output variables in a dataset. However, the presence of noise and outliers changes the relationship. The main spirit of resistant regression is not to completely discard the outliers in the dataset, but to reduce the influence of these outliers on the final estimator. These robust regression problems were also investigated in the machine learning field [15,16,17,18,19]. The resistant regressors using the least trimmed squares (LTS) approach is particularly notable because of its simplicity and ease of use.
Fuzzy neural networks (FNNs) possess the advantages of both fuzzy systems [20,21] and neural networks [22,23], do not require accurate mathematical models, and have good learning ability which can approximate a wide range of nonlinear functions. FNNs have been widely used as machine learning models to deal with regression problems [24,25]. Next, we give a brief introduction to hierarchical fuzzy systems (HFSs) [26,27,28,29,30]. They made their debut in the control field and were devised to solve the problem of the curse of dimensionality (i.e., rule-explosion problem) in a fuzzy system. If a hierarchical fuzzy system is represented as network architecture, it is called a hierarchical fuzzy neural network (HFNN).
The main idea of HFSs is to appropriately partition input variables into groups so that each group becomes inputs of a low-dimensional fuzzy system. Each lower-dimensional fuzzy system is called a level in a hierarchical fuzzy system. Consider a simple hierarchical fuzzy neural network as shown in Figure 1. It has five inputs, a single output, and three levels. In this study, each low-dimensional fuzzy system will be replaced by an augmented fuzzy neural network (AFNN). Unlike the standard fuzzy network which has only a single hidden layer, an AFNN can have many hidden layers so that deep learning techniques can readily be applied to machine learning problems. Moreover, to enhance the representing power of the HFNN, each low-dimensional fuzzy system may have more than one output.
To the best of our knowledge, hierarchical fuzzy neural networks have never been used as resistant learning machines in past research. Furthermore, there have only been a few studies on traditional hierarchical fuzzy systems [28,29,30]. The authors proved the universal approximation property of the hierarchical fuzzy systems. Roughly speaking, this means that given a function g with some kind of smoothness property and a positive constant ϵ , no matter how small, there is a hierarchical fuzzy system f (treated as a crisp nonlinear map), with the number of fuzzy rules unrestricted, such that the maximum absolute difference between g and f on a given compact set is less than ϵ . Of course, this is an important property for any class of learning machines in successful application to real problems. It should be noted that the low-dimensional fuzzy system in each level is a Takagi–Sugeno–Kang (TSK) fuzzy system and the THEN parts of the fuzzy rules are all crisp polynomial functions but not the usual linear functions in TSK models. In [30], the authors developed a class of special hierarchical fuzzy systems where the outputs of the previous layer are not used in the IF parts, but used only in the THEN parts of the fuzzy rules of the current layer. This class of special hierarchical fuzzy systems also has the universal approximation property. The goal of this study is to investigate the performance of the proposed resistant hierarchical fuzzy neural networks against the adverse effect induced by the outliers.

2. Fuzzy Neural Networks

Consider a standard fuzzy system with n inputs and p outputs as shown in Figure 1 in [19]. Suppose the m fuzzy rules in the rule base are given in canonical form as
R j : IF x 1 is A 1 j and x 2 is A 2 j and … and x n is A n j ,
 THEN y 1 is B j 1 and y 2 is B j 2 and … and y p is B j p ,
where j m ̲ = 1 , 2 , . . . , m . The input–output mapping of the fuzzy system with a singleton fuzzifier, product inference engine, center-average defuzzifier, and Gaussian membership functions for the fuzzy sets A i j can be written as [19]:
y k = j = 1 m w j k exp i = 1 n x i c i j 2 x i c i j 2 σ i j 2 σ i j 2 j = 1 m exp i = 1 n x i c i j 2 x i c i j 2 σ i j 2 σ i j 2 , k p ̲ , x = x 1 . . . x n T n ,
where, for i n ̲ , j m ̲ , and k p ̲ ,
w j k : center of the normal fuzzy set B j k
c i j : center of the Gaussian fuzzy set A i j
σ i j 2 : “variance” of the Gaussian fuzzy set A i j
In order to escape from the division by zero in (1), to satisfy the constraints σ i j 2 > 0 , and to allow the learning machine to handle broader types of data, a better parameterization is given by
y k = f o k x = f o k j = 1 m w j k exp j = 1 m x i c i j 2 exp v i j j = 1 m exp i = 1 n x i c i j 2 exp v i j , k p ̲ ,
where v i j = log 1 1 σ i j 2 σ i j 2 and f o k is the activation function of the kth output node.
For i n ̲ , j m ̲ , and k p ̲ , we define
u j = i = 1 n x i c i j 2 exp v i j , r j = exp u j ,
s k = j = 1 m w j k r j , g = j = 1 m r j ,
then
y k = f o k s k s k g g .
According to (3), the fuzzy system given in (2) can now be represented as a feedforward neural network [19], called a fuzzy neural network (FNN), as shown schematically in Figure 2. Note that there is only a single hidden layer in this neural network.
As mentioned earlier, to enhance the learning capability or predictive power of the proposed hierarchical fuzzy neural network model, it would be a good idea to allow many hidden layers in a fuzzy neural network. In that case, we call it an augmented fuzzy neural network (AFNN), which is shown schematically in Figure 3. In the hierarchical fuzzy neural networks studied here, each low-dimensional fuzzy system will be replaced by an AFNN.

3. Loss Function

There are many choices for the loss functions in the resistant machine learning problem. In the following development, we will use the least trimmed squares (LTS) loss function. The methodology by using other loss functions is similar and will not be repeated here. Now, we briefly review the LTS loss function.
Let X n and Y . Suppose the training data are given by
S : = x q , d q q = 1 l X × Y
.
Let y q , q l ̲ , denote the predicted response value corresponding to the predictor x q . Then, its residual is defined by e q = d q y q . The LTS approach to resistant regression is to search for the connection weights c i j , v i j , w j k of the neural network so as to minimize the following loss function:
J l t s = q = 1 l a R e q 2 e q 2 ,
where the score function a · (i.e., penalizing weight) is defined by
a q : = 1 , 1 q l * , 0 , l * < q l ,
and R e q 2 is the rank (from the smallest to the largest) of e q 2 in e 1 2 , …, e l 2 . We call h = l l * l l * l l the trimming percentage. It is the most important design parameter in LTS regression.

4. Partition of Input Variables

The biggest problem in using a hierarchical fuzzy neural network as a learning machine is how to appropriately partition the input variables so that each group becomes inputs of a low-dimensional fuzzy system. After several trials, we found that the correlation between the target and each of the predictors, i.e., correlation coefficient, may reliably be used to determine the input variables for the fuzzy neural network in each level of the hierarchical fuzzy neural network under consideration. Once this is done, the following rules of thumb may then be applied:
Rule 1:
Correlation is measured in absolute value (i.e., magnitude).
Rule 2:
Put the more correlated predictors in our hierarchical fuzzy neural network as early as possible.
Rule 3:
Variables with low correlation will not be used as predictors.
Rule 4:
Predictors with about the same level of correlation may be collected in the same group.

5. Illustrative Examples

To better evaluate the performances of the resistant hierarchical fuzzy neural network, the k-fold cross validation, one of the commonly used cross validation techniques, will be employed. First, the given data are randomly partitioned into k parts which are roughly equal in size. In each run of the method, one part is left as a validating set with the remaining used as the training set. After training with the training set, the pre-specified performance indicators of the validating set are computed. Those k values of the performance indicators from the k runs are then averaged to give estimates of the performance indicators. In this study, we use 10-fold cross validation (i.e., k = 10 ) and we take the median to average the values of the performance indicators. The numerical variables in the dataset are standardized before cross validation.
To test the robustness or resistance of the learning machine, we will add three kinds of noise to the output of the training data. These are random samples drawn from the following probability distributions:
(1)
normal distribution: μ = 0 , σ = 1 .
f x = 1 σ 2 π exp 1 2 x μ σ 2 , < x < .
(2)
Laplace distribution: μ = 0 , b = 1 .
f x = 1 2 b exp x μ b , < x < ( b > 0 ).
(3)
Uniform distribution: a = 1 , b = 1 .
f x = 1 1 b a b a , a x b , 0 , o t h e r w i s e .
The noise will not be added to the validating set.
In the following simulations using LTS loss function, we set the trimming percentage to be 0.2.
One natural question is that can HFNNs compete with the state-of-the-art artificial neural networks, e.g., densely connected artificial neural networks (DNNs)? Ten real-world datasets described in Table 1 are used in this study to compare the performances of the LTS-HFNNs, LTS-DNNs, and DNNs without noise. These datasets can be loaded from the UCI machine learning repository of the University of California at Irvine [31], LIBSVM of National Taiwan University [32], and Kaggle [33].
For example, the first dataset shown in Table 1 is “Airfoil Self-Noise” from NASA in the benchmark UCI. It is obtained from a series of aerodynamic and acoustic tests of two- and three-dimensional airfoil blade sections conducted in an anechoic wind tunnel. The number of instances (records) is 1503 and the number of attributes is 6. The five input predictors are frequency, angle of attack, chord length, free-stream velocity, and suction side displacement thickness. The only output is the scaled sound pressure level. In other words, we are performing a 5-input/1-output regression task for this dataset. All other datasets can be described similarly.
Next, we list the architecture of the HFNNs and DNNs, as shown in Table 2. For fair comparison, the DNN in each of the following simulations is configured such that the number of adjustable parameters of the DNN is about the same as that of the corresponding HFNN. The integers in the parentheses of the second and third columns are the number of adjustable parameters of the neural networks.
Example 1.
First, we consider adding the output noise from the normal distribution. Table 3 shows the 10-fold cross validated simulation results, where the values in parentheses are the standard errors (i.e., standard deviations) of the performance indicators. In the simulations, three performance indicators are used. In Table 3, “LOSS” denotes the value of the LTS loss function, “MSE” is the mean squared error, and “MAE” is the mean absolute error. In the last column of the table, we list the simulation results for DNNs without noise, where the loss function is the same as MSE. Those values serve as goal values for LTS-HFNNs and LTS-DNNs. As observed in Table 3, for the first dataset “Airfoil self-noise”, the MSE of LTS-HFNN is 0.2970 which is roughly the same as that of LTS-DNN, 0.2916. Both are higher than that of DNN, 0.1039, since in this method there are no Gaussian noises added on the output variable. One important statistic is the standard error of the 10-fold cross validation. Those values are 0.0600, 0.3971, and 0.4019. Smaller standard errors reveal more stable estimations. In fact, except for the “Boston housing” data, the proposed LTS-HFNNs have smaller standard errors for the remaining 9 datasets. Roughly speaking, the proposed LTS-HFNNs usually have slightly higher mean values of MSE and MAE, but smaller standard errors.
Example 2.
Next, we consider adding the output noise from the Laplace distribution. Table 4 shows the 10-fold cross validated simulation results. As observed in Table 4, the LOSS and MSE of LTS-HFNN for 4 out of 10 datasets are slightly smaller than for LTS-DNN, and for the remaining 6 datasets they are slightly higher than for LTS-DNN. The MAE of LTS-HFNN for 3 out of 10 datasets is smaller than for LTS-DNN, and for the remaining 7 datasets it is higher than for LTS-DNN. For all LOSS, MSE, and MAE values, the standard errors of LTS-HFNN are obviously much smaller than for LTS-DNN for all 10 datasets, except for MAE in “Boston housing”. In general, LTS-DNNs have slightly smaller mean values of MSE and MAE, but larger standard errors.
Example 3.
Finally, we consider adding the output noise from the uniform distribution. Table 5 shows the 10-fold cross validated simulation results. As observed in Table 5, the LOSS of LTS-HFNN for 6 out of 10 datasets are smaller than for LTS-DNN, and for the remaining 4 datasets it is higher than for LTS-DNN. The MSE and MAE of LTS-HFNN for 8 out of 10 datasets are smaller than for LTS-DNN, and for the remaining 2 datasets they are higher than for LTS-DNN. As in the previous two examples, LTS-FNNs usually have slightly higher mean values of MSE and MAE, but smaller standard errors.
In the above three examples, we have shown the robustness of the proposed FNNs through adding various types of noise to the datasets. In the simulations, we conducted cross validations to calculate the means and standard deviations of the testing errors as the performance indicators. As observed, the proposed LTS-HFNN and LTS-DNN both exhibited very good robustness.

6. Conclusions

In this study, we proposed resistant hierarchical fuzzy neural networks and investigated the associated deep learning problems. Correlations between the target and each of the predictors, together with some rules of thumb, were utilized to partition input variables into groups so that each group becomes the input variables of a fuzzy system in each level of the hierarchical fuzzy neural network. The least trimmed squared error was used as the cost function in order to enhance the resistance of the hierarchical fuzzy neural networks. Three different types of noise were added to the output node of the neural network to test the resistance of the neural networks. From the simulation results, it seems that the performance has very little connection with the types of the noise. The proposed LTS-HFNN was compared with LTS-DNN, and DNN without noises was also taken for comparison. Ten datasets were utilized for regression. Experimental results shows that the added noises for LTS-HFNN and LTS-DNN increase the loss. Furthermore, LTS-FNNs usually have slightly higher mean values of MSE and MAE, but smaller standard errors. Smaller standard errors correspond to stronger robustness of the models. Sacrificing a small accuracy for much stronger robustness is worth the cost when real-world datasets have a certain number of outliers. Because of the proposed hierarchical structure together with the augmented fuzzy neural units and the least trimmed squared error, robust fuzzy neural networks are obtained in our works.
The most important future work of hierarchical fuzzy neural networks is to develop more powerful techniques to partition the input variables. One immediate technique we can think of is the decision tree. It may help us to appropriately partition the input variables in a particular way. Moreover, the input variables used in earlier levels can also be used in later levels. Another interesting topic is to investigate the resistance of hierarchical fuzzy neural networks against the x-space outliers. These x-space outliers can be generated by adding various types of noises in the input variables.

Author Contributions

Conceptualization, J.-G.H.; methodology, J.-G.H. and F.G.; validation, F.G. and J.-H.J.; formal analysis, J.-G.H. and J.-H.J.; writing—original draft preparation, Y.-S.K.; writing—review and editing, F.G., J.-G.H., J.-H.J. and Y.-S.K.; visualization, F.G.; supervision, J.-H.J. and J.-G.H.; project administration, J.-G.H. and J.-H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors acknowledge the Ministry of Science and Technology, Taiwan, through grant number MOST-109-2221-E-214-013 and MOST 110-2221-E-214-019.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

p ̲ p ̲ : = 1 , 2 , , p
n n-dimensional real space
X × Y Cartesian product of sets X and Y

References

  1. Knorr, E.M.; Ng, R.T.; Tucakov, V. Distance-based outliers: Algorithms and applications. VLDB J. 2000, 8, 237–253. [Google Scholar] [CrossRef]
  2. Leys, C.; Ley, C.; Klein, O.; Bernard, P.; Licata, L. Detecting outliers: Do not use standard deviation around mean, use absolute deviation around the median. J. Exp. Soc. Psychol. 2013, 49, 764–766. [Google Scholar] [CrossRef] [Green Version]
  3. Aguinis, H.; Gottfredson, R.K.; Joo, H. Best-practice recommendations for defining, identifying, and handling outliers. Organ. Res. Methods 2013, 16, 270–301. [Google Scholar] [CrossRef]
  4. Alfons, A.; Croux, C.; Gelper, S. Sparse least trimmed squares regression for analyzing high-dimensional large data sets. Ann. Appl. Stat. 2013, 7, 226–248. [Google Scholar] [CrossRef]
  5. Febrero-Bande, M.; Galeano, P.; Gonzále-Manteiga, W. Functional principal component regression and functional partial least-squares regression: An overview and a comparative study. Int. Stat. Rev. 2017, 85, 61–83. [Google Scholar] [CrossRef]
  6. Lecué, G.; Lerasle, M. Robust machine learning by median-of-means: Theory and practice. Ann. Stat. 2017, 48, 906–931. [Google Scholar] [CrossRef]
  7. Abonazel, M.R.; Saber, O.M. A comparative study of robust estimators for Poisson regression model with outliers. J. Stat. Appl. Probab 2020, 9, 279–286. [Google Scholar] [CrossRef]
  8. Lathuilière, S.; Mesejo, P.; Alameda-Pineda, X.; Horaud, R. Deepgum: Learning deep robust regression with a gaussian-uniform mixture model. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 202–217. [Google Scholar]
  9. Klivans, A.; Kothari, P.K.; Meka, R. Efficient algorithms for outlier-robust regression. In Proceedings of the Conference On Learning Theory, Stockholm, Sweden, 6–9 July 2018; pp. 1420–1430. [Google Scholar]
  10. Schmidt, A.F.; Chris, F. Linear regression and the normality assumption. J. Clin. Epidemiol. 2018, 98, 146–151. [Google Scholar] [CrossRef] [Green Version]
  11. Thompson, E.D.; Bowling, B.V.; Markle, R.E. Predicting student success in a major’s introductory biology course via logistic regression analysis of scientific reasoning ability and mathematics scores. Res. Sci. Educ. 2018, 48, 151–163. [Google Scholar] [CrossRef]
  12. Mahaboob, B.; Praveen, J.P.; Rao, B.A.; Haranadh, Y.; Narayana, C.; Prakash, G.B. A study on multiple linear regression using matrix calculus. Adv. Math. Sci. J. 2020, 9, 4863–4872. [Google Scholar] [CrossRef]
  13. Steward, P.R.; Dougill, A.J.; Thierfelder, C.; Pittelkow, C.M.; Stringer, L.C.; Kudzala, M.; Shackelford, G.E. The adaptive capacity of maize-based conservation agriculture systems to climate stress in tropical and subtropical environments: A meta-regression of yields. Agric. Ecosyst. Environ. 2018, 251, 194–202. [Google Scholar] [CrossRef]
  14. Aspuru, J.; Ochoa-Brust, A.; Félix, R.A.; Mata-López, W.; Mena, L.J.; Ostos, R.; Martínez-Peláez, R. Segmentation of the ECG signal by means of a linear regression algorithm. Sensors 2019, 19, 775. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Hsieh, J.-G.; Lin, Y.-L.; Jeng, J.-H. Preliminary study on Wilcoxon learning machines. IEEE Trans. Neural Netw. 2008, 19, 201–211. [Google Scholar] [CrossRef] [PubMed]
  16. Rusiecki, A. Robust learning algorithm based on LTA estimator. Neurocomputing 2008, 120, 624–632. [Google Scholar] [CrossRef]
  17. Lin, Y.-L.; Hsieh, J.-G.; Jeng, J.-H.; Cheng, W.-C. On least trimmed squares neural networks. Neurocomputing 2015, 161, 107–112. [Google Scholar] [CrossRef]
  18. Rusiecki, A. Trimmed categorical cross-entropy for deep learning with label noise. Electron. Lett. 2019, 55, 319–320. [Google Scholar] [CrossRef]
  19. Hsieh, J.-G.; Jeng, J.-H.; Lin, Y.-H.; Kuo, Y.-S. Single index fuzzy neural networks using locally weighted polynomial regression. Fuzzy Sets and Systems. Fuzzy Sets Syst. 2019, 368, 82–100. [Google Scholar] [CrossRef]
  20. Deng, Z.; Jiang, Y.; Choi, K.S.; Chung, F.L.; Wang, S. Knowledge-leverage-based TSK fuzzy system modeling. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 1200–1212. [Google Scholar] [CrossRef]
  21. Shihabudheen, K.V.; Pillai, G.N. Recent advances in neuro-fuzzy system: A survey. Knowl.-Based Syst. 2018, 152, 136–162. [Google Scholar] [CrossRef]
  22. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018, 4, e00938. [Google Scholar] [CrossRef] [Green Version]
  23. Khodabandelou, G.; Ebadzadeh, M.M. Fuzzy neural network with support vector-based learning for classification and regression. Soft Comput. 2019, 23, 12153–12168. [Google Scholar] [CrossRef]
  24. Tang, J.; Liu, F.; Zhang, W.; Ke, R.; Zou, Y. Lane-changes prediction based on adaptive fuzzy neural network. Expert Syst. Appl. 2018, 91, 452–463. [Google Scholar] [CrossRef]
  25. Khashei, M.; Hamadani, A.Z.; Bijari, M. A novel hybrid classification model of artificial neural networks and multiple linear regression models. Expert Syst. Appl. 2012, 39, 2606–2620. [Google Scholar] [CrossRef]
  26. Zheng, Y.; Jeon, B.; Xu, D. Image segmentation by generalized hierarchical fuzzy C-means algorithm. J. Intell. Fuzzy Syst. 2015, 28, 961–973. [Google Scholar] [CrossRef]
  27. Deng, Y.; Ren, Z.; Kong, Y.; Bao, F.; Dai, Q. A hierarchical fused fuzzy deep neural network for data classification. IEEE Trans. Fuzzy Syst. 2015, 25, 1006–1012. [Google Scholar] [CrossRef]
  28. Wang, L.X. Universal approximation by hierarchical fuzzy systems. Fuzzy Sets Syst. 1998, 93, 223–230. [Google Scholar] [CrossRef]
  29. Chen, W.; Wang, L.X. A note on universal approximation by hierarchical fuzzy systems. Inf. Sci. 2000, 123, 241–248. [Google Scholar] [CrossRef]
  30. Joo, M.G.; Lee, J.S. Universal approximation by hierarchical fuzzy system with constraints on the fuzzy rule. Fuzzy Sets Syst. 2002, 130, 175–188. [Google Scholar] [CrossRef]
  31. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/index.php (accessed on 26 October 2020).
  32. LIBSVM: A Library for Support Vector Machines. Available online: http://www.csie.ntu.edu.tw/~cjlin/libsvm (accessed on 5 November 2020).
  33. Kaggle. Available online: https://www.kaggle.com/datasets (accessed on 12 November 2018).
Figure 1. A simple hierarchical fuzzy neural network.
Figure 1. A simple hierarchical fuzzy neural network.
Electronics 11 00598 g001
Figure 2. Fuzzy neural network with a single hidden layer.
Figure 2. Fuzzy neural network with a single hidden layer.
Electronics 11 00598 g002
Figure 3. Augmented fuzzy neural network.
Figure 3. Augmented fuzzy neural network.
Electronics 11 00598 g003
Table 1. Description of the datasets.
Table 1. Description of the datasets.
DatasetNo. of CasesNo. of PredictorsSource
Airfoil self-noise15035UCI
Boston housing50613Kaggle
Combined cycle power plant95684UCI
Concrete compressive strength10308UCI
Cpusmall819212LIBSVM
Mg13856LIBSVM
Parkinsons telemonitoring (motor UPDRS)587516UCI
Parkinsons telemonitoring (total UPDRS)587516UCI
QSAR fish toxicity9086UCI
Space-GA31076LIBSVM
Table 2. Architecture of HFNNs and DNNs.
Table 2. Architecture of HFNNs and DNNs.
DatasetHFNNDNN
Airfoil self-noise(1,4,3,2)
(4,4,3,2)
(4,4,3,1)
(159)
5-10-6-4-1
(159)
Boston housing(3,4,3,2)
(5,4,3,2)
(6,4,3,1)
(5,4,3,1)
(269)
13-10-7-5-1
(263)
Combined cycle power plant(1,4,3,2)
(3,4,3,2)
(4,4,3,1)
(151)
4-8-8-4-1
(153)
Concrete compressive strength(2,4,3,2)
(5,4,3,2)
(5,4,3,1)
(183)
8-8-7-5-1
(181)
Cpusmall(2,4,3,2)
(6,4,3,2)
(6,4,3,2)
(4,4,3,1)
(261)
12-10-8-4-1
(259)
Mg(2,4,3,2)
(4,4,3,2)
(4,4,3,1)
(167)
6-8-7-5-1
(165)
Parkinsons telemonitoring (motor UPDRS)(2,4,3,2)
(4,4,3,2)
(5,4,3,1)
(175)
16-6-5-5-1
(173)
Parkinsons telemonitoring (total UPDRS)(1,4,3,2)
(4,4,3,2)
(4,4,3,1)
(159)
16-6-5-3-1
(159)
QSAR fish toxicity(1,4,3,2)
(4,4,3,2)
(3,4,3,2)
(4,4,3,1)
(213)
6-9-8-7-1
(214)
Space-GA(1,4,3,2)
(3,4,3,2)
(5,4,3,2)
(3,4,3,1)
(213)
6-9-8-7-1
(214)
Table 3. Simulation results in Example 1.
Table 3. Simulation results in Example 1.
LTS-HFNNLTS-DNNDNN without Noise
Airfoil self-noise
LOSS0.0919 (±0.0201)0.0901 (±0.1667)
MSE0.2970 (±0.0600)0.2916 (±0.3971)0.1039 (±0.4019)
MAE0.4118 (±0.0410)0.4022 (±0.2255)0.2374 (±0.2563)
Boston housing
LOSS0.3005 (±0.0691)0.2220 (±0.0500)
MSE0.9877 (±0.3232)0.5845 (±0.2183)0.1017 (±0.0981)
MAE0.7745 (±0.0841)0.5928 (±0.0817)0.2359 (±0.0581)
Combined cycle power plant
LOSS0.0260 (±0.0036)0.0275 (±0.1481)
MSE0.0734 (±0.0107)0.0744 (±0.2897)0.0570 (±0.4011)
MAE0.2133 (±0.0124)0.2144 (±0.2031)0.1850 (±0.2910)
Concrete compressive strength
LOSS0.1481 (±0.0352)0.1947 (±0.1487)
MSE0.3964 (±0.0743)0.4904 (±0.3362)0.1393 (±0.2216)
MAE0.4986 (±0.0478)0.5574 (±0.1814)0.2746 (±0.1489)
Cpusmall
LOSS0.0330 (±0.0048)0.0327 (±0.0248)
MSE0.8737 (±0.1438)0.1296 (±0.4685)0.0253 (±0.0035)
MAE0.3928 (±0.0343)0.2518 (±0.1383)0.1141 (±0.0062)
Mg
LOSS0.1538 (±0.0211)0.1507 (±0.1192)
MSE0.4595 (±0.0747)0.4139 (±0.2331)0.2920 (±0.2131)
MAE0.5262 (±0.0435)0.5047 (±0.1335)0.4157 (±0.1297)
Parkinsons telemonitoring (motor UPDRS)
LOSS0.3616 (±0.0319)0.4102 (±0.0979)
MSE0.9389 (±0.0684)0.9262 (±0.1149)0.7059 (±0.0996)
MAE0.7757 (±0.0329)0.7952 (±0.0801)0.6746 (±0.0637)
Parkinsons telemonitoring (total UPDRS)
LOSS0.3333 (±0.0404)0.3729 (±0.0578)
MSE0.9158 (±0.0686)0.9454 (±0.0937)0.7011 (±0.1500)
MAE0.7536 (±0.0393)0.7744 (±0.0501)0.6689 (±0.0728)
QSAR fish toxicity
LOSS0.2175 (±0.0419)0.2016 (±0.0671)
MSE0.6589 (±0.1362)0.6178 (±0.2001)0.4719 (±0.1418)
MAE0.6108 (±0.0463)0.5904 (±0.0793)0.4779 (±0.0923)
Space-GA
LOSS0.1474 (±0.0276)0.1145 (±0.0623)
MSE0.5002 (±0.1766)0.3512 (±0.1806)0.2716 (±0.2544)
MAE0.5074 (±0.0478)0.4469 (±0.0890)0.3911 (±0.1295)
Table 4. Simulation results in Example 2.
Table 4. Simulation results in Example 2.
LTS-HFNNLTS-DNNDNN without Noise
Airfoil self-noise
LOSS0.1096 (±0.0273)0.2661 (±0.1869)
MSE0.3451 (±0.0621)0.6195 (±0.4015)0.1039 (±0.4019)
MAE0.4478 (±0.0478)0.6258 (±0.2294)0.2374 (±0.2563)
Boston housing
LOSS0.3600 (±0.1021)0.2086 (±0.0765)
MSE1.0135 (±0.2562)0.8298 (±0.3050)0.1017 (±0.0981)
MAE0.7595 (±0.0897)0.6845 (±0.0892)0.2359 (±0.0581)
Combined cycle power plant
LOSS0.0247 (±0.0016)0.0265 (±0.1501)
MSE0.0703 (±0.0057)0.0711 (±0.2893)0.0570 (±0.4011)
MAE0.2041 (±0.0061)0.2103 (±0.2067)0.1850 (±0.2910)
Concrete compressive strength
LOSS0.1658 (±0.0425)0.1507 (±0.1461)
MSE0.4358 (±0.0972)0.4265 (±0.3083)0.1393 (±0.2216)
MAE0.5256 (±0.0569)0.5078 (±0.1754)0.2746 (±0.1489)
Cpusmall
LOSS0.0205 (±0.0063)0.0217 (±0.0279)
MSE0.4521 (±0.2958)0.0929 (±0.4073)0.0253 (±0.0035)
MAE0.2949 (±0.0819)0.2052 (±0.1373)0.1141 (±0.0062)
Mg
LOSS0.1449 (±0.0367)0.1374 (±0.1370)
MSE0.4620 (±0.0654)0.3965 (±0.2542)0.2920 (±0.2131)
MAE0.5013 (±0.0443)0.4823 (±0.1512)0.4157 (±0.1297)
Parkinsons telemonitoring (motor UPDRS)
LOSS0.3612 (±0.0380)0.3407 (±0.0875)
MSE0.9004 (±0.0912)0.8740 (±0.1118)0.7059 (±0.0996)
MAE0.7605 (±0.0387)0.7441 (±0.0763)0.6746 (±0.0637)
Parkinsons telemonitoring (total UPDRS)
LOSS0.3287 (±0.0300)0.3946 (±0.0578)
MSE0.8618 (±0.0619)0.9758 (±0.0921)0.7011 (±0.1500)
MAE0.7409 (±0.0305)0.7943 (±0.0489)0.6689 (±0.0728)
QSAR fish toxicity
LOSS0.2311 (±0.0651)0.1886 (±0.1202)
MSE0.7317 (±0.1558)0.5801 (±0.2875)0.4719 (±0.1418)
MAE0.6655 (±0.0762)0.5735 (±0.1287)0.4779 (±0.0923)
Space-GA
LOSS0.1379 (±0.0077)0.1069 (±0.0862)
MSE0.4689 (±0.1199)0.3195 (±0.2381)0.2716 (±0.2544)
MAE0.5055 (±0.0152)0.4410 (±0.1197)0.3911 (±0.1295)
Table 5. Simulation results in Example 3.
Table 5. Simulation results in Example 3.
LTS-HFNNLTS-DNNDNN without Noise
Airfoil self-noise
LOSS0.0878 (±0.0127)0.0801 (±0.1175)
MSE0.2787 (±0.0543)0.2720 (±0.2748)0.1039 (±0.4019)
MAE0.3961 (±0.0273)0.3829 (±0.1545)0.2374 (±0.2563)
Boston housing
LOSS0.1799 (±0.0378)0.1248 (±0.0521)
MSE0.6684 (±0.2669)0.4135 (±0.1432)0.1017 (±0.0981)
MAE0.5968 (±0.0777)0.4861 (±0.0773)0.2359 (±0.0581)
Combined cycle power plant
LOSS0.0271 (±0.0019)0.0295 (±0.1466)
MSE0.0761 (±0.0063)0.0780 (±0.3029)0.0570 (±0.4011)
MAE0.2155 (±0.0072)0.2185 (±0.2044)0.1850 (±0.2910)
Concrete compressive strength
LOSS0.1160 (±0.0262)0.0897 (±0.1117)
MSE0.3350 (±0.0706)0.2479 (±0.2392)0.1393 (±0.2216)
MAE0.4414 (±0.0468)0.3816 (±0.1371)0.2746 (±0.1489)
Cpusmall
LOSS0.0267 (±0.0062)0.0311 (±0.0157)
MSE0.8325 (±0.0999)0.1182 (±0.3180)0.0253 (±0.0035)
MAE0.3724 (±0.0357)0.2518 (±0.0899)0.1141 (±0.0062)
Mg
LOSS0.1213 (±0.0249)0.1249 (±0.1376)
MSE0.3826 (±0.0528)0.3638 (±0.2730)0.2920 (±0.2131)
MAE0.4662 (±0.0338)0.4643 (±0.1566)0.4157 (±0.1297)
Parkinsons telemonitoring (motor UPDRS)
LOSS0.3412 (±0.0275)0.2980 (±0.0684)
MSE0.9623 (±0.0748)0.8299 (±0.0921)0.7059 (±0.0996)
MAE0.7576 (±0.0291)0.7097 (±0.0522)0.6746 (±0.0637)
Parkinsons telemonitoring (total UPDRS)
LOSS0.3184 (±0.0318)0.3892 (±0.0752)
MSE0.9121 (±0.0766)0.9835 (±0.1188)0.7011 (±0.1500)
MAE0.7326 (±0.0324)0.7991 (±0.0644)0.6689 (±0.0728)
QSAR fish toxicity
LOSS0.1507 (±0.0441)0.1246 (±0.0489)
MSE0.4986 (±0.1436)0.4540 (±0.1513)0.4719 (±0.1418)
MAE0.5347 (±0.0696)0.5097 (±0.0838)0.4779 (±0.0923)
Space-GA
LOSS0.1224 (±0.0073)0.0980 (±0.0114)
MSE0.4284 (±0.0993)0.2986 (±0.0790)0.2716 (±0.2544)
MAE0.4792 (±0.0176)0.4196 (±0.0284)0.3911 (±0.1295)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, F.; Hsieh, J.-G.; Kuo, Y.-S.; Jeng, J.-H. Study on Resistant Hierarchical Fuzzy Neural Networks. Electronics 2022, 11, 598. https://doi.org/10.3390/electronics11040598

AMA Style

Gao F, Hsieh J-G, Kuo Y-S, Jeng J-H. Study on Resistant Hierarchical Fuzzy Neural Networks. Electronics. 2022; 11(4):598. https://doi.org/10.3390/electronics11040598

Chicago/Turabian Style

Gao, Fengyu, Jer-Guang Hsieh, Ying-Sheng Kuo, and Jyh-Horng Jeng. 2022. "Study on Resistant Hierarchical Fuzzy Neural Networks" Electronics 11, no. 4: 598. https://doi.org/10.3390/electronics11040598

APA Style

Gao, F., Hsieh, J. -G., Kuo, Y. -S., & Jeng, J. -H. (2022). Study on Resistant Hierarchical Fuzzy Neural Networks. Electronics, 11(4), 598. https://doi.org/10.3390/electronics11040598

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop