1. Introduction
By mimicking the memorizing and information processing activities of neuronal networks, artificial neural networks (ANNs) give computer and information systems human-like classification and approximation capability. Such a capability is further enhanced by advances in computing and networking technologies, giving rise to advanced ANNs that grow in size and complexity. This Special Issue is intended to provide technical details of the construction and training of advanced ANNs. These details will hold great interest for researchers in computer science, artificial intelligence, soft computing, and information management, as well as for practicing engineers. This Special Issue features a balance between state-of-the-art research and practical applications. This Special Issue also provides a forum for researchers and practitioners to review and disseminate quality research work on advanced ANNs and critical issues for further development.
2. Special Issue
All submissions to this Special Issue have been reviewed by at least three experts in the ANN area. After a strict review, six papers were accepted, as introduced below.
Senthilnath et al. [
1] built a Bayesian extreme learning machine Kohonen network (BELMKN) to solve the clustering problem of nonlinearly separable datasets. The BELMKN had three layers. In the first level, the extreme learning machine (ELM)-based feature learning approach was applied to map the data onto a
d-dimensional space to capture the nonlinearity in the data distribution. Subsequently, in the second level, the Bayesian information criterion (BIC) was applied to the ELM-based feature extracted data to estimate the number of clusters. Finally, in the third level, the feature-extracted data associated with the number of clusters were processed with a Kohonen Network to optimize the clustering accuracy.
In the second paper, Tian et al. [
2] built a random vector functional link neural network (RVFLNN) to predict the temperature inside a metro station. Inputs to the RVFLNN included temperatures in different places, passenger flow, and metro arrival frequency. Various numbers of nodes in the hidden layer have been tried. In addition, the results of a parametric analysis showed that the ranges of weights and biases did not affect the minimization of the prediction error.
Qin et al. [
3] proposed a neuro-dynamic programming (NDP) method to simultaneously optimize fuel economy and battery state of charge (SOC). In the NDP approach, the critic network was a multi-resolution wavelet neural network based on the Meyer wavelet function, and the action network was a wavelet neural network based on the Morlet function. The values of weights and parameters in both networks were derived using a backward propagation algorithm.
To improve the feature recognition rate and reduce the time-cost (i.e., the time to unlock precisely) of convolutional neural networks (CNNs), Yang and Yang [
4] proposed a modified CNN approach based on dropout and the stochastic gradient descent (SGD) optimizer (MCNN-DS). The proposed methodology and three existing methods, including the weighted CNN (WCNN), multilayer perceptron CNN (MLP-CNN), and support vector machine-optimized ELM (SVM-ELM), have been applied to several benchmark cases for a comparison. The experimental results supported the superiority of the MCNN-DS approach over the existing methods.
In the fifth paper, Zhang and Tao [
5] constructed a back propagation neural network to fit the degradation model of 6205 deep groove ball bearings, so as to identify and classify the different fault states of the bearing. To further enhance the classification accuracy, the polynomial fitting principle and Pearson correlation coefficient were applied to fuse features, especially skewness and other features.
In the last paper, Cui et al. [
6] built a hybrid deep neural network that was composed of a convolutional auto-encoder and a complementary convolutional neural network to classify remote sensing data from heterogeneous sources. In the proposed methodology, the convolutional auto-encoder was used to extract features from and reduce the dimensions of the remote sensing data. Then, the convolutional neural network classified the remote sensing data based on the extracted features.