Next Article in Journal
An Ensemble Model for Consumer Emotion Prediction Using EEG Signals for Neuromarketing Applications
Next Article in Special Issue
Investigation of Phase Shifts Using AUC Diagrams: Application to Differential Diagnosis of Parkinson’s Disease and Essential Tremor
Previous Article in Journal
Effect of Source Mispositioning on the Repeatability of 4D Vertical Seismic Profiling Acquired with Distributed Acoustic Sensors
Previous Article in Special Issue
AI-Based Detection of Aspiration for Video-Endoscopy with Visual Aids in Meaningful Frames to Interpret the Model Outcome
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Structure and Base Analysis of Receptive Field Neural Networks in a Character Recognition Task

Faculty of Electrical Engineering and Information Technology, Slovak University of Technology, Ilkovicova 3, 812 19 Bratislava, Slovakia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(24), 9743; https://doi.org/10.3390/s22249743
Submission received: 27 October 2022 / Revised: 5 December 2022 / Accepted: 8 December 2022 / Published: 12 December 2022

Abstract

:
This paper explores extensions and restrictions of shallow convolutional neural networks with fixed kernels trained with a limited number of training samples. We extend the work recently done in research on Receptive Field Neural Networks (RFNN) and show their behaviour using different bases and step-by-step changes within the network architecture. To ensure the reproducibility of the results, we simplified the baseline RFNN architecture to a single-layer CNN network and introduced a deterministic methodology for RFNN training and evaluation. This methodology enabled us to evaluate the significance of changes using the (recently widely used in neural networks) Bayesian comparison. The results indicate that a change in the base may have less of an effect on the results than re-training using another seed. We show that the simplified network with tested bases has similar performance to the chosen baseline RFNN architecture. The data also show the positive impact of energy normalization of used filters, which improves the classification accuracy, even when using randomly initialized filters.

1. Introduction

Convolutional neural networks bring, along with considerable advantages, a number of challenging problems. One of them is the choice of hyperparameters for the individual layers. From previous research, we know the effects of the activation function, depth of the network, different types of normalization, preprocessing, and optimization methods on the overall accuracy achieved [1]. The open problem is still the choice of convolutional kernel sizes and their numbers in the individual layers. The popular answer is a minimum symmetric kernel of size 3 × 3 pixels [2], which also contains a small number of parameters. However, this solution is memory-intensive in terms of calculating convolution with high resolution input images, and, therefore, the convolutional kernel can be enlarged with respect to physical constraints [3]. Another reason for changing the kernel size is that it directly affects the size of the receptive field, as the larger filter takes into account the larger surroundings and captures higher-level image features. However, enlarging the size of a kernel quadratically increases the number of its possible parameters and may not improve network accuracy due to generalization problems [4,5]. Moreover, rapid reduction of dimensionality by convolution with a larger kernel leads to shallower architectures. The practical solution is the use of a grid search [6], as there is no general solution yet, and the architectural design of these networks requires some experience.
Another not less-important problem of classic deep networks is their representative complexity, where it is not entirely clear what transformation the network performs and what the individual features mean. Therefore, some authors also believe the direction that research should take is to create simpler rather than deeper models [7,8]. This trend of architecture modifications towards parameter reduction and simplification has become an active area of research in recent years. In this case, it is important to maintain the ability of networks to produce representative features with an emphasis on achieved accuracy.
Last but not least, the need for a large amount of data for the training process itself significantly limits the wider deployment of deep learning. Although there are ways to achieve good results on smaller datasets, such as using transfer learning and expanding the training set by image augmentation [9], these techniques should be implemented with caution. The use of transfer learning is, in many cases, application-specific [10], and image augmentation techniques can create small image artefacts that must be considered in critical applications such as medical diagnostics [11].

1.1. Simplification of Convolutional Neural Networks

With regard to the success of deep neural networks, the main motivations for their simplification are deployment on devices with limited computing resources, as well as a deeper understanding and easier interpretation of the created models. One of the major simplifications was tested by Iandola et al. [12], who proposed the SqeezeNet architecture. In addition to the introduction of Fire Modules that reduce the dimensionality of data in the first step, followed by an expansion phase, the same authors also introduced complex bypass connections. These bridging connections are similar to those in the ResNet [13] architecture but with an introduction of the point-wise convolution operation. The overall architecture has 50-fold fewer parameters while retaining similar accuracy as the AlexNet [3] architecture. Another point of interest is that the fully connected layers were not used in the Sqeezenet architecture intended for classification. Instead of the common approach, one feature map for each corresponding category of the classification task was generated. On top of these feature maps, the average of each feature map was taken separately, resulting in an output vector that was then directly fed into the softmax layer. This approach is generally called global average pooling and was introduced by Lin et al. [14] when proposing Network in Network architecture. On the success of previous architectures, Howard et al. [15] presented an efficient convolutional neural network designed for deployment on mobile devices based on depthwise separable convolution [16].
One of the other approaches is the replacement of some convolutional kernels by approximated Gabor filters. In convolutional computational layers, these filters are used as fixed kernels for the extraction of intrinsic features. Some authors used Gabor filters only in the first convolutional layer [17,18], while Sarwar et al. [19] also looked at their use in other layers with an increasing depth of the network. Adding fixed filters to deeper layers of the network, however, results in a decrease in overall accuracy. Therefore, the same authors proposed a solution that uses a combination of fixed and trainable kernels. The interim results show that this method leads to a significant improvement in energy savings, shortened training time, and reduced memory requirements with a negligible drop in classification accuracy. The combination of learned free-form filters with structured ones also bring the ability to modify the size of convolutional kernels in the training process [20]. It can lead to an effective expansion of the total number of convolutional kernels [21] or reduction of parameters by creating efficient layers for feature extraction [22,23].
In addition to these methods, there are many approaches based on compression, regularization, and pruning [24,25].

1.2. Path towards Receptive Field Neural Networks

One of the possible solutions that could at least partially solve these problems is to replace standard convolutional kernels with a linear combination of predefined and fixed filters in convolutional neural network architectures. This influential concept, proposed by Jacobsen et al. [26], defines Receptive Field Neural Networks (RFNNs) as a group of networks that do not handle convolutional kernels as individual pixels but as functions in scale-space using a finite Gaussian derivative basis. The architecture of RFNNs allows the use of any unspecified base, which creates efficient filters designed for feature extraction. It is therefore not surprising that models based on fixed filters according to the original Receptive Field Neural Network (RFNN) architecture have aroused the interest of researchers in recent years. Schlimbach [27] showed that filter size has an impact on network performance, and it was therefore proposed to include scale as a parameter in the learning process itself. This idea was later elaborated by Pintea et al. [28], who created a network that effectively learns the dynamic size of the convolutional kernel. A similar approach was developed by Tabernik et al. [22,23], who created a Displaced Aggregation Units (DAUs) capable of learning the receptive field size through spatial displacement. By extending the original model with the Gaussian base of multiple scales and orientations, the authors of [29] created a Multi Scale and Orientation Structured Receptive Neural Network (MSO-RFNN) inspired by the DenseNet architecture [30].
Although the architecture of RFNNs are specific to their Gaussian derivative basis, other bases have also been tested. Verkes [31] proposed using the Gabor kernels, where real and imaginary parts are rotated to match a reference Gaussian basis accordingly, and achieved similar results. Discrete directional Parseval frames with compact support were tested in the role of a base by Labate et al. [32]. With the introduction of sparsity constraints, the Geometric-Biased CNN (GBCNN) was created and is capable of solving the problem of hyperspectral image classification. By incorporating Parseval frames into RFNN Resnet-like models, the ResnetRF architecture was proposed by Karantzas et al. [33]. Ulicny et al. [34] proposed using spectral filters based on a Discrete Cosine Transform (DCT), thus creating harmonic neural networks achieving fine accuracy in classification and segmentation tasks.
Although several bases have been proposed (Table 1), a valid comparison is still lacking, especially in the area in which these models excel: training with a small number of training samples. This implies our intention to verify the possibility of using different bases in the RFNN architecture under deterministic conditions.
In this work, we analyse and extend RFNN to study the consequences of changes in the individual compact bases. We restrict RFNN into shallow architectures to look more closely at the impact of different parameters and introduce a deterministic methodology for obtaining reproducible results and evaluation in the form of a Correlated Bayesian t-test.
The paper is structured as follows. In Section 2, we review in detail the RFN layer, which is the crucial layer for the RFNN architecture. In Section 3, we present the reference architecture chosen as the baseline for the experiments. The methodological changes that enable reproducibility of experiments are proposed in Section 4. The performed experiments and their results are presented in Section 5. The obtained results are discussed in Section 6. We conclude our contribution in Section 7.

2. Receptive Field Neural Layer

In neural network architectures, a convolution layer is commonly used for feature extraction. It may differ in implementation, but usually this layer performs a discrete convolution denoted by the ∗ operator between the input tensor h x R 1 × D x × W x × H x and the set of convolutional kernels K R N K × D x × W K × H K , which, at the same time represent the set of parameters of this computational layer. The output is a multidimensional feature map h y R 1 × N K × W x W K + 1 × H x H K + 1 , which is created by applying convolution kernels in all mutual positions to the input where the kernel fits entirely within the input boundaries. Equation (1) shows the calculation of the j-th feature map h j y R 1 × 1 × W x W K + 1 × H x H K + 1 , j { 0 , 1 , 2 , , N K 1 } in a standard 2D convolutional layer, where q { 0 , 1 , 2 , , W x W K } and p { 0 , 1 , 2 , , H x H K } are their horizontal and vertical indices, respectively.
h j y p , q = 2 D C o n v h x , K j p , q = h x × K j p , q = i = 0 D x 1 u = 0 H K 1 v = 0 W K 1 h x i , p + u , q + v · K j i , H K u 1 , W K v 1
We can decompose the regular convolution into two simpler operations: extraction of features in the spatial plane using dephtwise convolution and subsequent channel recombination using pointwise convolution, effectively creating a depthwise separable convolutional layer [44,45]. Depthwise convolution is convolution in the spatial domain applied independently over every channel of an input. The calculation of the feature maps h ˜ y R 1 × N K D x × W x W K + 1 × H x H K + 1 by the 2D depthwise convolution is described by Equation (2), where i { 0 , 1 , 2 , , D x 1 } and k { 0 , 1 , 2 , , N K 1 } . The output is a three-dimensional tensor, where the index o { 0 , 1 , 2 , , N K D x 1 } represents a specific channel in the depth plane, and p { 0 , 1 , 2 , , H x H K } and q { 0 , 1 , 2 , , W x W K } are their vertical and horizontal indices, respectively.
h ˜ y o , p , q = 2 D D e p t h w i s e C o n v h x , Θ o , p , q = u = 0 H K 1 v = 0 W K 1 h x i , p + u , q + v · Θ k , i , H K u 1 , W K v 1
Subsequent pointwise convolution can be seen as regular convolution with a 1 × 1 kernel size, recombining the channels along depth and projecting them onto a new channel space and creating an output j-th feature map h j y , j { 0 , 1 , 2 , , D Y 1 } (Equation (3)).
h j y p , q = P o i n t w i s e C o n v h ˜ y , α j p , q = o = 0 N K D x 1 h ˜ y o , p , q · α j o
The original set of parameters K is hereby replaced by Θ R N K × D x × W K × H K as a set of effective filters and α R D Y × N K D x × 1 × 1 for the subsequent recombination of the created feature maps.
The main idea of the RFNN computational layer is to replace the parameters of 2D spatial filters Θ with N K fixed-base filters θ . This makes it possible to introduce information into the network a priori in the form of fixed spatial 2D convolutional filters for the extraction of intrinsic features, which are then linearly combined to create effective feature maps h y at the output of the RFNN layer (Equation (4)).
h y = P o i n t w i s e C o n v 2 D D e p t h w i s e C o n v h x , θ , α
By introducing fixed filters from a certain domain, the convolutional kernels can be treated as functions, while the network parameters are also reduced, as only α parameters are learned during the training process. The detailed architecture of the RFNN layer is shown in Figure 1.
The hyperparameters of this layer are:
  • A set of fixed two-dimensional kernels ( N K —number of kernels; W K , H K —width and height of kernels);
  • The convolution parameters used for feature extraction (padding, stride, and possibly other additional parameters);
  • The number of output feature maps ( D Y —depth dimension of the output tensor).

3. Choice of Reference Architecture

As a baseline for comparative evaluation of RFNN networks, we selected the methodology and results from Jacobsen et al. [26], as described below. We re-implemented their proposed RFNN network to reproduce the originally published results and to ensure correct implementation using the PyTorch [46] framework. We refer to this implementation as R F N N r e f . Since the original article contains several experiments and architectural modifications, we decided to concentrate on results where the MNIST [47] dataset was used (Figure 5 in the original paper [26]).
The baseline architecture is a convolutional neural network with three layers of receptive fields and is used to classify the known problem of handwritten digits into ten classes. A detailed description of the individual layers together with the output dimensions and number of parameters is given in Table 2.
Fixed convolution kernels from the Gaussian basis were formed in the same way for all layers with a difference in size. In the first layer, the convolutional kernels have a size of 11 × 11 pixels and σ = 1.5 , while in the second and third layers the kernels are identical with a size of 7 × 7 pixels with σ = 1 . In each receptive fields layer, filtered outputs after depthwise convolution are recombined into 64 output feature maps.

4. Proposed Changes

Although we managed to verify the performance of the baseline architecture (see Section 5.1) and even achieve better results (see Section 5.2) than the original authors using the baseline architecture, we noticed the shortcomings of such an evaluation, which is common in the field of neural networks. The results may be affected by the setting of hyperparameters and training methodology, randomness of selection and order of samples, as well as stochastic regularization (e.g., dropout) during training. These effects are more pronounced with a smaller training set, which, in our case, was manifested by an increase in variance. We encountered the difficulties when reproducing the original results. Therefore we propose to change to a methodology that ensures reproducible results. This methodology is also used in our experiments, starting in Section 5.3:
  • All used frameworks were set to deterministic mode;
  • All parameters were initialised based on one chosen experimental seed;
  • To ensure variability, the selected experimental seed was used to generate a seed vector for all runs of the experiment.
Based on the recommendations in [48,49] we chose the Bayesian-based comparison in the form of a Correlated Bayesian t-test [50]. We used an implementation of this test from the Baycomp library [48]. The results of this test are three probabilities: that the models are practically equivalent, that model C1 achieves better results than model C2, and vice versa. For all tests, we chose the standard 1% as the Region of Practical Equivalence (ROPE), where if the difference in the metrics of the compared models A and B is less than or equal to 1%, we consider the models to be equivalent. Based on previous adjustments and evaluations, we decided to change the methodology of the experiments. Subsequently, in each experiment, all frameworks using randomness were set to the deterministic mode. We used stratified 10-fold cross-validation repeated 10 times with different initial seedings (a total of 100 simulations). The same master seed is used for each experiment, so both the selection and the order of the training samples are maintained for all models on the same dataset. The statistical evaluation of the model in the form of Correlated Bayesian t-test was performed on the validation part of the training set within the cross validation. However, a numerical evaluation was performed on the excluded test set containing the same 10,000 samples to obtain an objective view of the real accuracy of the model. A possible alternative is to combine the training and test sets into one dataset. We chose not to use this procedure in order to be able to compare the results obtained on a dedicated test set with previous research in this area. The final training and evaluation methodology for ensuring repeatability is shown in Figure 2.

5. Experiments and Results

In this section, we describe our experiments with RFNN architecture using the MNIST database. In Section 5.1, we describe our procedure for repeating the results from the original work [26]. Next, in Section 5.2, we experiment with training methodology by applying only one change at a time (random or stratified sampling, early stopping, or different optimizer) to analyse its influence on the accuracy. To see the impact of changes in the RFNN architecture, we simplified the reference network. This simplification is specified in Section 5.3. Next, in Section 5.4, we analyse the impact of sample selection and energy normalization on the achieved accuracy. In Section 5.5, we experiment with different bases instead of a Gaussian compact base, and last, in Section 5.6, we remove the pooling layer and provide corresponding learning curves.

5.1. Experiment 1—Reference Architecture Evaluation

Receptive field neural networks excel when not enough data are available to use common deep learning models. We used a limited number of training samples in the range from 300 to 60,000 samples, i.e., the whole training set. For each experiment, the entire MNIST test set, consisting of 10,000 samples, was used to evaluate the classification accuracy. Apart from normalizing data to the range <0,1>, no other form of sample pre-processing was used. The network was trained by a standard backpropagation algorithm using an average cross-entropy loss with a mini-batch size of 25 images. An Adadelta optimizer with decay rate ρ = 0.95 , stability constant ϵ = 1 × 10 6 and linear learning rate decay was applied. The initial value of the learning rate was set to l r i = 5 and decreased to the final value of l r f = 0.05 during the training, while a fixed number of epochs were taken from the original article. The values are listed in Table 3 [26].
The resulting classification accuracy of the reference architecture R F N N r e f representing the average of three experiments is depicted in Figure 3. The original and our result lines approximately match, but there is apparent difference in the last point for the 300 training samples. The variability of these results even if the original methodology is followed may be due to various factors such as the random number generator (RNG) initial state, hardware parameters, software versions of used frameworks and others. These data are missing in the majority of published studies [51], which makes the verification of original results more difficult.
Since we were unable to achieve the original published values for three repetitions, we decided to repeat each experiment 300 times and evaluate the results by displaying the full range < m i n , m a x > of the achieved accuracy. The resulting interval is also shown in Figure 3. For reference, we compared the R F N N r e f model with standard machine learning classifiers such as k-nearest neighbors (KNN) and support vector machine (SVM). We used an SVM classifier with a radial basis function kernel (RBF), a unit regularization parameter and an automatic kernel coefficient gamma. A KNN classifier with a uniform weight function and 5 neighbors was trained using a brute-force search with standard Euclidean distance as a metric. The range of achieved values is represented in the following part of the article by a violin plot [52]. In addition to the interquartile range, this type of graph also shows the estimated distribution of data. In our representation, we marked the median and displayed the entire range of values from min to max.
We can confirm that we have managed to achieve comparable results to those presented by Jacobsen et al. [26]. The R F N N r e f reference architecture achieves a significant improvement in classification accuracy compared to standard machine learning models in our reproduction test on the MNIST database if the number of training samples is limited.

5.2. Experiment 2—Reference Architecture and Limited Number of Training Samples

In this section, we focus on the most important contribution of the RFNN architecture, which is the ability to achieve high accuracy with a limited number of training samples. Analysing the results, we observed a high variability of classification accuracy for the smallest number of training samples, which is evident from the violin graph (Figure 3). From these results, it is not possible to determine whether the high variance is due to sampling, the characteristics of the chosen model, or by specific training parameters. To analyse the variance and impact of parameter changes, we tested several adjustments and modifications of the training methodology. Network training and pre-processing were done exactly the same as in the previous section with a random selection of 300 training samples. We repeated the experiment N = 300 times, making only one change in the methodology at a time. The results are shown in Figure 4 and also numerically in Table 4. Note, the leftmost case in Figure 4 is the same as the violin plot (rightmost case) in Figure 3.

5.2.1. Stratified Sampling

First, we replaced random with stratified sampling as a way to eliminate sampling bias and thus a possible source of variance. The resulting violin graph in Figure 4 is denoted as “(orig) + Stratification”. We achieved comparable results to those of the reference methodology, and thus, the possible class imbalance did not have a significant impact on the observed variance.
Figure 4. Comparison of various modifications of the methodology on the MNIST database when selecting 300 random training samples (i.e., it corresponds to a value of 300 on the x-axis in Figure 3).
Figure 4. Comparison of various modifications of the methodology on the MNIST database when selecting 300 random training samples (i.e., it corresponds to a value of 300 on the x-axis in Figure 3).
Sensors 22 09743 g004

5.2.2. Early Stopping

In the following experiment, in addition to the original methodology, we used early stopping as follows: 10% of the 300 random samples were selected via stratified sampling in the validation set. The remaining 90% of samples were used for training. During training, the cross-entropy loss on the validation set was evaluated after each epoch. If this loss was reduced, we saved the current parameters of the model and continued training. If the validation loss did not improve during p = M m a x e p o c h s , then we stopped the training and used the best model whose parameters we had saved. We performed two different experiments for the conditions of early stopping, p = 20 and 100, to evaluate the impact of this parameter. When testing a longer stopping condition p = 100 , better results were obtained, but they had still deteriorated. The results show that a longer stop condition generally produces better results, which is consistent with the results of Prechelt [53]. Figure 4 shows that the use of early stopping can increase the variance of the results, especially if the stop condition is too strict. The deterioration of the results may be affected by the small number of samples in the validation set and by the use of a dropout in the architecture. However, their impact on the overall result would need to be examined in more detail. Due to mixed opinions on the use of early stopping and deteriorating results in our particular case, we decided not to use early stopping in further experiments.

5.2.3. Optimizer Modification

Choi et al. [54] showed that not only does the choice of the optimizer but also the setting of its parameters has a fundamental influence on the model accuracy. However, since tuning the optimizer hyperparameters would require a separate optimization process, and a grid search would be computationally intensive, we decided to test the latest optimizers with the default settings recommended by their respective authors without using the learning rate decay. We used Adadelta [55], Adam [56], AdamW [57], AdaBound [58] and Nadam [59] gradient descent optimization algorithms. The results of our tests (Figure 4 and Table 4) show that the optimizers are almost as successful—only AdaBound and Adadelta lag behind at this setting. In the last comparison, we decided to use the learning rate decay for the most successful optimizer, AdamW, as well as in the original experiments, where we managed to surpass the original results.
Table 4. Results of various modifications of the training methodology for analysis of variance. The better results obtained compared to the reference values in the first row are highlighted in bold.
Table 4. Results of various modifications of the training methodology for analysis of variance. The better results obtained compared to the reference values in the first row are highlighted in bold.
MethodologyMean
Accuracy [ % ]
Min
Accuracy [ % ]
Max
Accuracy [ % ]
Variance [ % 2 ]
Adadelta + lr decay (orig)95.2693.7096.630.3033
(orig) + Stratification95.3093.6496.540.2684
(orig) + Early Stopping (20)92.3984.7895.032.7026
(orig) + Early Stopping (100)93.2884.7895.762.3816
AdamW94.6493.0896.390.3854
Nadam94.8892.9396.180.3993
Adam94.6592.6796.170.3889
Adadelta94.0891.1595.780.5439
AdaBound92.9490.5994.760.6310
AdamW + lr decay95.6794.0896.920.2204

5.3. Experiment—Simplification of RFNN Architecture

For a valid comparison of the impact of changes in the RFNN architecture, we simplified the reference network as much as possible. Due to the possible mutual influence of parameters in deep architectures, we omitted the normalization and regularization layers and simplified the neural network to a shallow architecture with one computational RFConv layer. The architecture is described in detail in Table 5.
Another simplification was the removal of computational layers (local response normalization and dropout) for which we could not ensure repeatability when training on a GPU. We kept the base and number of filters according to the original architecture. For training, we used the same settings as in the previous experiments except for changing the optimizer to AdamW using a fixed learning rate with a value of 1 × 10 3 . For a reference, we evaluated the simplified 1-layer architecture R F N N L 1 against the original 3-layer architecture R F N N r e f in the form of 10 times repeated 10-fold cross-validation using a Correlated Bayesian t-test for the random selection of 300 samples.
Figure 5 shows the achieved accuracies on a test set in the form of a violin graph and posterior probability distribution of a Correlated Bayesian t-test evaluated on a validation set in each cross-validation split. Compared to the original architecture, the results deteriorated significantly, which was expected. For a ROPE of 1%, the probability that R F N N r e f was better than R F N N L 1 was 99.03%, the probability that both models were practically equivalent was 0.95%, and the probability that R F N N L 1 was better than R F N N r e f was 0.02%.

5.4. Experiment 4—Sample Selection and Energy Normalization

In this experiment, we analysed the impact of sample selection and Energy Normalization on the achieved accuracy of the created model. The settings and methodology of the training were the same as in the previous section. We chose R F N N L 1 as the reference architecture.

5.4.1. Sample Selection

We conditioned the selection of samples by changing the deterministic seed that affected the result. We trained again on a random selection of 300 samples and re-evaluated classification accuracies using the same methodology. A comparison of the results obtained with the seed change is shown in Figure 6.
The results differ due to a different selection of samples for the training set despite the same R F N N L 1 architecture. The probability that the model R F N N L 1 s e e d # 1 was better than R F N N L 1 s e e d # 2 was 39.87%, the probability that R F N N L 1 s e e d # 2 was better than R F N N L 1 s e e d # 1 was 31.69%, and the probability that both models were practically equivalent was 28.44 % = 100 % ( 39.87 % + 31.69 % ) .

5.4.2. Basis Energy Normalization

Normalization in neural network architectures can help to improve both convergence and generalization [60]. Different types of preprocessing and normalization are introduced to eliminate sharp differences between network parameters or extracted feature maps [61]. This effect can be even more pronounced when using fixed filters. For this reason, we evaluated the impact of energy normalization on the achieved accuracy. By energy normalization, we mean normalization to unit energy, where the sum of all squared k-th filter elements θ k R W K × H K is equal to one (Equation (5)).
θ k n o r m m , n = θ k m , n u = 0 H K 1 v = 0 W K 1 θ k 2 u , v
where m { 0 , 1 , 2 , , H K 1 } and n { 0 , 1 , 2 , , W K 1 } are their corresponding vertical and horizontal indices. The achieved accuracies on a test set in the form of a violin graph and posterior probability distribution of a Correlated Bayesian t-test evaluated on a validation set in each cross-validation split are shown in Figure 7 for the Gaussian derivative basis, and in Figure 8 for randomly initialized kernels.
The results of both experiments show that base energy normalization helped to improve classification accuracy. When using the original Gaussian derivative base, we found that R F N N L 1 G a u s s i a n was better than R F N N L 1 G a u s s i a n n o r m a l i z e d with a probability of 6.75%, the probability that both models were practically equivalent was 55.06%, and the probability that R F N N L 1 G a u s s i a n n o r m a l i z e d was better than R F N N L 1 G a u s s i a n was 38.19%. In the case of using a random base, the probability that R F N N L 1 R a n d o m was better than R F N N L 1 R a n d o m n o r m a l i z e d was 21.15%, the probability that both models were practically equivalent was 49.21%, and the probability that R F N N L 1 R a n d o m n o r m a l i z e d was better than R F N N L 1 R a n d o m was 29.64%.

5.5. Experiment 5—Basis-Related Experiments

In the original article [26], the Gaussian compact base was used for feature extraction. The aim of our experiment was to verify whether we can replace the given base and what effect this change will have on the achieved classification accuracy. We again chose R F N N L 1 as the reference architecture, while the parameters and evaluation methodology remained unchanged. In each experiment, we kept all parameters constant except for the base used, which had a size of 11 × 11 pixels. Since the reference Gaussian base contained 10 filters, we selected these filters along the primary diagonal based on a triangular selection according to Ulicny et al. [34]. We compared the reference Gaussian derivative base (Gaussian), the orthonormal Discrete Cosine base (DCTII), the orthonormal Discrete Hartley base (DHT), the randomly initialized kernels with normal distribution (RND), and the random orthonormal base (ORTHN_RND). The results for the selection of 300 samples are shown in Figure 9.
We evaluated neural models using a Correlated Bayesian t-test with respect to the simplified reference architecture R F N N L 1 with a Gaussian base for each model change separately (Figure 10). The results showed that for all tested bases, the hypothesis that both compared models are equivalent has the highest probability, with the one exception being a random base, which worsened the results and had the greatest impact on achieved classification accuracy. The probability that the R F N N L 1 model of a network with a Gaussian base is better than a model with a random kernels is about 47.1%, which far exceeds the other results. Compared to the other bases tested, we obtained the best results using discrete Hartley and discrete cosine bases. In both cases, the results of the statistical test were that the models were practically equivalent, with a probability of 51.06% and 44.77%, respectively.

5.6. Experiment 6—Learning Curve Analysis and Max-Pooling Removal

In the previous experiment in Section 5.5, we analysed the influence of the type of fixed basis used. As we can see in Figure 9 and Figure 10, the random kernels also performed relatively well, which could be due to the simplicity of the MNIST database used, or excessive downsampling in max-pooling layer. Based on the obtained results, we decided to explore the influence of the max-pooling layer on the classification accuracy. We compared the network performance with and without the max-pooling layer on the more complex Kuzushiji-MNIST [62] database. The R F N N L 1 model was trained for a fixed number of 50 epochs using the AdamW optimizer with default setting while following the same methodology and preprocessing as in the previous experiments. We show the corresponding representative results in the form of the learning curves in Figure 11. It can be seen from the results, that the difference between randomly initialized filters and structured bases widens significantly when max-pooling is omitted.

6. Discussion

The outcomes of this paper have provided insight into the Receptive Field Neural Network architecture, which is specific in that it uses a linear combination of a predefined fixed kernels to create a set of effective filters instead of learning entire convolutional kernels pixel-by-pixel. The results in Section 5.1 confirm the previously published results on the MNIST database (Figure 3) by Jacobsen et al. [26] and demonstrate the possibility of replacing the original Gaussian derivative base with other structured bases intended for the extraction of intrinsic features within the shallow RFNN architecture. Since the RFNN excels compared to other architectures when training with a limited amount of data; the experiments focused on the random selection of N = 300 training samples only while maintaining repeatability of results by introducing a deterministic evaluation methodology (Figure 2). The methodology change was introduced to ensure reliability of findings with respect to the high variability of the achieved classification accuracy on the test set (Section 5.1 and Section 5.2), which we could not reduce by the introduction of sample stratification, the use of early stopping during training, or by changing the optimizer (Figure 4 and Table 4), as shown in Section 5.2. We modified the training methodology to ensure repeatability and evaluated the proposed step-by-step changes in the form of a Correlated Bayesian t-test using 10 times repeated 10-fold cross-validation. By switching to this evaluation methodology, we ensured a reproducible environment for reliable comparison of results where not only the selection but also the order of the samples during training remained the same. To ensure repeatability within the neural network, we simplified the original R F N N r e f (Table 2) architecture to a single-layer network R F N N L 1 (Table 5) with removal of normalization and regularization layers for which we could not guarantee determinism within the used framework. At the cost of a significant deterioration of R F N N L 1 classification accuracy compared to the baseline (Section 5.3, Figure 5), we obtained reproducible conditions for further experiments (Section 5.4 and Section 5.5). When evaluating the influence of the seed for the pseudorandom number generator, on which depends samples selection and the initialization of parameters within the R F N N L 1 network in the form of a Correlated Bayesion t-test, we found that the probability that the same model with different initial seedings and training sets are practically equivalent was only 28.44 % (Figure 6b). This probability was the lowest among all experiments, meaning that this change had the greatest impact on the achieved classification accuracy. The results are in line with the findings of Crane et al. [51], who pointed out that unreported effects in evaluation methodology can substantially influence the achieved results. The data also show the positive impact of energy normalization (Section 5.4), which improved the classification accuracy not only for the Gaussian derivative base (Figure 7) but also for randomly initialized fixed filters (Figure 8). The next experiment was a comparison of specifically structured bases within the R F N N L 1 architecture (Section 5.5). Although we managed to outperform the original Gaussian derivative basis using DCT and DHT bases (Figure 9), the more important result is that the Bayesian Correlated t-test (Figure 10) did not show a significant difference between individual bases in terms of achieved classification accuracy with the exception of randomly initialized filters (Table 6). In the last experiment (Section 5.6, Figure 11), we tested the impact of the max-pooling layer. The results show that omitting max-pooling increased the difference between structured bases and randomly initialized kernels. Further research could be devoted to initialization with structured filters based on various transforms, which could, in theory, improve convergence, as shown recently by Li et al. [63]. Since the feature extraction in our reference RFNN architecture is divided into filtering using a set of filters along the spatial dimensions and the subsequent combination of channels via pointwise convolution, representation of the extracted features and the subsequent visual display can be easier to interpret compared to the classical convolutional layers. Therefore, it would be interesting to examine the visual properties of these architectures with respect to the extraction of intrinsic features for better understanding and deployment in areas with a lack of training data, in which they excel.

7. Conclusions, Limitations, and Future Research Work

Although these results confirm the added value of structured filters within the RFNN architecture, the differences between the specific bases chosen in the context of neural networks need further investigation, mainly in the context of the various constraints introduced. Deployments of complex neural networks in critical applications and devices with limited computing power are the main reasons for simplification and deeper analysis of created models. In this work, we analysed the Receptive Field Neural Network (RFNN) as a promising simplified model that uses a Gaussian derivative basis inspired by the scale-space theory. When validating the original results, we identified a problem with a large variation in the achieved test classification accuracy, especially with a small amount of training data. To enable the reproducibility of the results by introducing deterministic methodology, we simplified the baseline RFNN architecture to a single-layer CNN network whose input of computational layers was a set of fixed filters of any size and number. Subsequently, we experimentally verified that bases other than Gaussian can be used as fixed filters within the RFNN architecture. We found that a change in base may have less of an effect on the results obtained than re-training the network using another seed in failing to ensure repeatability of results. We also verified the positive impact of energy normalization of used filters, which improves the achieved classification accuracy even when using randomly initialized kernels. Further research is needed to establish specific differences between individual bases and the influence of various hyperparameters within the RFNN architecture.

Author Contributions

Conceptualization, J.G. and R.V.; methodology, J.G. and R.V.; software, J.G.; validation, R.V.; investigation, J.G., R.V., J.P., M.O. and S.K.; data curation, J.G. and R.V.; writing—original draft preparation, J.G., R.V. and J.P.; writing—review and editing, J.P., M.O. and S.K.; visualization, J.G. and R.V.; supervision, J.P., M.O. and S.K.; funding acquisition, J.P. and M.O. All authors have read and agreed to the published version of the manuscript.

Funding

The research described in this paper was funded by project No. 1/0202/23 of the Slovak Grant Agency VEGA and the Internal FEI STU Bratislava project grant No. 2022-23-04 to support young excellent research teams.

Data Availability Statement

The experiments in this paper were performed on the public “KMNIST Dataset” [62] (created by CODH), adapted from the “Kuzushiji Dataset” (created by NIJL and others): http://doi.org/10.20676/00000341 and MNIST [47] datasets.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mishkin, D.; Sergievskiy, N.; Matas, J. Systematic Evaluation of Convolution Neural Network Advances on the Imagenet. Comput. Vis. Image Underst. 2017, 161, 11–19. [Google Scholar] [CrossRef] [Green Version]
  2. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A Review on Deep Learning Techniques Applied to Semantic Segmentation. arXiv 2017. [Google Scholar] [CrossRef]
  3. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  4. Ozturk, S.; Ozkaya, U.; Akdemir, B.; Seyfi, L. Convolution Kernel Size Effect on Convolutional Neural Network in Histopathological Image Processing Applications. In Proceedings of the 2018 International Symposium on Fundamentals of Electrical Engineering, ISFEE 2018, Bucharest, Romania, 1–3 November 2018. [Google Scholar]
  5. Karatzoglou, A.; Schnell, N.; Beigl, M. Applying Depthwise Separable and Multi-Channel Convolutional Neural Networks of Varied Kernel Size on Semantic Trajectories. Neural Comput. Appl. 2020, 32, 6685–6698. [Google Scholar] [CrossRef]
  6. Chollet, F. Deep Learning with Python; Manning: Shelter Island, NY, USA, 2017; p. 384. ISBN 9781617294433. [Google Scholar]
  7. Yao, H.; Chuyi, L.; Dan, H.; Weiyu, Y. Gabor Feature Based Convolutional Neural Network for Object Recognition in Natural Scene. In Proceedings of the Proceedings—2016 3rd International Conference on Information Science and Control Engineering, ICISCE 2016, Beijing, China, 8–10 July 2016. [Google Scholar]
  8. Rudin, C. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [Green Version]
  9. Shorten, C.; Khoshgoftaar, T.M. A Survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  10. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A Survey of Transfer Learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef] [Green Version]
  11. Hussain, Z.; Gimenez, F.; Yi, D.; Rubin, D. Differential Data Augmentation Techniques for Medical Imaging Classification Tasks. AMIA Annu. Symp. Proc. 2017, 2017, 979. [Google Scholar]
  12. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size. arXiv 2016. [Google Scholar] [CrossRef]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; Volume 2016. [Google Scholar]
  14. Lin, M.; Chen, Q.; Yan, S. Network in Network. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, 14–16 April 2014. Conference Track Proceedings. [Google Scholar]
  15. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  16. Sifre, L.; Mallat, S. Rotation, Scaling and Deformation Invariant Scattering for Texture Discrimination. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  17. Chang, S.Y.; Morgan, N. Robust CNN-Based Speech Recognition with Gabor Filter Kernels. In Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Singapore, 14–18 September 2014. [Google Scholar]
  18. Li, J.; Wang, T.; Zhou, Y.; Wang, Z.; Snoussi, H. Using Gabor Filter in 3D Convolutional Neural Networks for Human Action Recognition. In Proceedings of the Chinese Control Conference, CCC, Dalian, China, 26–28 July 2017. [Google Scholar]
  19. Sarwar, S.S.; Panda, P.; Roy, K. Gabor Filter Assisted Energy Efficient Fast Learning Convolutional Neural Networks. In Proceedings of the International Symposium on Low Power Electronics and Design, Taipei, Taiwan, 24–26 July 2017. [Google Scholar]
  20. Shelhamer, E.; Wang, D.; Darrell, T. Efficient Receptive Field Learning by Dynamic Gaussian Structure. In Proceedings of the ICLR 2019 Workshop, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  21. Luan, S.; Chen, C.; Zhang, B.; Han, J.; Liu, J. Gabor Convolutional Networks. IEEE Trans. Image Process. 2018, 27, 4357–4366. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Tabernik, D.; Kristan, M.; Leonardis, A. Spatially-Adaptive Filter Units for Deep Neural Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  23. Tabernik, D.; Kristan, M.; Leonardis, A. Spatially-Adaptive Filter Units for Compact and Efficient Deep Neural Networks. Int. J. Comput. Vis. 2020, 128, 2049–2067. [Google Scholar] [CrossRef]
  24. Li, J.Y.; Zhao, Y.K.; Xue, Z.E.; Cai, Z.; Li, Q. A Survey of Model Compression for Deep Neural Networks. Gongcheng Kexue Xuebao/Chinese J. Eng. 2019, 41, 1229–1239. [Google Scholar]
  25. Blalock, D.; Gonzalez Ortiz, J.J.; Frankle, J.; Guttag, J. What is the state of neural network pruning? In Proceedings of the 3rd MLSys Conference, Austin, TX, USA, 2–4 March 2020.
  26. Jacobsen, J.H.; Van Gemert, J.; Lou, Z.; Smeulders, A.W.M. Structured Receptive Fields in CNNs. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; Volume 2016. [Google Scholar]
  27. Schlimbach, R.J. Investigating Scale in Receptive Fields Neural Networks. Bachelor’s Thesis, University of Amsterdam, Amsterdam, The Netherlands, 2018. [Google Scholar]
  28. Pintea, S.L.; Tomen, N.; Goes, S.F.; Loog, M.; Van Gemert, J.C. Resolution Learning in Deep Convolutional Networks Using Scale-Space Theory. IEEE Trans. Image Process. 2021, 30, 8342–8353. [Google Scholar] [CrossRef]
  29. Hilbert, A.; Veeling, B.S.; Marquering, H.A. Data-Efficient Convolutional Neural Networks for Treatment Decision Support in Acute Ischemic Stroke 2018. In Proceedings of the 1st Conference on Medical Imaging with Deep Learning (MIDL 2018), Amsterdam, The Netherlands, 4–6 July 2018. [Google Scholar]
  30. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the Proceedings—30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; Volume 2017. [Google Scholar]
  31. Verkes, G. Receptive Fields Neural Networks Using the Gabor Kernel Family. Bachelor’s Thesis, University of Amsterdam, Amsterdam, The Netherlands, 2017. [Google Scholar]
  32. Labate, D.; Safaripoorfatide, M.; Karantzas, N.; Prasad, S.; Foroozandeh Shahraki, F. Structured Receptive Field Networks and Applications to Hyperspectral Image Classification. In Proceedings of the Wavelets Sparsity XVIII, San Diego, CA, USA, 11–15 August 2019; Volume 11138, pp. 218–226. [Google Scholar] [CrossRef]
  33. Karantzas, N.; Safari, K.; Haque, M.; Sarmadi, S.; Papadakis, M. Compactly Supported Frame Wavelets and Applications in Convolutional Neural Networks. In Proceedings of the Wavelets Sparsity XVIII, San Diego, CA, USA, 11–15 August 2019; Volume 11138, pp. 152–164. [Google Scholar] [CrossRef]
  34. Ulicny, M.; Krylov, V.A.; Dahyot, R. Harmonic Convolutional Networks Based on Discrete Cosine Transform. Pattern Recognit. Pattern Recognit. 2022, 129, 108707. [Google Scholar] [CrossRef]
  35. Ulicny, M.; Krylov, V.A.; Dahyot, R. Harmonic Networks for Image Classification. In Proceedings of the British Machine Vision Conference (BMVC), Cardiff, UK, 9–12 September 2019. [Google Scholar]
  36. Ulicny, M.; Krylov, V.A.; Dahyot, R. Harmonic Networks with Limited Training Samples; In Proceedings of the 27th European Signal Processing Conference, EUSIPCO, Coruña, Spain, 2–6 September 2019.
  37. Kumawat, S.; Raman, S. Depthwise-STFT Based Separable Convolutional Neural Networks. In Proceedings of the 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3337–3341. [Google Scholar] [CrossRef] [Green Version]
  38. Tomen, N.; Pintea, S.-L.; Gemert, J. Van Deep Continuous Networks. In Proceedings of the 38th International Conference on Machine Learning, Virtual Event. 18–24 July 2021. PMLR 139:10324-110335. [Google Scholar]
  39. Saldanha, N.; Pintea, S.L.; Van Gemert, J.C.; Tomen, N. Frequency Learning for Structured CNN Filters with Gaussian Fractional Derivatives. arXiv 2021, arXiv:2111.06660. [Google Scholar]
  40. Lindeberg, T. Scale-Covariant and Scale-Invariant Gaussian Derivative Networks. J. Math. Imaging Vis. 2022, 64, 223–242. [Google Scholar] [CrossRef]
  41. Elmoataz, A.; Fadili, J.; Quéau, Y.; Rabin, J.; Simon, L. (Eds.) Scale Space and Variational Methods in Computer Vision: 8th International Conference, SSVM 2021, Virtual Event, May 16–20, 2021, Proceedings; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; Volume 12679, pp. 3–14. [Google Scholar] [CrossRef]
  42. Fukuzaki, S.; Ikehara, M. Principal Components of Neural Convolution Filters. IEEE Access 2022, 10, 104328–104336. [Google Scholar] [CrossRef]
  43. Penaud–Polge, V.; Velasco-Forero, S.; Angulo, J. Fully Trainable Gaussian Derivative Convolutional Layer. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2421–2425. [Google Scholar] [CrossRef]
  44. Wei, H.; Wang, Z.; Hua, G. Dynamically Mixed Group Convolution to Lighten Convolution Operation. In Proceedings of the 2021 4th International Conference on Artificial Intelligence and Big Data, ICAIBD 2021, Chengdu, China, 28–31 May 2021. [Google Scholar]
  45. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the Proceedings—30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  46. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury Google, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the 33rd Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar] [CrossRef]
  47. Yann, L.; Corinna, C. Burges Christopher THE MNIST DATABASE of Handwritten Digits. Courant Inst. Math. Sci. 1998. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 1 October 2022).
  48. Benavoli, A.; Corani, G.; Demšar, J.; Zaffalon, M. Time for a Change: A Tutorial for Comparing Multiple Classifiers through Bayesian Analysis. J. Mach. Learn. Res. 2017, 18, 2653–2688. [Google Scholar]
  49. Corani, G.; Benavoli, A.; Demšar, J.; Mangili, F.; Zaffalon, M. Statistical Comparison of Classifiers through Bayesian Hierarchical Modelling. Mach. Learn. 2017, 106, 1817–1837. [Google Scholar] [CrossRef] [Green Version]
  50. Nilsson, A.; Smith, S.; Ulm, G.; Gustavsson, E.; Jirstrand, M. A Performance Evaluation of Federated Learning Algorithms. In Proceedings of the DIDL 2018—Proceedings of the 2nd Workshop on Distributed Infrastructures for Deep Learning, Part of Middleware 201, Rennes, France, 10 December 2018. [Google Scholar]
  51. Crane, M. Questionable Answers in Question Answering Research: Reproducibility and Variability of Published Results. Trans. Assoc. Comput. Linguist. 2018, 6, 241–252. [Google Scholar] [CrossRef] [Green Version]
  52. Hintze, J.L.; Nelson, R.D. Violin Plots: A Box Plot-Density Trace Synergism Statistical Computing and Graphics Violin Plots: A Box Plot-Density Trace Synergism. Source Am. Stat. 1998, 52, 181–184. [Google Scholar]
  53. Prechelt, L. Early Stopping—But When? Lect. Notes Comput. Sci. 2012, 7700, 55–69. [Google Scholar] [CrossRef]
  54. Choi, D.; Shallue, C.J.; Nado, Z.; Lee, J.; Maddison, C.J.; Dahl, G.E. On Empirical Comparisons of Optimizers for Deep Learning. arXiv 2019. [Google Scholar] [CrossRef]
  55. Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012. [Google Scholar] [CrossRef]
  56. Kingma, D.P.; Ba, J.L. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  57. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  58. Luo, L.; Xiong, Y.; Liu, Y.; Sun, X. Adaptive Gradient Methods with Dynamic Bound of Learning Rate. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  59. Dozat, T. Incorporating Nesterov Momentum into Adam. In Proceedings of the 4th International Conference on Learning Representations, ICLR 2016 —Conference Track Proceedings, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  60. Shao, J.; Hu, K.; Wang, C.; Xue, X.; Raj, B. Is Normalization Indispensable for Training Deep Neural Networks? In Proceedings of the Advances in Neural Information Processing Systems, Online, 6–12 December 2020; Volume 2020. [Google Scholar]
  61. Huang, L.; Qin, J.; Zhou, Y.; Zhu, F.; Liu, L.; Shao, L. Normalization Techniques in Training DNNs: Methodology, Analysis and Application. arXiv 2020. [Google Scholar] [CrossRef]
  62. Clanuwat, T.; Bober-Irizar, M.; Kitamoto, A.; Lamb, A.; Yamamoto, K.; Ha, D. Deep Learning for Classical Japanese Literature. arXiv 1998. Available online: https://github.com/rois-codh/kmnist (accessed on 1 December 2022).
  63. Li, L.; Wang, K.; Li, S.; Feng, X.; Zhang, L. LST-Net: Learning a Convolutional Neural Network with a Learnable Sparse Transform. Lect. Notes Comput. Sci. 2020, 12355, 562–579. [Google Scholar] [CrossRef]
Figure 1. Receptive Field Neural Layer. The set of N K fixed filters is first copied D x times to equal the number of input channels. Subsequently, intrinsic features are extracted using the depthwise convolution, and the output feature maps are created in the learning process by a linear combination using a pointwise convolution.
Figure 1. Receptive Field Neural Layer. The set of N K fixed filters is first copied D x times to equal the number of input channels. Subsequently, intrinsic features are extracted using the depthwise convolution, and the output feature maps are created in the learning process by a linear combination using a pointwise convolution.
Sensors 22 09743 g001
Figure 2. The training and evaluation methodology consisted of stratification and repeated cross-validation while ensuring repeatability by setting a deterministic mode in all parts of the experiment. The final evaluation is realized in the form of a Correlated Bayesian t-test.
Figure 2. The training and evaluation methodology consisted of stratification and repeated cross-validation while ensuring repeatability by setting a deterministic mode in all parts of the experiment. The final evaluation is realized in the form of a Correlated Bayesian t-test.
Sensors 22 09743 g002
Figure 3. Reproduction of the originally published results on the MNIST database. The average of the three experiments is shown in green for all samples listed in Table 3. Accuracies of the standard KNN and SVM machine learning classifiers are given as separate curves. The graph also shows the range of results obtained for repeating the experiment N = 300 times for the reference architecture R F N N r e f . For the smallest number of training samples, the estimated distribution of all repeated experimental results shown as a violin plot is also displayed separately (best viewed in colour).
Figure 3. Reproduction of the originally published results on the MNIST database. The average of the three experiments is shown in green for all samples listed in Table 3. Accuracies of the standard KNN and SVM machine learning classifiers are given as separate curves. The graph also shows the range of results obtained for repeating the experiment N = 300 times for the reference architecture R F N N r e f . For the smallest number of training samples, the estimated distribution of all repeated experimental results shown as a violin plot is also displayed separately (best viewed in colour).
Sensors 22 09743 g003
Figure 5. Comparison of R F N N r e f and R F N N L 1 neural model on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N r e f and R F N N L 1 of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Figure 5. Comparison of R F N N r e f and R F N N L 1 neural model on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N r e f and R F N N L 1 of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Sensors 22 09743 g005
Figure 6. Comparison of R F N N L 1 neural models trained on different samples based on two deterministic seeds on the MNIST database using proposed methodology. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 s e e d # 1 and R F N N L 1 s e e d # 2 of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Figure 6. Comparison of R F N N L 1 neural models trained on different samples based on two deterministic seeds on the MNIST database using proposed methodology. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 s e e d # 1 and R F N N L 1 s e e d # 2 of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Sensors 22 09743 g006
Figure 7. Comparison of R F N N L 1 neural models with Gaussian derivative basis with and without energy normalization according to Equation (5) on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 G a u s s i a n and the same model with the energy normalized base R F N N L 1 G a u s s i a n n o r m a l i z e d of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Figure 7. Comparison of R F N N L 1 neural models with Gaussian derivative basis with and without energy normalization according to Equation (5) on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 G a u s s i a n and the same model with the energy normalized base R F N N L 1 G a u s s i a n n o r m a l i z e d of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Sensors 22 09743 g007
Figure 8. Comparison of R F N N L 1 neural models with Randomly initialized kernels with normal distribution with and without energy normalization according to Equation (5) on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 R a n d o m and the same model with the energy normalized base R F N N L 1 R a n d o m n o r m a l i z e d of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Figure 8. Comparison of R F N N L 1 neural models with Randomly initialized kernels with normal distribution with and without energy normalization according to Equation (5) on the MNIST database using proposed methodology for random selection of 300 training samples. The evaluation of the results is shown for test classification accuracy represented by violin graph (a) and posterior probability distribution between R F N N L 1 R a n d o m and the same model with the energy normalized base R F N N L 1 R a n d o m n o r m a l i z e d of a Correlated Bayesian t-test for the validation classification accuracy at 1% ROPE (b).
Sensors 22 09743 g008
Figure 9. Comparison of different bases of simplified R F N N L 1 architecture. Each violin graph represents test classification accuracy of R F N N L 1 neural model with different compact basis on the MNIST database. The test was performed for a random selection of 300 training samples using proposed methodology.
Figure 9. Comparison of different bases of simplified R F N N L 1 architecture. Each violin graph represents test classification accuracy of R F N N L 1 neural model with different compact basis on the MNIST database. The test was performed for a random selection of 300 training samples using proposed methodology.
Sensors 22 09743 g009
Figure 10. Posterior probability distributions of a Correlated Bayesian t-test between reference simplified R F N N L 1 with Gaussian basis and (a) orthonormal discrete cosine basis (DCTII), (b) orthonormal discrete Hartley basis (DHT), (c) randomly initialized kernels with normal distribution (RND), and (d) orthonormal random basis (ORTH_RND). The vertical lines define a region of practical equivalence where the mean difference in accuracy is within ± 1 % . The test was performed for a random selection of 300 training samples using the proposed methodology.
Figure 10. Posterior probability distributions of a Correlated Bayesian t-test between reference simplified R F N N L 1 with Gaussian basis and (a) orthonormal discrete cosine basis (DCTII), (b) orthonormal discrete Hartley basis (DHT), (c) randomly initialized kernels with normal distribution (RND), and (d) orthonormal random basis (ORTH_RND). The vertical lines define a region of practical equivalence where the mean difference in accuracy is within ± 1 % . The test was performed for a random selection of 300 training samples using the proposed methodology.
Sensors 22 09743 g010
Figure 11. Learning curves of R F N N L 1 neural models with and without max-pooling (WMP) layer displayed for Randomly initialized base, DCT, and Gaussian derivative basis on the Kuzushiji-MNIST database using proposed methodology for random selection of 300 training samples (a) and 1000 training samples, respectively (b).
Figure 11. Learning curves of R F N N L 1 neural models with and without max-pooling (WMP) layer displayed for Randomly initialized base, DCT, and Gaussian derivative basis on the Kuzushiji-MNIST database using proposed methodology for random selection of 300 training samples (a) and 1000 training samples, respectively (b).
Sensors 22 09743 g011
Table 1. A comparison of proposed fixed kernels from a predefined dictionary and their applications within different convolutional neural network architectures inspired by the Receptive Field Neural Network in recent years.
Table 1. A comparison of proposed fixed kernels from a predefined dictionary and their applications within different convolutional neural network architectures inspired by the Receptive Field Neural Network in recent years.
AuthorsProposed NetworkReference ArchitectureFixed Kernels
Jacobsen et al. (2016) [26]Receptive Field
Neural Network
multi-layer CNN,
Network in Network
Gaussian derivative basis
11 × 11 , 7 × 7
Verkes (2017) [31]Gabor RFNNmulti-layer CNNGabor kernel family
7 × 7
Schlimbach (2018) [27]Gaussian RFNNmulti-layer CNNGaussian derivative basis
9 × 9 , 7 × 7 , 5 × 5 , 3 × 3
Hilbert et al. (2018) [29]Multi-Scale and
-Orientation RFNN
DenseNetGaussian derivative basis
11 × 11 , 5 × 5
Karantzas et al.
(2019) [33]
ResnetRFResNet18, ResNet34,
ResNet50
directional
Parseval frames
with compact support
7 × 7 , 3 × 3
Labate et al. (2019) [32]Geometric-Biased
CNN
multi-layer CNNdirectional
Parseval frames
with compact support
5 × 5 , 3 × 3
Ulicny et al.
(2019, 2019, 2022) [34,35,36]
Harmonic Networkmulti-layer CNN,
Wide residual network,
SE-ResNeXt,
Resnet-50, Resnet-101
DCT basis
(various)
Kumawat et al.
(2020) [37]
Depthwise-STFT
Based Separable
Convolutional
Neural Network
MobileNet,
ShuffleNet, ReLPU
Short-Term
Fourier
Transform kernels
5 × 5 , 3 × 3
Pintea et al. (2021) [28]N-JetNetNetwork in Network,
ALLCNN,
Resnet-32, Resnet-110,
EfficientNet
Gaussian derivative basis
(dynamic size)
Tomen et al. (2021) [38]Deep Continuous
Network
ResNet-34,
ODE-Net
Gaussian derivative basis
(dynamic size)
Saldanha et al. (2021) [39]FracSRF
Network
Network in Network,
Resnet-32,
EfficientNet-b0
Gaussian derivative basis
(dynamic size)
Lindeberg (2021, 2022) [40,41]Scale-covariant
and Scale-invariant
Gaussian Derivative
Network
multi-layer CNNscale-normalized
Gaussian derivative basis
(various)
Fukuzaki et al. (2022) [42]OtXVGG-16, DenseNet-121,
EfficientNetV2-S,
NFNet- F0
Principal Component
Analysis of well-trained
filter weights 3 × 3
Penaud–Polge et al. (2022) [43]Fully Trainable
Gaussian Derivative
Network
VGG-16,
U-Net
oriented and shifted
Gaussian derivative kernels
(various)
Table 2. Baseline reference neural network architecture ( R F N N r e f ) with 3 receptive field convolutional layers used in the classification task on the MNIST dataset. The output dimension of individual computational layers is denoted in the form N @ H × W × D , where N is the number of objects with height, width and depth, accordingly.
Table 2. Baseline reference neural network architecture ( R F N N r e f ) with 3 receptive field convolutional layers used in the classification task on the MNIST dataset. The output dimension of individual computational layers is denoted in the form N @ H × W × D , where N is the number of objects with height, width and depth, accordingly.
Layer Number
and Type
Inner Computational
Layers
HyperparametersOutput ShapeLearnable
Parameters
#0 Input-- N @ 28 × 28 × 1 0
#1 RFConvDepthwise convolutionfixed Gaussian kernels
of shape 10 @ 11 × 11 × 1 ,
stride = 1, padding = 5
N @ 28 × 28 × 10 0
Pointwise convolutionnumber of kernels = 64,
stride = 1, padding = 0
N @ 28 × 28 × 64 640
Max poolingkernel size = 3 × 3 ,
stride = 2
N @ 13 × 13 × 64 0
ReLu- N @ 13 × 13 × 64 0
Local response
normalization
size = 9, α = 1 × 10 4 ,
β = 0.75 , k = 2.0
N @ 13 × 13 × 64 0
Dropoutp = 0.2 N @ 13 × 13 × 64 0
#2 RFConvDepthwise convolutionfixed Gaussian kernels
of shape 6 @ 7 × 7 × 64 ,
stride = 1, padding = 6
N @ 19 × 19 × 384 0
Pointwise convolutionnumber of kernels = 64,
stride = 1, padding = 0
N @ 19 × 19 × 64 24,576
Max poolingkernel size = 3 × 3 ,
stride = 2
N @ 9 × 9 × 64 0
ReLu- N @ 9 × 9 × 64 0
Local response normalizationsize = 9, α = 1 × 10 4 , β = 0.75 , k = 2.0 N @ 9 × 9 × 64 0
Dropoutp = 0.2 N @ 9 × 9 × 64 0
#3 RFConvDepthwise convolutionfixed Gaussian kernels
of shape 6 @ 7 × 7 × 64 ,
stride = 1, padding = 6
N @ 15 × 15 × 384 0
Pointwise convolutionnumber of kernels = 64,
stride = 1, padding = 0
N @ 15 × 15 × 64 24,576
Max poolingkernel size = 3 × 3 ,
stride = 2
N @ 7 × 7 × 64 0
ReLu- N @ 7 × 7 × 64 0
Local response
normalization
size = 9, α = 1 × 10 4 ,
β = 0.75 , k = 2.0
N @ 7 × 7 × 64 0
Dropoutp = 0.7 N @ 7 × 7 × 64 0
#4 FlattenReshape- N @ 3136 0
#5 LinearFully-connected- N @ 10 31,360
#6 OutputSoftmax- N @ 10 0
Table 3. Determined number of epochs of individual experiments for a limited number of samples on the MNIST database.
Table 3. Determined number of epochs of individual experiments for a limited number of samples on the MNIST database.
Number of samples60,00040,00020,00010,000500020001000300
Number of epochs1001001001001502003001000
Table 5. Simplified neural network architecture R F N N L 1 with one receptive field convolutional layer.
Table 5. Simplified neural network architecture R F N N L 1 with one receptive field convolutional layer.
Layer Number and TypeInner Computational LayersHyperparametersOutput ShapeLearnable
Parameters
#0 Input-- N @ 28 × 28 × 1 0
#1 RFConvDepthwise convolutionfixed Gaussian kernels
of shape 10 @ 11 × 11 × 1 ,
stride = 1, padding = same
N @ 28 × 28 × 10 0
Pointwise convolutionnumber of kernels = 64,
stride = 1, padding = 0
N @ 28 × 28 × 64 640
Max poolingkernel size = 3 × 3 ,
stride = 2
N @ 13 × 13 × 64 0
ReLu- N @ 13 × 13 × 64 0
#4 FlattenReshape- N @ 10 , 816 0
#5 LinearFully connected- N @ 10 108,160
#6 OutputSoftmax- N @ 10 0
Table 6. Results of test classification accuracies on MNIST database obtained using simplified R F N N L 1 architecture with various fixed bases of the same size for 10 selected filters. The best results are highlighted in bold.
Table 6. Results of test classification accuracies on MNIST database obtained using simplified R F N N L 1 architecture with various fixed bases of the same size for 10 selected filters. The best results are highlighted in bold.
BaseMean
Accuracy [ % ]
Min
Accuracy [ % ]
Max
Accuracy [ % ]
Variance [ % 2 ]
Gaussian89.6786.3091.751.0131
Discrete Cosine (DCTII)90.2188.4591.860.7268
Discrete Hartley (DHT)90.2088.4791.760.5958
Random (RND)88.4886.1590.090.9813
Orthonormal random
(ORTHN_RND)
89.3287.3690.650.5528
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Goga, J.; Vargic, R.; Pavlovicova, J.; Kajan, S.; Oravec, M. Structure and Base Analysis of Receptive Field Neural Networks in a Character Recognition Task. Sensors 2022, 22, 9743. https://doi.org/10.3390/s22249743

AMA Style

Goga J, Vargic R, Pavlovicova J, Kajan S, Oravec M. Structure and Base Analysis of Receptive Field Neural Networks in a Character Recognition Task. Sensors. 2022; 22(24):9743. https://doi.org/10.3390/s22249743

Chicago/Turabian Style

Goga, Jozef, Radoslav Vargic, Jarmila Pavlovicova, Slavomir Kajan, and Milos Oravec. 2022. "Structure and Base Analysis of Receptive Field Neural Networks in a Character Recognition Task" Sensors 22, no. 24: 9743. https://doi.org/10.3390/s22249743

APA Style

Goga, J., Vargic, R., Pavlovicova, J., Kajan, S., & Oravec, M. (2022). Structure and Base Analysis of Receptive Field Neural Networks in a Character Recognition Task. Sensors, 22(24), 9743. https://doi.org/10.3390/s22249743

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop