Next Article in Journal
A Novel Hyperchaotic 2D-SFCF with Simple Structure and Its Application in Image Encryption
Previous Article in Journal
Optimized Ship-Radiated Noise Feature Extraction Approaches Based on CEEMDAN and Slope Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Enhanced Scheme for Reducing the Complexity of Pointwise Convolutions in CNNs for Image Classification Based on Interleaved Grouped Filters without Divisibility Constraints

by
Joao Paulo Schwarz Schuler
1,*,
Santiago Romani Also
1,
Domenec Puig
1,
Hatem Rashwan
1 and
Mohamed Abdel-Nasser
1,2
1
Departament d’Enginyeria Informatica i Matemátiques, Universitat Rovira i Virgili, 43007 Tarragona, Spain
2
Electronics and Communication Engineering Section, Electrical Engineering Department, Aswan University, Aswan 81528, Egypt
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(9), 1264; https://doi.org/10.3390/e24091264
Submission received: 24 July 2022 / Revised: 1 September 2022 / Accepted: 5 September 2022 / Published: 8 September 2022
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
In image classification with Deep Convolutional Neural Networks (DCNNs), the number of parameters in pointwise convolutions rapidly grows due to the multiplication of the number of filters by the number of input channels that come from the previous layer. Existing studies demonstrated that a subnetwork can replace pointwise convolutional layers with significantly fewer parameters and fewer floating-point computations, while maintaining the learning capacity. In this paper, we propose an improved scheme for reducing the complexity of pointwise convolutions in DCNNs for image classification based on interleaved grouped filters without divisibility constraints. The proposed scheme utilizes grouped pointwise convolutions, in which each group processes a fraction of the input channels. It requires a number of channels per group as a hyperparameter C h . The subnetwork of the proposed scheme contains two consecutive convolutional layers K and L, connected by an interleaving layer in the middle, and summed at the end. The number of groups of filters and filters per group for layers K and L is determined by exact divisions of the original number of input channels and filters by C h . If the divisions were not exact, the original layer could not be substituted. In this paper, we refine the previous algorithm so that input channels are replicated and groups can have different numbers of filters to cope with non exact divisibility situations. Thus, the proposed scheme further reduces the number of floating-point computations (11%) and trainable parameters (10%) achieved by the previous method. We tested our optimization on an EfficientNet-B0 as a baseline architecture and made classification tests on the CIFAR-10, Colorectal Cancer Histology, and Malaria datasets. For each dataset, our optimization achieves a saving of 76%, 89%, and 91% of the number of trainable parameters of EfficientNet-B0, while keeping its test classification accuracy.

1. Introduction

In 2012, Krizhevsky et al. [1] reported a breakthrough in the ImageNet Large Scale Visual Recognition Challenge [2] using their AlexNet architecture, which contains 5 convolutional layers and 3 dense layers. Since 2012, many other architectures have been introduced, like ZFNet [3], VGG [4], GoogLeNet [5], ResNet [6] and DenseNet [7]. Since the number of layers of proposed convolutional neural networks has increased from 5 to more than 200, those models are usually referred to as “Deep Learning” or DCNN.
In 2013, Min Lin et al. introduced the Network in Network architecture (NiN) [8]. It has 3 spatial convolutional layers with 192 filters, separated by pairs of pointwise convolutional layers. These pointwise convolutions enable the architecture to learn patterns without the computational cost of a spatial convolution. In 2016, ResNet [6] was introduced. Following VGG [4], all ResNet spatial filters have 3 × 3 pixels. Their paper conjectures that deeper CNNs have exponentially low convergence rates. To deal with this problem, they introduced skip connections every 2 convolutional layers. In 2017, Ioannou et al. [9] adapted the NiN architecture to use 2 to 16 convolutional groups per layer for classifying the CIFAR-10 dataset.
A grouped convolution separates input channels and filters into groups. Each filter processes only input channels entering its group. Each group of filters can be understood as an independent (parallel) path for information flow. This aspect drastically reduces the number of weights in each filter and, therefore, reduces the number of floating-point computations. Grouping 3 × 3 and 5 × 5 spatial convolutions, Ioannou et al. were able to decrease the number of parameters by more than 50% while keeping the NiN classification accuracy. Ioannou et al. also adapted the Resnet-50, Resnet-200, and GoogleLeNet architectures applying 2 to 64 groups per layer when classifying the ImageNet dataset, obtaining parameter reduction while maintaining or improving the classification accuracy. Also in 2017, an improvement for the ResNet architecture called ResNeXt [10] was introduced, replacing the spatial convolutions with parallel paths (groups), reducing the number of parameters.
Several studies have also reported the creation of parameter-efficient architectures with grouped convolutions [11,12,13,14,15]. In 2019, Mingxing Tan et al. [16] developed the EfficientNet architecture. At that time, their EfficientNet-B7 variant was 8.4 times more parameter-efficient and 6.1 times faster than the best existing architecture, achieving 84.3% top-1 accuracy on ImageNet. More than 90% of the parameters of EfficientNets come from standard pointwise convolutions. This aspect opens an opportunity for a huge reduction in several parameters and floating-point operations, as we have exploited in the present paper.
Most parameters in DCNNs are redundant [17,18,19,20,21]. Pruning methods remove connections and neurons found to be irrelevant by different techniques. After training the original network with the full set of connections, the removal is carried out [22,23,24,25,26,27]. Our method differs from pruning as we reduce the number of connections before the training starts, while pruning does after training. Therefore, our method can save computing resources during training time.
In previous works [28,29], we proposed replacing standard pointwise convolutions with a sub-architecture that contains two grouped pointwise convolutional layers (K and L), an interleaving layer that mixes channels from layer K before feeding the layer L, and a summation at the end that sums the results from both convolutional layers. Our original method accepts a hyperparameter C h , which denotes the number of input channels fed to each group of filters. Then, our method computes the number of groups of filters and filters per group according to the division of original input channels and filters by C h . Our original method avoided substituting the layers where the divisions were not exact.
In this paper, we propose an enhanced scheme to allow computing the number of groups in a flexible manner, in the sense that the divisibility constraints do not have to be considered anymore. By applying our method to all pointwise convolutional layers of an EfficientNet-B0 architecture, we are able to reduce a huge amount of resources (trainable parameters, floating-point computations) while maintaining the learning capacity.
This paper is structured as follows: Section 2 details our improved solution for grouping pointwise convolutions while skipping the constraints of divisibility found in our previous method. Section 3 details the experiments carried out for testing our solution. Section 4 summarizes the conclusions and limitations of our proposal.

2. Methodology

2.1. Mathematical Ground for Regular Pointwise Convolutions

Let X i = { x 1 i , x 2 i , , x I c i i } be a set of input feature maps (2D lattices) for a convolutional layer i in a DCNN, where I c i denotes the number of input channels for this layer. Let W i = { w 1 i , w 2 i , , w F i i } be a set of filters containing the weights for convolutions, where F i denotes the number of filters at layer i, which is also the number of output channels of this layer. Following the notation proposed in [30], a regular DCNN convolution can be mathematically expressed as in Equation (1):
X i + 1 = W i X i = { w 1 i X i , w 2 i X i , , w F i i X i }
where the ⨂ operator indicates that filters in W i are convolved with feature maps in X i , using the ∗ operator to indicate a 3D tensor multiplication and shifting of a filter w j i across all patches of the size of the filter in all feature maps. For simplicity, we are ignoring the bias terms. Consequently, X i + 1 will contain F i feature maps that will feed the next layer i + 1 . The tensor shapes of involved elements are the following:
X i R H × W × I c i W i R F i × S × S × I c i w j i R S × S × I c i X i + 1 R H × W × F i
where H × W is the size (height, width) of feature maps, and S × S is the size of a filter (usually square). In this paper we work with S = 1 because we are focused on pointwise convolutions. In this case, each filter w j i carries I c i weights. The total number of weights P i in layer i is obtained with a simple multiplication:
P i = I c i · F i

2.2. Definition of Grouped Pointwise Convolutions

For expressing a grouped pointwise convolution, let us split the input feature maps and the set of filters in G i groups, as X i = X 1 i , X 2 i , , X G i i and W i = W 1 i , W 2 i , , W G i i . Assuming that both I c i and F i are divisible by G i , the elements in X i and W i can be evenly distributed through all their subset X j i and W j i . Then, Equation (1) can be reformulated as Equation (4):
X i + 1 = W 1 i X 1 i , W 2 i X 2 i , , W G i i X G i i
The shapes of the subsets are the following:
X m i R H × W × I c i G i W m i R F g i × 1 × 1 × I c i G i w j i , m R 1 × 1 × I c i G i
where F g i is the number of filters per group, namely, F g i = F i G i . Since each filter w j i , m only convolves on a fraction of input channels ( I c i G i ) , the total number of weights per subset W m i is F i G i · I c i G i . Multiplying the last expression by the number of groups provides the total number of weights P i ¯ in a grouped pointwise convolutional layer i:
P i ¯ = ( I c i · F i ) / G i
Equation (6) shows that the number of trainable parameters is inversely proportional to the number of groups. However, grouping has the evident drawback that it prevents the filters to be connected with all input channel, which reduces the possible connections of input channels for learning new patterns. As it may lead to a lower learning capacity of the DCNN, one must be cautious with using such grouping technique.

2.3. Improved Scheme for Reducing the Complexity of Pointwise Convolutions

Two major limitations of our previous method were inherited from constraints found in most deep learning APIs:
  • The number of input channels I c i must be multiple of the number of groups G i .
  • The number of filters F i must be multiple of the number of groups G i .
The present work circumvents the first limitation by replicating channels from the input. The second limitation is circumvented by adding a second parallel path with another pointwise grouped convolution when required. Figure 1 shows an example of our updated architecture.
Details of this process are described below, which is applied to substitute each pointwise convolutional layer i found in the original architecture. To explain the method, we start detailing the construction of the layer K shown in Figure 1. For simplicity, we drop the index i and use the index K to refer to the original hyperparameters, i.e., we use I c K instead of I c i , F K instead of F i . Also, we will use the indexes K 1 and K 2 to refer the parameters of the two parallel paths that may exist in layer K.
First of all, we must manually specify the value of the hyperparameter C h . In the graphical example shown in Figure 1, we set C h = 4 . The rest of hyperparameters such as number of groups in layers K and L are determined automatically by the rules of our algorithm, according to the chosen value of C h , the number of input channels I c K and the number of filters F K . We do not have a procedure to find the optimal value of C h , hence we must apply ablation studies on a range of C h values as shown in the results section. For the example in Figure 1, we have chosen the value of C h to obtain a full variety of situations that must be tackled by our algorithm, i.e., non-divisibility conditions.

2.4. Definition of Layer K

The first step of the algorithm is to compute the number of groups in branch K1, as in Equation (7):
G K 1 = I c K C h
Since the number of input channels I c K may not be divisible by Ch, we use the ceiling operator on the division to obtain an integer number of groups. In the example, G K 1 = 14 / 4 = 4 . Thus, the output of filters in branch K1 can be defined as in (8):
K 1 = W 1 K 1 X 1 K , W 2 K 1 X 2 K , , W G K 1 K 1 X G K 1 K
The subsets X m K are composed of input feature maps x j , collected in a sorted manner, i.e., X 1 K = x 1 , x 2 , , x C h , X 2 K = x C h + 1 , x C h + 2 , , x 2 C h , etc. Equation (9) provides a general definition of which feature maps x j are included in any feature subset X m K :
X m K = x a + 1 , x a + 2 , , x a + C h , a = ( m 1 ) · C h
However, if I c K is not divisible by C h , the last group m = G K 1 would not have C h channels. In this case, the method will complete this last group replicating C h b initial input channels, where b is computed as stated in Equation (10):
X G K 1 K = x a + 1 , x a + 2 , , x a + b , x 1 , x 2 , , x C h b , a = G K 1 1 · C h , b = G K 1 · C h I c K
It can be proved that b will always be less or equal than C h , since b is the excess of the integer division I c K / C h , i.e., G K 1 · C h will always be above or equal to I c K , but less than I c K + C h , because otherwise G K 1 would increase its value (as a quotient of I c K / C h ). In the example, b = 2 , hence X 4 K 1 = x 13 , x 14 , x 1 , x 2 .
Then, the method calculates the number of filters per group F g K 1 as in (11):
F g K 1 = F K G K 1
To avoid divisibility conflicts, this time we have chosen the floor integer division. For the first path K1, each of the filter subsets shown in (8) will contain the following filters:
W m K 1 = w 1 K 1 , m , w 2 K 1 , m , , w F g K 1 K 1 , m w j K 1 , m R 1 × 1 × C h
For the first path of the example, the number of filters per group is F g K 1 = 10 / 4 = 2 . So, the first path has 4 groups ( G K 1 ) of 2 filters ( F g K 1 ), each filter being connected to 4 input channels ( C h ).
If F K is not divisible by C h , a second path K2 will provide as many groups as filters not provided in K1, with one filter per group, to complete the total number of filters F K :
G K 2 = F K F g K 1 · G K 1 F g K 2 = 1
In the example, G K 2 = 2 . The required input channels for the second path is C h · G K 2 . The method obtains those channels reusing the same subsets of input feature maps X m K shown in (9). Hence, the output of filters in path K 2 can be defined as in (14):
K 2 = w 1 K 2 X 1 K , w 2 K 2 X 2 K , , w G K 2 K 2 X G K 2 K
where w j K 2 R 1 × 1 × C h . Therefore, each filter in K 2 operates on exactly the same subset of input channels than the corresponding subset of filters in K 1 . Hence, each filter in the second path can be considered as belonging to one of the groups of the first path.
It must be noticed that G K 2 will always be less than G K 1 . This is true because G K 2 is the reminder of the integer division F K / G K 1 , as can be deduced from (11) and (13). This property warranties that there will be enough subsets X m K for this second path.
After defining paths K1 and K2 in layer K, the output of this layer is the concatenation of both paths:
K = K 1 , K 2
The total number of channels after the concatenation is equal to F K = G K 1 · F g K 1 + G K 2 .

2.5. Interleaving Stage

As mentioned above, grouped convolutions inherently face a limitation: each parallel group of filters computes its output from their own subset of input channels, preventing combinations of channels connected to different groups. To alleviate this limitation, we propose to interleave the output channels from the convolutional layer K.
The interleaving process simply consists in arranging the odd channels first and the even channels last, as noted in Equation (16):
I K = { k 1 , k 3 , k 5 , , k 2 c 1 , k 2 , k 4 , k 6 , , k 2 c } c = F K / 2
Here we are assuming that F K is even. Otherwise, the list of odd channels will include an extra channel k 2 c + 1 .

2.6. Definition of Layer L

The interleaved output feeds the grouped convolutions in layer L to process data coming from more than one group from the preceding layer K.
To create layer L, we apply the same algorithm as for layer K, but now the number of input channels is equal to F K instead of I c K .
The number of groups in path L1 is computed as:
G L 1 = F K C h
Note that G L 1 may not be equal to G K 1 . In the example, G L 1 = 10 / 4 = 3 .
Then, the output of L1 is computed as in (18), where the input channel groups I m K come from the interleaving stage. Each group is composed of C h channels, whose indexes are generically defined in (19):
L 1 = W 1 L 1 I 1 K , W 2 L 1 I 2 K , , W G L 1 K 1 I G L 1 K
I m K = i a + 1 K , i a + 2 K , , i a + C h K , a = ( m 1 ) · C h
Again, the last group of indexes may not contain C h channels due to a non-exact division condition in (17). Similar to path K1, for path L1 the missing channels in the last group will be supplied by replicating C h b initial interleaved channels, where b is computed as stated in Equation (20):
I G L 1 K = i a + 1 K , i a + 2 K , , i a + b K , i 1 K , i 2 K , , i C h b K , a = G L 1 1 · C h , b = G L 1 · C h F K
The number of filters per group F g L 1 is computed as in (21):
F g L 1 = F K G L 1
In the example, F g L 1 = 10 / 3 = 3 . Each group of filters W m L 1 shown in (18) can be defined as in (22), each one containing F g L 1 convolutional filters of C h inputs:
W m L 1 = w 1 L 1 , m , w 2 L 1 , m , , w F g L 1 L 1 , m w j L 1 , m R 1 × 1 × C h
It should be noted that if the division in (21) is not exact, the number of output channels from layer L may not reach the required F K outputs. In this case, a second path L2 will be added, with the following parameters:
G L 2 = F K F g L 1 · G L 1 F g L 2 = 1
In the example, G L 2 = 1 . The output of path L2 is computed as in (24), defining one extra convolutional filter for some initial groups of interleaved channels declared in (18) and (19), taking into account that G L 2 will always be less than G L 1 according to the same reasoning done for G K 2 and G K 1 :
L 2 = w 1 L 2 I 1 K , w 2 L 2 I 2 K , , w G L 2 L 2 I G L 2 K
The last step in defining the output of layer L is to join the outputs of paths L1 and L2:
L = L 1 , L 2

2.7. Joining of Layers

Finally, the output of both convolutional layers K and L are summed to create the output of the original layer:
X i + 1 = K + L
Compared to concatenation, summation has the advantage of allowing a residual learning in the filters of layer L, because gradient can be backpropagated through L filters or directly to K filters. In other words, residual layers provide more learning capacity with low degree of downsides due to increasing the number of layers (i.e., overfitting, longer training time, etc.) In the results section, we present an ablation study that contains experiments done without the interleaving and the L layers (rows labeled with “no L”). These experiments empirically prove that the interleaving mechanism and the secondary L layer help in improving the sub-architecture accuracy, with low impact.
It is worth mentioning that we only add the layer L an the interleaving when the number of input channels is bigger or equal to the number of filters in layer K.

2.8. Computing the Number of Parameters

We can compute the total number of parameters of our sub-architecture. First, Equation (27) shows that the number of filters in layer K is equal to the number of filters in layer L, which in turn is equal to the total number of filters in the original convolutional layer F i :
F g K 1 · G K 1 + G K 2 = F g L 1 · G L 1 + G L 2 = F i
Then, the total number of parameters P i ¯ ¯ is twice the number of original filters multiplied by the number of input channels per filter:
P i ¯ ¯ = 2 ( F i · C h )
Therefore, comparing Equation (28) with (3), it is clear that C h must be significantly less than I c i / 2 to reduce the number of parameters of a regular pointwise convolutional layer. Also, comparing Equation (28) with (6), our sub-architecture provides a parameter reduction similar to a plain grouped convolutional layer when C h is around I c i / 2 · G i , although we cannot specify a general G i term because of the complexity of our pair of layers with possibly two paths per layer.
The requirement for a low value of C h is also necessary to ensure that divisions in Equations (7) and (17) provide quotients above one, otherwise our method will not create grouping. Hence, C h must be less or equal to I c i / 2 and F i / 2 . These are the only two constraints that our method is restricted by.
As shown in Table 1, pointwise convolutional layers found in real networks such as EfficientNet-B0 have significant Figures for I c i and F i , either hundreds or thousands. Therefore, values of C h less or equal than 32 will ensure a good ratio of parameter reduction for most of these pointwise convolutional layers.
EfficientNet is one of the most complex (but efficient) architectures that can be found in the literature. To our method, the degree of complexity of a DCNN is mainly related to the maximum number of input channels and output features in any pointwise convolutional layer. Our method does not care about the number of layers, neither in depth nor in parallel, because it works on each layer independently. Therefore, the degree of complexity of EfficientNet-B0 can be considered significantly high, taking into account the values shown in the last row of Table 1. Arguably, other versions of EfficientNet (B1, B2, etc.) and other types of DCNN can exceed those values. In such cases, higher values of C h may be necessary, but we cannot provide any rule to forecast its optimum value for the configuration of any pointwise convolutional layer.

2.9. Activation Function

In 2018, Prajit et al. [31] tested a number of activation functions. In their experimentation, they found that the best performing one was the so-called “swish”, shown in Equation (29).
f ( x ) = x · s i g m o i d ( β x )
In previous works [28,29], we used the ReLU activation function. In this work, we use the swish activation function. This change gives us better results in our ablation experiments shown on Table 5.

2.10. Implementation Details

We tested our optimization by replacing original pointwise convolutions in the EfficientNet-B0 and named it as “kEffNet-B0 V2”. With CIFAR-10, we tested an additional modification that skips the first 4 convolutional strides, allowing input images with 32 × 32 pixels instead of the original resolution of 224 × 224 pixels.
In all our experiments, we saved the trained network from the epoch that achieved the lowest validation loss for testing with the test dataset. Convolutional layers are initialized with Glorot’s method [32]. All experiments were trained with RMSProp optimizer, data augmentation [33] and cyclical learning rate schedule [34]. We worked with various configurations of hardware with NVIDIA video cards. Regarding software, we did our experiments with K-CAI [35] and Keras [36] on the top of Tensorflow [37].
Our source code is publicly available at: https://github.com/joaopauloschuler/kEffNetV2/, accessed on 1 September 2022.

2.11. Horizontal Flip

In some experiments, we run the model twice with the input image and its horizontally flipped version. The output from the softmax from both runs is summed before class prediction. In these experiments, the number of floating-point computations doubles, although the number of trainable parameters remains the same.

3. Results and Discussion

In this section, we present and discuss the results of the proposed scheme with three image classification datasets: CIFAR-10 dataset [38], Malaria dataset, and colorectal cancer histology dataset [39,40].

3.1. Results on the CIFAR-10 Dataset

The CIFAR-10 dataset [38] is a subset of [41] and consists of 60k 32 × 32 images belonging to 10 different classes: airplane, automobile, bird, cat, deer, dog, frog, horse ship and truck. These images are taken from natural and uncontrolled lightning environment. They contain only one prominent instance of the object to which the class refers to. The object may be partially occluded or seen from an unusual viewpoint. This dataset has 50k images for training and 10k images for test. We picked 5k images for validation and left the training set with 45k images. We run experiments with 50 and 180 epochs.
On Table 2 we compare kEffNet-B0 V1 (our previous method) and V2 (our current method), for two values of C h . We can see that our V2 models has slightly more reduction in both number of parameters and floating-point computations than the V1 counterpart models, while achieving slightly higher accuracy. Specifically, V2 models save 10% of the parameters (from 1,059,202 to 950,650) and 11% of the floating-point computations (from 138,410,206 to 123,209,110) of V1 models. All of our variants obtain similar accuracy to the baseline with a remarkable reduction of resources (at least 26.3% of trainable parameters and 35.5% of computations).
As the scope of this work is limited to small datasets and small architectures, we only experimented with the smallest EfficientNet variant (EfficientNet-B0) and our modified variant (kEffNet-B0). Nevertheless, Table 3 provides the number of trainable parameters of the other EfficientNet variants (original and parameter-reduced). Equation (3) indicates that the number of parameters grows with the number of filters and the number of input channels. Equation (6) indicates that the number of parameters decreases with the number of groups. As we create more groups when the number of input channels grows, we expect to find bigger parameter savings on larger models. This saving can be seen on Table 3.
We also tested our kEffNet-B0 with 2, 4, 8, 16 and 32 channels per group for 50 epochs as shown in Table 4. As expected, the test classification accuracy increases when allocating more channels per group: from 84.26% for C h = 2 to 93.67% for C h = 32. Also, the resource saving decreases as the number of channels per group increase: from 7.8% of parameters and 11.4% of computations for C h = 2 to 23.6% of parameters and 31.6% of computations for C h = 32 (compared to the baseline). For CIFAR-10, if we aim to achieve an accuracy comparable to the baseline, we must choose at least 16 channels per group. If we add an extra run per image sample with horizontal flipping when training kEffNet-B0 V2 32ch, the classification accuracy increases from 93.67% to 94.01%.
Table 5 replicates most of the results shown in Table 4, but comparing the effect of not including layer L and interleaving, and also substituting the swish activation function with the typical ReLU. As can be observed, disabling layer L has a noticeable degradation on test accuracy when the values of C h are smaller. For example, when C h = 4, the performance drops more than 5%. On the other hand, when C h = 32 the drop is less than 0.5%. This is logical taking into account that, the more channels are included per group, the more chances are to combine input features in the filters. Therefore, a second layer and the corresponding interleaving is not as crucial as when the filters of layer K are fed with fewer channels.
In the comparison of activation functions, the same effect can be appreciated: the swish function works better than the ReLU function, but provides less improvement for larger number of channels per group. Nevertheless, the gain in the least difference case (32 ch) is still profitable, with more than 1.5% of extra test accuracy when using the swish activation function.
Table 6 shows the effect in accuracy when classifying the CIFAR-10 dataset with EfficientNet-B0 and our kEffNet-B0 V2 32ch variant for 180 epochs instead of 50 epochs. The additional training epochs assign slightly higher test accuracy to the baseline than to our core variant. When adding horizontal flipping, our variant has slightly surpassed the baseline results. Nevertheless, all three results can be considered similar to each other, but our variant offers a significant saving in parameters and computations. Although the H flipping doubles the computational cost of our core variant, it still remains only a fraction (63.3%) of the baseline computational cost.

3.2. Results on the Malaria Dataset

The Malaria dataset [40] has 27,558 cell images from infected and healthy cells separated into 2 classes. There is the same number of images for healthy and infected cells. From the original 27,558 images set, we separated 10% of the images (2756 images) for validation and another 10% for testing. On all training, validation, and test subsets, there are 50% of healthy cell images. We quadruplicated the number of validation images by flipping these images horizontally and vertically, resulting in 11,024 images for validation.
On this dataset, we tested our kEffNet-B0 with 2, 4, 8, 12, 16, and 32 channels per group, as well as the baseline architecture, as shown in Table 7. Our variants have from 7.5% to 23.5% of the trainable parameters and from 15.7% to 42.2% of the computations allocated by the baseline architecture. Although the worst classification accuracy was found with the smallest variant (2ch), its classification accuracy is less than 1% inferior to the best performing variant (16ch) and only 0.69% below the baseline performance. With only 8 channels per group, our method equals the baseline accuracy with a small portion of the parameters (10.8%) and computations (22.5%) required by the baseline architecture. Curiously, our 32ch variant is slightly worse than the 16ch variant, but still better than the baseline. It is an example that a rather low complexity of the input images may require less channels per filter (and more parallel groups of filters), to optimally capture the relevant features of images.

3.3. Results on the Colorectal Cancer Histology Dataset

The collection of samples in colorectal cancer histology dataset [39] contains 5000 150 × 150 images separated into 8 classes: adipose, complex, debris, empty, lympho, mucosa, stroma, and tumor. Similar to what we did with the Malaria dataset, we separated 10% of the images for validation and another 10% for testing. We also quadruplicated the number of validation images by flipping these images horizontally and vertically.
On this dataset, we tested our kEffNet-B0 with 2, 4, 8, 12, and 16 channels per group, as well as the baseline architecture, as shown in Table 8. Similar to the Malaria dataset, higher values of channels per group do not lead to better performance. In this case, the variants with the highest classification accuracy are 4ch and 8ch, achieving 98.02% of classification accuracy, outperforming the baseline accuracy in 0.41%. The 16ch variant has obtained the same accuracy than the 2ch variant, but doubling the required resources. Again, it indicates that the complexity of the images plays a role in the selection of the optimal number of channels per group. In other words, simpler images may require less channels per group. Unfortunately, the only method we know to find out this optimal value is performing theses scanning experiments.

4. Conclusions and Future Work

This paper presented an efficient scheme for decreasing the complexity of pointwise convolutions in DCNNs for image classification based on interleaved grouped filters with no divisibility constraints. From our experiments, we can conclude that connecting all input channels from the previous layer to all filters is unnecessary: grouped convolutional filters can achieve the same learning power with a small fraction of resources (1/3 of floating-point computations, 1/4 of parameters). Our enhanced scheme avoids the divisibility contraints, furter reducing the required resources (up to 10% less) while maintaining or slightly surpassing the accuracy of our previous method.
We have made ablation studies to obtain the optimal number of channels per group for each dataset. For colorectal cancer dataset, this number is surprisingly low (4 channels per group). On the other side, for CIFAR-10 the best results require at least 16 channels per group. This fact indicates that the complexity of the input images affects the optimal configuration of our sub-architecture.
As the main limitation of our method, it cannot determine the optimal number of channels per group automatically, according to the complexity of each pointwise convolutional layer to be substituted and the complexity of input images. A second limitation is that the same number of channels per group is applied to all pointwise convolutional layers of the target architecture, regardless of the specific complexity of each layer. This limitation could be easily tackled by setting C h as a fraction of the total number of parameters of each layer. This is a straightforward task for future research. Besides, we will apply our method to different problems, such as instance and semantic image segmentation, developing an efficient deep learning-based seismic acoustic impedance inversion method [42], object detection, and forecasting.

Author Contributions

Conceptualization, J.P.S.S. and S.R.A.; methodology, J.P.S.S. and S.R.A.; software, J.P.S.S.; validation, S.R.A.; formal analysis, S.R.A.; investigation, J.P.S.S.; resources, D.P.; data curation, J.P.S.S.; writing—original draft preparation, J.P.S.S. and S.R.A.; writing—review and editing, J.P.S.S., S.R.A. and M.A.-N.; visualization, J.P.S.S.; supervision, S.R.A., M.A.-N., H.R. and D.P.; project administration, D.P.; funding acquisition, D.P. All authors have read and agreed to the published version of the manuscript.

Funding

The Spanish Government partly supported this research through Project PID2019-105789RB-I00.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets used in this study are publicly available: CIFAR-10 [38], Colorectal cancer histology [39] and Malaria [40]. Software APIs are also publicly available: K-CAI [35] and Keras [36]. Our source code and raw experiment results are publicly available: https://github.com/joaopauloschuler/kEffNetV2, accessed on 1 September 2022.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
APIApplication Programming Interface
DCNNDeep Convolutional Neural Network
NiNNetwork in Network

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 1097–1105. [Google Scholar]
  2. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef]
  3. Zeiler, M.D.; Fergus, R. Visualizing and Understanding Convolutional Networks. In Computer Vision— ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 818–833. [Google Scholar]
  4. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR, San Diego, CA, USA, 7–9 May 2015; Conference Track Proceedings. Bengio, Y., LeCun, Y., Eds.; 2015. [Google Scholar]
  5. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; IEEE Computer Society: Los Alamitos, CA, USA, 2015; pp. 1–9. [Google Scholar] [CrossRef]
  6. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: Los Alamitos, CA, USA, 2016; pp. 770–778. [Google Scholar] [CrossRef]
  7. Huang, G.; Liu, Z.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
  8. Lin, M.; Chen, Q.; Yan, S. Network In Network. arXiv 2014, arXiv:cs.NE/1312.4400. [Google Scholar]
  9. Ioannou, Y.; Robertson, D.P.; Cipolla, R.; Criminisi, A. Deep Roots: Improving CNN Efficiency with Hierarchical Filter Groups. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5977–5986. [Google Scholar] [CrossRef]
  10. Xie, S.; Girshick, R.B.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  11. Zhang, T.; Qi, G.; Xiao, B.; Wang, J. Interleaved Group Convolutions for Deep Neural Networks. arXiv 2017, arXiv:1707.02725. [Google Scholar]
  12. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  13. Sun, K.; Li, M.; Liu, D.; Wang, J. IGCV3: Interleaved Low-Rank Group Convolutions for Efficient Deep Neural Networks. In Proceedings of the BMVC, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  14. Huang, G.; Liu, S.; Maaten, L.v.d.; Weinberger, K.Q. CondenseNet: An Efficient DenseNet Using Learned Group Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2752–2761. [Google Scholar] [CrossRef]
  15. Yu, C.; Xiao, B.; Gao, C.; Yuan, L.; Zhang, L.; Sang, N.; Wang, J. Lite-HRNet: A Lightweight High-Resolution Network. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10435–10445. [Google Scholar] [CrossRef]
  16. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. Int. Conf. Mach. Learn. 2019, 97, 6105–6114. [Google Scholar]
  17. Denil, M.; Shakibi, B.; Dinh, L.; Ranzato, M.; de Freitas, N. Predicting Parameters in Deep Learning. In NIPS’13, Proceedings of the 26th International Conference on Neural Information Processing Systems, Lake Tahoe, CA, USA, 5–10 December 2013; Curran Associates Inc.: Red Hook, NY, USA, 2013; Volume 2, pp. 2148–2156. [Google Scholar]
  18. Cheng, Y.; Yu, F.X.; Feris, R.S.; Kumar, S.; Choudhary, A.N.; Chang, S. An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; IEEE Computer Society: Washington, DC, USA, 2015; pp. 2857–2865. [Google Scholar] [CrossRef]
  19. Yang, W.; Jin, L.; Sile, W.; Cui, Z.; Chen, X.; Chen, L. Thinning of Convolutional Neural Network with Mixed Pruning. IET Image Process. 2019, 13, 779–784. [Google Scholar] [CrossRef]
  20. Kahatapitiya, K.; Rodrigo, R. Exploiting the Redundancy in Convolutional Filters for Parameter Reduction. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 1409–1419. [Google Scholar] [CrossRef]
  21. Liebenwein, L.; Baykal, C.; Carter, B.; Gifford, D.; Rus, D. Lost in Pruning: The Effects of Pruning Neural Networks beyond Test Accuracy. Proc. Mach. Learn. Syst. 2021, 3, 93–138. [Google Scholar]
  22. LeCun, Y.; Denker, J.; Solla, S. Optimal Brain Damage. In Advances in Neural Information Processing Systems; Touretzky, D., Ed.; Morgan-Kaufmann: Burlington, MA, USA, 1989; Volume 2. [Google Scholar]
  23. Reed, R. Pruning algorithms-a survey. IEEE Trans. Neural Netw. 1993, 4, 740–747. [Google Scholar] [CrossRef] [PubMed]
  24. Zhuang, Z.; Tan, M.; Zhuang, B.; Liu, J.; Guo, Y.; Wu, Q.; Huang, J.; Zhu, J. Discrimination-aware Channel Pruning for Deep Neural Networks. In Advances in Neural Information Processing Systems 31; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; pp. 881–892. [Google Scholar]
  25. Han, S.; Mao, H.; Dally, W.J. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. In Proceedings of the International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  26. Baykal, C.; Liebenwein, L.; Gilitschenski, I.; Feldman, D.; Rus, D. Data-Dependent Coresets for Compressing Neural Networks with Applications to Generalization Bounds. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  27. Liebenwein, L.; Baykal, C.; Lang, H.; Feldman, D.; Rus, D. Provable Filter Pruning for Efficient Neural Networks. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  28. Schuler, J.; Romaní, S.; Abdel-nasser, M.; Rashwan, H.; Puig, D. Grouped Pointwise Convolutions Significantly Reduces Parameters in EfficientNet; IOS Press: Amsterdam, The Netherlands, 2021; pp. 383–391. [Google Scholar] [CrossRef]
  29. Schwarz Schuler, J.P.; Romani, S.; Abdel-Nasser, M.; Rashwan, H.; Puig, D. Grouped Pointwise Convolutions Reduce Parameters in Convolutional Neural Networks. MENDEL 2022, 28, 23–31. [Google Scholar] [CrossRef]
  30. Wang, X.; Kan, M.; Shan, S.; Chen, X. Fully Learnable Group Convolution for Acceleration of Deep Neural Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 9041–9050. [Google Scholar] [CrossRef]
  31. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  32. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. JMLR Workshop Conf. Proc. 2010, 9, 249–256. [Google Scholar]
  33. Shorten, C.; Khoshgoftaar, T. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  34. Smith, L.N. Cyclical Learning Rates for Training Neural Networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 464–472. [Google Scholar] [CrossRef]
  35. Schuler, J.P.S. K-CAI NEURAL API. 2021. Available online: https://zenodo.org/record/5810093#.YxnEvbRBxPY (accessed on 4 September 2022).
  36. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 1 January 2022).
  37. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. Software. 2015. Available online: tensorflow.org (accessed on 1 January 2022).
  38. Krizhevsky, A. Learning Multiple Layers of Features from Tiny Images; Technical Report; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
  39. Kather, J.N.; Zöllner, F.G.; Bianconi, F.; Melchers, S.M.; Schad, L.R.; Gaiser, T.; Marx, A.; Weis, C.A. Collection of Textures in Colorectal Cancer Histology. 2016. Available online: https://zenodo.org/record/53169#.YxnFTLRBxPY (accessed on 1 January 2022).
  40. Rajaraman, S.; Antani, S.; Poostchi, M.; Silamut, K.; Hossain, M.; Maude, R.; Jaeger, S.; Thoma, G. Pre-trained convolutional neural networks as feature extractors toward improved malaria parasite detection in thin blood smear images. PeerJ 2018, 6, e4568. [Google Scholar] [CrossRef] [PubMed]
  41. Torralba, A.; Fergus, R.; Freeman, W.T. 80 Million Tiny Images: A Large Data Set for Nonparametric Object and Scene Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1958–1970. [Google Scholar] [CrossRef] [PubMed]
  42. Shahbazi, A.; Monfared, M.S.; Thiruchelvam, V.; Fei, T.K.; Babasafari, A.A. Integration of knowledge-based seismic inversion and sedimentological investigations for heterogeneous reservoir. J. Asian Earth Sci. 2020, 202, 104541. [Google Scholar] [CrossRef]
Figure 1. A schematic diagram of our pointwise convolution replacement. This example replaces a pointwise convolution with 14 input channels and 10 filters. It contains two convolutional layers, K and L, one interleaving, and one summation layer. Channels surrounded by a red border represent replicated channels.
Figure 1. A schematic diagram of our pointwise convolution replacement. This example replaces a pointwise convolution with 14 input channels and 10 filters. It contains two convolutional layers, K and L, one interleaving, and one summation layer. Channels surrounded by a red border represent replicated channels.
Entropy 24 01264 g001
Table 1. For a standard pointwise convolution with I c input channels, F filters, P parameters and a given number of channels per group C h , this Table shows the calculated parameters for layers K and L: the number of groups G < l a y e r > < p a t h > and the number of filters per group F g < l a y e r > < p a t h > . The last 2 columns show the total number of parameters and its percentage with respect to the original layer.
Table 1. For a standard pointwise convolution with I c input channels, F filters, P parameters and a given number of channels per group C h , this Table shows the calculated parameters for layers K and L: the number of groups G < l a y e r > < p a t h > and the number of filters per group F g < l a y e r > < p a t h > . The last 2 columns show the total number of parameters and its percentage with respect to the original layer.
Original SettingsLayer KLayer LK+L Params
Ic F P Ch G K 1 Fg K 1 G K 2 G L 1 Fg L 1 G L 2 Total%
141014044223318057.14%
1603840614,4001610384000061,44010.00%
3257680000122,88020.00%
1921152221,184161296000018,4328.33%
326192000036,86416.67%
1152320368,64016724322016010,2402.78%
32368321032020,4805.56%
38406402,457,6001624021604016020,4800.83%
321205402032040,9601.67%
Table 2. Comparing EfficientNet-B0, kEffNet-B0 V1 and kEffNet-B0 V2 with CIFAR-10 dataset after 50 epochs.
Table 2. Comparing EfficientNet-B0, kEffNet-B0 V1 and kEffNet-B0 V2 with CIFAR-10 dataset after 50 epochs.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,020,358100.0%389,969,098100.0%93.33%
kEffNet-B0 V1 16ch639,70215.9%84,833,89021.8%92.46%
kEffNet-B0 V2 16ch623,22615.5%82,804,37421.2%92.61%
kEffNet-B0 V1 32ch1,059,20226.3%138,410,20635.5%93.61%
kEffNet-B0 V2 32ch950,65023.6%123,209,11031.6%93.67%
Table 3. Number of trainable parameters for EfficientNet, kEffNet V2 16ch and kEffNet V2 32ch with a 10 classes dataset.
Table 3. Number of trainable parameters for EfficientNet, kEffNet V2 16ch and kEffNet V2 32ch with a 10 classes dataset.
VariantEfficientNetkEffNet V2 16ch%kEffNet V2 32ch%
B04,020,358623,22615.50%950,65023.65%
B16,525,994968,71014.84%1,389,06221.29%
B27,715,084983,19812.74%1,524,59019.76%
B310,711,6021,280,61211.96%2,001,43018.68%
B417,566,5461,858,44010.58%2,911,05216.57%
B528,361,2742,538,8708.95%4,011,62614.14%
B640,758,7543,324,6548.16%5,245,14012.87%
B763,812,5704,585,1547.19%7,254,62611.37%
Table 4. Ablation study done with the CIFAR-10 dataset for 50 epochs, comparing the effect of varying the number of channels per group. It also includes the improvement achieved by double training kEffNet-B0 V2 32ch with original images and horizontally flipped images.
Table 4. Ablation study done with the CIFAR-10 dataset for 50 epochs, comparing the effect of varying the number of channels per group. It also includes the improvement achieved by double training kEffNet-B0 V2 32ch with original images and horizontally flipped images.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,020,358100.0%389,969,098100.0%93.33%
kEffNet-B0 V2 2ch311,9947.8%44,523,28611.4%84.36%
kEffNet-B0 V2 4ch354,8188.8%49,487,88612.7%87.66%
kEffNet-B0 V2 8ch444,34611.1%60,313,52615.5%90.53%
kEffNet-B0 V2 16ch623,22615.5%82,804,37421.2%92.61%
kEffNet-B0 V2 32ch950,65023.6%123,209,11031.6%93.67%
kEffNet-B0 V2 32ch + H Flip950,65023.6%246,418,22063.3%94.01%
Table 5. Extra experiments made for kEffNet-B0 V2 4ch, 8ch, 16ch and 32ch variants. Rows labeled with “no L” indicate experiments done using only layer K, i.e., disabling layer L and the interleaving. Rows labeled with “ReLU” replace the swish activation function by ReLU.
Table 5. Extra experiments made for kEffNet-B0 V2 4ch, 8ch, 16ch and 32ch variants. Rows labeled with “no L” indicate experiments done using only layer K, i.e., disabling layer L and the interleaving. Rows labeled with “ReLU” replace the swish activation function by ReLU.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,020,358100.0%389,969,098100.0%93.33%
kEffNet-B0 V2 4ch354,8188.8%49,487,88612.7%87.66%
kEffNet-B0 V2 4ch no L342,0708.5%48,064,09812.3%82.44%
kEffNet-B0 V2 4ch ReLU354,8188.8%47,595,91412.2%85.34%
kEffNet-B0 V2 8ch444,34611.1%60,313,52615.5%90.53%
kEffNet-B0 V2 8ch no L422,88610.5%57,466,37014.7%89.27%
kEffNet-B0 V2 8ch ReLU444,34611.1%58,421,55415.0%88.82%
kEffNet-B0 V2 16ch623,22615.5%82,804,37421.2%92.61%
kEffNet-B0 V2 16ch no L584,93414.6%77,356,80219.8%91.52%
kEffNet-B0 V2 16ch ReLU623,22615.5%80,912,40620.8%91.16%
kEffNet-B0 V2 32ch950,65023.6%123,209,11031.6%93.67%
kEffNet-B0 V2 32ch no L879,75021.9%112,684,70628.9%93.21%
kEffNet-B0 V2 32ch ReLU950,65023.7%121,317,14231.1%92.00%
Table 6. Results obtained with the CIFAR-10 dataset after 180 epochs.
Table 6. Results obtained with the CIFAR-10 dataset after 180 epochs.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,020,358100.0%389,969,098100.0%94.86%
kEffNet-B0 V2 32ch950,65023.6%123,209,11031.6%94.45%
kEffNet-B0 V2 32ch + H Flip950,65023.6%246,418,22063.3%94.95%
Table 7. Results obtained with the Malaria dataset after 75 epochs.
Table 7. Results obtained with the Malaria dataset after 75 epochs.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,010,110100.0%389,958,834100.0%97.39%
kEffNet-B0 V2 2ch301,7467.5%61,196,07015.7%96.70%
kEffNet-B0 V2 4ch344,5708.6%69,691,35817.9%96.95%
kEffNet-B0 V2 8ch434,09810.8%87,725,25422.5%97.39%
kEffNet-B0 V2 12ch524,02613.1%106,199,56627.2%97.31%
kEffNet-B0 V2 16ch612,97815.3%124,672,93432.0%97.61%
kEffNet-B0 V2 32ch940,40223.5%164,422,95042.2%97.57%
Table 8. Results obtained with the colorectal cancer dataset after 1000 epochs.
Table 8. Results obtained with the colorectal cancer dataset after 1000 epochs.
ModelParameters%Computations%Test acc.
EfficientNet-B0 baseline4,017,796100.0%389,966,532100.0%97.61%
kEffNet-B0 V2 2ch355,0648.8%61,203,76815.7%97.62%
kEffNet-B0 V2 4ch397,8889.9%69,699,05617.9%98.02%
kEffNet-B0 V2 8ch487,41612.1%87,732,95222.5%98.02%
kEffNet-B0 V2 12ch531,71213.2%106,207,26427.2%97.22%
kEffNet-B0 V2 16ch620,66415.4%124,680,63232.0%97.62%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Schwarz Schuler, J.P.; Also, S.R.; Puig, D.; Rashwan, H.; Abdel-Nasser, M. An Enhanced Scheme for Reducing the Complexity of Pointwise Convolutions in CNNs for Image Classification Based on Interleaved Grouped Filters without Divisibility Constraints. Entropy 2022, 24, 1264. https://doi.org/10.3390/e24091264

AMA Style

Schwarz Schuler JP, Also SR, Puig D, Rashwan H, Abdel-Nasser M. An Enhanced Scheme for Reducing the Complexity of Pointwise Convolutions in CNNs for Image Classification Based on Interleaved Grouped Filters without Divisibility Constraints. Entropy. 2022; 24(9):1264. https://doi.org/10.3390/e24091264

Chicago/Turabian Style

Schwarz Schuler, Joao Paulo, Santiago Romani Also, Domenec Puig, Hatem Rashwan, and Mohamed Abdel-Nasser. 2022. "An Enhanced Scheme for Reducing the Complexity of Pointwise Convolutions in CNNs for Image Classification Based on Interleaved Grouped Filters without Divisibility Constraints" Entropy 24, no. 9: 1264. https://doi.org/10.3390/e24091264

APA Style

Schwarz Schuler, J. P., Also, S. R., Puig, D., Rashwan, H., & Abdel-Nasser, M. (2022). An Enhanced Scheme for Reducing the Complexity of Pointwise Convolutions in CNNs for Image Classification Based on Interleaved Grouped Filters without Divisibility Constraints. Entropy, 24(9), 1264. https://doi.org/10.3390/e24091264

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop