Next Article in Journal
Three-Dimensional Reconstruction of Light Field Based on Phase Similarity
Next Article in Special Issue
Multiclass Image Classification Using GANs and CNN Based on Holes Drilled in Laminated Chipboard
Previous Article in Journal
Over-Actuated Underwater Robots: Configuration Matrix Design and Perspectives
Previous Article in Special Issue
A Heterogeneous RISC-V Processor for Efficient DNN Application in Smart Sensing System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Convolutional Autoencoder Topology for Classification in High-Dimensional Noisy Image Datasets

by
Emmanuel Pintelas
1,*,
Ioannis E. Livieris
2 and
Panagiotis E. Pintelas
1
1
Department of Mathematics, University of Patras, 26500 Patras, Greece
2
Core Innovation and Technology O.E., 11745 Athens, Greece
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7731; https://doi.org/10.3390/s21227731
Submission received: 12 October 2021 / Revised: 14 November 2021 / Accepted: 18 November 2021 / Published: 20 November 2021
(This article belongs to the Collection Machine Learning and AI for Sensors)

Abstract

:
Deep convolutional neural networks have shown remarkable performance in the image classification domain. However, Deep Learning models are vulnerable to noise and redundant information encapsulated into the high-dimensional raw input images, leading to unstable and unreliable predictions. Autoencoders constitute an unsupervised dimensionality reduction technique, proven to filter out noise and redundant information and create robust and stable feature representations. In this work, in order to resolve the problem of DL models’ vulnerability, we propose a convolutional autoencoder topological model for compressing and filtering out noise and redundant information from initial high dimensionality input images and then feeding this compressed output into convolutional neural networks. Our results reveal the efficiency of the proposed approach, leading to a significant performance improvement compared to Deep Learning models trained with the initial raw images.

1. Introduction

Nowadays, convolutional neural networks (CNNs) have considerably flourished mainly because they have shown noticeable classification performance in image classification and computer vision tasks [1,2]. However, robustness and stability are some major problems in which Deep Learning (DL) models are prone, since it is proved that they can be fooled even by a tiny amount of perturbation, exhibiting poor and unreliable performance in these cases [3,4].
Moreover, in Machine Learning (ML) image classification tasks when dealing with high-dimensional data, which usually contain a lot of redundant information and noise, the reliable knowledge feature extraction procedure deteriorates [5]. The extraction of only the most important features compresses the initial feature space, leading to a stable and robust latent image representation [2,5,6]. Thus, it is necessary to capture only the most relevant information.
Training a supervised DL model with high dimensionality and low-quality image data can lead to overfitting and/or unstable behavior, especially when the training instances are limited or unbalanced. In other words, small pixel changes can lead the model to change its predictions, which implies that it has not exploited the information in the training data, and it exhibits poor and inefficient performance [7]. Additionally, it is worth highlighting that another significant problem is that the higher the dimension of the input images, the more the network is affected by the presence of noise, even if the amount of noise is small. By taking into consideration these difficulties and constraints, the application of a preproccessing step, which will attempt to reduce the noise in the image data while simultaneously reduce their dimension is considered essential for improving the performance of the DL model.
A traditional approach for image denoising is the transformation of the image from pixel intensities into another representation in order to capture the image’s statistical regularities more easily and effectively [8]. In the literature, a variety of approaches have been proposed for image denoising such as Gaussian scale mixture (GSM) models [9,10] and the more elegant Markov random fields (MRF)-based methodologies [11,12]. Nevertheless, the main disadvantage of these approaches is the considerable computational cost for parameter estimation and the fact that their values significantly affect the denoising performance quality. For example, naive methods of learning MRF-based models require the calculation of the partition function as well as a normalization factor which is generally intractable for high image dimensions [8]. To this end, a significant amount of research has been devoted to approximate MRF learning and inference techniques, which are usually computationally inefficient; thus, the parameters estimation is a significantly hard task. In addition, even if a traditional method is successfully applied, the noise in the images will be considerably reduced in the best-case scenario, but not totally removed. This implies that the high dimension of the input images, together with any amount of noise, will probably lead to unsatisfying performance, although the amount of noise may be small.
Convolutional autoencoder (CAE) [7] models constitute neural-network-based models, which have been proposed for dimensionality reduction and representation learning in a variety of tasks [13,14,15]. These models avoid the computational cost drawback of image denoising by posing the task within the statistical framework of regression, which constitutes a more tractable computation; thus, it permits greater representational power than density estimation [8]. More specifically, the process of image denoising with a CAE model can be formulated as a learning problem of training the model; therefore, the parameter estimation is performed by a modification of the well-known backpropagation algorithm [16]. The novelty in CAE for image processing is the utilization of convolutional layers which are able to create more abstract representations of the initial inputs by removing noise and redundant information. Therefore, these layers have been characterized as one of the major frontiers in deep learning and image analysis [7].
The utilization of convolutional layers allows CAEs to filter out noise and create robust and stable feature representations [5,17] while simultaneously reducing the input dimension size, making them suitable for dealing with high-dimensional noisy images. It is worth mentioning that an attractive property of CAEs over traditional dense autoencoders for image processing is that generally a huge loss of information is noticed when stacking and slicing the data. Instead of stacking the data as in classical autoencoders, the convolutional layers of CAEs are able to efficiently retain the spatial information of the input image data and gently extract information. In other words, CAEs can learn compressed image latent representations [5,17,18], therefore preserving the spatial locality of the input in a manner similar to other CNNs [19].
In this work, we propose an ML topology for performing efficient and robust classification in high-dimensional and noisy input data images. As a first preprocessing step, we utilize a CAE model in order to compress and filter out noise and redundant information, maintaining also the spatial feature characteristics from the initial image data and capturing only the most relevant and useful feature information. As a result, the initial high-dimensional input is transformed via the CAE into a compressed and compact 2D spatial image representation. In the second step, the CAE outputs are fed into a powerful pre-trained image classification model in order to successfully address classification tasks.
The main contribution of this research lies in proposing and implementing CAE as a general unsupervised learning data preprocessing method for creating robust and compressed feature representations. The rationale behind our approach is to improve state-of-the-art DL models to perform stable and accurate predictions on classification tasks when high-dimensional and noisy image input datasets are involved. The aim of this work is to tackle the problem of Deep Learning models being vulnerable in noise and redundant information tied into the raw input images, especially from high-dimensional inputs, leading to unstable and unreliable predictions. To this end, our approach is based on exploiting the advantages of CAEs as an efficient pre-processing noise filtering and compression method in order to improve CNN models in terms of robustness and accuracy when dealing with noisy high-dimensional inputs.
In order to validate the efficacy and the efficiency of the proposed method, we performed extensive experimental simulations utilizing high-dimensional image datasets concerning three different application domains. The first dataset concerns the problem of plant disease detection, the second concerns skin cancer detection, and the third DeepFake detection.
The rest of this paper is organized as follows: in Section 2, we present the state of the art image classification models and our proposed topology, while in Section 3, we present the technical details concerning the utilized datasets. Section 4 demonstrates our experimental results, and finally, Section 5 sketches our conclusive remarks and the future directions.

2. Materials and Methods

2.1. Related Work

During the last decades, a remarkable number of methodologies have been proposed in order to remove noise and redundant information from input images for creating robust and efficient image representations. Tian et al. [20] presented an excellent review regarding conventional machine learning methods and deep learning technologies for image denoising. The authors presented the most decent works proposed in the literature, focusing on the advantages of each approach. Finally, they discussed some promising research directions for image denoising based on deep learning technologies. In general, image denoising approaches can be separated into the traditional ones, such as GSM models and MRF-based methodologies, and the most elegant—the CAE-based approaches.
Portilla et al. [10] proposed a new methodology for noise removal from digital images which was based on a local GSM model in an overcomplete oriented pyramid representation. In their proposed methodology, they computed the full optimal local Bayesian least squares solution (LSS), as opposed to first approximating the local variance, and then utilized it to estimate the coefficient. Additionally, we utilized the vectorial form of the LLS solution in order to exploit all the information provided by the covariance modeling of signal and noise. The authors provided empirical evidence that these enhancements considerably improve the denoising performance.
Tappen et al. [11] presented a new approach for training a Gaussian Conditional Random Field (GCRF) model for image denoising, which is able to outperform the non-convex Field of Experts model. The rationale behind their approach of focusing on discrete-valued and non-convex MRF models was that GSM models tend to over-smooth images and blur edges. An advantage of the GCRF model is that its parameters can be optimized efficiently on relatively large images. Based on their experimental analysis, the authors stated that their proposed approach constitutes an attractive option for image and vision processing applications.
Barbu [21] proposed an interesting approach in which it was demonstrated that the process of training an MRF/CRF model together with a very fast inference algorithm could offer promising results relative to both speed and accuracy. The key idea of the proposed approach was that a validation set can be utilized to estimate the generalization performance of the trained system. Their experiment was performance on 256 × 256 images which presented that the proposed approach obtained an improved performance as well as a 1000–3000 times speedup compared to the state-of-the-art Field of Experts MRF trained with contrastive divergence.
Zhang et al. [22] proposed a novel Gaussian mixture Markov random field model (GM-MRF) which can be efficiently utilized as a very expressive prior model for image denoising and reconstruction. The proposed method forms a global image model by merging together individual GSM mixture models for image patches. Furthermore, the authors analytically presented a framework for computing MAP estimates with the GM-MRF model through the construction of exact surrogate functions. Their experimental analysis included a demonstration of the efficiency of their approach for denoising of dual-energy CT images.
Chen et al. [23] proposed a new framework which was based on a convolutional autoencoder model for creating unsupervised representations for images of lung nodule. More specifically, their approach was composed by a two-stage training procedure: at the first phase, the CAE was trained in an unsupervised way utilizing unlabeled data for image features learning; in the second phase the CAE is merged with a dense neural network and the resulting model was trained in a supervised way utilizing labeled data. An attractive property of the proposed approach comparing to a supervised one, is that it requires a small amount of labeled data for efficient feature learning applied in classification tasks. Additionally, the authors provided evidence that their proposed methodology can be extended for similarity measurement tasks of lung nodules images.
Seyfioğlu et al. [24] proposed a three-layer CAE topology for radar-based classification of similar aided and unaided human activities. After the unsupervised training procedure of the CAE, the decoder was removed, and it was substituted by dense layers and an output softmax layer. The develoved convolutional-based classification model was then trained in a supervised way. Their experimental analysis showed that their proposed methodology was superior compared to other deep learning classification models, support vector machines, extreme gradient boosting and random forest.
The main difference of our proposed approach compared to the previous state-of-the-art approaches lies in the fact that a CAE is trained for filtering out noise and creating robust and stable feature representations while simultaneously reducing the input image dimension size. In the sequence, the output of the encoder is utilized for developing a training set to fit a powerful pre-trained image classification model. The rationale behind our approach is to enhance the predictive power of pre-trained neural network classification model by developing a higher quality training dataset. For this purpose, we utilized a CAE for simultaneously compressing an image and filtering out noise and redundant information while also maintaining the spatial feature characteristics from the initial image data.

2.2. State-of-Art Pre-Trained CNN Classification Models

Large Deep Learning models trained on over millions of images, composed by a large variety of various CNN architectures, topologies such as VGG and ResNet, are considered as the mainstream approaches for addressing image classification applications [1,2]. In fact, these networks are utilized as pre-trained feature extraction models transferring their knowledge into new small non-trained networks (main transfer learning approach) in order to specialize in new specific image classification problems.
VGG [19] adopted its name by the team Visual Geometry Group at the University of Oxford and is applied in computer vision tasks. This deep neural network is constituted by multiple ( 3 × 3 ) convolution filters, which are proved to be more efficient compared to its prior network AlexNet. On the other hand, AlexNet was composed by kernel-sized filters (11 and 5 in the first and second layer, respectively).
ResNet [25], also called Residual Network, is a Deep Learning model which utilizes identity connections in order to address the degradation problem which is caused by very large network depths (such as over 18 layers). In particular, these connections take the input directly to the end of each residual block, while each residual block is constituted by 3 × 3 and 1 × 1 convolution filters.
DenseNet [26] constitutes an updated version of ResNet and is implemented using dense blocks, which connect each layer to every other layer in a feedforward way. The main advantages of utilizing these blocks are feature reuse, implicit deep supervision, and parameter efficiency.
MobileNet [27] is a computationally efficient state of the art CNN topology designed for application mainly to mobiles. MobileNet’s topology is based on an inverted residual structure, while the input and output of the residual block are thin bottleneck layers in contrast to classic residual networks.

2.3. Convolutional Autoencoders

Convolutional autoencoders (CAEs) are unsupervised dimensionality reduction models composed by convolutional layers capable of creating compressed image representations [28]. In general, CAEs are mainly utilized for reducing and compressing the input dimension size, removing noise while simultaneously keeping all useful information and extracting robust features [5,17,18].
The main difference between convolutional AE and traditional AE is the utilization of convolutional layers. It is worth mentioning that these layers are characterized by their attractive property of extracting knowledge and learning the internal representation of image data.
More specifically, CAEs are composed by two CNN models, the Encoder and Decoder, as presented in Figure 1. The Encoder is mainly used for encoding the initial input image into a latent representation which has lower dimension. On the other hand, the Decoder is responsible for reconstructing the compressed latent representation creating an output image being as much similar with the initial one.
Mathematically, let x = X i n i t i a l M × H × W × 3 denote the initial input image dataset, where H, W are the number of pixels corresponding to the Height and Width of every image X, and M is the number of samples. Additionally, E and D denote the Encoder and Decoder, respectively. Then, the encoded representation y = X c o m p r e s s e d M × h × w × 3 , where h and w are the width and depth dimensions of the compressed spatial 2D representation, and the decoded reconstruction x ^ = X i n i t i a l M × H × W × 3 consist of the output of the encoder and decoder, respectively, that is:
y = E ( x ) x ^ = D ( y )
The performance of the convolutional autoencoder can be measured by the reconstruction error e C A E , which is defined by:
e C A E = L C A E ( x ^ ( k ) ) , x ( k )
The function L C A E denotes a measurement of difference such as the widely used square Euclidean distance defined as:
L C A E x ^ ( k ) , x ( k ) = 1 2 x ^ ( k ) x ( k ) 2 .
Then, the cost function in its general form can be formulated as follows:
J C A E = 1 M k = 1 M L C A E D ( E ( x ( k ) ) ) , x ( k )
By minimizing the cost function J C A E , we attempt to find the optimal weight parameters for the convolutional autoencoder.
In our implementation, the cost function (1) was minimized utilizing Adam optimization algorithm [29], and the value of the learning rate was set to 10 3 .

2.4. Proposed Topology

Figure 2 presents the main pipeline of the proposed convolutional autoencoder–convolutional neural network (CAE-CNN) topology. In our approach, initially a CAE is trained with the initial training dataset. When the CAE finishes its training procedure, then the decoder component is discarded, while the encoder is used for compressing the initial high-dimensional image dataset into a compressed image dataset. Finally, the output of the CAE’s encoder (compressed image dataset) is used for feeding and training a CNN classification model, such as ResNet, VGG, etc.
In the sequel, let us denote C as the CNN classification model and l = { l 1 , l 2 , , l N } , where l i { 0 , 1 } i N , as the target output of N total classes with respect to the classification problem. The initial training dataset x is transformed via the encoder E into an encoded compressed 2D representation y. The raw output l ^ = { l ^ 1 , l ^ 2 , , l ^ N } of the CNN classification model is given as follows:
l ^ = C ( y )
The performance of the CNN classification model can be measured by the reconstruction error:
e C N N = L C N N l ^ ( k ) , l ( k )
The function L C N N denotes a measurement of difference such as the widely used Cross Entropy loss function [30]. Then, the cost function in its general form can be formulated as follows:
J C N N = 1 M k = 1 M L C A E D ( E ( l ( k ) ) ) , l ( ( k ) )
Finally, by minimizing the cost function J C N N , we obtain the optimal weight parameters for the CNN classification model with respect to the classification task.

2.5. Proposed Convolutional Autoencoder’s Architectural Design

Figure 3 presents the proposed CAE topology of our architecture, while Figure 4 presents the compressed image output of the CAE of some examples regarding the three utilized image datasets. Finally, Table 1 presents in a detailed way the CAE parameters’ settings configurations setup. The proposed CAE has a symmetric architecture with four batches of 2D convolutional and deconvolutional layers followed by a Rectified Linear Unit (ReLU) activation function. The deconvolution [31] (also called as transposed convolution) is performing the reverse operation of the convolutional layer. In particular, it maps the input from a low-dimensional space to a high-dimensional one.
More specifically, the raw input image with dimensions H × W × 3 is fed into the first layer (2D Conv1-ReLU1), which is also the CAE’s Encoder’s input and creates 32 down-sampled spatial feature maps of dimensions H / 2 × W / 2 , utilizing a 4 × 4 kernel size with 32 filters. Subsequently, this output is fed into the second layer (2D Conv2-ReLU2), which is the Encoder’s output and creates the compressed image representation of dimensions H / 4 × W / 4 × 3 , utilizing a 2 × 2 kernel size with 3 filters. Since the first layer’s output feature maps have a lower dimensional size comparing to the input image, it is reasonable to utilize a smaller kernel size in the second layer. Similarly, the third and fourth layers (2D Deconv3-ReLU3, 2D Deconv4-ReLU4) of CAE’s decoder component perform in a symmetric way the reverse operation of this of the encoder’s.

3. Case Study Applications/Datasets

Next, we present the characteristics of the datasets, which were utilized in this study in order to evaluate the efficiency of the proposed architectural topology. The first application concerns the problem of plant disease detection, the second concerns the skin cancer detection, and the third the DeepFake detection.
These three application domains have attained very high interest in the last years [32,33,34,35,36,37,38,39] for many different reasons. More specifically, regarding the plant disease problem, due to the massive agriculture improvement, it is necessary to automate plant disease detection by using technologies such as air drones and cure ill plants fast and accurately [40]. Regarding skin cancer detection, the fast recognition and treatment in its earliest stages is crucial for its treatment and curing process. The DeepFake faces image detection problem recently has attained very high interest [33,34,35,36,37,38,39]. The recent invention and the continuous development of the Generative Adversarial Networks (GANs) [40] technology has made possible the generation and the creation of high quality and extremely realistic fake images and videos being very hard even for experts to recognize them. These fake images/videos can be extremely harmful for human rights, especially when the deepfakes are used maliciously as a source of misinformation, manipulation, and harassment.
Plant disease. Concerning the plant disease detection problem, the utilized training dataset is composed of 464, 81, 559, and 532 of “healthy”, “multiple-disease”, “rust”, and “scab” labeled plants, respectively, while the testing dataset consists of 52, 10, 63, and 60 of “healthy”, “multiple-disease”, “rust”, and “scab” labeled plants, respectively. All images had initial resolution of 1024×1024. This dataset is available at: https://drive.google.com/drive/folders/12DIF3KF-ZOnpBzsA6inRRjNNOBLqJhhG (accessed on 14 November 2021).
Skin cancer. This dataset was obtained by ISIC (International Skin Imaging Collaboration, https://www.isic-archive.com (accessed on 14 November 2021)), which aims to assist in reducing melanoma mortality through the application of digital skin cancer imaging. The utilized training dataset consists of 1754 and 178, while the testing dataset consists of 168 and 22 of “Benign” and “Malignant” diagnosed patients, respectively. All images had initial resolution of 1024 × 1024 .
DeepFake experts. Concerning the DeepFake detection problem, we utilized a balanced dataset created by expert photoshop designers, conducted by the Computational Intelligence and Photography Lab in the Department of Computer Science at Yonsei University. In particular, they forged and replaced persons’ facial marks such as eyes and mouth into other different person faces. Thus, such images are manipulated and considered as fake instances. The utilized dataset includes images, which vary from easy, mid, and hard recognition difficulty. The training dataset constitutes 924 and 973, while the testing dataset constitutes 121 and 108 “Fake” and “Real” labeled face images, respectively. All images had initial resolution of 600 × 600 . This dataset is available at: https://www.kaggle.com/ciplab/real-and-fake-face-detection (accessed on 14 November 2021).

4. Experimental Results

In this section, we validate the efficiency and robustness of the proposed topology by performing comprehensive experimental simulations utilizing various state of the art CNN topologies. The measurement of quality is based on the well-known widely used evaluation metrics: Accuracy (Acc), Geometric Mean (GM), and the Area Under the Curve (AUC) [41]. Notice that the performance metrics GM and AUC present the information provided by a confusion matrix in compact form [42,43]; hence, these two metrics constitute the proper ones to evaluate if a prediction model has not overfitted the training data. The best performance for each state-of-the-art model and performance metric is highlighted in bold. The implementation code was written in Python 3.7 utilizing FastAI library [44], while the hyper-parameters were defined under exhaustive experimentation.
In our implementation, the cost function (1) was minimized utilizing Adam optimization algorithm [29], and the value of the learning rate was set to 10 3 . Additionally, in order to avoid overfitting and maximize the efficiency of the proposed CAE, 10% of training data was used for validation and early stopping technique based on “validation loss’’ was used.
The evaluated approaches are the “Traditional”, “Means Denoising (MD)”, and the proposed one, “CAE”. In the Traditional approach, the images are directly fitted to a CNN classification model, while in the MD approach, a non-local MD algorithm [45] was applied on every input image, as a first preprocessing step, before feeding the images into the CNN classification models. The motivation behind this approach was to provide a more comprehensive experimental comparison by evaluating the proposed methodology against a classical approach which uses an image processing denoising technique. Finally, in our proposed CAE approach, the images are initially compressed and transformed via the CAE model and then fitted to a CNN classification model.
Table 2, Table 3 and Table 4 summarize the performance of the evaluated approaches regarding plant disease, skin cancer, and DeepFake experts, respectively. Notice that each approach has been evaluated utilizing VGG, ResNet, DenseNet, and MobileNet as a pre-trained CNN classification model. Notice that all pre-trained models were trained on ImageNet dataset. The interpretation of Table 2, Table 3 and Table 4 reveals that the incorporation of CAE managed to significantly increase the performance of all CNN models in all case study scenarios, especially for the VGG model in which is observed a considerable improvement. In addition, it managed to outperform the MD image pre-processing technique for every utilized pre-trained CNN. More specifically, the proposed CAE approach managed to significantly increase the performance of the ResNet, DenseNet, and MobileNet CNN model producing the best results for all datasets. Moreover, it managed to considerably increase the performance of the VGG model for all utilized datasets. Furthermore, the DenseNet CNN model managed to deliver the best results overall for all utilized approaches (Traditional, MD, CAE) and all utilized datasets, while the VGG model reported the worst results. However, utilizing the proposed CAE approach, the VGG model managed to achieve a decent and great performance, similar to the other CNN models.
Finally, it is worth mentioning that in our preliminary experiments, we fitted the CNN models with the original images, and we utilized the reconstructed images on the testing phase. However, the performance of all models was similar or slightly degraded.
Another significant finding is that the incorporation of CAE considerably improved the performance of the CNN models in the DeepFake Experts dataset. This is probably due to the fact that this dataset is considered as a very noisy dataset with considerable amount of redundant information in every image. This means that the CAE managed to capture the relevant information (fake signs on every face for this case study scenario), filtering out the noise. Therefore, the CNN classification models had to focus only on the most relevant information, thus leading to this great performance improvement.
Clearly, the purpose of this study was not to propose a deep learning classifier but to demonstrate a complete topology for addressing hard image classification problems in which the training image data has a high-dimensional size and contains noise. As a result, it obviously leads to some computational cost increase for training the CAE, but on the other hand, reducing the noise leads to a considerable reduction in computational cost for training the DL model (smaller figures).
To summarize, our results demonstrated that the incorporation of convolutional autoencoders as an image preprocessing technique could improve the performance of CNN models leading to robust and accurate results. Therefore, it can be considered as a promising tool on high-dimensional and noisy dataset applications.
Next, we attempt to provide statistical evidences about the efficiency of our proposed approach. More analytically, we investigate if the hypothesis H 0 that all evaluated approaches, i.e., “Traditional”, “Means Denoising (MD)”, and the proposed one, “CAE”, performed equally well for a given level. For this purpose, we used the non-parametric Friedman Aligned Ranking (FAR) [46] test. Furthermore, for examining if the differences in the performance of the utilized pre-trained CNN models are statistically significant, we applied the post hoc Finner test [47] with significance level α = 5 % .
Table 5, Table 6 and Table 7 report the statistical analysis, performed by nonparametric multiple comparison, relative to Accuracy, GM, and AUC performance metrics, respectively. Clearly, the results presented in Table 5, Table 6 and Table 7 provide statistical evidence that the proposed approach reported the highest probability-based ranking, outperforming the other approaches.

5. Conclusions

In this work, we proposed and suggested the incorporation of convolutional autoencoders as a general unsupervised learning data preprocessing method for creating robust and compressed feature representations in order to improve CNN performance on image classification tasks.
The utilized CNN models (ResNet, DenseNet, MobileNet, and VGG) are generally considered as widely used state-of-the-art (SoA) image classification models. Furthermore, our scope in this research work is not to prove directly that our method leads to the highest performance results comparing with other SoA image classification models. In contrast, our aim is to prove that the proposed method is able to improve any SoA CNN model via the combination and incorporation of the CAE component. Therefore, we consider that the utilization of four SoA image classification models is a sufficient number in model selection in order to lead to reliable and robust experimental results.
In order to validate the efficiency of the proposed methodology, we utilized datasets from three very popular and totally different application domains, the plant disease, skin cancer, and DeepFake detection problems, applying state-of-the-art CNN model architectures such as ResNet and DenseNet. Based on our experimental results, the proposed methodology was significantly superior compared to every other utilized approach. Note that it is possible that the prediction ability of the proposed approach could be further improved by including more sophisticated DL tools, loss and activation functions, class weighted approaches, etc. (see [48] and the references therein). This is to be included and fully investigated in our future research.
Nevertheless, the limitation of the proposed framework lies in the fact that there is no mathematical proof that it outperforms traditional methodologies nor that the conditions of the Nyquist theorem are satisfied. The conclusions and findings of this work can be demonstrated only from experimental and qualitative reasoning. Furthermore, the utilized real-world datasets did not provide us with any information about the distribution and the kind of noise. To this end, we were not able to estimate the expense of error or the image quality improvement. These also constitute general limitations to all prior works [8,15,23,24], which proposed denoising schemes based on the use of CAEs.
Additionally, CAEs are usually prone to overfitting which implies that in such cases, they may not create high-quality data for fitting a Deep Learning model. From our numerical experiments, we have not noticed such cases; however, more experiments are needed. This is probably dependent on the problem at hand and the quality of the original images.
Another limitation of the proposed work is that the level of compression is dependent on the CAE’s architecture and more specifically on the output of the encoder. Note that the smaller the output of the encoder, the higher the level compression. However, for identifying the optimal level of compression, we believe that more experimentation is needed in order to examine and evaluate the performance of the proposed approach using different architectures and eventually different levels of compression.
In future work, we aim to investigate ensemble learning methods [49,50,51,52] such as Bagging and Stacking and combine them with our proposed topology, aiming to create more robust and accurate composite classification models. In addition, we intend to explore further the efficacy and efficiency of the proposed network in a variety of datasets of various sizes and complexity from real-world application domains. The vigorous development of the Internet and the widespread adoption of electronic medical records have led to the development of large repositories of labeled and mostly of unlabeled biomedical images. Nevertheless, the process of correctly labeling new, unlabeled instances frequently requires the efforts of specialized personnel, which incur a lot of time and high monetary costs. To deal with this problem, Semi-Supervised Learning algorithms constitute the appropriate machine learning methodology for extracting useful knowledge from both labeled and unlabeled data in order to build efficient prediction models. To this end, another interesting idea could be the adoption of the proposed autoencoder methodology to Semi-Supervised Learning (SSL) techniques [53] for addressing challenging biomedical classification tasks. SSL is a new state-of-the-art data mining area, which focuses on applications where the labeled data are limited and require much effort and cost to obtain. Since the autoencoders do not require by default labeled data for their training procedure, the incorporation of the proposed framework in the SSL area could provide us with promising models. Finally, we also aim to incorporate intrinsic interpretability using explainable features [54,55,56] in order to provide some degree of explainability.

Author Contributions

I.E.L., E.P., and P.E.P. conceived of the idea, designed and performed the experiments, analyzed the results, drafted the initial manuscript, and revised the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

List of Acronyms and Abbreviations

AccAccuracy
AUCArea under the curve
CAEConvolutional autoencoders
CNNConvolutional neural networks
CAE-CNNConvolutional autoencoder–convolutional neural network
DLDeep learning
GCRFGaussian conditional random field
GMGeometric mean
GM-MRFGaussian mixture Markov random field model
GSMGaussian scale mixture
ISICInternational skin imaging collaboration
LSSLeast squares solution
MDMeans denoising
MLMachine learning
MRFMarkov random fields

References

  1. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Shao, L.; Zhu, F.; Li, X. Transfer learning for visual categorization: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2014, 26, 1019–1034. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, C.; Zhang, K.; Li, Y. A causal view on robustness of neural networks. arXiv 2020, arXiv:2005.01095. [Google Scholar]
  4. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  5. Wickramasinghe, C.S.; Marino, D.L.; Manic, M. ResNet Autoencoders for Unsupervised Feature Learning From High-Dimensional Data: Deep Models Resistant to Performance Degradation. IEEE Access 2021, 9, 40511–40520. [Google Scholar] [CrossRef]
  6. Zabalza, J.; Ren, J.; Zheng, J.; Zhao, H.; Qing, C.; Yang, Z.; Du, P.; Marshall, S. Novel segmented stacked autoencoder for effective dimensionality reduction and feature extraction in hyperspectral imaging. Neurocomputing 2016, 185, 1–10. [Google Scholar] [CrossRef] [Green Version]
  7. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for image denoising using convolutional neural network: A review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  8. Jain, V.; Seung, S. Natural image denoising with convolutional networks. In Proceedings of the 22nd Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–10 December 2008. [Google Scholar]
  9. Lyu, S.; Simoncelli, E.P. Statistical modeling of images with fields of Gaussian scale mixtures. In Proceedings of the 20th Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006. [Google Scholar]
  10. Portilla, J.; Strela, V.; Wainwright, M.J.; Simoncelli, E.P. Image denoising using scale mixtures of Gaussians in the wavelet domain. IEEE Trans. Image Process. 2003, 12, 1338–1351. [Google Scholar] [CrossRef] [Green Version]
  11. Tappen, M.F.; Liu, C.; Adelson, E.H.; Freeman, W.T. Learning gaussian conditional random fields for low-level vision. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  12. Weiss, Y.; Freeman, W.T. What makes a good model of natural images? In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  13. Arai, H.; Chayama, Y.; Iyatomi, H.; Oishi, K. Significant dimension reduction of 3D brain MRI using 3D convolutional autoencoders. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5162–5165. [Google Scholar]
  14. Guo, X.; Liu, X.; Zhu, E.; Yin, J. Deep clustering with convolutional autoencoders. In Proceedings of the International Conference on Neural Information Processing, Guangzhou, China, 14–18 November 2017; pp. 373–382. [Google Scholar]
  15. Kucharski, D.; Kleczek, P.; Jaworek-Korjakowska, J.; Dyduch, G.; Gorgon, M. Semi-supervised nests of melanocytes segmentation method using convolutional autoencoders. Sensors 2020, 20, 1546. [Google Scholar] [CrossRef] [Green Version]
  16. Da Silva, I.N.; Spatti, D.H.; Flauzino, R.A.; Liboni, L.H.B.; dos Reis Alves, S.F. Artificial neural network architectures and training processes. In Artificial Neural Networks; Springer: Cham, Switzerland, 2017; pp. 21–28. [Google Scholar]
  17. Sun, Y.; Mao, H.; Guo, Q.; Yi, Z. Learning a good representation with unsymmetrical auto-encoder. Neural Comput. Appl. 2016, 27, 1361–1367. [Google Scholar] [CrossRef]
  18. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial–spectral feature learning by 3D convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  19. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  20. Tian, C.; Xu, Y.; Fei, L.; Yan, K. Deep learning for image denoising: A survey. In Proceedings of the International Conference on Genetic and Evolutionary Computing, Changzhou, China, 14–17 December 2018; pp. 563–572. [Google Scholar]
  21. Barbu, A. Learning real-time MRF inference for image denoising. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1574–1581. [Google Scholar]
  22. Zhang, R.; Bouman, C.A.; Thibault, J.B.; Sauer, K.D. Gaussian mixture Markov random field for image denoising and reconstruction. In Proceedings of the 2013 IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; pp. 1089–1092. [Google Scholar]
  23. Chen, M.; Shi, X.; Zhang, Y.; Wu, D.; Guizani, M. Deep features learning for medical image analysis with convolutional autoencoder neural network. IEEE Trans. Big Data 2017. [Google Scholar] [CrossRef]
  24. Seyfioğlu, M.S.; Özbayoğlu, A.M.; Gürbüz, S.Z. Deep convolutional autoencoder for radar-based classification of similar aided and unaided human activities. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 1709–1723. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  27. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  28. Pintelas, E.; Livieris, I.E.; Barotsis, N.; Panayiotakis, G.; Pintelas, P. An Autoencoder Convolutional Neural Network Framework for Sarcopenia Detection Based on Multi-frame Ultrasound Image Slices. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, Crete, Greece, 25–27 June 2021; pp. 209–219. [Google Scholar]
  29. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  30. Martinez, M.; Stiefelhagen, R. Taming the cross entropy loss. In Proceedings of the German Conference on Pattern Recognition, Stuttgart, Germany, 9–12 October 2018; pp. 628–637. [Google Scholar]
  31. Xu, L.; Ren, J.S.; Liu, C.; Jia, J. Deep convolutional neural network for image deconvolution. Adv. Neural Inf. Process. Syst. 2014, 27, 1790–1798. [Google Scholar]
  32. Singh, A.; Ganapathysubramanian, B.; Singh, A.K.; Sarkar, S. Machine learning for high-throughput stress phenotyping in plants. Trends Plant Sci. 2016, 21, 110–124. [Google Scholar] [CrossRef] [Green Version]
  33. Sadeghi, M.; Lee, T.K.; McLean, D.; Lui, H.; Atkins, M.S. Detection and analysis of irregular streaks in dermoscopic images of skin lesions. IEEE Trans. Med. Imaging 2013, 32, 849–861. [Google Scholar] [CrossRef]
  34. Li, L.; Bao, J.; Zhang, T.; Yang, H.; Chen, D.; Wen, F.; Guo, B. Face X-ray for more general face forgery detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5001–5010. [Google Scholar]
  35. Dolhansky, B.; Bitton, J.; Pflaum, B.; Lu, J.; Howes, R.; Wang, M.; Canton Ferrer, C. The deepfake detection challenge dataset. arXiv 2020, arXiv:2006.07397. [Google Scholar]
  36. Rossler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Nießner, M. Faceforensics++: Learning to detect manipulated facial images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 1–11. [Google Scholar]
  37. Agarwal, S.; Farid, H.; Gu, Y.; He, M.; Nagano, K.; Li, H. Protecting World Leaders Against Deep Fakes. In Proceedings of the CVPR Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  38. Durall, R.; Keuper, M.; Pfreundt, F.J.; Keuper, J. Unmasking deepfakes with simple features. arXiv 2019, arXiv:1911.00686. [Google Scholar]
  39. Qian, Y.; Yin, G.; Sheng, L.; Chen, Z.; Shao, J. Thinking in frequency: Face forgery detection by mining frequency-aware clues. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 86–103. [Google Scholar]
  40. Pan, Z.; Yu, W.; Yi, X.; Khan, A.; Yuan, F.; Zheng, Y. Recent progress on generative adversarial networks (GANs): A survey. IEEE Access 2019, 7, 36322–36333. [Google Scholar] [CrossRef]
  41. Powers, D. Evaluation: From Precision, Recall and F-Measure to ROC, Informedness, Markedness & Correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
  42. Livieris, I.E.; Pintelas, E.; Kiriakidou, N.; Stavroyiannis, S. An advanced deep learning model for short-term forecasting US natural gas price and movement. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, Neos Marmaras, Greece, 5–7 June 2020; pp. 165–176. [Google Scholar]
  43. Livieris, I.E.; Kiriakidou, N.; Stavroyiannis, S.; Pintelas, P. An advanced CNN-LSTM model for cryptocurrency forecasting. Electronics 2021, 10, 287. [Google Scholar] [CrossRef]
  44. Howard, J.; Gugger, S. Deep Learning for Coders with Fastai and PyTorch; O’Reilly Media: Newton, MA, USA, 2020. [Google Scholar]
  45. Buades, A.; Coll, B.; Morel, J.M. Non-local means denoising, image processing on line. Image Process. On Line 2011, 1, 208–212. [Google Scholar] [CrossRef] [Green Version]
  46. Hodges, J.L.; Lehmann, E.L. Rank methods for combination of independent experiments in analysis of variance. Ann. Math. Stat. 1962, 33, 482–497. [Google Scholar] [CrossRef]
  47. Finner, H. On a monotonicity problem in step-down multiple test procedures. J. Am. Stat. Assoc. 1993, 88, 920–923. [Google Scholar] [CrossRef]
  48. Lomacenkova, A.; Arandjelovic, O. Whole Slide Pathology Image Patch Based Deep Classification: An Investigation of the Effects of the Latent Autoencoder Representation and the Loss Function Form. In Proceedings of the 2021 IEEE EMBS International Conference on Biomedical and Health Informatics (BHI), Athens, Greece, 27–30 July 2021. [Google Scholar] [CrossRef]
  49. Livieris, I.E.; Iliadis, L.; Pintelas, P. On ensemble techniques of weight-constrained neural networks. Evol. Syst. 2020, 12, 155–167. [Google Scholar] [CrossRef]
  50. Pintelas, P.; Livieris, I.E. Special issue on ensemble learning and applications. Algorithms 2020, 13, 140. [Google Scholar] [CrossRef]
  51. Zhang, C.; Ma, Y. Ensemble Machine Learning: Methods and Applications; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  52. Polikar, R. Ensemble learning. In Ensemble Machine Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–34. [Google Scholar]
  53. Van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef] [Green Version]
  54. Pintelas, E.; Liaskos, M.; Livieris, I.E.; Kotsiantis, S.; Pintelas, P. Explainable machine learning framework for image classification problems: Case study on glioma cancer prediction. J. Imaging 2020, 6, 37. [Google Scholar] [CrossRef]
  55. Pintelas, E.; Liaskos, M.; Livieris, I.E.; Kotsiantis, S.; Pintelas, P. A novel explainable image classification framework: Case study on skin cancer and plant disease prediction. Neural Comput. Appl. 2021, 33, 15171–15189. [Google Scholar] [CrossRef]
  56. Pintelas, E.; Livieris, I.E.; Pintelas, P. A grey-box ensemble model exploiting black-box accuracy and white-box intrinsic interpretability. Algorithms 2020, 13, 17. [Google Scholar] [CrossRef] [Green Version]
Figure 1. General pipeline of a CAE model.
Figure 1. General pipeline of a CAE model.
Sensors 21 07731 g001
Figure 2. Main pipeline of the proposed CAE-CNN topology.
Figure 2. Main pipeline of the proposed CAE-CNN topology.
Sensors 21 07731 g002
Figure 3. Architectural presentation of our CAE topology.
Figure 3. Architectural presentation of our CAE topology.
Sensors 21 07731 g003
Figure 4. (A) Initial high-dimensional image. (B) Compressed image using CAE.
Figure 4. (A) Initial high-dimensional image. (B) Compressed image using CAE.
Sensors 21 07731 g004
Table 1. Parameter settings of the utilized CAE topology.
Table 1. Parameter settings of the utilized CAE topology.
LayersInput SizeKernel SizeStrideOutput Size
Encoder2D Conv
(E. Input)
H × W × 3 4 × 4 × 32 2 × 2 × 1 H / 2 × W / 2 × 32
ReLU H / 2 × W / 2 × 32 H / 2 × W / 2 × 32
2D Conv H / 2 × W / 2 × 32 2 × 2 × 3 2 × 2 × 1 H / 4 × W / 4 × 3
Sigmoid
(E. Output )
H / 4 × W / 4 × 3 H / 4 × W / 4 × 3
Decoder2D Deconv3
(D. Input)
H / 4 × W / 4 × 3 2 × 2 × 32 2 × 2 × 1 H / 2 × W / 2 × 32
ReLU H / 2 × W / 2 × 32 H / 2 × W / 2 × 32
2D Deconv4 H / 2 × W / 2 × 32 4 × 4 × 3 2 × 2 × 1 H × W × 3
Sigmoid
(D. Output)
H × W × 3 H × W × 3
Table 2. Performance results for plant disease benchmark. The best results are displayed in bold.
Table 2. Performance results for plant disease benchmark. The best results are displayed in bold.
ApproachesCNN ModelAccGMAUC
Traditional 80.0%8520.866
MDVGG83.2%9060.896
CAE 87.6%9130.917
Traditional 90.3%10270.935
MDResNet91.4%10440.946
CAE 92.4%10660.949
Traditional 93.0%11870.953
MDDenseNet92.9%11880.954
CAE 93.5%11400.956
Traditional 94.6%10310.963
MDMobileNet92.9%10750.957
CAE 93.5%10870.956
Table 3. Performance results for skin cancer benchmark. The best results are displayed in bold.
Table 3. Performance results for skin cancer benchmark. The best results are displayed in bold.
ApproachesCNN ModelAccGMAUC
Traditional 66.3%380.631
MDVGG67.4%390.637
CAE 70.0%400.750
Traditional 73.7%370.633
MDResNet74.7%390.642
CAE 77.9%420.697
Traditional 75.8%420.685
MDDenseNet75.3%410.681
CAE 80.0%410.689
Traditional 75.2%350.603
MDMobileNet74.2%360.612
CAE 72.0%390.638
Table 4. Performance results for DeepFake experts benchmark. The best results are displayed in bold.
Table 4. Performance results for DeepFake experts benchmark. The best results are displayed in bold.
ApproachesCNN ModelAccGMAUC
Traditional 73.3%830.732
MDVGG76.4%870.774
CAE 82.0%930.869
Traditional 80.8%920.881
MDResNet81.2%930.892
CAE 82.1%940.897
Traditional 84.3%960.905
MDDenseNet83.4%950.903
CAE 86.1%980.926
Traditional 79.0%900.853
MDMobileNet79.5%910.857
CAE 81.0%920.889
Table 5. FAR test and Finner post hoc test based on Accuracy metric.
Table 5. FAR test and Finner post hoc test based on Accuracy metric.
SeriesFriedmanFinner Post Hoc Test
Rankingp-Value H 0
CAE9.4167
MD21.66670.004399Rejected
Traditional24.41670.000975Rejected
Table 6. FAR test and Finner post hoc test based on GM metric.
Table 6. FAR test and Finner post hoc test based on GM metric.
SeriesFriedmanFinner Post Hoc Test
Rankingp-Value H 0
CAE11.5000
MD17.83330.140894Not rejected
Traditional26.16670.00065Rejected
Table 7. FAR test and Finner post hoc test based on AUC metric.
Table 7. FAR test and Finner post hoc test based on AUC metric.
SeriesFriedmanFinner Post Hoc Test
Rankingp-Value H 0
CAE7.58330
MD22.04170.000775Rejected
Traditional25.87500.000021Rejected
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pintelas, E.; Livieris, I.E.; Pintelas, P.E. A Convolutional Autoencoder Topology for Classification in High-Dimensional Noisy Image Datasets. Sensors 2021, 21, 7731. https://doi.org/10.3390/s21227731

AMA Style

Pintelas E, Livieris IE, Pintelas PE. A Convolutional Autoencoder Topology for Classification in High-Dimensional Noisy Image Datasets. Sensors. 2021; 21(22):7731. https://doi.org/10.3390/s21227731

Chicago/Turabian Style

Pintelas, Emmanuel, Ioannis E. Livieris, and Panagiotis E. Pintelas. 2021. "A Convolutional Autoencoder Topology for Classification in High-Dimensional Noisy Image Datasets" Sensors 21, no. 22: 7731. https://doi.org/10.3390/s21227731

APA Style

Pintelas, E., Livieris, I. E., & Pintelas, P. E. (2021). A Convolutional Autoencoder Topology for Classification in High-Dimensional Noisy Image Datasets. Sensors, 21(22), 7731. https://doi.org/10.3390/s21227731

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop