Next Article in Journal
Distributed Detection of Malicious Android Apps While Preserving Privacy Using Federated Learning
Previous Article in Journal
A Reinforcement Learning Handover Parameter Adaptation Method Based on LSTM-Aided Digital Twin for UDN
Previous Article in Special Issue
A Deep Residual Neural Network for Image Reconstruction in Biomedical 3D Magnetic Induction Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of HEp-2 Staining Pattern Images Using Adapted Multilayer Perceptron Neural Network-Based Intra-Class Variation of Cell Shape

by
Khamael Al-Dulaimi
1,†,
Jasmine Banks
1,†,
Aiman Al-Sabaawi
2,*,†,
Kien Nguyen
1,†,
Vinod Chandran
1,† and
Inmaculada Tomeo-Reyes
3,†
1
School of Electrical Engineering and Robotics, Queensland University of Technology (QUT), Brisbane, QLD 4000, Australia
2
School of Computer Science, Queensland University of Technology (QUT), Brisbane, QLD 4000, Australia
3
School of Electrical Engineering and Telecommunications, University of New South Wales, Sydney, NSW 2052, Australia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(4), 2195; https://doi.org/10.3390/s23042195
Submission received: 13 January 2023 / Revised: 1 February 2023 / Accepted: 8 February 2023 / Published: 15 February 2023
(This article belongs to the Special Issue Machine Learning for Biomedical Imaging and Sensing II)

Abstract

:
There exists a growing interest from the clinical practice research communities in the development of methods to automate HEp-2 stained cells classification procedure from histopathological images. Challenges faced by these methods include variations in cell densities and cell patterns, overfitting of features, large-scale data volume and stained cells. In this paper, a multi-class multilayer perceptron technique is adapted by adding a new hidden layer to calculate the variation in the mean, scale, kurtosis and skewness of higher order spectra features of the cell shape information. The adapted technique is then jointly trained and the probability of classification calculated using a Softmax activation function. This method is proposed to address overfitting, stained and large-scale data volume problems, and classify HEp-2 staining cells into six classes. An extensive experimental analysis is studied to verify the results of the proposed method. The technique has been trained and tested on the dataset from ICPR-2014 and ICPR-2016 competitions using the Task-1. The experimental results have shown that the proposed model achieved higher accuracy of 90.3% (with data augmentation) than of 87.5% (with no data augmentation). In addition, the proposed framework is compared with existing methods, as well as, the results of methods using in ICPR2014 and ICPR2016 competitions.The results demonstrate that our proposed method effectively outperforms recent methods.

1. Introduction

The analysis and classification of HEp-2 cell staining patterns of histopathological images are important processes in diagnosing autoimmune diseases [1]. Computer-Aided Diagnosis (CAD) techniques have been introduced to reduce the issues of manual annotation and classification [2]. This can help to reduce the error rate of decisions during the stage of diagnosis disease [3]. A CAD system can also be used as an aid in the training and education of specialised medical personnel. Recently, DL techniques have been widely used in biomedical, biological and medical image analysis applications for CAD system development to support physician and pathologists in making an effective and accurate decision of diagnosing disease [4]. Since several international contests have been conducted in the last few years, many research studies have been proposed in relation to automatic pattern recognition and classification of HEp-2 staining microscopic images at cell and specimen levels [5,6]. The HEp-2 tests produces variety of staining patterns; therefore, the structure of a cell includes the cytoplasm, nucleus, chromosomes, and nucleoli. These types are different in terms of their number, shape, location, and size of the cell. This can help an expert to distinguish among staining patterns to differentiate autoimmune diseases [7,8].
In contrast, existing CAD systems involve five fundamental aspects [9]: (i) image acquisition; (ii) enhancement; (iii) segmentation (iv) extraction and selection features; and (v) design of classification models. The CAD procedure and its performance depend on these fundamental aspects while the classification performance is influenced by the segmentation and feature extraction processes. Recently, automatic HEp-2 cell classification based on the extraction different features has become an active area of histopathological imaging research. Features that have been used to classify HEp-2 cells, consist of: (i) geometrical features, including shape index histogram with donut-shaped spatial pooling [10]; (ii) texture features, including analysis of morphological and textural feature [11]; and (iii) colour feature, including grayscale representation of HEp-2 cell [12]. The classification methods of histopathological images face certain challenges due to the variations in cell patterns, using different stains, variations in shape due to transformation operations, and time-consuming. In addition, a single specimen often contains cells with different patterns. There are also other issues, such as large numbers of cells per image, poor-quality of images, and clustering of cells. In-homogeneous illumination of images can cause huge intra-class variations, which negatively impact HEp-2 cell recognition and classification. Another issue is the overfitting of features used for classification, due to the very high dimensionality of features compared with the relatively small number of images for model training. HEp-2 cell classification is therefore a crucial processing step.
This work is inspired by recent advances in research into feature representation schemes and multi-class classification. It is also motivated by the development of machine learning and DL techniques addressing the problems of HEp-2 cell classification in CAD system, especially overfitting of features, stained images, intra-class variation and large-scale data volume problems. In this paper, a Multilayer Perceptron (MLP) neural network is adapted by using two hidden layers to classify the Higher Order Spectra (HOS) features into six classes based on the shapes of the HEp-2 staining patterns at cell level. The first new hidden layer includes the L-moments functions, which calculate the variation in the mean, scale, skewness and kurtosis of features vector. The second layer is calculated based on a Softmax activation function, which returns the probabilities of each class, where the target class has the highest probability. The proposed method is compared with current techniques and the results of classification competitions hosted at ICPR2014 and ICPR2016. The contribution of the work is that: (i) provides a robust technique for addressing the challenges of HEp-2 cell classification; (ii) unpublished method for analysis and classification of histopathological HEp-2 cell images using DL technique; (iii) produces an accurate recognition and classification processes of HEp-2 staining pattern at cell level for increasing the accuracy of CAD system at early diagnosing stages; (iv) provides a method that can recognize six classes of Hep-2 cell “(homogeneous, speckled, nucleolar, centromere, nuclear membrane, and golgi)”; and (v) provides high performance and benchmarking for classification of Hep-2 cell comparing with state-of-the-art methods.
This organisation of paper is as: Section 2 includes the explanation of the related work; Section 3 explains the proposed method; experiments and data description is detailed in Section 4; result evaluation is discussed in Section 5; and Section 6 summarizes the conclusion of this paper.

2. Related Work

Different machine learning and Deep Learning (DL) techniques have been proposed widely in different fields, including biomedical and medical images, remote sensing, biometric recognition, health informatics applications and so more. For instance, diagnosis of breast cancer from histopathological images [13], determination of Autoantibodies against HEp-2 cells (BCA) [14], medical image analysis [15], detection of COVID-19 from chest x-ray images [16], detection of biomedical imaging [17] white blood cell segmentation [18,19], classification of white blood cells [20,21], and HEp-2 cell segmentation from histopathological images [22]. Problems related to the classification of HEp-2 cell staining patterns from histopathological images have attracted the attention of many researchers in terms of benchmarking and comparison, and particularly the contests held at international conferences, such as pattern recognition ICPR2012, ICPR2014 and ICPR2016 as well as International Conference Image Processing (ICIP) [23].
Machine learning methods have been previously proposed to classify HEp-2 cells. The research of [24] used vector of hierarchically method for HEp-2 classification. A K-NN classifier had been been used [25] to classify HEp-cells. A linear Support Vector Machine (SVM) learning strategy and majority was proposed in [26] for specimen-HEp-2 level classification based on three types of features SID, SIFT and SURF [27]. A Gaussian mixture model has proposed in [28] for HEp-2 cell-level classification using different groups of features, including texture, statistical, spectral, binary, and intensity features. In [29], they used “region-level classification and majority voting for classification method”. They extract several sets of features for classification, including the adjacent local binary patterns, the covariance of the intensity, the local projection coefficients, and the morphological features. Recently, DL techniques have been proposed and merged with different techniques to classify HEp-2 cells. In [30], QRFIRMLP technique was used for the recognition of HEp-2 cells using a class of temporal processing neural networks based on finite impulse response filters. In [31], a DCAE technique has been proposed to perform feature extraction via an encoding–decoding scheme. They are automatically discriminated, during the feature learning process, and representations produce by the DCAE. CNNs, pattern histograms and linear SVMs were proposed in [32] for the classification of HEp-2 specimen patterns. In [33], deep CNNs were used to extract features directly from the pixel values of the cell image in a hierarchical way, without requirement to resort to hand-crafted features in order to classify HEp-2 cells [33]. The results showed that DL methods were successed against various large-scale benchmarks for the classification of HEp-2 specimen pattern images. However, all above methods still have challenges affecting the classification process and computation processing time, for example: (i) these methods lose discriminated information when using a fixed input image size, specifically with deformed edge in the segmented images; (ii) building a visual dictionary is complicated process such using the method in [32], especially when using a large number of training images in the model due to using data augmentation; and (iii) the segmentation process may affect negatively the performance of the classification process.

3. Proposed Method

The proposed method consists of various processes for classifying HEp-2 cell level patterns, as shown in Figure 1.

3.1. Segmentation of HEp-2 Cell Staining Images

HEp-2 cells were segmented by [22,34]. The segmentation method was proposed the level set method via Geometric Active Contours (GACs) to detect the HEp-2 cells shape information from poor quality microscope images.

3.2. Feature Extraction

Feature extraction is an important process in the automatic classification of the HEp-2 cells. To achieve the final feature extraction task, three steps are carried out from [20]. The steps are:
(1)
Radon projection: We use a Radon projection to convert a two dimension image into one dimension vector. In this paper, R [ θ ] is calculated using the procedure from research of [20].
(2)
Bispectrum: is the product of the Fourier coefficients at component frequencies [35]. The bispectrum S in the frequency domain is then written:
S ( k 1 , k 2 ) = F ( k 1 ) F ( k 2 ) F ( k 1 + k 2 )
“where the Fourier transform is represented by F ( k ) of R at each θ in the range [ 0 , 180 ] degrees. k 1 and k 2 are the normalised frequencies divided by one and half of the sampling frequency, and are in the range [0, 1] ” [35].
(3)
Bispectral invariants: “is a complex value that contains the information of the integrated bispectrum along a straight line a which gives a slope. a means the bispectral invariant feature of each θ , and is known as the phase.” A more detailed explanation of this method can be found in [20,35].

3.3. MultiLayer Perceptron

A MLP is one of a supervised DL technique that can learn f ( · ) : R m R o function. It is trained on a dataset, where the number of dimensions of the input is represented by m and the number of dimensions of the output is represented by o. Suppose a set of features X = x 1 , x 2 , , x m is given where the output is y, it can learn a non-linear function approximation for either classification or regression. In this paper, the hidden layers are based on two functions: (a) L-Moment measuring and (b) Softmax Activation Function, as described below:

3.3.1. L-Moment Measuring

The L-moment measures location, while the L-moment ratio measures the scale, skewness and kurtosis. A hidden layer in the MLP classifier is L-moments that calculation is used from [21]. Where the data H ( a ) θ are in ascending order, and a is the size of the individual projections (the length of the vector used to collect the results of each line integral). L-moment is used H ( a ) which indicated bispectral invariant feature vector, “L-Mean, L-Scale, L-Skewness and L-Kurtosis” are written as follows:
-
L-Mean which considers location features of cell, where L M = L 1 , and L 1 = β 0 ,
-
L-Scale which measures variation in scaling of the cell, where L S = L 2 / L 1 , and L 2 = 2 β 1 β 0 ,
-
L-Skewness which measures variation in concavity of cell, where L S K = L 3 / L 2 , and L 3 = 6 β 2 6 β 1 + β 0 ,
-
L-Kurtosis which measures variation in sharpness of cell, where L K = L 4 / L 2 , and L 4 = 20 β 3 30 β 2 + 12 β 1 β 0 .
β 0 , β 1 , β 2 and β 3 are written in Equations (4)–(7), where n is θ from 0 to 90 to measure the variation of cell shape
β 0 = 1 n j = 1 n H j
β 1 = 1 n j = 2 n H j [ ( j 1 ) / ( n 1 ) ]
β 2 = 1 n j = 3 n H j [ ( j 1 ) ( j 2 ) / ( n 1 ) ( n 2 ) ]
β 3 = 1 n j = 4 n H j [ ( j 1 ) ( j 2 ) ( j 3 ) / ( n 1 ) ( n 2 ) ( n 3 ) ]
The first hidden layer applies L-moments function, which calculates “ L-mean, L-scale, L-skewness and L-kurtosis”. The hidden layers can perform nonlinear transformations of the inputs entered into the network. Each hidden layer function is specialized to produce good features by calculating L-moments.

3.3.2. Softmax Activation Function

For multiple-classes, x is a hidden vector of features of L M , L S , L K S , L K containing K classes. It can pass through the function instead of passing through the logistic function. The node that has the highest value can be the input member of its class. The Softmax Activation Function corresponds to the neural network output representing the probability value that the input belongs to the certain class. This is written as:
P ( y = j | z i ) = softmax ( z ) i = exp ( z i ) l = 1 k exp ( z l )
where z is defined in follow equation:
z = w 1 x 1 + + w m x m + b = l = 1 m w l x l + b = w T x + b
where w is a weight vector, x is a feature vector of one training sample, and b is a bias unit. The role of Softmax function “ is computing the probability that this training sample x i can belong to class j by giving the weight and net input z i . The probability “ p ( y = j x ( i ) ; w j ) for each class label in j = 1 , , k can be computed. ”

3.3.3. Cross-Entropy Function

The mean squared error cost function can be used here to optimise the cost when a Softmax activation function is used at the output layer. The cross-entropy (C) in multi-class classification problems, is known to outperform the gradient descent function and is computed as follows: First, we need to define a cost function J ( · ) that represents the average of all cross-entropies over the n training samples that requires to minimise:
J ( W ; b ) = 1 n i = 1 n C ( Y i , O i ) ,
C ( Y i , O i ) = m Y i · log ( O i )
Y is the “target”, i.e., the true class labels, and the O is the probability computed via Softmax (output), but it is not the predicted class label. The derivative cost is computed as below equation in order to train our Softmax model and determine the weight coefficients via a gradient descent method:
w j J ( W ; b ) = 1 n i = 0 n x ( i ) O i Y i
Then, the cost derivative is used to update the weights w in opposite direction of the cost gradient with learning rate η for each class j { 0 , 1 , , k } [36]:
w j : = w j η w j J ( W ; b )
where w j is the weight vector of the class y = j . Then, the bias units are updated as Equation (12):
b j : = b j η 1 n i = 0 n O i T i .
We add an additional bias and a regularisation term “to reduce the variance of the model and decrease the degree of overfitting”, such as the L 2 term with the regularisation parameter λ : L 2 : λ 2 | | w | | 2 2 in Equation (13) where
| | w | | 2 2 = l = 0 m j = 0 k w i , j
The cost function becomes:
J ( W ; b ) = 1 n i = 1 n C ( T i , O i ) + λ 2 | | w | | 2 2
The “regularised” weight update is defined as:
w j : = w j η w j J ( W ) + λ w j .
The regularization parameter λ considers an input to our model, which reduces overfitting, by reducing the variance of our estimated regression parameters. Increasing λ results with less overfitting, but also has highest bias. One approach is randomly sub-sampled data number of times and looked at the variation in our estimate. Then, repeating the process for a slightly higher value of λ may affect the variability of our estimate model. We used a small value that can help achieving comparable regularization on the whole data set. In Figure 2, the leftmost layer, which is known as the input layer, consists of a set of neurons representing the input features x = x 1 , x 2 , , x m which represent the HOS features H ( a ) . The hidden layer consists of two layers which transform each neuron in the values from the previous layer, using a weighted linear summation w 1 x 1 + w 2 x 2 + + w m x m : The first hidden layer applies L-moments function, which calculates L-mean, L-scale, L-skewness and L-kurtosis. This is followed by a Softmax activation function R Y 6 . The softmax function calculates the probabilities variation of mean, scale, skewness and kurtosis of each target class over all possible target classes. The range value of Softmax will be from 0 to 1, and the sum of all the probabilities value should be equal to 1. For multi-classification model, the softmax function returns the value of probabilities of each class, and the target class will have the highest value probability. The output layer receives the values from the last hidden layer and transforms them into output values [37].

4. Experimental Analysis

Our experimental results are analyzed as follows:

4.1. Description of Dataset

The data set for Task-1 collected at the Sullivan Nicolaides pathology (SNP) laboratory from evaluations of positive serum samples for 419 patients. The dataset includes two classes: positive cells and intermediate cells, as shown Figure 3.
“The six patterns of cell staining in the dataset are: homogeneous (Hm), speckled (Sp), nucleolar (Nu), centromere (Cn), nuclear membrane (Nm), and Golgi (Gl), as shown in Figure 4. In the cell level classification of staining patterns, there are a total of 13,596 categorised cell images which each class has: homogeneous (2494), speckled (2831), nucleolar (2598), centromere (2741), Golgi (724) and nuclear membrane (2208).” The training set is augmented by rotation with an angle step of (90) and (30) to increase the number of images in the dataset, and to take the overfitting issue into consideration. The trained images are cropped to a size of 50 × 50 in order to standardise the size of the images before the rotation process is applied.

4.2. Implementation of Proposed Method

In this paper, classifying HEp-2 cell is implemented using MATLAB 2022b, and the process is illustrated in Figure 1. The proposed technique is used 80% of the dataset for the training set (10,833 images) and the remaining 20% for the testing set (2717 images). The steps of the proposed method are:
1
Pre-processing and segmentation:
Pre-processing is done by adjusting the intensity of image for increasing the contrast of the image. A level set method via edge-based GACs is then applied to detect the HEp-2 cell shape information from original microscope images used by [22].
2
Feature extraction using HOS:
The HOS technique is applied to the results of segmentation to extract features. The segmented HEp-2 cell image has been then converted to a set of 1D vectors using the “MATLAB Radon projection function”. This function can produce a Radon vector R for each angel from 0 to 180 degrees. A total of 256 features have been extracted, and the length of FFT used for each Radon projection is 1024. Finally, we obtained a set of 23,040 features for each image.
3
MLP using Softmax regression via gradient descent:
For the work presented in this paper, the neural network is implemented using Python 3.5.8. The MLP classifier model has four layers. Firstly, we encode the class labels into a certain format. One-hot encoding is applied, in which a sample belonging to Class-1 has the 1 value in the first cell of its row; a sample belonging to Class-2 has the 1 value in the second cell of its row, and so on. The input layer is a vector of (23,040) features multiplying by training images size of 10,833 × 23,040. Then, we initialise the parameter of weight matrix size of 10,833 × 23,040 × 6 (one column for each class and one row feature), where k represents four weights for each node. For example, the first row the matrix of dimensional weights is [0.1 0.2 0.3 0.4 0.5 0.6]. We construct a neural network with two hidden layers. The first hidden layer is calculated by summing the L-moments function, which includes the “ L-mean, L-scale, L-skewness and L-kurtosis”. We multiply this sum by the weight matrix w, and add the bias unit, which is [0.01 0.1 0.1 0.1 0.1 0.1], the result will be a 10,833 × 1024 matrix. The second hidden layer is calculated using a Softmax activation function. Following this, we find the average of all cross-entropies for 10 , 833 training images in order to learn our Softmax model, determining the weight coefficients (“regularised” weight) using gradient descent method. The learning rate (eta) is between [0.0, 1.0], and has a default value 0.01 . Using parameters I t e r a t i o n = 500 and C o s t = 0.06 , the prediction label is then created. The output layer is a vector of six class. Figure 5 shows an adapted MLP classifier using Softmax based gradient descent classification features using data augmentation and no data augmentation. Figure 6 shows an adapted MLP classifier using Softmax based gradient descent calculation cost and iteration and the best result is on i t e r a t i o n = 500 .

5. Discussion

5.1. Evaluation Results

The Mean Class Accuracy ( M C A ) is used for measuring the performance of classification classes and is adopted to score and compare the methods based on Correct Classification Rate C C R for each class. The prediction C C R n is calculated using two parameters from a confusion matrix, as illustrated in below equation:
C C R n = 1 M n ( T P n + T N n )
where T N n and T P n are the total number of True Negatives and True Positives for class n, respectively. M n represents the number of images belonging to the specfic n t h class. The average value of C C R n therefore considers the MCA value:
M C A = 1 n k = 1 n C C R n
Two experiments have done: In the first experiment, the testing and training datasets are randomly selected, with 2717 images used for testing and 10,833 for training. Average MCA values are 87.5 % with no data augmentation, and 90.03 % with data augmentation. These results show that the proposed method with data augmentation reduces the overfitting of features and results in better accuracy than the model with no data augmentation. In the second experiment, training sets have been created separately for the positive and intermediate cells. We train the same model configurations and parameters initialisation of the both categories to classify cells into positive or intermediate. we have submitted the method based on the networks error function “cross-entropy loss”. Then, the prediction label is then compared with the actual label, and the results are shown in Table 1. The results show that 3982 cell images are classified as positive and 4762 as intermediate cells. However, the difference between the values of validation and testing accuracy shows that this model may not be able to better generalise, as it sometimes suffers from overfitting to the training data, even when using the parameters chosen with the validation data. The 244,530 input features, 10,868 hidden with maximum iterations of 500 have been selected using Softmax activation to get 6 output class. “MLP performance is shown after the removal of the these features (one-by-one) at the train and test phases. If the removal of features had a negative impact on the performance of the classifier, it would be considered as highly important features for the separation of the 6 classes of HEp-2 cell”.

5.2. Benchmarking and Comparison with Other Techniques

Table 2 shows a comparison of the MCA values for our proposed method and other recent methods and the benchmarked dataset methods based on the Task-1 training dataset, which considered the average classification results for staining pattern images at cell level. We use the same training set of the competitions, and the results indicate that our proposed classification method outperforms all of the other methods (the first 16 rows) except the method that had done by [38] for which the result slightly higher than our proposed method. Our model achieves accuracy of 87.5% with no augmentation, and 90.03% with augmentation, and these result are higher than the best nine methods in Table 2. In addition, we implement plain MLP with the same HOS features and obtain MCA value of (84.32%. The result shows that adapted MLP is better than plain MLP.
The confusion matrix values are presented in Table 3. The confusion matrix has shown that the accuracy results of the proposed method with multi-class MLP for classification the cell into six classes are improved in comparison with values of plain MLP and another work in [7]. Furthermore, the confusion matrix values of the proposed method are different from those in [47]. For example, the speckled cell yields high accuracy of 90.00% when using the proposed method as shown in Figure 7, in contrast to the other methods, in which the features are more likely to be overlapped with the features of the centromere and homogeneous cells [47]. This difference is due to a failure to capture the cell shape information and an insufficient number of cells used to train and test resulting in the overlapped features being incorporated as true features by the proposed scenario. However, the method of research [38] results in slightly better than our proposed method by 91.02% due to discriminating between texturally similar patterns, such as Homogeneous and Speckled.

6. Conclusions and Future Work

In this paper, we present an effective and practical method for the classification histopathlogical images of HEp-2 staining patterns at cell level. A multi-class MLP is adapted by adding two hidden layers to automatically calculate the variation in the mean, scale, kurtosis and skewness of the input vector features based on HOS, and the Softmax regression algorithm is used to calculate the probability distribution of the variation in the mean, scale, kurtosis and skewness for each class. The neural network is trained and a predicted output vector is generated using a test dataset. The proposed algorithm is shown to perform well in practice, compared to existing methods in the literature and the state-of-the-art methods arising from benchmarked dataset. The proposed method achieved mean class accuracy of 87.50% with no augmentation and 90.03% with rotation augmentation. The performance of the proposed classification algorithm is also effective, both with and without data augmentation, and it is shown that the proposed bispectral invariant features and adapted multi-class MLP with data augmentation result in a higher classification accuracy (90.03%) than the 15 other methods reported in the two contests.

6.1. Proposed Methodology Advances

The results demonstrate that the proposed method is relatively invariant to the shape, rotation, scaling and shifting of cells, and is therefore robust against intra-class variation, overfitting and large-scale data volume phenomena. It has also been shown that the features are robust to variation in the mean, scale, kurtosis and skewness, and to discrimination between the different classes. The neural network is also trained to recognise the data set as positive or intermediate cells. The results demonstrate that proposed method can recognise 3982 cell images as positive from 4513, with an approximate accuracy of 88.23%, and 4762 as intermediate cell images from 5583, with an approximate accuracy of 85.30%. These results are satisfactory in comparison with the other methods investigated. The results show that the proposed method has an excellent adaptability across variations in scale, mean, skewness and kurtosis of the higher order spectra features, which is highly desirable for classification under different lab conditions.

6.2. Proposed Methodology Limitation

At present we have not calculated time performance, as our approach is implemented on different CPU processors and in different programming environments to other methods. Another limitation identified in the introduction is that the proposed method lose discriminative information when using fixed size images. Finally, the impact of this work is to enhance the decision of pathologists and the efficiency of CAD system, and particularly to discriminate between classes which is challenging in clinical practice. We believe that our proposed method has the potential to benefit patients for faster and more accurate diagnosis of diseases. In the future, our proposed method can be useful for designing computationally efficient HEp-2 classification method by reducing the number of network parameters ad running more data augmentation, and also designing the networks to be trained with smaller datasets. The work will study Scalability, Runtime, Memory, and Sensitivity analysis of proposed method as well as statistical analysis using T-test, as done in [48,49]. In addition, future work will study the combined Recurrent Neural Networks (RNNs) with One-vs-One classification and investigate their suitability for HEp-2 cells classification.

Author Contributions

K.A.-D., J.B., A.A.-S., K.N., V.C. and I.T.-R. have equally contributed this paper, including Conceptualization; methodology; software; validation; formal analysis; investigation, writing—original draft preparation; and writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study does not require ethical approval.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

The data is not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HEp-2Human Epithelial Type-2
CADComputer Aided Diagnoses
MLPMultilayer Perceptron
GACsGeometric Active Contours
SNPSullivan Nicolaides pathology
MCAMean Class Accuracy
SVMsSupport Vector Machines
CNNsConvolutional Neural Networks
DCAEDeep Convolutional AutoEncoder
FFTFast Fourier Transform
RNNsRecurrent Neural Networks
CCRCorrect Classification Rate
TPTrue Positive
TNTrue Negative

References

  1. Di Cataldo, S.; Tonti, S.; Bottino, A.; Ficarra, E. ANAlyte: A modular image analysis tool for ANA testing with indirect immunofluorescence. Comput. Methods Programs Biomed. 2016, 128, 86–99. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Manivannan, S.; Li, W.; Akbar, S.; Wang, R.; Zhang, J.; McKenna, S.J. An automated pattern recognition system for classifying indirect immunofluorescence images of HEp-2 cells and specimens. Pattern Recognit. 2016, 51, 12–26. [Google Scholar] [CrossRef] [Green Version]
  3. Shen, L.; Jia, X.; Li, Y. Deep cross residual network for HEp-2 cell staining pattern classification. Pattern Recognit. 2018, 82, 68–78. [Google Scholar] [CrossRef]
  4. Srinidhi, C.L.; Kim, S.W.; Chen, F.D.; Martel, A.L. Self-supervised driven consistency training for annotation efficient histopathology image analysis. Med. Image Anal. 2022, 75, 102256. [Google Scholar] [CrossRef]
  5. Al-Dulaimi, K.; Chandran, V.; Nguyen, K.; Banks, J.; Tomeo-Reyes, I. Benchmarking HEp-2 Specimen Cells Classification Using Linear Discriminant Analysis on Higher Order Spectra Features of Cell Shape. Pattern Recognit. Lett. 2019, 125, 534–541. [Google Scholar] [CrossRef]
  6. Hobson, P.; Lovell, B.C.; Percannella, G.; Vento, M.; Wiliem, A. Benchmarking human epithelial type 2 interphase cells classification methods on a very large dataset. Artif. Intell. Med. 2015, 65, 239–250. [Google Scholar] [CrossRef] [Green Version]
  7. Hobson, P.; Lovell, B.C.; Percannella, G.; Saggese, A.; Vento, M.; Wiliem, A. HEp-2 staining pattern recognition at cell and specimen levels datasets, algorithms and results. Pattern Recognit. Lett. 2016, 82, 12–22. [Google Scholar] [CrossRef] [Green Version]
  8. Ponomarev, G.V.; Arlazarov, V.L.; Gelfand, M.S.; Kazanov, M.D. ANA HEp-2 cells image classification using number, size, shape and localization of targeted cell regions. Pattern Recognit. 2014, 47, 2360–2366. [Google Scholar] [CrossRef] [Green Version]
  9. Ensafi, S.; Lu, S.; Kassim, A.A.; Tan, C.L. Automatic CAD System for HEp-2 Cell Image Classification. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 3321–3326. [Google Scholar]
  10. Larsen, A.B.L.; Vestergaard, J.S.; Larsen, R. HEp-2 cell classification using shape index histograms with donut-shaped spatial pooling. IEEE Trans. Med. Imaging 2014, 33, 1573–1580. [Google Scholar] [CrossRef]
  11. Di Cataldo, S.; Bottino, A.; Islam, I.U.; Vieira, T.F.; Ficarra, E. Subclass discriminant analysis of morphological and textural features for hep-2 staining pattern classification. Pattern Recognit. 2014, 47, 2389–2399. [Google Scholar] [CrossRef]
  12. Cordelli, E.; Soda, P. Methods for greyscale representation of HEp-2 colour images. In Proceedings of the IEEE 23rd International Symposium on Computer-Based Medical Systems (CBMS), Bentley, Australia, 12–15 October 2010; pp. 383–388. [Google Scholar]
  13. Carvalho, E.D.; Antonio Filho, O.; Silva, R.R.; Araujo, F.H.; Diniz, J.O.; Silva, A.C.; Paiva, A.C.; Gattass, M. Breast cancer diagnosis from histopathological images using textural features and CBIR. Artif. Intell. Med. 2020, 105, 101845. [Google Scholar] [CrossRef] [PubMed]
  14. de Melo Cruvinel, W.; Andrade, L.E.C.; von Muhlen, C.A.; Dellavance, A.; Ximenes, A.C.; Bichara, C.D.; Bueno, C.; Mangueira, C.L.P.; Bonfa, E.; de Almeida Brito, F.; et al. V Brazilian consensus guidelines for detection of anti-cell autoantibodies on hep-2 cells. Adv. Rheumatol. 2022, 59, 28. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Jungo, A.; Scheidegger, O.; Reyes, M.; Balsiger, F. pymia: A Python package for data handling and evaluation in deep learning-based medical image analysis. Comput. Methods Programs Biomed. 2021, 198, 105796. [Google Scholar] [CrossRef] [PubMed]
  16. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  17. Zhou, X.; Li, Z.; Xue, Y.; Chen, S.; Zheng, M.; Chen, C.; Yu, Y.; Nie, X.; Lin, X.; Wang, L.; et al. CUSS-Net: A Cascaded Unsupervised-based Strategy and Supervised Network for Biomedical Image Diagnosis and Segmentation. IEEE J. Biomed. Health Inform. 2023. [Google Scholar] [CrossRef]
  18. AL-Dulaimi, K.; Al-Sabaawi, A.; Resen, R.D.; Stephan, J.J.; Zwayen, A. Using adapted JSEG algorithm with Fuzzy C Mean for segmentation and counting of white blood cell and nucleus images. In Proceedings of the 6th IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE), Melbourne, Australia, 9–11 December 2019; Volume 2019, pp. 1–7. [Google Scholar]
  19. AL-Dulaimi, K.; Tomeo-Reyes, I.; Banks, J.; Chandran, V. Evaluation and benchmarking of level set-based three forces via geometric active contours for segmentation of white blood cell nuclei shape. Comput. Biol. Med. 2020, 116, 103568. [Google Scholar] [CrossRef]
  20. AL-Dulaimi, K.; Chandran, V.; Banks, J.; Tomeo-Reyes, I.; Nguyen, K. Classification of white blood cells using bispectral invariant features of nuclei shape. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Canberra, ACT, Australia, 10–13 December 2018; pp. 19–26. [Google Scholar]
  21. Al-Dulaimi, K.; Nguyen, K.; Banks, J.; Chandran, V.; Tomeo-Reyes, I. Classification of White Blood Cells Using L-Moments Invariant Features of Nuclei Shape. In Proceedings of the International Conference on Image and Vision Computing New Zealand (IVCNZ), Auckland, New Zealand, 19–21 November 2018; pp. 1–6. [Google Scholar]
  22. AL-Dulaimi, K.; Banks, J.; Tomeo-Reyes, I.; Chandran, V. Automatic segmentation of HEp-2 cell Fluorescence microscope images using level set method via geometric active contours. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 81–83. [Google Scholar]
  23. Hobson, P.; Percannella, G.; Vento, M.; Wiliem, A. International Competition on Cells Classification by Fluorescent Image Analysis. In Proceedings of the 20th IEEE International Conference on Image Processing (ICIP), Melbourne, Australia, 15–18 September 2013. [Google Scholar]
  24. Kastaniotis, D.; Fotopoulou, F.; Theodorakopoulos, I.; Economou, G.; Fotopoulos, S. HEp-2 cell classification with Vector of Hierarchically Aggregated Residuals. Pattern Recognit. 2017, 65, 47–57. [Google Scholar] [CrossRef]
  25. Stoklasa, R.; Majtner, T.; Svoboda, D. Efficient K-NN based HEp-2 cells classifier. Pattern Recognit. 2014, 47, 2409–2418. [Google Scholar] [CrossRef]
  26. Ensafi, S.; Lu, S.; Kassim, A.A.; Tan, C.L. Accurate HEp-2 cell classification based on sparse coding of superpixels. Pattern Recognit. Lett. 2016, 82, 64–71. [Google Scholar] [CrossRef]
  27. Gragnaniello, D.; Sansone, C.; Verdoliva, L. Cell image classification by a scale and rotation invariant dense local descriptor. Pattern Recognit. Lett. 2016, 82, 72–78. [Google Scholar] [CrossRef]
  28. Sarrafzadeh, O.; Rabbani, H.; Dehnavi, A.M.; Talebi, A. Analyzing features by SWLDA for the classification of HEp-2 cell images using GMM. Pattern Recognit. Lett. 2016, 82, 44–55. [Google Scholar] [CrossRef]
  29. Sakrapee, P.; Chunhua, S.; van den Hengel, A. A scalable stagewise approach to large-margin multiclass loss-based boosting. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 1002–1013. [Google Scholar]
  30. Codrescu, C. Quadratic recurrent finite impulse response MLP for indirect immunofluorescence image recognition. In Proceedings of the 2014 1st Workshop on Pattern Recognition Techniques for Indirect Immunofluorescence Images, Stockholm, Sweden, 24 August 2014; pp. 49–52. [Google Scholar]
  31. Vununu, C.; Lee, S.H.; Kwon, K.R. A Strictly Unsupervised DL Method for HEp-2 Cell Image Classification. Sensors 2020, 20, 2717. [Google Scholar] [CrossRef] [PubMed]
  32. Li, H.; Huang, H.; Zheng, W.-S.; Xie, X.; Zhang, J. HEp-2 specimen classification via deep CNNs and pattern histogram. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 2145–2149. [Google Scholar]
  33. Gao, Z.; Wang, L.; Zhou, L.; Zhang, J. HEp-2 cell image classification with deep convolutional neural networks. IEEE J. Biomed. Health Inform. 2017, 21, 416–428. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. AL-Dulaimi, K.; Tomeo-Reyes, I.; Banks, J.; .Chandran, V. White blood cell nuclei segmentation using level set methods and geometric active contours. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, QLD, Australia, 30 November–2 December 2016; pp. 1–7. [Google Scholar]
  35. Vinod, C.; Brett, C.; Boualem, B.; Steve, E. Pattern recognition using invariants defined from higher order spectra: 2-D image inputs. Trans. Image Process. 1997, 6, 703–712. [Google Scholar]
  36. Raschka, S. Softmax Regression-Gradient Descent. 2014–2019. Available online: http://rasbt.github.io/mlxtend/user_guide/classifier/SoftmaxRegression/ (accessed on 10 December 2022).
  37. Gibson, E.; Li, W.; Sudre, C.; Fidon, L.; Shakir, D.I.; Wang, G.; Eaton-Rosen, Z.; Gray, R.; Doel, T.; Hu, Y.; et al. NiftyNet: A deep-learning platform for medical imaging. Comput. Methods Programs Biomed. 2018, 158, 113–122. [Google Scholar] [CrossRef]
  38. Oei, R.W.; Hou, G.; Liu, F.; Zhong, J.; Zhang, J.; An, Z.; Xu, L.; Yang, Y. Convolutional neural network for cell classification using microscope images of intracellular actin networks. PLoS ONE 2019, 14, e0213626. [Google Scholar] [CrossRef]
  39. Qi, X.; Zhao, G.; Chen, J.; Pietikäinen, M. HEp-2 cell classification: The role of gaussian scale space theory as a pre-processing approach. Pattern Recognit. Lett. 2016, 82, 36–43. [Google Scholar] [CrossRef] [Green Version]
  40. Han, X.H.; Chen, Y.W.; Xu, G. Integration of spatial and orientation contexts in local ternary patterns for HEp-2 cell classification. Pattern Recognit. Lett. 2016, 82, 23–27. [Google Scholar] [CrossRef]
  41. Nanni, L.; Lumini, A.; dos Santos, F.L.C.; Paci, M.; Hyttinen, J. Ensembles of dense and dense sampling descriptors for the HEp-2 cells classification problem. Pattern Recognit. Lett. 2016, 82, 28–35. [Google Scholar] [CrossRef]
  42. Cascio, D.; Taormina, V.; Cipolla, M.; Bruno, S.; Fauci, F.; Raso, G. A multi-process system for HEp-2 cells classification based on SVM. Pattern Recognit. Lett. 2016, 82, 56–63. [Google Scholar] [CrossRef]
  43. Theodorakopoulos, I.; Kastaniotis, D.; Economou, G.; Fotopoulos, S. HEp-2 cells classification using morphological features and a bundle of local gradient descriptors. In Proceedings of the 1st Workshop on Pattern Recognition Techniques for Indirect Immunofluorescence Images, Stockholm, Sweden, 24 August 2014; pp. 33–36. [Google Scholar]
  44. Meng, N.; Lam, E.Y.; Tsia, K.K.; So, H.K.H. Large-scale multi-class image-based cell classification with deep learning. IEEE J. Biomed. Health Inform. 2018, 23, 2091–2098. [Google Scholar] [CrossRef]
  45. Nguyen, L.D.; Gao, R.; Lin, D.; Lin, Z. Biomedical image classification based on a feature concatenation and ensemble of deep CNNs. J. Ambient Intell. Humaniz. Comput. 2019, 1–13. [Google Scholar] [CrossRef]
  46. Jorgensen, B.; AL-Dulaimi, K.; Banks, J. HEp-2 Specimen Cell Detection and Classification Using Very Deep Convolutional Neural Networks-Based Cell Shape. In Proceedings of the International Conference on Digital Image Computing: Techniques and Applications (DICTA), Goldcoast, Australia, 30 November–2 December 2021; Volume 2021. [Google Scholar]
  47. Manivannan, S.; Li, W.; Akbar, S.; Wang, R.; Zhang, J.; McKenna, S.J. HEp-2 cell classification using multi-resolution local patterns and ensemble SVMs. In Proceedings of the 1st Workshop on Pattern Recognition Techniques for Indirect Immunofluorescence Images, Stockholm, Sweden, 24 August 2014; pp. 37–40. [Google Scholar]
  48. El-kenawy, E.S.M.; Abutarboush, H.F.; Mohamed, A.W.; Ibrahim, A. Advance artificial intelligence technique for designing double T-shaped monopole antenna. Comput. Mater. Contin. 2021, 69, 2983–2995. [Google Scholar] [CrossRef]
  49. Yu, Z.; Guindani, M.; Grieco, S.F.; Chen, L.; Holmes, T.C.; Xu, X. Beyond t test and ANOVA: Applications of mixed-effects models for more rigorous statistical analysis in neuroscience research. Neuron 2022, 110, 21–35. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Proposed method processes to classify HEp-2 cell level patterns, including segmentation, feature extraction and multi-classes classification using adapted MLP neural network with two layers.
Figure 1. Proposed method processes to classify HEp-2 cell level patterns, including segmentation, feature extraction and multi-classes classification using adapted MLP neural network with two layers.
Sensors 23 02195 g001
Figure 2. A MLP network architecture shows input layer, hidden layer and output layer.
Figure 2. A MLP network architecture shows input layer, hidden layer and output layer.
Sensors 23 02195 g002
Figure 3. Two under-classes can be defined: (Row 1) positive cells (high intensity) and (Row 2) intermediate cells (low intensity).
Figure 3. Two under-classes can be defined: (Row 1) positive cells (high intensity) and (Row 2) intermediate cells (low intensity).
Sensors 23 02195 g003
Figure 4. Cell staining included six pattern in the dataset: (a) homogeneous (Hm), (b) speckled (Sp), (c) nucleolar (Nu), (d) centromere (Cn), (e) nuclear membrane (Nm), and (f) Golgi (Gl).
Figure 4. Cell staining included six pattern in the dataset: (a) homogeneous (Hm), (b) speckled (Sp), (c) nucleolar (Nu), (d) centromere (Cn), (e) nuclear membrane (Nm), and (f) Golgi (Gl).
Sensors 23 02195 g004
Figure 5. An adapted MLP classifier using Softmax based gradient descent classification features using no data augmentation (first row) and data augmentation (second row), where x-axis is training vectors, containing the number of samples and the number of features, and y-axis is the number of samples to plot decision borders.
Figure 5. An adapted MLP classifier using Softmax based gradient descent classification features using no data augmentation (first row) and data augmentation (second row), where x-axis is training vectors, containing the number of samples and the number of features, and y-axis is the number of samples to plot decision borders.
Sensors 23 02195 g005
Figure 6. A adapted MLP classifier using Softmax based gradient descent calculation cost and iteration.
Figure 6. A adapted MLP classifier using Softmax based gradient descent calculation cost and iteration.
Sensors 23 02195 g006
Figure 7. Correct classification rates (CCR) for each cell class and mean value of methods comparing with our proposed method to show the variation in classification of each class.
Figure 7. Correct classification rates (CCR) for each cell class and mean value of methods comparing with our proposed method to show the variation in classification of each class.
Sensors 23 02195 g007
Table 1. Accuracy of positive/intermediate cells for each class and the number of images which is classified correctly from test images.
Table 1. Accuracy of positive/intermediate cells for each class and the number of images which is classified correctly from test images.
Actual ClassPositive CellIntermediate Cell
ImagesRate%ImagesRate%
Homogeneous722/81588.59%902/105585.50%
Speckled900/109282.42%900/103087.38%
Nucleola450/50090.00%1000/124880.13%
Centromere955/103392.45%950/102292.96%
Nuclear Membrane655/70792.65%810/94885.44%
Golgi300/36681.97%200/28071.43%
Overall3982/451388.23%4762/558385.30%
Table 2. Results were achieved by the participants compared in terms of MCA to the recent methods, ICPR2014 and ICPR2016 contests (the first sixteenth rows of the table) and our adapted (MLP) classifier (the remaining rows of the table in bold text) over Task-1 training dataset.
Table 2. Results were achieved by the participants compared in terms of MCA to the recent methods, ICPR2014 and ICPR2016 contests (the first sixteenth rows of the table) and our adapted (MLP) classifier (the remaining rows of the table in bold text) over Task-1 training dataset.
ReferencesFeature Extraction and SelectionClassifierData AugmentationTrain SetTest Set
[30]Trainable featuresQR-FIRMLPMirroring and rotation98.9474.68
[29]CoALBP, STR, LPCMulticlass boostingRotation100.0081.50
[26]SIFT and SURF with BoWLinear SVMs & Majority voting98.0780.84
[27]SID with soft BoWLinear SVM (one-vs-all)95.4783.85
[33]Trainable featuresDeep CNNsRotation89.0276.26
[39]LOAD with IFVLinear SVM (one-vs-all)99.9184.26
[2]Multi-resolution LP & Root-SIFTSVMs with Platte re-scalingRotation9587.42
[40]RICWLTPLinear SVM (one-vs-all)94.6868.37
[41]LCP, RIC-LBP, ELBP, PLBP, STRSVM with Kernel RBFResizing & Rotation100.0079.91
[42]Geometry, morphology & entropySVM (one-vs-one) cell level90.2580.45
[43]Morphological & textural featuresLinear SVM (one-vs-all)93.8283.06
[28]Statistical, spectral & LDAGaussian mixture model88.5973.78
[24]SIFT descriptorsVector of hierarchically residuals82.80
[38]CNN-based Softmaxrotation, cropping & flipping95.3291.33
[44]CNNs94.0189.52
[45]feature concatenation & ensembleCNNs96.5689.00
[46]Very deep CNNsRotation 89.36
MLP methodHigher order spectraPlain MLPNo augmentation90.2284.32
Our proposed methodHigher order spectraAMLP based L-momentNo augmentation95.8287.55
Our proposed methodHigher order spectraAMLP based L-momentRotation97.1190.83
Table 3. Confusion matrix parameters Classification results of comparable methods of ICPR2014, ICPR2016 and recent techniques using the Task-1 data set for each class.
Table 3. Confusion matrix parameters Classification results of comparable methods of ICPR2014, ICPR2016 and recent techniques using the Task-1 data set for each class.
ReferencesHmSpCnNuGoNm
[30]69.1672.5968.6867.0894.2176.15
[29]75.8482.9376.475.5694.5183.78
[26]75.5381.4376.6173.794.1883.57
[27]83.0282.2685.3778.495.3778.68
[33]80.7964.6573.5167.6285.5273.3
[39]89.9180.6786.8481.5385.580.11
[2]87.4780.5183.0491.0189.8492.09
[40]68.5853.5163.0364.7483.0277.36
[41]89.6976.2170.3178.0584.4680.05
[42]89.1976.370.1777.9586.2882.33
[43]91.1179.2475.0578.0687.2987.12
[28]80.7263.6271.1166.684.8375.15
[24]88.9377.779.2583.0090.4077.05
[38]93.2890.0188.0891.3691.5693.57
[44]92.1287.8086.5188.0591.6291.01
[45]84.5290.0191.3380.0491.9796.31
[46]------
plain MLP82.1382.1490.1288.1390.0882.10
Proposed method-184.5385.3491.3280.1391.8091.15
Proposed method-291.9189.8189.5185.9791.6796.16
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Al-Dulaimi, K.; Banks, J.; Al-Sabaawi, A.; Nguyen, K.; Chandran, V.; Tomeo-Reyes, I. Classification of HEp-2 Staining Pattern Images Using Adapted Multilayer Perceptron Neural Network-Based Intra-Class Variation of Cell Shape. Sensors 2023, 23, 2195. https://doi.org/10.3390/s23042195

AMA Style

Al-Dulaimi K, Banks J, Al-Sabaawi A, Nguyen K, Chandran V, Tomeo-Reyes I. Classification of HEp-2 Staining Pattern Images Using Adapted Multilayer Perceptron Neural Network-Based Intra-Class Variation of Cell Shape. Sensors. 2023; 23(4):2195. https://doi.org/10.3390/s23042195

Chicago/Turabian Style

Al-Dulaimi, Khamael, Jasmine Banks, Aiman Al-Sabaawi, Kien Nguyen, Vinod Chandran, and Inmaculada Tomeo-Reyes. 2023. "Classification of HEp-2 Staining Pattern Images Using Adapted Multilayer Perceptron Neural Network-Based Intra-Class Variation of Cell Shape" Sensors 23, no. 4: 2195. https://doi.org/10.3390/s23042195

APA Style

Al-Dulaimi, K., Banks, J., Al-Sabaawi, A., Nguyen, K., Chandran, V., & Tomeo-Reyes, I. (2023). Classification of HEp-2 Staining Pattern Images Using Adapted Multilayer Perceptron Neural Network-Based Intra-Class Variation of Cell Shape. Sensors, 23(4), 2195. https://doi.org/10.3390/s23042195

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop