Next Article in Journal
How Little Do We Know about HIV and STIs Prevention? Results from a Web-Based Survey among the General Population
Previous Article in Journal
Assessment of Parental Knowledge towards Space Maintainer as an Essential Intervention after Premature Extraction of Primary Teeth
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma

1
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakakah 72312, Saudi Arabia
2
School of Information Technology & Engineering, Vellore Institute of Technology, Vellore 632001, India
3
School of Computer Science (SCS), Taylor’s University, Subang Jaya 47500, Malaysia
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(6), 1058; https://doi.org/10.3390/healthcare10061058
Submission received: 19 April 2022 / Revised: 4 June 2022 / Accepted: 6 June 2022 / Published: 8 June 2022
(This article belongs to the Section Artificial Intelligence in Medicine)

Abstract

:
Lung cancer is among the most hazardous types of cancer in humans. The correct diagnosis of pathogenic lung disease is critical for medication. Traditionally, determining the pathological form of lung cancer involves an expensive and time-consuming process investigation. Lung cancer is a leading cause of mortality worldwide, with lung tissue nodules being the most prevalent way for doctors to identify it. The proposed model is based on robust deep-learning-based lung cancer detection and recognition. This study uses a deep neural network as an extraction of features approach in a computer-aided diagnosing (CAD) system to assist in detecting lung illnesses at high definition. The proposed model is categorized into three phases: first, data augmentation is performed, classification is then performed using the pretrained CNN model, and lastly, localization is completed. The amount of obtained data in medical image assessment is occasionally inadequate to train the learning network. We train the classifier using a technique known as transfer learning (TL) to solve the issue introduced into the process. The proposed methodology offers a non-invasive diagnostic tool for use in the clinical assessment that is effective. The proposed model has a lower number of parameters that are much smaller compared to the state-of-the-art models. We also examined the desired dataset’s robustness depending on its size. The standard performance metrics are used to assess the effectiveness of the proposed architecture. In this dataset, all TL techniques perform well, and VGG 16, VGG 19, and Xception for 20 epoch structure are compared. Preprocessing functions as a wonderful bridge to build a dependable model and eventually helps to forecast future scenarios by including the interface at a faster phase for any model. At the 20th epoch, the accuracy of VGG 16, VGG 19, and Xception is 98.83 percent, 98.05 percent, and 97.4 percent.

1. Introduction

Cancer is a significant concern across the globe, and it is the second foremost reason for death. Lung cancer stands in third place among cancer deaths. Lung cancer is broadly classified into small and non-small-cell lung cancer. Various subtypes in non-small-cell lung cancer are adenocarcinoma, squamous cell cancer, and large cell carcinoma. Non-small-cell lung cancer is frequently observed, but small cell lung cancer spreads faster and is often fatal. Changes to the affected person’s voice, chest pain, shortness of breath, and wheezing are a few symptoms to list, and more painful symptoms also prevail [1,2]. Chemotherapy, targeted drugs, and immunotherapy are the treatments that have been approved over the decades [3].
Google Trends show the potential research going on in the field of cancer, and this is graphed in Figure 1. Several progressive methods are being devised for earlier detection with the help of artificial intelligence concepts.
Our main contribution to the work is as follows:
  • We have designed a model which can classify the patients’ level of lung carcinoma by applying the TL process, and it is the first of the type to be carried out with this dataset [4].
  • Using CT images as the system’s input, it can predict the level and helps to take contour action at the earliest time.
  • We have applied three TL approaches here, namely VGG16, VGG19, and Xception, with 20 epochs in the Google Colab platform, and it was shown to be a better model to use for future prediction.
  • Based on the experimental performance for lung carcinoma, VGG16 gives maximum accuracy of 98.83%, whereas Xception shows an accuracy rate of 97.4%.
The work is carried out in a phased manner. Section 2 of the paper discusses the recent work carried out in this domain. Section 3 details the workflow with architecture diagram illustration and fine-tuned hyperparameter values. Section 4 elaborates on the experimental discussion, followed by a comparison with recent results, conclusion, and future work in Section 5.

2. Related Work

Early diagnosis of the cancer cell is the required parameter irrespective of the organ. The initiation and growth of the tumor cell makes the life of the patient very difficult. Data filter algorithm for preprocessing, region growing algorithm for image segmentation, and feature selection followed by convolutional neural networks for classification are the various steps performed on computed tomography (CT) images to obtain proof of whether there is malignancy or not in the case of lungs [5]. Clinical attributes and the images of the work are used to predict the ailments at the earlier stage. Recent research in China has established a rapid blood test that screens lung cancer at the early stage. Smoking, radon exposure, lethal chemicals, particle pollution, and genetic factors are leading influences on lung cancer [4].
The EL-CAP dataset used in the work was created by Cornell University in 2003 and holds 500 low-dose images. The images were segregated in two folders based on the benign and malignant classes. For the work, 75% was considered as the training set and 25% as the testing set. A hybrid approach with a support vector machine and a feed-forward back propagation network was utilized to reduce the computational complexity in the classification process and produce accuracy of 98.08% [6]. The work was carried out with the enhanced spiral CT and MRI images of lung cancer obtained from 74 patients considered to be highly suspected of having lung cancer. CT and MRI accuracy was 94.6% and 89.2%, respectively. Data enhancement was performed. The implementation flow dumps data were followed by pretrained model loading and VGG-based TL accompanied by hyperparameter fine-tuning [7].
The risk factor associated with lung cancer was analyzed with the help of a database constructed from 1000 patients with 23 features related to the symptom and target as 3 values: low, medium, and high. Data visualization was performed, followed by decision trees and the random forest approach, to predict the lung cancer risk level. The accuracy obtained in this model was 93.33% [8]. The dataset considered was from sick people with non-small-cell lung cancer, who were treated with anti-programmed death therapy. Various attributes pertaining to the characteristics of patients, mutations, and results were retrieved from the laboratory. Out of several attributes, 19 were used to build the model, using machine learning algorithms such as ridge regression, multilayer neural network, LightBGM, XGBoost, linear discriminant analysis, and ridge regression. Feature engineering accompanied by machine learning shows that LightBGM serves well to build a highly reliable model for predicting lung cancer [9].
Artificial intelligence helps in diagnosis of the tumors at the earliest stage with the images collected from the cancer-prone area. Images captured via computed tomography, magnetic resonance, ultrasound, endoscopy procedure, and biopsies aid in better prediction. Breast, lung, thyroid, gastric, oral, skin, and liver cancer are interpreted with principles of artificial intelligence and expert inputs [10]. The deep transfer neural network and extreme learning machine were integrated and experimented on over the lung image database consortium and image database resource initiative dataset, and accuracy of 94.94% was achieved [11]. The cancer incidence data for this study were taken from the Surveillance, Epidemiology, and End Results (SEER) program. Statistics data are provided to combat the burden of cancer in the US at the early stages. Work carried out using ANN, RNN, and CNN paved the way for performance measures, and accuracy is 71.18% [12].
Several sources were used to compile a dataset of chest computed tomography scans with lung nodules and their pathologic diagnoses. We created, trained, and verified two machine learning algorithms. The support vector machine model was employed in the first algorithm, and the convolutional neural network was used in the second. The performance of the classification on the test dataset was evaluated using receiver operating characteristic analysis [13].
This work used the enhanced profuse clustering method (IPCT) and the deep learning with instantaneously trained neural networks (DITNN) strategy to evaluate lung CT images for predicting lung cancer. The lung CT pictures were first obtained from the Cancer Imaging Archive (CIA) collection and were then improved by constructing a weighted mean function that substituted the pixel using probability distribution and the cumulative distribution procedure. After improving the image’s representation, the damaged area was segregated using the pixel similarity value followed by the clusters that were established [14].
In order to provide prompt therapy to lung cancer patients, accurate detection of the disease is critical. Artificial neural networks (ANNs) are a recently developed machine learning algorithm that can be applied to both large and small datasets. An ensemble of weight optimized neural networks with maximum likelihood boosting for LCD in huge data is investigated in this research. Feature selection and ensemble classification are the two stages of the proposed technique. To reduce classification time, the important attributes were chosen in the first step using an integrated Newton–Raphson maximum likelihood and minimum redundancy (MLMR) preprocessing model. The patient was then classified using the boosted weighted optimized neural network ensemble classification algorithm in the second stage [15].

3. Workflow Architecture

Analysis of the cancer dataset with the TL was very informative and yielded a good model based on the various performance measure perspectives. Capturing the lung cancer image dataset was followed by preprocessing using the image data generator, then the TL model, and finally the prediction of the type of abnormality was achieved, as is portrayed in Figure 2.

3.1. Dataset Description

In the fall of 2019, the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases (IQ-OTH/NCCD) collected data for three months. The dataset comprised CT scans captured from patients in different stages and scans captured from patients in different stages and healthy people in DICOM format. The following CT protocol was utilized for reading: 120 kV, 1 mm slice thickness, window widths varying from 350 to 1200 HU, and window centers ranging from 50 to 600. An oncologist and radiologist working in the centers marked the slides. From 110 cases, 1190 images were retrieved. All the cases were classified into 3 classes, namely normal, benign, and malignant. In total, 55 cases were classified as normal, 15 cases were classified as benign, and 40 cases were classified as malignant.
Prior to analysis, all photos were de-identified. The oversight review board waived written consent. The institutional review boards of the participating medical centers approved the study. There were multiple slices in each scan. The number of slices varied between 80 and 200, and each one displayed an image of the human chest from various sides and perspectives. Gender, age, educational attainment, residency area, and living status were all different in the 110 cases. Some worked for the Iraqi ministries of transportation and oil, while others were farmers and gainers. Most of them were from Iraq’s central area, specifically the provinces of Baghdad, Wasit, Diyala, Salahuddin, and Babylon [16].

3.2. Image Data Generator

To expand the size of the dataset image, augmentation phase is deployed. In Keras, image data generator is used to achieve the augmentation. Different transformation is applied to the original to make it sizable in applying the deep learning structure. It also integrates variation in the images so that the built model will be reliable. This process requires reduced memory only. In our work, the training data following the augmentation technique have values as follows: rescale = 1.255, shear_range = 0.2, zoom_range = 0.2, and horizontal_flip = True. Upon analysis with various options, we found that this was most suitable.

3.3. TL

To the created image dataset pool, TL algorithms, namely, VGG 16, VGG 19, and Xception were applied with fine-tuning hyperparameters to structure the network for training and testing in turn, in order to help in future prediction for new records. First, work was carried out with deep neural networks, meant for object detection. The alternative to deep learning is the TL method, and it works on four transfer methods, namely, instance-based, feature-based, parameter-based, and relation-based. Applications of TL are substantial and this article provides insight [17,18]. GPU support in the Google Colab platform provided the helping hand to overcome the computation glitches and showed good throughput.

3.4. Fine-Tuned Hyperparameters

Model hyperparameters were used to optimize the model, loss used was categorical_crossentropy, optimizer was adam. Fit_generator was the iterator method integrated to fit the data on the neural network simulation. Parameters were target_size = (224,224), batch_size = 32, and epochs = 20, and class_mode was categorical, based (normal, benign, malignant) on the nature of the dataset. In total, 75% of data were for training purposes. Varying hyperparameters were tried and executed to obtain optimized results.
The following performance measures were used to study the model consistency. The training and testing criteria showed varied performance [19,20].

3.4.1. Accuracy

Accuracy is an ideal classification metric and is easy to understand [21,22]. It is the proportion of true results to total number of results. It is easily suited to binary and multiclass classification scenarios. It is a valid evaluation method for problems where data are not skewed and are well-balanced. Below is the confusion matrix to substantiate the 3-class performance measure.
Accuracy = T P + T N T P + F P + F N + T N
ACTUAL VALUES
NormalBenignMalignant
PREDICTED VALUESNormal+ve
1
−ve
2
−ve
3
Benign−ve
4
+ve
5
−ve
6
Malignant−ve
7
−ve
8
+ve
9
Normal
  • TP = Cell1
  • FP = Cell2 + Cell3
  • TN = Cell5 + Cell6 + Cell8 + Cell9
  • FN = Cell4 + Cell7
Benign
  • TP = Cell5
  • FP = Cell4 + Cell6
  • TN = Cell1 + Cell3 + Cell7 + Cell9
  • FN = Cell2 + Cell8
Malignant
  • TP = Cell9
  • FP = Cell7 + Cell8
  • TN = Cell1 + Cell2 + Cell4 + Cell5
  • FN = Cell3 + Cell6
  • TP—True Positive—predict yes and have lung carcinoma
  • TN—True Negative—correctly predict that they have no lung carcinoma
  • FP—False Positive—incorrectly predict having lung carcinoma but no lung carcinoma (type 1 error)
  • FN—False Negative—predict no lung carcinoma but have lung carcinoma (type 2 error)

3.4.2. Loss

It is a function employed to compute distance between the current output of the algorithm and the expected output. It evaluates the performance of an algorithm’s ability to model data. There are two groups, one for classification and another for regression. The commonly used loss functions are cross-entropy, exponential loss, log loss, and hinge loss.

3.4.3. AUC

It refers to area under the receiver operating characteristic curve. It is a performance indicator for how well the positive class probabilities are separated from negative class probabilities. It is scale-invariant. It indicates how well-ranked the predictions are, rather than the absolute values of predictions. Another benefit of employing it is that it is classification-threshold-invariant such that it measures the prediction’s quality irrespective of the threshold chosen.

3.4.4. Precision

Precision is the ratio of relevant outcomes predicted to actually relevant outcomes [23,24]. It is also called positive prediction value. It takes false positives into account.
Precision = T P T P + F P

3.4.5. Recall

Recall gives the ratio of outcomes that were actually relevant to correctly predicted outcomes. It is also called sensitivity. It takes false negatives into account.
Recall = T P T P + F N

3.4.6. F1 Score

F1 Score is another measure that provides the harmonic mean of precision and recall.
F 1   Score = 2 Precision Recall Precision + Recall

4. Experimental Discussion and Analysis

4.1. VGG 16

VGG 16 is a CNN model for visual recognition proposed by Karen Simonyan and Andrew Zisserman of Visual Geometry Group Lab at Oxford University in 2014. VGG 16 won first place in detecting objects within 200 classes and won second place in classifying images labelled 1 of 1000 categories at the ImageNet Large Scale Visual Recognition Challenge in 2014. Applying this model on the ImageNet dataset containing 1000 classes with 14 million images yielded 92.7% top-5 test accuracy.
It takes image input in dimensions (224,224,3). It has a total of sixteen layers. The first two layers are of 3∗3 filter size and have 64 channels with the same padding. Then there is the (2,2) stride max pool layer followed by two 256 filters and (3,3) filter size convolution layers. This is followed by a (2,2) stride max pool layer. Then there are two (3,3) size and 256 filters convolution layers. It is followed by two sets of three 512 filters and (3,3) size convolution layer and a max pool layer all of the same padding. The image is then passed to the two convolution layers stack. Padding of 1-pixel is used to prevent spatial features of the image after each convolution layer.
The (7,7512) feature map is generated once the image passes through all the layers of VGG 16. This feature map is flattened to obtain the (1,25088) feature vector. There are three fully connected layers. The first obtains input from the last feature vector and gives a (1,4096) vector. The second also produces the same output as the first. While the third outputs 1000 channels for 1000 classes. The classification vector is then normalized, bypassing the output form third layer to softmax layer [25,26,27]. The architecture is pictured in Figure 3.
The performance measures for the various epochs over VGG 16 are tabulated and plotted in Table 1 for training and Table 2 for testing data and are shown in Figure 4, Figure 5, Figure 6 and Figure 7. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. By the start of the 5th epoch, the accuracy started to show outstanding value of above 95%. Similarly, AUC shows the ideal curve. Without compromising, precision and recall also show the upper trend.

4.2. VGG 19

A VGG model consisting of 19 layers is VGG 19. It has sixteen convolution layers, three fully connected layers, five maxpool layers, and a softmax layer. VGG 19 is conceptually the same as VGG 16 while the number of layers supported by it is different. Nineteen and sixteen denote the number of layers in the model. This model is trained on millions of images present in the ImageNet database. It can classify pictures into 1000 object categories such as pencil, keyboard, car, and many animals.
The input RGB image is given in the fixed matrix dimension of (224,224,3). It is preprocessed by subtracting RGB mean value from each pixel. Kernels of (3,3) size with 1-pixel stride is used to cover the whole image notion. Spatial resolution of the image is preserved by the usage of spatial padding. The (2,2) max pooling is performed with two-pixel stride. This is followed by rectified linear unit (ReLu) to incorporate non-linearity in the model, which makes the model classify better and reduces computational time. This model proved to be much better than previously used models such as tanh or sigmoid function. It has three completely connected layers, of which the first two are of size 4096. A layer with 1000 channels for 1000 object categories of ILSVRC classification is also present, followed by a final softmax function, as illustrated in Figure 8.
Primarily, it was developed to win ILSVRC but later it was used as classification architecture for image datasets. Its models were made available to the public by the author, so it can be used as it is or with any required modifications for similar purposes. It can be used to perform TL tasks such as facial recognition. The VGG 19 model was able to perform facial recognition with higher accuracy even with mask usage [28].
Performance measures for the various epochs over VGG 19 are tabulated and plotted in Table 3 for training and Table 4 for testing data and are shown in Figure 9, Figure 10, Figure 11 and Figure 12. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. By the start of the 10th epoch, the accuracy started to show outstanding value of above 95%. Similarly, AUC shows the ideal curve. Without compromising, precision and recall also show the upper trend.

4.3. Xception

It is a convolutional neural network model, with Google assisting object detection and image analysis. It is an extreme version of inception. It is based on layers of depth-wise separable convolution, which performs better than inception. It is also hypothesized that mapping of spatial correlations and cross-channel correlations in convolutional neural network feature maps can be completely decoupled. This hypothesis is stronger than the one present in inception.
Xception architecture possesses thirty-six convolution layers giving rise to a base where feature extraction can be performed. The thirty-six layers are separated into fourteen modules with linear residual connections surrounding them. Linear residual connections are exceptions in the first and last module. In short, there are depth-wise separable layers of convolution with residual connections in a linear stack. Hence, it is an easily definable and modifiable model. It can be modified by using just a few lines of code and high-level libraries such as Keras or TensorFlow.
The presence of residual connections helps in convergence; it makes it a better performer in terms of classification and speed. The image first goes into entry flow, followed by middle flow, where the same process is repeated eight times. Finally, it goes through exit flow. All depth-wise separable convolution layers and convolution layers are followed by batch normalization. All layers have a depth multiplier of one, such that there is no expansion of depth. Xception models perform well in convolutional neural networks, and Figure 13 depicts the flow [29].
The performance measures for the various epochs over Xception are tabulated and plotted in Table 5 for training and Table 6 for testing data and are shown in Figure 14, Figure 15, Figure 16 and Figure 17. Gradually loss decreases, indicating the building of a good model; other performance measures also indicate the building of a good model and other performance measures show the increasing trend. From the beginning itself, accuracy, AUC, precision, and recall are good, and as mentioned, this model performs well in the initial stage itself. Because of the similar values, the Figure 17 graph merges precision vs. recall together.

5. Conclusions and Future Work

On comparison with the recent works carried out using this same dataset, our model yields good results. With the same dataset work carried out using SVM classification, the accuracy achieved is 89.8876% [30]. In work carried out with the AlexNet architecture, accuracy is 93.548% [31].
All TL approaches perform well in this dataset and compare VGG 16, VGG 19, and Xception for the 20-epoch structure. For any model, preprocessing acts as the great bridge to construct a reliable model and eventually helps predict a reliable model and eventually helps in predicting future cases by incorporating the interface at the faster phase. Cumulatively, accuracy of VGG 16, VGG 19, and Xception is 98.83%, 98.05%, and 97.4%, respectively, at the 20th epoch. Loss is somewhat high in the case of Xception over testing data, and accuracy of VGG 16, VGG 19, and Xception is 83.39%, 80.97%, and 89.68%, respectively, at the 20th epoch. In comparison, VGG16 and VGG19 outperform.
Utilization of other clinical data along with this will provide even more clarity. Additional demographic information helps in taking additional steps to create awareness in more prone locations. Normally, lifestyle factors such as eating habits, cultural habits, and social living have a great impact on many diseases and cumulating the same will aid greatly in decision-making.

Author Contributions

Conceptualization, M.H. and R.S.; methodology, R.S.; software, S.N.A. and M.H.; validation, R.S., N.Z.J. and S.N.A.; formal analysis, R.S. and M.H.; investigation, R.S. and M.H.; resources, M.H.; data curation, M.H. and R.S.; writing—original draft preparation, M.H. and R.S.; writing—review and editing, R.S., M.H. and N.Z.J.; visualization, M.H.; supervision, N.Z.J.; project administration, N.Z.J.; funding acquisition, S.N.A. and M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Deanship of Scientific Research at Jouf University under grant no. (DSR-2021-02-0330).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be furnished on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nall, R. What to Know about Lung Cancer. Medical News Today. 2018. Available online: https://www.medicalnewstoday.com/articles/323701 (accessed on 2 April 2022).
  2. Saeed, S.; Jhanjhi, N.Z.; Naqvi, M.; Humyun, M.; Ahmad, M.; Gaur, L. Optimized Breast Cancer Premature Detection Method with Computational Segmentation: A Systematic Review Mapping. In Approaches and Applications of Deep Learning in Virtual Medical Care; IGI Global: Harrisburg, PA, USA, 2022; pp. 24–51. [Google Scholar]
  3. Facts About Lung Cancer. Available online: https://www.lungcancerresearchfoundation.org/lung-cancer-facts/ (accessed on 27 March 2022).
  4. Shukla, D. New Test May Quickly Detect Early-Stage Lung Cancer. Medical News Today. 2022. Available online: https://www.medicalnewstoday.com/articles/new-test-may-quickly-detect-early-stage-lung-cancer (accessed on 28 March 2022).
  5. Zhou, Y.; Lu, Y.; Pei, Z. Accurate diagnosis of early lung cancer based on the convolutional neural network model of the embedded medical system. Microprocess. Microsyst. 2021, 81, 103754. [Google Scholar] [CrossRef]
  6. Nanglia, P.; Kumar, S.; Mahajan, A.N.; Singh, P.; Rathee, D. A hybrid algorithm for lung cancer classification using SVM and Neural Networks. ICT Express 2021, 7, 335–341. [Google Scholar] [CrossRef]
  7. Bai, Y.; Li, D.; Duan, Q.; Chen, X. Analysis of high-resolution reconstruction of medical images based on deep convolutional neural networks in lung cancer diagnostics. Comput. Methods Programs Biomed. 2022, 217, 106592. [Google Scholar] [CrossRef]
  8. Ahmad, A.S.; Mayya, A.M. A new tool to predict lung cancer based on risk factors. Heliyon 2020, 6, e03402. [Google Scholar] [CrossRef]
  9. Ahn, B.-C.; So, J.-W.; Synn, C.-B.; Kim, T.H.; Kim, J.H.; Byeon, Y.; Kim, Y.S.; Heo, S.G.; Yang, S.D.; Yun, M.R.; et al. Clinical decision support algorithm based on machine learning to assess the clinical response to anti–programmed death-1 therapy in patients with non–small-cell lung cancer. Eur. J. Cancer 2021, 153, 179–189. [Google Scholar] [CrossRef]
  10. Agarwal, S.; Yadav, A.S.; Dinesh, V.; Vatsav, K.S.S.; Prakash, K.S.S.; Jaiswal, S. By artificial intelligence algorithms and machine learning models to diagnosis cancer. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  11. Huang, X.; Lei, Q.; Xie, T.; Zhang, Y.; Hu, Z.; Zhou, Q. Deep transfer convolutional neural network and extreme learning machine for lung nodule diagnosis on CT images. Knowl.-Based Syst. 2020, 204, 106230. [Google Scholar] [CrossRef]
  12. Doppalapudi, S.; Qiu, R.G.; Badr, Y. Lung cancer survival period prediction and understanding: Deep learning approaches. Int. J. Med. Inform. 2021, 148, 104371. [Google Scholar] [CrossRef]
  13. Ashraf, S.F.; Yin, K.; Meng, C.X.; Wang, Q.; Wang, Q.; Pu, J.; Dhupar, R. Predicting benign, preinvasive, and invasive lung nodules on computed tomography scans using machine learning. J. Thorac. Cardiovasc. Surg. 2022, 163, 1496–1505. [Google Scholar] [CrossRef]
  14. Shakeel, P.M.; Burhanuddin, M.A.; Desa, M.I. Lung cancer detection from CT image using improved profuse clustering and deep learning instantaneously trained neural networks. Measurement 2019, 145, 702–712. [Google Scholar] [CrossRef]
  15. ALzubi, J.A.; Bharathikannan, B.; Tanwar, S.; Manikandan, R.; Khanna, A.; Thaventhiran, C. Boosted neural network ensemble classification for lung cancer disease diagnosis. Appl. Soft Comput. 2019, 80, 579–591. [Google Scholar] [CrossRef]
  16. Alyasriy, H. The IQ-OTHNCCD Lung Cancer Dataset. Mendeley Data, Version 1. 2020. Available online: https://data.mendeley.com/datasets/bhmdr45bh2/1 (accessed on 28 March 2022).
  17. Szegedy, C.; Toshev, A.; Erhan, D. Deep neural networks for object detection. In Advances in Neural Information Processing Systems, Proceedings of the 26th International Conference on Neural Information Processing Systems, Nevada City, CA, USA, 5–10 December 2013; Morgan Kaufmann Publishers: San Francisco, CA, USA, 2013; Volume 26. [Google Scholar]
  18. Yu, X.; Wang, J.; Hong, Q.Q.; Teku, R.; Wang, S.H.; Zhang, Y.D. Transfer learning for medical images analyses: A survey. Neurocomputing 2022, 489, 230–254. [Google Scholar] [CrossRef]
  19. Hand, D.J.; Till, R.J. A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  20. Ferri, C.; Hernández-Orallo, J.; Modroiu, R. An experimental comparison of performance measures for classification. Pattern Recognit. Lett. 2009, 30, 27–38. [Google Scholar] [CrossRef]
  21. Humayun, M.; Alsayat, A. Prediction Model for Coronavirus Pandemic Using Deep Learning. Comput. Syst. Sci. Eng. 2022, 40, 947–961. [Google Scholar] [CrossRef]
  22. Khalil, M.I.; Humayun, M.; Jhanjhi, N.; Talib, M.; Tabbakh, T.A. Multi-class Segmentation of Organ at Risk from Abdominal CT Images: A Deep Learning Approach. In Intelligent Computing and Innovation on Data Science; Peng, S.L., Hsieh, S.Y., Gopalakrishnan, S., Duraisamy, B., Eds.; Springer: Berlin, Germany, 2021; pp. 425–434. [Google Scholar]
  23. Gouda, W.; Almurafeh, M.; Humayun, M.; Jhanjhi, N.Z. Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare 2022, 10, 343. [Google Scholar] [CrossRef]
  24. Khalil, M.I.; Tehsin, S.; Humayun, M.; Jhanjhi, N.; AlZain, M.A. Multi-Scale Network for Thoracic Organs Segmentation. CMC-Comput. Mater. Contin. 2022, 70, 3251–3265. [Google Scholar] [CrossRef]
  25. Cheng, S.; Zhou, G. Facial expression recognition method based on improved VGG convolutional neural network. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2056003. [Google Scholar] [CrossRef]
  26. Tammina, S. Transfer learning using VGG-16 with deep convolutional neural network for classifying images. Int. J. Sci. Res. Publ. IJSRP 2019, 9, 143–150. [Google Scholar] [CrossRef]
  27. VGG-16: CNN Model. Available online: https://www.geeksforgeeks.org/vgg-16-cnn-model/ (accessed on 27 March 2022).
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  30. Kareem, H.F.; AL-Husieny, M.S.; Mohsen, F.Y.; Khalil, E.A.; Hassan, Z.S. Evaluation of SVM Performance in the Detection of Lung Cancer in Marked CT Scan Dataset. Indones. J. Electr. Eng. Comput. Sci. 2021, 21, 1731–1738. [Google Scholar] [CrossRef]
  31. Amiripallia, S.S.; Rao, G.N.; Beharaa, J.; Sanjay, K. Mineral Rock Classification Using Convolutional Neural Network. In Recent Trends in Intensive Computing; IOS Press: Amsterdam, The Netherlands, 2021. [Google Scholar]
Figure 1. Cancer research.
Figure 1. Cancer research.
Healthcare 10 01058 g001
Figure 2. TL-based lung carcinoma classification.
Figure 2. TL-based lung carcinoma classification.
Healthcare 10 01058 g002
Figure 3. VGG 16 architecture.
Figure 3. VGG 16 architecture.
Healthcare 10 01058 g003
Figure 4. Loss vs. Epoch.
Figure 4. Loss vs. Epoch.
Healthcare 10 01058 g004
Figure 5. Accuracy vs. Epoch.
Figure 5. Accuracy vs. Epoch.
Healthcare 10 01058 g005
Figure 6. AUC vs. Epoch.
Figure 6. AUC vs. Epoch.
Healthcare 10 01058 g006
Figure 7. Precision vs. Recall.
Figure 7. Precision vs. Recall.
Healthcare 10 01058 g007
Figure 8. VGG 19 architecture.
Figure 8. VGG 19 architecture.
Healthcare 10 01058 g008
Figure 9. Loss vs. Epoch.
Figure 9. Loss vs. Epoch.
Healthcare 10 01058 g009
Figure 10. Accuracy vs. Epoch.
Figure 10. Accuracy vs. Epoch.
Healthcare 10 01058 g010
Figure 11. AUC vs. Epoch.
Figure 11. AUC vs. Epoch.
Healthcare 10 01058 g011
Figure 12. Precision vs. Recall.
Figure 12. Precision vs. Recall.
Healthcare 10 01058 g012
Figure 13. Xception architecture.
Figure 13. Xception architecture.
Healthcare 10 01058 g013
Figure 14. Loss vs. Epoch.
Figure 14. Loss vs. Epoch.
Healthcare 10 01058 g014
Figure 15. Accuracy vs. Epoch.
Figure 15. Accuracy vs. Epoch.
Healthcare 10 01058 g015
Figure 16. AUC vs. Epoch.
Figure 16. AUC vs. Epoch.
Healthcare 10 01058 g016
Figure 17. Precision vs. Recall.
Figure 17. Precision vs. Recall.
Healthcare 10 01058 g017
Table 1. Performance measures over VGG 16 for training data.
Table 1. Performance measures over VGG 16 for training data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
11.14750.66410.83330.67110.65620.6636
50.15220.95310.9960.96010.94010.9500
100.08260.98310.99930.98310.98310.9831
150.06250.98440.99950.98560.98310.9843
200.05080.98830.99940.98830.98570.9870
Table 2. Performance measures over VGG 16 for testing data.
Table 2. Performance measures over VGG 16 for testing data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
10.49630.86450.94090.8820.84350.8623
50.46260.82740.94540.83940.81770.8284
100.48510.86450.94720.86950.85970.8646
150.62480.81450.9230.81390.81130.8126
200.59970.83390.9390.83630.83230.8343
Table 3. Performance measures over VGG 19 for training data.
Table 3. Performance measures over VGG 19 for training data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
11.06430.6810.84860.70930.66410.6860
50.28210.89840.97830.9130.8880.9003
100.1150.96480.99780.97370.96350.9686
150.07850.98570.99920.98570.98570.9857
200.06580.98050.99920.98040.97660.9785
Table 4. Performance measures over VGG 19 for testing data.
Table 4. Performance measures over VGG 19 for testing data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
11.54130.30320.59080.2890.26610.2771
50.45450.89030.94940.89880.87420.8863
100.56330.81290.92960.8230.80970.8163
150.60810.79680.93190.7980.78390.7909
200.65240.80970.92620.8110.80970.8103
Table 5. Performance measures over Xception for training data.
Table 5. Performance measures over Xception for training data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
10.42470.95830.98050.95830.95830.9583
50.14180.97920.99280.97920.97920.9792
100.20220.97530.98970.97530.97530.9753
150.12180.97790.99460.97790.97790.9779
200.12380.9740.99270.9740.9740.974
Table 6. Performance measures over Xception for training data.
Table 6. Performance measures over Xception for training data.
EpochsLossAccuracyAUCPrecisionRecallF1 Score
13.55710.79350.87090.79350.79350.7935
52.75820.85970.92360.85970.85970.8597
104.07570.81290.88930.81290.81290.8129
153.26050.86450.92070.86430.86430.8643
203.56820.89680.93380.89680.89680.8968
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Humayun, M.; Sujatha, R.; Almuayqil, S.N.; Jhanjhi, N.Z. A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma. Healthcare 2022, 10, 1058. https://doi.org/10.3390/healthcare10061058

AMA Style

Humayun M, Sujatha R, Almuayqil SN, Jhanjhi NZ. A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma. Healthcare. 2022; 10(6):1058. https://doi.org/10.3390/healthcare10061058

Chicago/Turabian Style

Humayun, Mamoona, R. Sujatha, Saleh Naif Almuayqil, and N. Z. Jhanjhi. 2022. "A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma" Healthcare 10, no. 6: 1058. https://doi.org/10.3390/healthcare10061058

APA Style

Humayun, M., Sujatha, R., Almuayqil, S. N., & Jhanjhi, N. Z. (2022). A Transfer Learning Approach with a Convolutional Neural Network for the Classification of Lung Carcinoma. Healthcare, 10(6), 1058. https://doi.org/10.3390/healthcare10061058

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop