Next Article in Journal
Neuromodulation Induced by Sitagliptin: A New Strategy for Treating Diabetic Retinopathy
Next Article in Special Issue
Displacement of Gray Matter and Incidence of Seizures in Patients with Cerebral Cavernous Malformations
Previous Article in Journal
Reduction of Rapid Proliferating Tumour Cell Lines by Inhibition of the Specific Glycine Transporter GLYT1
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Deep Convolutional Neural Networks for Enhanced Ultrasonographic Image Diagnosis of Differentiated Thyroid Cancer

1
Division of Endocrinology and Metabolism, Department of Internal Medicine, Chang Gung Memorial Hospital, College of Medicine, Chang Gung University, Taoyuan 33302, Taiwan
2
Professional Master Program in Artificial Intelligence in Medicine, College of Medicine, Taipei Medical University, Taipei 10675, Taiwan
*
Author to whom correspondence should be addressed.
Biomedicines 2021, 9(12), 1771; https://doi.org/10.3390/biomedicines9121771
Submission received: 18 October 2021 / Revised: 17 November 2021 / Accepted: 22 November 2021 / Published: 26 November 2021
(This article belongs to the Special Issue Artificial Intelligence in Biological and Biomedical Imaging)

Abstract

:
Differentiated thyroid cancer (DTC) from follicular epithelial cells is the most common form of thyroid cancer. Beyond the common papillary thyroid carcinoma (PTC), there are a number of rare but difficult-to-diagnose pathological classifications, such as follicular thyroid carcinoma (FTC). We employed deep convolutional neural networks (CNNs) to facilitate the clinical diagnosis of differentiated thyroid cancers. An image dataset with thyroid ultrasound images of 421 DTCs and 391 benign patients was collected. Three CNNs (InceptionV3, ResNet101, and VGG19) were retrained and tested after undergoing transfer learning to classify malignant and benign thyroid tumors. The enrolled cases were classified as PTC, FTC, follicular variant of PTC (FVPTC), Hürthle cell carcinoma (HCC), or benign. The accuracy of the CNNs was as follows: InceptionV3 (76.5%), ResNet101 (77.6%), and VGG19 (76.1%). The sensitivity was as follows: InceptionV3 (83.7%), ResNet101 (72.5%), and VGG19 (66.2%). The specificity was as follows: InceptionV3 (83.7%), ResNet101 (81.4%), and VGG19 (76.9%). The area under the curve was as follows: Incep-tionV3 (0.82), ResNet101 (0.83), and VGG19 (0.83). A comparison between performance of physicians and CNNs was assessed and showed significantly better outcomes in the latter. Our results demonstrate that retrained deep CNNs can enhance diagnostic accuracy in most DTCs, including follicular cancers.

1. Introduction

Most thyroid tumors are incidentally discovered via palpation by clinical physicians. It has been estimated that the prevalence of thyroid cancer can reach 65% [1] and is more common among females. Fortunately, most tumors are benign thyroid nodules; that is, only a small number of them are malignant [2]. Roughly 5–10% of these tumors are identified as thyroid cancer. In Taiwan, thyroid cancer is becoming increasingly common with most cases identified in individuals between 40 and 65 years old, and it is currently the fourth most prevalent form of cancer among women, as well as the most common cancer of the endocrine system. The Health Promotion Administration of Taiwan has reported a 9.67% annual increase in the number of newly diagnosed cases of thyroid cancer. This may be due in part to advances in ultrasound and imaging technology over the past decade, which have greatly facilitated diagnostic procedures [3], particularly when dealing with tumors measuring less than 1 cm. The prognosis in cases of thyroid cancer is generally good. At present, the long-term prognosis after standard treatment for differentiated thyroid cancers (DTC) is excellent, with a 10 year survival rate of 96% [4].
Risk factors of thyroid cancer include exposure to radiation in one’s youth, a history of thyroid goiter, and a family history of cancer [5]. The histopathology of thyroid cancers can be classified according to the state of follicular epithelial cells as DTC, poorly differentiated thyroid carcinoma, and anaplastic thyroid carcinoma (ATC) [6], as well as medullary thyroid cancer (MTC) derived from C cells [7]. DTC is the most common form of thyroid cancer, which appears as papillary thyroid carcinoma (PTC) in 90–92% of cases and the follicular variant of papillary thyroid carcinoma (FVPTC; the most common subtype of PTC). Rare forms include follicular thyroid carcinoma (FTC) and Hürthle cell carcinoma (HCC), both of which have proven difficult to diagnose [8]. At present, the prognosis for rare variants other than differentiated thyroid cancer is very poor, due to a lack of effective treatment options [6].
Currently, the gold standard for the clinical diagnosis of thyroid cancers is ultrasound-guided fine-needle aspiration or core-needle biopsy combining cytological and pathological analysis [9]. Unfortunately, this method is applicable only to PTC. Identifying other types of DTC (FVPTC, FTC, and HTC) is hampered by a lack of distinguishing cytological characteristics in ultrasound images (Figure 1), such as hypo-echogenicity, irregular margins, or microcalcifications [10]. In many cases, surgical resection of the tumor is necessary to confirm DTC subtypes, such as FTC and HCC. Molecular biology and genetic analysis can be used to facilitate diagnosis [11]; however, the tools and expertise required for such analysis are generally available only in medical centers. Note that the invasive nature of fine-needle aspiration and core-needle biopsy inevitably leads to complications, such as bleeding or infection. Furthermore, the effectiveness of the procedures depends largely on the experience and skill of the operator [12].
Artificial intelligence (AI) is a key technology in the on-going personalization and development of precision medicine. Musko [13] claimed that artificial intelligence (AI) allows doctors and researchers to make predictions of greater accuracy, thereby making it easier to identify the treatment and prevention strategies best suited to a particular disease and/or groups of patients. Deep learning algorithms are increasingly being used to facilitate the diagnosis of tumors. Chi [14] reported that the retrained GoogleNet outperformed conventional machine learning approaches, such as support vector machine (SVM). Using InceptionV3, Song [15] achieved diagnostic performance comparable to that of experienced professional radiologists. Various deep learning models have also been trained to differentiate between malignant and benign thyroid tumors [16,17,18,19,20,21,22,23]. A number of studies have addressed the issue of training AI systems in the analysis of thyroid ultrasound images; however, most of this work has focused on PTC. Diagnosing other pathological types (FVPTC, FTC, and HCC) is hindered by their rarity in clinical practice and their similarity to benign lesions in ultrasound images. The prognosis for DTCs should be similar to that of PTC as long as they are identified early. Ideally, clinicians should be able to confirm diagnosis prior to surgical intervention.
In this study, transfer learning was used to train a deep convolutional neural network (CNN) [24] for the analysis of ultrasound images with the aim of differentiating between malignant and benign thyroid lesions, and facilitating the identification of other DTCs (e.g., FTC). We anticipate that such CNN models could help to eliminate unnecessary invasive examinations or surgical interventions.

2. Materials and Methods

2.1. Data Sources

This retrospective study was based on medical records and clinical data from the cancer registry of Chang Gung Memorial Hospital (Linko branch) covering the period from January 2008 to July 2020. Patients aged >20 years with surgically confirmed diagnosis of thyroid cancer were enrolled. Furthermore, patients with surgically confirmed benign thyroid nodules between January 2016 and July 2020 were also enrolled. Patients with non-DTC were excluded (Figure 2). Patients who had not undergone an ultrasound examination within 12 months prior to surgical intervention were also excluded, as were patients without recognizable cancer lesions in ultrasound images. This study was approved by the Institutional Review Board of the Chang Gung Medical Foundation (IRB No. 202001440B0, 31 August 2020). The requirement for informed consent was waived due to the retrospective nature of this analysis.

2.2. Data Collection

Demographic and clinical data included the age of the patient at the time of diagnosis, gender, lesion location (left, right, both, or isthmus), ultrasound manufacturer (e.g., Aloka, Hitachi, and Siemens) (Table 1), the distribution of pathological groups as a function of ultrasound brand (Table 2), and histopathological data (Table 3). All images were downloaded and stored in TIFF format. Every patient included in the study presented at least one thyroid tumor in ultrasound images (longitudinal or horizontal view) when assessed using the models of multiple ultrasound manufacturers. An ultrasound image may be formed by a nodule in two views as long as they were saved in double-view mode. After a manual review of the examination data, researchers collected 1791 ultrasound images for analysis. Note that the regions of interest (ROIs) in the ultrasound images were marked by the author as rectangle bounding boxes (Figure 3). The ROI was meant to include the entire tumor except in cases where the tumor exceeded the image boundary, such that the bounding box included only the visible part of the tumor. Ultrasound images were subsequently cropped according to the bounding boxes, which resulted in 2308 images of nodules for training (Figure 4). The images were divided into a training set (80%) and a test set (20%). Images from each patient were placed in either the training set or the test set, but not in both. Image data underwent preprocessing to compensate for the relatively small number of images and reduce the likelihood of overfitting. As shown in Figure 5, data augmentation based on histogram equalization/normalization and horizontal flipping increased the number of images by four times, as follows: original image, image with histogram equalization, image with horizontal flipping, and images with histogram equalization and horizontal flipping. The training set included 3316 images showing malignant tumors and 4044 images showing benign tumors. All images showing Hürthle cell adenoma (HA) were included in the training set. The test set (used to assess diagnostic performance) included 204 images showing malignant tumors and 264 images showing benign tumors.

2.3. Study Design

The diagnoses of all tumors in this study were subject to surgical and pathological confirmation; therefore, training was implemented as supervised learning. Transfer learning and fine-tuning of hyperparameters were implemented on three pretrained CNNs, namely, InceptionV3, ResNet101, and VGG19. Note that the classification accuracy of these CNNs has been demonstrated in the ImageNet Large-Scale Visual Recognition Challenge (ILSVRC). The MATLAB 2021a platform was used for the retraining of the three CNNs to classify benign and malignant thyroid tumors in ultrasound images. The size of the input images was adjusted according to the CNN settings. Stochastic gradient descent with momentum (SGDM) was applied as the solver. The maximum epochs were as follows: InceptionV3 (26), ResNet101 (21), and VGG19 (32). The learning rate was as follows: InceptionV3 (0.001), ResNet101 (0.001), and VGG19 (0.0001). Fivefold cross-validation was used to ensure the stability of the results.

2.4. Statistical Analysis

This study compared the diagnostic capability of CNNs with that of two endocrinologists with over 20 years of experience in performing fine-needle aspiration and the interpretation of ultrasound images on the test set. In estimating the diagnosis performance of physicians, the images were classified as malignant and benign according to sonographic patterns and estimated risk of malignancy, as suggested in the American Thyroid Association (ATA) classification system [25]. The CNNs were assessed in terms of accuracy, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV), as well as the receiver operating characteristic (ROC) curve, area under the curve (AUC), and confusion matrix. We also assessed accuracy in identifying tumors with various histopathologies. Continuous variables were presented as the mean and standard deviation (SD), as indicated. Categorical data were expressed in terms of actual frequencies and percentages. Statistical analysis was performed using the chi-square test and analysis of variance (ANOVA). p-Values < 0.05 were considered significant. All statistical analysis was conducted using the SAS Suite, version 9.4 (SAS Institute, Cary, NC, USA).

3. Results

3.1. Study Population

A total of 791 patients were identified in our initial analysis (Figure 1). From this group, 17 patients were excluded due to non-DTC, including anaplastic cancer (n = 6), medullary cancer (n = 7), and metastatic cancer (n = 4). Patients who had not undergone ultrasound examinations within 12 months prior to surgery were also excluded, as were those without recognizable lesions (n = 353). This left 421 DTC patients and 391 patients with benign thyroid nodules who met the enrollment criteria for this study.

3.2. Demographics

As shown in Table 1, the patients were divided into a malignant group (comprising a PTC group, FTC group, FVPTC group, and HCC group) and a benign group. The mean age of patients was 44.9–54.2 years old. The average age at the time of diagnosis was lower in the malignant groups (p < 0.0001). We observed a higher proportion of females in all groups; however, the female-to-male ratio between groups did not differ significantly. We observed statistically significant between-group differences in terms of lesion location (p = 0.0017) with very few instances of bilateral lesion. In malignant groups, the PTC group presented the highest proportion of simultaneous bilateral lesions (5.14%), whereas the FTC group presented the highest proportion of isthmus lesions (4.29%). We observed statistically significant between-group differences in terms of ultrasound manufacturer (p < 0.0001). The most common brands in the PTC group were GE Healthcare (37.12%) and Siemens (28.82%), and the most common brand in the other groups was GE Healthcare. Table 2 lists the distribution (percentages) of pathological groups as a function of ultrasound brand. Statistically significant differences were observed between all pathological groups as a function of ultrasound brand (p < 0.001).
Table 3 lists the histopathological distribution of tumors among groups. On the basis of ultrasonic features, the malignant group was divided into PTC and FTC subgroups. The PTC subgroup included classic PTC, the diffuse sclerosing variant, the tall cell variant, the cribriform morular variant, and the encapsulated variant. The FTC sub-group included FTC, FVPTC, HCC, and the encapsulated follicular variant of PTC. The benign group included nodular hyperplasia (NH), follicular adenoma (FA), cysts, and HA. Classic PTC and FVPTC were the most common pathology types in the PTC and FTC subgroups, respectively. Nodular hyperplasia was the most common feature in the benign group.

3.3. Performance Assessment on CNNs and Physician

The training set contained a total of 7360 nodule images after data augmentation, including 1744 in the PTC group, 852 in the FVPTC group, 568 in the FTC group, 152 in the HCC group, and 4044 in the benign group. Following the completion of transfer learning, the test set was used to assess the performance of the CNNs and obtain a confusion matrix, as shown in Figure 6. Table 4 presents the accuracy of the CNNs and physicians in terms of diagnostic performance, as follows: InceptionV3 (76.5%), ResNet101 (77.6%), VGG19 (76.1%), Endocrinologist 1 (58.8%), and Endocrinologist 2 (62%). The sensitivity was as follows: InceptionV3 (83.7%), ResNet101 (72.5%), VGG19 (66.2%), Endocrinologist 1 (38.7%), and Endocrinologist 2 (35.3%). The specificity was as follows: InceptionV3 (83.7%), ResNet101 (81.4%), VGG19 (76.9%), Endocrinologist 1 (72.4%), and Endocrinologist 2 (82.6%). A confusion matrix illustrating multiclass classification using ResNet101 in the test set is shown in Figure S1. Due to the small number of cases in the malignant groups (e.g., FVPTC, FTC, and HCC), the accuracy of the CNN was only 65%.
Table 5 lists the accuracy of the CNNs and physicians in identifying tumors with various pathological types. In the identification of malignant tumors, the highest accuracy in diagnosing PTC was 81.4% (InceptionV3), the highest accuracy in diagnosing FVPTC was 74.6% (ResNet101), the highest accuracy in diagnosing FTC was 72.7% (InceptionV3), and the highest accuracy in diagnosing HCC was 66.7% (InceptionV3 and ResNet101). In identifying benign tumors, the highest accuracy in diagnosing NH was 82.4% (VGG19), the highest accuracy in diagnosing FA was 80% (ResNet101), and the highest accuracy in diagnosing cysts was 95% (VGG19). In terms of the performance of physician diagnosis, Endocrinologist 1 showed better accuracy in diagnosing PTC (58.8% vs. 53.6%), FVPTC (20.3% vs. 17%), HCC (13.3% vs. 6.7%), and FA (80% vs. 75%), while Endocrinologist 2 showed higher accuracy in FTC (30.3% vs. 27.3%), NH (81.9% vs. 73%), and cyst (90% vs. 80%). Figure 7 presents the ROC curve of CNNs and performance of physicians. As shown in Table 3, the area under the curve (AUC) was as follows: InceptionV3 (0.82), ResNet101 (0.83), and VGG19 (0.83).

4. Discussion

In this study, deep convolutional neural networks (CNNs) were used to classify thyroid tumors as malignant or benign. Note that the accuracy achieved in the current study was slightly lower than in previous studies [15,16,17,21,22]. To the best of our knowledge, this was the first study focusing on the use of CNNs for the classification of DTCs other than PTC (FVPTC, FTC, and HCC). Diagnosing malignant thyroid tumors (e.g., FTC) prior to surgical intervention remains an unresolved problem [26]. A definitive diagnosis of FTC requires surgical intervention. This issue is largely due to similarities between the ultrasonic features of malignant and benign nodules [27]. From a microscopic point of view, the major difference between malignant FTC and benign follicular tumors (FAs) is the occurrence of vascular or capsular invasion. Furthermore, cytopathological analysis is often inconclusive due to a lack of distinguishing characteristics, such as the “clear cell border” and “pseudo-inclusion body” observed in PTC cells. Note that most of these cases would eventually be classified as follicular neoplasms (FNs) [28]. Technical advances in molecular biology and genetic engineering have revealed a link between BRAF mutations and PTC, as well as a link between RAS mutations and FTC [11]. Unfortunately, the cost of molecular and genetic testing is prohibitive in most cases and unavailable except in the best-equipped medical centers. Due to the low incidence of malignant thyroid tumor, the inclusion of advanced diagnostics in routine thyroid examinations is unreasonable. More importantly, diagnostic accuracy is strongly influenced by the number of successful punctures in fine-needle aspiration.
In one recent study, machine learning methods proved highly effective in diagnosing FNs [29]. In fact, a thyroid CAD named AmCAD-UT has already been approved by the United States Food and Drug Administration (FDA) and Taiwan Medical Device Marketing Approval for the assessment of thyroid tumors using feature extraction/selection. AI is proving highly effective in overcoming the difficulties associated with the diagnosis of malignant tumors.
The ImageNet project has been instrumental in advancing computer vision and deep learning research. ImageNet provides an image database based on the WordNet hierarchy, and data are freely available to researchers for noncommercial applications [30]. The database has been manually annotated with more than 14 million images. Between 2010 and 2017, ImageNet held an annual competition (referred to as ILSVRC) to evaluate algorithms used in object detection and image classification. The CNNs used in this study for transfer learning achieved the highest classification accuracy in 2014 (Inception), the highest detection results in 2014 (VGG), and the highest classification accuracy in 2015 (ResNet). Since 2015, the accuracy of deep learning image classification has exceeded 95%, far exceeding the capabilities of humans. In the intervening years, new CNNs (e.g., SEnet) have achieved even higher accuracy; however, the difference is negligible. Most of the previous thyroid cancer imaging studies using Inception, ResNet, and VGG achieved acceptable accuracy and were, therefore, deemed suitable for transfer learning in the current study. We discovered that the less complex CNNs (Inception and ResNet) were slightly faster than VGG in terms of training and classification; however, overall classification accuracy was nearly identical.
In identifying cases of PTC, we were unable to achieve the 90% accuracy observed in previous studies [17,22], due primarily to an insufficient number of cases. Note that most of the previous CNN studies on classic PTC included at least 1000 patients. According to the results in previous studies, it appears likely that accuracy in identifying PTC could be increased simply by including a larger number of cases. It is also possible that the performance of the CNN algorithms was hindered by unacceptably low image resolution after cropping. There is also a possibility that the apparatus used for ultrasonic imaging played a role in algorithm performance, due to subtle differences in image fineness, brightness, contrast, and texture output from different ultrasound manufacturers.
In our analysis, accuracy in identifying FVPTC reached 74.6% (ResNet101), which is similar to the results obtained for classic PTC. FVPTC is the most common PTC variant and the second largest group in the current study. The ultrasonic characteristics of FVPTC differ considerably from those of classic PTC and, in many respects, are similar to those of benign tumors [31]. Our results demonstrated that accuracy in identifying FTC was only 63.6–72.7% and accuracy in identifying FA was only 65–80%, regardless of the CNN. Accuracy in identifying HCC was only 60–66.7%, due largely to the small number of cases in the database. The CNNs seemed not to provide much benefit in the identification or diagnosis of FTC or HCC, due largely to a lack of cases resulting from low incidence and prevalence. Note that there is only a slight difference between FTC, HCC, and FA in terms of gross structure and ultrasound features [32,33].
The performance of classification by retrained CNNs was a lot better than that of the participating physicians, especially in the malignant groups. The poor diagnostic performance of physicians in dealing with malignant tumors resulted in poor sensitivity. In clinical practice, the endocrinologist or radiologist usually considers malignant features suggested by the image as a whole, not just the cropped area surrounding the tumor with low resolution. However, with the help of fine-needle aspiration and cytopathological analysis, the sensitivity of physicians may be comparable to CNNs retrained by ultrasound images alone. Remarkably low accuracy in the classification of malignant tumors by physicians also indicates the difficulty in clinical diagnosis, particularly in cases of FTC, FVPTC, and HCC. Overall, it was demonstrated that the diagnostic performance of the CNNs exceeded that of the physicians.
Overall, InceptionV3 achieved the highest sensitivity, whereas ResNet101 and VGG19 achieved higher specificity. The concurrent application of all three CNNs appears to be a viable possibility. InceptionV3 could be used to confirm a diagnosis of malignancy, whereas ResNet101 and VGG19 could be used to confirm that lesions are indeed benign.
There are numerous situations in which CAD could advantageously be implemented in conjunction with AI. For example, many developing countries lack the medical resources, professional radiologists, and endocrinologists required to obtain a reliable diagnosis of thyroid lesions. CAD could be used to screen for potential thyroid cancers for referral to a medical center. Even in medical centers, CAD could be used to facilitate the training of medical students and inexperienced physicians. More importantly, CAD could provide helpful advice in dilemmatic cases with inconclusive cytology results. From the perspective of healthcare and therapeutics, AI has also been shown to play an important role in treatment quality. Fionda [34] reported that the use of AI-based predictive models and decision support systems for radiation oncology and interventional radiotherapy can alleviate many time-consuming repetitive tasks, thereby enabling a corresponding decrease in healthcare costs.
In the current study, image ROIs were manually cropped by the author using a bounding box. This method is precise but time-consuming, and different physicians would no doubt differ in their approach to cropping. Deep learning models with auto-detection or auto-segmentation (e.g., YOLOV3 and R-CNN) could be developed to increase the speed of ROI framing before undergoing a manual review and adjustment. In the clinical application of CAD, it is also necessary to establish a graphical user-interface (GUI) capable of automating the process of ROI framing and classification.
This study was subject to a number of limitations. Firstly, the retrospective design of this study made selection bias inevitable. Secondly, an extended acquisition period was required to obtain a usable number of samples in the malignant groups. Thus, it was inevitable that the collection period for malignant cases would far exceed that of the benign control group. Thirdly, the small number of cases in our test set may have also had a negative effect on accuracy. Fourthly, the process of image selection was complicated and slow. Unlike computed tomography and magnetic resonance imaging, ultrasound images do not present a unified arrangement and require extensive manual preprocessing. In this study, the author had to review all of the ultrasound images in a search for the target lesions identified surgically. Ultrasound images also tend to vary considerably in terms of size and zooming ratio. This made it impossible to measure the tumor sizes retrospectively, leading to instances of missing data and/or mismatch with pathology reports. Note that this was the reason for the omission of tumor size in this study. Fifthly, we opted not to include the rarer forms of malignant tumor, such as undifferentiated thyroid cancers and metastatic cancers. Note that applying the current CAD in clinical practice would no doubt raise concerns about missed diagnoses. Sixthly, most of the images selected for CNN training presented identifiable single nodules. Thus, the diagnostic power in dealing with multinodular goiters with ill-defined margins remains unclear. Lastly, differences in the output algorithms of ultrasound machines can have a profound effect on training and classification. However, without raw data, there is no simple way to standardize images. Thus, the only viable approach to balancing the datasets is to apply histogram equalization or collect more images from different ultrasound manufacturers.

5. Conclusions

Advanced deep CNN models that are fine-tuned using transfer learning show considerable potential as a noninvasive approach to the diagnosis of DTCs, including FTC. Clinicians should be able to diagnose thyroid cancer more easily by combining ultrasound with CAD. Anticipated advances in ultrasound technology and larger databases will greatly enhance the efficacy of these methods.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/biomedicines9121771/s1, Figure S1: Confusion matrix of multi-class classification using ResNet101 in test set. B: Benign, F: Follicular thyroid carcinoma, FV: Follicular variant of papillary thyroid carcinoma, H: Hürthle cell carcinoma, P: Papillary thyroid carcinoma.

Author Contributions

Conceptualization, W.-K.C. and J.-H.S.; methodology, W.-K.C.; software, S.-J.P.; validation, J.-H.S., M.-J.L. and W.-Y.C.; resources, F.-H.L. and S.-T.C.; data curation, Y.-R.L.; writing—original draft preparation, W.-K.C.; writing—review and editing, J.-H.S. and S.-J.P.; supervision, S.-J.P.; project administration, S.-J.P. All authors read and agreed to the published version of the manuscript.

Funding

This research was financially supported in part by the Ministry of Science and Technology, Taiwan, under the project MOST 110-2221-E-038-008.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of the Chang Gung Medical Foundation (IRB No. 202001440B0).

Informed Consent Statement

Informed consent was waived because of the retrospective nature of the study and the use of deidentified clinical data.

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

The authors wish to thank all the members of the Cancer Center, Chang Gung Memorial Hospital, for their invaluable help.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Dean, D.S.; Gharib, H. Epidemiology of thyroid nodules. Best Pract. Res. Clin. Endocrinol. Metab. 2008, 22, 901–911. [Google Scholar] [CrossRef]
  2. Uppal, A.; White, M.G.; Nagar, S.; Aschebrook-Kilfoy, B.; Chang, P.J.; Angelos, P.; Kaplan, E.L.; Grogan, R.H. Benign and Malignant Thyroid Incidentalomas Are Rare in Routine Clinical Practice: A Review of 97,908 Imaging Studies. Cancer Epidemiol. Prev. Biomark. 2015, 24, 1327–1331. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Zevallos, J.P.; Hartman, C.M.; Kramer, J.R.; Sturgis, E.M.; Chiao, E.Y. Increased thyroid cancer incidence corresponds to increased use of thyroid ultrasound and fine-needle aspiration: A study of the Veterans Affairs health care system. Cancer 2015, 121, 741–746. [Google Scholar] [CrossRef]
  4. Ganly, I.; Nixon, I.J.; Wang, L.Y.; Palmer, F.L.; Migliacci, J.C.; Aniss, A.; Sywak, M.; Eskander, A.E.; Freeman, J.L.; Campbell, M.J.; et al. Survival from Differentiated Thyroid Cancer: What Has Age Got to Do with It? Thyroid 2015, 25, 1106–1114. [Google Scholar] [CrossRef] [Green Version]
  5. Carling, T.; Udelsman, R. Thyroid cancer. Annu. Rev. Med. 2014, 65, 125–137. [Google Scholar] [CrossRef] [PubMed]
  6. Bible, K.C.; Kebebew, E.; Brierley, J.; Brito, J.P.; Cabanillas, M.E.; Clark, T.J., Jr.; Di Cristofano, A.; Foote, R.; Giordano, T.; Kasperbauer, J.; et al. 2021 American Thyroid Association Guidelines for Management of Patients with Anaplastic Thyroid Cancer. Thyroid 2021, 31, 337–386. [Google Scholar] [CrossRef]
  7. Wells, S.A., Jr.; Asa, S.L.; Dralle, H.; Elisei, R.; Evans, D.B.; Gagel, R.F.; Lee, N.; Machens, A.; Moley, J.F.; Pacini, F.; et al. Revised American Thyroid Association guidelines for the management of medullary thyroid carcinoma. Thyroid 2015, 25, 567–610. [Google Scholar] [CrossRef]
  8. Kushchayeva, Y.; Duh, Q.Y.; Kebebew, E.; D’Avanzo, A.; Clark, O.H. Comparison of clinical characteristics at diagnosis and during follow-up in 118 patients with Hurthle cell or follicular thyroid cancer. Am. J. Surg. 2008, 195, 457–462. [Google Scholar] [CrossRef] [PubMed]
  9. Wong, K.T.; Ahuja, A.T. Ultrasound of thyroid cancer. Cancer Imaging 2005, 5, 157–166. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Han, K.; Ha, H.J.; Kong, J.S.; Kim, J.S.; Myung, J.K.; Koh, J.S.; Park, S.; Shin, M.S.; Song, W.T.; Seol, H.S.; et al. Cytological Features That Differentiate Follicular Neoplasm from Mimicking Lesions. J. Pathol. Transl. Med. 2018, 52, 110–120. [Google Scholar] [CrossRef]
  11. Ferrari, S.M.; Fallahi, P.; Ruffilli, I.; Elia, G.; Ragusa, F.; Paparo, S.R.; Ulisse, S.; Baldini, E.; Giannini, R.; Miccoli, P.; et al. Molecular testing in the diagnosis of differentiated thyroid carcinomas. Gland. Surg. 2018, 7, S19–S29. [Google Scholar] [CrossRef] [PubMed]
  12. Polyzos, S.A.; Anastasilakis, A.D. Clinical complications following thyroid fine-needle biopsy: A systematic review. Clin. Endocrinol. 2009, 71, 157–165. [Google Scholar] [CrossRef]
  13. Mesko, B. The role of artificial intelligence in precision medicine. Expert Rev. Precis. Med. Drug Dev. 2017, 2, 239–241. [Google Scholar] [CrossRef]
  14. Chi, J.; Walia, E.; Babyn, P.; Wang, J.; Groot, G.; Eramian, M. Thyroid Nodule Classification in Ultrasound Images by Fine-Tuning Deep Convolutional Neural Network. J. Digit. Imaging 2017, 30, 477–486. [Google Scholar] [CrossRef]
  15. Song, J.; Chai, Y.J.; Masuoka, H.; Park, S.W.; Kim, S.J.; Choi, J.Y.; Kong, H.J.; Lee, K.E.; Lee, J.; Kwak, N.; et al. Ultrasound image analysis using deep learning algorithm for the diagnosis of thyroid nodules. Medicine 2019, 98, e15133. [Google Scholar] [CrossRef] [PubMed]
  16. Abdolali, F.; Kapur, J.; Jaremko, J.L.; Noga, M.; Hareendranathan, A.R.; Punithakumar, K. Automated thyroid nodule detection from ultrasound imaging using deep convolutional neural networks. Comput. Biol. Med. 2020, 122, 103871. [Google Scholar] [CrossRef]
  17. Guan, Q.; Wang, Y.; Du, J.; Qin, Y.; Lu, H.; Xiang, J.; Wang, F. Deep learning based classification of ultrasound images for thyroid nodules: A large scale of pilot study. Ann. Transl. Med. 2019, 7, 137. [Google Scholar] [CrossRef]
  18. Liu, C.; Xie, L.; Kong, W.; Lu, X.; Zhang, D.; Wu, M.; Zhang, L.; Yang, B. Prediction of suspicious thyroid nodule using artificial neural network based on radiofrequency ultrasound and conventional ultrasound: A preliminary study. Ultrasonics 2019, 99, 105951. [Google Scholar] [CrossRef] [PubMed]
  19. Liu, R.; Zhou, S.; Guo, Y.; Wang, Y.; Chang, C. Nodule Localization in Thyroid Ultrasound Images with a Joint-Training Convolutional Neural Network. J. Digit. Imaging 2020, 33, 1266–1279. [Google Scholar] [CrossRef]
  20. Liu, Z.; Zhong, S.; Liu, Q.; Xie, C.; Dai, Y.; Peng, C.; Chen, X.; Zou, R. Thyroid nodule recognition using a joint convolutional neural network with information fusion of ultrasound images and radiofrequency data. Eur. Radiol. 2021, 31, 5001–5011. [Google Scholar] [CrossRef] [PubMed]
  21. Park, V.Y.; Han, K.; Seong, Y.K.; Park, M.H.; Kim, E.K.; Moon, H.J.; Yoon, J.H.; Kwak, J.Y. Diagnosis of Thyroid Nodules: Performance of a Deep Learning Convolutional Neural Network Model vs. Radiologists. Sci. Rep. 2019, 9, 17843. [Google Scholar] [CrossRef]
  22. Wang, L.; Yang, S.; Yang, S.; Zhao, C.; Tian, G.; Gao, Y.; Chen, Y.; Lu, Y. Automatic thyroid nodule recognition and diagnosis in ultrasound imaging with the YOLOv2 neural network. World J. Surg. Oncol. 2019, 17, 12. [Google Scholar] [CrossRef] [Green Version]
  23. Akkus, Z.; Cai, J.; Boonrod, A.; Zeinoddini, A.; Weston, A.D.; Philbrick, K.A.; Erickson, B.J. A Survey of Deep-Learning Applications in Ultrasound: Artificial Intelligence-Powered Ultrasound for Improving Clinical Workflow. J. Am. Coll. Radiol. 2019, 16, 1318–1328. [Google Scholar] [CrossRef]
  24. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Haugen, B.R.; Alexander, E.K.; Bible, K.C.; Doherty, G.M.; Mandel, S.J.; Nikiforov, Y.E.; Pacini, F.; Randolph, G.W.; Sawka, A.M.; Schlumberger, M.; et al. 2015 American Thyroid Association Management Guidelines for Adult Patients with Thyroid Nodules and Differentiated Thyroid Cancer: The American Thyroid Association Guidelines Task Force on Thyroid Nodules and Differentiated Thyroid Cancer. Thyroid 2016, 26, 1–133. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. McHenry, C.R.; Phitayakorn, R. Follicular adenoma and carcinoma of the thyroid gland. Oncologist 2011, 16, 585–593. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Brito, J.P.; Gionfriddo, M.R.; Al Nofal, A.; Boehmer, K.R.; Leppin, A.L.; Reading, C.; Callstrom, M.; Elraiyah, T.A.; Prokop, L.J.; Stan, M.N.; et al. The accuracy of thyroid nodule ultrasound to predict thyroid cancer: Systematic review and meta-analysis. J. Clin. Endocrinol. Metab. 2014, 99, 1253–1263. [Google Scholar] [CrossRef]
  28. Liu, F.H.; Liou, M.J.; Hsueh, C.; Chao, T.C.; Lin, J.D. Thyroid follicular neoplasm: Analysis by fine needle aspiration cytology, frozen section, and histopathology. Diagn. Cytopathol. 2010, 38, 801–805. [Google Scholar] [CrossRef]
  29. Wu, M.-H.; Chen, K.-Y.; Hsieh, M.-S.; Chen, A.; Chen, C.-N. Risk Stratification in Patients With Follicular Neoplasm on Cytology: Use of Quantitative Characteristics and Sonographic Patterns. Front. Endocrinol. 2021, 12, 350. [Google Scholar] [CrossRef]
  30. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  31. Yoon, J.H.; Kim, E.K.; Hong, S.W.; Kwak, J.Y.; Kim, M.J. Sonographic features of the follicular variant of papillary thyroid carcinoma. J. Ultrasound Med. 2008, 27, 1431–1437. [Google Scholar] [CrossRef] [PubMed]
  32. Li, P.; Liu, P.; Zhang, H. Ultrasonic diagnosis for thyroid Hürthle cell tumor. Cancer Biomark. 2017, 20, 235–240. [Google Scholar] [CrossRef] [PubMed]
  33. Sillery, J.C.; Reading, C.C.; Charboneau, J.W.; Henrichsen, T.L.; Hay, I.D.; Mandrekar, J.N. Thyroid follicular carcinoma: Sonographic features of 50 cases. AJR Am. J. Roentgenol. 2010, 194, 44–54. [Google Scholar] [CrossRef] [PubMed]
  34. Fionda, B.; Boldrini, L.; D’Aviero, A.; Lancellotta, V.; Gambacorta, M.A.; Kovacs, G.; Patarnello, S.; Valentini, V.; Tagliaferri, L. Artificial intelligence (AI) and interventional radiotherapy (brachytherapy): State of art and future perspectives. J. Contemp. Brachyther. 2020, 12, 497–500. [Google Scholar] [CrossRef]
Figure 1. Ultrasound images of (a) papillary thyroid carcinoma, (b) follicular variant of papillary thyroid carcinoma, (c) follicular thyroid carcinoma, (d) Hürthle cell carcinoma, and (e) benign thyroid nodule.
Figure 1. Ultrasound images of (a) papillary thyroid carcinoma, (b) follicular variant of papillary thyroid carcinoma, (c) follicular thyroid carcinoma, (d) Hürthle cell carcinoma, and (e) benign thyroid nodule.
Biomedicines 09 01771 g001
Figure 2. Flowchart of study population. A total of 421 differentiated thyroid cancer patients and 391 patients with benign nodules enrolled after excluding other cancer types and patients without recognizable lesion in sonography. Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC), follicular variant of papillary thyroid carcinoma (FVPTC), Hürthle cell carcinoma (HCC).
Figure 2. Flowchart of study population. A total of 421 differentiated thyroid cancer patients and 391 patients with benign nodules enrolled after excluding other cancer types and patients without recognizable lesion in sonography. Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC), follicular variant of papillary thyroid carcinoma (FVPTC), Hürthle cell carcinoma (HCC).
Biomedicines 09 01771 g002
Figure 3. Ultrasound image of PTC. ROI cropped by a rectangle frame drawn by the author. Region of interest (ROI), papillary thyroid carcinoma (PTC).
Figure 3. Ultrasound image of PTC. ROI cropped by a rectangle frame drawn by the author. Region of interest (ROI), papillary thyroid carcinoma (PTC).
Biomedicines 09 01771 g003
Figure 4. Process of training and classification. After preprocessing and ROI cropping of nodule images, 2308 images were split into a training set and test set. Data augmentation based on histogram equalization/normalization and horizontal flipping increased the number of images fourfold. Transferred learning by three pretrained CNNs was performed with fivefold cross-validation; then, the diagnostic performance using the test set was evaluated. Region of interest (ROI), convolution neural network (CNN).
Figure 4. Process of training and classification. After preprocessing and ROI cropping of nodule images, 2308 images were split into a training set and test set. Data augmentation based on histogram equalization/normalization and horizontal flipping increased the number of images fourfold. Transferred learning by three pretrained CNNs was performed with fivefold cross-validation; then, the diagnostic performance using the test set was evaluated. Region of interest (ROI), convolution neural network (CNN).
Biomedicines 09 01771 g004
Figure 5. (a) Cropped image of PTC nodule. (b) Cropped image was preprocessed with horizontal flipping and histogram normalization. Papillary thyroid carcinoma (PTC).
Figure 5. (a) Cropped image of PTC nodule. (b) Cropped image was preprocessed with horizontal flipping and histogram normalization. Papillary thyroid carcinoma (PTC).
Biomedicines 09 01771 g005
Figure 6. Confusion matrix of CNNs in test set. (a) InceptionV3, (b) ResNet101, and (c) VGG19. M: malignant group, B: benign group.
Figure 6. Confusion matrix of CNNs in test set. (a) InceptionV3, (b) ResNet101, and (c) VGG19. M: malignant group, B: benign group.
Biomedicines 09 01771 g006
Figure 7. ROC curve of CNN models and performance of endocrinologists. AUC of InceptionV3 = 0.81, AUC of ResNet101 = 0.81, and AUC of VGG19 = 0.80. Receiver operating characteristic (ROC), convolution neural network (CNN), area under the curve (AUC).
Figure 7. ROC curve of CNN models and performance of endocrinologists. AUC of InceptionV3 = 0.81, AUC of ResNet101 = 0.81, and AUC of VGG19 = 0.80. Receiver operating characteristic (ROC), convolution neural network (CNN), area under the curve (AUC).
Biomedicines 09 01771 g007
Table 1. Baseline demography of five pathology types divided into malignant and benign groups.
Table 1. Baseline demography of five pathology types divided into malignant and benign groups.
Pathological Types
Malignant Group
(n = 421)
Benign Group (n = 391)
PTCFVPTCFTCHCCBenignp-Value
Number of patients2141147023391
Age (years), mean (SD)47.36 ± 13.7044.90 ± 15.1246.61 ± 17.1051.17 ± 15.6254.18 ± 13.15<0.0001
Sex (n, %)
Male53 (25)28 (25)15 (21)4 (17)80 (20)0.6985
Female161 (75)86 (75)55 (79)19 (83)311 (80)
Number of US images47021513138937
Number of cropped images533272175531275
Location (n, %)
Left88 (41.12)52 (45.61)39 (55.71)13 (56.52)143 (36.57)0.0017
Right109 (50.93)57 (50.00)27 (38.57)10 (43.48)192 (49.10)
Both (Left + Right)11 (5.15)4 (3.51)1 (1.43)0 (0.00)47 (12.03)
Isthmus6 (2.80)1 (0.88)3 (4.29)0 (0.00)9 (2.30)
Ultrasound brands (%)
Aloka3.932.466.670.007.57<0.0001
GE Healthcare37.1262.3064.0178.2645.87
Hitachi7.422.461.330.007.34
Philips4.370.828.008.702.06
Siemens28.8210.665.338.7018.81
Toshiba17.9016.3913.334.3418.12
Others0.444.911.330.000.23
Continuous data are expressed as the mean and standard deviation (SD); categorical data are expressed as a percentage (%). Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC), follicular variant of papillary thyroid carcinoma (FVPTC), Hürthle cell carcinoma (HCC), ultrasound (US).
Table 2. Distribution of pathological groups as a function of ultrasound brands.
Table 2. Distribution of pathological groups as a function of ultrasound brands.
AlokaGE HealthcareHitachiPhilipsSiemensToshibaOthersp-Value
PTC18.0019.9032.0735.7239.5227.1511.11<0.0001
FVPTC6.0017.805.663.577.7813.2566.67<0.0001
FTC10.0011.241.8921.432.406.6211.11<0.0001
HCC0.004.220.007.141.200.660.00<0.0001
Benign66.0046.8460.3832.1449.1052.3211.11<0.0001
Categorical data are expressed as a percentage (%). Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC), follicular variant of papillary thyroid carcinoma (FVPTC), Hürthle cell carcinoma (HCC).
Table 3. Distribution of histopathological and ultrasonic features among groups.
Table 3. Distribution of histopathological and ultrasonic features among groups.
HistopathologyNumber (%)
Malignant group (n = 421)PTC features
(n = 214)
Classic PTC208 (97.19)
Diffuse sclerosing variant3 (1.40)
Tall cell variant1 (0.47)
Cribriform morular variant1 (0.47)
Encapsulated variant1 (0.47)
FTC features
(n = 207)
Follicular variant of PTC106 (51.21)
Follicular carcinoma, minimally invasive70 (33.82)
Hürthle cell carcinoma23 (11.11)
Encapsulated follicular variant of PTC8 (3.86)
Benign group (n = 391)Nodular hyperplasia289 (73.91)
Follicular adenoma48 (12.28)
Cyst47 (12.02)
Hürthle cell adenoma7 (1.79)
Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC).
Table 4. Performance of CNNs and endocrinologists in test set.
Table 4. Performance of CNNs and endocrinologists in test set.
SensitivitySpecificityPPVNPVAccuracyAUC
InceptionV376.076.971.880.676.50.82
ResNet10172.581.475.179.377.60.83
VGG1966.283.775.876.276.10.83
Endocrinologist 138.774.253.761.158.8-
Endocrinologist 235.382.661.062.362.0-
Positive predictive value (PPV), negative predictive value (NPV), area under curve (AUC).
Table 5. Diagnostic accuracy of CNNs and endocrinologists on different pathological types in test set.
Table 5. Diagnostic accuracy of CNNs and endocrinologists on different pathological types in test set.
Malignant GroupBenign Group
PTCFVPTCFTCHCCNHFAC
InceptionV381.472.972.766.775.065.092.5
ResNet10173.274.669.766.779.480.090.0
VGG1964.971.263.660.082.475.095.0
Endocrinologist 158.820.327.313.373.080.080.0
Endocrinologist 253.617.030.36.781.975.090.0
Papillary thyroid carcinoma (PTC), follicular thyroid carcinoma (FTC), follicular variant of papillary thyroid carcinoma (FVPTC), Hürthle cell carcinoma (HCC), Nodular hyperplasia (NH), Follicular adenoma (FA), Cyst (C).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chan, W.-K.; Sun, J.-H.; Liou, M.-J.; Li, Y.-R.; Chou, W.-Y.; Liu, F.-H.; Chen, S.-T.; Peng, S.-J. Using Deep Convolutional Neural Networks for Enhanced Ultrasonographic Image Diagnosis of Differentiated Thyroid Cancer. Biomedicines 2021, 9, 1771. https://doi.org/10.3390/biomedicines9121771

AMA Style

Chan W-K, Sun J-H, Liou M-J, Li Y-R, Chou W-Y, Liu F-H, Chen S-T, Peng S-J. Using Deep Convolutional Neural Networks for Enhanced Ultrasonographic Image Diagnosis of Differentiated Thyroid Cancer. Biomedicines. 2021; 9(12):1771. https://doi.org/10.3390/biomedicines9121771

Chicago/Turabian Style

Chan, Wai-Kin, Jui-Hung Sun, Miaw-Jene Liou, Yan-Rong Li, Wei-Yu Chou, Feng-Hsuan Liu, Szu-Tah Chen, and Syu-Jyun Peng. 2021. "Using Deep Convolutional Neural Networks for Enhanced Ultrasonographic Image Diagnosis of Differentiated Thyroid Cancer" Biomedicines 9, no. 12: 1771. https://doi.org/10.3390/biomedicines9121771

APA Style

Chan, W. -K., Sun, J. -H., Liou, M. -J., Li, Y. -R., Chou, W. -Y., Liu, F. -H., Chen, S. -T., & Peng, S. -J. (2021). Using Deep Convolutional Neural Networks for Enhanced Ultrasonographic Image Diagnosis of Differentiated Thyroid Cancer. Biomedicines, 9(12), 1771. https://doi.org/10.3390/biomedicines9121771

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop