Next Article in Journal
Implementation of an Intelligent Exam Supervision System Using Deep Learning Algorithms
Next Article in Special Issue
Assessing Cognitive Workload Using Cardiovascular Measures and Voice
Previous Article in Journal
Test and Analysis of Vegetation Coverage in Open-Pit Phosphate Mining Area around Dianchi Lake Using UAV–VDVI
Previous Article in Special Issue
Artificial Neural Networks Combined with the Principal Component Analysis for Non-Fluent Speech Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predictions for Three-Month Postoperative Vocal Recovery after Thyroid Surgery from Spectrograms with Deep Neural Network

1
Division of Biomedical Informatics, Seoul National University Biomedical Informatics (SNUBI), Seoul National University College of Medicine, Seoul 110799, Korea
2
Department of Otolaryngology, Thyroid/Head & Neck Cancer Center, The Dongnam Institute of Radiological & Medical Sciences (DIRAMS), Busan 46033, Korea
3
Samsung Electronics Co., Ltd., 130 Samsung-ro, Yeongtong-gu, Suwon-si 16677, Korea
4
Microsoft, Redmond, WA 98052, USA
5
Department of Radiology, Pusan National University Yangsan Hospital, Yangsan 50612, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6387; https://doi.org/10.3390/s22176387
Submission received: 7 June 2022 / Revised: 11 August 2022 / Accepted: 17 August 2022 / Published: 24 August 2022
(This article belongs to the Special Issue Biomedical Signal and Image Processing in Speech Analysis)

Abstract

:
Despite the lack of findings in laryngeal endoscopy, it is common for patients to undergo vocal problems after thyroid surgery. This study aimed to predict the recovery of the patient’s voice after 3 months from preoperative and postoperative voice spectrograms. We retrospectively collected voice and the GRBAS score from 114 patients undergoing surgery with thyroid cancer. The data for each patient were taken from three points in time: preoperative, and 2 weeks and 3 months postoperative. Using the pretrained model to predict GRBAS as the backbone, the preoperative and 2-weeks-postoperative voice spectrogram were trained for the EfficientNet architecture deep-learning model with long short-term memory (LSTM) to predict the voice at 3 months postoperation. The correlation analysis of the predicted results for the grade, breathiness, and asthenia scores were 0.741, 0.766, and 0.433, respectively. Based on the scaled prediction results, the area under the receiver operating characteristic curve for the binarized grade, breathiness, and asthenia were 0.894, 0.918, and 0.735, respectively. In the follow-up test results for 12 patients after 6 months, the average of the AUC values for the five scores was 0.822. This study showed the feasibility of predicting vocal recovery after 3 months using the spectrogram. We expect this model could be used to relieve patients’ psychological anxiety and encourage them to actively participate in speech rehabilitation.

1. Introduction

Voice disorders, the most common discomfort, are known to occur in 40% to 80% of patients who undergo thyroid surgery [1,2]. Vocal cord paralysis, which is caused by nerve damage, including damage to the recurrent laryngeal nerve, receives the most medical attention. However, it is very rare and found only in about 1% of patients, while functional vocal problems that occur without special anatomical abnormalities after thyroid surgery have been reported to occur about 30% of the time [1,2]. After a thyroidectomy, although patients may have normal findings in terms of their laryngeal endoscopy, it is common that they may complain of vocal dysfunction [3,4]. In many cases, it is difficult to find the cause of the change in speech by way of visual observations such as performing a laryngoscope examination.
The typical symptoms of patients who complain of vocal dysfunction after thyroid surgery are vocal fatigue, difficulty producing treble, difficulty maintaining vocalization, difficulty breathing, mild hoarseness, and problems with swallowing [5]. Functional voice disorders are characterized by dysphonia in the presence of apparently normal vocal fold anatomy and movement. Choi et al., reported that even in the absence of laryngeal nerve damage, half of the patients may experience vocal changes, which require a 6-month recovery period [1]. The ability to identify early changes would facilitate an early referral for a comprehensive voice evaluation that aims at improving the patient’s quality of life, preventing secondary injuries, and identifying those who might benefit from vocal fold augmentation [6].
The GRBAS scale has been established as the most compact perceptual grading system [7]. The perceptual GRBAS scale is reliable, and has been applied worldwide. Its use has enabled a better level of communication among clinicians and speech-language pathologists concerning their patients’ voices, which constitutes a major advantage for an assessment tool [8].
In otorhinolaryngology, because the vocal samples are collected using the same protocol, quantitative analysis is possible in terms of using objective and quantitative data [9,10]. The acoustic sound generated through acoustic-voice inspection can be used as various types of data by preprocessing it. Studies reported that the normalization of speech data into spectrograms can be used for classification and regression [11]. There have been several studies using a spectrogram that transforms voice data into an image form using a convolutional neural network (CNN) algorithm [12,13,14]. Moreover, the GRBAS score, a subjective measure among the vocal analysis methods, was predicted by researchers using machine learning techniques [15].
With the advent of machine learning technology, the use of artificial intelligence (AI) in medicine has become essential [16]. Particularly in thyroid cancer, deep learning has been widely used for the diagnosis of disease and the detection of lesions [17,18,19]. Because vocal diseases have a great impact on a patient’s quality of life, predicting the prognosis is crucial [20,21]. However, despite advances in acoustic-acoustic testing, AI studies using audio data are relatively rare in medicine.
This study aimed to predict vocal recovery after 3 months by using a deep neural network algorithm from preoperative and postoperative vocal spectrograms. This can improve the patient’s quality of life and allow them to receive treatments in advance. The performance of the model was verified through internal consecutive split validation, and the possibility of an application to AI’s negative disease has been examined.

2. Related Works

Our study has several differences compared with previous related studies on postoperative outcomes [22,23,24]. There have been studies to find prognostic factors related to the voice outcome. However, these were studies analyzing the prognostic factors related to voice, not AI-based prognosis prediction models. In this study, the goal was to predict long-term outcomes in advance using the patient’s pre- and postoperative voices with deep learning. Furthermore, previous studies on using acoustic samples for voice disorder prediction have focused on the diagnosis of the current condition [25,26,27]. By contrast, in our study, we predicted the patient’s prognosis and quickly suggested treatment to patients with poor prognoses. Providing these clinical decisions will help improve patients’ quality of life.

3. Materials and Methods

3.1. Patients and Vocal Data

The protocol for this retrospective study was approved by the Ethics Committee of the Institutional Review Board (D-1801-003-002) at Dongnam Institute of Radiological & Medical Sciences (DIRAMS). Written or oral informed consent was not obtained from the patients because this study had a nonintrusive retrospective design and the IRB waived the need for individual informed consent because all data were anonymously analyzed. All experiments were performed in accordance with relevant guidelines and regulations, and all experimental protocols were approved by DIRAMS. The vocal data analyzed during the current study are not publicly available, but are available from the corresponding author upon reasonable request.
Our exclusion criteria were as follows: first, the patients with previous laryngeal or vocal fold paralysis; second, patients with pulmonary diseases; third, patients who had previous neck surgery; fourth, patients who showed post-thyroidectomy evidence of recurrent laryngeal nerve; and, finally, patients with an injury of the external branch of the superior laryngeal nerve palsy. Patients with preoperative, postoperative, and after-3-months voice samples were recruited in this study.
We consecutively enrolled 114 patients with a thyroidectomy who were diagnosed with thyroid cancer and who had visited the hospital from January 2018 to December 2019. The average age was 47.4 years, and the sex split was 26 men and 88 women. The distribution of grade scores for all patients can be found in Table 1.
To measure the performance of the model, internal consecutive split validation was performed, and the test set was divided by time. Of the total data, 92 sample data, corresponding to 80%, were used for data training and for the parameter tuning set, and 22 sample data recruited last, corresponding to 20%, were used as a test set.
To evaluate the degree of patients’ speech impairment, the GRBAS scale, a perceptual evaluation method, was applied to compare the degree of speech impairment in each session [28]. The GRBAS system for describing vocal quality contains five well-defined parameters: G (overall grade of hoarseness), R (roughness), B (breathiness), A (asthenic), and S (the strained) quality of the voice [7].

3.2. Preprocessing the Vocal Data

The speech signals in the vocal data—preoperative, postoperative, and 3 months postoperative—were all generated through the same protocol. The speech signal in the voice data with durations ranging from less than 1 s to about 20 s was sampled at 11,025 Hz. Each sentence was labeled with the GRBAS score by an experienced phonologist. We calculated the spectrograms for all vocal data. We ignored the rest and used a frequency range of 0–6 KHz. The DFT was calculated through a frame size of 1024. The DFT data were converted to the log-power spectrum. Finally, through RGBA mapping, spectrogram images of the same size were created. The model architecture and the process of generating the spectrogram are included in the Supplementary Material.

3.3. The Two Stages of Deep Neural Networks

To predict the amount of vocal recovery after 3 months using pre- and postoperative vocal data, we first developed a pretrained model to predict the GRBAS score. To evaluate the performance of predicting the GRBAS score from the voice sample, the performance was measured using a 20% random split. The architecture of the deep neural network was based on EfficientNet-B4, which is a recent state of the art structure [29]. We added two fully connected layers to the last layer of the CNN model, composed of 1024 and 5 units for the final layer, respectively. The mean squared error (MSE) was used as the loss function, and the model was compiled to predict the GRBAS score as a label for the final layer with five units. The model was used with the parameters pretrained from ImageNet data, and the fine-tuning of the entire model was performed using all data.
The above model was used as an encoding model to extract the important features to predict the GRBAS score as a pretrained model. Two CNN models were connected by separately applying models that extracted 1024 features to pre- and postoperative vocal data. We concatenated two feature vectors to compose the input of the next long short-term memory (LSTM) layer, which is a variant of the recurrent neural network involving time series by introducing a three gates operation (input, output, and a forget gates) [30]. Then, we used the LSTM recurrent neural network for the time-series prediction at the postoperative and 3-month times from pre- and postoperative spectrograms [31]. Finally, this model architecture was fine-tuned to predict the postoperative GRBAS scores and GRBAS score after 3 months without a spectrogram after 3 months with an MSE loss of 25. The scheme of the overall model structure is shown in Figure 1, and the detailed structure and source code of the model are shown in Supplementary Materials.
To avoid the overfitting, online augmentation was performed by transforming the images while learning. During the learning process, the rotation range within 20 degrees, the 20% range for the width and height, and the shear zoom were randomly applied to the training images. This process was performed using ImageDataGenerator supported by Tensorflow Keras (Google, Mountain View, CA, USA). However, the flip, which could change the meaning of the spectrogram, was not used.
We used the Keras deep-learning framework included in Tensorflow version 2.3.0 (Google, Mountain View, CA, USA). We trained the model using the Adam optimizer to accelerate the convergence of network parameters with a learning rate of 0.001. The learning rate was gradually reduced through the call-back function. The batch size was 24. A total of 200 epochs were performed, and the weight with the lowest validation loss was used. We used EfficientNet-B4 as the backbone architecture of the CNN (https://github.com/qubvel/efficientnet—accessed on 16 January 2022).

3.4. The Activation Heatmap of the Convolution Layer

We inferred the activation heatmap of the model that predicted the GRBAS score to identify the important location in the spectrograms. To derive the activation heatmap, we used the gradient-weighted class activation mapping (Grad-CAM) method using gradient-based localization [32]. The heatmap was derived by computing the gradients for class with respect to the feature map of the selected convolutional layer. The activation heatmaps were derived from the convolutional layer in the second and third residual block, corresponding to 56 × 56 and 28 × 28 of the model, respectively.

3.5. Performance Evaluation

To measure the performance of the model, internal consecutive split validation was performed with the test set divided by time. The final output was the GRBAS score value, the MSE values were calculated for each label in 22 patients, and the Spearman’s rank correlation test was performed. To measure the area under the receiver operating characteristic curve (AUROC), the GRBAS scores were divided into binary values of 0 or greater than 1.
In Equation (1), n and i denote the total number of patients and the index of the patient’s vocal sample, respectively. t and s denote the time point of the vocal samples (preoperative, postoperative, and 3-months-after vocal sample). This metric was calculated in terms of the five scores representing the GRBAS. The Spearman’s rank correlation coefficient, used for the performance evaluation of numeric labels, was computed for the derived parameters according to Equation (2), where n and d denote the number of voice samples and the rank difference between the GRBAS score and predicted GRBAS score, respectively.
R M S E t s = 1 n i = 1 n y i t s y ˜ i t s 2
ρ = 1 6 d i 2 n n 2 1

4. Results and Discussion

4.1. The Prediction Performance of the Deep Neural Network

Table 2 summarizes the performance of the deep neural network predicting GRBAS after 3 months of surgical treatment using pre- and postoperative voice spectrograms. The overall root mean square error (RMSE) value of the GRBAS score of the deep neural network was 0.3798. The RMSE values for each GRBAS score was 0.399, 0.365, 0.409, 0.469, and 0.203. In the Spearman’s rank correlation test, the grade was 0.741, roughness was 0.153, breathiness was 0.766, and asthenia was 0.433; the strain could not be calculated because all test samples had zeros. Only two patients had a value of one for roughness, and all others were zero. Where the Spearman’s rank correlation results were concerned, scores that had a statistically significant correlation were grade, breathiness, and asthenia (p < 0.01). For the roughness of the test sample, only two samples had a score of one, and the rest were zero. Figure 2 shows the relationship between the prediction results for grade, breathiness, and asthenia for the test set and the actual score.
We used EfficientNet-B4 among the CNN architectures. The changes in performance when using other CNN models are summarized in Supplementary Table S1. The efficientnet-b4 architecture had the lowest RMSE value in terms of the average value of all scores.

4.2. Binary Classification Performance for Prognostic Prediction

Because RMSE and correlation have different meanings that depend on the range and scale of the data, they do not represent the performance of an objective model. Therefore, we divided the grade, breathiness, and asthenia scores into binary values of zero or greater than one, and measured AUROC scores using these (Figure 3). The AUROC values were 0.894, 0.918, and 0.735 for grade, breathiness, and asthenia, respectively. The roughness score was more than one among 22 patients used as the test set, and the AUROC value was 0.575, which was poorly predicted. The strain score was zero for all patients in the test set. Because the strain score was zero for all patients in the test set, it was impossible to calculate the AUROC value.
The optimal thresholds corresponding to the Youden index yielded 76.5% and 100.0% sensitivity and specificity for the grade score. For breathiness and asthenia, sensitivity and specificity were 70.6% and 100.0%, and 55.6% and 92.3%, respectively.

4.3. Prognosis in Patients after 6 Months

Among the consecutively recruited patients, vocal samples after 6 months were obtained for 12 of 22 patients in the test set. The distribution of GRBAS scores of patients 6 months after surgery is shown in Supplementary Table S2. The AUROC values were 0.852, 0.800, 0.688, 0.861, and 0.909 for grade, roughness, breathiness, asthenia, and strain, respectively (Figure 4).
The optimal thresholds given by the Youden index, sensitivity and specificity of the grade score, were 78.8% and 100.0%, respectively. For roughness and breathiness, sensitivity and specificity were 60.0% and 100.0%, and 62.5% and 75.0%, respectively. For asthenia and strain, sensitivity and specificity were 100.0% and 83.3%, and 100.0% and 90.9%, respectively.

4.4. Scores for the Activation Heatmap

To visualize the important features of the scores in the spectrograms, an activation heatmap was constructed using the second and third residual block of the EfficientNet model (Figure 5). The voice spectrogram of a patient with high-degree grade was shown in Figure 5A. This spectrogram is from one of the patients with very poor voice quality. In the heatmap visualization through Grad-CAM, the highlighted part shows an important imaging feature to predict the grade score. The spectrogram of a normal grade voice is shown in Figure 5B. In this patient, the amplitude widely spreads over various frequencies. The region where the heatmap is activated in the spectrogram is very narrow.

4.5. Discussion

This study aimed to predict patients who will have problems with long-term vocal recovery. We obtained patient voice samples for those who underwent surgical treatment as well as their GRBAS scores from experienced phonologists for three points in time: preoperative, postoperative, and 3 months after surgery. Using this data, we developed a model that predicts, using deep learning, the GRBAS score of the patient’s voice 3 months after the operation. For model performance, the RMSE value was 0.3798, and the results for the Spearman’s rank correlation analysis for the grade, breathiness, and asthenia scores were 0.796, 0.784, and 0.602, respectively. When each score was binarized, the AUROC values for grade, breathiness, and asthenia were 0.894, 0.918, and 0.735, respectively. Through these results, we showed the possibility of predicting a patient’s long-term vocal disorder using the pre- and postoperative voice samples. This method is expected to help relieve patients’ psychological anxiety and encourage patients to actively participate in speech rehabilitation.
There have been studies conducted to predict voice quality and voice disorders using deep learning from voice samples [33,34,35]. The results of these studies have demonstrated the potential of using predictive models in clinical settings, which have shown high performance comparable to human evaluation. This study aimed to predict voice quality after 3 months from pre- and postoperative vocal samples. Long-term voice change after 3 months can be used for prognosis prediction, a problem that cannot be evaluated by humans. To the best of our knowledge, this is the first study predicting the prognosis of long-term voice quality after surgery.
Evaluating the objective patient’s voice is a challenging task because it is laborious, time-consuming, and cost-intensive [36]. Additionally, it is difficult to objectively measure voice samples given that interobserver variability may exist depending on the phonetician, and intraobserver variability may exist depending on the condition [7,37]. To provide an objective measure for this, there have been various methods that predict GRBAS by using machine learning. In our study, we implemented a deep-learning model that predicts the voice after 3 months postsurgery based on a pretrained model, which is the feature extractor based on the GRBAS score.
The results for predicting vocal diseases 3 months after a patient’s surgery can be regarded as a prognostic marker. Long-term vocal defects are a major factor in a patient’s quality of life, and it is almost impossible to predict them in advance. In this study, we found that by using internal consecutive split validation, AI models could predict long-term vocal defects. This means that the patient’s voice, as generated by the same protocol, can be used for objective data. Therefore, we suggest the possibility that a patient’s voice data can be applied to various purposes to make a prognosis for the patient.
Our study has some limitations. Our data ere from 112 people, which is not enough to train a deep neural network. Next, we are planning a validation study by recruiting more patients with long-term follow-up. We did not use factors that could have affected interpersonal phonetic outcomes, such as the patient’s surgical range, method, or age. Because the variability in the GRBAS score varies depending on the person who measures it, predicting it cannot completely cover the vocal condition. To find an objective answer to this problem, we first constructed a pretrained model that extracts features based on GRBAS, and we attempted to avoid the problem by applying a deep neural network to pre- and postoperative data.
In conclusion, we developed a model for predicting the recovery of a patient’s voice after 3 months by using a deep neural network algorithm from pre- and postoperative voice spectrograms. This approach can help physicians select patients with long-term vocal disorders for whom intensive care should be applied.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22176387/s1, Table S1: RMSE values for prognosis prediction according to the architecture of CNN model; Table S2: Distribution of GRBAS scores in patients 6 months after surgery.

Author Contributions

H.Y.S. conceived, coordinated, and directed all processes of this study; J.H.L. conducted the voice data analysis, data structure conversion, statistical analysis, and deep learning model development and wrote the manuscript; C.Y.L. supervised the patient’s voice data generation, visualization of the results, and writing of the manuscript; J.S.E. and M.P. conducted the comprehensive development of the CNN-RNN deep learning model, and H.S.J. conducted a comprehensive interpretation of the predictions for 3-month postoperative vocal recovery. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The protocol for this retrospective study was approved by the Ethics Committee of the Institutional Review Board (D-1801-003-002) at the Dongnam Institute of Radiological & Medical Sciences (DIRAMS).

Informed Consent Statement

Written or oral informed consent was not obtained from the patients because this study had a nonintrusive retrospective design, and IRB waived the need for individual in-formed consent because all data were anonymously analyzed. All experiments were performed in accordance with relevant guidelines and regulations and all experimental protocols were approved by DIRAMS.

Data Availability Statement

The vocal data analyzed during the current study are not publicly available, but are available to the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, J.-S.; Jeong, J.-I.; Jang, M.-S.; Son, Y.-I. Voice Changes after Thyroidectomy without Recurrent Laryngeal Nerve Injury. J. Korean Soc. Laryngol. Phoniatr. Logop. 2010, 21, 37–41. [Google Scholar]
  2. Lombardi, C.P.; Raffaelli, M.; De Crea, C.; D’Alatri, L.; Maccora, D.; Marchese, M.R.; Paludetti, G.; Bellantone, R. Long-Term Outcome of Functional Post-Thyroidectomy Voice and Swallowing Symptoms. Surgery 2009, 146, 1174–1181. [Google Scholar] [CrossRef] [PubMed]
  3. Roy, N.; Barkmeier-Kraemer, J.; Eadie, T.; Sivasankar, M.P.; Mehta, D.; Paul, D.; Hillman, R. Evidence-Based Clinical Voice Assessment: A Systematic Review. Am. J. Speech-Lang. Pathol. 2013, 22, 212–226. [Google Scholar] [CrossRef]
  4. Randolph, G.W. The Importance of Pre-and Postoperative Laryngeal Examination for Thyroid Surgery. Thyroid 2010, 20, 453–458. [Google Scholar] [CrossRef] [PubMed]
  5. Chandrasekhar, S.S.; Randolph, G.W.; Seidman, M.D.; Rosenfeld, R.M.; Angelos, P.; Barkmeier-Kraemer, J.; Benninger, M.S.; Blumin, J.H.; Dennis, G.; Hanks, J.; et al. Clinical Practice Guideline: Improving Voice Outcomes after Thyroid Surgery. Otolaryngol. Head Neck Surg. 2013, 148, S1–S37. [Google Scholar] [CrossRef]
  6. Henry, L.R.; Helou, L.B.; Solomon, N.P.; Howard, R.S.; Gurevich-Uvena, J.; Coppit, G.; Stojadinovic, A. Functional Voice Outcomes after Thyroidectomy: An Assessment of the Dsyphonia Severity Index (DSI) after Thyroidectomy. Surgery 2010, 147, 861–870. [Google Scholar] [CrossRef]
  7. De Bodt, M.S.; Wuyts, F.L.; de Heyning, P.H.; Croux, C. Test-Retest Study of the GRBAS Scale: Influence of Experience and Professional Background on Perceptual Rating of Voice Quality. J. Voice 1997, 11, 74–80. [Google Scholar] [CrossRef]
  8. Wuyts, F.L.; De Bodt, M.S.; de Heyning, P.H. Is the Reliability of a Visual Analog Scale Higher than an Ordinal Scale? An Experiment with the GRBAS Scale for the Perceptual Evaluation of Dysphonia. J. Voice 1999, 13, 508–517. [Google Scholar] [CrossRef]
  9. Bhuta, T.; Patrick, L.; Garnett, J.D. Perceptual Evaluation of Voice Quality and Its Correlation with Acoustic Measurements. J. Voice 2004, 18, 299–304. [Google Scholar] [CrossRef]
  10. Xie, Z.; Gadepalli, C.; Farideh, J.; Cheetham, B.M.G.; Homer, J.J. Machine Learning Applied to GRBAS Voice Quality Assessment. Adv. Sci. Technol. Eng. Syst. J. 2018, 3, 329–338. [Google Scholar] [CrossRef]
  11. Dennis, J.; Tran, H.D.; Li, H. Spectrogram Image Feature for Sound Event Classification in Mismatched Conditions. IEEE Signal Process. Lett. 2010, 18, 130–133. [Google Scholar] [CrossRef]
  12. Costa, Y.M.G.; Oliveira, L.S.; Silla Jr, C.N. An Evaluation of Convolutional Neural Networks for Music Classification Using Spectrograms. Appl. Soft Comput. 2017, 52, 28–38. [Google Scholar] [CrossRef]
  13. Sakashita, Y.; Aono, M. Acoustic Scene Classification by Ensemble of Spectrograms Based on Adaptive Temporal Divisions. In Technical report, Detection and Classification of Acoustic Scenes and Events (DCASE) Challenge; DCASE Community: Washington, DC, USA, 2018. [Google Scholar]
  14. Wyse, L. Audio Spectrogram Representations for Processing with Convolutional Neural Networks. arXiv 2017, arXiv:1706.09559. [Google Scholar]
  15. Xie, S.; Yan, N.; Yu, P.; Ng, M.L.; Wang, L.; Ji, Z. Deep Neural Networks for Voice Quality Assessment Based on the GRBAS Scale. In Proceedings of the Interspeech, San Francisco, CA, USA, 8–12 September 2016; pp. 2656–2660. [Google Scholar]
  16. Hamet, P.; Tremblay, J. Artificial Intelligence in Medicine. Metabolism 2017, 69, S36–S40. [Google Scholar] [CrossRef]
  17. Lee, J.H.; Baek, J.H.; Kim, J.H.; Shim, W.H.; Chung, S.R.; Choi, Y.J.; Lee, J.H. Deep Learning—Based Computer-Aided Diagnosis System for Localization and Diagnosis of Metastatic Lymph Nodes on Ultrasound: A Pilot Study. Thyroid 2018, 28, 1332–1338. [Google Scholar] [CrossRef]
  18. Lee, J.H.; Ha, E.J.; Kim, J.H. Application of Deep Learning to the Diagnosis of Cervical Lymph Node Metastasis from Thyroid Cancer with CT. Eur. Radiol. 2019, 29, 5452–5457. [Google Scholar] [CrossRef]
  19. Lee, J.H.; Ha, E.J.; Kim, D.; Jung, Y.J.; Heo, S.; Jang, Y.; An, S.H.; Lee, K. Application of Deep Learning to the Diagnosis of Cervical Lymph Node Metastasis from Thyroid Cancer with CT: External Validation and Clinical Utility for Resident Training. Eur. Radiol. 2020, 30, 3066–3072. [Google Scholar] [CrossRef]
  20. Alhussein, M.; Muhammad, G. Voice Pathology Detection Using Deep Learning on Mobile Healthcare Framework. IEEE Access 2018, 6, 41034–41041. [Google Scholar] [CrossRef]
  21. Gama, A.C.C.; Cardoso, F.E.C.; Reis, C.A.d.C.; Bassi, I.B. Idiopathic Parkinson’s Disease: Vocal and Quality of Life Analysis. Arq. Neuropsiquiatr. 2012, 70, 674–679. [Google Scholar]
  22. Ongkasuwan, J.; Espinosa, M.C.L.; Hollas, S.; Devore, D.; Procter, T.; Bassett, E.; Schwabe, A. Predictors of Voice Outcome in Pediatric Non-Selective Laryngeal Reinnervation. Laryngoscope 2020, 130, 1525–1531. [Google Scholar] [CrossRef]
  23. Lee, D.Y.; Lee, K.J.; Hwang, S.M.; Oh, K.H.; Cho, J.-G.; Baek, S.-K.; Kwon, S.-Y.; Woo, J.-S.; Jung, K.-Y. Analysis of Temporal Change in Voice Quality after Thyroidectomy: Single-Institution Prospective Study. J. Voice 2017, 31, 195–201. [Google Scholar] [CrossRef]
  24. Tedla, M.; Chakrabarti, S.; Suchankova, M.; Weickert, M.O. Voice Outcomes after Thyroidectomy without Superior and Recurrent Laryngeal Nerve Injury: VoiSS Questionnaire and GRBAS Tool Assessment. Eur. Arch. Oto-Rhino-Laryngol. 2016, 273, 4543–4547. [Google Scholar] [CrossRef]
  25. Tulics, M.G.; Szaszák, G.; Mészáros, K.; Vicsi, K. Using ASR Posterior Probability and Acoustic Features for Voice Disorder Classification. In Proceedings of the 2020 11th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Mariehamn, Aland, 23–25 September 2020; pp. 155–160. [Google Scholar]
  26. Tulics, M.G.; Szaszák, G.; Mészáros, K.; Vicsi, K. Artificial Neural Network and Svm Based Voice Disorder Classification. In Proceedings of the 2019 10th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Naples, Italy, 23–25 October 2019; pp. 307–312. [Google Scholar]
  27. Sauder, C.; Bretl, M.; Eadie, T. Predicting Voice Disorder Status from Smoothed Measures of Cepstral Peak Prominence Using Praat and Analysis of Dysphonia in Speech and Voice (ADSV). J. Voice 2017, 31, 557–566. [Google Scholar] [CrossRef]
  28. Fex, B.; Fex, S.; Shiromoto, O.; Hirano, M. Acoustic Analysis of Functional Dysphonia: Before and after Voice Therapy (Accent Method). J. Voice 1994, 8, 163–167. [Google Scholar] [CrossRef]
  29. Tan, M.; Le, Q. V EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
  30. Van Houdt, G.; Mosquera, C.; Nápoles, G. A Review on the Long Short-Term Memory Model. Artif. Intell. Rev. 2020, 53, 5929–5955. [Google Scholar] [CrossRef]
  31. Mikolov, T.; Kombrink, S.; Burget, L.; Černocký, J.; Khudanpur, S. Extensions of Recurrent Neural Network Language Model. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5528–5531. [Google Scholar]
  32. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-Cam: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  33. Arias-Londoño, J.D.; Gómez-García, J.A.; Godino-Llorente, J.I. Multimodal and Multi-Output Deep Learning Architectures for the Automatic Assessment of Voice Quality Using the Grb Scale. IEEE J. Sel. Top. Signal Process. 2019, 14, 413–422. [Google Scholar] [CrossRef]
  34. Fujimura, S.; Kojima, T.; Okanoue, Y.; Shoji, K.; Inoue, M.; Omori, K.; Hori, R. Classification of Voice Disorders Using a One-Dimensional Convolutional Neural Network. J. Voice 2020, 36, 15–20. [Google Scholar] [CrossRef]
  35. Verde, L.; De Pietro, G.; Sannino, G. Voice Disorder Identification by Using Machine Learning Techniques. IEEE Access 2018, 6, 16246–16255. [Google Scholar] [CrossRef]
  36. Hegde, S.; Shetty, S.; Rai, S.; Dodderi, T. A Survey on Machine Learning Approaches for Automatic Detection of Voice Disorders. J. Voice 2019, 33, 947.e11–947.e33. [Google Scholar] [CrossRef]
  37. Poburka, B.J.; Patel, R.R.; Bless, D.M. Voice-Vibratory Assessment with Laryngeal Imaging (VALI) Form: Reliability of Rating Stroboscopy and High-Speed Videoendoscopy. J. Voice 2017, 31, 513.e1–513.e14. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Workflow scheme of the data-learning process for acoustic vocal samples of patients with thyroid surgery using an artificial intelligence model. (A) Inclusion criteria and steps to train a pretrained model that extracts important features from vocal data. (B) The development of a deep learning model to predict the GRBAS score after three weeks from pre- and postoperative vocal samples with the pretrained model.
Figure 1. Workflow scheme of the data-learning process for acoustic vocal samples of patients with thyroid surgery using an artificial intelligence model. (A) Inclusion criteria and steps to train a pretrained model that extracts important features from vocal data. (B) The development of a deep learning model to predict the GRBAS score after three weeks from pre- and postoperative vocal samples with the pretrained model.
Sensors 22 06387 g001
Figure 2. The prediction results for the (A) grade, (B) breathiness, and (C) asthenia scores of the test set. The x-axis represents the observed GRBAS score of the patient, and the y-axis represents the predicted value by the deep learning model. The blue line is the regression line to see the relationship between the predicted value and the actual value.
Figure 2. The prediction results for the (A) grade, (B) breathiness, and (C) asthenia scores of the test set. The x-axis represents the observed GRBAS score of the patient, and the y-axis represents the predicted value by the deep learning model. The blue line is the regression line to see the relationship between the predicted value and the actual value.
Sensors 22 06387 g002
Figure 3. For the grade, breathiness, and asthenia scores, we divided the patient based on 0 or not, and the ROC was calculated.
Figure 3. For the grade, breathiness, and asthenia scores, we divided the patient based on 0 or not, and the ROC was calculated.
Sensors 22 06387 g003
Figure 4. ROC curve for vocal samples of patients 6 months after surgery.
Figure 4. ROC curve for vocal samples of patients 6 months after surgery.
Sensors 22 06387 g004
Figure 5. A spectrogram example and its visualization results using gradient-based localization to predict the grade score in second and third residual block of EfficientNet. (A) Spectrogram with heatmap visualization of a patient with high-degree grade score. (B) A patient with normal grade.
Figure 5. A spectrogram example and its visualization results using gradient-based localization to predict the grade score in second and third residual block of EfficientNet. (A) Spectrogram with heatmap visualization of a patient with high-degree grade score. (B) A patient with normal grade.
Sensors 22 06387 g005
Table 1. Grade score distribution.
Table 1. Grade score distribution.
GradePre op.Post op.3 Months Post op.
G0432531
G1616067
G292414
G3152
Table 2. Prediction performance for the GRBAS score.
Table 2. Prediction performance for the GRBAS score.
ClassRMSERhop Value
Grade0.3990.796<0.001
Roughness0.3650.1490.509
Breathiness0.4090.784<0.001
Asthenia0.4690.6020.003
Strain0.203NANA
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, J.H.; Lee, C.Y.; Eom, J.S.; Pak, M.; Jeong, H.S.; Son, H.Y. Predictions for Three-Month Postoperative Vocal Recovery after Thyroid Surgery from Spectrograms with Deep Neural Network. Sensors 2022, 22, 6387. https://doi.org/10.3390/s22176387

AMA Style

Lee JH, Lee CY, Eom JS, Pak M, Jeong HS, Son HY. Predictions for Three-Month Postoperative Vocal Recovery after Thyroid Surgery from Spectrograms with Deep Neural Network. Sensors. 2022; 22(17):6387. https://doi.org/10.3390/s22176387

Chicago/Turabian Style

Lee, Jeong Hoon, Chang Yoon Lee, Jin Seop Eom, Mingun Pak, Hee Seok Jeong, and Hee Young Son. 2022. "Predictions for Three-Month Postoperative Vocal Recovery after Thyroid Surgery from Spectrograms with Deep Neural Network" Sensors 22, no. 17: 6387. https://doi.org/10.3390/s22176387

APA Style

Lee, J. H., Lee, C. Y., Eom, J. S., Pak, M., Jeong, H. S., & Son, H. Y. (2022). Predictions for Three-Month Postoperative Vocal Recovery after Thyroid Surgery from Spectrograms with Deep Neural Network. Sensors, 22(17), 6387. https://doi.org/10.3390/s22176387

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop