Next Article in Journal
Development of Machine Learning Models Predicting Estimated Blood Loss during Liver Transplant Surgery
Next Article in Special Issue
Identification of Early Esophageal Cancer by Semantic Segmentation
Previous Article in Journal
Interferon-Inducible Protein-10 as a Marker to Detect Latent Tuberculosis Infection in Patients with Inflammatory Rheumatic Diseases
Previous Article in Special Issue
Transparent Quality Optimization for Machine Learning-Based Regression in Neurology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning to Detect Triangular Fibrocartilage Complex Injury in Wrist MRI: Retrospective Study with Internal and External Validation

1
Department of Orthopedics, Tri-Service General Hospital, National Defense Medical Center, No. 325, Sec. 2, Chenggong Rd., Neihu District, Taipei 11490, Taiwan
2
Department of Surgery, Tri-Service General Hospital Penghu Branch, National Defense Medical Center, Penghu 88056, Taiwan
3
Graduate Institute of Technology, Innovation and Intellectual Property Management, National Cheng Chi University, Taipei 11605, Taiwan
4
School of Public Health, National Defense Medical Center, Taipei 11490, Taiwan
5
Department of Animal Science, National Pingtung University of Science and Technology, Pingtung 91201, Taiwan
*
Author to whom correspondence should be addressed.
J. Pers. Med. 2022, 12(7), 1029; https://doi.org/10.3390/jpm12071029
Submission received: 16 May 2022 / Revised: 15 June 2022 / Accepted: 21 June 2022 / Published: 23 June 2022
(This article belongs to the Special Issue Artificial Intelligence Application in Health Care System)

Abstract

:
Objective: To use deep learning to predict the probability of triangular fibrocartilage complex (TFCC) injury in patients’ MRI scans. Methods: We retrospectively studied medical records over 11 years and 2 months (1 January 2009–29 February 2019), collecting 332 contrast-enhanced hand MRI scans showing TFCC injury (143 scans) or not (189 scans) from a general hospital. We employed two convolutional neural networks with the MRNet (Algorithm 1) and ResNet50 (Algorithm 2) framework for deep learning. Explainable artificial intelligence was used for heatmap analysis. We tested deep learning using an external dataset containing the MRI scans of 12 patients with TFCC injuries and 38 healthy subjects. Results: In the internal dataset, Algorithm 1 had an AUC of 0.809 (95% confidence interval—CI: 0.670–0.947) for TFCC injury detection as well as an accuracy, sensitivity, and specificity of 75.6% (95% CI: 0.613–0.858), 66.7% (95% CI: 0.438–0.837), and 81.5% (95% CI: 0.633–0.918), respectively, and an F1 score of 0.686. Algorithm 2 had an AUC of 0.871 (95% CI: 0.747–0.995) for TFCC injury detection and an accuracy, sensitivity, and specificity of 90.7% (95% CI: 0.787–0.962), 88.2% (95% CI: 0.664–0.966), and 92.3% (95% CI: 0.763–0.978), respectively, and an F1 score of 0.882. The accuracy, sensitivity, and specificity for radiologist 1 were 88.9, 94.4 and 85.2%, respectively, and for radiologist 2, they were 71.1, 100 and 51.9%, respectively. Conclusions: A modified MRNet framework enables the detection of TFCC injury and guides accurate diagnosis.

Graphical Abstract

1. Introduction

Triangular fibrocartilage complex (TFCC) is a vital structure in sports medicine, and its complete observation is difficult because it is situated in a narrow space between the lunate, triquetrum, and ulnar head [1,2,3,4]. According to Ng, A.W., et al., it is difficult to diagnose TFCC injury in MRI with its intrinsic small and thin structure. High-resolution MRI, preferably contrast-enhanced or sometimes with arthrogram, is an ideal detection tool. [5] Observer experience was also an important effect in the diagnosis of TFCC injury in Philip E.’s study, which showed that the difference in the overall accuracy rate for the prediction of a TFCC lesion and its location was 32%. [6] Nowadays, deep learning algorithms for lesion detection allow for the interpretation of medical images with a possibly higher precision and accuracy than the radiologists. [7] Artificial intelligence (AI), especially deep learning algorithms [8,9,10,11,12,13,14,15,16], can identify subtle signs and provide radiologists with more rapid and accurate diagnoses. [11,17,18] For instance, a deep ResNet model, which contains residual links as shortcuts connecting the input and output of a block of convolutional layers [19,20], has been used as the framework to detect the subtle anterior cruciate ligament injuries. [18,21,22] Similarly, an ensemble MRNet has been proposed to detect knee injuries such as cruciate ligaments and meniscal injuries from MRI orthopedic scans. [23] However, there have been no reports using the AI deep learning system to assist in the diagnosis of TFCC injuries on MRI.
As pioneer research, we developed two modified TFCC injury detection algorithms based on the ResNet and MRNet framework and explainable AI to learn MR images from 332 clinical cases and to detect the presence of TFCC injuries.

2. Materials and Methods

2.1. Study Design

We used the wrist MRI database maintained by the Medical Record Department of the general hospital to conduct a retrospective study involving 332 participants. This study was approved by the Institutional Review Board of our institution (name blinded for peer review). Given its retrospective nature, no patient participated in the design, implementation, and reports of this study, and therefore the requirement for informed consent was waived.

2.2. MRI Datasets and Radiologist Reports

From the 332 wrist MRI scans, 143 exhibited TFCC injuries, whereas 189 exhibited healthy hands. The 143 patients with TFCC injuries were diagnosed using arthroscopy following the corresponding MRI examinations. The MRI scans of healthy hands were obtained from patients undergoing routine health checks, and orthopedic surgeons confirmed the absence of TFCC injuries. We retrospectively studied MRI scans obtained over 11 years and 2 months between 1 January 2009 and 29 February 2019. Cases with and without TFCC injuries were randomly assigned to three datasets: 72.8% (n = 242) to the training set, 13.6% (n = 45) to the validation set, and 13.6% (n = 45) to the test set. The size of each dataset was determined based on the minimum requirements for yielding meaningful statistics during and after training. In the internal dataset, we used 13,377 images for the training set, 994 images for the validation set and 3681 images for the test set; in the external dataset, we used 3301 images for the test set (Figure 1).
To compare the differences between the diagnoses provided by AI algorithms and human physicians, we invited two diagnostic radiologists and used the evaluated algorithms to diagnose the MRI scans from the test sets. The task was to determine the presence of TFCC injuries. Among the cases diagnosed with TFCC injuries, we distinguished those where the radiologist detected and classified an injury correctly, detected an injury correctly but classified the injury incorrectly, failed to detect an existing injury, and correctly judged the absence of injury. These results were compared with the AI-generated results for the performance evaluation. The algorithms were evaluated using both the local and external test sets. The external test set was collected at TSGH Song-Shan Branch (Figure 1) and comprised the MRI scans of 12 patients with TFCC injuries and 38 subjects with healthy hands. The scans included coronal T1-weighted, coronal T2-weighted with fat saturation, sagittal proton-density-weighted, sagittal T2-weighted with fat saturation, and axial proton-density-weighted with fat saturation. All scans were acquired in a 3.0 T magnetic field at 3 mm intervals with 16–20 images per slice.

2.3. Preprocessing

For image preprocessing, model execution, and performance evaluation, we used Python (version 3.8) programs with the PyDICOM library (version 2.0.0) and OpenCV-Python (version 4.2.0.34) running on a computer equipped with a 2.10 GHz 64 core Intel-Xeon Gold-6130 processor with 276 GB DDR4-SDRAM, 4 T V100 32 GB graphical processing units, and executing a Linux system. The visual data extracted from the DICOM (Digital Imaging and Communications in Medicine) files measured 256 × 256 pixels and were converted into the PNG (portable network graphics) format. To ensure the validity of the analysis, we used a standardization algorithm to convert the pixel values into values between 0 and 255, which is the range of PNG images.

2.4. Algorithms and Model Architectures

We evaluated two framework models, ResNet50 and MRNet, using the PyTorch (version 1.8.0) and TorchVision (version 0.8.1) libraries. The code used in this study is available at https://github.com/akousist/tfcc-detection (accessed on 10 May 2022). Algorithm 1 was based on ResNet50 (Upper part of Figure 2A), which is essential in many applications relying on image classification. We used the ResNet50 algorithm (Figure 2B) to explain the model predictions. ResNet50 contains 16 residual blocks, each consisting of three convolutional layers. Each convolutional layer was followed by batch normalization, and ReLU (rectified linear unit) activation was applied between the convolutional layers and after summing the identity map from the residual link. A convolutional layer was added in front of the ResNet50 architecture to transform eight images along the z-axis into one 3-channel 2D image using a design known as 2.5D CNN to incorporate the 3D information. According to the principle of CNN, information can be exchanged between different channels. This is different from the Siamese network-style design of MRNet (described below), where images along the z-axis do not exchange information. The effect of integrating z-axis imagery can be integrated during model operation. After the z-cropping, the input tensor becomes the size of n × 8 × 256 × 256 (n is the batch size). This tensor passes through the z-compression network, compresses the z-direction to n × 3 × 256 × 256, conforming to the ResNet50 input form, before entering the main architecture of ResNet50. The output of ResNet50 (n × 1000) passes through the classifier to obtain a prediction of “with TFCC tears” or “without TFCC tears” (n × 2). We applied Adam optimization to accelerate the evolution of stochastic gradient descent. With weights pretrained on the ImageNet dataset (i.e., by applying transfer learning), we used a learning rate of 1 × 10−5 and batch size of 16 and trained the model for 20 epochs. A two-class cross-entropy function was employed as the loss function.
Algorithm 2 was derived from MRNet, an ensemble model for diagnosis using a 3D MRI series (lower part of Figure 2A). The MRNet architecture consists of three AlexNets, each of which processes MRI images in one direction (sagittal, coronal, axial). AlexNet was originally a model that processed a single 2D image (Figure 2C). For AlexNet to process the 3D imagery, the batch dimension of the original model was treated as a z-dimension using a Siamese network-like approach, using the same set of model weights to process “a single patient with a set of MRI images in the same scan direction”. Each set of MRI images is first divided into three groups of images according to the scanning direction, then each image will be resized, adjusted to 3 × 256 × 256 (3 means the RGB three color channels), and then each image is arranged in the z-direction into the tensor of s × 3 × 256 × 256 (s is the number of images in the z-direction of the group). The size of this tensor conforms to the input form of AlexNet, which is ready to enter AlexNet. The output of AlexNet becomes a tensor of s × 256. This tensor will enter a z-max pooling to integrate the information of each picture and obtains a 256-dimensional vector. This 256-dimensional vector passes through another classifier to obtain two predictions of “with TFCC tears” or “without TFCC tears”.

2.5. Statistical Analysis

Using a pair of validation and test sets, we evaluated the algorithm processing of the MRI scans and the training performance. We evaluated the performance of the final algorithm using the internal test set. To evaluate the model precision, we used the scikit-learn (version 0.19, available at scikit-learn.org) library to calculate the F1 scores and receiver operating characteristic (ROC) curve.
The precision was defined as the number of true positives divided by the number of false positives plus the number of true positives, while the recall was defined as the number of true positives divided by the number of true positives plus the number of false negatives. The F1 score is the harmonic mean of the precision and recall (F1 score = 2/((1/precision) + (1/recall))) and the ROC curve is the softmax function of the absence of a feature.

3. Results

3.1. Model Performance on the Internal and External Test Sets

The 332 MRI scans were obtained from the same number of subjects (mean age, 36.8 years; 215 men, 64.8%; 117 women, 35.2%). Overall, Algorithm 2 (MRNet framework) outperformed Algorithm 1 (ResNet50 framework) in the detection of TFCC injuries. In Figure 3, Algorithm 1 yielded an AUC of 0.809 (95% CI: 0.670, 0.947) for the TFCC injury detection, as depicted in the receiver operating characteristic curves (ROC curves).

3.2. Interpretation and Visualization

An integrated gradient was used after the last convolutional layer of the model, and the results were overlaid on the MRI scans to show the relevance of specific areas for TFCC injury detection. As shown in Figure 4, the algorithm focuses on the region of the contrast media for evaluation and classification. Hence, the Algorithm learned to evaluate the features of the contrast media instead of learning the complete MRI scan with TFCC injury. Figure 4 also shows examples of the integrated grading discrepancies that included MRI scans with poor quality from the original dataset, possibly leading to incorrect predictions.

4. Discussion

In our study, two algorithms, MRNet and ResNet50, were proposed to process unlabeled MRI images. Under the classification of only the image data with or without tears of the TFCC, the feature learning of the aforementioned two deep learning algorithms was used to analyze the internal dataset and verify the external dataset. Unlike other studies that spent a lot of time on labeling, this study is the first deep learning exploration on unlabeled MRI images of the TFCC tear.
A well-designed AI-based diagnostic algorithm can reduce human errors in medicine, allowing physicians to provide more precise diagnoses. However, no algorithm has been devised to provide a universal and comprehensive solution for MRI interpretation to date [24,25,26]. The use of deep learning for disease detection in MRI is currently less studied and can be particularly challenging because it often requires the analysis of different image datasets, complicating the process of analysis [17]. In several of Liu, et al.’s studies, labeling data with the CNN training model can improve the accuracy and efficiency of the discrimination between the cartilage and bone injury or cruciate ligament injury in the knee joint [27,28]. To reduce the time costs for labeling and increase the information in the MRI of the TFCC tear, which physicians could miss, we used unlabeled data with deep-learning algorithms to improve these problems.
We developed and evaluated two algorithms based on the ResNet50 and MRNet frameworks to detect TFCC injuries. The results showed that Algorithm 2 with the MRNet framework substantially outperformed Algorithm 1 with the ResNet50 framework. Specifically, Algorithm 2 had a higher AUC of 0.871 than Algorithm 1 (AUC of 0.809). Algorithm 2 was also superior to Algorithm 1 in terms of sensitivity (88.2 vs. 66.7%) and specificity (92.3 vs. 81.5%). These results are attributable to the operation of MRNet, which is based on an ensemble method combining models to solve a single prediction problem. This method of transforming the old model can properly use the existing pretrained weights on the 2D image, and in the case of limited MRI data, the best results can be achieved with the least training resources.
The MRNet architecture in this study was a trained 3D CNN with logistic regression to combine the results for the final prediction. Although training was performed separately and not in an end-to-end approach, the architecture suitably processed MRI scans comprising 3D data from three orientations. In the ResNet50 architecture, we first selected the plane with the most obvious signs for detection, that is, the coronal plane, and obtained eight core images, yielding eight images with resolution of 256 × 256 pixels. Then, we used eight as the feature dimension to treat the scan as a plane with eight RGB channels. By processing the scan through a convolutional layer, we converted it into the common 3 × 256 × 256 dimensions for ResNet-based prediction. We believe that prediction without first combining the algorithm results does not reflect the optimal solution. Therefore, the ResNet50 model can further be improved regarding accuracy.
We used explainable AI to build a glass box model that provides a visual outcome of the algorithm reasoning without resorting to self-interpretable models with less predictive power. Contrast media enhancement in the MRI datasets of TFCC injuries will be focused because of its saliency region. Thus, without labeling by physicians, the algorithm decision on TFCC injury was affected by the contrast media. In other words, the images showed sites of contrast media instead of those of TFCC injuries. Despite our objective being to investigate how explainable AI detects TFCC injury hotspots, hotspots of contrast media were identified. Although using explainable AI to detect TFCC injury hotspots in MR imaging did not yield obvious results, these results may guide future research, and new findings may be unveiled if explainable AI is used to directly detect TFCC injury hotspots in MRI data without the influence of contrast media.
Overall, the proposed algorithms showed high performance in detecting TFCC injuries, which can facilitate diagnosis. However, indicating the TFCC injury sites is challenging if the contrast media provide important contributions to diagnosis. The results indicate that radiologist interpretation is required for AI algorithms to conduct a detailed structural evaluation that identifies soft-tissue injuries in the presence of contrast media. Therefore, the existing deep learning algorithms should be further developed, and the effects of deep learning on diagnostic radiology should be explored in future work.
Some limitations of this study include the following: First, model training was based on the interpretation of radiologists. Although the interpretations of radiologists are the best reference for applications, their interpretations may vary. Second, although the deep learning algorithms considered 3D image evaluation, additional data and patient inquiry are required for a more complete TFCC injury evaluation in clinical practice. Third, the algorithms neglected potential abnormalities associated with hand TFCC such as avascular necrosis and bone abnormalities. Finally, image data obtained from other machine learning algorithms and images from additional external sources were not available, possibly undermining diagnostic performance.
To sum up, we verified the feasibility of applying two deep learning algorithms, the ResNet50 and MRNet frameworks, to detect hand TFCC injuries from features extracted from the MRI scans. The comparison between the performances of the two algorithms further suggests the advantages of the MRNet framework as it works well with 3D images and is competent in integrating the information from the three axes. Overall, we propose an innovative approach on TFCC diagnosis with deep learning algorithms, which we believe will benefit both the radiologists and the surgeons clinically.

5. Key Points

An MRNet framework enables the detection of a TFCC injury and guides accurate diagnosis.
The flow direction of the contrast media will affect the convergence of the heatmap.

Author Contributions

Conceptualization, K.-Y.L., Y.-T.L. and T.-T.Y.; methodology, J.-Y.H.; software, J.-Y.H.; validation, S.-Y.P. and C.-C.W.; formal analysis, C.-M.C.; investigation, K.-Y.L. and Y.-T.L.; resources, S.-Y.P., C.-C.W. and T.-T.Y.; data curation, K.-Y.L. and Y.-T.L.; writing—original draft preparation, K.-Y.L.; writing—review and editing, Y.-T.L.; visualization, C.-C.W.; supervision, T.-T.Y.; project administration, K.-Y.L. and T.-T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was approved by the Institutional Review Board of Tri-Service General Hospital (Approval no.: C202005150 and date of approval: 3 December 2020).

Informed Consent Statement

Informed consent was waived due to the retrospective nature of study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AI: Artificial Intelligence; CNNs: convolutional neural networks; MRI: magnetic resonance imaging; ROC: receiver operating characteristic; TFCC: triangular fibrocartilage complex.

References

  1. Brahim, A.; Jennane, R.; Riad, R.; Janvier, T.; Khedher, L.; Toumi, H.; Lespessailles, E. A decision support tool for early detection of knee OsteoArthritis using X-ray imaging and machine learning: Data from the OsteoArthritis Initiative. Comput. Med. Imaging Graph. 2019, 73, 11–18. [Google Scholar] [CrossRef] [PubMed]
  2. Üreten, K.; Arslan, T.; Gültekin, K.E.; Demir, A.N.D.; Özer, H.F.; Bilgili, Y. Detection of hip osteoarthritis by using plain pelvic radiographs with deep learning methods. Skeletal. Radiol. 2020, 49, 1369–1374. [Google Scholar] [CrossRef] [PubMed]
  3. Urakawa, T.; Tanaka, Y.; Goto, S.; Matsuzawa, H.; Watanabe, K.; Endo, N. Detecting intertrochanteric hip fractures with orthopedist-level accuracy using a deep convolutional neural network. Skeletal. Radiol. 2019, 48, 239–244. [Google Scholar] [CrossRef] [PubMed]
  4. Fong, R.C.; Vedaldi, A. Interpretable explanations of black boxes by meaningful perturbation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 3429–3437. [Google Scholar]
  5. Ng, A.W.H.; Griffith, J.F.; Fung, C.S.Y. MR imaging of the traumatic triangular fibrocartilaginous complex tear. Quant. Imaging Med. Surg. 2017, 7, 443–460. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Blazar, P.E.; Chan, P.S.; Kneeland, J.B. The effect of observer experience on magnetic resonance imaging interpretation and localization of triangular fibrocartilage complex lesions. J. Hand. Surg. Am. 2001, 26, 742–748. [Google Scholar] [CrossRef] [PubMed]
  7. Oren, O.; Gersh, B.J.; Bhatt, D.L. Artificial intelligence in medical imaging: Switching from radiographic pathological data to clinically meaningful endpoints. Lancet Digit. Health 2020, 2, e486–e488. [Google Scholar] [CrossRef]
  8. Von Schacky, C.E.; Sohn, J.; Foreman, S.C.; Liu, F.; Jungmann, P.M.; Nardo, L.; Ozhinsky, E.; Posadzy, M.; Nevitt, M.C.; Link, T.M.; et al. Development and performance comparison with radiologists of a multitask deep learning model for severity grading of hip osteoarthritis features on radiographs. Osteoarthr. Cartil. 2020, 28, S306–S308. [Google Scholar] [CrossRef]
  9. Ha, C.W.; Kim, S.H.; Lee, D.H.; Kim, H.; Park, Y.B. Predictive validity of radiographic signs of complete discoid lateral meniscus in children using machine learning techniques. J. Orthop. Res. 2020, 38, 1279–1288. [Google Scholar] [CrossRef]
  10. Namiri, N.K.; Flament, I.; Astuto, B.; Shah, R.; Tibrewala, R.; Caliva, F.; Link, T.M.; Pedoia, V.; Majumdar, S. Hierarchical severity staging of anterior cruciate ligament injuries using deep learning with MRI images. arXiv 2020, arXiv:2003.09089. [Google Scholar]
  11. Ambellan, F.; Tack, A.; Ehlke, M.; Zachow, S. Automated segmentation of knee bone and cartilage combining statistical shape knowledge and convolutional neural networks: Data from the Osteoarthritis Initiative. Med. Image Anal. 2019, 52, 109–118. [Google Scholar] [CrossRef]
  12. Zhou, Z.; Zhao, G.; Kijowski, R.; Liu, F. Deep convolutional neural network for segmentation of knee joint anatomy. Magn. Reson. Med. 2018, 80, 2759–2770. [Google Scholar] [CrossRef] [PubMed]
  13. Prasoon, A.; Petersen, K.; Igel, C.; Lauze, F.; Dam, E.; Nielsen, M. Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network. Med. Image Comput. Comput-Assist. Interv. 2013, 16, 246–253. [Google Scholar] [PubMed]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  15. Badgeley, M.A.; Zech, J.R.; Oakden-Rayner, L.; Glicksberg, B.S.; Liu, M.; Gale, W.; McConnell, M.V.; Percha, B.; Snyder, T.M.; Dudley, J.T. Deep learning predicts hip fracture using confounding patient and healthcare variables. NPJ Digit. Med. 2019, 2, 31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Zhou, Q.Q.; Wang, J.; Tang, W.; Zhang-Chun, H.; Zi-Yi, X.; Xue-Song, L.; Rongguo, Z.; Xindao, Y.; Bing, Z.; Zhang, H. Automatic detection and classification of rib fractures on thoracic CT using convolutional neural network: Accuracy and feasibility. Korean J. Radiol. 2020, 21, 869–879. [Google Scholar] [CrossRef] [PubMed]
  17. Pedoia, V.; Norman, B.; Mehany, S.N.; Bucknor, M.D.; Link, T.M.; Majumdar, S. 3D convolutional neural networks for detection and severity staging of meniscus and PFJ cartilage morphological degenerative changes in osteoarthritis and anterior cruciate ligament subjects. J. Magn. Reson. Imaging 2019, 49, 400–410. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, L.; Li, M.; Zhou, Y.; Lu, G.; Zhou, Q. Deep learning approach for anterior cruciate ligament lesion detection: Evaluation of diagnostic performance using arthroscopy as the reference standard. J. Magn. Reason. Imaging 2020, 52, 1745–1752. [Google Scholar] [CrossRef]
  19. Weikert, T.; Noordtzij, L.A.; Bremerich, J.; Stieltjes, B.; Parmar, V.; Cyriac, J.; Sommer, G.; Sauter, A.W. Assessment of a deep learning algorithm for the detection of rib fractures on whole-body trauma computed tomography. Korean J. Radiol. 2020, 21, 891–899. [Google Scholar] [CrossRef]
  20. Deniz, C.M.; Xiang, S.; Hallyburton, R.S.; Welbeck, A.; Babb, J.S.; Honig, S.; Cho, K.; Chang, G. Segmentation of the proximal femur from MR images using deep convolutional neural networks. Sci. Rep. 2018, 8, 16485. [Google Scholar]
  21. Tiulpin, A.; Thevenot, J.; Rahtu, E.; Lehenkari, P.; Saarakkala, S. Automatic knee osteoarthritis diagnosis from plain radiographs: A deep learning-based approach. Sci. Rep. 2018, 8, 1727. [Google Scholar] [CrossRef]
  22. Chang, P.D.; Wong, T.T.; Rasiej, M.J. Deep learning for detection of complete anterior cruciate ligament tear. J. Digit. Imaging 2019, 32, 980–986. [Google Scholar] [CrossRef]
  23. Bien, N.; Rajpurkar, P.; Ball, R.; Irvin, J.; Park, A.; Jones, E.; Bereket, M.; Patel, B.N.; Yeom, K.W.; Shpanskaya, K.; et al. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet. PloS Med. 2018, 15, e1002699. [Google Scholar] [CrossRef] [PubMed]
  24. Lindsey, R.; Daluiski, A.; Chopra, S.; Lachapelle, A.; Mozer, M.; Sicular, S.; Hanel, D.; Gardner, M.; Gupta, A.; Hotchkiss, R.; et al. Deep neural network improves fracture detection by clinicians. Proc. Natl. Acad. Sci. USA 2018, 115, 11591–11596. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Miller, D.D.; Brown, E.W. Artificial intelligence in medical practice: The question to the answer? Am. J. Med. 2018, 131, 129–133. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, F.; Zhou, Z.; Samsonov, A.; Blankenbaker, D.; Larison, W.; Kanarek, A.; Lian, K.; Kambhampati, S.; Kijowski, R. Deep learning approach for evaluating knee MR Images: Achieving high diagnostic performance for cartilage lesion detection. Radiology 2018, 289, 160–169. [Google Scholar] [CrossRef]
  27. Liu, F.; Zhou, Z.; Jang, H.; Samsonov, A.; Zhao, G.; Kijowski, R. Deep convolutional neural network and 3D deformable approach for tissue segmentation in musculoskeletal magnetic resonance imaging. Magn. Reson. Med. 2018, 79, 2379–2391. [Google Scholar] [CrossRef]
  28. Liu, F.; Guan, B.; Zhou, Z.; Samsonov, A.; Rosas, H.; Lian, K.; Sharma, R.; Kanarek, A.; Kim, J.; Guermazi, A.; et al. Fully Automated Diagnosis of Anterior Cruciate Ligament Tears on Knee MR Images by Using Deep Learning. Radiol. Artif. Intell. 2019, 1, 180091. [Google Scholar] [CrossRef]
Figure 1. The local and external datasets for model training, validation, and testing.
Figure 1. The local and external datasets for model training, validation, and testing.
Jpm 12 01029 g001
Figure 2. (A) An overview of both algorithms used in this study. The topmost data flow shows how Algorithm 1 (ResNet50) processes images along the coronal axis. ResNet50 itself was originally a model for processing 2D images. For ResNet50 to integrate the information from the z-axis, the MR images would first go through the z-direction cropping, which selects the middle eight of a set of MR images. Z-compression then transforms the eight images into three channels. In the ResNet50 architecture, this study uses the model’s channel dimension as the z-dimension. Other data flows show how Algorithm 2 (MRNet) processes images along three axes. Training of MRNet is two-stage. Three AlexNets are optimized respectively, indicated by the three log-likelihood maximization. In the second stage training, their outputs are then passed to a logistic regression classifier for an ensemble result. In the design for the MRNet, where the backbone AlexNet was also designed for 2D images, batch dimension was used for the z-dimension, and information along the z-axis was integrated with z-max pooling. (B)The architecture of ResNet50, the backbone of Algorithm 1 in this study. ResNet50 features residual links, indicated by the jumping arrows to the right of the layer stacks, facilitating the passing of information into a very deep network. Each convolution layer (colored block) is followed by a batch normalization and ReLU activation function, which are not shown in this Figure 2. (C) The architecture of AlexNet, the main backbone of MRNet in this study. The feature extraction of AlexNet consists of five convolution layers and three max-pooling layers, in the depicted order. In the original AlexNet, feature extraction ends with global average pooling to 256 × 6 × 6 and the output tensor were flattened to a 9216-dimensional vector for the downstream classifier. Here, we followed the MRNet’s design and average-pool the output to 256 × 1 × 1 before entering the fully-connected network.
Figure 2. (A) An overview of both algorithms used in this study. The topmost data flow shows how Algorithm 1 (ResNet50) processes images along the coronal axis. ResNet50 itself was originally a model for processing 2D images. For ResNet50 to integrate the information from the z-axis, the MR images would first go through the z-direction cropping, which selects the middle eight of a set of MR images. Z-compression then transforms the eight images into three channels. In the ResNet50 architecture, this study uses the model’s channel dimension as the z-dimension. Other data flows show how Algorithm 2 (MRNet) processes images along three axes. Training of MRNet is two-stage. Three AlexNets are optimized respectively, indicated by the three log-likelihood maximization. In the second stage training, their outputs are then passed to a logistic regression classifier for an ensemble result. In the design for the MRNet, where the backbone AlexNet was also designed for 2D images, batch dimension was used for the z-dimension, and information along the z-axis was integrated with z-max pooling. (B)The architecture of ResNet50, the backbone of Algorithm 1 in this study. ResNet50 features residual links, indicated by the jumping arrows to the right of the layer stacks, facilitating the passing of information into a very deep network. Each convolution layer (colored block) is followed by a batch normalization and ReLU activation function, which are not shown in this Figure 2. (C) The architecture of AlexNet, the main backbone of MRNet in this study. The feature extraction of AlexNet consists of five convolution layers and three max-pooling layers, in the depicted order. In the original AlexNet, feature extraction ends with global average pooling to 256 × 6 × 6 and the output tensor were flattened to a 9216-dimensional vector for the downstream classifier. Here, we followed the MRNet’s design and average-pool the output to 256 × 1 × 1 before entering the fully-connected network.
Jpm 12 01029 g002
Figure 3. The ROC curves of algorithms (a) 1 and (b) 2 for the internal dataset. The ROC curves of algorithms (c) 1 and (d) 2 for the external dataset. ROC, receiver operating characteristic.
Figure 3. The ROC curves of algorithms (a) 1 and (b) 2 for the internal dataset. The ROC curves of algorithms (c) 1 and (d) 2 for the external dataset. ROC, receiver operating characteristic.
Jpm 12 01029 g003
Figure 4. The activation heatmaps overlaid on the original MRI scans. The white arrows indicate the overlapping area of the contrast media and heatmaps.
Figure 4. The activation heatmaps overlaid on the original MRI scans. The white arrows indicate the overlapping area of the contrast media and heatmaps.
Jpm 12 01029 g004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, K.-Y.; Li, Y.-T.; Han, J.-Y.; Wu, C.-C.; Chu, C.-M.; Peng, S.-Y.; Yeh, T.-T. Deep Learning to Detect Triangular Fibrocartilage Complex Injury in Wrist MRI: Retrospective Study with Internal and External Validation. J. Pers. Med. 2022, 12, 1029. https://doi.org/10.3390/jpm12071029

AMA Style

Lin K-Y, Li Y-T, Han J-Y, Wu C-C, Chu C-M, Peng S-Y, Yeh T-T. Deep Learning to Detect Triangular Fibrocartilage Complex Injury in Wrist MRI: Retrospective Study with Internal and External Validation. Journal of Personalized Medicine. 2022; 12(7):1029. https://doi.org/10.3390/jpm12071029

Chicago/Turabian Style

Lin, Kun-Yi, Yuan-Ta Li, Juin-Yi Han, Chia-Chun Wu, Chi-Min Chu, Shao-Yu Peng, and Tsu-Te Yeh. 2022. "Deep Learning to Detect Triangular Fibrocartilage Complex Injury in Wrist MRI: Retrospective Study with Internal and External Validation" Journal of Personalized Medicine 12, no. 7: 1029. https://doi.org/10.3390/jpm12071029

APA Style

Lin, K. -Y., Li, Y. -T., Han, J. -Y., Wu, C. -C., Chu, C. -M., Peng, S. -Y., & Yeh, T. -T. (2022). Deep Learning to Detect Triangular Fibrocartilage Complex Injury in Wrist MRI: Retrospective Study with Internal and External Validation. Journal of Personalized Medicine, 12(7), 1029. https://doi.org/10.3390/jpm12071029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop