Next Article in Journal
Enabling Green Crowdsourced Social Delivery Networks in Urban Communities
Next Article in Special Issue
Untangling Computer-Aided Diagnostic System for Screening Diabetic Retinopathy Based on Deep Learning Techniques
Previous Article in Journal
Deep-Learning-Based Resource Allocation for Time-Sensitive Device-to-Device Networks
Previous Article in Special Issue
Mammography Image-Based Diagnosis of Breast Cancer Using Machine Learning: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears

by
Mazhar Javed Awan
1,2,*,
Mohd Shafry Mohd Rahim
1,
Naomie Salim
1,
Amjad Rehman
3 and
Begonya Garcia-Zapirain
4,*
1
Faculty of Engineering, School of Computing, Universiti Teknologi Malaysia (UTM), Skudai 81310, Johor, Malaysia
2
Department of Software Engineering, University of Management and Technology, Lahore 54770, Pakistan
3
Artificial Intelligence and Data Analytics Laboratory, College of Computer and Information Sciences (CCIS), Prince Sultan University, Riyadh 11586, Saudi Arabia
4
eVIDA Lab, University of Deusto, 48007 Bilbao, Spain
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(4), 1552; https://doi.org/10.3390/s22041552
Submission received: 20 January 2022 / Revised: 12 February 2022 / Accepted: 14 February 2022 / Published: 17 February 2022
(This article belongs to the Special Issue Machine Learning and AI for Medical Data Analysis)

Abstract

:
The anterior cruciate ligament (ACL) is one of the main stabilizer parts of the knee. ACL injury leads to causes of osteoarthritis risk. ACL rupture is common in the young athletic population. Accurate segmentation at an early stage can improve the analysis and classification of anterior cruciate ligaments tears. This study automatically segmented the anterior cruciate ligament (ACL) tears from magnetic resonance imaging through deep learning. The knee mask was generated on the original Magnetic Resonance (MR) images to apply a semantic segmentation technique with convolutional neural network architecture U-Net. The proposed segmentation method was measured by accuracy, intersection over union (IoU), dice similarity coefficient (DSC), precision, recall and F1-score of 98.4%, 99.0%, 99.4%, 99.6%, 99.6% and 99.6% on 11451 training images, whereas on the validation images of 3817 was, respectively, 97.7%, 93.8%,96.8%, 96.5%, 97.3% and 96.9%. We also provide dice loss of training and test datasets that have remained 0.005 and 0.031, respectively. The experimental results show that the ACL segmentation on JPEG MRI images with U-Nets achieves accuracy that outperforms the human segmentation. The strategy has promising potential applications in medical image analytics for the segmentation of knee ACL tears for MR images.

1. Introduction

A well-functionally healthy knee is essential for mobility. One can understand the intricacies of the movement of the knee by getting familiar with the joints and the bones [1]. The knee is a complex joint of different structures, including bones, tendons, ligaments, and muscles [2]. Ligaments are strong bands of tissues that connect one bone to another. The anterior cruciate ligament (ACL), one of two ligaments that cross in the middle of the knee, connects the thigh bone (femur) to the shinbone (tibia) and helps stabilize your knee joint [3]. The ACL tear ruptured causes ACL injuries, osteoporosis, and osteoarthritis, common among spots players who have to execute start–stop [4]. Osteoporosis (OP) is a condition that leads to weakened bones, causing them to break more efficiently [5]. Osteoarthritis is a prevalent long-term condition affecting hundreds of millions of people worldwide. While it can happen at any age, it most commonly develops after fifty. In osteoarthritis, the cartilage wears down over time, and the joint surface becomes rougher, leading to inflammation. It most commonly affects the knee, hips, hands, feet, and lower back [6,7].
The knee joint divides into four main parts, (1) knee bones are involved in the femur, tibia, patella, and fibula. (2) Knee cartilage is the tissue that protects and shields the bones in the joint. (3) Knee ligaments are bands of tough, flexible, fibrous tissues that connect the bones’ ends and hold joints together, and the (4) knee patellar tendon connects the patella kneecap to the tibia [8]. The knee joint anatomy taxonomy [9] is as shown in Figure 1.
ACL injuries can lead to developing osteoarthritis of the knee irrespective of treatment. The ACL injuries divide into three grades [10]. The grade I tear is a short-term symptoms sprain or a slight ligament stretch. The grade III tear is a complete rupture that usually requires surgical reconstruction. Grade II is a partial ACL tear or an incomplete tear or injury to the ACL, one of the major ligaments in the knee [11]. The partial ACL injuries occur during sports involving sudden stops or direction changes [12,13]. Four ACL tears in the knee stabilize the knee joint, as shown in Figure 2 [14].
The 3D MRI is a non-invasive and non-ionizing radiation method of studying the soft-tissue structures for ACL tears. It is complicated to detect magnetic resonance imaging (MRI) to determine about the partial tear ligament is damaged. It has very good spatial interpose, submillimeter, and spatial resolution [16].
Image Segmentation is the image processing technique. It provides information about the various regions of interest in the image [17]. It involves classifying each pixel into one or more classes [18,19]. The grouping of pixels together is based on specific characteristics. Image segmentation aims to partition an image into meaningful segments that can be used mainly for medical and other perspectives [20,21]. The elements can often lead to a different type of image segmentation, which can divide into object detection [22] and semantic segmentation. Semantic segmentation refers to the process of linking each pixel [23,24].
An expert radiologist performs the segmentation task manually, but it can be very time-consuming and have variability among radiologists. Moreover, manual delineation is not exactly possible [25]. Segmentation is the critical step for a medical pipeline. It is essential to diagnose, monitor, and treat a patient.
Recently many deep learning approaches have been proposed to solve the problem of segmentation through radiology images in the medical field. The U-Net [26] is a convolutional neural network that was initially designed for semantic segmentation for medical images [27,28].
Although many models have been proposed and implemented, more accurate segmentation approaches are still required for anterior cruciate ligament tears. Therefore, our work aims to present a deep learning framework for MR images mask generation of anterior cruciate ligament tear and segmentation automatically. The modified U-Net was performed results of segmentation above 96% dice coefficient score. The key contributions of this work are abridged in the following key points:
  • Make the first effort to prepare pickle MR images into mask and JPEG in the study for segmentation purposes.
  • Develop a U-Net CNN architecture after adjusting hyperparameters to ensure the successful segmentation of ACL tears.
  • Extensive experiments were performed to calculate scores of accuracy, intersection over union, dice coefficient, precision, recall, F1-score scores and also evaluated through accuracy and dice coefficient loss metrics on training and test values.
  • The predicted segment images could classify efficient detection for ACL injury cases.
This paper is organized as follows: In Section 2, recent research related to the segmentation of knee bone diseases regarding MRI is discussed. In Section 3, the dataset and methodology of the proposed U-Net CNN model. In Section 4, some of the experiments result from our segmentation network. The benchmark studies and comparisons with limitations were discussed in Section 5. The last Section 6 conclusion is drawn.

2. Research Background

This section delineates several works relevant to our research of semantic segmentation other than ACL, deep learning methods of knee MRI segmentation on various components. The machine learning algorithms were performed better for classification of various diseases instead of segmentation [29,30,31,32,33]. The recent studies have incorporated U-Net CNN architecture for above-knee ACL segmentation.
Prasoon [34] studied segmented knee cartilage with three CNNs on 120,000 training voxels. The method was used 2D features of CNN for each voxel of three planes xy, yz and zx. The performance was measured with dice similarity coefficient, accuracy, sensitivity, and specificity on 114 unseen test data, respectively, 0.8249%, 99.93%, 81.92%, and 99.97%. The dataset of training voxels was limited and did not use the stage layer to the pre-trained CNN model. The advantage of this model was that it trained in a single process of three CNNs.
Deniz et al.’s [35] study was trained on two different architectures; the first was 2D CNN, so-called U-Net [26], that extracted trabecular bone probability map in the proximal femur. The second was 3D CNN into three-dimensional for volumetric segmentation with dilation. The purpose of dilated convolutions was to enlarge the image. The output of the model was performed precision, recall, and dice-similarity score 0.95 after four-fold cross-validation to obtain proximal femur. The AUC value was 0.998 in the case of 2D CNN u-net and 3D CNN with dilated feature maps of 6, 32 with 4 layers. Moreover, the study had institutional review board approval and written informed consent was obtained from 86 subjects of 48 coronal slices for segmentation tasks covering the proximal femur. However, this study was time-consuming, manually segmented by an expert of 86 subjects.
Zhou et al.’s [36] hybrid method was trained on deep CNN based on Convolutional encoder–decoder (CED), conditional random field (CRF), and deformable modeling on musculoskeletal tissue segmentation. CRF aims to identify the contextual relationship among voxels within the same and different classes of tissues. The purpose of deformation was used to refine the output from the 3D CRF. The deep convolutional encoder network pipeline was designed on VGG16 convolutional layers. The encoder network for performing rapid and accurate comprehensive tissue segmentation of the knee joint. The dataset taken from segmentation of knee images 2010 (SK110) image segmentation challenges hosted by the MICCAI [37] training on 19 subjects and 13 classes. The performance was measured by a mean dice coefficient between 0.7 to 0.9. The results performed better in the case of femur and tibia due to the large volume shape. The advantage was less sensitive against the overfitting data. However, the study’s limitations were less accurate than those found in the case of small structures like a meniscus.
Ambellan et al.’s [38] combined approach incorporated statistical shape models (SSMs) [39] with 2D CNN and 3D CNNs to segment femoral bone (FB) and tibial bone (TB) via voting way. The 2D CNN variant of U-Net created segmentation masks of FB and TB, SSMs. Furthermore, fitted on these masks, refinement with 3D CNN to segment MRI taken from SSMs, post-processing SSM to enhance the result of CNN regions on knee FC and TC and the last step applied 3D CNN for segmentation in TC and FC. The time of training was reduced by a factor of 6. The approach was trained on sagittal MRI cartilage subjects of 88 Osteoarthritis Initiative (OAI) Imorphics [40], 507 OAI ZIB [41], and 60 SK110. The accuracy of the model was evaluated on 40 validations and 50 submission datasets of SK110. The DSC scores on the OAI Imorphics baseline dataset performed were 89.4%, 86.1%, 90.4% bones cartilages of femur cartilage, medial tibial cartilage, and lateral tibial cartilage, respectively. The OAI ZIB dataset was shown DSC 98.6% for FB, 98.5% for TB, 89.9% for FC, and 85.6% for TC. However, the total score of DSC has remained 75.73% for SK110 dataset. The processing time will take 43 weeks in a single node when the data set becomes greater than 50,000.
Xu and Niethammer [42] study was used DeepAtlas for jointly deep CNN networks for weakly supervised image registration on unlabeled MR images via an anatomy similarity loss and customized light 3D U-Net. The experimental results were tested on 3D knee MRIs from the Osteoarthritis Initiative (OAI) dataset and 3D brain MRI. The approach was beneficial in case of a lack of manual segmentations. The segmentation of Semi-DeepAtlas DSC was showed 96.80% for bones (femur and tibia) and 77.63% in the case of cartilages (femoral and tibial). The limitation of the study was not investigated multitask.
Burton et al.’s [43] study was trained through 2D, and 3D CNN models for automatic segmentation related to biomechanics orthopedics of 36 labeled MRI a statistical shape model of the knee [44]. The classes for the annotated scan were associated with the femur, femur cartilage, patella, patella cartilage, tibia, and tibial cartilages of 20 males and 16 female subjects. The 51 unlabeled MRIs from the OAI dataset [45] used U-Net architecture.
Liu et al.’s [46] study was applied on cartilage tissues of the femur, tibia, femoral cartilage, and tibial to segment with a 2D convolutional encoder network of a Visual Geometry Group 16 (VGG16) architecture [47,48]. The datasets consisted of 175 patients of MRI cartilage lesion within 17395 image slices. The segmentation results were performed through mean Dice confidence femur, tibia, femoral cartilage, and tibial cartilage, respectively, 0.96, 0.95, 0.81 and 0.82. The average training was taken 20 s for a patient on all image sections with a trained network of cartilage segmentation. However, in the case of the segmentation network, it took 6.2 h in each fold of image data sets. However, one-time training of cartilage classification was a long process. However, the training time of the entire network for diagnostics was rapid. The study has some limitations: the study did not evaluate segmentation and classification on patellar cartilages, the reference standard for presence and absence of articular cartilage was not evaluated by arthroscopy, the study would not be possible to assess through arthroscopy because the exact location of cartilage is not described correctly. The clinical detection system can be applied to the test set for cartilage in the future.
Tack et al.’s [49] study applied convolutional neural network 2D and 3D U-Net in combination with statistical shape models (SSM) [50] of menisci sub-volumes has image the dimensions (384 × 384), (48 × 72 × 16), respectively, on sagittal MRI. These networks were trained on IGS datasets provided by Imorphics on two-fold cross-validation. The segmentation was improved by principal component analysis on three meniscus, lateral, medical, and combined. The segmentation results with dice similarity coefficient were 83.8% for medial menisci (MM) and 88.9%. The study’s limitation was that coronal DESS MRI accuracy did not increase, and cartilage denudation decreased tibia and meniscus extrusion.
Raj’s [51] study proposed a µ-Net based on U-Net 3D-CNN architecture for knee cartilage to segment high resolution 100 3D MR volumes on a MICCAI SKI10 [37] dataset after re-sampled the labels to 128 × 128 × 64 voxels. The model was also validated on the Osteoarthritis Initiative (OAI) datasets [52] of 176 knee 3D MR volumes of 88 patients. The 3D-DESS MR Slice has generated four segmentations by µ-Net femoral cartilage, patellar cartilage, meniscus, and tibial cartilage. The dice score of femoral cartilage and tibial cartilage was 0.834 and 0,825, respectively. The patellar cartilage was a lower score of DSC due to the absence of ground truth.
Pedoia et al.’s [53] study was used 2D U-Net for the cartilage and meniscus segmentation of 11 classes. The exact approach 3D U-Net CNN architecture was used to detect severity staging patellar and meniscal cartilage bounding box in a cascaded manner. The study was taken using 3D FSE CUBE knee MRI of 1481 samples, OA patients of 173 and after anterior cruciate ligament injury 129. The average cropped meniscus volume region was 39 × 79 × 44 voxels out of 5912 meniscus volume of interest. The ROC for binary meniscus lesion of training, validation, and testing was 0.95, 0.84 and 0.89, respectively, whereas in the case of binary lesion, it was 0.99, 0.89, and 0.88 with a ratio of 65, 20, and 15. However, the specificity of meniscus lesion and cartilage was 81.98% and 82.27%, respectively. The limitation of the study was the lack of image annotation uncertainty and the lack of the actual gold standard.
Norman et al. [54] automatic segmentation method was end-to-end, with no extensive pipeline for image registration. The U-Net CNN architecture was applied on 638 MR imaging volumes with a weighted cross-entropy function to handle the class imbalance. The authors used two different datasets, included patients with ACL injuries, patients with OA, (a) the 464 3.0 T MRI T1p-weighted (b) 174 3D double-echo steady-state (DESS) that were taken from OAI. The automatic segmentation performance was calculated by dice coefficient scores for both datasets of six knee MR parts FC, LTC, MTC, PC, LM, and MM. The DSC range remained 0.632 to 0.699 only for the case of T1p-weighted on validation data with a processing time of 2.5 s. In the case of DESS, the DSC range was higher at 0.878 to 0.753 with a lower computational time of 8 s. The study’s limitation was a lack of ground truth, and that it did not segment more parts of the meniscus and cartilages.
Pedoia, Norman, Mehany, Bucknor, Link and Majumdar’s [53] study used 2D U-Net for the cartilage and meniscus segmentation of 11 classes. The same approach 3D U-Net CNN architecture was used to detect severity staging patellar and meniscal cartilage bounding box in a cascaded manner. The study was taken using 3D FSE CUBE knee MRI of 1481 samples, OA patients of 173 and after anterior cruciate ligament injury 129. The average cropped meniscus volume region was 39 × 79 × 44 voxels out of 5912 meniscus volume of interest. The ROC for binary meniscus lesion of training, validation, and testing was 0.95, 0.84 and 0.89, respectively, whereas in the case of binary lesion, it was 0.99, 0.89, and 0.88 with a ratio of 65, 20, and 15. However, the specificity of meniscus lesion and cartilage was 81.98% and 82.27%, respectively. The limitation of the study was the lack of image annotation uncertainty and the lack of an actual gold standard from human interpret images.
Norman, Pedoia and Majumdar’s [54] automatic segmentation method was end-to-end, with no extensive pipeline for image registration. The U-Net CNN architecture was applied on 638 MR imaging volumes with a weighted cross-entropy function to handle the class imbalance. The authors used two different datasets, included patients with ACL injuries, patients with OA, (a) the 464 3.0 T MRI T1p-weighted (b) 174 3D double-echo steady-state (DESS) that were taken from OAI. The datasets were divided into training, validation, and time-point testing into the ratio of 70:20:10. The automatic segmentation performance was calculated by dice coefficient scores for both datasets of six knee MR parts FC, LTC, MTC, PC, LM, and MM. The DSC range has remained 0.632 to 0.699 only for the case of T1p-weighted on validation data with the processing time of 2.5 s. In the case of DESS, the DSC range was higher at 0.878 to 0.753 with the lower computational time of 8 s. The study’s limitation was a lack of ground truth, and that it did not segment more parts of the meniscus and cartilages.
Flannery et al.’s [55] study was an automated segmentation method MR imaging data total of 246 sagittal Constructive Interference in Steady State (CISS) scans BEAR I and BEAR II of intact ACLs. The U-net CNN architecture was a configured segmentation by symmetric down-sampling and depth 5, kernel size 5 × 5, and batch normalization. The model’s performance was evaluated on 29 samples through Dice coefficient, precision, and sensitivity of 0.84, 0.82, and 0.85, respectively. The limitations of the study were that the U-Net model was not trained on low-resolution MRI sequences and other sequences. Furthermore, the model has not used transfer learning to segment ACL reconstruction and repairs.
Flannery et al.’s [56] study was an automatic segmentation quantitatively of reconstructed anterior cruciate ligament and graft after surgery MRIs. The U-Net 2D CNN transfer learning approach was used on the large dataset named Bridge Enhanced ACL Repair (Bear I and Bear II) [57,58,59] of 76 subjects and 45 ACLR issues. Firstly, the base U-Net model was trained on 3480 sagittal slices out of 4920 for segmenting intact ACL and evaluated results through Dice coefficient, sensitivity, and precision of 0.84, 0.85 and 0.82, respectively. Secondly, the transfer learning approach, frozen four layers on BEAR Dataset, performed a lower dice coefficient, sensitivity, and precision of 0.80, 0.82, and 0.79. Thirdly, the ACL graft segmentation results were 0.78, 0.80, and 0.78 on the same evaluation metrics after 200 sagittal slices for testing out of 2400. Besides performance gain after applying transfer learning in both ACL repair and graft segmentation cases, this study was limited only to a MR CISS sequence on a single make scanner.
Almajalid et al.’s [60] used a customized U-Net model to identify the knee bone segmentation of tibia, femur, and patella bones. The Imorphics dataset OAI of 99 knee MRI cases with 160 2D slices was used. The study was trained on only 69 patients, a small dataset with modification of U-net with adam optimizer and softmax activation function. The segmentation result achieved a dice coefficient of 96.94% and a similarity index of 93.98% on testing 15 cases. The study has the limitation that data labeling was time-consuming for segmentation tasks and evaluated on MRI DESS sequences only.
In this section, some of the research studies are introduced that were conducted to segment knee various parts using deep learning. It is clear from the above literature that the accuracy, similarity, and dice coefficient score obtained in individual research work is not currently satisfactory. Most of the studies state the limited number of samples for segmentation and lack of ACL tear segmentation.

3. Materials and Methods

This section presents the methods and materials used in this study. Section 3.1 is about the three classes datasets of MR images description used in the proposed method. The proposed CNN is presented in Section 3.2.

3.1. DataSet

The original database used in this study was acquired from the knee 917 ACL samples of MRI Stajduhar et al. [61], sponsored by the clinical hospital center in Rijeka, Croatia. The patient age and gender was anonymous in the dataset. The three classes consisted of 690 healthy, 172 normal and 55 complete rupture tears of MR images. The original image size was 330 × 330 × 32 respectively, in width, height and depth. Each slice ACL was manually segmented and served as the ground truth.

3.2. Proposed Segmentation Framework

Our proposed segmentation framework is composed of three main phases as follows:
  • Phase 1: The pickle ACL MR images were converted into JPEG using an algorithm described in Section 3.2.1.
  • Phase 2: The knee mask or ground truth images generation and Javascript object notations (JSON) file creation process were explained in Section 3.2.2.
  • Phase 3: Section 3.2.3 was explained with the proposed model based on U-Net CNN.
The detailed framework of our segmentation is shown in Figure 3.

3.2.1. Data Preparation Conversion of JPEG Images

The dataset is in pickle format and such format is encrypted. The short name of the pickle file, which is encountered while reading its format. A pickle file is a unique format. It can hold a minimum of 8 images and a maximum of 32 shots. The method of decrypting pickle files is called un-pickling, so it is used for decryption techniques. After un-pickling images, they are converted into grayscale JPEG format. The process of turning pickle files into JPEG files by python script of pickling algorithm is described in Table 1.

3.2.2. Data Preparation Conversion of Knee Masking

After unpickling all the pickle files into JPEG, we have to label the images into masking with the classes given in the CSV metafile. For this, each JPEG image was loaded into VGG image annotator (VIA) standalone manual annotation software [62]. For labeling the ROI area, the ROI script file was used to identify which particular image contained the data. The VIA tool marked out the labeling box of the JPEG. It created the knee mask and allotted their class labels into JSON (JavaScript Object Notation) format file. The sample jpeg images and knee mask images dimensions are 320 × 320 × 3, shown in Figure 4. The total number of knee JPEG and knee mask images was 15,268 slices.

3.2.3. Our U-Net Convolutional Neural Network Architecture

U-Net is the auto-encoder based is convolutional neural architecture for fast and precise segmentation of images. The Architecture consists of a contracting path and an expensive path [26]. The U-Net combines the context information from the downsampling path with the localization information. There is no dense layer, so images of different sizes can be used as input. The use of massive data augmentation is important in domain-like biomedical segmentation. The three main stages are explained below:
  • Contracting/downsampling path
The contracting path is composed of 4 blocks. Each block is composed of two blocks of 3 × 3 convolutional. After this, convolutional 2 × 2 max pooling is applied.
  • Bottleneck
There is a bottleneck layer between the contracting and expanding path. The bottleneck layer is simply 2 convolutional layers.
  • Expanding/ upsampling path
The expanding layer is the decoder layer. The expanding path is also composed of 4 blocks. Each of these blocks is taken of deconvolutional layer with stride 2. The corresponding cropped feature map is concatenated from the contracting path. At the end, again two blocks of 3 × 3 convolutional, Relu activation function with batch normalization are used.
All of the JPEG and knee mask images were resized into 128 × 128 × 3 before being entered into the input layer, and the filters were fed 64, 128, 256, 512 and 1024, respectively. The first step of two convolution layers (Conv 1) output and kernel size was 128 × 128 × 64, 3 × 3, respectively, with rectified linear unit (ReLU). The padding same represent extra pixels on the edges and output image is the same as input. The next step was the max pooling with stride 2 × 2, the dimensions become half 64 × 64 × 64 of 128 by 128 as used 2 × 2 kernel. The two convolutional layers (Conv 2) output dimensions 64 × 64 × 128 where 128 filter. Repeat this process until bottleneck sage. At this stage, the pooling layer (Pool 4) dimension was 8 × 8 × 512, and on the right side 8 × 8 × 1024. The upsampling (UP6) again used stride 2 × 2 generated the dimension of 16 × 16 × 1024 after concatenating. Then couple of convolutional layers, and the result was 16 × 16 × 152. The batch normalization was used act as a regularizer, eliminating the need for Dropout [63]. Repeat this process until final layer of convolution, which was 128 × 128 × 1. Figure 5 illustrates the complete step by step approach of our U-Net.
At the end, we adjusted the hyper-parameter while compiling the model. We used optimizer adaptive moment estimation (Adam) [64] instead of stochastic gradient descent optimizer and changed the activation function from softmax to sigmoid [65]. The sigmoid function is defined in Equation (1).
y = 1 1 + e x
The other adjustment is loss function in the metrics, we used binary cross entropy and dice loss (BCE). This loss combines Dice loss with the standard binary cross-entropy (BCE) loss that is generally the default for segmentation models [66].
The hyper-parameters adjustment details with their values are described in Table 2.

4. Experimental Results

In this section we described the experiment setup, splitting of data and evaluation metrics.

4.1. Experimental Setup

The experiments were carried out on Google Colab with Python version 3.8.8. The processor was Intel(R) Xeon(R) Silver 4110 CPU 8 cores with 16 GB RAM. The execution is totally based on CPUs rather than on GPU. The research architectures have been implemented using Keras implementation in Tensorflow 2.7.0. Moreover, Jupyter Notebook has been used in Anaconda on Windows 10 Operating System.

4.2. Train/Test Split

To prepare the data for our model, we randomly divided the 75% cases into training, and 25% test sets for both cases of JPEG and mask knee slices. The details of ratio of the samples are described in Table 3.

4.3. Evaluation Metrics

The segmentation tasks need to output an image the same size as the input image, and the network architecture needs to be adapted accordingly with this output. The value of each pixel or voxel represents the segmentation class of that voxel. The output segmentation is measured against a ground truth often an expert delineation using the metrics. In segmentation, we want to classify whether it is background or foreground. The foreground pixels belonged to the class. We calculated score through accuracy, Intersection over union (IoU), Dice similarity coefficient (DSC), Precision, Recall, and F1-Score. Furthermore, the loss value we also evaluated through Binary Cross Entropy and Dice Loss. The explanation of evaluations metrics is described below:
  • Accuracy
The average accuracy of the model is calculated by the fraction of total area of foreground plus background by the foreground. The accuracy Equation (2) is as below.
A c c u a c y = F o r e g r o u n d   a r e a   o f   k n e e   M a s k T o t a l   a r e a   o f   f o r e g r o u n d + T o t a l   b a c k g r o u n d   a r e a
2.
Intersection over Union
The intersection is the area of the intersection between label (A) and the prediction (B), and it should be maximizing (True Positive). The union should be minimizing because some part of the union is outside the label, in which occur errors (False Positive and False Negative). The IoU Equation (3) is as below.
I O U = I n t e r s e c t i o n ( A B )   U n i o i n   ( A     B )
3.
Dice Coefficient
It is two times the intersection divided by the sum of the sizes of two segmentation regions. Both these metrics are zero when there is no overlapping between the predicted and ground truth regions and one when there is perfect overlap. The dice coefficient score Equation (4) is as below.
D i c e   C o e f f i c i e n t =   2 × I n t e r s e c t i o n ( A B )   U n i o i n   ( A     B ) + I n t e r s e c t i o n ( A B )  
4.
Precision
The fraction of corrected segmented positive images by a total number of true positive and number of false positive. The precision Equation (5) is as below.
P r e c i s o n = T o t a l   n u m b e r   o f   c o r r e c t e d   p r e d i c t i o n   ( T P ) T P + F P  
5.
Recall
The fraction of all positive images in three classes is correctly predicted as positive by the classifier. The recall Equation (6) is as below.
R e c a l l = ( T P )     T P + F N  
6.
F1 score
It combines precision and recalls through harmonic means. The F1 score Equation (7) is as below.
F 1   S c o r e = 1 1 p recision + 1 recall
Figure 6 further shows the curves of training vs. testing accuracy, Iou, dice_coeff, precision, recall and F1 score.
7.
Binary Cross Entropy Dice Loss (BCE-Dice Loss)
It is widely used for classification objective, and as segmentation is pixel level classification, it works well. This loss combines Dice loss with the standard binary cross-entropy (BCE) loss that is generally the default for segmentation models. Combining the two methods allows for some diversity in the loss, while benefitting for Binary Cross Entropy and Dice Loss between a true and a predicted classification. The Equation (8) for Binary cross entropy with dice coefficient is as below.
B C E   l o s s ( A , B ) = 1 / N   i = 1 N ( A   l o g ( B ) + ( 1 A )     l o g ( 1 B ) )
8.
Dice Similarity Loss (DSC)
The Dice coefficient is a widely used metric for the imbalance category to calculate the similarity between two images. If the DSC loss will have a negative impact on the back propagation. The formula of dice is as below in Equation (9)
DiceLoss = 1 2 i = 1   N   A i   B i i = 1 N A i 2 +   i = 1 N B i 2
Figure 7 shows the plot of training and validation loss of BCE-Dice loss and dice loss after 30 epochs.

5. Discussion

In this study, we focused on the semantic segmentation deep learning approach of U-Net CNN architecture to segment the ACL MR images automatically. The ACL segmentation is difficult in the case of pickle images, for this we prepared these images into mask images and then segment it. The experiment results are shown in Figure 8 after evaluation of modification in the U-Net model.
Our results demonstrate in the figure above through deep learning there is not much difference in the accuracy score between the training and testing, respectively, 98.48% and 97.75%. However, the IoU, dice, coeff, revall, precision and F1 score difference was between 5% to 3%. Therefore, our model performance score has shown very good results.
Furthermore, the error loss between the training and test are shown in Figure 9. The test error dice loss is much less as compared with BCE-dice-Loss, respectively, 0.00318 and 0.0849.
Figure 10 shows our predicted images after segmentation of some samples after testing with true images.
Table 4 describes the state-of-art work comparisons with our segmentation model, and evaluation metrics.
The above Table 4 clearly shows that our result of segmentation was performed with good results.
There are five studies in Table 4 applying CNN model without U-Net architecture but having problems: limited dataset of training voxels, not use stage layer to the pre-trained CNN model [31], manual segmentation was time-consuming [33], less accuracy was found even in cases of small structures of knee [35], processing time took much time for segmentation automatically [39], and not investigated multitask learning for segmentation and registration to decease model size [43].
There are six studies that apply U-Net CNN architecture but have limitations: less performance, absence of ground truth [47], lack of image annotation uncertainty [49], lack of gold standard [50], segmentation is performed on only a few parts of knee [50], low resolution of MR images [51], limited only for MR CISS sequences in a single make scanner [52] and data labeling is time consuming [56].
There arethree limitations of our study. Firstly, the data labeling is time-consuming because the bounding box was generated through a VGG annotator tool. Secondly, the segmentation algorithm is also time-consuming to train our U-Net model. Thirdly, the dataset was limited only to ACL tear segmentation.

6. Conclusions

The semantic segmentation of medical images is a more advanced and complicated task. A fully automated modified U-Net convolutional neural network method was applied and segmented the ACL tear of MR images. From the public knee MRI dataset, 917 volumes and 15,268 3D knee MRI slices have been used in this study after converting knee masking. It is the first attempt of dataset preparation into JPEG and knee masking. Without any human intervention, the trained system takes the 11451 MRI slices as input and predicts outputs of the segmentation. In the experiments, the intersection over union, dice coefficient, precision, recall and F1 score was 93.83%, 96.82%, 97.31%, 96.53% and 96.92%, respectively, on the testing dataset for whole knee ACL segmentation. We further conducted binary cross entropy combined with dice coefficient loss and dice loss on training and test datasets. These losses were 0.0849 and 0.0318 on test data after 30 epochs. In addition, we compared the propose model with other state-of-art of all knee components and models. It clearly shown that our result was very promising as compared to other studies. In the future, we could apply classification models on predicted knee images. The instant segmentation can also be applied on knee mask images.

Author Contributions

Conceptualization, M.J.A. and M.S.M.R.; methodology, M.J.A.; software, A.R.; validation, M.J.A. and M.S.M.R.; formal analysis, M.J.A.; investigation, M.J.A. and A.R.; resources, M.S.M.R. and N.S.; data curation, M.J.A. and M.S.M.R.; writing—original draft preparation, M.J.A.; writing—review and editing, M.J.A., M.S.M.R. and B.G.-Z.; visualization, M.J.A. and B.G.-Z.; supervision, M.S.M.R., N.S. and A.R.; project administration, M.S.M.R. and N.S.; funding acquisition, B.G.-Z. All authors have read and agreed to the published version of the manuscript.

Funding

The publication funding has been supported by the University of Deusto (eVIDA group, under the “Contrato Programa” and “Hazitek” Program from Basque Government.

Institutional Review Board Statement

This study used knee MRI and was approved by the Ethics Committee of CHC Rijeka on 30 August 2014. Moreover, on 23 May 2017 it received the Ethics Committee’s approval to make the data publicly available.

Informed Consent Statement

Patient consent was anonymized. Knee MR images were taken for treatment use and there was no identifiable patient information.

Data Availability Statement

We are using this dataset kneeMRI dataset available online: http://www.riteh.uniri.hr/~istajduh/projects/KneeMRI/ (accessed on: 1 March 2017) in our work from Clinical Hospital Centre Rijeka, under reference [61].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ajdaroski, M.; Tadakala, R.; Nichols, L.; Esquivel, A. Validation of a Device to Measure Knee Joint Angles for a Dynamic Movement. Sensors 2020, 20, 1747. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Gupton, M.; Imonugo, O.; Terreberry, R.R. Anatomy, Bony Pelvis and Lower Limb, Knee; StatPearls Publishing: Treasure Island, FL, USA, 2018. [Google Scholar]
  3. Arnoczky, S.P. Anatomy of the anterior cruciate ligament. Clin. Orthop. Relat. Res. 1983, 19–25. [Google Scholar] [CrossRef]
  4. Templeton, K. Musculoskeletal disorders: Sex and gender evidence in anterior cruciate ligament injuries, osteoarthritis, and osteoporosis. In How Sex and Gender Impact Clinical Practice; Elsevier: Amsterdam, The Netherlands, 2021; pp. 207–227. [Google Scholar] [CrossRef]
  5. Awan, M.; Rahim, M.; Salim, N.; Ismail, A.; Shabbir, H. Acceleration of knee MRI cancellous bone classification on google colaboratory using convolutional neural network. Int. J. Adv. Trends Comput. Sci. 2019, 8, 83–88. [Google Scholar] [CrossRef]
  6. Primorac, D.; Molnar, V.; Rod, E.; Jeleč, Ž.; Čukelj, F.; Matišić, V.; Vrdoljak, T.; Hudetz, D.; Hajsok, H.; Borić, I. Knee osteoarthritis: A review of pathogenesis and state-of-the-art non-operative therapeutic considerations. Genes 2020, 11, 854. [Google Scholar] [CrossRef] [PubMed]
  7. Felson, D.T. Osteoarthritis as a disease of mechanics. Osteoarthr. Cartil. 2013, 21, 10–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Abulhasan, J.F.; Grey, M.J. Anatomy and physiology of knee stability. J. Funct. Morphol. Kinesiol. 2017, 2, 34. [Google Scholar] [CrossRef] [Green Version]
  9. McGonagle, D.; Gibbon, W.; Emery, P. Classification of inflammatory arthritis by enthesitis. Lancet 1998, 352, 1137–1140. [Google Scholar] [CrossRef]
  10. Javed Awan, M.; Mohd Rahim, M.S.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient Detection of Knee Anterior Cruciate Ligament from Magnetic Resonance Imaging Using Deep Learning Approach. Diagnostics 2021, 11, 105. [Google Scholar] [CrossRef]
  11. Chen, W.-T.; Shih, T.-F.; Tu, H.-Y.; Chen, R.-C.; Shau, W.-Y. Partial and complete tear of the anterior cruciate ligament: Direct and indirect MR signs. Acta Radiol. 2002, 43, 511–516. [Google Scholar]
  12. Frank, J.S.; Gambacorta, P.L. Anterior Cruciate Ligament Injuries in the Skeletally Immature Athlete: Diagnosis and Management. J. Am. Acad. Orthop. Surg. 2013, 21, 78–87. [Google Scholar] [CrossRef]
  13. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Rehman, A.; Nobanee, H.; Shabir, H. Improved Deep Convolutional Neural Network to Classify Osteoarthritis from Anterior Cruciate Ligament Tear Using Magnetic Resonance Imaging. J. Pers. Med. 2021, 11, 1163. [Google Scholar] [CrossRef] [PubMed]
  14. Sherman, M.F.; Lieber, L.; Bonamo, J.R.; Podesta, L.; Reiter, I. The long-term followup of primary anterior cruciate ligament repair: Defining a rationale for augmentation. Am. J. Sports Med. 1991, 19, 243–255. [Google Scholar] [CrossRef] [PubMed]
  15. DiFelice, G.S.; Villegas, C.; Taylor, S. Anterior Cruciate Ligament Preservation: Early Results of a Novel Arthroscopic Technique for Suture Anchor Primary Anterior Cruciate Ligament Repair. Arthrosc. J. Arthrosc. Relat. Surg. 2015, 31, 2162–2171. [Google Scholar] [CrossRef] [PubMed]
  16. Maqsood, M.; Nazir, F.; Khan, U.; Aadil, F.; Jamal, H.; Mehmood, I.; Song, O.-Y. Transfer Learning Assisted Classification and Detection of Alzheimer’s Disease Stages Using 3D MRI Scans. Sensors 2019, 19, 2645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Yogamangalam, R.; Karthikeyan, B. Segmentation techniques comparison in image processing. Int. J. Eng. Technol. (IJET) 2013, 5, 307–313. [Google Scholar]
  18. Rakhmadi, A.; Othman, N.Z.S.; Bade, A.; Rahim, M.S.M.; Amin, I.M. Connected Component Labeling Using Components Neighbors-Scan Labeling Approach. J. Comput. Sci. 2010, 6, 1099–1107. [Google Scholar] [CrossRef] [Green Version]
  19. Rad, A.E.; Rahim, M.S.M.; Kolivand, H.; Amin, I.B.M. Morphological region-based initial contour algorithm for level set methods in image segmentation. Multimed. Tools Appl. 2017, 76, 2185–2201. [Google Scholar] [CrossRef] [Green Version]
  20. Norouzi, A.; Rahim, M.S.M.; Altameem, A.; Saba, T.; Rad, A.E.; Rehman, A.; Uddin, M. Medical Image Segmentation Methods, Algorithms, and Applications. IETE Tech. Rev. 2014, 31, 199–213. [Google Scholar] [CrossRef]
  21. Afza, F.; Khan, M.A.; Sharif, M.; Rehman, A. Microscopic skin laceration segmentation and classification: A framework of statistical normal distribution and optimal feature selection. Microsc. Res. Tech. 2019, 82, 1471–1488. [Google Scholar] [CrossRef]
  22. Fulkerson, B.; Vedaldi, A.; Soatto, S. Class segmentation and object localization with superpixel neighborhoods. In Proceedings of the 2009 IEEE 12th international conference on computer vision, Kyoto, Japan, 29 September–2 October 2009; pp. 670–677. [Google Scholar]
  23. Ko, T.-Y.; Lee, S.-H. Novel Method of Semantic Segmentation Applicable to Augmented Reality. Sensors 2020, 20, 1737. [Google Scholar] [CrossRef] [Green Version]
  24. Rehman, A.; Harouni, M.; Karimi, M.; Saba, T.; Bahaj, S.A.; Awan, M.J. Microscopic retinal blood vessels detection and segmentation using support vector machine and K-nearest neighbors. Microsc. Res. Tech. 2022, in press. [CrossRef]
  25. Barnouin, Y.; Butler-Browne, G.; Voit, T.; Reversat, D.; Azzabou, N.; Leroux, G.; Behin, A.; McPhee, J.S.; Carlier, P.G.; Hogrel, J.Y. Manual segmentation of individual muscles of the quadriceps femoris using MRI: A reappraisal. J. Magn. Reson. Imaging 2014, 40, 239–247. [Google Scholar] [CrossRef] [PubMed]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical image computing and computer-assisted intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  27. Sun, M.; Lu, L.; Hameed, I.A.; Kulseng, C.P.S.; Gjesdal, K.-I. Detecting Small Anatomical Structures in 3D Knee MRI Segmentation by Fully Convolutional Networks. Appl. Sci. 2022, 12, 283. [Google Scholar] [CrossRef]
  28. Islam, K.T.; Wijewickrema, S.; O’Leary, S. A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images. Sensors 2022, 22, 523. [Google Scholar] [CrossRef] [PubMed]
  29. Nabeel, M.; Majeed, S.; Awan, M.J.; Ud-Din, H.M.; Wasique, M.; Nasir, R. Review on Effective Disease Prediction through Data Mining Techniques. Int. J. Electr. Eng. Inform. 2021, 13, 717–733. [Google Scholar] [CrossRef]
  30. Gupta, M.; Jain, R.; Arora, S.; Gupta, A.; Javed Awan, M.; Chaudhary, G.; Nobanee, H. AI-enabled COVID-9 Outbreak Analysis and Prediction: Indian States vs. Union Territories. Comput. Mater. Contin. 2021, 67, 933–950. [Google Scholar] [CrossRef]
  31. Aftab, M.O.; Awan, M.J.; Khalid, S.; Javed, R.; Shabir, H. Executing Spark BigDL for Leukemia Detection from Microscopic Images using Transfer Learning. In Proceedings of the 2021 1st International Conference on Artificial Intelligence and Data Analytics (CAIDA), Riyadh, Saudi Arabia, 6–7 April 2021; pp. 216–220. [Google Scholar]
  32. Ali, Y.; Farooq, A.; Alam, T.M.; Farooq, M.S.; Awan, M.J.; Baig, T.I. Detection of Schistosomiasis Factors Using Association Rule Mining. IEEE Access 2019, 7, 186108–186114. [Google Scholar] [CrossRef]
  33. Awan, M.J.; Bilal, M.H.; Yasin, A.; Nobanee, H.; Khan, N.S.; Zain, A.M. Detection of COVID-19 in Chest X-ray Images: A Big Data Enabled Deep Learning Approach. Int. J. Environ. Res. Public Health 2021, 18, 10147. [Google Scholar] [CrossRef]
  34. Prasoon, A.; Petersen, K.; Igel, C.; Lauze, F.; Dam, E.; Nielsen, M. Deep features learning for knee carliage segmentation using a triplanar convolutional neural network. In Proceedings of the International conference on medical image computing and computer-assisted intervention, Nagoya, Japan, 22–26 September 2013; pp. 246–253. [Google Scholar]
  35. Deniz, C.M.; Xiang, S.; Hallyburton, R.S.; Welbeck, A.; Babb, J.; Honig, S.; Cho, K.; Chang, G. Segmentation of the Proximal Femur from MR Images using Deep Convolutional Neural Networks. Sci. Rep. 2018, 8, 16485. [Google Scholar] [CrossRef] [Green Version]
  36. Zhou, Z.; Zhao, G.; Kijowski, R.; Liu, F. Deep convolutional neural network for segmentation of knee joint anatomy. Magn. Reson. Med. 2018, 80, 2759–2770. [Google Scholar] [CrossRef]
  37. Heimann, T.; Morrison, B.J.; Styner, M.A.; Niethammer, M.; Warfield, S. Segmentation of knee images: A grand challenge. In Proceedings of the MICCAI Workshop on Medical Image Analysis for the Clinic, the 13th International Conference on Medical Image Computing and Computer Assisted Intervention, Beijing, China, 20–24 September 2010; pp. 207–214. [Google Scholar]
  38. Ambellan, F.; Tack, A.; Ehlke, M.; Zachow, S. Automated segmentation of knee bone and cartilage combining statistical shape knowledge and convolutional neural networks: Data from the Osteoarthritis Initiative. Med. Image Anal. 2019, 52, 109–118. [Google Scholar] [CrossRef] [PubMed]
  39. Kainmueller, D. Deformable Meshes for Medical Image Segmentation: Accurate Automatic Segmentation of Anatomical Structures; Springer: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  40. Vincent, G.; Wolstenholme, C.; Scott, I.; Bowes, M. Fully automatic segmentation of the knee joint using active appearance models. In Medical Image Analysis for the Clinic: A Grand Challenge; CreateSpace Independent Publishing: Scotts Valley, CA, USA, 2010; Volume 1, p. 224. [Google Scholar]
  41. Seim, H.; Kainmueller, D.; Lamecker, H.; Bindernagel, M.; Malinowski, J.; Zachow, S. Model-based auto-segmentation of knee bones and cartilage in MRI data. In Proceedings of the 13th International Conference on Medical Image Computing and Computer Assisted Intervention, Beijing, China, 20–24 September 2010. [Google Scholar]
  42. Xu, Z.; Niethammer, M. DeepAtlas: Joint semi-supervised learning of image registration and segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 420–429. [Google Scholar]
  43. Burton, W., 2nd; Myers, C.; Rullkoetter, P. Semi-supervised learning for automatic segmentation of the knee from MRI with convolutional neural networks. Comput. Methods Programs Biomed. 2020, 189, 105328. [Google Scholar] [CrossRef] [PubMed]
  44. Smoger, L.M.; Fitzpatrick, C.K.; Clary, C.W.; Cyr, A.J.; Maletsky, L.P.; Rullkoetter, P.J.; Laz, P.J. Statistical modeling to characterize relationships between knee anatomy and kinematics. J. Orthop. Res. 2015, 33, 1620–1630. [Google Scholar] [CrossRef] [PubMed]
  45. Peterfy, C.; Schneider, E.; Nevitt, M. The osteoarthritis initiative: Report on the design rationale for the magnetic resonance imaging protocol for the knee. Osteoarthr. Cartil. 2008, 16, 1433–1441. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Liu, F.; Zhou, Z.; Samsonov, A.; Blankenbaker, D.; Larison, W.; Kanarek, A.; Lian, K.; Kambhampati, S.; Kijowski, R. Deep Learning Approach for Evaluating Knee MR Images: Achieving High Diagnostic Performance for Cartilage Lesion Detection. Radiology 2018, 289, 160–169. [Google Scholar] [CrossRef]
  47. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  48. Awan, M.J.; Masood, O.A.; Mohammed, M.A.; Yasin, A.; Zain, A.M.; Damaševičius, R.; Abdulkareem, K.H. Image-Based Malware Classification Using VGG19 Network and Spatial Convolutional Attention. Electronics 2021, 10, 2444. [Google Scholar] [CrossRef]
  49. Tack, A.; Mukhopadhyay, A.; Zachow, S. Knee menisci segmentation using convolutional neural networks: Data from the Osteoarthritis Initiative. Osteoarthr. Cartil. 2018, 26, 680–688. [Google Scholar] [CrossRef]
  50. Lamecker, H. Variational and statistical shape modeling for 3D geometry reconstruction. Ph.D Thesis, Freie Universität Berlin, Berlin, Germany, 2008. [Google Scholar]
  51. Raj, A.; Vishwanathan, S.; Ajani, B.; Krishnan, K.; Agarwal, H. Automated knee cartilage segmentation using fully volumetric convolutional neural netowrks for evaluation of osteoarthritis. In Proceedings of the IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 851–854. [Google Scholar]
  52. Nevitt, M.; Felson, D.; Lester, G. The osteoarthritis initiative. Protoc. Cohort Study 2006, 1. [Google Scholar]
  53. Pedoia, V.; Norman, B.; Mehany, S.N.; Bucknor, M.D.; Link, T.M.; Majumdar, S. 3D convolutional neural networks for detection and severity staging of meniscus and PFJ cartilage morphological degenerative changes in osteoarthritis and anterior cruciate ligament subjects. J. Magn. Reson. Imaging 2019, 49, 400–410. [Google Scholar] [CrossRef]
  54. Norman, B.; Pedoia, V.; Majumdar, S. Use of 2D U-Net Convolutional Neural Networks for Automated Cartilage and Meniscus Segmentation of Knee MR Imaging Data to Determine Relaxometry and Morphometry. Radiology 2018, 288, 177–185. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Flannery, S.W.; Kiapour, A.M.; Edgar, D.J.; Murray, M.M.; Fleming, B.C. Automated magnetic resonance image segmentation of the anterior cruciate ligament. J. Orthop. Res. 2021, 39, 831–840. [Google Scholar] [CrossRef] [PubMed]
  56. Flannery, S.W.; Kiapour, A.M.; Edgar, D.J.; Murray, M.M.; Beveridge, J.E.; Fleming, B.C. A transfer learning approach for automatic segmentation of the surgically treated anterior cruciate ligament. J. Orthop. Res. 2021, 40, 277–284. [Google Scholar] [CrossRef] [PubMed]
  57. Murray, M.M.; Kalish, L.A.; Fleming, B.C.; BEAR Trial Team; Flutie, B.; Freiberger, C.; Henderson, R.N.; Perrone, G.S.; Thurber, L.G.; Proffen, B.L.; et al. Bridge-Enhanced Anterior Cruciate Ligament Repair: Two-Year Results of a First-in-Human Study. Orthop. J. Sports Med. 2019, 7, 2325967118824356. [Google Scholar] [CrossRef]
  58. Murray, M.M.; Fleming, B.C.; Badger, G.J.; BEAR Trial Team; Freiberger, C.; Henderson, R.; Barnett, S.; Kiapour, A.; Ecklund, K.; Proffen, B.; et al. Bridge-Enhanced Anterior Cruciate Ligament Repair Is Not Inferior to Autograft Anterior Cruciate Ligament Reconstruction at 2 Years: Results of a Prospective Randomized Clinical Trial. Am. J. Sports Med. 2020, 48, 1305–1315. [Google Scholar] [CrossRef]
  59. Murray, M.M.; Kiapour, A.M.; Kalish, L.A.; Ecklund, K.; BEAR Trial Team; Freiberger, C.; Henderson, R.; Kramer, D.; Micheli, L.; Yen, Y.-M.; et al. Predictors of Healing Ligament Size and Magnetic Resonance Signal Intensity at 6 Months After Bridge-Enhanced Anterior Cruciate Ligament Repair. Am. J. Sports Med. 2019, 47, 1361–1369. [Google Scholar] [CrossRef]
  60. Almajalid, R.; Zhang, M.; Shan, J. Fully Automatic Knee Bone Detection and Segmentation on Three-Dimensional MRI. Diagnostics 2022, 12, 123. [Google Scholar] [CrossRef]
  61. Štajduhar, I.; Mamula, M.; Miletić, D.; Ünal, G. Semi-automated detection of anterior cruciate ligament injury from MRI. Comput. Methods Programs Biomed. 2017, 140, 151–164. [Google Scholar] [CrossRef]
  62. Dutta, A.; Zisserman, A. The VIA annotation software for images, audio and video. In Proceedings of the 27th ACM international conference on multimedia, Nice, France, 21–25 October 2019; pp. 2276–2279. [Google Scholar]
  63. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International conference on machine learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  64. Wichrowska, O.; Maheswaranathan, N.; Hoffman, M.W.; Colmenarejo, S.G.; Denil, M.; Freitas, N.; Sohl-Dickstein, J. Learned optimizers that scale and generalize. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3751–3760. [Google Scholar]
  65. Han, J.; Moraga, C. The influence of the sigmoid function parameters on the speed of backpropagation learning. In Proceedings of the International workshop on artificial neural networks, Torremolinos, Spain, 7–9 June 1995; pp. 195–201. [Google Scholar]
  66. Yeung, M.; Sala, E.; Schönlieb, C.-B.; Rundo, L. Unified Focal loss: Generalising Dice and cross entropy-based losses to handle class imbalanced medical image segmentation. arXiv 2021, arXiv:2102.04525. [Google Scholar] [CrossRef]
Figure 1. Taxonomy of knee joint anatomy.
Figure 1. Taxonomy of knee joint anatomy.
Sensors 22 01552 g001
Figure 2. The four types of ACL tears; (a) type 1 avulsion of femur wall; (b) type II middle segment in 2nd ligament;(c) type III middle segment in 3rd of the ligament; and (d) type IV avulsion in the tibial part [15].
Figure 2. The four types of ACL tears; (a) type 1 avulsion of femur wall; (b) type II middle segment in 2nd ligament;(c) type III middle segment in 3rd of the ligament; and (d) type IV avulsion in the tibial part [15].
Sensors 22 01552 g002
Figure 3. The segmentation framework of our proposed approach with phase 1 of data preparation, phase 2 of knee mask generation and phase 3 of model building with training and testing on our U-Net CNN architecture.
Figure 3. The segmentation framework of our proposed approach with phase 1 of data preparation, phase 2 of knee mask generation and phase 3 of model building with training and testing on our U-Net CNN architecture.
Sensors 22 01552 g003
Figure 4. The sample true JPEG images and their knee mask images.
Figure 4. The sample true JPEG images and their knee mask images.
Sensors 22 01552 g004
Figure 5. The complete process of modified U-Net Model with eight convolutional layers, four maximum pooling and upsample layers training and prediction.
Figure 5. The complete process of modified U-Net Model with eight convolutional layers, four maximum pooling and upsample layers training and prediction.
Sensors 22 01552 g005
Figure 6. The evaluation score after 30 epochs training vs. validation plots; (a) Accuracy score curve; (b) IoU score curve; (c) dice_coeff score curve; (d) Precision score curve; (e) recall score curve; and (f) F1 score curve.
Figure 6. The evaluation score after 30 epochs training vs. validation plots; (a) Accuracy score curve; (b) IoU score curve; (c) dice_coeff score curve; (d) Precision score curve; (e) recall score curve; and (f) F1 score curve.
Sensors 22 01552 g006aSensors 22 01552 g006bSensors 22 01552 g006cSensors 22 01552 g006d
Figure 7. The loss value after 30 epochs training vs. test curves; (a) BCE-Dice Loss curve; and (b) Dice Similarity Loss curve.
Figure 7. The loss value after 30 epochs training vs. test curves; (a) BCE-Dice Loss curve; and (b) Dice Similarity Loss curve.
Sensors 22 01552 g007
Figure 8. The performance charts of accuracy, IoU, dice_coeff, recall, precision and F1 score.
Figure 8. The performance charts of accuracy, IoU, dice_coeff, recall, precision and F1 score.
Sensors 22 01552 g008
Figure 9. The Error Loss of Dice Loss and BCE_Dice-Loss on training and test dataset.
Figure 9. The Error Loss of Dice Loss and BCE_Dice-Loss on training and test dataset.
Sensors 22 01552 g009
Figure 10. After testing actual knee mask and predicted mask.
Figure 10. After testing actual knee mask and predicted mask.
Sensors 22 01552 g010
Table 1. The algorithm of turning pickles into JPEG images.
Table 1. The algorithm of turning pickles into JPEG images.
Input: Load all pickle files into data
for id in enumerate(data):
reshape each d into image till last image
save each image into JPEG form
Output: show all images
Table 2. The hyper-parameters adjustment and values.
Table 2. The hyper-parameters adjustment and values.
Hyper-Parameters AdjustmentValue
Input Image 128 × 128 × 3
Batch Size32
Number of Epochs30
Learning rate 0.0001
Optimizers Adam
Loss FunctionBinary cross-entropy and Dice loss
Table 3. The training and test split ratio of samples of knee MR images.
Table 3. The training and test split ratio of samples of knee MR images.
Table 11451Knee JPEG ImagesKnee Mask Images
Training Data 1145111451
Test Data 38173817
Table 4. The segmentation of various knee comparisons with our ACL segmentation.
Table 4. The segmentation of various knee comparisons with our ACL segmentation.
Author, Year Technique/ModelSegment Part ACL Yes/NOSegment PartEvaluation
DSCIoURecall PrecisionF1-Score
Prasoon [34], 20132D CNN NoTC0.824 -0.819--
Deniz, Xiang, Hallyburton, Welbeck, Babb, Honig, Cho and Chang [35] P, 20183D CNN
U-Net
NoPF0.950-0.9500.950-
Zhou, Zhao, Kijowski and Liu [36], 2018CNNVGG16 NoTF, muscle,
non-spec tissue
0.910----
Ambellan, Tack, Ehlke and Zachow [38], 2019CNNNo,
OAI Imorphics
FC0.894----
MTC0.861 ----
LTC0.904----
Xu and Niethammer [42], 2019CNNNoBone0.968----
Cartilages0.776----
knee part other0.872----
Burton, Myers and Rullkoetter [43], 2020U-Net NoFemur, FC, TC, PC, Tibia, petella0.9890.971---
Liu, Zhou, Samsonov, Blankenbaker, Larison, Kanarek, Lian, Kambhampati and Kijowski [46], 20182D CNN NoFemur0.96 ----
Tibia0.95 ----
FC0.81----
TC0.82----
Tack, Mukhopadhyay and Zachow [49], 2018U-NetNoLM0.889----
MM0.838----
Raj [51], 2018U-NetNoOAI: FC0.849 ----
LM0.849 ----
LTC0.856----
MM0.801----
MTC0.806 ----
PC0.784----
SK110:FC0.834 ----
TC0.825----
Pedoia, Norman, Mehany, Bucknor, Link and Majumdar [53], 2019U-Net NoMeniscus--0.899--
Cartilage--0.801--
Normal lesion--0.807--
Norman, Pedoia and Majumdar [54], 2018U-NetNoFC0.878----
LTC0.820----
MTC0.795----
PC0.767----
LM0.809----
MM0.753----
Flannery, Kiapour, Edgar, Murray and Fleming [55], 2021U-NetYes
repair BEAR
ACL0.840-0.8500.821-
Flannery, Kiapour, Edgar, Murray, Beveridge and Fleming [56], 2021U-Net YesACL Intact BEAR0.840-0.8500.820-
ACL graft 0.780-0.8010.781-
Almajalid, Zhang and Shan [60]U Net No
Imorphics OAI
Tibia0.963 -0.9950.988-
Femur0.979-0.9960.988-
petella0.928-0.9710.992-
Dice Similarity Coefficient= DSC, Femur Cartilage = FC, Tibial Cartilage = TC, Proximal Femur = PF, Lateral Meniscus = LM, Medical Meniscus = MM, Lateral Tibial Cartilage = LTC, Medial Tibial Cartilage = MTC.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Awan, M.J.; Rahim, M.S.M.; Salim, N.; Rehman, A.; Garcia-Zapirain, B. Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears. Sensors 2022, 22, 1552. https://doi.org/10.3390/s22041552

AMA Style

Awan MJ, Rahim MSM, Salim N, Rehman A, Garcia-Zapirain B. Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears. Sensors. 2022; 22(4):1552. https://doi.org/10.3390/s22041552

Chicago/Turabian Style

Awan, Mazhar Javed, Mohd Shafry Mohd Rahim, Naomie Salim, Amjad Rehman, and Begonya Garcia-Zapirain. 2022. "Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears" Sensors 22, no. 4: 1552. https://doi.org/10.3390/s22041552

APA Style

Awan, M. J., Rahim, M. S. M., Salim, N., Rehman, A., & Garcia-Zapirain, B. (2022). Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears. Sensors, 22(4), 1552. https://doi.org/10.3390/s22041552

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop