Next Article in Journal
Fracture-Related Infection in Bicolumnar Acetabular Fracture: A Case Report
Next Article in Special Issue
Theory and Practice of Integrating Machine Learning and Conventional Statistics in Medical Data Analysis
Previous Article in Journal
Viscosity Plane-Wave UltraSound (Vi PLUS) in the Evaluation of Thyroid Gland in Healthy Volunteers—A Preliminary Study
Previous Article in Special Issue
Deep Learning Assisted Automated Assessment of Thalassaemia from Haemoglobin Electrophoresis Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Deep Learning to IVC Filter Detection from CT Scans

1
Department of Computer Science, University of Wisconsin-Eau Claire, Eau Claire, WI 54701, USA
2
Interventional Radiology, Mayo Clinic Health System, Eau Claire, WI 54703, USA
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(10), 2475; https://doi.org/10.3390/diagnostics12102475
Submission received: 10 September 2022 / Revised: 5 October 2022 / Accepted: 10 October 2022 / Published: 13 October 2022
(This article belongs to the Special Issue Artificial Intelligence in Clinical Medical Imaging Analysis)

Abstract

:
IVC filters (IVCF) perform an important function in select patients that have venous blood clots. However, they are usually intended to be temporary, and significant delay in removal can have negative health consequences for the patient. Currently, all Interventional Radiology (IR) practices are tasked with tracking patients in whom IVCF are placed. Due to their small size and location deep within the abdomen it is common for patients to forget that they have an IVCF. Therefore, there is a significant delay for a new healthcare provider to become aware of the presence of a filter. Patients may have an abdominopelvic CT scan for many reasons and, fortunately, IVCF are clearly visible on these scans. In this research a deep learning model capable of segmenting IVCF from CT scan slices along the axial plane is developed. The model achieved a Dice score of 0.82 for training over 372 CT scan slices. The segmentation model is then integrated with a prediction algorithm capable of flagging an entire CT scan as having IVCF. The prediction algorithm utilizing the segmentation model achieved a 92.22% accuracy at detecting IVCF in the scans.

1. Introduction

Venous thromboembolism (VTE) are blood clots that begin in vein, a common disorder that affects 1 in 1000 people annually [1]. The standard treatment are anticoagulants (blood thinners) which are highly effective at reducing the risks of ongoing and recurrent VTE. However, some patients cannot tolerate anticoagulation. In these cases, the insertion of a IVCF is recommended to reduce the risk of a pulmonary embolism [1,2]. They are a second-line intervention as they do not treat an existing VTE and instead act as prevention against a pulmonary embolism.
Originally developed in a permanent form in the 1970s, a retrievable version was approved for use by the FDA in 2003 [3], while greater than 85% of filters are initially placed with the intention of temporary use, a retrospective review found that up to two-thirds were never retrieved [4]. In theory, not retrieving an IVCF should not be a problem as all filters are FDA-approved for permanent use in addition to temporary indications. Unfortunately, numerous studies have shown the potential for serious complications resulting from long dwell times of filters, with the risk of complication increasing almost linearly with time [5,6]. This was highlighted in an FDA communication, released in August 2010 and updated in May 2014, that recommended retrieval as soon as possible and estimated the risk/benefit begins to favor retrieval between 29 and 54 days after placement unless there was need for prolonged IVC filtration [7]. Importantly, that communication explicitly placed the responsibility of evaluating the ongoing need for filtration on the implanting physician. The communication also had the side effect of drastically increasing IVCF-related litigation [8].
Retrieval of IVCF is important to avoid potential long-term complications that may arise. These complications include filter migration, caval thrombosis, filter fracture and caval penetration [2,9,10]. The risk of these complications occurring increases with longer dwell times [2,10]. However, the study in [4] also suggested that “loss to follow-up” or discontinued care and lack of a tracking program account for at least 20% of the filters that were not retrieved.
It is now basically mandatory that all physician groups who place IVCF, primarily Interventional Radiologists, have a method for tracking patients and removing filters that are no longer indicated. This is commonly done using a spreadsheet, which has the significant limitation that the information is not shared even across physician groups within the same enterprise. As patients transfer their care from one physician group to another it is common to become “lost to follow-up.” Improvements in tracking beyond the simple spreadsheet certainly have many advantages but cannot help identify patients whose filter was placed elsewhere but are now under the purview of a new health system or physician group.
Fortunately, IVCF are readily visible on CTs of the abdomen and patients receive this type of imaging commonly and for reasons unrelated to the filter itself including injury, abdominal pain, screening (e.g., CT colonography instead of a colonoscopy) or cancer staging/surveillance. Radiologists usually comment if a filter has any finding of a complication but often do not make any comment if a filter appears normal. This results in poor reliability of automated natural language processing (NLP) methods to identify previously unknown filters. Given that there are approximately 400 eligible CT scans performed every day within the Mayo enterprise and low prevalence of IVCF in the general population, a detection algorithm with low accuracy would place an untenable burden on clinician review of the inevitable false positive results. Therefore, an algorithm with greater accuracy, and most importantly one with high specificity, is needed.
Within the broader construct of Artificial Intelligence (AI), machine learning algorithms are used to learn patterns, gather statistics on input data and use that information for prediction of future unseen data. The performance and prediction accuracy of these algorithms is dependent on quality and presentation of raw input. The concept of better representation of raw data to the machine learning models is termed as feature engineering [11,12,13,14] and a significant amount of time must be spent in manually creating important features from the existing ones [15,16,17]. Deep learning algorithms, which are a subset of machine learning, offer an advantage in this domain as they can automatically extract relevant features without the requirement of human input using a layered and hierarchical structure [18]. A class of deep learning architecture known as a Convolutional Neural Network (CNN) will be explored in this study due to its ability to accurately classify images by incorporating their spatial aspect [19,20].
A CNN consists of convolutional layers stacked one after another. There are three basic steps to convolution which are repeated several times: convolution, activation, and pooling. In convolution, an image is analyzed by sliding several square grids of a certain pixel size across the entire image. This process is referred to as a sliding window approach and each of these sliding windows are called filters or kernels [21]. Each kernel extracts certain features in the image. The output of a layer is next passed through an activation function introducing non-linearity into the model, which is the one of the biggest strengths of a neural network. Rectified Linear Unit (ReLU) [22] is the most commonly used activation function in this stage. In pooling, an image is analyzed in multiple resolutions to detect low, medium and high-level features. Max-pooling is commonly employed in this stage wherein the images are downscaled multiple times. Since the convolution operation repeats on the image in multiple resolutions, we can extract features from the image with superior accuracy [23].
This project aims to develop a simple yet effective AI-based algorithm with significant accuracy to automatically analyze all CT scans and produce a binary output indicating the presence or absence of an IVCF. Due to the narrow focus and retrospective nature of the task, it can also serve as an example for rapidly developing an AI tool and deploying it without any risk to patients. With almost 811,487 IVCF placed between 2005 and 2012 according to data from the Agency for Healthcare Research and Quality Health Care Utilization Project Nationwide Inpatient Sample, it is imperative that a follow-up system is made to ensure improved patient care [24].

2. Artificial Intelligence and Medical Imaging

Recent developments in technology have brought about the digitization of healthcare with up to 84% of general medical hospitals in the United States having adopted the use of electronic health record (EHR) systems as of 2015 [25]. This abundance of digitized data has allowed AI to revolutionize the healthcare industry and improve patient care. One of these revolutionary techniques is the utilization of deep learning models, specifically CNNs, for medical image analysis [26] to diagnose medical conditions.
Image segmentation and classification tasks are of especially high priority in the medical imaging field [1]. CNNs are a prime tool for these tasks due to its ability to learn highly discriminative features present in the images [27,28]. UNet is a type of CNN architecture which was initially developed for biomedical image segmentation purpose. This architecture outperformed other image segmentation methods of its times including the previous best method [29]. Since then, variants of the UNet architecture have been used extensively for medical image analysis. This includes but is not limited to diagnosing cancer (liver, lung, cervical, etc.) as well as segmenting hard and soft tissues in CT scans [30]. UNet architecture was also used to segment the IVC lumen in intracardiac echocardiography images as part of a proposed pipeline for image-guided vascular navigation [31]. CNNs have successfully been used to detect the presence of lung cancer in CT scans. In [32], authors suggested that with appropriate preprocessing steps, CNNs are able to classify CT scans as positive or negative for lung cancer with high accuracy. Similarly, Ref. [33] proposed a two-module network to detect the presence of malignant pulmonary nodules. The first module is a 3D-region proposal network that detects suspicious nodules in the scan while the second module evaluates the cancer probability based on the most suspicious nodules using a leaky noisy-or model. CNNs have also been used extensively for detection of tuberculosis [20,34,35], cancer detection [36,37,38,39,40] as well as COVID-19 [41,42,43,44]. Domain adaptation using unsupervised machine learning have also been successfully applied for knowledge extraction and organ segmentation [45,46,47].
In regards to medical image analysis involving IVCF, research has previously been done on classifying the type of IVCF present in radiographs using CNNs. One research focused on classifying 14 different IVCF types in manually cropped radiographic images. A 50-layer ResNet architecture was used with a modified final fully connected layer to perform this classification [48]. Another research attempted to build on this approach by creating an architecture that is able to classify three different IVCF types without the need to crop and manipulate the radiographs to be centered on the filter [49]. Prior research by Dr. Wildenberg, developed a purely image-processing algorithm that can detect a filter from CT scans with a sensitivity and specificity of about 80% each. Given the approximately 400 eligible CT scans performed every day within the Mayo enterprise and low prevalence of IVCF in the general population, this accuracy would place an untenable burden on clinician review of the inevitable false positive results. An automated deep learning approach provides a feasible solution and to our knowledge the domain of real-time IVCF presence/absence detection from CT scans remains unexplored.

3. Materials and Methods

The schematic diagram of the proposed approach is shown in Figure 1. It consists of an image augmentation phase, a deep learning model phase, followed by a prediction phase utilizing the deep learning models.

3.1. Dataset

Institutional Review Board (IRB) approval was obtained for this project. Approximately 100 known positive (IVCF present) and 100 known normal CT scans were provided by Dr. Wildenberg using the internal list of known patients with filters out of which 90 IVCF and 90 normal scans were finally used. These scans were anonymized by removing any protected health information (PHI) attached to the scan. Slice thickness and width distribution is shown in Figure 2. Initial preprocessing of the CT images was guided utilizing a priori knowledge about IVCF (e.g., composed of metal, located approximately within the center of the body, small relative to other structures) to maximize the relevance of the data submitted to the CNN.
A total of 465 CT image slices were used during the training and validation phase from the 90 CT scans with IVCF. These slices had visible IVCF in them. Segmentation of IVCF pixels were generated by the research group under supervision of Dr. Wildenberg and Dr. Gomes to ensure they represent a proper mask for the CNN to train on. CNN development, training, and validation was then performed. The most significant aspect of this project involved exploration of the exact architecture and parameters to yield a high-performing CNN for filter recognition.
The set of 90 normal scans were later used in the IVCF prediction algorithm to validate the performance of the model in the realm of false positives. These scans were not used during the segmentation phase.

3.2. Spatial Cropping

Data augmentation both pre-training and during training were conducted with diverse objectives. Prior to training, all scans were spatially cropped to remove as much of background possible giving the deep learning model significantly less extraneous information for better performance. This spatial cropping of 20% was applied across all four edges. Hence the original CT scans of 512 × 512 were reduced to 307 × 307 followed by resampling to 256 × 256 . The spatial cropping was followed by removal of CT image slices per scan that were 40 cm below the craniocaudal region. The choice of 40 cm as a cut-off was ideal since nearly all IVCF are located inside that region. Figure 3 shows the distribution of slices per CT scan before and after this reduction. The 40 cm cut-off was able to reduce the total number of CT image slices by 19.01%.

3.3. Normalization

Two variants of the IVCF dataset were generated with and without intensity normalization. For soft normalization scheme, CT images were normalized using minimum and maximum Hounsfield Units (HU) values. For Hard Normalization, the maximum and minimum HU values were set to 1 and 2500, respectively. The purpose of this preprocessing step was to ascertain if window width increases the contrast of IVCF making them more prominent for the CNN models. Figure 4 shows the result of the application of the aforementioned intensity normalization.

3.4. Image Augmentation

Data augmentation during the training phase satisfied the removal of training bias. Any machine learning model is sensitive to location and orientation of features. As such, the proposed model performed cropping, flipping, and rotation. CT scan and the corresponding label were first resized by a certain margin before applying a random cropping function. This random cropping simulated a patch based analysis where the CNN would randomly extract a subset of the entire CT image slice. This ensured that the model would not receive the same image while training in each epoch. The CT image slices were also randomly flipped from left to right to simulate differences in horizontal orientation of CT image acquisition. These steps ensure that CNN predictions stay robust even if they are used across multiple scanning platforms with different settings. To ensure that image labels are consistent in the resampling process, Nearest Neighbor Interpolation was used compared to Bilinear interpolation for the CT scan images. An example of image augmentation is shown in Figure 5.

3.5. Network Architecture

A modified version of UNet was used as a base architecture for this research. This framework was built using online resources available from TensorFlow [50]. The inception of deep learning and image segmentation began with the UNet architecture proposed by Ronneberger et al. [29]. Here each CT image slice is analyzed using a combination of image filters (also known as kernels), and in multiple resolutions. The kernels applied over the image are responsible for identifying hidden pattern in the dataset. Proper application of these kernels including their sizes and dilation rate can have a significant impact on accuracy as well as optimization features [51,52]. For example, a kernel that could be used to delineate edges in the image is the Sobel filter [53]. These filters act as weights during training process allowing the deep learning model to train itself and identify patterns associated with the IVCF. The non-linear approach significantly boosts model performance over traditional machine learning algorithms. The hierarchical structure of the UNet architecture is made possible by modifying the image resolution. Application of kernels at multiple resolutions further allows extraction of diverse features associated with the IVCF. Thus, the underlying architecture can process complex data and extract relevant features in different levels of abstraction.

3.6. Training Parameters

The underlying framework of this architecture is visible in Figure 6. Approximately 25 million parameters were used for training. Adam optimizer was used for gradient descent. Leaky Relu activation was used in the downsampling phase to allow negative gradients and further optimize weights for better training. The model was trained for 500 epochs using a batch size of 20. Training and validation accuracy were recorded with 80% of the total images (i.e., 372 slices) being used for training and 20% for validation (i.e., 93 slices). Since this is a segmentation, Sparse Categorical Crossentropy loss was used. The training loss decreased steadily from 0.3276 to 0.0016 for hard normalization scheme. The training loss also decreased steadily from 0.1651 to 0.0015 for soft normalization scheme. Since dropout is not an effective approach in convolution networks [54], it was only applied in the first upsampling layer with a value of 0.5. Batch normalization was used regularly after convolution layers. Deep learning models are computationally intensive. Hence all processing was carried out using the BOSE cluster made available through the Blugold Center for High Performance Computing. The Graphics Processing Unit for this research was NVIDIA Tesla V100S with 32 GB memory.

3.7. IVCF Prediction Pipeline

The proposed UNet architecture above is used to develop a model that is able to successfully segment IVCF pixels in the slices of a CT scan. To integrate the tool and enable rapid diagnosis, an IVCF prediction algorithm is proposed. This algorithm generates a comprehensive report for the clinicians about the IVCF and its location. The flagging algorithm should have the potential to reduce false positives that may arise from bone or surgical implants which share similar HU values with IVCF. It uses two parameters (a) sig_count and (b) sequence to flag a CT scan as having IVCF. Using sig_count the clinician sets a threshold for the number of pixels that may belong to an IVCF per CT image slice. The sequence parameter takes a spatial approach by looking at how many consecutive slices exceed the sig_count. For example, if the sig_count = 200 and sequence = 5, the prediction pipeline will alert the clinician when five consecutive CT slices have 200 IVCF segmented pixels or more. The clinician can then review the results to see if any IVCF is present. The clinician can also modify the sig_count and sequence values to fine tune the prediction outputs based on patient and CT scanner characteristics. For example, if a patient has surgical implants, the clinician may feel comfortable increasing these parameters to reduce detection of other devices placed in the body. If the prediction does not satisfy any of the two parameters, the CT scan will be classified as having no IVCF.
The process begins by accepting the segmented prediction from UNet and passing it through scikit-image image processing library in Python [55]. Region properties of the segments are analyzed to remove any spurious segmentation that may arise occasionally. The sub-section of the algorithm for this prediction pipeline is shown in Algorithm 1. The algorithm takes in as input the images, labels, sig_count, and sequence. The images represents a processed NumPy array of a CT scan of a patient, while the labels contain a predicted mask for that CT scan. Both these arrays have a dimension of ( s l i c e s × w i d t h × h e i g h t ) . The default dimensions of the processed NumPy version of a CT scan used for prediction is ( 128 × 256 × 256 ) . The algorithm also calls a separate function called displayImage which is responsible for printing the scans and segmentation masks of the IVCF sequence. The combination of sequence and sig_count plays an excellent role in IVCF detection and will be discussed in the next section.
Algorithm 1: IVCF prediction pipeline using the deep learning model
Diagnostics 12 02475 i001

4. Results

4.1. Segmentation Evaluation

Two separate UNet variants were evaluated. Table 1 highlights the data variants used for training. These variants differ in the hard and soft normalization techniques as highlighted in Figure 4. The mean average precision determines the accuracy between ground truth segmentation mask compared to the prediction masks of IVCF pixels. The Dice score was used to analyze the mean average precision. An example of Dice metric analysis is shown in Figure 7. Here one rectangle references the segmentation and the other is for prediction.
The first variant of the UNet model performed significantly well with Dice score of 0.817 for IVCF on training data and 0.798 for validation data. The number of false positives were significantly low for both training and validation IVCF pixels. During training, only 7.97% of IVCF pixels were misclassified as background and 0.12% of background pixels were classified as IVCF. The training sensitivity was about 0.92 and specificity was almost 0.99. For the hard normalized data, a SegNet variant of the model was also used for comparative analysis. We noticed slightly lower Dice scores at 0.743 for training data. This may be due to inherent architecture of SegNet [56] which only transfers max-pooling indices from the encoder phase. Due to this performance, UNet model was selected in the final prediction pipeline.
The second variant of the UNet model utilized the soft normalized data and followed a similar pattern of experiments as the first variant. The Dice score for IVCF was 0.71 for training and 0.70 for validation data. Here the Dice scores were lower than the first variant. Furthermore, 17.26% of the total IVCF pixels were classified as background and false positives was around 0.15% of the background pixels.
Figure 8 is a sample prediction of CT image slices with IVCF. Column two shows the hard normalized segmentation followed by column three with the soft normalized segmentation outcome. As it can be observed, the segmentation of hard and soft normalization are very comparable to each other. Due to the better Dice scores for hard normalization, it was selected to be used in the IVCF prediction pipeline.
Since the binary segmentation is highly imbalanced, precision-recall curves were generated over ROC plots. These results shown in Figure 9. Comparing the Dice scores for soft against hard normalization revealed that hard normalization is able to help the UNet model to attain precision in its segmentation approach.

4.2. IVCF Prediction Pipeline Evaluation

The prediction algorithm was then used on the segmentation output from hard normalized CT scans. Using Algorithm 1, several variations of sig_count and sequence were used to test the IVCF classification accuracy on all 90 CT scans with IVCF. The combination of three of the most significant parameters are shown in Table 2. Confusion matrices in Figure 10 represents this information as true positives. We noticed that increasing the number of sequential CT slices reduces the classification accuracy of the prediction algorithm. This is mostly due to scanner heterogeneity as some scans with a higher resolution show more of the IVCF profile. A 92.22% detection accuracy from the 90 IVCF scans was observed using sig_count = 300 IVCF pixels and sequence = 7.
A validation test was conducted using 90 normal CT scans that were not used during segmentation phase. We noticed some limitations with regards to false positives as some of the scans were classified as IVCF. Confusion matrices in Figure 10 show that the highest number of false positives were 13 when using a sequence of 5 but only 3 when using sequence of 9. Upon closer inspection it was observed that these scans were associated with calcification around spinal foramen region near the IVCF. Since it assumes a ring-like appearance the segmentation model would mislabel the pixels as IVCF. This limitation requires further investigation. We also noticed superior performance of the segmentation model as none of bone or surgical implants were segmented as false positives even though they share similar signature as the IVCF.
To summarize, results from the proposed classifier indicates that the optimum combination for the current dataset is sig_count = 300 IVCF pixels and sequence = 7 sequential slices in a CT scan. Using this combination produces we achieved a 91.67% overall accuracy using 180 CT scans. Only 8 normal scans are flagged as patients having IVCF. A detailed metric from this analysis is shown in Table 2. Other combinations of sig_count and sequence were also explored to validate the efficacy of the proposed approach as shown in Table A2. It was noticeable that decreasing the sig_count to 200 and sequence to 5 produced a 98.8% accuracy in detecting IVCF in scans. However, it also increased the number of false positives.

5. Conclusions

In this research we have used concepts of deep learning to create an IVCF detection pipeline. Two UNet models were trained to identify the foundation of our segmentation approach. A prediction algorithm for flagging CT scans with IVCF was constructed. The pipeline returned promising results with prediction time per scan being as low as 43 seconds. Future work will include training the model on a much larger dataset to investigate the possibility of reducing false positives caused by calcification around the spine. The feasibility of model deployment in a testing phase is being explored at Mayo Clinic. Preliminary results show promise for future improvements.

Author Contributions

Conceptualization, methodology, and supervision, R.G. and J.W.; validation, R.G., J.W. and C.K. data curation and software, J.W., R.G., C.S., P.D.M. and C.K.; writing—original draft preparation, J.W., R.G., P.D.M. and J.L.; writing—review and editing, R.G. and J.W.; visualization, R.G., C.S., P.D.M. and C.K; funding acquisition, J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Mayo-UWEC Research Innovation Council. The computational resources of the study were provided by the Blugold Center for High-Performance Computing under NSF grant CNS-1920220.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of Mayo Clinic (#19-001923 on 27 May 2019). The University of Wisconsin-Eau Claire IRB has approved this study as Exempt (#202120049) based on the Code of Federal Regulations 45 Part 46.104(d)(4)(ii) on 28 April 2021.

Informed Consent Statement

Not applicable.

Data Availability Statement

Links to our scripts used for analysis can be found here https://github.com/rahulgomes19/IVC_2D (accessed on 14 September 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
IVCFInferior Vena Cana Filters
HUHounsfield Unit
IRBInstitutional Review Board
ReLURectified Linear Unit
VTEVenous thromboembolism

Appendix A

The image augmentation algorithm used while training the UNet model is shown in Algorithm A1.
Algorithm A1: IVCF image augmentation algorithm for the deep learning model
Diagnostics 12 02475 i002
The architecture of the proposed UNet model is shown in Table A1.
Table A1. Modified UNet architecture for generating IVCF segmentation map. Sequential layers are represented as downsample_x, and upsample_x, where ’x’ represents layer number.
Table A1. Modified UNet architecture for generating IVCF segmentation map. Sequential layers are represented as downsample_x, and upsample_x, where ’x’ represents layer number.
LayerOutput ShapeParam #
input_1(20, 256, 256, 1)0
downsample_1(20, 128, 128, 64)1024
downsample_2(20, 64, 64, 128)131,584
downsample_3(20, 32, 32, 256)525,312
downsample_4(20, 16, 16, 512)2,099,200
downsample_5(20, 8, 8, 1024)8,392,704
upsample_1(20, 16, 16, 512)8,390,656
concatenate_1(20, 16, 16, 1024)0
upsample_2(20, 32, 32, 256)4,195,328
concatenate_2(20, 32, 32, 512)0
upsample_3(20, 64, 64, 128)1,049,088
concatenate_3(20, 64, 64, 256)0
upsample_4(20, 128, 128, 64)262,400
concatenate_4(20, 128, 128, 128)0
conv2d_
transpose_1
(20, 256, 256, 64)131,136
conv2d_1(20, 256, 256, 2)2050
Total Parameters25,180,482
Table A2. Prediction pipeline performance for hard normalized segmentation using a combination of different parameters.
Table A2. Prediction pipeline performance for hard normalized segmentation using a combination of different parameters.
Sequence
Number,
Sig_Count
IVCF ScansNormal Scans
Scans
Flagged with
IVCF
(Best Is 90)
% Scans
Flagged
Correctly
Scans Not
Flagged
with IVCF
(Best Is 90)
% Scans
Flagged
as Normal
(5, 200)8998.87280
(7, 200)8998.87886.67
(9, 200)8291.18392.22
(5, 300)8695.567785.56
(7, 300)8392.228291.11
(9, 300)7482.228796.67
(5, 400)7987.788190
(7, 400)6976.678796.67
(9, 400)5561.118796.67

References

  1. Uberoi, R.; Tapping, C.R.; Chalmers, N.; Allgar, V. British Society of Interventional Radiology (BSIR) inferior vena cava (IVC) filter registry. Cardiovasc. Interv. Radiol. 2013, 36, 1548–1561. [Google Scholar] [CrossRef] [Green Version]
  2. Salei, A.; Raborn, J.; Manapragada, P.P.; Stoneburner, C.G.; Aal, A.K.A.; Gunn, A.J. Effect of a dedicated inferior vena cava filter retrieval program on retrieval rates and number of patients lost to follow-up. Diagn. Interv. Radiol. 2020, 26, 40. [Google Scholar] [CrossRef] [PubMed]
  3. Kaufman, J.A. Retrievable vena cava filters. Tech. Vasc. Interv. Radiol. 2004, 7, 96–104. [Google Scholar] [CrossRef] [PubMed]
  4. Jia, Z.; Fuller, T.A.; McKinney, J.M.; Paz-Fumagalli, R.; Frey, G.T.; Sella, D.M.; Van Ha, T.; Wang, W. Utility of retrievable inferior vena cava filters: A systematic literature review and analysis of the reasons for nonretrieval of filters with temporary indications. Cardiovasc. Interv. Radiol. 2018, 41, 675–682. [Google Scholar] [CrossRef] [PubMed]
  5. Durack, J.C.; Westphalen, A.C.; Kekulawela, S.; Bhanu, S.B.; Avrin, D.E.; Gordon, R.L.; Kerlan, R.K. Perforation of the IVC: Rule rather than exception after longer indwelling times for the Günther Tulip and Celect retrievable filters. Cardiovasc. Interv. Radiol. 2012, 35, 299–308. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, W.; Zhou, D.; Obuchowski, N.; Spain, J.; An, T.; Moon, E. Fracture and migration of Celect inferior vena cava filters: A retrospective review of 741 consecutive implantations. J. Vasc. Interv. Radiol. 2013, 24, 1719–1722. [Google Scholar] [CrossRef] [PubMed]
  7. Medwatch, F. Filters: Initial Communication: Risk of Adverse Events with Long Term Use. 2010. Available online: https://www.eeworldonline.com/inferior-vena-cava-ivc-filters-initial-communication-risk-of-adverse-events-with-long-term-use/ (accessed on 10 August 2022 ).
  8. Ahmed, O.; Jilani, S.; Heussner, D.; Khan, M. Trapped by controversy: Inferior vena cava filters and the law. J. Vasc. Interv. Radiol. 2017, 6, 886–888. [Google Scholar] [CrossRef] [PubMed]
  9. Gyang, E.; Zayed, M.; Harris, E.J.; Lee, J.T.; Dalman, R.L.; Mell, M.W. Factors impacting follow-up care after placement of temporary inferior vena cava filters. J. Vasc. Surg. 2013, 58, 440–445. [Google Scholar] [CrossRef] [Green Version]
  10. Grewal, S.; Chamarthy, M.R.; Kalva, S.P. Complications of inferior vena cava filters. Cardiovasc. Diagn. Ther. 2016, 6, 632. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Zheng, A.; Casari, A. Feature Engineering for Machine Learning: Principles and Techniques for Data Scientists; O’Reilly Media, Inc.: Sebastopol, CA, USA, 2018. [Google Scholar]
  12. Nargesian, F.; Samulowitz, H.; Khurana, U.; Khalil, E.B.; Turaga, D.S. Learning Feature Engineering for Classification. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2529–2535. [Google Scholar]
  13. Gomes, R.; Paul, N.; He, N.; Huber, A.F.; Jansen, R.J. Application of Feature Selection and Deep Learning for Cancer Prediction Using DNA Methylation Markers. Genes 2022, 13, 1557. [Google Scholar] [CrossRef]
  14. Ahsan, M.; Gomes, R.; Chowdhury, M.M.; Nygard, K.E. Enhancing Machine Learning Prediction in Cybersecurity Using Dynamic Feature Selector. J. Cybersecur. Priv. 2021, 1, 199–218. [Google Scholar] [CrossRef]
  15. Domingos, P. A few useful things to know about machine learning. Commun. ACM 2012, 55, 78–87. [Google Scholar] [CrossRef] [Green Version]
  16. Jogin, M.; Madhulika, M.; Divya, G.; Meghana, R.; Apoorva, S. Feature extraction using convolution neural networks (CNN) and deep learning. In Proceedings of the 2018 3rd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bengaluru, India, 18–19 May 2018; pp. 2319–2323. [Google Scholar]
  17. Dara, S.; Tumma, P. Feature extraction by using deep learning: A survey. In Proceedings of the 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 29–31 March 2018; pp. 1795–1801. [Google Scholar]
  18. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–27 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  19. O’Shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  20. Ahsan, M.; Gomes, R.; Denton, A. Application of a convolutional neural network using transfer learning for tuberculosis detection. In Proceedings of the 2019 IEEE International Conference on Electro Information Technology (EIT), Brookings, SD, USA, 20–22 May 2019; pp. 427–433. [Google Scholar]
  21. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  22. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  23. Markham, B.L.; Barker, J.L. Spectral characterization of the Landsat Thematic Mapper sensors. Int. J. Remote Sens. 1985, 6, 697–716. [Google Scholar] [CrossRef] [Green Version]
  24. Shah, M.; Alnabelsi, T.; Patil, S.; Reddy, S.; Patel, B.; Lu, M.; Chandorkar, A.; Perelas, A.; Arora, S.; Patel, N.; et al. IVC filters—Trends in placement and indications, a study of 2 populations. Medicine 2017, 96, e6449. [Google Scholar] [CrossRef]
  25. Henry, J.; Pylypchuk, Y.; Searcy, T.; Patel, V. Adoption of electronic health record systems among US non-federal acute care hospitals: 2008–2015. ONC Data Brief 2016, 35, 2008–2015. [Google Scholar]
  26. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  27. Kim, M.; Yun, J.; Cho, Y.; Shin, K.; Jang, R.; Bae, H.J.; Kim, N. Deep learning in medical imaging. Neurospine 2019, 16, 657. [Google Scholar] [CrossRef] [Green Version]
  28. Setio, A.A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; Van Riel, S.J.; Wille, M.M.W.; Naqibullah, M.; Sánchez, C.I.; Van Ginneken, B. Pulmonary nodule detection in CT images: False positive reduction using multi-view convolutional networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  30. Siddique, N.; Paheding, S.; Elkin, C.P.; Devabhaktuni, V. U-net and its variants for medical image segmentation: A review of theory and applications. IEEE Access 2021, 9, 82031–82057. [Google Scholar] [CrossRef]
  31. Nisar, H.; Carnahan, P.K.; Fakim, D.; Akhuanzada, H.; Hocking, D.; Peters, T.M.; Chen, E.C. Towards ultrasound-based navigation: Deep learning based IVC lumen segmentation from intracardiac echocardiography. In Medical Imaging 2022: Image-Guided Procedures, Robotic Interventions, and Modeling; SPIE: Bellingham, WA, USA, 2022; Volume 12034, pp. 467–476. [Google Scholar]
  32. Alakwaa, W.; Nassef, M.; Badr, A. Lung cancer detection and classification with 3D convolutional neural network (3D-CNN). Lung Cancer 2017, 8, 409. [Google Scholar] [CrossRef] [Green Version]
  33. Liao, F.; Liang, M.; Li, Z.; Hu, X.; Song, S. Evaluate the malignancy of pulmonary nodules using the 3-d deep leaky noisy-or network. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3484–3495. [Google Scholar] [CrossRef] [PubMed]
  34. Hooda, R.; Sofat, S.; Kaur, S.; Mittal, A.; Meriaudeau, F. Deep-learning: A potential method for tuberculosis detection using chest radiography. In Proceedings of the 2017 IEEE International Conference on Signal and Image Processing Applications (ICSIPA), Kuching, Malaysia, 12–14 September 2017; pp. 497–502. [Google Scholar]
  35. Kant, S.; Srivastava, M.M. Towards automated tuberculosis detection using deep learning. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 1250–1253. [Google Scholar]
  36. Liu, B.; Liu, Y.; Pan, X.; Li, M.; Yang, S.; Li, S.C. DNA methylation markers for pan-cancer prediction by deep learning. Genes 2019, 10, 778. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Diamant, A.; Chatterjee, A.; Vallières, M.; Shenouda, G.; Seuntjens, J. Deep learning in head & neck cancer outcome prediction. Sci. Rep. 2019, 9, 2764. [Google Scholar]
  38. Tiwari, M.; Bharuka, R.; Shah, P.; Lokare, R. Breast Cancer Prediction Using Deep Learning and Machine Learning Techniques. 2020. Social Science Research Network. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3558786 (accessed on 1 August 2022).
  39. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A semi-supervised deep learning method based on stacked sparse auto-encoder for cancer prediction using RNA-seq data. Comput. Methods Programs Biomed. 2018, 166, 99–105. [Google Scholar] [CrossRef] [PubMed]
  40. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A deep learning-based multi-model ensemble method for cancer prediction. Comput. Methods Programs Biomed. 2018, 153, 1–9. [Google Scholar] [CrossRef] [PubMed]
  41. Shorten, C.; Khoshgoftaar, T.M.; Furht, B. Deep Learning applications for COVID-19. J. Big Data 2021, 8, 18. [Google Scholar] [CrossRef]
  42. Gomes, R.; Kamrowski, C.; Langlois, J.; Rozario, P.; Dircks, I.; Grottodden, K.; Martinez, M.; Tee, W.Z.; Sargeant, K.; LaFleur, C.; et al. A Comprehensive Review of Machine Learning Used to Combat COVID-19. Diagnostics 2022, 12, 1853. [Google Scholar] [CrossRef]
  43. El-Rashidy, N.; Abdelrazik, S.; Abuhmed, T.; Amer, E.; Ali, F.; Hu, J.W.; El-Sappagh, S. Comprehensive survey of using machine learning in the COVID-19 pandemic. Diagnostics 2021, 11, 1155. [Google Scholar] [CrossRef]
  44. Dairi, A.; Harrou, F.; Zeroual, A.; Hittawe, M.M.; Sun, Y. Comparative study of machine learning methods for COVID-19 transmission forecasting. J. Biomed. Inform. 2021, 118, 103791. [Google Scholar] [CrossRef]
  45. Hong, J.; Zhang, Y.D.; Chen, W. Source-free unsupervised domain adaptation for cross-modality abdominal multi-organ segmentation. Knowl.-Based Syst. 2022, 250, 109155. [Google Scholar] [CrossRef]
  46. Hong, J.; Yu, S.C.H.; Chen, W. Unsupervised domain adaptation for cross-modality liver segmentation via joint adversarial learning and self-learning. Appl. Soft Comput. 2022, 121, 108729. [Google Scholar] [CrossRef]
  47. Hong, J.; Cheng, H.; Wang, S.H.; Liu, J. Improvement of cerebral microbleeds detection based on discriminative feature learning. Fundam. Inform. 2019, 168, 231–248. [Google Scholar] [CrossRef]
  48. Ni, J.C.; Shpanskaya, K.; Han, M.; Lee, E.H.; Do, B.H.; Kuo, W.T.; Yeom, K.W.; Wang, D.S. Deep learning for automated classification of inferior vena cava filter types on radiographs. J. Vasc. Interv. Radiol. 2020, 31, 66–73. [Google Scholar] [CrossRef]
  49. Park, B.J.; Sotirchos, V.S.; Adleberg, J.; Stavropoulos, S.W.; Cook, T.S.; Hunt, S.J. Feasibility and visualization of deep learning detection and classification of inferior vena cava filters. medRxiv 2020. [Google Scholar] [CrossRef]
  50. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  51. Vesal, S.; Ravikumar, N.; Maier, A. SkinNet: A deep learning framework for skin lesion segmentation. In Proceedings of the 2018 IEEE Nuclear Science Symposium and Medical Imaging Conference Proceedings (NSS/MIC), Sydney, Australia, 10–17 November 2018; pp. 1–3. [Google Scholar]
  52. Gomes, R.; Rozario, P.; Adhikari, N. Deep learning optimization in remote sensing image segmentation using dilated convolutions and ShuffleNet. In Proceedings of the 2021 IEEE International Conference on Electro Information Technology (EIT), Mt. Pleasant, MI, USA, 14–15 May 2021; pp. 244–249. [Google Scholar]
  53. Irwin, S.; Feldman, G. An isotropic 3 × 3 image gradient operator. Present. Stanf. Proj. 2014. Available online: https://www.researchgate.net/publication/281104656_An_Isotropic_3x3_Image_Gradient_Operator (accessed on 10 August 2022).
  54. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  55. Van der Walt, S.; Schönberger, J.L.; Nunez-Iglesias, J.; Boulogne, F.; Warner, J.D.; Yager, N.; Gouillart, E.; Yu, T. Scikit-image: Image processing in Python. PeerJ 2014, 2, e453. [Google Scholar] [CrossRef]
  56. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the proposed approach to IVCF detection.
Figure 1. Schematic diagram of the proposed approach to IVCF detection.
Diagnostics 12 02475 g001
Figure 2. Spatial resolution of 180 CT scans used in the research where (a) Slice Thickness and (b) Slice Width.
Figure 2. Spatial resolution of 180 CT scans used in the research where (a) Slice Thickness and (b) Slice Width.
Diagnostics 12 02475 g002
Figure 3. CT Image Slice Count Distribution before and after removal of scans 40 cm below the craniocaudal region.
Figure 3. CT Image Slice Count Distribution before and after removal of scans 40 cm below the craniocaudal region.
Diagnostics 12 02475 g003
Figure 4. Example of CT scan slices along the axial plane with IVCF. Top row (a,b) are without application of intensity adjustment (Soft Normalization). Scans (c,d) have HU values adjusted by setting minimum as 1 and maximum as 2500 (Hard Normalization).
Figure 4. Example of CT scan slices along the axial plane with IVCF. Top row (a,b) are without application of intensity adjustment (Soft Normalization). Scans (c,d) have HU values adjusted by setting minimum as 1 and maximum as 2500 (Hard Normalization).
Diagnostics 12 02475 g004
Figure 5. Image augmentation during training process for a CT image slice. These augmentations like random cropping, rotation, and flip reduces spatial bias introduced by the dataset. The masks also undergo similar modifications during training.
Figure 5. Image augmentation during training process for a CT image slice. These augmentations like random cropping, rotation, and flip reduces spatial bias introduced by the dataset. The masks also undergo similar modifications during training.
Diagnostics 12 02475 g005
Figure 6. UNet model architecture used for the training phase to generate a segmented map of the IVCF in the CT image slices. The downsampling and upsampling phase were evaluated as separate functions and called using the gradient tape function in TensorFlow. Numbers indicate the (width, height, kernels) used in the layer. Arrows indicate the operation performed while training.
Figure 6. UNet model architecture used for the training phase to generate a segmented map of the IVCF in the CT image slices. The downsampling and upsampling phase were evaluated as separate functions and called using the gradient tape function in TensorFlow. Numbers indicate the (width, height, kernels) used in the layer. Arrows indicate the operation performed while training.
Diagnostics 12 02475 g006
Figure 7. IoU metric evaluation (Dice Score) for IVCF pixels in the CT images. Image by Adrian Rosebrock, distributed under a CC BY-SA 4.0 license.
Figure 7. IoU metric evaluation (Dice Score) for IVCF pixels in the CT images. Image by Adrian Rosebrock, distributed under a CC BY-SA 4.0 license.
Diagnostics 12 02475 g007
Figure 8. Segmentation output from UNet model variants on a CT scan using hard normalized data.
Figure 8. Segmentation output from UNet model variants on a CT scan using hard normalized data.
Diagnostics 12 02475 g008
Figure 9. Precision-Recall curves for UNet segmentation approach using (a) UNet Model 1 and (b) UNet Model 2.
Figure 9. Precision-Recall curves for UNet segmentation approach using (a) UNet Model 1 and (b) UNet Model 2.
Diagnostics 12 02475 g009
Figure 10. Confusion matrices obtained from IVCF prediction algorithm using (a) sig_count = 300, sequence = 5, (b) sig_count = 300, sequence = 7 and (c) sig_count = 300, sequence = 9. Here 1 represents IVCF scan and 0 represents normal scan.
Figure 10. Confusion matrices obtained from IVCF prediction algorithm using (a) sig_count = 300, sequence = 5, (b) sig_count = 300, sequence = 7 and (c) sig_count = 300, sequence = 9. Here 1 represents IVCF scan and 0 represents normal scan.
Diagnostics 12 02475 g010
Table 1. Dice score for different model variants.
Table 1. Dice score for different model variants.
UNet
Model 
NormalizeDatasetDice Scores
BackgroundIVCF
1HardTraining0.99810.8168
Validation0.99790.7981
2SoftTraining0.99690.7153
Validation0.99700.7082
Table 2. Prediction pipeline performance for IVCF detection metrics from Figure 9.
Table 2. Prediction pipeline performance for IVCF detection metrics from Figure 9.
CombinationAccuracySensitivitySpecificityPrecisionF1 Score
300, 590.56%0.95560.85560.86870.9101
300, 791.67%0.920.91110.91210.9171
300, 989.44%0.82220.96670.9610.8862
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gomes, R.; Kamrowski, C.; Mohan, P.D.; Senor, C.; Langlois, J.; Wildenberg, J. Application of Deep Learning to IVC Filter Detection from CT Scans. Diagnostics 2022, 12, 2475. https://doi.org/10.3390/diagnostics12102475

AMA Style

Gomes R, Kamrowski C, Mohan PD, Senor C, Langlois J, Wildenberg J. Application of Deep Learning to IVC Filter Detection from CT Scans. Diagnostics. 2022; 12(10):2475. https://doi.org/10.3390/diagnostics12102475

Chicago/Turabian Style

Gomes, Rahul, Connor Kamrowski, Pavithra Devy Mohan, Cameron Senor, Jordan Langlois, and Joseph Wildenberg. 2022. "Application of Deep Learning to IVC Filter Detection from CT Scans" Diagnostics 12, no. 10: 2475. https://doi.org/10.3390/diagnostics12102475

APA Style

Gomes, R., Kamrowski, C., Mohan, P. D., Senor, C., Langlois, J., & Wildenberg, J. (2022). Application of Deep Learning to IVC Filter Detection from CT Scans. Diagnostics, 12(10), 2475. https://doi.org/10.3390/diagnostics12102475

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop