Next Article in Journal
Evaluation of Respiratory Particle Emission during Otorhinolaryngological Procedures in the Context of the SARS-CoV-2 Pandemic
Next Article in Special Issue
An Enhanced Transfer Learning Based Classification for Diagnosis of Skin Cancer
Previous Article in Journal
Comparison of Simple RNA Extraction Methods for Molecular Diagnosis of Hepatitis C Virus in Plasma
Previous Article in Special Issue
A Radiation-Free Classification Pipeline for Craniosynostosis Using Statistical Shape Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Are We There Yet? The Value of Deep Learning in a Multicenter Setting for Response Prediction of Locally Advanced Rectal Cancer to Neoadjuvant Chemoradiotherapy

1
Department of Diagnostic and Interventional Radiology, University Hospital Bonn, 53127 Bonn, Germany
2
Computer Assisted Clinical Medicine, Mannheim Institute for Intelligent Systems in Medicine, Medical Faculty Mannheim, Heidelberg University, 68167 Mannheim, Germany
3
Data Analysis and Modeling, Mannheim Institute for Intelligent Systems in Medicine (MIISM), Medical School Mannheim, Central Institute for Scientific Computing (IWR), Central Institute for Computer Engineering (ZITI), CZS Heidelberg Center for Model-Based AI, Heidelberg University, 69047 Heidelberg, Germany
4
Clinical Functional Imaging, Department of Diagnostic and Interventional Radiology, University Hospital Bonn, 53127 Bonn, Germany
5
Department of Radiotherapy and Oncology, University Hospital Frankfurt, 60596 Frankfurt am Main, Germany
6
Department of Medicine III, Medical Faculty Mannheim, Heidelberg University, 68167 Mannheim, Germany
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(7), 1601; https://doi.org/10.3390/diagnostics12071601
Submission received: 14 June 2022 / Revised: 26 June 2022 / Accepted: 28 June 2022 / Published: 30 June 2022
(This article belongs to the Special Issue Artificial Intelligence in Clinical Medical Imaging Analysis)

Abstract

:
This retrospective study aims to evaluate the generalizability of a promising state-of-the-art multitask deep learning (DL) model for predicting the response of locally advanced rectal cancer (LARC) to neoadjuvant chemoradiotherapy (nCRT) using a multicenter dataset. To this end, we retrained and validated a Siamese network with two U-Nets joined at multiple layers using pre- and post-therapeutic T2-weighted (T2w), diffusion-weighted (DW) images and apparent diffusion coefficient (ADC) maps of 83 LARC patients acquired under study conditions at four different medical centers. To assess the predictive performance of the model, the trained network was then applied to an external clinical routine dataset of 46 LARC patients imaged without study conditions. The training and test datasets differed significantly in terms of their composition, e.g., T-/N-staging, the time interval between initial staging/nCRT/re-staging and surgery, as well as with respect to acquisition parameters, such as resolution, echo/repetition time, flip angle and field strength. We found that even after dedicated data pre-processing, the predictive performance dropped significantly in this multicenter setting compared to a previously published single- or two-center setting. Testing the network on the external clinical routine dataset yielded an area under the receiver operating characteristic curve of 0.54 (95% confidence interval [CI]: 0.41, 0.65), when using only pre- and post-therapeutic T2w images as input, and 0.60 (95% CI: 0.48, 0.71), when using the combination of pre- and post-therapeutic T2w, DW images, and ADC maps as input. Our study highlights the importance of data quality and harmonization in clinical trials using machine learning. Only in a joint, cross-center effort, involving a multidisciplinary team can we generate large enough curated and annotated datasets and develop the necessary pre-processing pipelines for data harmonization to successfully apply DL models clinically.

1. Introduction

Colorectal cancer is the third most common cancer, accounting for 10% of all diagnosed cancers, and is the second leading cause of cancer death worldwide [1]. Rectal cancer constitutes about one-third of colorectal malignancies [2,3]. Within the last two decades, optimized imaging has enabled significant advances toward stage-specific and individualized therapy options for patients with rectal cancer.
The current standard-of-care treatment for locally advanced rectal cancer (LARC) involves neoadjuvant radiotherapy or chemoradiotherapy (nCRT) followed by total mesorectal excision (TME) [4]. Beyond reducing the local recurrence rate, improving resectability, and potentially preserving sphincter function in tumors of the lower rectum third, nCRT harbors the potential for an organ-preserving approach in patients exhibiting a clinical complete remission (cCR) [5]. Using standard chemoradiotherapy in locally advanced tumors, about 10% of patients demonstrate cCR [6]. Total neoadjuvant therapy (TNT), comprising 3–4.5 months of additional chemotherapy with 5-fluorouraicl and oxaliplatin before or after radio(chemo)therapy, increases the rate of cCR to about 30% [7]. Besides identifying patients who qualify for a watch-and-wait strategy, the reliable detection of treatment response is valuable for assessing overall survival [8,9].
Accurate staging/restaging using magnetic resonance imaging (MRI) to identify patients with a pathological complete response (pCR) prior to surgical resection is essential. cCR is used as a surrogate for pCR. However, treatment response assessment after nCRT by MRI is challenging and often entails overstaging due to mixed signal intensities or irregular fibrosis on T2-weighted (T2w) images and false positive residual lymph nodes [10]. The MRI-derived tumor regression grade (TRG) exhibits high specificity (93%) for pCR, but suboptimal sensitivity (32%) [11].
Considering these limitations of morphological MRI in evaluating treatment response, several studies have since investigated the potential of functional MRI like perfusion imaging [12,13] and, in particular, diffusion-weighted imaging (DWI) [14,15]. While DWI has been shown to increase diagnostic accuracy in detecting lymph node metastases and circumferential resection margin (CRM) involvement, its value in predicting treatment response is controversial [16,17,18]. Several models have been devised to quantitatively correlate the DWI signal with microstructural features of healthy and cancerous tissue in the rectum.
The current clinical standard is the so-called apparent diffusion coefficient (ADC), which simply describes the DWI signal as decaying monoexponentially with increasing diffusion weighting [19]. Low pretherapeutic ADC values and an early increase in mean tumor ADC values during therapy were found to correlate with a good treatment response [17,20,21,22,23]. However, other studies failed to reproduce this correlation or even demonstrated an inverse correlation linking pretherapeutic high ADC values to a good response to nCRT and initial low ADC values to increased tumor aggressiveness [14,17,24,25,26,27,28,29,30]. Focal diffusion restriction at DWI may be observed in up to 50% of patients at the site of proven pCR [10]. Furthermore, there is a high overlap of ADC values in patients with pCR and partial regression [31].
With the increasing availability of graphical processing units, the application of Deep Learning (DL) to train neural networks holds a very promising, now feasible approach to improve the hitherto insufficient characterization of therapy response relying on T2w and DWI [6]. In a single-center study with external validation, Jin et al. designed two Siamese subnetworks joined at multiple layers to allow for simultaneous tumor segmentation and response prediction [32]. After training their model using pre- and post-therapeutic T2w, diffusion-weighted (DW) and T1w with and without contrast enhancement images of 321 Asian LARC patients from one center, they prospectively obtained a dataset of 160 Asian LARC patients treated at the same center for internal validation. In an independent, external validation cohort of 141 Asian LARC patients from another institution, they achieved high accuracy in predicting pCR with an area under the receiver operating characteristic curve (AUC) of 0.92 (95% confidence interval (CI): 0.87–0.96). Integrating blood-based tumor markers further improved prediction accuracy to an AUC of 0.97 (95% CI: 0.93–0.99).
Before translating such DL approaches into clinical routine for screening, treatment response assessment and surveillance, their predictive performance remains to be verified in a multicenter setting with heterogeneous data of different ethnic groups. Facing the challenge of varying data quality and uniformity, in this study we test a DL network similar to the approach by Jin et al. [32] on a multicenter dataset from a Western population. To ensure clinical feasibility, we train and validate our network using only pre- and post-therapeutic T2w images, DW images at b = 800 s/mm2 and ADC-maps of 83 LARC patients from four different centers and test our trained network on an independent, external, clinical routine dataset of 46 LARC patients. In this work, we aim to share our initial experiences and, more importantly, challenges of applying a state-of-the-art DL model to a multicenter dataset. Only through international collaborations and joint efforts will we achieve broad applicability of DL that will have a clinical impact.

2. Materials and Methods

2.1. Patient Cohort

With the approval of the institutional review board, this retrospective study enrolled 162 LARC patients (59 female mean age 63 years +/− 11 years 2 standard deviations, 103 male mean age 60 years +/− 10 years 2 standard deviations) of the CAO-ARO-AIO-12 study [33,34] acquired between the years 2015 and 2018 (Figure 1). The following centers participated in this study: Center 1 (Frankfurt), Center 2 (Regensburg), Center 3 (Würzburg), Center 4a (Mannheim). Furthermore, a clinical routine dataset acquired at Center 4b (Mannheim) from 2009 to 2013 without study conditions of 46 LARC patients was retrospectively included as an external test dataset in our DL approach.
The inclusion criteria of this study were: (a) histopathologically proven LARC, (b) treatment with nCRT followed by TME and pathological assessment of the TRG, (c) availability of pre- and post-nCRT MRI examinations including axial, high-resolution T2w- and DWI-sequences with the calculation of ADC maps. Patients had to be excluded if their MRI dataset was either incomplete or the pathologic TRG was not available.

2.2. Image Acquisition and Radiologic Assessment

All patients underwent a clinically indicated MRI examination before and after nCRT. As per the study protocol, MR images were acquired with at least a 1.5 T, preferably a 3.0 T MRI scanner in the supine position and with a phased-array body coil wrapped on the surface of the patient to cover the entire pelvis. Suggested parameters for the MR sequences relevant to this study were:
Axial T2w-TSE: Repetition time (TR) 5600 ms, echo time (TE) 110 ms, max. 3 mm slice thickness, voxel size 0.8 × 0.8 × 3 mm.
Axial DWI-EPI: TR 6500 ms, TE 76 ms, max. 5 mm slice thickness, voxel size 1.8 × 1.8 × 4 mm, acquired b-values 50/400/800 s/mm2 and a calculated b-value of 1400 s/mm2.
All MRI examinations before and after nCRT were clinically assessed and evaluated by trained, board-certified radiologists at the respective centers prior to surgery, avoiding knowledge of histopathology.
The tumor region on pre- and post-therapeutic, axial, high-resolution T2w images was segmented by a trained radiologist with 3 years of experience using the freely available, open-source software ITK-SNAP [35,36].

2.3. Pathological Assessment of Tumor Regression

After TME, the surgically resected tumor specimens were histopathologically examined and pathologic TRG was determined according to the Dworak classification [37]. The grading system was established as follows: TRG 0—no regression; TRG 1—dominant tumor mass with obvious fibrosis and/or vasculopathy; TRG 2—dominantly fibrotic changes with few tumor cells or groups (easy to find); TRG 3—very few (difficult to find microscopically) tumor cells in fibrotic tissue with or without mucous substance; TRG 4—no tumor cells, only fibrotic mass (total regression or response) [37].

2.4. Data Processing

Three image processing and harmonization steps were applied toward enhancing the robustness of the model. To correct for B1 field inhomogeneities, bias correction was performed using the N4 bias correction [38] implemented in SimpleITK [39,40]. Subsequently, rigid image registration of pre- and post-therapeutic T2w images was performed using Advanced Neuroimaging Tools (ANTs) [41] with standard parameters and without a mask. Their exact transformation was then applied to the DW images with SimpleITK. Alignment between T2w and DW images was visually assessed. Center-cropping and normalization was performed using the code provided by Jin et al. [32].

2.5. Deep Learning Model

For the classification task at hand, we used the approach developed by Jin et al. [32]. The code corresponding to the paper was taken from the repository and only the number of input channels was changed to match our training data of pre- and post-therapeutic T2w and DW images. The model was trained in two configurations: (1) using only T2w images as input, (2) using T2w and DW images comprising a b800 DW image and an ADC map as input.
The model utilizes multitask learning to improve its performance. Two U-Nets with shared parameters are used to segment the tumor in the pre- and post-therapeutic images. Features are then extracted from both networks at multiple levels. To evaluate treatment response, differences between these features are used for classification after applying a depth-wise convolution. Patients were classified into two classes: pCR (TRG 4) and non-pCR (TRG 0–3).

2.6. Training

The model was trained on all patients of Centers 1, 2, and 3 as described by Jin et al. [32], i.e., for 1000 epochs using focal loss and an initial learning rate of 0.01, which was halved when the loss did not improve for 30 epochs. To prevent overfitting, the weights of the epoch with the highest AUC on the validation set were used. An exponential moving average was applied to the validation score to remove random fluctuations. Due to the small size of the dataset, the training cohort was evaluated using 5-fold cross-validation. For each fold, the data of Centers 1, 2, and 3 were split into 68% for training, 12% for validation and 20% for testing, resulting in five trained networks. To test the performance of our approach, the data of Center 4a, which were previously unseen by the network, were applied to each of the previously trained networks for prediction. To evaluate model generalizability, the network was tested on a clinical routine dataset of 46 LARC patients from Center 4b, which were outside the training distribution (Figure 2). Again, these data were applied on the five previously trained networks for prediction.

2.7. Statistical Analysis

The confidence interval for the AUC of each fold was calculated using bootstrapping with 1000 samples. Random samples (allowing for duplicates), with the same number of patients as the original set, were drawn from the scores predicted by the network. Samples that did not include patients with pCR were skipped. The results were used for error calculation of the AUC.
The fast implementation of De Long’s method was used to calculate if there were significant differences in the AUC between the models [42]. The significance of the differences in patients’ characteristics was calculated using the Mann–Whitney U-Test with the null hypothesis that the distributions are equal. A p-value of 0.05 was considered significant.

3. Results

3.1. Patients’ Characteristics

There were significant differences in the distribution of patients’ characteristics, including age, pre- and post-nCRT MRI staging, tumor localization and time interval between initial MRI staging, re-staging and the date of the operation. For a comprehensive overview of patients’ characteristics in the training, validation and test cohorts, please refer to Table 1.

3.2. Imaging Characteristics

Apart from the clinical routine test dataset, various scanners of different manufacturers with dissimilar field strengths were used at each center, and in many cases the same patient was examined on differing scanners before and after nCRT. In addition, over the course of the CAO-ARO-AIO-12 study, selected scanners were upgraded, e.g., the Avanto scanner at Center 1 was upgraded to the Avanto-fit scanner. Assuming that each center used only one scanner with the same model name, 15 different scanners were used at 4 centers (see Table 2).
Furthermore, we observed significant differences in imaging parameters between the training/validation cohort and the test cohort, some of which also differed significantly from the prespecified study protocol (Table 3). Figure 2 exemplifies the distribution of selected parameters in T2w imaging across the different centers. It is of note that even within a single-center, parameters sometimes varied significantly. Consequently, the imaging data of the included patients from the CAO-ARO-AIO-12 study were very heterogeneous in terms of image quality.
Figure 2. Distribution of selected imaging parameters per center for the T2w images.
Figure 2. Distribution of selected imaging parameters per center for the T2w images.
Diagnostics 12 01601 g002

3.3. Data Processing

Rigid registration between the T2w pre- and post-therapeutic images was overall successful by visual inspection. However, we noticed a mismatch of the DW to the T2w images despite them being acquired in the same session. Upon closer examination, we found that, in addition to motion artifacts, misaligned slice positioning and angulation between the T2w and DW images played a major role in this. Attempts to correct this misalignment by rigid registration failed, presumably due to low SNR and distortions afflicting all DW images probably caused by susceptibilities affecting the EPI acquisition.

3.4. Model Performance

The AUC for pCR prediction averaged over the five models obtained via cross validation using only pre- and post-therapeutic T2w images as input was 0.68 (95% confidence interval [CI]: 0.59, 0.77). Combining pre- and post-therapeutic T2w, DW images, and ADC maps as input yielded an AUC of 0.43 (95% CI: 0.35, 0.52). Figure 3a and c depict the receiver operator characteristic (ROC) curves for each of the five models and the average over all folds.
Testing our implementation of the network of Jin et al. [32] on the external dataset (Center 4b) returned an AUC for pCR prediction of 0.54 (95% CI: 0.41, 0.65), when using only pre- and post-therapeutic T2w images as input. Using the combination of pre- and post-therapeutic T2w, DW images, and ADC maps as input, the AUC for pCR prediction was 0.60 (95% CI: 0.48, 0.71). Figure 3b and d depict the ROC curves for each of the five models and the average over all folds.
For the different ROC curves depicted in Figure 3, the AUC was calculated and plotted in Figure 4. In the training cohort, classification using only the T2w images performed significantly better than using both T2w and DW images (p < 0.01 using the DeLong method). In the external dataset, using both T2w and DW images yielded the best results with an AUC of 0.68, but was not significantly better than using only T2w images (p = 0.3 using the DeLong method).
To review model performance, we overlaid the different input images with the segmentation mask originally annotated on the high-resolution T2w images (Figure 5). As described above, the morphological images and the DW images were at times misaligned, so that the transformed segmentations incompletely or incorrectly labeled the tumor region on the DW images. In these cases of misalignment, the performance of the algorithm dropped considerably as illustrated by Figure 5.

4. Discussion

Reliably predicting treatment response/pCR preoperatively after completion of nCRT using MRI holds the potential to offer LARC patients a watch-and-wait strategy instead of surgical resection and to better assess overall survival. DL provides the ability to train a neural network to identify patients with pCR using morphological and functional imaging data as input [6]. While initial results from single-center studies are promising [32,43], further studies using multicenter real-world data are necessary before such networks can be widely applied clinically. In this study, we retrained a promising, state-of-the-art multitask DL model [32] using a heterogenous dataset consisting of T2w, DW images and ADC maps acquired before and after nCRT and assessed its performance in a multicenter trial.
Our results show that the performance of the proposed DL model decreases significantly using a multicenter dataset. Three key takeaways can be concluded from our study:
  • Data quality and uniformity are pivotal features to be addressed in clinical trials involving machine learning, requiring the development of a dedicated pre-processing pipeline.
  • If no homogeneous data are available, the sample size for training the DL approach needs to be drastically increased to mitigate artifacts related to image inhomogeneity.
  • Translating DL models into potentially useful clinical tools requires cross-center involvement of a multidisciplinary team.
We aimed to answer the question to what extent translation of promising DL models into the clinic is possible. Thus, to ensure clinical feasibility, we decided to use only pre- and post-therapeutic T2w and DW images as inputs for the proposed DL model and to refrain from collecting additional parameters such as blood values. Despite these minimal requirements, 78 of 162 patients of the CAO-ARO-AIO-12 study dataset already had to be excluded due to missing or incomplete image datasets that had been provided to us.
To train the network, we used patients enrolled under study conditions from the CAO-ARO-AIO-12 study [33], which has certain standards for image acquisition. Yet, significant differences in the composition of the patient population as well as in the distribution of the acquisition parameters were evident in these data. These differences were even more pronounced when comparing the dataset acquired under study conditions with the external, clinical dataset that we used for testing. Differences in image acquisition parameters result in varying image quality. The problem of data heterogeneity is compounded by the fact that the image datasets in this study were acquired on a total of 15 different scanner types of two vendors with varying field strengths. Often an individual patient was even examined on different scanners before and after nCRT. Particularly, the DW images often exhibited poor and heterogeneous quality, and moreover, were often acquired in a dissimilar orientation relative to the T2w images. These factors, including a low signal-to-noise ratio, various artifacts, and low resolution, led to the fact that we failed to register the DW images sufficiently with the high-resolution T2w images [44]. This proved especially challenging as tumor segmentation was performed on the high-resolution T2w images. Notably, for those patients misclassified by the network, the correspondence of the segmentation to the tumor region in the DW images was poor. It should also be pointed out that, especially in the internal multicenter training dataset, the performance of the network decreased from an AUC of 0.68 to 0.43 when the DWI data were added as an input to the pre- and post-therapeutic T2w images. Interestingly, this was not the case in the single-center test dataset.
These observations are consistent with the findings of Schurink et al. [45], who found significant intercenter variation in multicenter rectal MRI data, with greater variation in DW images and ADC maps compared to T2w images, mainly related to hardware and image acquisition protocols. They further state that such variations will likely negatively influence subsequent analysis if not corrected for [45]. Even though imaging processing and data harmonization were applied in this study prior to network training and testing, our results indicate that further development of a dedicated data pre-processing pipeline is necessary that specifically addresses the inhomogeneities of a multicenter dataset [46,47]. As an analysis by Glocker et al. suggests, current approaches to harmonizing data are not able to correct for scanner-specific biases and harmonization of data remains an open challenge [48].
Furthermore, to mitigate artifacts related to image inhomogeneity, much larger datasets are required for network training [49,50,51]. Yet, the availability of image data seems to be one of the biggest hurdles for the implementation of artificial intelligence in the clinical setting [52], as supported by our study. In addition, data preparation and annotation are very costly and time consuming [52]. Small datasets from small geographic areas are prone to selection bias that can lead to algorithms with limited utility and poor generalizability [52]. While new approaches such as federated learning might be able to address the problem of data availability, data curation and annotation accounted for a significant portion of our work, which pose barriers to a successful implementation of such algorithms in clinical practice [52]. In this regard, we consider it very important to emphasize that, beyond the arguments already mentioned, the basis or prerequisite for a successful translation of AI algorithms into the clinic is a cross-center, interdisciplinary team. This begins with careful planning by a multidisciplinary team and the specification of acquisition parameters that are implemented by radiographers, continues with standardized reporting by radiologists, cross-center image processing and preparation by data scientists, and finally ends with interdisciplinary interpretation of the results. Only through a joint effort will a sustainable implementation of such promising DL models in the clinic succeed.
Our study has certain limitations. Image registration of the image data turned out to be challenging due to distortions and slice misalignment within a patient rather than between time points. So far, we only used a rigid registration to align pre- and post-therapeutic images similar to Jin et al. [32]. In their work, it seems that the issue of distortions and misalignment between T2w and DWI was either not dealt with or minor. Due to the small size of our dataset and its heterogeneity, an elastic image registration would probably enhance our classification results which will be future research.
In comparison to the study by Jin et al. [32], we reduced the number of network inputs, partly because the CAO-ARO-AIO-12 study dataset, for instance, did not include pre- and post-contrast T1w images for all patients, and partly to ensure clinical feasibility, as already written above. Furthermore, we were limited in our ability to test the generalizability of the DL model [32] because the weights of the pre-trained network were not shared with us. This is an obstacle in the applicability of many published networks.
Lastly, we did not optimize the hyperparameters of the network, which might boost the performance since we aimed at translating the model “as is” to our local settings in the best possible way.

5. Conclusions

In this study, we re-trained a promising state-of-the-art multitask DL model using a heterogeneous multicenter dataset and assessed its performance in predicting pCR of LARC to nCRT using an external, real-world clinical dataset. Our results demonstrate that despite dedicated data pre-processing, the performance of the DL network drops significantly in the multicenter setting compared to the original single-center study with external validation. This finding highlights the importance of data quality and harmonization in clinical trials using machine learning. Only in a joint, cross-center effort, involving a multidisciplinary team can we generate large enough curated and annotated datasets and develop the necessary pre-processing pipelines for data harmonization to successfully apply DL models clinically.

Author Contributions

Conceptualization, U.I.A., F.G.Z. and J.H.; methodology, B.D.W., S.A. and W.Z.; validation, B.D.W. and S.A.; formal analysis, B.D.W., S.A. and W.Z.; resources, A.M., C.R. and R.-D.H.; data curation, B.D.W., A.M. and S.A.; writing—original draft preparation, B.D.W., S.A. and F.G.Z.; writing—review and editing, B.D.W., S.A., W.Z., A.M., C.R., R.-D.H., J.H., F.G.Z. and U.I.A.; supervision, U.I.A., F.G.Z. and J.H.; project administration, U.I.A., F.G.Z. and J.H.; funding acquisition, U.I.A., F.G.Z. and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

The authors gratefully acknowledge the data storage service SDS@hd supported by the Ministry of Science, Research and the Arts Baden-Württemberg (MWK) and the German Research Foundation (DFG) through grant INST 35/1314-1 FUGG and INST 35/1503-1 FUGG and support from the DFG through grant 428149221.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board II of Heidelberg University (protocol code 2017-571N-MA, 4th November 2019) as well as the Institutional Review Board of the University of Bonn (protocol code 110/21, 10 March 2021).

Informed Consent Statement

Written patient consent was waived in this retrospective study.

Data Availability Statement

Restrictions apply to the availability of these data. Data was obtained from German Rectal Cancer Study Group (EudraCT No.: 2011-006310-13). The data are not publicly available due to patient privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. Robert Koch-Institut und die Gesellschaft der epidemiologischen Krebsregister in Deutschland e.V. (Ed.) Krebs in Deutschland für 2017/2018, 13rd ed.; Robert Koch-Institute: Berlin, Germany, 2021. [Google Scholar]
  3. Siegel, R.L.; Miller, K.D.; Goding Sauer, A.; Fedewa, S.A.; Butterly, L.F.; Anderson, J.C.; Cercek, A.; Smith, R.A.; Jemal, A. Colorectal cancer statistics. CA Cancer J. Clin. 2020, 70, 145–164. [Google Scholar] [CrossRef] [Green Version]
  4. Benson, A.B.; Venook, A.P.; Al-Hawary, M.M.; Cederquist, L.; Chen, Y.J.; Ciombor, K.K.; Cohen, S.; Cooper, H.S.; Deming, D.; Engstrom, P.F.; et al. Rectal Cancer, Version 2.2018, NCCN Clinical Practice Guidelines in Oncology. J. Natl. Compr. Cancer Netw. 2018, 16, 874–901. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Fokas, E.; Appelt, A.; Glynne-Jones, R.; Beets, G.; Perez, R.; Garcia-Aguilar, J.; Rullier, E.; Joshua Smith, J.; Marijnen, C.; Peters, F.P.; et al. International consensus recommendations on key outcome measures for organ preservation after (chemo)radiotherapy in patients with rectal cancer. Nat. Rev. Clin. Oncol. 2021, 18, 805–816. [Google Scholar] [CrossRef] [PubMed]
  6. Koh, D.-M. Using Deep Learning for MRI to Identify Responders to Chemoradiotherapy in Rectal Cancer. Radiology 2020, 296, 65–66. [Google Scholar] [CrossRef]
  7. Bahadoer, R.R.; Dijkstra, E.A.; van Etten, B.; Marijnen, C.A.M.; Putter, H.; Kranenbarg, E.M.; Roodvoets, A.G.H.; Nagtegaal, I.D.; Beets-Tan, R.G.H.; Blomqvist, L.K.; et al. Short-course radiotherapy followed by chemotherapy before total mesorectal excision (TME) versus preoperative chemoradiotherapy, TME, and optional adjuvant chemotherapy in locally advanced rectal cancer (RAPIDO): A randomised, open-label, phase 3 trial. Lancet Oncol. 2021, 22, 29–42. [Google Scholar] [CrossRef]
  8. Loos, M.; Quentmeier, P.; Schuster, T.; Nitsche, U.; Gertler, R.; Keerl, A.; Kocher, T.; Friess, H.; Rosenberg, R. Effect of preoperative radio(chemo)therapy on long-term functional outcome in rectal cancer patients: A systematic review and meta-analysis. Ann. Surg. Oncol. 2013, 20, 1816–1828. [Google Scholar] [CrossRef]
  9. Rouanet, P. Tailored treatment strategy for locally advanced rectal carcinoma: Five-year results of the French phase II, randomized, multicenter GRECCAR4 trial. J. Clin. Oncol. 2020, 38, 4014. [Google Scholar] [CrossRef]
  10. van der Sande, M.E.; Beets, G.L.; Hupkens, B.J.; Breukink, S.O.; Melenhorst, J.; Bakers, F.C.; Lambregts, D.M.; Grabsch, H.I.; Beets-Tan, R.G.; Maas, M. Response assessment after (chemo)radiotherapy for rectal cancer: Why are we missing complete responses with MRI and endoscopy? Eur. J. Surg. Oncol. 2019, 45, 1011–1017. [Google Scholar] [CrossRef]
  11. Jang, J.K.; Choi, S.H.; Park, S.H.; Kim, K.W.; Kim, H.J.; Lee, J.S.; Kim, A.Y. MR tumor regression grade for pathological complete response in rectal cancer post neoadjuvant chemoradiotherapy: A systematic review and meta-analysis for accuracy. Eur. Radiol. 2020, 30, 2312–2323. [Google Scholar] [CrossRef]
  12. Gaa, T.; Neumann, W.; Sudarski, S.; Attenberger, U.I.; Schonberg, S.O.; Schad, L.R.; Zollner, F.G. Comparison of perfusion models for quantitative T1 weighted DCE-MRI of rectal cancer. Sci. Rep. 2017, 7, 12036. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Sudarski, S.; Henzler, T.; Floss, T.; Gaa, T.; Meyer, M.; Haubenreisser, H.; Schoenberg, S.O.; Attenberger, U.I. Variability and Reproducibility of 3(rd)-generation dual-source dynamic volume perfusion CT Parameters in Comparison to MR-perfusion Parameters in Rectal Cancer. Sci. Rep. 2018, 8, 6868. [Google Scholar] [CrossRef] [PubMed]
  14. Attenberger, U.I.; Ong, M.M.; Rathmann, N.; Doyon, F.; Kienle, P.; Hofheinz, R.D.; Pilz, L.R.; Dinter, D.J.; Schoenberg, S.O.; Hausmann, D. mMRI at 3.0 T as an Evaluation Tool of Therapeutic Response to Neoadjuvant CRT in Patients with Advanced-stage Rectal Cancer. Anticancer Res. 2017, 37, 215–222. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Attenberger, U.I.; Pilz, L.R.; Morelli, J.N.; Hausmann, D.; Doyon, F.; Hofheinz, R.; Kienle, P.; Post, S.; Michaely, H.J.; Schoenberg, S.O.; et al. Multi-parametric MRI of rectal cancer-do quantitative functional MR measurements correlate with radiologic and pathologic tumor stages? Eur. J. Radiol. 2014, 83, 1036–1043. [Google Scholar] [CrossRef]
  16. Park, M.J.; Kim, S.H.; Lee, S.J.; Jang, K.M.; Rhim, H. Locally advanced rectal cancer: Added value of diffusion-weighted MR imaging for predicting tumor clearance of the mesorectal fascia after neoadjuvant chemotherapy and radiation therapy. Radiology 2011, 260, 771–780. [Google Scholar] [CrossRef]
  17. Hötker, A.M.; Garcia-Aguilar, J.; Gollub, M.J. Multiparametric MRI of rectal cancer in the assessment of response to therapy: A systematic review. Dis. Colon Rectum 2014, 57, 790–799. [Google Scholar] [CrossRef]
  18. Mizukami, Y.; Ueda, S.; Mizumoto, A.; Sasada, T.; Okumura, R.; Kohno, S.; Takabayashi, A. Diffusion-weighted magnetic resonance imaging for detecting lymph node metastasis of rectal cancer. World J. Surg. 2011, 35, 895–899. [Google Scholar] [CrossRef]
  19. Le Bihan, D. Apparent diffusion coefficient and beyond: What diffusion MR imaging can tell us about tissue structure. Radiology 2013, 268, 318–322. [Google Scholar] [CrossRef]
  20. Intven, M.; Reerink, O.; Philippens, M.E. Diffusion-weighted MRI in locally advanced rectal cancer: Pathological response prediction after neo-adjuvant radiochemotherapy. Strahlenther. Onkol. 2013, 189, 117–122. [Google Scholar] [CrossRef]
  21. Jung, S.H.; Heo, S.H.; Kim, J.W.; Jeong, Y.Y.; Shin, S.S.; Soung, M.G.; Kim, H.R.; Kang, H.K. Predicting response to neoadjuvant chemoradiation therapy in locally advanced rectal cancer: Diffusion-weighted 3 Tesla MR imaging. J. Magn. Reason. Imaging 2012, 35, 110–116. [Google Scholar] [CrossRef]
  22. Dzik-Jurasz, A.; Domenig, C.; George, M.; Wolber, J.; Padhani, A.; Brown, G.; Doran, S. Diffusion MRI for prediction of response of rectal cancer to chemoradiation. Lancet 2002, 360, 307–308. [Google Scholar] [CrossRef]
  23. Cai, G.; Xu, Y.; Zhu, J.; Gu, W.L.; Zhang, S.; Ma, X.J.; Cai, S.J.; Zhang, Z. Diffusion-weighted magnetic resonance imaging for predicting the response of rectal cancer to neoadjuvant concurrent chemoradiation. World J. Gastroenterol. 2013, 19, 5520–5527. [Google Scholar] [CrossRef] [PubMed]
  24. Ippolito, D.; Monguzzi, L.; Guerra, L.; Deponti, E.; Gardani, G.; Messa, C.; Sironi, S. Response to neoadjuvant therapy in locally advanced rectal cancer: Assessment with diffusion-weighted MR imaging and 18FDG PET/CT. Abdom. Imaging 2012, 37, 1032–1040. [Google Scholar] [CrossRef]
  25. DeVries, A.F.; Kremser, C.; Hein, P.A.; Griebel, J.; Krezcy, A.; Ofner, D.; Pfeiffer, K.P.; Lukas, P.; Judmaier, W. Tumor microcirculation and diffusion predict therapy outcome for primary rectal carcinoma. Int. J. Radiat. Oncol. Biol. Phys. 2003, 56, 958–965. [Google Scholar] [CrossRef]
  26. Musio, D.; De Felice, F.; Magnante, A.L.; Ciolina, M.; De Cecco, C.N.; Rengo, M.; Redler, A.; Laghi, A.; Raffetto, N.; Tombolini, V. Diffusion-weighted magnetic resonance application in response prediction before, during, and after neoadjuvant radiochemotherapy in primary rectal cancer carcinoma. Biomed. Res. Int. 2013, 2013, 740195. [Google Scholar] [CrossRef] [PubMed]
  27. Monguzzi, L.; Ippolito, D.; Bernasconi, D.P.; Trattenero, C.; Galimberti, S.; Sironi, S. Locally advanced rectal cancer: Value of ADC mapping in prediction of tumor response to radiochemotherapy. Eur. J. Radiol. 2013, 82, 234–240. [Google Scholar] [CrossRef] [PubMed]
  28. Curvo-Semedo, L.; Lambregts, D.M.; Maas, M.; Beets, G.L.; Caseiro-Alves, F.; Beets-Tan, R.G. Diffusion-weighted MRI in rectal cancer: Apparent diffusion coefficient as a potential noninvasive marker of tumor aggressiveness. J. Magn. Reason. Imaging 2012, 35, 1365–1371. [Google Scholar] [CrossRef]
  29. Barbaro, B.; Vitale, R.; Valentini, V.; Illuminati, S.; Vecchio, F.M.; Rizzo, G.; Gambacorta, M.A.; Coco, C.; Crucitti, A.; Persiani, R.; et al. Diffusion-weighted magnetic resonance imaging in monitoring rectal cancer response to neoadjuvant chemoradiotherapy. Int. J. Radiat. Oncol. Biol. Phys. 2012, 83, 594–599. [Google Scholar] [CrossRef]
  30. Elmi, A.; Hedgire, S.S.; Covarrubias, D.; Abtahi, S.M.; Hahn, P.F.; Harisinghani, M. Apparent diffusion coefficient as a non-invasive predictor of treatment response and recurrence in locally advanced rectal cancer. Clin. Radiol. 2013, 68, e524–e531. [Google Scholar] [CrossRef]
  31. Joye, I.; Deroose, C.M.; Vandecaveye, V.; Haustermans, K. The role of diffusion-weighted MRI and (18)F-FDG PET/CT in the prediction of pathologic complete response after radiochemotherapy for rectal cancer: A systematic review. Radiother. Oncol. 2014, 113, 158–165. [Google Scholar] [CrossRef] [Green Version]
  32. Jin, C.; Yu, H.; Ke, J.; Ding, P.; Yi, Y.; Jiang, X.; Duan, X.; Tang, J.; Chang, D.T.; Wu, X.; et al. Predicting treatment response from longitudinal images using multi-task deep learning. Nat. Commun. 2021, 12, 1851. [Google Scholar] [CrossRef] [PubMed]
  33. Fokas, E.; Allgauer, M.; Polat, B.; Klautke, G.; Grabenbauer, G.G.; Fietkau, R.; Kuhnt, T.; Staib, L.; Brunner, T.; Grosu, A.L.; et al. Randomized Phase II Trial of Chemoradiotherapy Plus Induction or Consolidation Chemotherapy as Total Neoadjuvant Therapy for Locally Advanced Rectal Cancer: CAO/ARO/AIO-12. J. Clin. Oncol. 2019, 37, 3212–3222. [Google Scholar] [CrossRef] [PubMed]
  34. ClinicalTrials.gov. Available online: https://clinicaltrials.gov/ct2/show/NCT02363374 (accessed on 23 June 2022).
  35. Yushkevich, P.A.; Piven, J.; Hazlett, H.C.; Smith, R.G.; Ho, S.; Gee, J.C.; Gerig, G. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. Neuroimage 2006, 31, 1116–1128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. ITK-SNAP. Available online: www.itksnap.org (accessed on 23 June 2022).
  37. Dworak, O.; Keilholz, L.; Hoffmann, A. Pathological features of rectal cancer after preoperative radiochemotherapy. Int. J. Colorectal Dis. 1997, 12, 19–23. [Google Scholar] [CrossRef] [PubMed]
  38. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [Green Version]
  39. Lowekamp, B.C.; Chen, D.T.; Ibanez, L.; Blezek, D. The Design of SimpleITK. Front. Neuroinform. 2013, 7, 45. [Google Scholar] [CrossRef] [Green Version]
  40. Yaniv, Z.; Lowekamp, B.C.; Johnson, H.J.; Beare, R. SimpleITK Image-Analysis Notebooks: A Collaborative Environment for Education and Reproducible Research. J. Digit. Imaging 2018, 31, 290–303. [Google Scholar] [CrossRef] [Green Version]
  41. Avants, B.B.; Tustison, N.J.; Song, G.; Cook, P.A.; Klein, A.; Gee, J.C. A reproducible evaluation of ANTs similarity metric performance in brain image registration. Neuroimage 2011, 54, 2033–2044. [Google Scholar] [CrossRef] [Green Version]
  42. DeLong, E.R.; DeLong, D.M.; Clarke-Pearson, D.L. Comparing the areas under two or more correlated receiver operating characteristic curves: A nonparametric approach. Biometrics 1988, 44, 837–845. [Google Scholar] [CrossRef]
  43. Zhang, X.-Y.; Wang, L.; Zhu, H.-T.; Li, Z.-W.; Ye, M.; Li, X.-T.; Shi, Y.-J.; Zhu, H.-C.; Sun, Y.-S. Predicting Rectal Cancer Response to Neoadjuvant Chemoradiotherapy Using Deep Learning of Diffusion Kurtosis MRI. Radiology 2020, 296, 56–64. [Google Scholar] [CrossRef]
  44. Huizinga, W.; Poot, D.H.; Guyader, J.M.; Klaassen, R.; Coolen, B.F.; van Kranenburg, M.; van Geuns, R.J.; Uitterdijk, A.; Polfliet, M.; Vandemeulebroucke, J.; et al. PCA-based groupwise image registration for quantitative MRI. Med. Image Anal. 2016, 29, 65–78. [Google Scholar] [CrossRef] [PubMed]
  45. Schurink, N.W.; van Kranen, S.R.; Roberti, S.; van Griethuysen, J.J.M.; Bogveradze, N.; Castagnoli, F.; El Khababi, N.; Bakers, F.C.H.; de Bie, S.H.; Bosma, G.P.T.; et al. Sources of variation in multicenter rectal MRI data and their effect on radiomics feature reproducibility. Eur. Radiol. 2022, 32, 1506–1516. [Google Scholar] [CrossRef] [PubMed]
  46. Blumberg, S.B.; Palombo, M.; Khoo, C.S.; Tax, C.M.W.; Tanno, R.; Alexander, D.C. Multi-stage Prediction Networks for Data Harmonization. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019; Springer: Cham, Switzerland, 2019; pp. 411–419. [Google Scholar]
  47. Cackowski, S.; Barbier, E.L.; Dojat, M.; Christen, T. ImUnity: A generalizable VAE-GAN solution for multicenter MR image harmonization. arXiv 2021, arXiv:2109.06756. [Google Scholar]
  48. Glocker, B.; Robinson, R.; Castro, D.C.; Dou, Q.; Konukoglu, E. Machine learning with multi-site imaging data: An empirical study on the impact of scanner effects. arXiv 2019, arXiv:1910.04597. [Google Scholar]
  49. Dutta, S.; Gros, E. Evaluation of the impact of deep learning architectural components selection and dataset size on a medical imaging task. In Medical Imaging 2018: Imaging Informatics for Healthcare, Research, and Applications; SPIE: Bellingham, WA, USA, 2018; Volume 10579. [Google Scholar]
  50. Raj, A.; Tollens, F.; Hansen, L.; Golla, A.-K.; Schad, L.R.; Nörenberg, D.; Zöllner, F.G. Deep Learning-Based Total Kidney Volume Segmentation in Autosomal Dominant Polycystic Kidney Disease Using Attention, Cosine Loss, and Sharpness Aware Minimization. Diagnostics 2022, 12, 1159. [Google Scholar] [CrossRef] [PubMed]
  51. Kohli, M.D.; Summers, R.M.; Geis, J.R. Medical Image Data and Datasets in the Era of Machine Learning—Whitepaper from the 2016 C-MIMI Meeting Dataset Session. J. Digit. Imaging 2017, 30, 392–399. [Google Scholar] [CrossRef] [Green Version]
  52. Willemink, M.J.; Koszek, W.A.; Hardell, C.; Wu, J.; Fleischmann, D.; Harvey, H.; Folio, L.R.; Summers, R.M.; Rubin, D.L.; Lungren, M.P. Preparing Medical Imaging Data for Machine Learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the inclusions and exclusions of LARC patients of the CAO-ARO-AIO-12 study.
Figure 1. Flowchart of the inclusions and exclusions of LARC patients of the CAO-ARO-AIO-12 study.
Diagnostics 12 01601 g001
Figure 3. Receiver operator characteristic (ROC) of all obtained models (i.e., folds) and combination of input data for the training (a,b) and external validation dataset (c,d). (a,b) depict ROC curves for the training cohort: (a) using only T2w images as input vs. (b) using T2w and DW images as input. (c,d) depict the corresponding results for the external validation dataset: (c) using only T2w images as input vs. (d) using T2w and DW images as input. In each panel, the bold brown-colored curve depicts the average over the five folds.
Figure 3. Receiver operator characteristic (ROC) of all obtained models (i.e., folds) and combination of input data for the training (a,b) and external validation dataset (c,d). (a,b) depict ROC curves for the training cohort: (a) using only T2w images as input vs. (b) using T2w and DW images as input. (c,d) depict the corresponding results for the external validation dataset: (c) using only T2w images as input vs. (d) using T2w and DW images as input. In each panel, the bold brown-colored curve depicts the average over the five folds.
Diagnostics 12 01601 g003
Figure 4. The area under the receiver operating characteristic curve (AUC) was calculated for the different receiver operating characteristic curves depicted in Figure 3. The AUCs are shown for the training (colored red) and test cohorts (colored blue). For the training cohort, classification with only T2w images as inputs performed significantly better than with T2w and DW images.
Figure 4. The area under the receiver operating characteristic curve (AUC) was calculated for the different receiver operating characteristic curves depicted in Figure 3. The AUCs are shown for the training (colored red) and test cohorts (colored blue). For the training cohort, classification with only T2w images as inputs performed significantly better than with T2w and DW images.
Diagnostics 12 01601 g004
Figure 5. Examples of correctly and incorrectly classified patients. Each row shows input images of one patient before and after nCRT overlayed with the segmentation mask. Note the slight misalignment of the segmentation mask with the tumor region on the DW images, which might contribute to misclassification.
Figure 5. Examples of correctly and incorrectly classified patients. Each row shows input images of one patient before and after nCRT overlayed with the segmentation mask. Note the slight misalignment of the segmentation mask with the tumor region on the DW images, which might contribute to misclassification.
Diagnostics 12 01601 g005
Table 1. Characteristics of patients in the training, validation and test cohort disaggregated by centers.
Table 1. Characteristics of patients in the training, validation and test cohort disaggregated by centers.
Training CohortValidation CohortTest
Cohort
Significant Differences
CharacteristicCenter 1Center 2Center 3Center 4aCenter 4b
Acquisition (years)2015–20182016–20172015–20172015–20172009–20130.00
Age (mean ± std)60 ± 1061 ± 1160 ± 666 ± 664 ± 110.07
Sex 0.15
Male26148937
Female116729
Pre-nCRT T-stage (MRI) 0.02
T000000
T100000
T230208
T3 28111538
T440100
Not specified219160
Pre-nCRT N-stage (MRI) 0.00
N-220125
N+3518151021
CRM (initial staging) 0.06
Minimal distance to mesorectal fascia (MRF) in mm12 ± 9 2 ± 5 3 ± 3
MRF involvement20027
Not specified1320090
Tumor location 0.0
lower third1855410
middle third13910525
upper third000011
location not specified66020
Post-nCRT T-stage (MRI) 0.03
T010300
T120102
T21301027
T3 17010017
T440010
Not specified0200100
Post-nCRT N-stage (MRI) 0.68
N−1702039
N+2001325
Not specified020092
pCR 0.15
pCR79125
Non-pCR301114941
Time in days (mean ± std)
Initial Staging to OP146 ± 12146 ± 11142 ± 8177 ± 35123 ± 200.0
Post-nCRT MRI to OP13 ± 107 ± 38 ± 332 ± 1529 ± 140.0
Unless otherwise indicated, data are number of patients. Significance of differences between the training/validation and test cohort were calculated using the Mann–Whitney U-test; a p-value below 0.05 was considered significant.
Table 2. Overview of the different scanners used for imaging the patients at each center.
Table 2. Overview of the different scanners used for imaging the patients at each center.
VendorModel NameTeslaNumber of Patients before nCRTNumber of Patients after nCRT
Center 1SiemensPrisma_fit3.0618
SiemensAvanto1.5140
SiemensAvanto_fit1.5214
SiemensSymphonyTim1.5121
SiemensAera1.513
SiemensEspree1.501
PhilipsIngenia1.510
SiemensSpectra3.010
Center 2SiemensSkyra3.01710
SiemensAvanto1.5310
Center 3SiemensPrisma_fit3.0119
SiemensSkyra3.046
Center 4aSiemensSkyra3.048
SiemensTrioTim3.063
SiemensAvanto1.510
Center 4bSiemensTrioTim1.54646
Table 3. Differences between the training and test dataset.
Table 3. Differences between the training and test dataset.
Training T2wTest T2wTraining DWITest DWI
Slice thickness (mm)3.3 ± 0.53.1 ± 0.25.1 ± 1.15.0 ± 0.0
Repetition time (ms)4994.9 ± 1913.53971.5 ± 708.16463.1 ± 1539.14121.9 ± 562.7
Pixel bandwidth (Hz)215.9 ± 48.9201.6 ± 8.81792.5 ± 351.61735.0 ± 8.3
Flip angle (°)137.0 ± 16.5148.7 ± 5.690.0 ± 0.090.0 ± 0.0
Echo time (ms)97.8 ± 14.5101.9 ± 4.767.2 ± 13.473.2 ± 2.2
Field strength (T)2.5 ± 0.73.0 ± 0.02.5 ± 0.73.0 ± 0.0
In-plane resolution (mm)0.6 ± 0.20.6 ± 0.01.6 ± 0.42.0 ± 0.1
Each column shows the mean and standard deviation of the different acquisition parameters of the T2w and DW sequences for the training and test cohort, respectively. Only the flip angle (90°) and pixel bandwidth of DWI were not significantly different between cohorts.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wichtmann, B.D.; Albert, S.; Zhao, W.; Maurer, A.; Rödel, C.; Hofheinz, R.-D.; Hesser, J.; Zöllner, F.G.; Attenberger, U.I. Are We There Yet? The Value of Deep Learning in a Multicenter Setting for Response Prediction of Locally Advanced Rectal Cancer to Neoadjuvant Chemoradiotherapy. Diagnostics 2022, 12, 1601. https://doi.org/10.3390/diagnostics12071601

AMA Style

Wichtmann BD, Albert S, Zhao W, Maurer A, Rödel C, Hofheinz R-D, Hesser J, Zöllner FG, Attenberger UI. Are We There Yet? The Value of Deep Learning in a Multicenter Setting for Response Prediction of Locally Advanced Rectal Cancer to Neoadjuvant Chemoradiotherapy. Diagnostics. 2022; 12(7):1601. https://doi.org/10.3390/diagnostics12071601

Chicago/Turabian Style

Wichtmann, Barbara D., Steffen Albert, Wenzhao Zhao, Angelika Maurer, Claus Rödel, Ralf-Dieter Hofheinz, Jürgen Hesser, Frank G. Zöllner, and Ulrike I. Attenberger. 2022. "Are We There Yet? The Value of Deep Learning in a Multicenter Setting for Response Prediction of Locally Advanced Rectal Cancer to Neoadjuvant Chemoradiotherapy" Diagnostics 12, no. 7: 1601. https://doi.org/10.3390/diagnostics12071601

APA Style

Wichtmann, B. D., Albert, S., Zhao, W., Maurer, A., Rödel, C., Hofheinz, R. -D., Hesser, J., Zöllner, F. G., & Attenberger, U. I. (2022). Are We There Yet? The Value of Deep Learning in a Multicenter Setting for Response Prediction of Locally Advanced Rectal Cancer to Neoadjuvant Chemoradiotherapy. Diagnostics, 12(7), 1601. https://doi.org/10.3390/diagnostics12071601

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop