Next Article in Journal
[6]-Gingerol-Derived Semi-Synthetic Compound SSi6 Inhibits Tumor Growth and Metastatic Dissemination in Triple-Negative Breast Cancer Xenograft Models
Next Article in Special Issue
Feasibility, Safety and Impact on Overall Survival of Awake Resection for Newly Diagnosed Supratentorial IDH-Wildtype Glioblastomas in Adults
Previous Article in Journal
Bispecific T Cell Engagers for the Treatment of Multiple Myeloma: Achievements and Challenges
Previous Article in Special Issue
Motor Evoked Potential Warning Criteria in Supratentorial Surgery: A Scoping Review
 
 
cancers-logo
Article Menu
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Glioblastoma Surgery Imaging—Reporting and Data System: Standardized Reporting of Tumor Volume, Location, and Resectability Based on Automated Segmentations

by
Ivar Kommers
1,2,†,
David Bouget
3,†,
André Pedersen
3,
Roelant S. Eijgelaar
1,2,
Hilko Ardon
4,
Frederik Barkhof
5,6,
Lorenzo Bello
7,
Mitchel S. Berger
8,
Marco Conti Nibali
7,
Julia Furtner
9,
Even H. Fyllingen
10,11,
Shawn Hervey-Jumper
8,
Albert J. S. Idema
12,
Barbara Kiesel
13,
Alfred Kloet
14,
Emmanuel Mandonnet
15,
Domenique M. J. Müller
1,2,
Pierre A. Robe
16,
Marco Rossi
7,
Lisa M. Sagberg
17,
Tommaso Sciortino
7,
Wimar A. van den Brink
18,
Michiel Wagemakers
19,
Georg Widhalm
13,
Marnix G. Witte
20,
Aeilko H. Zwinderman
21,
Ingerid Reinertsen
3,10,
Ole Solheim
17,22,* and
Philip C. De Witt Hamer
1,2,*
add Show full author list remove Hide full author list
1
Department of Neurosurgery, Amsterdam University Medical Centers, Vrije Universiteit, 1081 HV Amsterdam, The Netherlands
2
Cancer Center Amsterdam, Brain Tumor Center, Amsterdam University Medical Centers, 1081 HV Amsterdam, The Netherlands
3
Department of Health Research, SINTEF Digital, NO-7465 Trondheim, Norway
4
Department of Neurosurgery, Twee Steden Hospital, 5042 AD Tilburg, The Netherlands
5
Department of Radiology and Nuclear Medicine, Amsterdam University Medical Centers, Vrije Universiteit, 1081 HV Amsterdam, The Netherlands
6
Institutes of Neurology and Healthcare Engineering, University College London, London WC1E 6BT, UK
7
Neurosurgical Oncology Unit, Department of Oncology and Hemato-Oncology, Humanitas Research Hospital, Università Degli Studi di Milano, 20122 Milano, Italy
8
Department of Neurological Surgery, University of California San Francisco, San Francisco, CA 94143, USA
9
Department of Biomedical Imaging and Image-Guided Therapy, Medical University Vienna, 1090 Wien, Austria
10
Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, NO-7491 Trondheim, Norway
11
Department of Radiology and Nuclear Medicine, St. Olav’s Hospital, Trondheim University Hospital, NO-7030 Trondheim, Norway
12
Department of Neurosurgery, Northwest Clinics, 1815 JD Alkmaar, The Netherlands
13
Department of Neurosurgery, Medical University Vienna, 1090 Wien, Austria
14
Department of Neurosurgery, Haaglanden Medical Center, 2512 VA The Hague, The Netherlands
15
Department of Neurological Surgery, Hôpital Lariboisière, 75010 Paris, France
16
Department of Neurology and Neurosurgery, University Medical Center Utrecht, 3584 CX Utrecht, The Netherlands
17
Department of Neurosurgery, St. Olav’s Hospital, Trondheim University Hospital, NO-7030 Trondheim, Norway
18
Department of Neurosurgery, Isala, 8025 AB Zwolle, The Netherlands
19
Department of Neurosurgery, University Medical Center Groningen, University of Groningen, 9713 GZ Groningen, The Netherlands
20
Department of Radiation Oncology, The Netherlands Cancer Institute, 1066 CX Amsterdam, The Netherlands
21
Department of Clinical Epidemiology and Biostatistics, Amsterdam University Medical Centers, University of Amsterdam, 1105 AZ Amsterdam, The Netherlands
22
Department of Neuromedicine and Movement Science, Norwegian University of Science and Technology, NO-7491 Trondheim, Norway
*
Authors to whom correspondence should be addressed.
These authors contributed equally.
Cancers 2021, 13(12), 2854; https://doi.org/10.3390/cancers13122854
Submission received: 30 April 2021 / Revised: 28 May 2021 / Accepted: 2 June 2021 / Published: 8 June 2021
(This article belongs to the Special Issue Perioperative Imaging and Mapping Methods in Glioma Patients)

Abstract

:

Simple Summary

Neurosurgical decisions for patients with glioblastoma depend on tumor characteristics in the preoperative MR scan. Currently, this is based on subjective estimates or manual tumor delineation in the absence of a standard for reporting. We compared tumor features of 1596 patients from 13 institutions extracted from manual segmentations by a human rater and from automated segmentations generated by a machine learning model. The automated segmentations were in excellent agreement with manual segmentations and are practically equivalent regarding tumor features that are potentially relevant for neurosurgical purposes. Standard reports can be generated by open access software, enabling comparison between surgical cohorts, multicenter trials, and patient registries.

Abstract

Treatment decisions for patients with presumed glioblastoma are based on tumor characteristics available from a preoperative MR scan. Tumor characteristics, including volume, location, and resectability, are often estimated or manually delineated. This process is time consuming and subjective. Hence, comparison across cohorts, trials, or registries are subject to assessment bias. In this study, we propose a standardized Glioblastoma Surgery Imaging Reporting and Data System (GSI-RADS) based on an automated method of tumor segmentation that provides standard reports on tumor features that are potentially relevant for glioblastoma surgery. As clinical validation, we determine the agreement in extracted tumor features between the automated method and the current standard of manual segmentations from routine clinical MR scans before treatment. In an observational consecutive cohort of 1596 adult patients with a first time surgery of a glioblastoma from 13 institutions, we segmented gadolinium-enhanced tumor parts both by a human rater and by an automated algorithm. Tumor features were extracted from segmentations of both methods and compared to assess differences, concordance, and equivalence. The laterality, contralateral infiltration, and the laterality indices were in excellent agreement. The native and normalized tumor volumes had excellent agreement, consistency, and equivalence. Multifocality, but not the number of foci, had good agreement and equivalence. The location profiles of cortical and subcortical structures were in excellent agreement. The expected residual tumor volumes and resectability indices had excellent agreement, consistency, and equivalence. Tumor probability maps were in good agreement. In conclusion, automated segmentations are in excellent agreement with manual segmentations and practically equivalent regarding tumor features that are potentially relevant for neurosurgical purposes. Standard GSI-RADS reports can be generated by open access software.

1. Introduction

The preoperative MR scan of a patient with a glioblastoma contains essential information that is interpreted by a neurosurgical team for a surgical strategy. Decisions on whether to perform a biopsy or a resection, estimations on how much tumor can be safely removed, the risks of complications and loss of brain functions, and judgements concerning the complexity of the surgery and ensuing pre- and intraoperative diagnostics are imperative for patient outcomes. In addition, the initial scan holds prognostic information, including tumor volume and location [1,2,3], which guides clinical decisions on radiotherapy and chemotherapy and serves patient counseling. In reports of surgical cohorts, multicenter trials, and registries, outcomes are customarily related to measurements of tumor characteristics on the initial scan and related to the outcomes and measurements of other teams [4,5,6,7,8,9,10,11,12,13,14,15]. Furthermore, these reports are pooled in meta-analyses enabling the identification of new patterns in the reported data to guide future clinical decisions [16,17]. Reliable measurements of tumor characteristics are therefore instrumental in patient care and in the development of glioblastoma treatment.
Whereas the response assessment of neuro-oncological treatment mainly focuses on changes in tumor volume over time [18,19] and radiotherapy planning on the clinical target volume on postoperative scans [20,21,22,23], pre-treatment tumor characteristics are of special interest for neurosurgical purposes. In addition to tumor volume, these include measurements of distance to and overlap with brain structures and expected resectability. The current standard is segmentation of the tumor in 3D, while qualitative description, measurement of tumor diameter, and bidimensional products are also in use [24]. These segmentations by human raters have disadvantages. Manual segmentations are time-consuming [25] and therefore expensive. It is common to have inexperienced students or junior investigators as raters for large numbers of segmentations. The level of experience of the rater is an important contributing factor to the accuracy of segmentations [26,27]. Certification of expert raters has not been established. The reproducibility of manual segmentations can be limited, probably due to human error, as attention may fluctuate in monotonous tasks [26,28,29,30,31,32,33]. In addition, segmentation updates or revisions take considerable time.
Automated segmentation algorithms have been developed and compared with manual segmentations as ground truth [34]. Convolutional neural networks [35], in particular employing U-Net [36], dominate the applications. Their performances have been benchmarked on a standardized image dataset (the Brain Tumor Image Segmentation, BraTS [32,34]), using a diagnostic accuracy approach with human rater segmentations as reference. In this approach, the spatial overlap of segmented voxels is typically reported as a Dice score, and the distance of segmentation surfaces as a Hausdorff metric. Nevertheless, this strictly determines the voxel-wise resemblance between an automated segmentation and the reference segmentation. This does not address the clinical utility of these segmentations, and the curated standardized image dataset is not representative for routine scans, which are often of suboptimal quality due to motion artefacts, missing sequences, and other image degradation. Furthermore, in routine scans, brains are not extracted, as is the case in the BraTS dataset.
Standard reporting and data systems (RADS) have been established for several solid tumors, including prostate cancer [37,38], hepatocellular carcinoma [39], head and neck squamous cell carcinoma [40], solitary bone tumors [41], bladder cancer [42], breast cancer [43], lymph node involvement by cancer [44], and lung cancer [45]. These RADS have enabled rules for imaging techniques, terminology for reports, definitions of tumor features, and treatment response, with less practice variation and reproducible tumor classification. Its broad implementation should facilitate collaborations and stimulate evaluation for development and improvement of RADS.
In this study, we determine the agreement in extracted tumor features between automated and manual segmentations from routine clinical MR scans before treatment and describe their discrepancies. We propose a standardized Glioblastoma Surgery Imaging Reporting and Data System (GSI-RADS) to automatically extract tumor features that are potentially relevant for glioblastoma surgery and demonstrate the use of a software module to create standard reports.

2. Materials and Methods

2.1. Patients and MR Images

We identified all patients of at least 18 years old with a newly diagnosed glioblastoma at first-time surgery between 1 January 2012 and 31 December 2013 from 13 hospitals: Northwest Clinics, Alkmaar, The Netherlands (ALK); Amsterdam University Medical Centers, location VU Medical Center, The Netherlands (AMS); University Medical Center Groningen, The Netherlands (GRO); Medical Center Haaglanden, The Hague, The Netherlands (HAG); Humanitas Research Hospital, Milano, Italy (MIL); Hôpital Lariboisière, Paris, France (PAR); University of California San Francisco Medical Center, US (SFR); Medical Center Slotervaart, Amsterdam, The Netherlands (SLO); St Elisabeth Hospital, Tilburg, The Netherlands (TIL); University Medical Center Utrecht, The Netherlands (UTR); Medical University Vienna, Austria (VIE); and Isala hospital, Zwolle, The Netherlands (ZWO), and between 2007 and 2018 from one hospital: St Olav’s hospital, Trondheim university Hospital, Norway (STO). Patients gave their informed consent for scientific use of their data, as required for each participating hospital. The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Medical Ethics Review Committee. Data and images for analysis were pseudonymized for analysis.
Patients were identified at each hospital by prospective electronic databases. Part of this cohort was reported earlier to address resectability and comparison of surgical decisions between institutes [46,47]. Descriptive information was collected from the electronical medical records, including age and gender.
Preoperative MR scans were acquired from the hospitals’ archival systems and included a 3D heavily T1-weighted gradient-echo pulse sequence at 1 mm isotropic resolution, obtained before and after administration of intravenous gadolinium, and a T2/FLAIR-weighted gradient-echo pulse sequence. MR scan protocols were standardized in hospitals but not identical between hospitals. Scanners from several vendors were in use, including Siemens, model Sonata, Avanto, Skyra, Prisma and mMR; GE medical systems, model Signa HDxt or DISCOVERY MR750; Toshiba, model Titan3T; and Philips, model Panorama HFO or Ingenuity with field strength of 1.5T or 3T. Detailed scan protocols have been described elsewhere [25,48].

2.2. Manual Tumor Segmentations

Tumors were manually segmented in 3D by trained raters using an initiation by either a region growing algorithm [26] (Brainlab SmartBrush, BrainLAB AG, Münich, Germany) or a grow cut algorithm [49] (3D Slicer, http://www.slicer.org, accessed on 3 June 2021) and subsequent manual editing. Trained raters were supervised by neuroradiologists and neurosurgeons. The tumor was defined as gadolinium-enhancing tissue on T1-weighted scans, including nonenhancing enclosed necrosis or cysts.

2.3. Automated Tumor Segmentations

A segmentation model was trained following a leave-one-hospital-out cross-validation strategy over the 1596 MRI volumes featured in our dataset, using the AGUNet architecture [50]. The model was trained from scratch, using the Dice Loss as cost function [51] and an Adam optimizer with an initial learning rate of 1e−3 and stopped after 30 epochs without validation loss improvement. Data augmentation was performed during training to improve generalization, such as random horizontal and vertical flip, rotation, and translation transforms.

2.4. Extracted Tumor Features

To correlate the tumor segmentations with standard anatomy, patient images were nonlinearly registered to a standard anatomical reference space, here consisting of the symmetric Montreal Neurological Institute ICBM2009a atlas, symmetric version 09a (MNI) [52,53], using symmetric image normalization as previously described [54,55]. From both the manual and the automated segmentation of each patient, the following measurements were extracted.
The laterality was defined as the main part of the tumor coinciding with either the left or right hemisphere, or none in the case where a tumor volume was not detected. Contralateral infiltration was defined as binary variable, true if any tumor voxel involved the contralateral hemisphere. The laterality index was defined as an index of tumor distribution between hemispheres, where −1 represents a tumor entirely located in the right hemisphere, 0 represents equal distribution of tumor between both hemispheres, and 1 represents a tumor completely located in the left hemisphere.
The native tumor volume in mL was defined as the number of tumor voxels in patient space times the volume of a tumor voxel in patient space. The normalized tumor volume in mL was defined as the number of tumor voxels in reference space times the volume of a tumor voxel in reference space.
Multifocality was defined as binary variable, true if more than one contrast-enhancing tumor component was observed and the second contrast-enhancing tumor component had a minimum volume of 0.1 mL and a minimum distance between the first and second largest tumor components of 5 mm. The number of foci was counted as the number of unconnected components.
The location profile of cortical structures is represented by the percentage of patients with a tumor per cortical parcel in a circular barplot [56]. We demonstrate the location profile of the cohort for two commonly used brain parcellations, Desikan’s brain parcellation with 96 parcels based on anatomy [57] and Schaefer’s brain parcellation with 17 network classes from 400 parcels based on functional connectivity using a resting state functional MRI [58,59]. Involvement of a patient’s tumor with a parcel was defined as any tumor voxel from a patient overlapping with that parcel.
The location profile of subcortical structures is represented by the percentage of patients with a tumor per white matter structure in a circular barplot [56]. The subcortical white matter structures deemed potentially relevant for surgery comprise a selection of tracts in each hemisphere, consisting of the corticospinal tract with a paracentral and three hand segments; the superior longitudinal fasciculus with three divisions; the arcuate fasciculus with a long, anterior, and posterior segment; the frontal aslant tract; the frontal striatal tract; the inferior fronto-occipital fasciculus; the uncinate fasciculus; the inferior longitudinal fascicle; and the optic radiation. The white matter structure definitions from the Brain Connectivity and Behaviour group were used [60]. The involvement of a patient’s tumor with a structure was defined as any tumor voxel overlapping with that white matter structure.
The expected residual tumor volume and the expected resectability index were calculated with a resection probability map of 451 patients with glioblastoma surgery in the left hemisphere and 464 patients in the right hemisphere, as reference, consisting of a subset of the current study population [46]. To calculate the resectability, the tumor segmentation masked the resection probability map. The resection probabilities of the masked voxels were summed to obtain the expected resectable volume. The preoperative tumor volume minus the expected resectable volume resulted in the expected residual tumor volume in mL. A division of the expected resectable volume by the preoperative tumor volume resulted in the expected resectability index, ranging from 0.0 to 1.0. This method has been detailed and validated elsewhere [46].
The tumor probability map was constructed for the whole population as 3D volume in standard brain space at 1 mm resolution. The fraction of tumors divided by the total number of patients was calculated voxel-wise.

2.5. Software Module and Standard Report

The proposed GSI-RADS software (https://github.com/SINTEFMedtek/GSI-RADS, accessed on 3 June 2021) enables the extraction of the described tumor features from a patient’s preoperative MR scan locally. The software has been developed in Python 3 and is compatible for use on Windows 10 (Microsoft Corp., Redmond, WA, USA), macOS (≥10.13; Apple Inc., Cupertino, CA, USA), and Ubuntu Linux 18.04 (Canonical Group Ltd., London, UK). A minimalistic GUI is provided to the user for specifying the required parameters and running the process. The input for the software consists of a 3D T1-weighted gadolinium-enhanced MRI volume provided as a DICOM sequence or NIfTI format. A manual segmentation of the tumor can be provided by the user (e.g., NIfTI format); if not, an automatic segmentation will be generated using the trained model. The output consists of a generated standard report in text (.txt) and CSV format, alongside multiple NIfTI files containing the tumor segmentation as a binary mask (in patient and MNI spaces), the registered MR scan in MNI space, and the anatomical region masks in patient space.
The standard report summarizes the extracted tumor features for each patient. These include the tumor laterality, contralateral infiltration, the laterality index, the native and normalized tumor volumes, the presence of multifocality and the number of foci, the percentage of tumor overlap with cortical parcels and subcortical structures, the expected residual tumor volume and expected resectability, and binary maps of the tumor segmentation in patient space and standard brain space.

2.6. Statistical Analysis

Differences in laterality, contralateral infiltration, multifocality, number of foci, and cortical and subcortical profiles between automated and manual segmentations were evaluated in contingency tables and tested for significance of paired data using McNemar’s test for two classes and Friedman’s test for more than two classes. The concordance as a percentage was calculated by dividing the sum of concordant classes over the total number of patients. Differences in native and normalized tumor volumes and expected residual volumes and resectability indices were tested for significance using the Wilcoxon signed-rank test for paired data. Agreement in laterality index, native and normalized tumor volumes, expected residual tumor volumes, and resectability indices between automated and manual segmentations was displayed in histograms, scatter plots, and Bland–Altman plots and calculated as an intraclass-correlation coefficient using a one-way model based on agreement with 95% confidence interval [61,62,63,64]. Equivalence in laterality, contralateral infiltration, native and normalized tumor volumes, multifocality, number of foci, expected residual tumor volumes, and resectability indices were tested using two one-sided tests for the smallest effect size of interest [65]. The smallest effect size of interest for equivalence bounds in proportions was considered to be 10%, for volumes two mL, for foci one focus, and for expected resectability indices 0.1. The product moment correlation coefficient with 95% confidence interval was calculated for the laterality indices, the native and normalized tumor volumes, expected residual tumor volumes, and expected resectability indices between automated and manual segmentations. Voxel-wise agreement was evaluated in tumor probability maps based on automated and manual segmentations. False discovery rates were calculated for the voxel-wise differences using a permutation test, as previously detailed [47,66].

3. Results

3.1. Patients

A total of 1596 patients were included in this analysis. No scans were excluded based on poor image quality or failed registration. A listing of the populations per hospital is provided in Table 1.

3.2. Agreement in Tumor Features between Manual and Automated Segmentations

3.2.1. Laterality, Contralateral Infiltration, and the Laterality Index

The automated and the manual segmentations, respectively, identified 785 (49.2%) and 794 (49.7%) patients with left-sided tumors, 792 (49.6%) and 799 (50.1%) patients with right-sided tumors, and 19 (1.2%) and 3 (0.2%) patients in whom no tumor volume was identified and hence were devoid of laterality, as listed in Table 2. Of the five discordant cases with opposing laterality, four were midline tumor with slightly dissimilar tumor voxel numbers in either hemisphere, and one scan was of poor quality with faint gadolinium enhancement of the tumor that the automated method failed to detect while a false positive segmentation of choroidal plexus was segmented contralaterally. In 17 (1.1%) patients, the automated segmentation did not identify a tumor, whereas the human rater did, due to minute tumor size, faint gadolinium enhancement, or poor scan quality. The observed laterality difference was statistically not different from zero (odds ratio: 0.98, 95% CI: 0.89–1.09; p-value = 0.744) and statistically equivalent to zero (95% CI: −0.029 to 0.030; Z = −5.59, p-value < 0.0001). The concordance was 98.6%.
Contralateral infiltration was observed in 430 (26.9%) patients based on the automated segmentations and in 469 (29.4%) based on the manual segmentations, as listed in Table 3. The observed difference in contralateral infiltration was statistically not different from zero (Z = 1.54, p-value = 0.125) and statistically not equivalent to zero (95% CI: −0.007 to 0.056; Z = −1.61, p-value = 0.0541). The concordance was 95.4%.
The distribution of the laterality indices determined by automated and manual segmentations and their correlation are shown in Figure 1A and the Bland–Altman plot in Figure 1B. The correlation coefficient was 0.998 (95% CI: 0.998–0.998). No bias was observed (0.00039, 95% CI: −0.0024 to 0.0032). The lower and upper 95% limits of agreement were −0.11 and 0.11.
This indicates excellent agreement to detect laterality, contralateral infiltration, and the laterality index between the segmentation methods.

3.2.2. Tumor Volumes

The difference between the native and normalized tumor volumes was plotted in Figure 2A,B. The median (interquartile range) of this difference for automated segmentations was −2.6 (6.8) mL and for manual segmentations −3.2 (7.5) mL. Apparently, the standard brain is somewhat larger than the brains of many patients. Therefore, we assessed normalized tumor volume in addition to native tumor volume.
The median (interquartile range) of the native tumor volumes was 26.5 (36.6) mL for automated segmentations and 26.6 (37.1) mL for manual segmentations, with a small but clinically negligible difference (0.4 mL, 95% CI: 0.4–0.5; p-value < 0.0001), well within the smallest effect size of interest of 2 mL (one-sided test for the upper bound t = −11.4, df = 1595, p-value < 0.0001 and for the lower bound t = 17.3, df = 1595, p-value < 0.0001).
The median (interquartile range) of the normalized tumor volumes was 30.1 (42.4) mL for the automated segmentations, and 31.2 (42.0) mL for the manual segmentations, again with a negligibly small difference (1.0 mL, 95% CI: 0.8–1.1; p-value < 0.0001), well within the smallest size of interest of 2 mL (one-sided test for the upper bound t = −4.9, df = 1595, p-value < 0.0001 and for the lower bound t = 17.3, df = 1595, p-value < 0.0001).
The intraclass correlation coefficient of the native tumor volumes was 98.2% (95% CI: 98.0–98.3%) and of the normalized tumor volumes 97.9% (95% CI: 97.7–98.1%), indicating excellent internal consistency.
In Figure 2C,E, the native and normalized tumor volumes based on automated and manual segmentations are plotted, indicating excellent agreement. In the Bland–Altman plots in Figure 2D,F, a small negligible systematic bias was observed between the automated and manual segmentations for native (0.4 mL, 95% CI: 0.1–0.7) and normalized tumor volumes (1.2 mL, 95% CI: 0.9–1.5). The limits of agreement were between −11.0 and 11.3 mL for the native tumor volumes and between −11.8 and 14.2 mL for the normalized tumor volumes.
This indicates excellent agreement, consistency, and equivalence in native and normalized tumor volume measurements between the automated and manual segmentations.

3.2.3. Multifocality and Number of Foci

Multifocality was identified in 320 (20.1%) patients for automated segmentations and in 374 (23.4%) for manual segmentations, as listed in Table 4. The observed multifocality difference was statistically different from zero (odds ratio 0.51, 95% CI: 0.37 to 0.72; p-value < 0.0001) and statistically equivalent to zero (95% CI: 0.010 to 0.058; Z = −4.54, p-value < 0.0001). The concordance was 89.5%.
The number of foci as determined by automated and manual segmentations is listed in Table 5. The observed number of foci was statistically different (Friedman chi-squared = 40.3, p-value < 0.0001). The concordance was 83.8%.
This indicates good agreement and equivalence in multifocality, but not in the number of foci between the automated and manual segmentations.

3.2.4. Location Profile of Cortical Parcels

The location profiles of the 96 cortical parcels from Desikan’s brain parcellation for the patient population are shown in Figure 3A,B according to the manual and automated segmentations. The well-known preferred locations of glioblastoma are apparent, and the incidence profiles of cortical involvement are almost identical between the segmentation methods. The correlation coefficient of the number of patients with parcel involvement as displayed in Figure 3C was 0.999 (95% CI: 0.999–0.999).
This indicates excellent agreement.
The location profiles of the 400 cortical parcels converging into 17 network classes from Schaefer’s brain parcellation for the patient population are shown in Figure 4A,B for the manual and automated segmentations. The incidence profiles of cortical involvement are almost identical between the segmentation methods. The correlation coefficient of the number of patients with parcel involvement as displayed in Figure 4C was 0.998 (95% CI: 0.998–0.999).
This indicates excellent agreement in cortical incidence profiles between the segmentation methods.

3.2.5. Location Profile of Subcortical Structures

The location profiles of 17 white matter tracts in either hemisphere for tumor overlap were compared for the whole population between the automated and manual segmentations in Figure 5A,B, respectively. The incidence profiles of cortical involvement are almost identical between the segmentation methods. The correlation coefficient of the number of patients with tract involvement was 0.999 (0.999–1.000), as displayed in Figure 5C.
This indicates excellent agreement between the segmentation methods.

3.2.6. Expected Residual Tumor Volume and Expected Resectability Index

The median (interquartile range) of the expected residual tumor volume was 4.5 (7.2) mL for automated segmentations and 4.7 (7.5) mL for manual segmentations, which have a small clinically negligible difference (0.2 mL, 95% CI: 0.2–0.3; p-value < 0.0001), within the smallest effect size of interest of 2 mL (one-sided test for the upper bound t = −35.6, df = 1575, p-value < 0.0001 and for the lower bound t = 56.7, df = 1575, p-value < 0.0001).
The median (interquartile range) of the expected resectability index was 0.857 (0.099) for automated segmentations and 0.849 (0.098) for manual segmentations, which have a small clinically negligible difference (−0.0025, 95% CI: −0.0035 to −0.0020; p-value < 0.0001), within the smallest effect size of interest of 0.1 (one-sided test for the upper bound t = −125, df = 1575, p-value < 0.0001 and for the lower bound t = 112, df = 1575, p-value < 0.0001).
Between automated and manual segmentations, the intraclass coefficient of the expected residual tumor volumes was 96.5% (95% CI: 96.2–96.8%), and the expected resectability index was 94.2% (95% CI: 93.6–94.7%), indicating excellent consistency.
In Figure 6A,C, the expected residual tumor volume and resectability index are plotted, indicating excellent correlation between the automated and manual segmentations. In the Bland–Altman plots in Figure 6B,D, a small negligible bias was observed between the automated and manual segmentations for the expected residual tumor volume (0.5 mL, 95% CI: 0.4–0.5) and for the expected resectability index (−0.005, 95% CI: −0.004 to −0.007). The limits of agreement were between −2.9 and 3.8 mL for the expected residual tumor volume and between −0.07 and 0.06 for the expected resectability index.
This indicates excellent agreement, consistency, and equivalence in expected residual tumor volume and resectability index between the segmentation methods.

3.2.7. Tumor Probability Map

The tumor probability maps based on automated and manual segmentations are provided in Figure 7. The maps were almost identical. Of 1.9 million brain voxels, none had an incidence difference with a false discovery rate below 20%.
This indicates excellent tumor probability map agreement between the segmentation methods.

3.3. Examples of Disagreement between Manual and Automated Segmentations

From inspection of the cases that showed lower agreement between automated and manual segmentations, four categories of disagreement emerged, as demonstrated in Figure 8: (i) false negative cystic tumor portions in the automated segmentations; (ii) false negative enhancing tumor volume, typically satellite lesions missed by the automated segmentation; (iii) mismatch in inclusion of nonenhancing tumor portions; and (iv) false positive vasculature structures or choroidal plexus, mistaken for a tumor.

3.4. GSI-RADS Software and Standard Report

An example of the generated output is shown in Figure 9 The numerical results are displayed as text in a window and can be exported in csv file format.

4. Discussion

The main finding of this study is that automated segmentations are in excellent agreement with manual segmentations regarding extracted tumor features, such as laterality, tumor volume, multifocality, location profiles of cortical parcels and subcortical structures, resectability, and tumor probability maps, which are potentially relevant for neurosurgical planning and reporting. This agreement supports at least equal validity of automated segmentations for these purposes. The generation of automated segmentations is more rapid and more reproducible than manual segmentations, as previously demonstrated [27]. We propose to substitute manual delineations with automated segmentation methods as standard in reports of patients with glioblastoma. To facilitate the distribution of these standard methods, we provide GSI-RADS as software to extract the most relevant tumor features from an MR scan, consisting of tumor laterality, volume, multifocality, location profiles of cortical and subcortical involvement, and resectability.
The use of a uniform method by the neurosurgical community to delineate a tumor and to extract tumor features would be an important step towards standardization across studies and between neurosurgical teams. A suitable segmentation method for neurosurgical use has several requirements: the method should be user friendly, rapid, scalable, accurate, reproducible, affordable, and valid [67]. The present software module is designed to minimize user interaction to import the DICOM scan. The processing duration of the automated method is a fraction of the manual method, which typically takes 30 min per patient [27], deterring to scale to cohorts larger than a few hundred patients. In absence of a ground truth for the exact tumor location, the accuracy of either method remains undetermined. Histopathological and molecular determination of tumor presence based on detailed multiregion sampling would theoretically be the ultimate ground truth [68]. This is infeasible for a patient cohort for obvious reasons. A second-best ground truth is postmortem investigation, although this would restrict a correlation to a recent last scan, and results may not extrapolate to the early stage of disease. An alternative ground truth could be an ensemble of segmentations by multiple expert raters, but this takes considerable time and expense restricted to a limited numbers of patients [69]. Therefore, we took a pragmatic approach and with equivalence between the segmentation methods, the question on the better method can remain unanswered. Automated segmentations are entirely reproducible and free, providing segmentations that can be updated through batch processing, whereas human raters are subject to disagreement between and within raters, yielding unreproducible data from a task that is not trivial in time and expense. In this study, we demonstrate that automated segmentations are equivalent to manual segmentations regarding neurosurgical tumor characteristics, hence they are equally valid. Either segmentation method may yield questionable results in a small subset of atypical tumors, characterized by faint contrast enhancement with large nonenhancing tumor portions, large cysts, or image artefacts. In the absence of a ground truth, we would argue that the reproducibility of an automated segmentation is preferable over arduous manual assessment, even in such less well-defined cases. Likewise, a pragmatic and reproducible standard for tumor volume, focality, location, and resectability based on automated segmentation is preferable over manual delineation.
Our finding that an automated processing by a ‘machine’ can replace a tedious and error-prone task by a ‘human’ adds to an already long list [70,71,72,73,74]. From this perspective, our findings are unsurprising and fit in the development of successful implementations of processes automated by deep learning.
Thus far, no other applications have been developed to extract tumor characteristics for use in glioblastoma surgery, although several applications were developed to segment the tumor in scans. The Brain Tumor Image Analysis tool (BraTumIA) has been developed to segment three brain tumor compartments using four scan sequences [33,75] and has been shown to have good agreement with manual tumor volumes on preoperative scans. The Pearson’s correlation coefficient between manual and BraTumIA tumor volumes was 0.8 based on 19 patients [75] and 0.88 based on 58 patients [76], albeit with a systematic overestimation. In addition, the BraTS challenge has been held yearly since 2012, which aims to improve disease diagnosis, treatment planning, monitoring, and clinical trials by means of reliable tumor segmentation. Participants have applied more than 200 models over the years. Many models were updated versions of previous submissions. As far as we are aware, none of these models has been used to generate tumor characteristics for neurosurgical practice. Therefore, the quest for the best performance in a common dataset by ranking of Dice score is not necessarily representative for clinical practice. In this study, we sought to address whether automation could replace manual labor without compromising validity in terms of tumor features and to make the software readily available for others to use and validate further, both clinically and technically. Future improvements of automated methods can be easily integrated in updated software.
A strength of this study is good external validity given the mixture of institutions, scanners, scan protocols, and patients. Until standardized scan acquisition protocols are implemented in neuro-oncological care [77], automated segmentation methods should resolve this practice variation. Another strength is the relatively large dataset for training the automated method. A limitation is that we used manual segmentations from one trained rater per tumor, although this probably represents current practice in neurosurgical reports of tumor characteristics.
A practical implication is that standard reports for glioblastoma surgery can now be generated by GSI-RADS. Obviously, improved patient outcomes cannot be expected from better reporting in itself. Indirectly, improved outcomes may result from more accurate data-driven decisions on the use of preoperative techniques such as DTI-based tractography, functional MRI, transcranial stimulation, and intraoperative stimulation mapping. Another indirect effect may be the facilitation of consultation between neurosurgeons and teams and possibly in referral patterns by better recognition of complex surgical cases regarding tumor location and eloquence. An example would be the identification of a more complex tumor near the arcuate fascicle, for instance, by a lower expected resectability index and infiltration of this tract, indicating additional preoperative diagnostics to detail the relation between the tumor and the tract and the use of intraoperative stimulation mapping to safely maximize tumor removal. As such, the automated methods hold potential for development of a quantitative standard for eloquence. Reliable definitions of pretreatment tumor characteristics from MR scans may also facilitate less biased comparisons across institutions, studies, or quality registries. Furthermore, prognostic information, surgical treatment evaluation, and response assessment may indirectly improve the risk stratification of patient cohorts. Finally, the standardized reports could speed up the learning curve and serve in the education and training of neurosurgeons.
In future efforts, several directions are important to explore. The automated segmentations can be extended to other pathology, such as lower-grade nonenhancing glioma, brain metastasis, and meningioma. Alternative automated methods can be benchmarked against the current results. The presented automated method can be trained with data from additional patients and institutions. New tumor features will be added to the standard report, such as different aspects of multifocality and the infiltration and disconnection of white matter pathways. These new measures should be compared with patient outcomes for evaluation of their clinical use [46]. This may, for instance, result in a quantitative assessment of risk for surgical complications and risk for early tumor progression. Other tumor compartments can be included, such as the T2/FLAIR hyperintense region, necrotic or ischemic tissue, hemorrhage, cyst fluid, and ultimately molecular heterogeneity and metabolic activity. In addition, reliable tumor segmentations over time and at different stages of disease would be instrumental to provide standardized reports of postsurgical evaluation and treatment response assessment. Finally, distribution of the software should be available for multiple platforms and environments, such as a standalone web-based application.

5. Conclusions

Automated segmentations are in excellent agreement with manual segmentations and are practically equivalent regarding tumor features that are potentially relevant for neurosurgical purposes. A standard GSI-RADS report is proposed for these tumor features, including the laterality, volume, multifocality, location, and resectability (https://github.com/SINTEFMedtek/GSI-RADS, accessed on 3 June 2021).

Author Contributions

Conceptualization, I.K., D.B., A.P., R.S.E., I.R., O.S. and P.C.D.W.H.; Data curation, I.K., D.B., H.A., F.B., L.B., M.S.B., M.C.N., J.F., E.H.F., S.H.-J., A.J.S.I., B.K., A.K., E.M., D.M.J.M., P.A.R., M.R., L.M.S., T.S., W.A.v.d.B., M.W., G.W., O.S. and P.C.D.W.H.; Formal analysis, I.K., D.B., A.P., R.S.E., A.H.Z., I.R., O.S. and P.C.D.W.H.; Funding acquisition, I.R., O.S. and P.C.D.W.H.; Investigation, I.K., D.B., R.S.E., F.B., L.B., D.M.J.M., M.W., A.H.Z., I.R., O.S. and P.C.D.W.H.; Methodology, I.K., D.B., A.P., R.S.E., H.A., F.B., L.B., M.S.B., J.F., E.H.F., S.H.-J., A.J.S.I., B.K., A.K., E.M., D.M.J.M., P.A.R., M.R., L.M.S., T.S., W.A.v.d.B., M.W., G.W., M.G.W., A.H.Z., I.R., O.S. and P.C.D.W.H.; Project administration, I.R., O.S. and P.C.D.W.H.; Resources, H.A., M.S.B., J.F., S.H.-J., A.J.S.I., B.K., A.K., E.M., P.A.R., M.R., W.A.v.d.B., M.W., G.W., O.S. and P.C.D.W.H.; Software, D.B., A.P., R.S.E., M.W., I.R., O.S. and P.C.D.W.H.; Supervision, I.R., O.S. and P.C.D.W.H.; Validation, I.K., D.B., H.A., F.B., L.B., M.S.B., M.C.N., J.F., E.H.F., S.H.-J., A.J.S.I., B.K., A.K., E.M., P.A.R., M.R., L.M.S., T.S., W.A.v.d.B., M.W., G.W., M.G.W., I.R., O.S. and P.C.D.W.H.; Visualization, I.K. and P.C.D.W.H.; Writing—original draft, I.K., A.P., R.S.E., I.R., O.S. and P.C.D.W.H.; Writing—review and editing, D.B., H.A., F.B., L.B., M.S.B., M.C.N., J.F., E.H.F., S.H.-J., A.J.S.I., B.K., A.K., E.M., D.M.J.M., P.A.R., M.R., L.M.S., T.S., W.A.v.d.B., M.W., G.W., M.G.W. and A.H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by an unrestricted grant of Stichting Hanarth fonds, “Machine learning for better neurosurgical decisions in patients with glioblastoma”; a grant for public-private partnerships (Amsterdam UMC PPP-grant) sponsored by the Dutch government (Ministry of Economic Affairs) through the Rijksdienst voor Ondernemend Nederland (RVO) and Topsector Life Sciences and Health (LSH), “Picturing predictions for patients with brain tumors”; a grant from the Innovative Medical Devices Initiative program, project number 10-10400-96-14003; The Netherlands Organisation for Scientific Research (NWO), 2020.027; a grant from the Dutch Cancer Society, VU2014-7113; the Anita Veldman foundation, CCA2018-2-17; and the Norwegian National Advisory Unit for Ultrasound and Image Guided Therapy.

Institutional Review Board Statement

IRB approval was obtained from the Medical Ethics Review Committee of VU University Medical Center (IRB00002991, 2014.336) and from the Norwegian regional ethics committee (REK ref. 2013/1348 and 2019/510).

Informed Consent Statement

Informed consent was obtained from patients as required for each participating hospital.

Data Availability Statement

The manual segmentation data can be found as a publicly archived dataset (https://doi.org/10.17026/dans-xam-j5aw, accessed on 3 June 2021). The data and code for analysis can be found as a public archive (https://gitlab.com/picture/gsi-rads, accessed on 3 June 2021). The open access software can be found as a public archive (https://github.com/SINTEFMedtek/GSI-RADS, accessed on 3 June 2021).

Acknowledgments

BrainLab® has generously provided us with their proprietary neuronavigational software, which was used for the manual segmentation. This work was in part conducted on the Dutch national e-infrastructure with the support of SURF Cooperative and the Translational Research IT (TraIT) project, an initiative from the Center for Translational Molecular Medicine (CTMM).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Verduin, M.; Primakov, S.; Compter, I.; Woodruff, H.C.; van Kuijk, S.M.J.; Ramaekers, B.L.T.; Dorsthorst, M.T.; Revenich, E.G.M.; Laan, M.T.; Pegge, S.A.H.; et al. Prognostic and predictive value of integrated qualitative and quantitative magnetic resonance imaging analysis in glioblastoma. Cancers 2021, 13, 722. [Google Scholar] [CrossRef] [PubMed]
  2. Kickingereder, P.; Burth, S.; Wick, A.; Götz, M.; Eidel, O.; Schlemmer, H.P.; Maier-Hein, K.H.; Wick, W.; Bendszus, M.; Radbruch, A.; et al. Radiomic profiling of glioblastoma: Identifying an imaging predictor of patient survival with improved performance over established clinical and radiologic risk models. Radiology 2016, 280, 880–889. [Google Scholar] [CrossRef] [PubMed]
  3. Kickingereder, P.; Neuberger, U.; Bonekamp, D.; Piechotta, P.L.; Götz, M.; Wick, A.; Sill, M.; Kratz, A.; Shinohara, R.T.; Jones, D.T.W.; et al. Radiomic subtyping improves disease stratification beyond key molecular, clinical, and standard imaging characteristics in patients with glioblastoma. Neuro-Oncol. 2018, 20, 848–857. [Google Scholar] [CrossRef]
  4. Graus, F.; Bruna, J.; Pardo, J.; Escudero, D.; Vilas, D.; Barceló, I.; Brell, M.; Pascual, C.; Crespo, J.A.; Erro, E.; et al. Patterns of care and outcome for patients with glioblastoma diagnosed during 2008–2010 in Spain. Neuro-Oncol. 2013, 15, 797–805. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. De Witt Hamer, P.C.; Ho, V.K.Y.; Zwinderman, A.H.; Ackermans, L.; Ardon, H.; Boomstra, S.; Bouwknegt, W.; van den Brink, W.A.; Dirven, C.M.; van der Gaag, N.A.; et al. Between-hospital variation in mortality and survival after glioblastoma surgery in the Dutch Quality Registry for Neuro Surgery. J. Neurooncol. 2019, 144, 313–323. [Google Scholar] [CrossRef] [Green Version]
  6. Mathiesen, T.; Peredo, I.; Lönn, S. Two-year survival of low-grade and high-grade glioma patients using data from the Swedish Cancer Registry. Acta Neurochir. 2011, 153, 467–471. [Google Scholar] [CrossRef]
  7. Wöhrer, A.; Waldhör, T.; Heinzl, H.; Hackl, M.; Feichtinger, J.; Gruber-Mösenbacher, U.; Kiefer, A.; Maier, H.; Motz, R.; Reiner-Concin, A.; et al. The Austrian Brain Tumour Registry: A cooperative way to establish a population-based brain tumour registry. J. Neurooncol. 2009, 95, 401–411. [Google Scholar] [CrossRef]
  8. Arrigo, R.T.; Boakye, M.; Skirboll, S.L. Patterns of care and survival for glioblastoma patients in the Veterans population. J. Neurooncol. 2012, 106, 627–635. [Google Scholar] [CrossRef]
  9. Sawaya, R.; Hammoud, M.; Schoppa, D.; Hess, K.R.; Wu, S.Z.; Shi, W.M.; Wildrick, D.M. Neurosurgical outcomes in a modern series of 400 craniotomies for treatment of parenchymal tumors. Neurosurgery 1998, 42, 1044–1055. [Google Scholar] [CrossRef]
  10. Chen, J.Y.; Hovey, E.; Rosenthal, M.; Livingstone, A.; Simes, J. Neuro-oncology practices in Australia: A Cooperative Group for Neuro-Oncology patterns of care study. Asia-Pac. J. Clin. Oncol. 2014, 10, 162–167. [Google Scholar] [CrossRef] [Green Version]
  11. Chang, S.M.; Parney, I.F.; Huang, W.; Anderson, F.A.; Asher, A.L.; Bernstein, M.; Lillehei, K.O.; Brem, H.; Berger, M.S.; Laws, E.R.; et al. Patterns of care for adults with newly diagnosed malignant glioma. JAMA 2005, 293, 557–564. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Chaichana, K.L.; Jusue-Torres, I.; Navarro-Ramirez, R.; Raza, S.M.; Pascual-Gallego, M.; Ibrahim, A.; Hernandez-Hermann, M.; Gomez, L.; Ye, X.; Weingart, J.D.; et al. Establishing percent resection and residual volume thresholds affecting survival and recurrence for patients with newly diagnosed intracranial glioblastoma. Neuro-Oncol. 2014, 16, 113–122. [Google Scholar] [CrossRef] [PubMed]
  13. Gramatzki, D.; Dehler, S.; Rushing, E.J.; Zaugg, K.; Hofer, S.; Yonekawa, Y.; Bertalanffy, H.; Valavanis, A.; Korol, D.; Rohrmann, S.; et al. Glioblastoma in the Canton of Zurich, Switzerland revisited: 2005 to 2009. Cancer 2016, 122, 2206–2215. [Google Scholar] [CrossRef] [PubMed]
  14. Rosenthal, M.A.; Drummond, K.J.; Dally, M.; Murphy, M.; Cher, L.; Ashley, D.; Thursfield, V.; Giles, G.G. Management of glioma in Victoria (1998–2000): Retrospective cohort study. Med. J. Aust. 2006, 184, 270–273. [Google Scholar] [CrossRef]
  15. Zinn, P.; Colen, R.R.; Kasper, E.M.; Burkhardt, J.-K. Extent of resection and radiotherapy in GBM: A 1973 to 2007 surveillance, epidemiology and end results analysis of 21,783 patients. Int. J. Oncol. 2013, 42, 929–934. [Google Scholar] [CrossRef] [Green Version]
  16. Brown, T.J.; Brennan, M.C.; Li, M.; Church, E.W.; Brandmeir, N.J.; Rakszawski, K.L.; Patel, A.S.; Rizk, E.B.; Suki, D.; Sawaya, R.; et al. Association of the Extent of Resection With Survival in Glioblastoma: A Systematic Review and Meta-analysis. JAMA Oncol. 2016, 2, 1460–1469. [Google Scholar] [CrossRef] [Green Version]
  17. De Witt Hamer, P.C.; Robles, S.G.; Zwinderman, A.H.; Duffau, H.; Berger, M.S. Impact of intraoperative stimulation brain mapping on glioma surgery outcome: A meta-analysis. J. Clin. Oncol. 2012, 30, 2559–2565. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Kickingereder, P.; Isensee, F.; Tursunova, I.; Petersen, J.; Neuberger, U.; Bonekamp, D.; Brugnara, G.; Schell, M.; Kessler, T.; Foltyn, M.; et al. Automated quantitative tumour response assessment of MRI in neuro-oncology with artificial neural networks: A multicentre, retrospective study. Lancet Oncol. 2019, 20, 728–740. [Google Scholar] [CrossRef] [Green Version]
  19. Chang, K.; Beers, A.L.; Bai, H.X.; Brown, J.M.; Ina Ly, K.; Li, X.; Senders, J.T.; Kavouridis, V.K.; Boaro, A.; Su, C.; et al. Automatic assessment of glioma burden: A deep learning algorithm for fully automated volumetric and bidimensional measurement. Neuro-Oncol. 2019, 21, 1412–1422. [Google Scholar] [CrossRef]
  20. Ermiş, E.; Jungo, A.; Poel, R.; Blatti-Moreno, M.; Meier, R.; Knecht, U.; Aebersold, D.M.; Fix, M.K.; Manser, P.; Reyes, M.; et al. Fully automated brain resection cavity delineation for radiation target volume definition in glioblastoma patients using deep learning. Radiat. Oncol. 2020, 15, 1–10. [Google Scholar] [CrossRef]
  21. Shusharina, N.; Söderberg, J.; Edmunds, D.; Löfman, F.; Shih, H.; Bortfeld, T. Automated delineation of the clinical target volume using anatomically constrained 3D expansion of the gross tumor volume. Radiother. Oncol. 2020, 146, 37–43. [Google Scholar] [CrossRef] [PubMed]
  22. Unkelbach, J.; Bortfeld, T.; Cardenas, C.E.; Gregoire, V.; Hager, W.; Heijmen, B.; Jeraj, R.; Korreman, S.S.; Ludwig, R.; Pouymayou, B.; et al. The role of computational methods for automating and improving clinical target volume definition. Radiother. Oncol. 2020, 153, 15–25. [Google Scholar] [CrossRef] [PubMed]
  23. Rahmat, R.; Brochu, F.; Li, C.; Sinha, R.; Price, S.J.; Jena, R. Semi-automated construction of patient individualised clinical target volumes for radiotherapy treatment of glioblastoma utilising diffusion tensor decomposition maps. Br. J. Radiol. 2020, 93, 20190441. [Google Scholar] [CrossRef]
  24. Berntsen, E.M.; Stensjøen, A.L.; Langlo, M.S.; Simonsen, S.Q.; Christensen, P.; Moholdt, V.A.; Solheim, O. Volumetric segmentation of glioblastoma progression compared to bidimensional products and clinical radiological reports. Acta Neurochir. 2020, 162, 379–387. [Google Scholar] [CrossRef]
  25. Fyllingen, E.H.; Stensjøen, A.L.; Berntsen, E.M.; Solheim, O.; Reinertsen, I. Glioblastoma Segmentation: Comparison of Three Different Software Packages. PLoS ONE 2016, 11, e0164891. [Google Scholar] [CrossRef] [Green Version]
  26. Huber, T.; Alber, G.; Bette, S.; Boeckh-Behrens, T.; Gempt, J.; Ringel, F.; Alberts, E.; Zimmer, C.; Bauer, J.S.; Boeckh-behrens, T.; et al. Reliability of Semi-Automated Segmentations in Glioblastoma. Clin. Neuroradiol. 2015. [Google Scholar] [CrossRef]
  27. Visser, M.; Müller, D.M.J.; van Duijn, R.J.M.; Smits, M.; Verburg, N.; Hendriks, E.J.; Nabuurs, R.J.A.; Bot, J.C.J.; Eijgelaar, R.S.; Witte, M.; et al. Inter-rater agreement in glioma segmentations on longitudinal MRI. NeuroImage Clin. 2019, 22, 101727. [Google Scholar] [CrossRef]
  28. Gooya, A.; Pohl, K.M.; Bilello, M.; Cirillo, L.; Biros, G.; Melhem, E.R.; Davatzikos, C. GLISTR: Glioma Image Segmentation and Registration. IEEE Trans. Med. Imaging 2012, 31, 1941–1954. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Cordova, J.S.; Schreibmann, E.; Hadjipanayis, C.G.; Guo, Y.; Shu, H.-K.G.; Shim, H.; Holder, C.A. Quantitative Tumor Segmentation for Evaluation of Extent of Glioblastoma Resection to Facilitate Multisite Clinical Trials. Transl. Oncol. 2014, 7, 40–47. [Google Scholar] [CrossRef] [Green Version]
  30. Kubben, P.L.; Postma, A.A.; Kessels, A.G.H.; van Overbeeke, J.J.; van Santbrink, H. Intraobserver and interobserver agreement in volumetric assessment of glioblastoma multiforme resection. Neurosurgery 2010, 67, 1329–1334. [Google Scholar] [CrossRef] [PubMed]
  31. Kleesiek, J.; Petersen, J.; Döring, M.; Maier-Hein, K.; Köthe, U.; Wick, W.; Hamprecht, F.A.; Bendszus, M.; Biller, A. Virtual Raters for Reproducible and Objective Assessments in Radiology. Sci. Rep. 2016, 6, 25007. [Google Scholar] [CrossRef] [Green Version]
  32. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS). IEEE Trans. Med. Imaging 2015, 34, 1993–2024. [Google Scholar] [CrossRef]
  33. Porz, N.; Bauer, S.; Pica, A.; Schucht, P.; Beck, J.; Verma, R.K.; Slotboom, J.; Reyes, M.; Wiest, R. Multi-modal glioblastoma segmentation: Man versus machine. PLoS ONE 2014, 9, e96873. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Ghaffari, M.; Sowmya, A.; Oliver, R. Automated Brain Tumor Segmentation Using Multimodal Brain Scans: A Survey Based on Models Submitted to the BraTS 2012-2018 Challenges. IEEE Rev. Biomed. Eng. 2020, 13, 156–168. [Google Scholar] [CrossRef]
  35. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  36. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  37. Weinreb, J.C.; Barentsz, J.O.; Choyke, P.L.; Cornud, F.; Haider, M.A.; Macura, K.J.; Margolis, D.; Schnall, M.D.; Shtern, F.; Tempany, C.M.; et al. PI-RADS Prostate Imaging—Reporting and Data System: 2015, Version 2. Eur. Urol. 2016, 69, 16–40. [Google Scholar] [CrossRef] [PubMed]
  38. Ahmed, H.U.; El-Shater Bosaily, A.; Brown, L.C.; Gabe, R.; Kaplan, R.; Parmar, M.K.; Collaco-Moraes, Y.; Ward, K.; Hindley, R.G.; Freeman, A.; et al. Diagnostic accuracy of multi-parametric MRI and TRUS biopsy in prostate cancer (PROMIS): A paired validating confirmatory study. Lancet 2017, 389, 815–822. [Google Scholar] [CrossRef] [Green Version]
  39. Chernyak, V.; Fowler, K.J.; Kamaya, A.; Kielar, A.Z.; Elsayes, K.M.; Bashir, M.R.; Kono, Y.; Do, R.K.; Mitchell, D.G.; Singal, A.G.; et al. Liver Imaging Reporting and Data System (LI-RADS) Version 2018: Imaging of Hepatocellular Carcinoma in At-Risk Patients. Radiology 2018, 289, 816–830. [Google Scholar] [CrossRef] [PubMed]
  40. Ashour, M.M.; Darwish, E.A.F.; Fahiem, R.M.; Abdelaziz, T.T. MRI Posttreatment Surveillance for Head and Neck Squamous Cell Carcinoma: Proposed MR NI-RADS Criteria. AJNR. Am. J. Neuroradiol. 2021. [Google Scholar] [CrossRef]
  41. Ribeiro, G.J.; Gillet, R.; Hossu, G.; Trinh, J.-M.; Euxibie, E.; Sirveaux, F.; Blum, A.; Teixeira, P.A.G. Solitary bone tumor imaging reporting and data system (BTI-RADS): Initial assessment of a systematic imaging evaluation and comprehensive reporting method. Eur. Radiol. 2021, 1–16. [Google Scholar] [CrossRef]
  42. Akcay, A.; Yagci, A.B.; Celen, S.; Ozlulerden, Y.; Turk, N.S.; Ufuk, F. VI-RADS score and tumor contact length in MRI: A potential method for the detection of muscle invasion in bladder cancer. Clin. Imaging 2021, 77, 25–36. [Google Scholar] [CrossRef]
  43. Kwon, M.-R.; Choi, J.S.; Won, H.; Ko, E.Y.; Ko, E.S.; Park, K.W.; Han, B.-K. Breast Cancer Screening with Abbreviated Breast MRI: 3-year Outcome Analysis. Radiology 2021, 299, 73–83. [Google Scholar] [CrossRef]
  44. Elsholtz, F.H.J.; Asbach, P.; Haas, M.; Becker, M.; Beets-Tan, R.G.H.; Thoeny, H.C.; Padhani, A.R.; Hamm, B. Introducing the Node Reporting and Data System 1.0 (Node-RADS): A concept for standardized assessment of lymph nodes in cancer. Eur. Radiol. 2021, 1–9. [Google Scholar] [CrossRef]
  45. Dyer, S.C.; Bartholmai, B.J.; Koo, C.W. Implications of the updated Lung CT Screening Reporting and Data System (Lung-RADS version 1.1) for lung cancer screening. J. Thorac. Dis. 2020, 12, 6966–6977. [Google Scholar] [CrossRef]
  46. Müller, D.M.J.; Robe, P.A.; Ardon, H.; Barkhof, F.; Bello, L.; Berger, M.S.; Bouwknegt, W.; Van den Brink, W.A.; Conti Nibali, M.; Eijgelaar, R.S.; et al. Quantifying eloquent locations for glioblastoma surgery using resection probability maps. J. Neurosurg. 2020, 1, 1–11. [Google Scholar] [CrossRef]
  47. Müller, D.M.J.; Robe, P.A.J.T.; Eijgelaar, R.S.; Witte, M.G.; Visser, M.; de Munck, J.C.; Broekman, M.L.D.; Seute, T.; Hendrikse, J.; Noske, D.P.; et al. Comparing Glioblastoma Surgery Decisions Between Teams Using Brain Maps of Tumor Locations, Biopsies, and Resections. JCO Clin. Cancer Inform. 2019, 2, 1–12. [Google Scholar] [CrossRef]
  48. Eijgelaar, R.S.; Visser, M.; Müller, D.M.J.; Barkhof, F.; Vrenken, H.; van Herk, M.; Bello, L.; Conti Nibali, M.; Rossi, M.; Sciortino, T.; et al. Robust Deep Learning–based Segmentation of Glioblastoma on Routine Clinical MRI Scans Using Sparsified Training. Radiol. Artif. Intell. 2020, 2, e190103. [Google Scholar] [CrossRef]
  49. Vezhnevets, V.; Konouchine, V. GrowCut—Interactive multi-label N-D image segmentation by cellular automata. In Proceedings of the Fifteenth International Conference on Computer Graphics and Applications (GraphiCon’2005), Novosibirsk Akademgorodok, Russia, 20–24 June 2005. [Google Scholar]
  50. Bouget, D.; Pedersen, A.; Hosainey, S.A.M.; Solheim, O.; Reinertsen, I. Meningioma segmentation in T1-weighted MRI leveraging global context and attention mechanisms. arXiv 2021, arXiv:2101.07715. [Google Scholar]
  51. Sudre, C.H.; Li, W.; Vercauteren, T.; Ourselin, S.; Jorge Cardoso, M. Generalised Dice Overlap as a Deep Learning Loss Function for Highly Unbalanced Segmentations. In Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Berlin/Heidelberg, Germany, 2017; Volume 10553 LNCS, pp. 240–248. ISBN 9783319675572. [Google Scholar]
  52. Fonov, V.; Evans, A.C.; Botteron, K.; Almli, C.R.; McKinstry, R.C.; Collins, D.L. Brain Development Cooperative Group Unbiased average age-appropriate atlases for pediatric studies. Neuroimage 2011, 54, 313–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Fonov, V.; Evans, A.; McKinstry, R.; Almli, C.; Collins, D. Unbiased nonlinear average age-appropriate brain templates from birth to adulthood. Neuroimage 2009, 47, S102. [Google Scholar] [CrossRef]
  54. Visser, M.; Petr, J.; Müller, D.M.J.; Eijgelaar, R.S.; Hendriks, E.J.; Witte, M.; Barkhof, F.; van Herk, M.; Mutsaerts, H.J.M.M.; Vrenken, H.; et al. Accurate MR Image Registration to Anatomical Reference Space for Diffuse Glioma. Front. Neurosci. 2020, 14. [Google Scholar] [CrossRef]
  55. Avants, B.B.; Epstein, C.; Grossman, M.; Gee, J. Symmetric diffeomorphic image registration with cross-correlation: Evaluating automated labeling of elderly and neurodegenerative brain. Med. Image Anal. 2008, 12, 26–41. [Google Scholar] [CrossRef] [Green Version]
  56. Gu, Z.; Gu, L.; Eils, R.; Schlesner, M.; Brors, B. circlize implements and enhances circular visualization in R. Bioinformatics 2014, 30, 2811–2812. [Google Scholar] [CrossRef] [Green Version]
  57. Desikan, R.S.; Ségonne, F.; Fischl, B.; Quinn, B.T.; Dickerson, B.C.; Blacker, D.; Buckner, R.L.; Dale, A.M.; Maguire, R.P.; Hyman, B.T.; et al. An automated labeling system for subdividing the human cerebral cortex on MRI scans into gyral based regions of interest. Neuroimage 2006, 31, 968–980. [Google Scholar] [CrossRef]
  58. Schaefer, A.; Kong, R.; Gordon, E.M.; Laumann, T.O.; Zuo, X.-N.; Holmes, A.J.; Eickhoff, S.B.; Yeo, B.T.T. Local-Global Parcellation of the Human Cerebral Cortex from Intrinsic Functional Connectivity MRI. Cereb. Cortex 2018, 28, 3095–3114. [Google Scholar] [CrossRef]
  59. Yeo, B.T.T.; Krienen, F.M.; Sepulcre, J.; Sabuncu, M.R.; Lashkari, D.; Hollinshead, M.; Roffman, J.L.; Smoller, J.W.; Zöllei, L.; Polimeni, J.R.; et al. The organization of the human cerebral cortex estimated by intrinsic functional connectivity. J. Neurophysiol. 2011, 106, 1125–1165. [Google Scholar] [CrossRef] [PubMed]
  60. Rojkova, K.; Volle, E.; Urbanski, M.; Humbert, F.; Dell’Acqua, F.; Thiebaut de Schotten, M. Atlasing the frontal lobe connections and their variability due to age and education: A spherical deconvolution tractography study. Brain Struct. Funct. 2016, 221, 1751–1766. [Google Scholar] [CrossRef] [PubMed]
  61. Bland, J.M.; Altman, D.G. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet 1986, 1, 307–310. [Google Scholar] [CrossRef]
  62. Bland, J.M.; Altman, D.G. Measuring agreement in method comparison studies. Stat. Methods Med. Res. 1999, 8, 135–160. [Google Scholar] [CrossRef] [PubMed]
  63. McGraw, K.O.; Wong, S.P. Forming inferences about some intraclass correlation coefficients. Psychol. Methods 1996, 1, 30–46. [Google Scholar] [CrossRef]
  64. Shrout, P.E.; Fleiss, J.L. Intraclass correlations: Uses in assessing rater reliability. Psychol. Bull. 1979, 86, 420–428. [Google Scholar] [CrossRef] [PubMed]
  65. Harms, C.; Lakens, D. Making “null effects” informative: Statistical techniques and inferential frameworks. J. Clin. Transl. Res. 2018, 3, 382–393. [Google Scholar] [CrossRef] [Green Version]
  66. Eijgelaar, R.; de Witt Hamer, P.C.; Peeters, C.F.W.; Barkhof, F.; van Herk, M.; Witte, M.G. Voxelwise statistical methods to localize practice variation in brain tumor surgery. PLoS ONE 2019, 14, e0222939. [Google Scholar] [CrossRef]
  67. Bossuyt, P.M.M.; Reitsma, J.B.; Linnet, K.; Moons, K.G.M. Beyond diagnostic accuracy: The clinical utility of diagnostic tests. Clin. Chem. 2012, 58, 1636–1643. [Google Scholar] [CrossRef]
  68. Verburg, N.; Koopman, T.; Yaqub, M.M.; Hoekstra, O.S.; Lammertsma, A.A.; Barkhof, F.; Pouwels, P.J.W.; Reijneveld, J.C.; Heimans, J.J.; Rozemuller, A.J.M.; et al. Improved detection of diffuse glioma infiltration with imaging combinations: A diagnostic accuracy study. Neuro-Oncol. 2020, 22, 412–422. [Google Scholar] [CrossRef]
  69. Warfield, S.K.; Zou, K.H.; Wells, W.M. Simultaneous Truth and Performance Level Estimation (STAPLE): An Algorithm for the Validation of Image Segmentation. IEEE Trans. Med. Imaging 2004, 23, 903–921. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Tschandl, P.; Codella, N.; Akay, B.N.; Argenziano, G.; Braun, R.P.; Cabo, H.; Gutman, D.; Halpern, A.; Helba, B.; Hofmann-Wellenhof, R.; et al. Comparison of the accuracy of human readers versus machine-learning algorithms for pigmented skin lesion classification: An open, web-based, international, diagnostic study. Lancet Oncol. 2019, 20, 938–947. [Google Scholar] [CrossRef]
  71. Chan, S.; Siegel, E.L. Will machine learning end the viability of radiology as a thriving medical specialty? Br. J. Radiol. 2019, 92, 20180416. [Google Scholar] [CrossRef] [PubMed]
  72. He, J.; Baxter, S.L.; Xu, J.; Xu, J.; Zhou, X.; Zhang, K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019, 25, 30–36. [Google Scholar] [CrossRef]
  73. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  74. Dreyer, K.J.; Geis, J.R. When Machines Think: Radiology’s Next Frontier. Radiology 2017, 285, 713–718. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  75. Porz, N.; Habegger, S.; Meier, R.; Verma, R.; Jilch, A.; Fichtner, J.; Knecht, U.; Radina, C.; Schucht, P.; Beck, J.; et al. Fully Automated Enhanced Tumor Compartmentalization: Man vs. Machine Reloaded. PLoS ONE 2016, 11, e0165302. [Google Scholar] [CrossRef] [PubMed]
  76. Zeppa, P.; Neitzert, L.; Mammi, M.; Monticelli, M.; Altieri, R.; Castaldo, M.; Cofano, F.; Borrè, A.; Zenga, F.; Melcarne, A.; et al. How reliable are volumetric techniques for high-grade gliomas? A comparison study of different available tools. Neurosurgery 2020, 87, E672–E679. [Google Scholar] [CrossRef]
  77. Ellingson, B.M.; Bendszus, M.; Boxerman, J.; Barboriak, D.; Erickson, B.J.; Smits, M.; Nelson, S.J.; Gerstner, E.; Alexander, B.; Goldmacher, G.; et al. Consensus recommendations for a standardized Brain Tumor Imaging Protocol in clinical trials. Neuro-Oncol. 2015, 17, 1188–1198. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Comparison between automated and manual segmentations for the laterality index in (A) a correlation plot with histograms in the margin and (B) a Bland–Altman plot. In the scatterplots, each dot represents the laterality indices of one patient. The diagonal indicates the identity line. The Bland–Altman plots of the mean of laterality indices versus the difference between the laterality indices. Each dot represents one patient. The bias is plotted as solid green line with 95% CI as dotted green lines. The limits of agreement are plotted as dotted red lines.
Figure 1. Comparison between automated and manual segmentations for the laterality index in (A) a correlation plot with histograms in the margin and (B) a Bland–Altman plot. In the scatterplots, each dot represents the laterality indices of one patient. The diagonal indicates the identity line. The Bland–Altman plots of the mean of laterality indices versus the difference between the laterality indices. Each dot represents one patient. The bias is plotted as solid green line with 95% CI as dotted green lines. The limits of agreement are plotted as dotted red lines.
Cancers 13 02854 g001
Figure 2. Comparison between manual and automated segmentations: (A) a histogram of absolute difference for the manual segmentations and (B) for the automated segmentations, (C) a correlation plot of the native tumor volumes with histograms in the margin, (D) a Bland–Altman plot for the native tumor volumes, (E) a correlation plot of the normalized tumor volumes with histograms in the margin, and (F) a Bland–Altman plot for the normalized tumor volumes. Each dot represents the volumes of one patient. The dotted diagonal in (C,E) indicates the identity line. The bias is plotted as solid green line with 95% CI as dotted green lines and the limits of agreement as dotted red lines in (D,F).
Figure 2. Comparison between manual and automated segmentations: (A) a histogram of absolute difference for the manual segmentations and (B) for the automated segmentations, (C) a correlation plot of the native tumor volumes with histograms in the margin, (D) a Bland–Altman plot for the native tumor volumes, (E) a correlation plot of the normalized tumor volumes with histograms in the margin, and (F) a Bland–Altman plot for the normalized tumor volumes. Each dot represents the volumes of one patient. The dotted diagonal in (C,E) indicates the identity line. The bias is plotted as solid green line with 95% CI as dotted green lines and the limits of agreement as dotted red lines in (D,F).
Cancers 13 02854 g002
Figure 3. Comparison of tumor location profiles of cortical structures between (A) automated and (B) manual segmentations in Desikan’s brain parcellation. In the circular bar plots, each bar represents one parcel from the Desikan’s brain parcellation categorized by lobe. Abbreviations are referring to anatomical parcels as detailed in the legend. The height of a bar represents the percentage of patients, indicated in grey, with tumor involvement in a parcel. The width of a bar corresponds with the relative volume of a parcel. (C) Correlation plot between the number of patients with parcel involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Figure 3. Comparison of tumor location profiles of cortical structures between (A) automated and (B) manual segmentations in Desikan’s brain parcellation. In the circular bar plots, each bar represents one parcel from the Desikan’s brain parcellation categorized by lobe. Abbreviations are referring to anatomical parcels as detailed in the legend. The height of a bar represents the percentage of patients, indicated in grey, with tumor involvement in a parcel. The width of a bar corresponds with the relative volume of a parcel. (C) Correlation plot between the number of patients with parcel involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Cancers 13 02854 g003
Figure 4. Comparison of tumor location profiles of cortical structures between (A) automated and (B) manual segmentations in Schaefer’s brain parcellation. In the circular bar plots, each bar represents one network class parcel. Abbreviations refer to classes as detailed in legend. The height of a bar represents the percentage of patients, indicated in grey, with tumor involvement in a class parcel. The width of a bar corresponds with the relative volume of a parcel. (C) Correlation plot between the number of patients with parcel involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Figure 4. Comparison of tumor location profiles of cortical structures between (A) automated and (B) manual segmentations in Schaefer’s brain parcellation. In the circular bar plots, each bar represents one network class parcel. Abbreviations refer to classes as detailed in legend. The height of a bar represents the percentage of patients, indicated in grey, with tumor involvement in a class parcel. The width of a bar corresponds with the relative volume of a parcel. (C) Correlation plot between the number of patients with parcel involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Cancers 13 02854 g004
Figure 5. Comparison of tumor location profiles of subcortical white matter structures between (A) automated and (B) manual segmentations. In the circular bar plots, each bar represents one tract or tract segment. Abbreviations refer to structures as detailed in legend. The height of a bar represents the percentage of patients with tumor involvement in a structure indicated in grey. The width of a bar corresponds with the relative volume of a structure. (C) Correlation plot between the number of patients with structure involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Figure 5. Comparison of tumor location profiles of subcortical white matter structures between (A) automated and (B) manual segmentations. In the circular bar plots, each bar represents one tract or tract segment. Abbreviations refer to structures as detailed in legend. The height of a bar represents the percentage of patients with tumor involvement in a structure indicated in grey. The width of a bar corresponds with the relative volume of a structure. (C) Correlation plot between the number of patients with structure involvement between the manual and automated segmentations. The dotted diagonal indicates the identity line.
Cancers 13 02854 g005
Figure 6. Comparison between automated and manual segmentations for (A,B) expected residual tumor volumes and (C,D) expected resectability index. In the scatterplots (A,C), each dot represents the data of one patient. The diagonal indicates the identity line. The boxplots display the distributions with median, 25% and 75% quartiles as hinges and 1.5 times the interquartile distance as whiskers. The Bland–Altman plots (B,D) of the mean of expectations versus the difference between expectations. Each dot represents one patient. The bias is plotted as solid line with 95% CI as dotted lines. The limits of agreement are plotted as dashed lines.
Figure 6. Comparison between automated and manual segmentations for (A,B) expected residual tumor volumes and (C,D) expected resectability index. In the scatterplots (A,C), each dot represents the data of one patient. The diagonal indicates the identity line. The boxplots display the distributions with median, 25% and 75% quartiles as hinges and 1.5 times the interquartile distance as whiskers. The Bland–Altman plots (B,D) of the mean of expectations versus the difference between expectations. Each dot represents one patient. The bias is plotted as solid line with 95% CI as dotted lines. The limits of agreement are plotted as dashed lines.
Cancers 13 02854 g006
Figure 7. Tumor probability maps for the automated and manual segmentations. Each voxel represents the tumor incidence in the study population with false discovery rates of the difference between the incidences, as specified in the legend.
Figure 7. Tumor probability maps for the automated and manual segmentations. Each voxel represents the tumor incidence in the study population with false discovery rates of the difference between the incidences, as specified in the legend.
Cancers 13 02854 g007
Figure 8. Examples of disagreement between manual and automated segmentation. (A) The automated and manual segmentation both included the tumor cyst with enhancing rim, but the automated segmentation did not include the cyst without enhancing rim nor some of the solid tumor extensions, as examples of false negative cyst detection. (B) The automated segmentation did not include the satellite lesion, as example of false negative enhancing tumor component. (C) The manual segmentation included tissue as tumor portions, whereas the automated segmentation did not, as example of either false positive inclusion of nonenhancing tissue by the manual segmentation or false negative exclusion of nonenhancing tumor exclusion by the automated segmentation. (D) Conversely, the manual segmentation excluded tissue from the tumor compartment, whereas the automated segmentation included this tissue, as example of either false negative exclusion of nonenhancing tissue by the manual segmentation or false positive inclusion of nonenhancing tumor exclusion by the automated segmentation. (E) The manual segmentation included a vascular structure, as example of a false positive vasculature structure. (F) Conversely, the automated segmentation included a vascular structure indicated by the yellow arrow, as another example of a false positive finding.
Figure 8. Examples of disagreement between manual and automated segmentation. (A) The automated and manual segmentation both included the tumor cyst with enhancing rim, but the automated segmentation did not include the cyst without enhancing rim nor some of the solid tumor extensions, as examples of false negative cyst detection. (B) The automated segmentation did not include the satellite lesion, as example of false negative enhancing tumor component. (C) The manual segmentation included tissue as tumor portions, whereas the automated segmentation did not, as example of either false positive inclusion of nonenhancing tissue by the manual segmentation or false negative exclusion of nonenhancing tumor exclusion by the automated segmentation. (D) Conversely, the manual segmentation excluded tissue from the tumor compartment, whereas the automated segmentation included this tissue, as example of either false negative exclusion of nonenhancing tissue by the manual segmentation or false positive inclusion of nonenhancing tumor exclusion by the automated segmentation. (E) The manual segmentation included a vascular structure, as example of a false positive vasculature structure. (F) Conversely, the automated segmentation included a vascular structure indicated by the yellow arrow, as another example of a false positive finding.
Cancers 13 02854 g008
Figure 9. Illustration of the GSI-RADS software and standard report. At the left, the standard report is displayed in text format. At the top right, the patient MRI scan and the patient MRI scan with overlayed automated tumor segmentation are displayed, and at the bottom right, the standard brain space and the registered patient MRI scan with overlayed automated tumor segmentation in standard brain space are demonstrated.
Figure 9. Illustration of the GSI-RADS software and standard report. At the left, the standard report is displayed in text format. At the top right, the patient MRI scan and the patient MRI scan with overlayed automated tumor segmentation are displayed, and at the bottom right, the standard brain space and the registered patient MRI scan with overlayed automated tumor segmentation in standard brain space are demonstrated.
Cancers 13 02854 g009
Table 1. Patient characteristics.
Table 1. Patient characteristics.
HospitalNWZSLZISALAPARISHUMMUWUMCG
n38497274758386
females, n (%)13 (34.2%)25 (51.0%)9 (12.5%)33 (44.6%)29 (38.7%)36 (43.4%)31 (36.0%)
median age in years (interquartile range)63.4 (17.4)63.6 (14.2)67.2 (20.7)59.0 (13.5)62.7 (16.3)67.3 (19.7)62.8 (12.4)
HospitalVUmcHMCUCSFETZUMCUSTOoverall
n971031341531714611596
females, n (%)35 (36.1%)38 (36.9%)49 (36.6%)50 (32.7%)63 (36.8%)189 (41.0%)600 (37.6%)
median age in years (interquartile range)64.0 (16.2)61.1 (18.1)64.2 (14.8)63.8 (12.2)66.2 (16.4)61.7 (14.4)63.2 (15.7)
Table 2. Contingency table of laterality between automated and manual segmentations.
Table 2. Contingency table of laterality between automated and manual segmentations.
Laterality by Automated Segmentation
Laterality by Manual Segmentationleftrightnonesubtotal
left782210794
right37897799
none0123
subtotal78579219
Table 3. Contingency table of contralateral infiltration between automated and manual segmentations.
Table 3. Contingency table of contralateral infiltration between automated and manual segmentations.
Contralateral Infiltration by Automated Segmentation
Contralateral Infiltration by Manual Segmentationnoyessubtotal
no1110171127
yes56413469
subtotal11664301596
Table 4. Contingency table of multifocality between automated and manual segmentations.
Table 4. Contingency table of multifocality between automated and manual segmentations.
Multifocality by Automated Segmentation
Multifocality by Manual Segmentationnoyessubtotal
no1165571222
yes111263374
subtotal12763201596
Table 5. Contingency table of multifocality between automated and manual segmentations.
Table 5. Contingency table of multifocality between automated and manual segmentations.
Number of Foci by Automated Segmentation
Number of Foci by Manual Segmentation012345subtotal
02100003
1131149525001219
24861481610255
301932362089
401873120
50122207
60000101
70000011
110000011
subtotal19125724266931596
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kommers, I.; Bouget, D.; Pedersen, A.; Eijgelaar, R.S.; Ardon, H.; Barkhof, F.; Bello, L.; Berger, M.S.; Conti Nibali, M.; Furtner, J.; et al. Glioblastoma Surgery Imaging—Reporting and Data System: Standardized Reporting of Tumor Volume, Location, and Resectability Based on Automated Segmentations. Cancers 2021, 13, 2854. https://doi.org/10.3390/cancers13122854

AMA Style

Kommers I, Bouget D, Pedersen A, Eijgelaar RS, Ardon H, Barkhof F, Bello L, Berger MS, Conti Nibali M, Furtner J, et al. Glioblastoma Surgery Imaging—Reporting and Data System: Standardized Reporting of Tumor Volume, Location, and Resectability Based on Automated Segmentations. Cancers. 2021; 13(12):2854. https://doi.org/10.3390/cancers13122854

Chicago/Turabian Style

Kommers, Ivar, David Bouget, André Pedersen, Roelant S. Eijgelaar, Hilko Ardon, Frederik Barkhof, Lorenzo Bello, Mitchel S. Berger, Marco Conti Nibali, Julia Furtner, and et al. 2021. "Glioblastoma Surgery Imaging—Reporting and Data System: Standardized Reporting of Tumor Volume, Location, and Resectability Based on Automated Segmentations" Cancers 13, no. 12: 2854. https://doi.org/10.3390/cancers13122854

APA Style

Kommers, I., Bouget, D., Pedersen, A., Eijgelaar, R. S., Ardon, H., Barkhof, F., Bello, L., Berger, M. S., Conti Nibali, M., Furtner, J., Fyllingen, E. H., Hervey-Jumper, S., Idema, A. J. S., Kiesel, B., Kloet, A., Mandonnet, E., Müller, D. M. J., Robe, P. A., Rossi, M., ... De Witt Hamer, P. C. (2021). Glioblastoma Surgery Imaging—Reporting and Data System: Standardized Reporting of Tumor Volume, Location, and Resectability Based on Automated Segmentations. Cancers, 13(12), 2854. https://doi.org/10.3390/cancers13122854

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop