Next Article in Journal
Concurrent Validity of Four Activity Monitors in Older Adults
Next Article in Special Issue
Cognitive Effort during Visuospatial Problem Solving in Physical Real World, on Computer Screen, and in Virtual Reality
Previous Article in Journal
Intrarow Uncut Weed Detection Using You-Only-Look-Once Instance Segmentation for Orchard Plantations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Feasibility and Accuracy of Holographic Navigation with Laser Crosshair Simulator Registration on a Mixed-Reality Display

1
Department of Neurosurgery, University of Marburg, Baldingerstrasse, 35043 Marburg, Germany
2
Department of Neurosurgery, First Medical Center of Chinese PLA General Hospital, Beijing 100853, China
3
Medical School of Chinese PLA, Beijing 100853, China
4
NCO School, Army Medical University, Shijiazhuang 050081, China
5
Department of Neurosurgery, Division of Medicine, Graduate School of Medical Sciences, Kanazawa University, Takara-machi 13-1, Kanazawa 920-8641, Japan
6
Center for Mind, Brain and Behavior (CMBB), 35043 Marburg, Germany
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(3), 896; https://doi.org/10.3390/s24030896
Submission received: 4 January 2024 / Revised: 21 January 2024 / Accepted: 23 January 2024 / Published: 30 January 2024

Abstract

:
Addressing conventional neurosurgical navigation systems’ high costs and complexity, this study explores the feasibility and accuracy of a simplified, cost-effective mixed reality navigation (MRN) system based on a laser crosshair simulator (LCS). A new automatic registration method was developed, featuring coplanar laser emitters and a recognizable target pattern. The workflow was integrated into Microsoft’s HoloLens-2 for practical application. The study assessed the system’s precision by utilizing life-sized 3D-printed head phantoms based on computed tomography (CT) or magnetic resonance imaging (MRI) data from 19 patients (female/male: 7/12, average age: 54.4 ± 18.5 years) with intracranial lesions. Six to seven CT/MRI-visible scalp markers were used as reference points per case. The LCS-MRN’s accuracy was evaluated through landmark-based and lesion-based analyses, using metrics such as target registration error (TRE) and Dice similarity coefficient (DSC). The system demonstrated immersive capabilities for observing intracranial structures across all cases. Analysis of 124 landmarks showed a TRE of 3.0 ± 0.5 mm, consistent across various surgical positions. The DSC of 0.83 ± 0.12 correlated significantly with lesion volume (Spearman rho = 0.813, p < 0.001). Therefore, the LCS-MRN system is a viable tool for neurosurgical planning, highlighting its low user dependency, cost-efficiency, and accuracy, with prospects for future clinical application enhancements.

1. Introduction

In neurosurgery, commercial navigation systems have emerged as a standard paradigm, instrumental in aligning medical images with the physical patient space [1,2,3]. They enable neurosurgeons to accurately track surgical instruments and maximize the extent of resection while safeguarding critical anatomical structures during interventions [4,5,6]. These systems have employed landmark-based and surface-based methods alongside automatic registration methods, each with unique operational principles [1,2,4,5,6] (see Figure 1A,B). While automatic methods are lauded for their low user dependency input and reduced registration error, they also show some drawbacks. These especially include their substantial hardware requirements and the necessity of specialized operation rooms, factors that significantly elevate the cost and limit their widespread adoption [7,8].
In response to these challenges, there has been a shift towards exploring cost-effective alternatives, particularly in augmented reality (AR) and mixed reality (MR) technologies [9,10,11,12,13,14,15,16,17,18,19,20]. These technologies, particularly when implemented through a head-mounted display (HMD), offer an innovative approach for navigation. They provide an immersive, interactive experience, allowing for intuitive preoperative planning and surgical guidance, a concept called mixed reality navigation (MRN) [7,8,11,12,21,22,23]. Unlike conventional systems, MRN incorporates user-friendly interfaces and leverages emerging technologies to enhance accuracy and adaptability in surgical procedures [12,21,22,24,25,26].
The development of the laser crosshair simulator (LCS) within the MRN framework marks a significant advancement [7]. The LCS-MRN system, mirroring the automatic registration paradigm of conventional systems, aims to (1) reproduce the relationship between the reference scan and the physical patient, (2) establish an initial virtual-to-physical space registration, and (3) maintain this alignment through continuous tracking (see Figure 1C). The LCS-MRN system offers simplicity in hardware and software configurations, addressing the complexities of handling virtual objects.
Despite these advantages, two critical aspects arise regarding the LCS-MRN system’s reliability and clinical applicability. The first involves identifying a universal and reliable methodology to validate the accuracy of this novel system. The second aspect pertains to the system’s versatility and ability to meet the diverse and extensive demands of various surgical procedures and conditions. While various parameters have been studied within the MRN context, specific metrics to evaluate the LCS-MRN system remain unreported. Furthermore, its application in neurosurgery has not been extensively explored.
Therefore, this study retrospectively utilized imaging data from patients with intracranial lesions to create life-sized head phantoms for evaluation. These phantoms, employed in simulated clinical scenarios, facilitated the validation of the LCS-MRN system’s feasibility and accuracy. This approach attempts to address the gap in the literature by providing a foundational assessment of this innovative system in a controlled yet clinically relevant setting. Further, it presents an advancement in MRN registration techniques through a more streamlined and user-friendly hardware and software configuration. This enhanced approach significantly improves accuracy and flexibility and contributes to developing cost-effective MRN systems in neurosurgery. The potential for improved surgical outcomes, especially in environments without standard navigation systems, marks a significant advancement in neurosurgical technology.

2. Materials and Methods

This study aims to develop and validate a low-cost, easy-to-use LCS-MRN system for neurosurgical planning and navigation. It addresses the dual objectives of creating this innovative system and establishing a universal, reliable evaluation methodology for its accuracy and applicability in diverse neurosurgical contexts. This section provides a detailed overview of the foundational concepts and structure of the LCS-MRN system, followed by an exploration of the techniques and methods used for image processing and phantom creation. The section concludes by detailing the metrics used to evaluate the quality of the LCS-MRN system. The schematic of the practical workflow is depicted in Figure 2.

2.1. Development of the LCS-MRN System

2.1.1. The Laser Crosshair Simulator (LCS)

The LCS, as depicted in Figure 3, was designed to address the challenges of scanner tracking in MRN. Conventional automatic registration systems often rely on a reference array attached to the scanner, aligning images with a universal coordinate framework such as the world coordinate system [7]. However, the LCS introduces a novel approach by employing dual laser emitters and a specialized MR interface to overcome disruptions in optical tracking caused by user movement or scanner activities in HMDs (see Figure 3B). These lasers project intersecting lines, simulating the crosshairs found in computed tomography (CT) or magnetic resonance imaging (MRI) scans, facilitating the precise transfer of scanning details for accurate image localization (see Figure 3A). The MR interface features a stainless-steel panel with a target image on both sides. When recognized by the HMD’s Vuforia software development kit (SDK) (Version 10.14, PTC, Inc., Boston, MA, USA), the image’s central point establishes the reference origin in virtual space, ensuring precise hologram positioning and enhancing spatial mapping accuracy in MRN systems (see Figure 3C).

2.1.2. The Mixed Reality Navigation (MRN)

The MRN system in this study utilizes the Microsoft HoloLens-2 (HL-2) (Microsoft, Redmond, WA, USA) as its primary hardware. The HL-2, a self-contained, portable, optical lightweight HMD, operates independently without external hosts or tracking systems. Equipped with a high-definition red–green–blue (RGB) camera, four Visible Light Cameras (VLC), a depth camera, and an Inertial Measurement Unit (IMU), the HL-2 excels in mapping its position and orientation through simultaneous localization and mapping (SLAM). It projects virtual objects onto a specific plane using waveguide imaging technology, offering a 43° × 29° field of view (FoV) color display, which ensures stable, realistic virtual representations regardless of the viewer’s perspective.
The software component of the MRN system is developed in Unity (Version 2021.3.4f1, Unity, San Francisco, CA, USA), integrated with image detection and tracking algorithms from the Vuforia SDK (Version 10.14, PTC, Inc., Boston, MA, USA). The interface, designed using the Mixed Reality Toolkit (MRTK) SDK (Version 2.8.3, Microsoft), employs C# scripts in Visual Studio (Version 16.11.26, Microsoft, 2019) for program logic and user interaction through voice and gesture commands. This setup facilitates an immersive experience, with the final application packaged as a Universal Windows Platform (UWP) app tailored for deployment on the HL-2. The combination of these advanced software elements enhances the overall functionality and user experience of the MRN system.

2.2. Image Data Acquisition and Image Processing

2.2.1. Subjects

The study retrospectively collected preoperative cranial MRI or CT data from 19 patients over four years (2018–2021) in two clinical hospitals (the first medical center of the Chinese PLA General Hospital, Beijing, and the Hainan Hospital of the Chinese PLA General Hospital, Sanya) who were diagnosed with intracranial lesions (neurological neoplasm, hypertensive cerebral hemorrhage) and were candidates for surgical intervention. Prior to study inclusion, written informed consent was provided by each patient or their legal representative, permitting the use of pseudonymized image data for this study. Given the study’s retrospective nature and the absence of invasive procedures, an ethical review was deemed unnecessary. The 19 enrolled patients met the following minimal inclusion criteria:
  • High-resolution and high-quality MRI or CT data with clear skin contours covering the cranial region;
  • Clear lesion boundaries visible in at least one image data set;
  • At least five skin adhesive markers were attached to the patient’s scalp before imaging.
In this study, all patients underwent surgery guided by conventional navigation systems, where the markers placed near the surgical area were used for traditional navigation registration [8,23]. The successful completion of these surgeries without significant complications validated the effectiveness of the neuroimaging and skin markers used for traditional navigation. This outcome is crucial, as it establishes the reliability of the preoperative data and marker placements, which are also essential for evaluating the quality of the MRN system. By confirming that these elements were effective in a conventional setting, one can reasonably infer their utility in evaluating the MRN system. Thus, the surgical procedures and their outcomes, although not directly involving MRN, provide a foundational basis for the MRN study, ensuring that the retrospective data used is accurate and applicable for this novel navigation system evaluation.

2.2.2. Image Acquisition

In the case of intracranial neoplasms, high-resolution image acquisition was conducted using a 1.5 T MRI scanner (Espree, Siemens, Erlangen, Germany), whereas, in the case of cerebral hemorrhage, image acquisition was performed using a 128 multislice CT scanner (SOMATOM, Siemens, Forchheim, Germany). All image data were acquired one day before or on the day of the surgery, serving as the reference image for reconstructing the medical imaging space’s geometric shape, precisely the scanning parameters the LCS-MRN system aimed to replicate and transfer in this study. They were thus termed “reference scans”, with the resulting images called “reference images”. In addition to the reference images, all patients with intracranial neoplasms underwent multimodal MRI acquisition days before surgery, including, depending on the location of the lesion, a T1-weighted contrast-enhanced (T1-CE), T2-weighted (T2) and/or Diffusion tensor imaging (DTI)) in clinical routine. Imaging data were provided in Digital Imaging and Communications in Medicine (DICOM) file format.

2.2.3. Creation of Head Phantoms

Life-sized head phantoms for all 19 subjects were 3D printed based on the individual MRI or CT-based reference image to evaluate the practicality of the suggested registration technique. Firstly, standard tessellation language (STL) files, a widely used 3D printing and modeling format, were generated from segmented skin surfaces in reference imaging using the “Segment Editor” extension within 3D Slicer software (Version 5.1.0) for 3D reconstruction. The process also involved model optimization, such as hollowing the head, smoothing the surface to eliminate noise, ensuring a flat bottom surface for stable printing, and refining sharp edges for the researcher’s safety to ensure high-quality, cost-effective, and safe 3D phantom creation. Utilizing the basic planes from the reference image, the laser crosshair mark lines (width: 1 mm) were reconstructed and incorporated into the design of the phantom (see Figure 4B). The replica was exported as an “.stl” file and manufactured via an A5S 3D printer (Shenzhen Aurora Technology Co., Ltd., Shenzhen, China) under the following conditions: extrusion temperature 210°C, bed temperature 50 °C, polylactic acid as print medium, layer height 0.3 mm, and infill density 10%, thereby producing a life-sized head phantom for assessment.

2.2.4. Image Segmentation

To rigorously evaluate the MRN system’s accuracy and reliability, a set of outlined objects was generated to create holograms using the 3D Slicer platform for this investigation (see Figure 4A). For each subject, intracranial lesions (neurological neoplasm or hematoma) were segmented following a strict protocol to evaluate the MRN system’s precision in lesion localization. Additionally, fiducial markers were segmented, and their centroids were marked to elaborate a series of accuracy metrics quantitatively (as indicated in Section 2.4.2). Furthermore, for each case, critical anatomical structures (such as ventricles, arteries, venous sinuses, pyramidal tracts, optic radiations, frontal sinuses, and scalp quadrants) upon availability of according data sets, as well as models for surgical guidance (such as puncture paths for reaching hemorrhage, endoscopic pathways, and the bone flaps suitable for port surgery with introducer assistance), were individually segmented [7,8,23,27], allowing to simulate the visualization environment in clinical settings closely and to evaluate the MRN system’s computational load capacity for rendering multiple models. It should be noted, however, that due to individual variations and different surgical plans, these structures were not used for accuracy assessment.
Markers and scalp quadrants were segmented within the T1-weighted data set. The segmentation of intracranial lesions, arteries, venous sinuses, and ventricles was carried out based on the T1-CE data. In contrast, the identification and segmentation of the optic radiations and pyramidal tracts were based on the DTI data. The latter utilized a specialized and distinctly different approach from conventional image segmentation, known as tractography [28,29,30].
Finally, co-registration was performed across all image data sets to transform all segmented structures from multiple modalities into a unified reference image coordinate system (RICS). Therefore, the reference images (i.e., the highest-resolution T1-weighted scan of the cranium with attached markers or the CT data set, respectively) were set as fixed images while all other data sets (T1-CE, T2, and DTI) were rigidly co-registered to this reference image. The registration process was facilitated through the “General Registration (Elastix)” extension available on the 3D Slicer platform, with the preset set to “Generic rigid (all)”. The calculated registration matrix was saved in the 3D Slicer scene files, enabling the transformation of various segments into a unified coordinate system.

2.2.5. Holograms Generation

Through the “Segmentations” extension on the 3D Slicer platform, pixels segmented on each two-dimensional (2D) slice were aligned, stacked, and voxelized in three-dimensional space to construct continuous 3D surfaces for holographic visualization (3D holograms). This process, known as 3D reconstruction, liberates the user’s observation of holograms from the confines of the original imaging planes. All models were exported in the OBJ file format.

2.3. LCS-MRN Registration Workflow

2.3.1. LCS Deployment

The generated head models and holograms were used for further analyses (see Figure 4A,B). Phantoms were securely affixed using self-tapping screws from their base to prevent any unknown deformations during fixation that might affect the accuracy of measurements (see Figure 4C) as the patient’s head in a clinical setup would be fixed with a head clamp too during navigation-supported surgery. Once the head phantom was securely positioned, the user turned on the laser emitters and fine-tuned the placement of the LCS (see Figure 4D). This adjustment continued until the LCS’s laser lines were precisely aligned with the marked lines on the phantom. Subsequently, the LCS’s orientation was fixed to preserve its alignment relative to the phantom.

2.3.2. Surgical Position Compatibility

The study developed a matrix transform approach encapsulated within the software to facilitate LCS deployment and practicality in varying surgical positions. Since reference imaging is usually conducted in the supine position, differing from potential prone or lateral surgical positions, using the original coordinates (i.e., the RICS) directly could lead to inconvenient LCS placement, possibly conflicting with surgical instruments or anesthesia equipment. For example, in a prone position, direct use of original coordinates might place the LCS in conflict with the patient’s head fixation system. Similarly, in a lateral position, the LCS could potentially interfere with the anesthesia tubing. The authors’ software-based matrix computation artificially adjusts the coordinate system—for instance, rotating the prone position around the vertical axis by 180° and the lateral position by ±90°. This allows for LCS placement that does not interfere with surgical operations while maintaining patient positioning and smooth medical procedures.
To ensure compatibility of the holograms with the LCS-MRN system, in addition to adapting for different surgical positions as mentioned earlier, it is also essential to address the disparities in scale between the Reference Image Coordinate System (RICS) and the Virtual Coordinate System (VCS) defined by the Vuforia SDK. For example, while RICS is usually defined in millimeters during imaging scans, the Vuforia SDK operates in centimeters, necessitating careful consideration of unit conversion and corresponding numerical transformations. Considering the differences in coordinate axis directions and coordinate scales, the Forward Engineering Matrix ( FEM ) is implemented to achieve compatibility with the Vuforia SDK. This step integrates both the direction and scale disparities to precisely position, orient, and dimension the virtual objects within the LCS-MRN system, ensuring accurate holographic visualization in varied surgical positions and setups (see Figure 5A–L). In other words, FEM harmonizes the physical and virtual environments for effective LCS deployment. Conversely, when analyzing holograms from the MRN system within 3D Slicer, involving the analysis coordinate system (ACS), a reciprocal operation equivalent to reversing the linear transformation, known as the Reverse Engineering Matrix ( REM ), becomes imperative. These transformations are contingent upon the patient’s head position during the use of the LCS, with various positions such as supine, prone, left-lateral, and right-lateral necessitating distinct FEM s and REM s. Notably, since RICS and ACS are right-handed coordinate systems while VCS is a left-handed system, a handedness conversion matrix ( HCM ) is required to invert the direction of the z-axis, a step encapsulated within the software; hence, in practice, the FEM is the transformation from RICS to an intermediate VCS (IVCS).
Before being integrated into the MR platform, all models were adapted using the FEM -based process via the “Transform” extension in the 3D Slicer software.

2.3.3. MRN Setup

The HL-2 was carefully configured to each user’s interpupillary distance (IPD), ensuring a custom-fit visual experience to initialize the holographic visualization. The device must be securely positioned on the user to prevent accidental shifts. Utilizing Vuforia SDK’s feature detection algorithm, the MRN system meticulously tracks the designated image markers within the LCS, thus accurately maintaining the real-time positioning of the LCS. This precise tracking facilitates the establishment of the RICS defined by the LCS’s orientation and position. Subsequently, the user employs gesture controls to introduce the holographic model into the RICS. The holograms were visualized on the static head phantom, signifying the completion of the MRN’s initial registration phase (see Figure 4E and Figure 6A).

2.3.4. Registration Update

Subsequently, users can maintain or modify the current registration. This is controlled by switching between the “Freeze” and “Unfreeze” modes on the MR platform. Activating “Freeze” secures the holographic display to its spatial anchor, achieved through the HL-2’s SLAM process, thus ensuring stability of the holographic image even during temporary tracking loss of the image target. Conversely, selecting “Unfreeze” resumes the tracking process, allowing for updates to the hologram to reflect any shifts in the patient’s head position relative to the global coordinate system.

2.4. Quality Assessments

2.4.1. Practicality Assessment for the Workflow

The workflow and the visualized scenes of the LCS-MRN for each case were retrospectively analyzed to assess the system’s practicality in a simulated clinical setting. This evaluation was facilitated through the HL-2’s ’Mixed Reality Capture (MRC)’ feature, which allowed for capturing the merged virtual and physical worlds into photo or video files using the built-in Photos-Videos (PV) camera. To mitigate any potential impact on the device’s performance and spatial positioning accuracy, MRC was primarily conducted via photography. The recorded images were then systematically reviewed to evaluate the ease of use, accuracy of virtual object placement, and overall user interaction within the MR environment. These factors are crucial in determining the practicality of the LCS-MRN system in real-world clinical scenarios. Multiple methods were available to activate the camera, including local control through gaze, hand gestures, and voice commands or remote operation via a web browser on a connected computer or smartphone, offering flexibility in documentation.

2.4.2. Assessment of Landmarks

The fiducial markers attached to the scalp, not included in the registration process, were selected as reference points for accurate measurement. In this study, the LCS-MRN system’s custom-developed “virtual probe” feature was utilized to acquire coordinates of the virtual (fiducial holograms) and the physical representation (3D-printed markers integrated with the head model) of these markers. The virtual probe, designed with a white line handle and a spherical tip, is a precise tool within the virtual space, enabling the user to position it accurately at spatial points of interest. As the user positions the probe tip at the chosen spatial locations, the LCS-MRN system relays the tip coordinates within the RICS on its virtual panel. It is imperative to note, as outlined in Section 2.2.5, that the coordinates collected for each case must transform the ( REM ) corresponding to the surgical position. This step is essential before proceeding to further analysis in 3D Slicer; failure to do so could lead to significant errors. The coordinates of both the virtual and physical fiducials, along with their centroid coordinates analyzed using 3D Slicer, are fundamental for calculating the system’s interpolative and extrapolative accuracy indices.
The marker’s centroids were considered the ground truth in the RICS with right–anterior–superior (RAS) coordinates (see Figure 6C), a standard medical coordinate system where the primary axes of right, anterior, and superior are considered positive directions. These centroids can be extracted using the “Segment statistics” module within 3D Slicer. For computational convenience, unique indices are established for each marker, allowing each case to have an ordered set of fiducial markers, C , facilitating rapid access and manipulation of specific points. Without loss of generality, the centroid coordinates of the i-th fiducial marker, C i , are denoted as:
C i = ( x C i , y C i , z C i )
For each marker, relying on stereoscopic vision, the user delicately maneuvered the virtual probe’s tip and sequentially placed it on each perceived virtual marker point, V i (see Figure 4E and Figure 6B). The MR platform instantaneously reported the converted coordinates of this point within the RICS in a panel as:
V i = ( x V i , y V i , z V i )
Similarly to the point set C , all virtual marker points compose an ordered set, V , based on the indices of the markers. The presented coordinates in the panel can be exported for subsequent analysis.
Following this, the users cleared all probes and repositioned the virtual probe’s tip sequentially on each perceived physical marker point, P i (see Figure 4G and Figure 6B). The MR platform also reported and exported P i ’s coordinates within the RICS, formulating another ordered set, P , expressed as:
P i = ( x P i , y P i , z P i )

2.4.3. Landmark-Based Evaluation of Accuracy

To objectively quantify the discrepancy between virtual images and the actual physical environment in the LCS-MRN system, the target registration error ( TRE ) was chosen as the measure of accuracy. The TRE is calculated based on the Euclidean distance between specific reference points in virtual space and their corresponding points in physical space. However, considering potential manual errors that might occur during the use of the virtual probe, the fiducial localization error ( FLE ) was additionally utilized as an indicator of the reliability of the assessment process. Furthermore, the fiducial registration error ( FRE ) and the Frobenius norm ( FN ) were analyzed as extrapolative indicators of overall head displacement, as the anatomical accuracy of the patient’s head itself, compared to the fiducials.
The FLE denotes the displacement vector from a marker’s centroid (ground truth) to the perceived virtual counterpart’s center, reflecting the perceptual error as the user moves and positions the virtual probe’s tip. A minimal FLE indicates a trustworthy spatial point measurement by the user. FLE at the i-th fiducial marker can thereby be computed as:
FLE i = V i C i
Its magnitude can be computed as:
FLE i 2 = ( x V i x C i ) 2 + ( y V i y C i ) 2 + ( z V i z C i ) 2
The term · 2 denotes the Euclidean norm of the displacement vector.
The TRE represents the displacement vector post-registration from a marker’s centroid (ground truth) to the perceived physical counterpart’s center, signifying the virtual-to-physical alignment accuracy. It acts as a validity metric for the data set. A lower magnitude of TRE insinuates accurate virtual–physical alignment by the MRN system. It can be formulated as follows:
TRE i = P i C i
Its magnitude can be computed as:
TRE i 2 = ( x P i x C i ) 2 + ( y P i y C i ) 2 + ( z P i z C i ) 2
The FRE describes the displacement vector between two registered ordered point sets. Specifically, when the markers’ centroid (as the ground truth, GT) set C is transformed to align the perceived physical point sets P using an algorithm (e.g., least squares), FRE measures the displacement vector between each point in P and its counterpart point in the transformed set P (see Figure 4H and Figure 6D–F) and gauges the congruence between the two sets of points. A reduced FRE indicates a high level of geometric shape consistency between the virtual and physical point sets, thereby suggesting the reliability of extrapolative analysis based on reference landmarks. The FRE at the i-th marker can be formulated as:
FRE i = P i P i
Given that the point sets C and P contain an equal number of points with corresponding indices, the optimal rigid linear transformation mapping from C to P is appropriately computed using a least squares algorithm. Calculation of the optimal transformation can be automatically performed in the “Image-guided therapy (IGT)” module of the 3D Slicer platform in the “Fiducial Registration Wizard” extension. The core algorithm of this extension, an overwriting of the “vtkLandmarkTransform()” function in the Visualization Toolkit (VTK) library, accomplishes the task of fitting the optimal rigid linear transformation between two ordered point sets, subsequently returning the optimal transformation matrix T C P R 4 × 4 to the user (see Figure 4I and Figure 6F). T C P can be represented as:
T C P = R C P t C P 0 1
The term R C P R 3 × 3 is the rotation matrix, t C P R 3 × 1 is the translation matrix, and 0 R 1 × 3 represents no affine transformation on the original points. One point P i in the post-transformation point set P can be expressed as:
P i = R C P C i + t C P
Based on the definition of FRE , the following relationship is derived:
FRE i = P i P i = R C P C i P i + t C P
Incorporating the coordinates P i and C i into Equation (11) facilitates the calculation of the components of FRE i along the three principal axes in the RICS and its Euclidean norm FRE i 2 .
The FN offers a method for quantifying the “size” of a displacement matrix, analogous to the Euclidean vector norm but applicable to matrices. In the context of precision assessment, the FN can be used to measure the discrepancy between matrix T C P and the identity matrix I R 4 × 4 across all elements. A larger FN of the difference between T C P and I indicates a greater divergence between T C P and I , signifying a higher mismatch between virtual elements and their physical space counterparts following MRN registration. The FN of the difference between T C P and I is calculated using the following formula:
F N = T C P I F = i = 1 4 j = 1 4 | ( T C P I ) i j | 2

2.4.4. Lesion-Based Evaluation of Accuracy

Precise lesion localization, encompassing both volume and boundaries, is essential for successful surgical planning and navigation. In this study, the Sørensen–Dice similarity coefficient (DSC) and the 95th percentile of the Hausdorff distance (HD95) were chosen to analyze the LCS-MRN system’s performance, focusing on lesion volume and shape accuracy. To assess the accuracy of the alignment between virtual and physical lesions, the ground truth (GT) and the transformed model (TM) of the lesion(s) in the physical scene were assessed using both volumetric and surface-based metrics. In this study, the TM is approximated by the lesion segmentation results upon which the transformation T C P is applied. The “Segment compare” extension in 3D Slicer facilitates the automatic computation of the DSC and HD95.
For analysis regarding the spatial volumetric overlap, the DSC is calculated as follows:
D S C = 2 | GT TM | | GT | + | TM |
Here, TM denotes the point set of the transformed model, and GT is the point set representing the ground truth. This metric ranges from 0 to 1, whereas a value of 1 indicates perfect congruence, while 0 denotes no spatial overlap.
To analyze the object’s shape similarity, the HD95 is used, calculating the maximum distance between corresponding points on the surface point sets, mitigating the effect of outliers by focusing on the 95th percentile range of the measurements. This distance is reflected in millimeters and is expressed through the formula:
H D 95 = max g t GT 95 % min t m TM t m g t 2
In this formula, max g t GT 95 % · . Calculate the maximum value of the first 95 % of the distance set, where TM denotes the set of contour points from the transformed model, and GT corresponds to the set of contour points from the ground truth.
To investigate whether these accuracy metrics related to the lesions correlate with the lesions’ intrinsic characteristics or initial scanning positions; in this study, the lesion volume, the lesion depth, and L2 norm are calculated for each lesion as follows:
  • The lesion volume is defined as the product of the number of voxels contained in the lesion’s segmentation and the physical volume of a single voxel, calculated using the “Segment statistics” module in the 3D Slicer software (Version 5.1.0).
  • The lesion depth is defined as the minimum radius of a sphere tangential to the skin surface with the lesion’s centroid as its center. Therefore, the centroid of the lesion (geometric center) is first calculated using the “Segment statistics” module. Subsequently, the radius of this tangential sphere is computed using the “Model to model distance” module in the 3D Slicer software (Version 5.1.0).
  • The L2 norm is defined as the Euclidean distance from the lesion’s centroid to the origin of the reference image space in the 3D Slicer software (Version 5.1.0).

2.4.5. Statistical Analysis

Statistical analyses were conducted using the Kruskal–Wallis test across all 19 cases for landmark-based metrics and all 21 lesions for volumetric assessments. The Mann–Whitney U-test was utilized further to investigate differences between groups in our post hoc analysis. The overall level of significance was set to p < 0.05. Considering multiple comparisons being performed (three different surgical positions), the Bonferroni correction was applied to adjust the alpha value to control for the overall Type I error rate (p < 0.017 (0.050/3)) to control for the overall Type I error rate in these comparisons. Correlations were analyzed separately for DSC and HD95 with lesion and scanning characteristics (volume, depth, and L2 norm) using Spearman’s rank correlation method. MATLAB (version R2022a, MathWorks, Apple Hill Campus, Natick, MA, USA) was utilized for all statistical computations and the generation of plots.

3. Results

This section encompasses a comprehensive quality evaluation of the LCS-MRN system. Section 3.1 details patient demographics and clinical profiles involved in the study. Section 3.2 discusses hologram generation and application for anatomical visualization to assess the system’s practicality. Section 3.3 focuses on a landmark-based accuracy evaluation, quantifying the virtual–physical discrepancies. Finally, Section 3.4 emphasizes lesion-based analysis, assessing the system’s precision in lesion localization for navigation, underlining the LCS-MRN system’s clinical utility.

3.1. Subject Demographics

In this cohort of 19 patients, comprising twelve males and seven females with an average age of 54.4 ± 18.5 years, 15 patients underwent preoperative MRI imaging in case of an intracerebral neoplasm (case 01 to 15), while four underwent CT imaging (case 16 to 20) in case of intracerebral hematoma. The patient’s demographic information, histopathological findings, and surgical positioning are shown in Table 1. Intracranial neoplasms included metastasis (n = 7), meningioma (n = 3), diffusion astrocytoma (n = 1), cavernous malformation (n = 1), diffuse large B-cell lymphoma (n = 1), aneurysmal alterations (n = 1) and high-grade glioma (n = 1). In total, 21 lesions were identified, with four lesions in the frontal, three in the parietal, two in the temporal and seven in the occipital lobe. Additionally, three lesions were located in the basal ganglia, and two were infratentorial. Furthermore, hematoma is identified in four subjects.
All their life-sized head phantoms with positioning lines were successfully created. Based on the surgical records and photo documentation, the surgical positioning of all head phantoms was accurately replicated. Eight phantoms were secured in a supine position, six in a prone position, and five in a lateral position, with three left-sided and two right-sided.

3.2. Practicality Assessment Results

A comprehensive set of holograms (n = 240) was generated for experimental purposes (see Figure 7 and Table 2). Specifically, the overall count of holograms created includes 21 instances of lesions, 15 pairs of arteries and venous sinuses, 19 ventricles, seven pairs of the optic radiation and pyramidal tract, four puncture paths, three frontal sinuses, and three pairs of endoscopic pathways and bone flaps. Additionally, 19 scalp quadrants serving as intuitive references for alignment evaluation and 124 markers serving as reference points for quantitative accuracy assessment were visualized.
For all 19 cases, 240 holograms were visualized successfully using MRN (see Figure 8). It is noteworthy that the image targets on the LCS’s MR interface were affixed on both sides, ensuring that regardless of the LCS being positioned on the user’s left (Cases 10, 11, 14) or right side (all other cases), the image targets were consistently recognized and tracked by the HL-2 system, yielding successful registration and visualization (see Figure 9).
The LCS-MRN approach effectively facilitated comprehensive intracranial anatomical visualization. The LCS deployment took 2 to 3 min, and the hologram loaded onto the MR platform for approximately 1 to 2 min.

3.3. Landmark-Based Evaluation of Accuracy

The landmark-based analysis compares the LCS-MRN system’s interpolated metrics ( FLE and TRE ) and extrapolated metrics ( FRE and FN ) across various surgical positions, including supine, prone, and lateral. Post hoc analysis was also conducted to assess statistical significance between different position groups.
In the case-oriented analysis, for interpolated metrics for accuracy, the FLE across all cases was 1.9 ± 0.4 mm, and the TRE was 3.0 ± 0.5 mm (see Figure 8 and Figure 10 and Table 3). The Kruskal–Wallis test revealed no statistically significant differences in FLE and TRE across different surgical positions (supine, prone, and lateral) ( FLE 2.0 ± 0.4 mm vs. 1.8 ± 0.2 mm vs. 1.8 ± 0.7 mm, p = 0.628; TRE 2.9 ± 0.6 mm vs. 3.1 ± 0.6 mm vs. 3.0 ± 0.4 mm, p = 0.745). For extrapolated metrics for accuracy, FRE for all cases was 2.1 ± 0.6 mm, and the FN was 3.4 ± 1.7 (see Figure 8 and Table 3). The Kruskal–Wallis test indicated no significant differences in FRE across surgical positions ( FRE 1.9 ± 0.6 mm vs. 2.4 ± 0.5 mm vs. 2.2 ± 0.4 mm, p = 0.154); however, significant differences were observed in the FN between groups (2.3 ± 0.7 vs. 4.2 ± 1.5 vs. 4.0 ± 2.2, p = 0.034). In post hoc analysis using the Mann–Whitney U-test, the alpha value was Bonferroni-adjusted to 0.017 (0.050/3) to control for the overall Type I error rate in multiple comparisons, but no pairwise comparison showed statistically significant differences (supine vs. prone, p = 0.020; supine vs. lateral, p = 0.653; prone vs. lateral, p = 0.662).

3.4. Lesion-Based Evaluation of Accuracy

The lesion-based analysis of the LCS-MRN system evaluates metrics such as DSC and HD95 alongside lesion characteristics, including lesion volume, lesion depth, and L2 norm. Statistical correlation analyses and post hoc tests were included to explore their impact on the LCS-MRN system’s performance.
In the lesion-oriented analysis, the volume for all 21 lesions was 23.9 ± 26.3 cm3, lesion depth was 3.8 ± 1.5 cm, L2 norm was 8.0 ± 2.3 cm, DSC was 0.83 ± 0.12, and HD95 was 0.76 ± 0.15 mm (see Figure 8 and Table 4). The Kruskal–Wallis test indicated statistically significant differences in DSC and HD95 among lesions in different surgical positions (DSC 0.90 ± 0.06 vs. 0.76 ± 0.15 vs. 0.81 ± 0.09, p = 0.005; HD95 1.7 ± 0.8 mm vs. 3.2 ± 0.6 mm vs. 1.9 ± 0.7 mm, p = 0.042). In post hoc analysis, the alpha value was Bonferroni-adjusted to 0.017 (0.050 / 3) to control for the overall Type I error rate for multiple comparisons. Post hoc analysis for HD95 revealed significant differences between prone and supine (p = 0.002) and between prone and lateral (p = 0.009), but no significant difference between supine and lateral positioning (p = 0.489). However, post hoc analysis using the Mann–Whitney U-test for DSC did not show statistically significant differences in pairwise comparisons (supine vs. prone, p = 0.029; supine vs. lateral, p = 0.059; prone vs. lateral, p = 0.534).
Additionally, the Kruskal–Wallis test demonstrated statistically significant differences in lesion depth and L2 norm across lesions in different surgical positions (depth, p = 0.012; L2 norm, p = 0.005) (see Figure 8 and Table 4). Although the variance in volume across different surgical positions was not statistically significant (p = 0.069), it is noteworthy that the mean values for the supine and prone groups were more than double that of the lateral group, indicating that lesion and scanning attributes were not evenly distributed among the surgical position groups. Therefore, these results do not imply that variations in surgical positions affect the magnitude of DSC and HD95.
The correlation analysis for DSC and HD95 with characteristics (lesion volume, lesion depth, and L2 norm), conducted using Spearman’s rank correlation method, revealed that apart from a statistically significant positive correlation between volume and DSC (rho = 0.813, p < 0.001), no other characteristics exhibited statistically significant relationships (see Table 5). When all characteristics were divided into groups based on greater than or less than the median value, only the volume size significantly impacted DSC (0.91 ± 0.04 vs. 0.76 ± 0.11, p < 0.001). The other groups’ differences were not significant (see Table 6). These results are consistent with the correlation analysis findings.
Figure 8 displays the extrapolated displacement and visualization results of the entire head and intracranial structures.

4. Discussion

Central to the study is evaluating an innovative MRN registration approach using an LCS, which addresses challenges such as user dependency and being cost-effective while enhancing accuracy in neurosurgical procedures. The method effectively replicates the scanner frame’s position relative to the patient and autonomously performs transformations, aligning tracking space coordinates with the imaging space. This study retrospectively utilized image data from a cohort of patients with intracranial lesions to fabricate life-sized head phantoms. These were then employed in simulated clinical environments to test the LCS-MRN system, focusing on registering image data and head phantoms to ascertain the system’s effectiveness and accuracy. The results demonstrated encouraging outcomes, laying the groundwork for future refinements and clinical applications.
The evolution of MRN can be traced through a series of advancements and challenges. Initial stages featured user-directed adjustments of virtual elements to correspond with real anatomical structures, a method necessitated recurrent recalibrations due to positional changes, thus being inefficient and inconsistent [11,12]. Advancements led to adopting a fiducial-based registration method, using distinct markers for more rapid configuration than the manual approach, albeit at the cost of increased FLE from marker displacement (e.g., skin shift) or equipment degradation (e.g., pointer abrasion) [8,21,23,31]. Meanwhile, some evolution shifted towards a markerless, surface-oriented strategy deploying computer vision (CV) algorithms, automating patient-image congruence, and minimizing physical interaction [32,33,34,35]. While streamlining the procedure during this phase, issues were encountered in maintaining registration accuracy due to sensitivity to low image quality, surface irregularities, feature scarcity, and image distortion, all demanding high computational power.
Amidst these developments, the quest for more efficient MRN methods persisted. The LCS-based registration method streamlines MRN by simplifying user involvement by aligning two laser cross-projections, surpassing the complexities of virtual object manipulation. This technique bypasses issues such as pointer degradation and the necessity for pre-defining virtual registration aids, as it employs pre-scanned, user-independent reference planes from DICOM images [7]. These orthogonally arranged planes ensure a consistent and globally representative registration process. The system is also user-friendly in assembly and operation, characterized by low-cost hardware fabrication and software processes that demand minimal computational power. A comparative analysis of registration techniques within state-of-the-art MRN paradigms is presented in Table 7, which elucidates the relative advantages and limitations of the LCS-MRN system against conventional systems.
This study’s integration of the LCS with the MRN system presents several navigational benefits. Technically, the LCS offers surgeons an intuitive physical positioning guide, complemented by the MR platform’s detailed visualization of anatomical structures, enhancing surgical area understanding. Visually, the LCS provides stable tracking in the physical space, while the MR platform maintains holographic consistency, even when surgeons adjust their perspectives during the procedure. The swift deployment of the LCS during surgery preparation, coupled with the MR platform’s visual aids, streamlines the surgical workflow. This “hybrid combination” approach, based on the principles of complementarity and compatibility, ensures that each system’s strengths are utilized while compensating for any limitations. This novel application of the LCS in establishing a consistent coordinate origin with CT or MRI data marks a significant advancement in MRN, offering improved integration and alignment accuracy. Bridging these technical and visual advantages, the dual-sided capability of the LCS’s MR interface marks a strategic innovation that optimizes spatial utility in the operating room (OR). This adaptability ensures the integration system’s functionality is maximized, allowing for flexible positioning that accommodates the surgeon’s needs and the constraints of the surgical environment.
Building on the advancements introduced by the LCS-MRN system, its versatility and scalability further extend its potential impact in the medical field. The scalability of the LCS-MRN system is highlighted by its effectiveness in various surgical positions, overcoming the limitations of traditional MRN paradigms, which are often restricted to supine positioning. The proposed system is equally effective in prone and lateral positions, broadening its applicability in diverse surgical contexts. Moreover, its foundational design centered on reference scans and integration with RICS presents the potential for compatibility with diverse imaging modalities, potentially including ultrasound and functional MRI. As long as modalities can align with RICS, they can integrate seamlessly with the LCS-MRN system, extending its applicability. Furthermore, the streamlined structural design of the LCS-MRN allows for adaptation across varied clinical settings, from advanced operating rooms to resource-limited environments, augmenting its global applicability and addressing a spectrum of medical requirements.
As a novel MRN approach, LCS-MRN aims to enhance neurosurgeons’ visual and spatial skills through perception and interaction. Theoretically, procedures that depend on these skills, such as lesion localization, understanding complex anatomical relationships, and planning surgical pathways for intraoperative guidance, will potentially benefit from this technology. In previous studies, these neurosurgical scenarios often employed in MRN evaluations serve to validate the system’s practicality and efficacy [8,11,12,22,39]. Within this study, the capabilities of the LCS-MRN system were analyzed in a diverse patient cohort of 19 human subjects, covering various pathologies, locations, and surgical objectives, including lesion resection, craniotomy planning, hemorrhage drainage, and endoscopic procedures. This analysis, conducted in a simulated clinical environment with patient imaging data, demonstrated the system’s technical performance and overall utility, confirming its effectiveness in supporting neurosurgical procedures.
The concept of accuracy in MRN systems is often variably interpreted. Still, the TRE remains a widely recognized measure for gauging navigation accuracy, applicable from initial registration stages to later phases of surgical intervention. Despite potential complications such as brain-shift-induced non-linear deformations during later surgical stages, accurate initial registration is paramount, as it fundamentally influences the accuracy and dependability of all subsequent procedural stages [26]. The TRE quantifies the discrepancy between specific points in the virtual environment and their real-world counterparts. The TRE of LCS-MRN in this study was evaluated in a three-dimensional setting. The analyses showed an average TRE of 3.0 ± 0.5 mm, ranging from 0.6 mm to 5.4 mm. This outcome aligns closely with the TRE (3.7 ± 1.7 mm) observed in the single-case pilot study of LCS-MRN [7], as well as another previous study on LCS-MRN using a fiducial-based approach showing an accuracy of 4.1 mm (interquartile range IQR 3.0–4.7 mm) [8], and is also in line with the reported TRE in conventional navigation systems applying fiducial-base registration (3.49 ± 1.09 mm) [45]. There appears to be a slight increase in accuracy with the proposed method employed in this study compared to the earlier approach, which might be related to automated approaches. However, it is important to note that these two approaches cannot be directly compared. The fiducial-based method’s accuracy was assessed using the projection error of lesions at the skin surface level as the metric for precision evaluation. While plenty of studies have examined the 3D TRE in MRN systems utilizing standalone HMDs, it is crucial to acknowledge that direct comparisons of TRE across these studies are not feasible, given the variations in their objectives and methodologies. The differences in these underline the unique aspects and context of each approach.
In mathematical terms, the TRE at specific points within this study’s context is considered a norm, necessitating the measurement of both virtual and physical points within the same coordinate space, either entirely virtual or physical. Distinct coordinate systems for each point prevent meaningful, direct measurements due to representation, scale, and direction differences. Most previous studies measured the TRE in the physical space using tools such as calipers, directly gauging the perceived distance between virtual and physical points [8]. While straightforward, this approach risks underestimating the TRE , as virtual targets could be located underneath the patient or phantom. For deep-seated targets, alternative methods involve marking virtual point locations with physical instruments (e.g., puncture needles, bone anchors) and re-registering these on post-operative images to calculate TRE [11,22,36,39,43]. This invasive approach, however, may alter the original structure and affect reliability.
In contrast, virtual space measurements offer repeatability and non-invasiveness [46,47,48]. The previous work regarding LCS-MRN developed a virtual probe for non-invasive, direct manual acquisition of actual marker points in virtual space, measuring the distance to virtual target counterparts [7]. However, moving and placing the virtual probe is subject to human error. To mitigate this, the study’s evaluations were conducted by a single neurosurgeon (Z.Q.) with extensive software/hardware and neurosurgery expertise. The FLE was introduced as a quality control parameter to reflect the potential inaccuracy due to manual operation. Worthy of note in the study’s design is that the placement of the point set V preceded that of point set P , allowing for the early acquisition of the reference metric FLE to prevent the accumulation of errors over time from affecting these baseline data, which reflect the quality of the assessment. Results indicated that FLE did not exceed TRE in any case, suggesting the manual error was well-monitored and did not compromise the credibility of TRE measurements.
Another issue worth noting is that the TRE indicates deviations at specific points, namely markers affixed to the skin surface, and does not necessarily represent the accuracy at the target level, such as lesions, eloquent structures, or proposed pathways, which remain inaccessible at that stage [7,12,49]. The study adopted alternative non-invasive approaches to address this limitation. By computing the transformation T C P from the fiducials’ centroids in RICS to their observed physical counterparts, the positions of anatomical structures beneath the surface in the physical world are extrapolated in the RICS. Consequently, this allows for using a set of similarity metrics, such as DSC and HD95, to represent the degree of overlap between virtual and physical models from volumetric and surface perspectives. While the two metrics are commonly used and effective in the context of image segmentation [50,51], their application in assessing the accuracy of AR or MR navigation systems, this methodology aligns with previous approaches [52,53], where DSC and Hausdorff Distance were similarly employed for assessment of navigation accuracy. Moreover, the study advocates using the “Segment compare” extension module in the open-source software platform 3D Slicer to calculate the DSC and HD95. This approach effectively ensures convenience and repeatability in the evaluation process.
The LCS-MRN system, designed as a markerless registration framework, presents an intriguing paradox in its validation approach, which employs physical markers on a head phantom for verification purposes. Researchers collected perceived virtual and real point sets from these specific markers. It is important to note that all quantitative accuracy data in this study, both interpolative and extrapolative, are derived from measurements of these markers, both physical and virtual. Subsequently, these measurements are extrapolated to estimate the overall accuracy of the head model. This methodology may raise concerns regarding the representativeness of the extrapolated whole-head data, as it is potentially influenced by the distribution of the markers used in the study. Since the extrapolation is based on a limited set of points, the accuracy and reliability of the whole-head data might be skewed or limited by the spatial configuration, number, and placement of these markers [54]. This concern is not unique to the LCS-MRN system, but is a common challenge in registration methods, including fiducial-based approaches. Several studies have explored the impact of maker distribution and placement on accuracy and reliability in fiducial-based registration methods [45,55]. In essence, the fidelity of the extrapolated data in representing the system’s performance across the entire head is contingent on the assumption that the markers provide a comprehensive and uniform representation of the head’s geometry. However, it is worth highlighting the success of these markers in facilitating accurate registration, evidenced not only in the context of the landmark-based MRN systems [8,23], but also in other applications, such as approaches based on an intraoperative scanner that assess TRE [1,2]. This success underscores their potential to be a reasonably reliable geometric reference for the entire head. This assertion is grounded in the premise that the strategic placement of markers, validated in earlier phases, might provide a global representativeness capable of inferring accuracy across the whole head. Extrapolation could be considered a pragmatic approach to gauge the system’s performance over the entire head region. It is important to note that, apart from landmark-based measures, other types of accuracy metrics may have been employed to assess the system’s performance comprehensively, such as angle difference [56,57], as well as parameters for nonlinear biomechanical modeling, such as Hausdorff distance [58]. Additionally, artificial intelligence (AI) and machine learning (ML) could serve as innovative strategies in image analysis, especially in object detection tasks [27], to assess and enhance the accuracy of the MRN system. For instance, regarding virtual-physical correspondence, AI can detect known physical points and their virtual counterparts, automatically computing registration quality parameters for error compensation, thereby improving its precision and reliability.
The LCS-MRN registration technique precisely projects and matches laser crosshairs on surfaces. A pilot study based on simulation modeling suggested that areas with smaller radii of curvature or more drastic curvature changes (e.g., nasal and zygomatic regions) provide more spatial information compared to larger or relatively flat curvature areas (e.g., temporal and occipital regions) [7]. This enhanced spatial information facilitates easier identification and correction of LCS deployment errors by users. However, it is important to consider the practical application of these findings in a clinical setting. In clinical practice, the visibility and accessibility of different regions on the patient’s head during image acquisition vary. For instance, the occipital region is typically not visible during image acquisition, which poses challenges when relying on crosshair lines in this area for registration. In contrast, registration in the supine and lateral positions can use crosshair lines marked in the nasal, zygomatic, and temporal regions, which are more visible and accessible. The heterogeneous curvature of the human head implies that different challenges and accuracies may arise when performing LCS-MRN registration and localization in various body positions. Landmark-based and lesion-based accuracy analyses indicated larger registration errors in the prone position, as evidenced by increased TRE , FN , HD95, and DSC values. Although post hoc analysis of the FN did not reveal pairwise statistically significant differences after Bonferroni-correction, the significantly heightened FN without correction for multiple comparisons in the prone compared to the supine position remains noteworthy, especially considering the conservative nature of Bonferroni correction, which increases the risk of Type II errors in statistical inference. A plausible explanation for this observation is that registration in supine and lateral positions typically utilizes crosshair lines marked in the nasal and zygomatic regions, as well as the temporal region, whereas, in the prone position, registration predominantly relies on crosshair lines in the occipital and temporal regions. The latter scenario involves greater use of flatter areas, potentially increasing the likelihood of LCS deployment errors. In the lesion-based accuracy analysis, the magnitudes of DSC and HD95 correspond well with the findings of the landmark-based analysis. However, despite some statistically significant differences in pairwise comparisons in the subgroup analysis, these differences cannot yet be considered engineering or clinical significance due to the imbalance in characteristics between the different positioning groups and small effect sizes.
Furthermore, the results indicated a positive correlation between the lesion volume and the DSC and a significant difference in LCS performance when localizing large lesions compared to smaller ones. The influence of lesion volume on DSC is frequently reported in automatic image segmentation [50]. However, to our knowledge, this is the first time this study has identified it in the MRN domain. These findings may suggest that the system is more effective in localizing larger lesions than smaller ones. No significant correlations were found between lesion depth and the distance of the lesion from the RICS origin. Future research might require an increased sample size or enhanced effect size to explore these aspects more thoroughly.
Some limitations in the introduced LCS-MRN system and within this study must be addressed in future work. Although the LCS-MRN system has demonstrated effectiveness in the controlled environment of a phantom study, translating these results to a clinical setting introduces additional complexities. The requirement to draw marker lines on the patient’s skin surface during the scanning process for LCS-MRN registration carries several limitations and potential negative effects. Skin movement across different body positions can affect the accuracy and stability of these markers, as the skin may not be in the same position during surgery as it was during scanning. This is particularly in cases where the surgical position differs from the scanning position, such as supine versus prone, which is not unique to the LCS-MRN, but also to conventional navigation systems [59]. Moreover, obtaining accurate marker lines for surgeries in prone or 3/4 prone positions can be challenging. When patients are scanned supine, the skin area required for the prone surgery markers is not exposed to the laser used for marking and is compressed or stretched, thereby hindering the line acquisition. Thus, the situation may introduce complexities not encountered in the controlled setting of a phantom study. Additional concerns include the risk of the lines being smudged or erased during patient transfer or preparation, which could compromise the registration process. All these factors underscore the need for a more robust and reliable method of establishing a reference frame for LCS-MRN registration in a clinical setting. Regarding this, a set of studies by Perkins et al. [60,61,62], conducted in the context of breast surgery, where skin movement and geometrical distortion are prevalent, opens a new pathway for enhancing LCS-MRN systems. Their conclusive evidence that patterns printed with magnetic ink can be imaged and identified via MRI paves the way for using temporary, flexible skin adhesives printed with this specialized ink as dual-visible skin markings. These markings promise to remain discernible to the naked eye and within MR imaging, potentially significantly improving current skin marking techniques. It can be hypothesized that applying such MR-visible markers and adhesives could substantially mitigate the geometric distortions caused by skin movement in neurosurgical procedures, which are similarly subject to changes in skin topology between imaging and surgery. This would preserve the integrity of surgical navigation markers and contribute to the precision of LCS-MRN registration. Moreover, incorporating MR-visible ink into the LCS-MRN workflow could potentially enhance procedural efficiency by reducing the need for manual line drawing and realignment, thereby shortening the preparation time and decreasing the potential for human error. Thus, further research and development are necessary to integrate MR ink and grid stickers into clinical LCS-MRN systems, including validation studies to confirm their effectiveness and safety.
The current workflow of the LCS-MRN system, although based on the 3D Slicer platform, involves a degree of complexity due to the reliance on multiple extension modules. Future research will aim to develop a dedicated extension module for the LCS-MRN to streamline this process. This specialized module would integrate key functionalities to enhance system efficiency and user experience. It would encompass the compatibility processing of holograms, linear transformations, and the evaluation of reverse engineering metrics. By consolidating these processes into a user-friendly module, the system’s overall complexity can be significantly reduced, leading to a more streamlined and efficient workflow.
Building on the existing LCS-MRN system’s framework, the next development phase will focus on enhancing real-time capabilities and addressing the current limitations in neurosurgical planning. The current LCS-MRN system and its workflow may not fully support real-time neurosurgical planning as traditional standard navigation systems do, such as real-time planning via a workstation with immediate feedback to the surgeon. This limitation arises from the inability of HL-2 to independently perform complex and computationally intensive tasks like image segmentation and 3D reconstruction. These processes are currently executed on separate computers before being transferred to HMDs for holographic visualization. However, the LCS-MRN system, as reported in this study, introduces an innovative approach through specialized MR interactions, notably via virtual probe placement for digitized marking. This feature, distinct from conventional navigation systems, enables a form of indirect ’real-time’ planning. It allows surgeons to non-invasively and non-destructively mark crucial points or pathways intraoperatively, computing their spatial information, which is a unique advantage over traditional navigation systems in real-time planning scenarios. This approach is particularly practical and promising for identifying and compensating for intraoperative brain shift.
The LCS system exhibits certain limitations, evident through its visualization outcomes, aligning with findings from previous studies. The system faces tracking instability, particularly at certain viewing angles, due to its dependence on the Vuforia SDK and the HL-2’s PV camera [63]. This instability is particularly problematic in the ‘Freeze’ mode, designed to stabilize the hologram’s position. In addition, there is an observed discrepancy between the positions used in preoperative imaging and those in the surgery, such as prone versus supine, leading to issues in registration accuracy. To address these challenges, a two-fold approach could be considered. Enhancing tracking stability might involve implementing alternative or supplementary tracking methods, which could provide stability across a wider range of viewing angles and overcome the limitations of the current hardware. Solutions for optimizing registration in various surgical positions could include aligning the scanning protocol of the preoperative imaging with the actual surgical positioning or integrating intraoperative imaging [59,64]. These methods would dynamically adapt the LCS system to the surgical environment, potentially further improving the accuracy and reliability of registration.
In addition to the broader challenges identified with the current methodology, the study is subject to some unique limitations. The current study is primarily a technical validation and proof of concept for the LCS-MRN system based on simulated neurosurgical procedures. It acknowledges the limitations in fully assessing its real-world effectiveness and applicability. Acknowledging this constraint, future research will involve methodical trials with human subjects better to understand the system’s performance in clinical scenarios and aim to conduct long-term studies involving actual surgical procedures. Thus, providing more comprehensive evidence of the system’s practical utility and reliability in real-world neurosurgical settings will be the next step. Moreover, while this initial investigation expands upon previous LCS-MRN research, it remains limited by an unbalanced sample size and a limited number of testers, which has constrained the study’s ability to extensively explore user experience and training. Recognizing this limitation, future studies are planned to more comprehensively evaluate the learning curve and user experience of the LCS-MRN system, especially in comparison to conventional paradigms, aiming to determine its ease of use and training needs for surgeons, which is vital for understanding its clinical practicality. In addition, this study did not include a comparative analysis with other state-of-the-art MRN paradigms due to objective constraints. Nevertheless, the evaluation methodology outlined, along with lesion-based and landmark-based metrics, facilitates future comparisons with various systems. Future research is anticipated to conduct these comparative studies, aiming for a comprehensive assessment of MRN technologies in neurosurgery.
Despite its challenges, the LCS-MRN system is promising and warrants continued refinement. The LCS-based registration improves earlier MRN registration techniques, offering a more straightforward, easy, and user-friendly process. Its streamlined hardware and software setup, coupled with the potential for increased accuracy and flexibility during potential surgical interventions, establishes this method as a noteworthy progression in developing cost-effective, user-friendly MRN systems for neurosurgical applications, leading to enhanced surgical results even in the absence of standard commercial navigation systems.

5. Conclusions

This study demonstrated the viability and accuracy of an LCS-MRN system through the utilization of phantoms and holograms based on patient data (MRI/CT imaging), retrospectively, particularly focusing on cases with intracranial lesions, investigating landmark- and lesion-based accuracy metrics. The findings indicate the LCS-MRN system’s potential as a technique with minimal user dependency, cost-effectiveness, and high reliability. The system’s operational efficiency and user-friendly design offer a noticeable improvement over conventional methods. With a straightforward setup in both hardware and software aspects, the system shows promise in improving accuracy and flexibility in surgical interventions. While further enhancements in precision can be anticipated, this study establishes a solid foundation for future development in accessible, user-friendly MRN technologies. Such advancements are pivotal in potentially elevating the quality of surgical procedures. Ongoing research is expected to refine further and validate these attributes in a wider range of clinical applications.

Author Contributions

Conceptualization, Z.Q., J.Z. and M.H.A.B.; methodology, Z.Q., C.N., X.C., Q.W., J.Z. and M.H.A.B.; software, Z.Q., Q.W., Z.G., S.Z., H.J., J.W., M.L., R.X., X.D. and X.C.; validation, Z.Q. and M.H.A.B.; formal analysis, Z.Q. and M.H.A.B.; investigation, Z.Q.; resources, X.C.; data curation, Z.Q. and M.H.A.B.; writing—original draft preparation, Z.Q.; writing—review and editing, M.H.A.B.; visualization, Z.Q. and M.H.A.B.; supervision, M.H.A.B., C.N., X.C. and J.Z.; project administration, M.H.A.B., X.C. and J.Z.; funding acquisition, X.C., J.Z., C.N. and M.H.A.B. All authors have read and agreed to the published version of the manuscript.

Funding

Open Access funding provided by the Open Access Publishing Fund of Philipps-Universität Marburg.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Written informed consent has been obtained from the patients to publish this paper.

Data Availability Statement

The data presented in this study are available upon reasonable request from the corresponding authors.

Acknowledgments

We would like to express our sincere appreciation to Hui Zhang for her invaluable assistance during this research.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
2DTwo-Dimensional
3DThree-Dimensional
ACSAnalysis Coordinate System
ARAugmented Reality
AIArtificial Intelligence
CTComputed Tomography
CVComputer vision
DICOMDigital Imaging and Communications in Medicine
DSCDice Similarity Coefficient
FEMForward Engineering Matrix
FLEFiducial Localization Error
FREFiducial Registration Error
FoVField of View
GTGround Truth
HCMHandedness Conversion Matrix
HD95the 95th percentile of the Hausdorff distance
HMDHead-Mounted Display
HL-2HoloLens-2
IGTImage-Guided Therapy
IMUInertial Measurement Unit
IPDInterpupillary Distance
LCSLaser Crosshair Simulator
MLMachine Learning
MRMixed Reality
MRCMixed Reality Capture
MRIMagnetic Resonance Imaging
MRNMixed Reality Navigation
MRTKMixed Reality Toolkit
OROperating Room
PVPhotos-Videos
RASRight-anterior-superior
REMReverse Engineering Matrix
RGBRed-Green-Blue
RICSReference Image Coordinate System
SDKSoftware Development Kit
SLAMSynchronous Localization and Mapping
STLstandard tessellation language
TMTransformed Model
T1T1-Weighted
T1-CET1-weighted Contrast-Enhanced
T2T2-weighted
DTIDiffusion Tensor Imaging
TRETarget Registration Error
VCSVirtual Coordinate System
VLCVisible Light Cameras
VTKVisualization Toolkit
UWPUniversal Windows Platform

References

  1. Bopp, M.H.; Corr, F.; Saß, B.; Pojskic, M.; Kemmling, A.; Nimsky, C. Augmented Reality to Compensate for Navigation Inaccuracies. Sensors 2022, 22, 9591. [Google Scholar] [CrossRef]
  2. Carl, B.; Bopp, M.; Saß, B.; Pojskic, M.; Gjorgjevski, M.; Voellger, B.; Nimsky, C. Reliable navigation registration in cranial and spine surgery based on intraoperative computed tomography. Neurosurg. Focus 2019, 47, E11. [Google Scholar] [CrossRef]
  3. Kiya, N.; Dureza, C.; Fukushima, T.; Maroon, J. Computer navigational microscope for minimally invasive neurosurgery. Min-Minim. Invasive Neurosurg. 1997, 40, 110–115. [Google Scholar] [CrossRef]
  4. Kantelhardt, S.R.; Gutenberg, A.; Neulen, A.; Keric, N.; Renovanz, M.; Giese, A. Video-assisted navigation for adjustment of image-guidance accuracy to slight brain shift. Oper. Neurosurg. 2015, 11, 504–511. [Google Scholar] [CrossRef]
  5. Stieglitz, L.H.; Fichtner, J.; Andres, R.; Schucht, P.; Krähenbühl, A.K.; Raabe, A.; Beck, J. The silent loss of neuronavigation accuracy: A systematic retrospective analysis of factors influencing the mismatch of frameless stereotactic systems in cranial neurosurgery. Neurosurgery 2013, 72, 796–807. [Google Scholar] [CrossRef]
  6. Watanabe, Y.; Fujii, M.; Hayashi, Y.; Kimura, M.; Murai, Y.; Hata, M.; Sugiura, A.; Tsuzaka, M.; Wakabayashi, T. Evaluation of errors influencing accuracy in image-guided neurosurgery. Radiol. Phys. Technol. 2009, 2, 120–125. [Google Scholar] [CrossRef]
  7. Qi, Z.; Bopp, M.H.; Nimsky, C.; Chen, X.; Xu, X.; Wang, Q.; Gan, Z.; Zhang, S.; Wang, J.; Jin, H.; et al. A Novel Registration Method for a Mixed Reality Navigation System Based on a Laser Crosshair Simulator: A Technical Note. Bioengineering 2023, 10, 1290. [Google Scholar] [CrossRef]
  8. Qi, Z.; Li, Y.; Xu, X.; Zhang, J.; Li, F.; Gan, Z.; Xiong, R.; Wang, Q.; Zhang, S.; Chen, X. Holographic mixed-reality neuronavigation with a head-mounted device: Technical feasibility and clinical application. Neurosurg. Focus 2021, 51, E22. [Google Scholar] [CrossRef]
  9. Mahvash, M.; Besharati Tabrizi, L. A novel augmented reality system of image projection for image-guided neurosurgery. Acta Neurochir. 2013, 155, 943–947. [Google Scholar] [CrossRef]
  10. Tabrizi, L.B.; Mahvash, M. Augmented reality–guided neurosurgery: Accuracy and intraoperative application of an image projection technique. J. Neurosurg. 2015, 123, 206–211. [Google Scholar] [CrossRef]
  11. Li, Y.; Chen, X.; Wang, N.; Zhang, W.; Li, D.; Zhang, L.; Qu, X.; Cheng, W.; Xu, Y.; Chen, W.; et al. A wearable mixed-reality holographic computer for guiding external ventricular drain insertion at the bedside. J. Neurosurg. 2018, 131, 1599–1606. [Google Scholar] [CrossRef]
  12. Incekara, F.; Smits, M.; Dirven, C.; Vincent, A. Clinical feasibility of a wearable mixed-reality device in neurosurgery. World Neurosurg. 2018, 118, e422–e427. [Google Scholar] [CrossRef]
  13. Shu, X.J.; Wang, Y.; Xin, H.; Zhang, Z.Z.; Xue, Z.; Wang, F.Y.; Xu, B.N. Real-time augmented reality application in presurgical planning and lesion scalp localization by a smartphone. Acta Neurochir. 2022, 164, 1069–1078. [Google Scholar] [CrossRef]
  14. Yavas, G.; Caliskan, K.E.; Cagli, M.S. Three-dimensional–printed marker–based augmented reality neuronavigation: A new neuronavigation technique. Neurosurg. Focus 2021, 51, E20. [Google Scholar] [CrossRef]
  15. Alves, M.O.; Dantas, D.O. Mobile augmented reality for craniotomy planning. In Proceedings of the 2021 IEEE Symposium on Computers and Communications (ISCC), Athens, Greece, 5–8 September 2021; pp. 1–6. [Google Scholar]
  16. De Almeida, A.G.C.; de Oliveira Santos, B.F.; Oliveira, J.L. A Neuronavigation System Using a Mobile Augmented Reality Solution. World Neurosurg. 2022, 167, e1261–e1267. [Google Scholar] [CrossRef]
  17. Cenzato, M.; Fratianni, A.; Stefini, R. Using a smartphone as an exoscope where an operating microscope is not available. World Neurosurg. 2019, 132, 114–117. [Google Scholar] [CrossRef]
  18. Deng, W.; Li, F.; Wang, M.; Song, Z. Easy-to-use augmented reality neuronavigation using a wireless tablet PC. Stereotact. Funct. Neurosurg. 2014, 92, 17–24. [Google Scholar] [CrossRef]
  19. Chiou, S.Y.; Zhang, Z.Y.; Liu, H.L.; Yan, J.L.; Wei, K.C.; Chen, P.Y. Augmented Reality Surgical Navigation System for External Ventricular Drain. Healthcare 2022, 10, 1815. [Google Scholar] [CrossRef]
  20. Abe, Y.; Sato, S.; Kato, K.; Hyakumachi, T.; Yanagibashi, Y.; Ito, M.; Abumi, K. A novel 3D guidance system using augmented reality for percutaneous vertebroplasty. J. Neurosurg. Spine 2013, 19, 492–501. [Google Scholar] [CrossRef]
  21. Van Doormaal, T.P.; van Doormaal, J.A.; Mensink, T. Clinical accuracy of holographic navigation using point-based registration on augmented-reality glasses. Oper. Neurosurg. 2019, 17, 588. [Google Scholar] [CrossRef]
  22. Li, Y.; Zhang, W.; Wang, N. Wearable mixed-reality holographic guidance for catheter-based basal ganglia hemorrhage treatment. Interdiscip. Neurosurg. 2023, 34, 101821. [Google Scholar] [CrossRef]
  23. Qi, Z.; Zhang, J.; Xu, X.; Gan, Z.; Xiong, R.; Zhang, S.; Wang, J.; Liu, M.; Li, Y.; Wang, Q.; et al. Implement of mixed reality navigation based on multimodal imaging in the resection of intracranial eloquent lesions. Zhonghua Wai Ke Za Zhi Chin. Surg. 2022, 60, 1100–1107. [Google Scholar]
  24. Meola, A.; Cutolo, F.; Carbone, M.; Cagnazzo, F.; Ferrari, M.; Ferrari, V. Augmented reality in neurosurgery: A systematic review. Neurosurg. Rev. 2017, 40, 537–548. [Google Scholar] [CrossRef]
  25. Gsaxner, C.; Li, J.; Pepe, A.; Jin, Y.; Kleesiek, J.; Schmalstieg, D.; Egger, J. The HoloLens in medicine: A systematic review and taxonomy. Med. Image Anal. 2023, 85, 102757. [Google Scholar] [CrossRef]
  26. Fick, T.; van Doormaal, J.A.; Hoving, E.W.; Willems, P.W.; van Doormaal, T.P. Current accuracy of augmented reality neuronavigation systems: Systematic review and meta-analysis. World Neurosurg. 2021, 146, 179–188. [Google Scholar] [CrossRef]
  27. Xu, X.; Zhang, J.; Yang, K.; Wang, Q.; Chen, X.; Xu, B. Prognostic prediction of hypertensive intracerebral hemorrhage using CT radiomics and machine learning. Brain Behav. 2021, 11, e02085. [Google Scholar] [CrossRef]
  28. Golby, A.J.; Kindlmann, G.; Norton, I.; Yarmarkovich, A.; Pieper, S.; Kikinis, R. Interactive diffusion tensor tractography visualization for neurosurgical planning. Neurosurgery 2011, 68, 496. [Google Scholar] [CrossRef]
  29. Zhang, F.; Noh, T.; Juvekar, P.; Frisken, S.F.; Rigolo, L.; Norton, I.; Kapur, T.; Pujol, S.; Wells, W., III; Yarmarkovich, A.; et al. SlicerDMRI: Diffusion MRI and tractography research software for brain cancer surgery planning and visualization. JCO Clin. Cancer Informatics 2020, 4, 299–309. [Google Scholar] [CrossRef]
  30. Pujol, S.; Wells, W.; Pierpaoli, C.; Brun, C.; Gee, J.; Cheng, G.; Vemuri, B.; Commowick, O.; Prima, S.; Stamm, A.; et al. The DTI challenge: Toward standardized evaluation of diffusion tensor imaging tractography for neurosurgery. J. Neuroimaging 2015, 25, 875–882. [Google Scholar] [CrossRef]
  31. Fick, T.; van Doormaal, J.; Hoving, E.; Regli, L.; van Doormaal, T. Holographic patient tracking after bed movement for augmented reality neuronavigation using a head-mounted display. Acta Neurochir. 2021, 163, 879–884. [Google Scholar] [CrossRef]
  32. Pepe, A.; Trotta, G.F.; Mohr-Ziak, P.; Gsaxner, C.; Wallner, J.; Bevilacqua, V.; Egger, J. A marker-less registration approach for mixed reality–aided maxillofacial surgery: A pilot evaluation. J. Digit. Imaging 2019, 32, 1008–1018. [Google Scholar] [CrossRef]
  33. Chien, J.C.; Tsai, Y.R.; Wu, C.T.; Lee, J.D. HoloLens-Based AR System with a Robust Point Set Registration Algorithm. Sensors 2019, 19, 3555. [Google Scholar] [CrossRef]
  34. Dewitz, B.; Bibo, R.; Moazemi, S.; Kalkhoff, S.; Recker, S.; Liebrecht, A.; Lichtenberg, A.; Geiger, C.; Steinicke, F.; Aubin, H.; et al. Real-time 3D scans of cardiac surgery using a single optical-see-through head-mounted display in a mobile setup. Front. Virtual Real. 2022, 3, 949360. [Google Scholar] [CrossRef]
  35. Li, W.; Fan, J.; Li, S.; Tian, Z.; Zheng, Z.; Ai, D.; Song, H.; Yang, J. Calibrating 3D Scanner in the Coordinate System of Optical Tracker for Image-To-Patient Registration. Front. Neurorobot. 2021, 15, 636772. [Google Scholar] [CrossRef]
  36. Gibby, J.T.; Swenson, S.A.; Cvetko, S.; Rao, R.; Javan, R. Head-mounted display augmented reality to guide pedicle screw placement utilizing computed tomography. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 525–535. [Google Scholar] [CrossRef]
  37. McJunkin, J.L.; Jiramongkolchai, P.; Chung, W.; Southworth, M.; Durakovic, N.; Buchman, C.A.; Silva, J.R. Development of a mixed reality platform for lateral skull base anatomy. Otol. Neurotol. Off. Publ. Am. Otol. Soc. Am. Neurotol. Soc. Eur. Acad. Otol. Neurotol. 2018, 39, e1137. [Google Scholar] [CrossRef]
  38. Marrone, S.; Scalia, G.; Strigari, L.; Ranganathan, S.; Travali, M.; Maugeri, R.; Costanzo, R.; Brunasso, L.; Bonosi, L.; Cicero, S.; et al. Improving mixed-reality neuronavigation with blue-green light: A comparative multimodal laboratory study. Neurosurg. Focus 2024, 56, E7. [Google Scholar] [CrossRef]
  39. Gibby, W.; Cvetko, S.; Gibby, A.; Gibby, C.; Sorensen, K.; Andrews, E.G.; Maroon, J.; Parr, R. The application of augmented reality–based navigation for accurate target acquisition of deep brain sites: Advances in neurosurgical guidance. J. Neurosurg. 2021, 137, 489–495. [Google Scholar] [CrossRef]
  40. Zhou, Z.; Yang, Z.; Jiang, S.; Zhuo, J.; Li, Y.; Zhu, T.; Ma, S.; Zhang, J. Validation of a surgical navigation system for hypertensive intracerebral hemorrhage based on mixed reality using an automatic registration method. Virtual Reality 2023, 27, 2059–2071. [Google Scholar] [CrossRef]
  41. Eom, S.; Sykes, D.; Rahimpour, S.; Gorlatova, M. NeuroLens: Augmented reality-based contextual guidance through surgical tool tracking in neurosurgery. In Proceedings of the 2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Singapore, 17–21 October 2022; pp. 355–364. [Google Scholar]
  42. Akulauskas, M.; Butkus, K.; Rutkūnas, V.; Blažauskas, T.; Jegelevičius, D. Implementation of augmented reality in dental surgery using HoloLens 2: An in vitro study and accuracy assessment. Appl. Sci. 2023, 13, 8315. [Google Scholar] [CrossRef]
  43. Liebmann, F.; Roner, S.; von Atzigen, M.; Scaramuzza, D.; Sutter, R.; Snedeker, J.; Farshad, M.; Fürnstahl, P. Pedicle screw navigation using surface digitization on the Microsoft HoloLens. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1157–1165. [Google Scholar] [CrossRef]
  44. Haxthausen, F.V.; Chen, Y.; Ernst, F. Superimposing holograms on real world objects using HoloLens 2 and its depth camera. Curr. Dir. Biomed. Eng. 2021, 7, 111–115. [Google Scholar] [CrossRef]
  45. Pfisterer, W.K.; Papadopoulos, S.; Drumm, D.A.; Smith, K.; Preul, M.C. Fiducial versus nonfiducial neuronavigation registration assessment and considerations of accuracy. Oper. Neurosurg. 2008, 62, 201–208. [Google Scholar] [CrossRef]
  46. Hayasaka, T.; Kawano, K.; Onodera, Y.; Suzuki, H.; Nakane, M.; Kanoto, M.; Kawamae, K. Comparison of accuracy between augmented reality/mixed reality techniques and conventional techniques for epidural anesthesia using a practice phantom model kit. BMC Anesthesiol. 2023, 23, 171. [Google Scholar] [CrossRef]
  47. Morita, S.; Suzuki, K.; Yamamoto, T.; Endo, S.; Yamazaki, H.; Sakai, S. Out-of-Plane Needle Placements Using 3D Augmented Reality Protractor on Smartphone: An Experimental Phantom Study. Cardiovasc. Interv. Radiol. 2023, 46, 675–679. [Google Scholar] [CrossRef]
  48. Li, C.; Lu, Z.; He, M.; Sui, J.; Lin, T.; Xie, K.; Sun, J.; Ni, X. Augmented reality-guided positioning system for radiotherapy patients. J. Appl. Clin. Med. Phys. 2022, 23, e13516. [Google Scholar] [CrossRef]
  49. Ivan, M.E.; Eichberg, D.G.; Di, L.; Shah, A.H.; Luther, E.M.; Lu, V.M.; Komotar, R.J.; Urakov, T.M. Augmented reality head-mounted display–based incision planning in cranial neurosurgery: A prospective pilot study. Neurosurg. Focus 2021, 51, E3. [Google Scholar] [CrossRef]
  50. Fick, T.; van Doormaal, J.A.; Tosic, L.; van Zoest, R.J.; Meulstee, J.W.; Hoving, E.W.; van Doormaal, T.P. Fully automatic brain tumor segmentation for 3D evaluation in augmented reality. Neurosurg. Focus 2021, 51, E14. [Google Scholar] [CrossRef]
  51. Farzaneh, N.; Williamson, C.A.; Jiang, C.; Srinivasan, A.; Bapuraj, J.R.; Gryak, J.; Najarian, K.; Soroushmehr, S.M.R. Automated Segmentation and Severity Analysis of Subdural Hematoma for Patients with Traumatic Brain Injuries. Diagnostics 2020, 10, 773. [Google Scholar] [CrossRef]
  52. Saß, B.; Carl, B.; Pojskic, M.; Nimsky, C.; Bopp, M. Navigated 3D Ultrasound in Brain Metastasis Surgery: Analyzing the Differences in Object Appearances in Ultrasound and Magnetic Resonance Imaging. Appl. Sci. 2020, 10, 7798. [Google Scholar] [CrossRef]
  53. Saß, B.; Zivkovic, D.; Pojskic, M.; Nimsky, C.; Bopp, M.H.A. Navigated Intraoperative 3D Ultrasound in Glioblastoma Surgery: Analysis of Imaging Features and Impact on Extent of Resection. Front. Neurosci. 2022, 16, 883584. [Google Scholar] [CrossRef]
  54. Steinmeier, R.; Rachinger, J.; Kaus, M.; Ganslandt, O.; Huk, W.; Fahlbusch, R. Factors influencing the application accuracy of neuronavigation systems. Stereotact. Funct. Neurosurg. 2000, 75, 188–202. [Google Scholar] [CrossRef]
  55. Smith, T.R.; Mithal, D.S.; Stadler, J.A.; Asgarian, C.; Muro, K.; Rosenow, J.M. Impact of fiducial arrangement and registration sequence on target accuracy using a phantom frameless stereotactic navigation model. J. Clin. Neurosci. 2014, 21, 1976–1980. [Google Scholar] [CrossRef]
  56. Shamir, R.R.; Joskowicz, L.; Spektor, S.; Shoshan, Y. Target and trajectory clinical application accuracy in neuronavigation. Oper. Neurosurg. 2011, 68, ons95–ons102. [Google Scholar] [CrossRef]
  57. Schneider, M.; Kunz, C.; Pal’a, A.; Wirtz, C.R.; Mathis-Ullrich, F.; Hlaváč, M. Augmented reality–assisted ventriculostomy. Neurosurg. Focus 2021, 50, E16. [Google Scholar] [CrossRef]
  58. Garlapati, R.R.; Roy, A.; Joldes, G.R.; Wittek, A.; Mostayed, A.; Doyle, B.; Warfield, S.K.; Kikinis, R.; Knuckey, N.; Bunt, S.; et al. More accurate neuronavigation data provided by biomechanical modeling instead of rigid registration. J. Neurosurg. 2014, 120, 1477–1483. [Google Scholar] [CrossRef]
  59. Furuse, M.; Ikeda, N.; Kawabata, S.; Park, Y.; Takeuchi, K.; Fukumura, M.; Tsuji, Y.; Kimura, S.; Kanemitsu, T.; Yagi, R.; et al. Influence of surgical position and registration methods on clinical accuracy of navigation systems in brain tumor surgery. Sci. Rep. 2023, 13, 2644. [Google Scholar] [CrossRef]
  60. Perkins, S.L.; Lin, M.A.; Srinivasan, S.; Wheeler, A.J.; Hargreaves, B.A.; Daniel, B.L. A mixed-reality system for breast surgical planning. In Proceedings of the 2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR-Adjunct), Nantes, France, 9–13 October 2017; pp. 269–274. [Google Scholar]
  61. Perkins, S.L.; Daniel, B.L.; Hargreaves, B.A. MR imaging of magnetic ink patterns via off-resonance sensitivity. Magn. Reson. Med. 2018, 80, 2017–2023. [Google Scholar] [CrossRef]
  62. Perkins, S.L.; Lin, M.A.; Srinivasan, S.; Wheeler, A.J.; Hargreaves, B.A.; Daniel, B.L. Perceptual accuracy of a mixed-reality system for MR-guided breast surgical planning in the operating room. In Proceedings of the 26th Annual Meeting of the ISMRM, Paris, France, 16–21 June 2018; Volume 607. [Google Scholar]
  63. Frantz, T.; Jansen, B.; Duerinck, J.; Vandemeulebroucke, J. Augmenting Microsoft’s HoloLens with vuforia tracking for neuronavigation. Healthc. Technol. Lett. 2018, 5, 221–225. [Google Scholar] [CrossRef]
  64. Dho, Y.S.; Kim, Y.J.; Kim, K.G.; Hwang, S.H.; Kim, K.H.; Kim, J.W.; Kim, Y.H.; Choi, S.H.; Park, C.K. Positional effect of preoperative neuronavigational magnetic resonance image on accuracy of posterior fossa lesion localization. J. Neurosurg. 2019, 133, 546–555. [Google Scholar] [CrossRef]
Figure 1. An illustration of general paradigms in conventional navigation systems. (A) Fiducial-based registration employs identifiable external markers placed on the patient’s skin. These are reference points to align preoperative images with the patient’s physical space during surgery. (B) Surface-based registration utilizes the contours of the patient’s exposed surfaces to create a spatial map that aligns preoperative images with the patient’s anatomy in the operating room. (C) Automatic registration effectively aligns the captured three-dimensional (3D) volumetric data (reference image) with its origin of capture, utilizing scanner tracking mechanisms such as those in intraoperative scanners to guarantee precise alignment for navigation purposes.
Figure 1. An illustration of general paradigms in conventional navigation systems. (A) Fiducial-based registration employs identifiable external markers placed on the patient’s skin. These are reference points to align preoperative images with the patient’s physical space during surgery. (B) Surface-based registration utilizes the contours of the patient’s exposed surfaces to create a spatial map that aligns preoperative images with the patient’s anatomy in the operating room. (C) Automatic registration effectively aligns the captured three-dimensional (3D) volumetric data (reference image) with its origin of capture, utilizing scanner tracking mechanisms such as those in intraoperative scanners to guarantee precise alignment for navigation purposes.
Sensors 24 00896 g001
Figure 2. Schematic of the study framework. LCS = laser crosshair simulator; MRN = mixed reality navigation.
Figure 2. Schematic of the study framework. LCS = laser crosshair simulator; MRN = mixed reality navigation.
Sensors 24 00896 g002
Figure 3. Functionality of the laser crosshair simulator (LCS). The LCS replicates a scanner’s laser crosshairs, projecting dual laser lines onto a patient’s head as seen in CT or MRI environments (A,B). Its primary function is to facilitate the translation of scanning parameters across various spatial and temporal settings to establish the position for the reference image coordinate system (C). The black curved arrow links the tracking and virtual environments via the MR interface. Detection and recognition of the target images by the HMD prompt the initialization of the virtual space, anchoring its origin at the predetermined position. ((A) provided by Alamy, has undergone minor modifications by the authors to suit the specific context of this work. Alamy has granted copyright authorization for the use and adaptation of the image). RICS = reference image coordinate system.
Figure 3. Functionality of the laser crosshair simulator (LCS). The LCS replicates a scanner’s laser crosshairs, projecting dual laser lines onto a patient’s head as seen in CT or MRI environments (A,B). Its primary function is to facilitate the translation of scanning parameters across various spatial and temporal settings to establish the position for the reference image coordinate system (C). The black curved arrow links the tracking and virtual environments via the MR interface. Detection and recognition of the target images by the HMD prompt the initialization of the virtual space, anchoring its origin at the predetermined position. ((A) provided by Alamy, has undergone minor modifications by the authors to suit the specific context of this work. Alamy has granted copyright authorization for the use and adaptation of the image). RICS = reference image coordinate system.
Sensors 24 00896 g003
Figure 4. The figure illustrates the technical validation process of the LCS-MRN system using the data of a 73-year-old female patient with a right parietal metastasis close to eloquent motor structures. (A) displays T1-weighted MRI data with key anatomical structures (green—lesion, bright red—markers, cyan—scalp quadrants, deep orange—arteries, dark blue—venous sinuses, light blue—ventricles, and magenta—pyramidal tract), markers, and laser positioning lines highlighted for 3D reconstruction. (B) shows the 3D-printed head phantom with integrated markers (blue arrows) and laser positioning lines (red arrows). (C) depicts the phantom affixed to a head clamp, indicating fixation points (red circles). (D) illustrates the deployment of the LCS and alignment of the laser crosshairs. (E) presents the interaction with holograms through the MR platform. (F,G) display the virtual probe positioning on virtual and physical markers, respectively. The visualization of the registration process and the subsequent data analysis, including the calculation of target registration error (TRE) and other metrics, is visually represented in (H,I). Detailed descriptions of each step, particularly the methodologies for measuring points and data processing, are discussed in the main text.
Figure 4. The figure illustrates the technical validation process of the LCS-MRN system using the data of a 73-year-old female patient with a right parietal metastasis close to eloquent motor structures. (A) displays T1-weighted MRI data with key anatomical structures (green—lesion, bright red—markers, cyan—scalp quadrants, deep orange—arteries, dark blue—venous sinuses, light blue—ventricles, and magenta—pyramidal tract), markers, and laser positioning lines highlighted for 3D reconstruction. (B) shows the 3D-printed head phantom with integrated markers (blue arrows) and laser positioning lines (red arrows). (C) depicts the phantom affixed to a head clamp, indicating fixation points (red circles). (D) illustrates the deployment of the LCS and alignment of the laser crosshairs. (E) presents the interaction with holograms through the MR platform. (F,G) display the virtual probe positioning on virtual and physical markers, respectively. The visualization of the registration process and the subsequent data analysis, including the calculation of target registration error (TRE) and other metrics, is visually represented in (H,I). Detailed descriptions of each step, particularly the methodologies for measuring points and data processing, are discussed in the main text.
Sensors 24 00896 g004
Figure 5. Compatibility processing of holograms for varying surgical positions. Panels (AD) display the head phantom in supine, prone, left-lateral, and right-lateral positions, respectively, highlighting the LCS-MRN system setup. Panels (EH) illustrate the different orientations of the RICS (brown) in comparison to the intermediate VCS (IVCS) (light green), the virtual coordinate system (VCS) (dark green), and the analysis coordinate system (ACS) (light blue) within the 3D Slicer software (Version 5.1.0) for each position, as well as their transformations (indicated by gradient color arrows): the Forward Engineering Matrix ( FEM ), Reverse Engineering Matrix ( REM ), and Handedness Conversion Matrix ( HCM ). Panels (IL), representing the FEM and REM , are customized for each surgical position to ensure accurate registration and analysis of the holograms within the LCS-MRN system. This process guarantees the precise display and visualization of MR content across different surgical scenarios.
Figure 5. Compatibility processing of holograms for varying surgical positions. Panels (AD) display the head phantom in supine, prone, left-lateral, and right-lateral positions, respectively, highlighting the LCS-MRN system setup. Panels (EH) illustrate the different orientations of the RICS (brown) in comparison to the intermediate VCS (IVCS) (light green), the virtual coordinate system (VCS) (dark green), and the analysis coordinate system (ACS) (light blue) within the 3D Slicer software (Version 5.1.0) for each position, as well as their transformations (indicated by gradient color arrows): the Forward Engineering Matrix ( FEM ), Reverse Engineering Matrix ( REM ), and Handedness Conversion Matrix ( HCM ). Panels (IL), representing the FEM and REM , are customized for each surgical position to ensure accurate registration and analysis of the holograms within the LCS-MRN system. This process guarantees the precise display and visualization of MR content across different surgical scenarios.
Sensors 24 00896 g005
Figure 6. An illustrative scheme demonstrates the quality assessment procedure. After registration of a set of holograms and the physical space (A) using the LCS, achieving theMRN registration, the tip of the virtual probe is positioned at the perceived centers of the virtual markers (red circles) and the physical markers (blue circles), respectively, and their 3D coordinates (automatically transformed into the reference image coordinate system (RICS)) were immediately reported and displayed (where ### represents the numerical values of R, A, or S) on the MR platform (B). Next, the coordinates reported underwent REM transformation. They were subtracted from the coordinates of the marker’s centroid, which were annotated in advance in 3D Slicer (C,D), resulting in two displacement vectors, i.e., fiducial localization error FLE and target registration error TRE (E). Moreover, the optimal rigid transformation T C P was calculated, ensuring the calculation of a set of extrapolative metrics, such as the fiducial registration error FRE and Frobenius norm FN. When T C P is applied to the original segmentation of the lesion (as the ground truth, GT), the transformed model (TM) is obtained, thereby allowing the calculation of the Sørensen–Dice similarity coefficient (DSC) and the 95% Hausdorff distance (HD95), focusing on the LCS-MRN accuracy regarding the lesion volume and shape (F).
Figure 6. An illustrative scheme demonstrates the quality assessment procedure. After registration of a set of holograms and the physical space (A) using the LCS, achieving theMRN registration, the tip of the virtual probe is positioned at the perceived centers of the virtual markers (red circles) and the physical markers (blue circles), respectively, and their 3D coordinates (automatically transformed into the reference image coordinate system (RICS)) were immediately reported and displayed (where ### represents the numerical values of R, A, or S) on the MR platform (B). Next, the coordinates reported underwent REM transformation. They were subtracted from the coordinates of the marker’s centroid, which were annotated in advance in 3D Slicer (C,D), resulting in two displacement vectors, i.e., fiducial localization error FLE and target registration error TRE (E). Moreover, the optimal rigid transformation T C P was calculated, ensuring the calculation of a set of extrapolative metrics, such as the fiducial registration error FRE and Frobenius norm FN. When T C P is applied to the original segmentation of the lesion (as the ground truth, GT), the transformed model (TM) is obtained, thereby allowing the calculation of the Sørensen–Dice similarity coefficient (DSC) and the 95% Hausdorff distance (HD95), focusing on the LCS-MRN accuracy regarding the lesion volume and shape (F).
Sensors 24 00896 g006
Figure 7. 3D-printed phantoms (rows 1, 3, 5) and corresponding holograms (rows 2, 4, 6) were generated from imaging data in all included 19 cases.
Figure 7. 3D-printed phantoms (rows 1, 3, 5) and corresponding holograms (rows 2, 4, 6) were generated from imaging data in all included 19 cases.
Sensors 24 00896 g007
Figure 8. Visual outcomes and corresponding metrics for all 19 cases utilizing the LCS-MRN system. Each case features a metrics list displayed on the left column; holograms derived from segmentation results are exhibited in the central column, with two-dimensional (2D) contour representations on the top and 3D models displayed within an MR setting on the bottom. On the right column of a case display, extrapolated correspondence errors of the registered holograms are illustrated using pseudocolor scale maps. For comparative ease, identical planes from 2D images and consistent perspectives within the 3D space are utilized. Within this figure, the legend positioned at the lower right corner elucidates the scale map legend, hologram labeling, and the significance of the metrics list. In the legend, the symbol “##” denotes a numerical value. A single asterisk “*” signifies one lesion, and a double asterisk “**” two in a case. It is important to note that for cases with two lesions (Case_08 and Case_14), the metrics reported outside the brackets pertain to the larger lesion, while values within the brackets correspond to the smaller lesion.
Figure 8. Visual outcomes and corresponding metrics for all 19 cases utilizing the LCS-MRN system. Each case features a metrics list displayed on the left column; holograms derived from segmentation results are exhibited in the central column, with two-dimensional (2D) contour representations on the top and 3D models displayed within an MR setting on the bottom. On the right column of a case display, extrapolated correspondence errors of the registered holograms are illustrated using pseudocolor scale maps. For comparative ease, identical planes from 2D images and consistent perspectives within the 3D space are utilized. Within this figure, the legend positioned at the lower right corner elucidates the scale map legend, hologram labeling, and the significance of the metrics list. In the legend, the symbol “##” denotes a numerical value. A single asterisk “*” signifies one lesion, and a double asterisk “**” two in a case. It is important to note that for cases with two lesions (Case_08 and Case_14), the metrics reported outside the brackets pertain to the larger lesion, while values within the brackets correspond to the smaller lesion.
Sensors 24 00896 g008
Figure 9. Dual-sided deployment of LCS. Panels (A,B) exhibit the LCS’s MR interface with Vuforia image targets attached on both sides, captured in a simulated operative context with the LCS positioned on different sides of the phantom head. The images clearly demonstrate successful registration and visualization of the holograms, indicating reliable recognition and tracking by the LCS-MRN system. Panels (C,D) offer schematic diagrams, emphasizing the LCS’s placement versatility around the user, accommodating both left-sided and right-sided arrangements.
Figure 9. Dual-sided deployment of LCS. Panels (A,B) exhibit the LCS’s MR interface with Vuforia image targets attached on both sides, captured in a simulated operative context with the LCS positioned on different sides of the phantom head. The images clearly demonstrate successful registration and visualization of the holograms, indicating reliable recognition and tracking by the LCS-MRN system. Panels (C,D) offer schematic diagrams, emphasizing the LCS’s placement versatility around the user, accommodating both left-sided and right-sided arrangements.
Sensors 24 00896 g009
Figure 10. Landmark-based accuracy results. The bar chart displays the magnitude of displacement vectors ( FLE , TRE , and FRE ) for each case (A). Histograms show the distribution of the magnitudes of each displacement vector across all measurements with a fitted normal curve (BD). 3D scatter plots are based on the components of each displacement vector along the three primary axes of the reference image coordinate system, with color gradients representing vector magnitude. (EG). Clustered bar charts illustrate comparisons of displacement metrics ( FLE , TRE , FRE , and FN ) grouped by surgical position (H). The whiskers in (A,H) on the plot extend to the smallest and largest data points that fall within a distance of 1.5 times the interquartile range (IQR) from the lower (Q1) and upper (Q3) quartiles, respectively.
Figure 10. Landmark-based accuracy results. The bar chart displays the magnitude of displacement vectors ( FLE , TRE , and FRE ) for each case (A). Histograms show the distribution of the magnitudes of each displacement vector across all measurements with a fitted normal curve (BD). 3D scatter plots are based on the components of each displacement vector along the three primary axes of the reference image coordinate system, with color gradients representing vector magnitude. (EG). Clustered bar charts illustrate comparisons of displacement metrics ( FLE , TRE , FRE , and FN ) grouped by surgical position (H). The whiskers in (A,H) on the plot extend to the smallest and largest data points that fall within a distance of 1.5 times the interquartile range (IQR) from the lower (Q1) and upper (Q3) quartiles, respectively.
Sensors 24 00896 g010
Table 1. Demographic data.
Table 1. Demographic data.
CaseSex [M/F]Age [Years]PathologyLocalizationSurgical Position
01M49MetastasisLeft temporalSupine
02M58Diffused astrocytomaRight temporalSupine
03M8Cavernous malformationLeft frontalSupine
04M62MeningiomaLeft cerebellarProne
05M41Diffuse large B-cell lymphomaLeft occipitalProne
06F27MeningiomaLeft frontalSupine
07M51MetastasisRight occipitalProne
08F66MetastasisBilateral occipitalProne
09F37AneurysmFourth ventricularProne
10F73MetastasisRight parietalLeft-lateral
11F54MeningiomaLeft parietalLeft-lateral
12M79MetastasisLeft occipitalRight-lateral
13F41High grade gliomaLeft parietalRight-lateral
14M74MetastasisRight frontal and occipitalLeft-lateral
15F58MetastasisLeft occipitalProne
16M84HematomaRight frontalSupine
17M63HematomaRight basal gangliaSupine
18M57HematomaRight basal gangliaSupine
19M51HematomaLeft basal gangliaSupine
Table 2. Case-specific reference images and individual hologram content.
Table 2. Case-specific reference images and individual hologram content.
CaseReference ImageL 1A 2VS 3V 4FS 5OR 6PT 7PP 8EP 9BF 10M 11SQ 12
01T1-----7 *
02T1-----7
03T1------7
04T1------7
05T1-----7
06T1-----7
07T1------7
08T1✔ (2)-----7
09T1-----6
10T1-----6
11T1-----6
12T1----6
13T1-----6
14T1✔ (2)----6
15T1-----6
16CT------7
17CT----6
18CT----7
19CT-----7
OverallCount2115151937743312419
1 Lesion(s); 2 arteries; 3 venous sinus; 4 ventricle; 5 frontal sinus; 6 optic radiation; 7 pyramidal tract; 8 puncture path; 9 endoscopic path; 10 bone flap; 11 fiducial markers; 12 scalp quadrants. “✔” = segmented; “-” = not segmented; * number of available markers; “(2)” = number of lesions.
Table 3. Landmark-based accuracy metrics among all surgical positions.
Table 3. Landmark-based accuracy metrics among all surgical positions.
CharacteristicFLE [mm]TRE [mm]FRE [mm]FN
Overall (n = 19)1.9 ± 0.43.0 ± 0.52.1 ± 0.63.4 ± 1.7
Supine (n = 8)2.0 ± 0.42.9 ± 0.61.9 ± 0.62.3 ± 0.7
Prone (n = 6)1.8 ± 0.23.1 ± 0.62.4 ± 0.54.2 ± 1.5
Lateral (n = 5)1.8 ± 0.73.0 ± 0.42.2 ± 0.44.0 ± 2.2
p-value *0.6280.7450.1540.034
Frobenius norm. * Kruskal–Wallis test.
Table 4. Lesion-based accuracy metrics among all surgical positions.
Table 4. Lesion-based accuracy metrics among all surgical positions.
CharacteristicVolume [cm3]Depth [cm]L2 Norm [cm]DSCHD95 [mm]
Overall lesions (n = 21)23.9 ± 26.33.8 ± 1.58.0 ± 2.30.83 ± 0.122.3 ± 1.0
Supine lesions (n = 8)38.1 ± 34.64.7 ± 1.76.2 ± 1.90.90 ± 0.061.7 ± 0.8
Prone lesions (n = 7)21.3 ± 18.64.0 ± 1.08.4 ± 1.20.76 ± 0.153.2 ± 0.6
Lateral lesions (n = 6)8.0 ± 7.02.6 ± 0.710.1 ± 1.80.81 ± 0.091.9 ± 0.7
p-value *0.0690.0120.0050.0050.042
* Kruskal–Wallis test.
Table 5. Correlation analysis between lesion-based accuracy metrics and lesion characteristics.
Table 5. Correlation analysis between lesion-based accuracy metrics and lesion characteristics.
Spearman’s Rank Correlationrho Valuep-Value
DSC—volume [cm3]0.813<0.001
DSC—depth [cm]0.3540.115
DSC—L2 norm [cm]−0.2010.380
HD95 [mm]—volume [cm3]−0.3770.092
HD95 [mm]—depth [cm]−0.1960.395
HD95 [mm]—L2 norm [cm]0.1590.490
Table 6. Subgroup accuracy analysis of radiological characteristics.
Table 6. Subgroup accuracy analysis of radiological characteristics.
HD95 [mm] DSC
Characteristic>Median≤Medianp-Value *>Median≤Medianp-Value *
Volume1.9 ± 1.02.6 ± 0.80.1120.91 ± 0.040.76 ± 0.11<0.001
Depth1.9 ± 0.82.6 ± 1.030.1480.88 ± 0.090.79 ± 0.120.084
L2 norm2.5 ± 1.02.1 ± 1.00.3060.80 ± 0.130.86 ± 0.100.245
* Mann–Whitney U-test.
Table 7. Comparative analysis of registration techniques in state-of-the-art MRN paradigms.
Table 7. Comparative analysis of registration techniques in state-of-the-art MRN paradigms.
Manual AlignmentFiducial-Based RegistrationMarkerless RegistrationLCS -Based Registration
AdvantagesSimplicity and portability; MR-HMD self-tracking; user perspective adaptabilitySpeed advantage; real-time AR marker detection; rigid tissue anchoringContact hazard elimination; automated surface data acquisitionSimplified process; low user-dependency; non-reliance on calibration objects; global averaging; simple assembly; cost-effectiveness
LimitationsTime consumption; lack of real-time support; spatial map coarseness; static nature; adjustment needs for spatial shiftsComplete user dependency; marker shift errors; environmental susceptibility; prone position inaccuracyrobustness challenges; noise sensitivity; high computational demand; prone position non-availabilityMarker line drawing difficulties; tracking angle and distance susceptibility
Cited worksLi et al., 2018 [11]; Li et al., 2023 [22]; Gibby et al., 2019 [36]; McJunkin et al., 2019 [37]; Marrone et al., 2024 [38]Qi et al., 2021 [8]; Qi et al., 2022 [23]; Gibby et al., 2021 [39]; Zhou et al., 2023 [40]; Eom et al., 2022 [41]; Akuluaskas et al., 2023 [42]Liebmann et al., 2019 [43]; Pepe et al., 2019 [32]; von Haxthausen et al., 2021 [44]Qi et al., 2023 [7]
Laser crosshair simulator.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, Z.; Jin, H.; Wang, Q.; Gan, Z.; Xiong, R.; Zhang, S.; Liu, M.; Wang, J.; Ding, X.; Chen, X.; et al. The Feasibility and Accuracy of Holographic Navigation with Laser Crosshair Simulator Registration on a Mixed-Reality Display. Sensors 2024, 24, 896. https://doi.org/10.3390/s24030896

AMA Style

Qi Z, Jin H, Wang Q, Gan Z, Xiong R, Zhang S, Liu M, Wang J, Ding X, Chen X, et al. The Feasibility and Accuracy of Holographic Navigation with Laser Crosshair Simulator Registration on a Mixed-Reality Display. Sensors. 2024; 24(3):896. https://doi.org/10.3390/s24030896

Chicago/Turabian Style

Qi, Ziyu, Haitao Jin, Qun Wang, Zhichao Gan, Ruochu Xiong, Shiyu Zhang, Minghang Liu, Jingyue Wang, Xinyu Ding, Xiaolei Chen, and et al. 2024. "The Feasibility and Accuracy of Holographic Navigation with Laser Crosshair Simulator Registration on a Mixed-Reality Display" Sensors 24, no. 3: 896. https://doi.org/10.3390/s24030896

APA Style

Qi, Z., Jin, H., Wang, Q., Gan, Z., Xiong, R., Zhang, S., Liu, M., Wang, J., Ding, X., Chen, X., Zhang, J., Nimsky, C., & Bopp, M. H. A. (2024). The Feasibility and Accuracy of Holographic Navigation with Laser Crosshair Simulator Registration on a Mixed-Reality Display. Sensors, 24(3), 896. https://doi.org/10.3390/s24030896

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop