Next Article in Journal
OPTILOD: Optimal Beacon Placement for High-Accuracy Indoor Localization of Drones
Next Article in Special Issue
Multi-Camera Calibration Using Far-Range Dual-LED Wand and Near-Range Chessboard Fused in Bundle Adjustment
Previous Article in Journal
An Ergonomics Analysis of Archers through Motion Tracking to Prevent Injuries and Improve Performance
Previous Article in Special Issue
Omnidirectional-Sensor-System-Based Texture Noise Correction in Large-Scale 3D Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation

Processing Speech and Images (PSI), Department of Electrical Engineering (ESAT), KU Leuven, 3001 Leuven, Belgium
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1864; https://doi.org/10.3390/s24061864
Submission received: 24 January 2024 / Revised: 9 March 2024 / Accepted: 10 March 2024 / Published: 14 March 2024
(This article belongs to the Special Issue Sensing and Processing for 3D Computer Vision: 3rd Edition)

Abstract

:
Radar data can provide additional depth information for monocular depth estimation. It provides a cost-effective solution and is robust in various weather conditions, particularly when compared with lidar. Given the sparse and limited vertical field of view of radar signals, existing methods employ either a vertical extension of radar points or the training of a preprocessing neural network to extend sparse radar points under lidar supervision. In this work, we present a novel radar expansion technique inspired by the joint bilateral filter, tailored for radar-guided monocular depth estimation. Our approach is motivated by the synergy of spatial and range kernels within the joint bilateral filter. Unlike traditional methods that assign a weighted average of nearby pixels to the current pixel, we expand sparse radar points by calculating a confidence score based on the values of spatial and range kernels. Additionally, we propose the use of a range-aware window size for radar expansion instead of a fixed window size in the image plane. Our proposed method effectively increases the number of radar points from an average of 39 points in a raw radar frame to an average of 100 K points. Notably, the expanded radar exhibits fewer intrinsic errors when compared with raw radar and previous methodologies. To validate our approach, we assess our proposed depth estimation model on the nuScenes dataset. Comparative evaluations with existing radar-guided depth estimation models demonstrate its state-of-the-art performance.

1. Introduction

With the rapid advancement of research on autonomous vehicles, various 3D algorithms have emerged to facilitate self-driving, encompassing approaches for depth estimation, 3D object detection, recognition, and segmentation. Among these, accurate outdoor depth estimation algorithms are crucial for improving 3D applications. Accordingly, researchers have proposed numerous monocular and stereo depth estimation algorithms [1,2,3,4,5,6,7,8] that exhibit notable improvements. However, each of these algorithms has its own set of weaknesses: stereo estimation is sensitive to homogeneous surfaces, environmental lighting, and texture conditions, while monocular depth estimation suffers from ill-posed and lack-of-depth characteristics. To address these limitations, approaches incorporating both camera images and lidar data as input have been suggested [9,10,11,12], leveraging the additional depth guidance from lidar to compensate for the limitations of camera features.
Despite lidar’s accurate depth measurements, its high cost and sensitivity to weather conditions limit its usage and performance. As an alternative, some researchers have explored integrating radar as an extra depth guidance into camera-based monocular depth estimation models [13,14,15,16]. While these models have shown promising improvements compared with camera-only models, integrating radar, especially using real data such as the measurements in the nuScenes dataset [17], presents challenges. The radar signal in nuScenes is two-dimensional, with only one beam positioned horizontally at a fixed height in the real world.
Additionally, radar data are extremely sparse compared with lidar or RGB image data, with less than 50 radar points on average for a corresponding 900 × 1600 image, resulting in a 0.003% density. Projecting radar points onto the corresponding image introduces inaccuracies due to the large width of the radar beam and the location difference between the camera and radar sensor, further complicating the fusion of radar in a depth estimation model. Existing methods employ preprocessing techniques for sparse radar to improve the performance of radar-guided depth estimation models by filtering the noisy measurements [13] or extending the resolution. In an earlier work, we proposed to extend each raw radar point to a vertical line to extend the resolution [14], while other methods offered to train a neural network to enhance the radar data [15,18,19]. However, these existing methods either take only spatial information into consideration or expand with a lidar-supervised neural network. Figure 1 illustrates lidar (a), raw radar (b), preprocessed radar from previous studies [14,15,18] (c–e), and our proposed radar (f). It is evident that our proposed radar offers improved density compared with others while remaining visually consistent with the corresponding image.
In this work, we propose a novel radar expansion method based on the joint bilateral filter to address sparsity and limited vertical view problems. Unlike previous methods that consider only spatial or range difference, our expansion method utilizes an algorithm based on the joint bilateral filter kernel to expand based on spatial and color intensity differences. While [15,18] suggested a lidar-supervised pretrained deep learning model to expand density and remove noisy radar points, our method relies solely on a corresponding camera image as a guidance signal for calculating joint bilateral filter confidence. In the current era of widespread neural network use, rather than relying on lidar-supervised neural networks to improve radar resolution, we are pioneering the adoption of a well-designed and established traditional image filter known as the joint bilateral filter for radar expansion. In contrast with conventional techniques that assign a weighted average of neighboring pixels to the current pixel, our approach expands sparse radar points by computing a confidence score grounded in the values of spatial and range kernels with sparse radar and reference images. Our proposed approach significantly increases sparse radar density while maintaining intrinsic errors within an acceptable range.
The main contributions of this work include the following:
  • The proposal suggests employing a joint bilateral filter and calculating a corresponding confidence map to facilitate radar expansion.
  • The use of a range-aware window size for expanding radar, providing a better expanding region and higher resolution compared with a fixed window.
  • The ability of the proposed expansion method to increase radar points by over 1000 times while minimizing intrinsic errors.
  • The flexibility of our method, which does not require lidar supervision during training and can be applied to lidar-free or unsupervised datasets.
  • Superior performance for depth estimation compared with previously proposed radar preprocessing methods under the same model settings across various evaluation metrics.
The structure of the paper is as follows: Section 2 reviews related works on lidar depth completion and radar-guided estimation. Section 3 introduces the bilateral filter and our proposed radar expansion method. Section 4 discusses experiments and results, and finally, Section 5 concludes the paper.

2. Related Works

In this section, we review representative works in the domains of monocular depth estimation, camera-lidar depth completion, and radar-guided depth estimation.

2.1. Monocular Depth Estimation

One of the essential challenges in the task of depth estimation is monocular depth estimation, where a model predicts depth based on monocular images. Early approaches to monocular depth estimation primarily utilized different encoder and decoder structures, regression constraints, or additional information inferred from camera images. Pioneering this field, Saxena et al. [20] proposed a model capturing depths and their relationships using Markov random fields, later extending their work to 3D model generation [21]. Eigen et al. [3,4] introduced a multiscale convolutional neural network to extract local and global features for monocular depth estimation, and since then, numerous works have been proposed, exploring different multiscale architectures [6,22], incorporating additional constraints such as extra semantic information [23] and optical flow [24]. Fu et al. [5] leveraged scene understanding and a ResNet [25] module, also reformulating depth estimation learning as an ordinal regression problem. While most works are based on convolutional backbones, Ranftl et al. [26] proposed a dense prediction transformer that employs a vision transformer [27] as a backbone encoder and decodes features into pixelwise dense depth. However, the ill-posed characteristic, where an infinite number of scenes can map to the same image, imposes limitations on the performance of monocular depth estimation models.

2.2. Camera-Lidar Depth Completion

To counter this, several researchers have proposed incorporating lidar as an additional guidance signal to compensate for the lack of depth information. Ma and Karaman [9] proposed concatenating low-resolution lidar with single images in an early-fusion manner to generate a dense output, and further extending the approach to an unsupervised version [10]. Wong et al. [28] employed spatial pyramid pooling to densify sparse lidar and learned topology shapes using synthetic data. Hu et al. [29] utilized a two-branch backbone for image and depth input, and merged their representations. Vangansbeke et al. [30] fused predictions based on confidence maps from both modalities. Li et al. [31] leveraged multiscale structure learning, proposing a model that takes input from different resolutions of sparse lidar, and is supervised by multiscale ground truth lidar. Qiu et al. [32] used surface normals as an intermediate representation. Jaritz et al. [33] learned to predict additional semantic segmentation to enhance depth completion. While lidar proves effective in providing supplementary depth information, its high cost and sensitivity properties create a gap between academic research and practical business usage.

2.3. Radar-Guided Monocular Depth Estimation

Due to the lower cost and robustness of radar in comparison with lidar, researchers have begun incorporating sparse radar data into monocular depth estimation models as an alternative source of depth information instead of lidar. However, the inherent characteristics of sparsity and noise in radar pose challenges, prompting most proposed approaches to address these issues before integrating radar data with images. Lin et al. [13] investigated the effects of various fusion approaches based on the sparse-to-dense model [9] and suggested a two-stage prediction method to filter noisy points in raw radar data. Lo and Vandewalle [14] extended raw radar points to a fixed height in real-world coordinates and projected them onto the image plane to address both sparsity and limited view issues. Lee et al. [16] employed a multitask strategy, generating additional 2D object detection and semantic segmentation outputs to enhance depth estimation performance. Long et al. [15] developed a neural network to densify radar depth, guiding depth estimation with additional confidence maps and expanded radar information. Huang et al. [18] enhanced sparse signals based on pixel intensity differences in RGB images and trained a superdensity neural network to address low-density and imbalanced distribution issues.
Singh et al. [19] introduced RadarNet, which initially maps an arbitrary number of radar points to object surfaces in an image, serving as the first stage to generate a quasi-dense radar depth. This is achieved through radar-camera correspondence from a single image and radar point cloud. Additionally, they proposed FusionNet, utilizing a gated fusion network as the second stage, to regulate the fusion of multimodal features and to estimate the final dense depth. Lo and Vandewalle [34] proposed a dedicated transformer module for radar input, reassembling features from radar and images instead of using readout tokens, as in [26,27].
These studies indicate that integrating radar data as additional depth guidance can improve the performance of depth estimation models. However, the existing methods face challenges. One issue is that prior efforts tend to concentrate solely on spatial information or require training with lidar, limiting their effectiveness. While expanding radar based on spatial information offers only a limited improvement in resolution, a lidar-supervised expansion network shows improved radar depth but lacks flexibility for adaptation to datasets without lidar information. Another issue stems from the inherent drawbacks of raw radar data, specifically their sparsity and limited vertical view characteristics, necessitating preprocessing. This preprocessing step becomes crucial for optimal performance in radar-guided depth estimation models. Our proposed solution is to expand sparse radar using both spatial and range information from its reference image, recognizing the fundamental correlation between camera images and radar data in autonomous vehicle datasets.

3. Methodology

Our objective is to develop a radar expansion methodology aimed at attaining an enhanced radar format and optimizing the performance of existing radar-guided depth estimation models. To address the inherent sparsity and limited field of view in radar data, we propose expanding the raw radar data, making use of the concept of a joint bilateral filter (JBF). The rationale is that radar data and images share the same object information but in different modalities. Additionally, the joint bilateral filter is a filter designed to enhance the resolution of a target image by utilizing information from a related reference image. Thus, we use images as the reference for expansion guidance since they offer abundant color intensity information about objects and surroundings. Instead of directly applying the joint bilateral filter, we propose computing a confidence map by integrating color information from a reference image and taking into account the Euclidean pixel distance in radar. Subsequently, radar expansion is performed based on the confidence scores obtained from the confidence map. Figure 2 depicts the procedural steps of our proposed joint bilateral filter radar expansion method. In this section, we initially introduce the bilateral filter and subsequently offer a comprehensive explanation of our proposed method, encompassing the computation of the confidence map and radar expansion.
Algorithm 1: Proposed Joint Bilateral Expansion.
  • function GET_REL_SIZE  ( d , w , h )
  •      u ( w × f u ) / d
  •      v ( h × f v ) / d
  •      return  u , v
  •  
  • for each point ( x , y , d ) in RADAR
  •      Start from point with a larger d
  •      u , v  GET_REL_SIZE  ( d , w , h )
  •      for i in range ( u / 2 , u / 2 )
  •           for j in range ( v / 2 , v / 2 )
  •                D s | | ( x , y ) ( x + i , y + j ) | |
  •                D r | I ( x , y ) I ( x + i , y + j ) |
  •                C J B F G σ s ( D s ) G σ r ( D r ) as in (5)
  •                if  C J B F Threshold  do
  •                     RADAR( x + i , y + j ) ←RADAR( x , y )
  • return  RADAR

3.1. Joint Bilateral Filter

The bilateral filter is a filter designed to preserve edges while effectively smoothing noise in images. Originally introduced by Tomasi and Manduchi [35], it has become a well-established technique in image processing [36]. The bilateral filter consists of a spatial kernel and a range kernel, taking into account differences not only in spatial distance but also in color values among neighbors. The main concept is that, for a pixel to influence its neighbors, it should be close and have similar values. The bilateral filter is defined as follows:
B F [ I ] p = 1 W p q S G σ s ( | | p q | | ) G σ r ( | I p I q | ) I q ,
where G σ denotes the 2D Gaussian kernel:
G σ ( x ) = 1 2 π σ 2 e x p x 2 2 σ 2 .
G σ s and G σ r are the spatial and range kernels, respectively, and σ s and σ r will specify the amount of filtering for the image I. I p is the image value at pixel position p. The kernels are calculated against all possible image locations q in the set of window size S. | | p q | | refers to the Euclidean distance between pixel locations p and q, and | I p I q | is the absolute difference in color between pixel locations p and q. W p is a normalizing factor that ensures pixel weights sum to 1.0:
W p = q S G σ s ( | | p q | | ) G σ r ( | I p I q | ) .
Note that the spatial and range kernels in the bilateral filter are multiplied, showing that both elements matter and that no smoothing occurs if either is close to zero.
With the edge-preserving property while smoothing, several works have introduced the joint bilateral filter that applies a second guidance image in the range filter to perform upsampling [37], depth reconstruction [38], and data fusion [39]. The joint bilateral filter is defined by the following:
J B F [ I ] p = 1 W p q S G σ s ( | | p q | | ) G σ r ( | I ˜ p I ˜ q | ) I q ,
where I ˜ is the guidance image.

3.2. Proposed Expansion Method

We propose expanding the sparse raw radar using the joint bilateral filter, leveraging both spatial and range kernels. However, in previous upsampling works employing the joint bilateral filter, the source was dense but with some missing pixels that needed to be filled. In our scenario, we contend with an extremely sparse source radar depth and a monocular corresponding reference image. Additionally, the bilateral filter traditionally involves a weighted average of neighboring pixels. Given our sparse source and the goal of expansion, calculating a weighted average based on neighboring pixels and assigning it to a specific pixel is not applicable. Instead, we perform expansion based on the confidence of how each radar point can contribute to its nearby pixel locations. Consequently, we calculate the JBF confidence C J B F , p of nearby points for every radar point q against given centered raw radar points p with both spatial and range kernels in window S, as follows:
C J B F , p = G σ s ( | | p q | | ) G σ r ( | I ˜ p I ˜ q | ) , for q S .
Algorithm 1 summarizes our proposed joint bilateral expansion method. The window size has to be defined before calculating the bilateral confidence. Instead of employing a fixed window size in the image plane as in [15], we opt for a fixed size in the real 3D world, which is an adaptive size in the image. Consequently, we initially project the real-world size onto the relative pixel size in the image plane based on the depth of a given point and the camera’s intrinsic value. The idea is that a closer radar point should have a larger window for expansion, while a more distant point should have a smaller window.
Subsequently, we compute the bilateral confidence of nearby points for every point in the sparse raw radar. Finally, we assign the depth of the current point to the nearby point if its joint bilateral confidence is equal to or larger than a predefined threshold. Figure 3 outlines the computation process of the JBF confidence map and the subsequent generation of an expanded radar depth based on the expansion map and sparse radar depth. Figure 4 shows samples of the raw radar and the proposed JBF expanded radar. It is clear that, based on both spatial and range constraints, our proposed method effectively captures the shape of objects.

3.3. Intrinsic Error

To show the performance of our proposed expansion method, we evaluate its intrinsic error by comparing the expanded radar against the ground truth sparse lidar. Error metrics are exclusively computed at locations where both lidar and radar data are available. The result of intrinsic error comparison with existing expansion methods is depicted in Table 1. Our proposed method outperforms all the other methods on density in that the number of points is expanded from an average of 39 points (0.01%) in raw radar to an average of 103,249 (28.68%) points per radar depth map. Since MER [15] is an expansion method that is supervised by ground truth lidar, it has the ability to cancel noisy measurements. As a result, JBF has lower errors compared with raw radar, height-extend radar [14], and S 3 radar [18] on δ and RMSE, but errors on these metrics are slightly higher compared with MER. It is noteworthy that the MER method necessitates the incorporation of both image and lidar data during its training process. In contrast, our proposed approach eliminates the need for training with lidar data, thereby facilitating its adaptability to other self-driving datasets.

4. Experiments

4.1. Implementation Details

NuScenes Dataset. We use the nuScenes dataset [17], consisting of 6 cameras, 5 FMCW radars, and a 32-beam Velodyne lidar, for conducting our experiments. The nuScenes dataset is currently one of the most comprehensive multimodal autonomous driving datasets, and most previous radar-guided depth estimation models are also based on this dataset. It consists of 1000 driving scenes captured in Boston and Singapore, and each scene contains roughly 40 manually synchronized key-frame samples from a 20 s recording of outdoor driving. The 1000 scenes are officially split into 700 training, 150 validation, and 150 testing scenes. We use the front-view data only among all views and train and test on training and validation samples with the same setting as in previous works.
Models Used in Experiments. Our aim is to propose a novel radar expansion method, so we directly use models from previous works [13,14,34], but the radar is replaced with our proposed radar. All the experiments are implemented in PyTorch [40] and trained on a Tesla V100 GPU on a DGX-1 server. The models are taken from the code released from the original papers, and we closely follow the same training and evaluation setting. To ease computation, the camera images, projected lidar depth, and radar depth are downsampled from the original shape of 900 × 1600 to a smaller shape. For radar-guided depth estimation experiments, we conduct experiments on DORN [14] and RCDPT [34]. The weights of ResNet-101 [25] in DORN are initialized via the pretrained model on ILSVRC [41,42]. The RGB images, lidar depth, and radar depth are downsampled from the original size of 900 × 1600 to 450 × 800 .
As the sky region contributes no depth values, the RGB images, lidar, and radar depth are further cropped into a shape of 350 × 800 as the training resolution. In RCDPT, both input camera images and radar depth are cropped into a size of 384 × 384 . For radar inference experiments, the input image is resized to 450 × 800 and also further cropped into a shape of 350 × 800 . For all experiments, a polynomial decay with a starting learning rate of 0.0001 and a power rate of 0.9 as the learning strategy is applied in the training phase. The batch size is set to 4, and momentum and weight decay are set to 0.9 and 0.0005, respectively. While training, we further use data augmentation for RGB images to improve the robustness as follows: gamma contrast in range (0.9, 1.1), brightness adjustment in range (0.9, 1.1), color adjustment in range (0.9, 1.1), and horizontal flipping with 0.5 probability. We train DORN, RCDPT, and S2D for 30 epochs on the nuScenes official training splits and test on the nuScenes official validation splits. The evaluation metrics used are the standard evaluation metrics also used in previous works, and calculations for all experiments are based on the size of 350 × 800 using ground truth sparse lidar with a maximum distance of 80 m. Note that the evaluation metrics are only calculated at the pixel locations with valid points in the ground truth sparse lidar. For radar expansion in our experiments, we set σ s = 25 , σ r = 10 , and the threshold is 0.05. We use a shape of 2.35 m × 2.35 m for the fixed expansion size in the real world, which is the average shape of vehicles from the released information of objects in the nuScenes dataset.

4.2. Evaluation Metrics

We use the standard metrics as in previous works to evaluate our results.
Threshold Accuracy ( δ n ): % of Y s.t.
m a x Y ^ Y , Y Y ^ = δ n < 1 . 25 n
Root Mean Square Error (RMSE):
1 N i = 1 N Y Y ^ 2 2
Absolute Relative Error (AbsRel):
1 N i = 1 N Y Y ^ Y ,
where i is pixels and N is the total number of pixels. Y and Y ^ are the dense prediction and the target depth, respectively.

4.3. Radar-Guided Depth Estimation

We train our proposed JBF radar with models from DORNradar [14] and RCDPT [34], for they are two of the state-of-the-art radar-guided depth estimation models. The quantitative results of the proposed radar and previous works on the nuScenes dataset with a depth range < 80 m are summarized in Table 2. We refer the interested readers to [13,14,15,18,19] for details of raw radar, height-extended radar, MER, S 3 , and RadarNet, respectively. In the bottom two rows of Table 2, we can see that our proposed radar with both DORNradar and RCDPT outperforms existing models with different radar formats in all evaluation metrics. It also shows that our proposed radar with the RCDPT model has better performance compared with DORNradar, which confirms the conclusion in previous works that the transformer backbone can yield better performance than the CNN backbone [26,34]. When comparing identical models trained with different radar formats, namely, DORNradar with height-extend versus JBF, and RCDPT with MER versus JBF, the results consistently indicate that the models exhibit better performance when trained with our proposed JBF radar. Figure 5 shows the qualitative results of our proposed radar compared with previous works. Although our proposed radar with RCDPT has better quantitative results, it only provides a little significant improvement in the qualitative results compared with previous models. Some structures can be captured slightly better with our proposed radar, but the results are comparable to [13,34] for the overall perceptions.

4.4. Radar Inference Experiments

We further conduct radar inference experiments as in [43] with our proposed JBF radar. The radar inference experiments aim to examine if a model can predict surroundings to a fair extent with only radar input and under lidar supervision during training. The models used in this experiment are DORNradar and S2Dradar, and the results of the comparison of our proposed radar and previous methods are shown in Table 3. It is clear that our proposed radar outperforms MER in all metrics on DORNradar, while it is comparable on S2Dradar. However, MER has higher δ and lower RMSE in the intrinsic error in Table 1 and is generated from a lidar-supervised model. It is fair for our proposed radar to have a comparable performance in one model and a better performance in another model. This also indicates that the intrinsic depth information that radar can contribute is indeed increased through our proposed expansion method comparing the results with the raw radar.

4.5. Selection of Spatial and Range Sigma

As the Gaussian kernel plays a fundamental role in our proposed method, both spatial sigma and range sigma in the joint bilateral filter significantly influence the expanded radar depth. The function of sigma in the Gaussian filter is to regulate the variation in the kernel: a larger sigma allows more variance, while a smaller sigma restricts variance. To understand the impact of different spatial and range sigma pairs on our proposed expansion method, we trained the RCDPT model under sparse lidar supervision with varying JBF expanded radar depths. Table 4 presents the evaluated results for spatial and range sigma pairs of (10, 5), (25, 10), and (50, 20), and the corresponding expanded results are depicted in Figure 6. In Figure 6, for the sigma pair ( σ s , σ r ) = (10, 5), the expansion is limited due to the smaller sigma values. Conversely, for the sigma pair ( σ s , σ r ) = (50, 20), it is evident that the sigma values are too large, causing many expanded radar points to extend beyond the objects around the original raw radar points. Therefore, we have selected the sigma pair ( σ s , σ r ) = (25, 10), as shown in the third column of Figure 6, which provides a balanced expansion of the raw radar to a fair extent. The evaluated results in Table 4 also confirm that the sigma pair ( σ s , σ r ) = (25, 10) yields the best performance among all three settings. Although the pair ( σ s , σ r ) = (50, 20) results in significantly more expanded points, these points introduce more noise, which leads to the worst performance as indicated in Table 4.

4.6. Effects of Employing Only a Single Kernel

Instead of utilizing both range and spatial kernels as in the joint bilateral filter, we further conduct experiments of expanding the raw radar data using either a single range kernel or a spatial kernel. This experiment aims to demonstrate the effectiveness of the JBF kernels. We set σ s to 25 for the spatial kernel expansion, σ r is set to 10 for range kernel expansion, and we use ( σ s , σ r ) = (25, 10) for expansion using both kernels. In Figure 7, the radar expanded with the range kernel considers only the differences in color intensity in the reference image, lacking spatial distance information, leading to overexpansion on objects. Conversely, the radar expanded with the spatial kernel takes only spatial distance into account, resulting in circular artifacts due to the absence of color information. The intrinsic error evaluated in Table 5 also reflects the same trend, indicating that the JBF-expanded radar exhibits lower errors compared with methods using a single kernel.

4.7. Impact of Using Additional Confidence Map

The JBF confidence map represents the degree of confidence for a given raw radar point to expand to all pixel locations within the expanding window. The expansion process is determined by the JBF confidence map and a threshold, as illustrated in Figure 3. Since the confidence map provides additional information on the expanded radar depth, we conducted an experiment by training the RCDPT model with our proposed JBF radar and confidence map as additional input modalities. Table 6 presents the evaluated results, comparing the performance with and without the use of the JBF confidence map as an additional input feature. The results indicate that a radar-guided depth estimation model, such as RCDPT, can fairly benefit from the inclusion of the confidence map. This additional information helps guide the model’s output more effectively for the expanded radar depth.

5. Conclusions

Compared with camera-lidar depth completion, radar-guided monocular depth estimation introduces several challenges related to the intrinsic properties of radar. Our proposed method draws inspiration from the combination of spatial and range kernels in the joint bilateral filter. We adapted the joint bilateral filter and proposed computing a confidence map based on spatial and range differences, followed by the subsequent expansion of radar data. Unlike preprocessing methods in previous works that solely use spatial information or require lidar training, our approach incorporates both spatial and range information. By utilizing only a reference image as prior information, our proposed method can be easily adapted to any other autonomous driving dataset. The experimental results demonstrate that our proposed radar method outperforms existing works across various evaluation metrics. Areas for future work include exploring more effective utilization of the confidence map in a well-designed depth estimation model with a dedicated encoder for specialized information extraction. Additionally, considering iterative applications of the bilateral filter could expand more points while maintaining low intrinsic error. In our current setting, σ s , σ r , and the threshold are fixed. However, dynamically determining these parameters based on the distance of the given point could lead to a better-expanded radar depth. The most notable issue in the bilateral filter is its complexity, which is O ( S 2 ) . Therefore, implementing a faster version is necessary to alleviate computational costs and reduce latency. Moreover, our proposed JBF radar demonstrates an improved resolution for objects and road features, surpassing the capabilities of raw radar data and previous methods using a deterministic and efficient expansion method that does not require training. Consequently, it can be employed in various autonomous driving tasks, such as 3D object detection and path planning, showing an even wider application domain than our initial purpose of depth estimation.

Author Contributions

Methodology, C.-C.L., P.V.; Software, C.-C.L.; Writing—original draft, C.-C.L.; Writing—review & editing, P.V.; Supervision, P.V. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by a KU Leuven-Taiwan MOE Scholarship and Internal Funds KU Leuven.

Data Availability Statement

Data are contained within the article. The autonomous vehicle dataset is available on the official nuscenes website (https://www.nuscenes.org/).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chang, J.; Chen, Y. Pyramid Stereo Matching Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 5410–5418. [Google Scholar]
  2. Zhang, F.; Prisacariu, V.; Yang, R.; Torr, P. GA-Net: Guided Aggregation Net for End-To-End Stereo Matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 185–194. [Google Scholar]
  3. Eigen, D.; Puhrsch, C.; Fergus, R. Depth Map Prediction from a Single Image using a Multi-Scale Deep Network. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, Canada, 8–13 December 2014; pp. 2366–2374. [Google Scholar]
  4. Eigen, D.; Fergus, R. Predicting Depth, Surface Normals, and Semantic Labels with a Common Multi-scale Convolutional Architecture. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Santiago, Chile, 7–13 December 2015; pp. 2650–2658. [Google Scholar]
  5. Fu, H.; Gong, M.; Wang, C.; Batmanghelich, K.; Tao, D. Deep Ordinal Regression Network for Monocular Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2002–2011. [Google Scholar]
  6. Lee, J.; Han, M.; Ko, D.; Suh, I. From big to small: Multi-scale local planar guidance for monocular depth estimation. arXiv 2019, arXiv:1907.10326. [Google Scholar]
  7. Godard, C.; Mac Aodha, O.; Brostow, G. Unsupervised Monocular Depth Estimation with Left-Right Consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6602–6611. [Google Scholar]
  8. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G. Digging into Self-Supervised Monocular Depth Prediction. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3827–3837. [Google Scholar]
  9. Ma, F.; Karaman, S. Sparse-to-Dense: Depth Prediction from Sparse Depth Samples, and a Single Image. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 4796–4803. [Google Scholar]
  10. Ma, F.; Cavalheiro, G.; Karaman, S. Self-Supervised Sparse-to-Dense: Self-Supervised Depth Completion from LiDAR and Monocular Camera. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 3288–3295. [Google Scholar]
  11. Siddiqui, S.; Vierling, A.; Berns, K. Multi-Modal Depth Estimation Using Convolutional Neural Networks. In Proceedings of the IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), Abu Dhabi, United Arab Emirates, 4–6 November 2020; pp. 354–359. [Google Scholar]
  12. Boettcher, W.; Hoyer, L.; Unal, O.; Li, K.; Dai, D. LiDAR Meta Depth Completion. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; pp. 7750–7756. [Google Scholar]
  13. Lin, J.; Dai, D.; Van Gool, L. Depth Estimation from Monocular Images, and Sparse Radar Data. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24–29 October 2020; pp. 10233–10240. [Google Scholar]
  14. Lo, C.-C.; Vandewalle, P. Depth Estimation From Monocular Images, and Sparse Radar Using Deep Ordinal Regression Network. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 3343–3347. [Google Scholar]
  15. Long, Y.; Morris, D.; Liu, X.; Castro, M.; Chakravarty, P.; Narayanan, P. Radar-Camera Pixel Depth Association for Depth Completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12502–12511. [Google Scholar]
  16. Lee, W.; Jovanov, L.; Philips, W. Semantic-Guided Radar-Vision Fusion for Depth Estimation, and Object Detection. In Proceedings of the 32th British Machine Vision Conference (BMVC), Online, 22–25 November 2021. [Google Scholar]
  17. Caesar, H.; Bankiti, V.; Lang, A.; Vora, S.; Liong, V.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11618–11628. [Google Scholar]
  18. Huang, Y.; Liu, Y.; Wu, T.; Su, H.; Chang, Y.; Tsou, T.; Wang, Y.; Hsu, W. S³: Learnable Sparse Signal Superdensity for Guided Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16701–16711. [Google Scholar]
  19. Singh, A.; Ba, Y.; Sarker, A.; Zhang, H.; Kadambi, A.; Soatto, S.; Srivastava, M.; Wong, A. Depth Estimation From Camera Image and mmWave Radar Point Cloud. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 9275–9285. [Google Scholar]
  20. Saxena, A.; Chung, S.; Ng, A. Learning depth from single monocular images. In Proceedings of the Advances in Neural Information Processing Systems 18 (NIPS 2005), Vancouver, BC, Canada, 5–8 December 2005. [Google Scholar]
  21. Saxena, A.; Sun, M.; Ng, A. Make3d: Learning 3d scene structure from a single still image. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 824–840. [Google Scholar] [CrossRef] [PubMed]
  22. Aich, S.; Vianney, J.; Islam, A.; Kaur, M.; Liu, B. Bidirectional Attention Network for Monocular Depth Estimation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 11746–11752. [Google Scholar]
  23. Zhu, S.; Brazil, G.; Liu, X. The edge of depth: Explicit constraints between segmentation and depth. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 13113–13122. [Google Scholar]
  24. Vijayanarasimhan, S.; Ricco, S.; Schmid, C.; Sukthankar, R.; Fragkiadaki, K. SfMNet: Learning of structure and motion from video. arXiv 2017, arXiv:1704.07804. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Ranftl, R.; Bochkovskiy, A.; Koltun, V. Vision Transformers for Dense Prediction. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 12159–12168. [Google Scholar]
  27. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 3–7 May 2021. [Google Scholar]
  28. Wong, A.; Cicek, S.; Soatto, S. Learning topology from synthetic data for unsupervised depth completion. IEEE Robot. Autom. Lett. 2021, 6, 1495–1502. [Google Scholar] [CrossRef]
  29. Hu, M.; Wang, S.; Li, B.; Ning, S.; Fan, L.; Gong, X. PENet: Towards Precise and Efficient Image Guided Depth Completion. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13656–13662. [Google Scholar]
  30. Vangansbeke, W.; Neven, D.; Debrabandere, B.; Van Gool, L. Sparse and Noisy LiDAR Completion with RGB Guidance and Uncertainty. In Proceedings of the International Conference on Machine Vision and Applications (MVA), Tokyo, Japan, 27–31 May 2019. [Google Scholar]
  31. Li, A.; Yuan, Z.; Ling, Y.; Chi, W.; Zhang, S.; Zhang, C. A Multi-Scale Guided Cascade Hourglass Network for Depth Completion. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Snowmass, CO, USA, 1–5 March 2020; pp. 32–40. [Google Scholar]
  32. Qiu, J.; Cui, Z.; Zhang, Y.; Zhang, X.; Liu, S.; Zeng, B.; Pollefeys, M. Deeplidar: Deep surface normal guided depth prediction for outdoor scene from sparse lidar data and single color image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3308–3317. [Google Scholar]
  33. Jaritz, M.; De Charette, R.; Wirbel, E.; Perrotton, X.; Nashashibi, F. Sparse and dense data with CNNs: Depth completion and semantic segmentation. In Proceedings of the International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 52–60. [Google Scholar]
  34. Lo, C.-C.; Vandewalle, P. RCDPT: Radar-Camera Fusion Dense Prediction Transformer. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  35. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Bombay, India, 7 January 1998; pp. 839–846. [Google Scholar]
  36. Paris, S.; Kornprobst, P.; Tumblin, J.; Durand, F. Bilateral Filtering: Theory and Applications; Now Foundations and Trends: Norwell, MA, USA, 2009; Available online: https://ieeexplore.ieee.org/document/8187212 (accessed on 18 August 2009).
  37. Kopf, J.; Cohen, M.; Lischinski, D.; Uyttendaele, M. Joint bilateral upsampling. ACM Trans. Graph. 2007, 26, 96–es. [Google Scholar] [CrossRef]
  38. Yang, Q.; Yang, R.; Davis, J.; Nister, D. Spatial-depth super resolution for range images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  39. Eisemann, E.; Durand, F. Flash photography enhancement via intrinsic relighting. ACM Trans. Graph. 2004, 23, 673–678. [Google Scholar] [CrossRef]
  40. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style High-Performance Deep Learning Library. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; pp. 8026–8037. [Google Scholar]
  41. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  42. Roy, A.; Todorovic, S. Monocular Depth Estimation Using Neural Regression Forest. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5506–5514. [Google Scholar]
  43. Lo, C.-C.; Vandewalle, P. How Much Depth Information can Radar Contribute to a Depth Estimation Model? In Proceedings of the IST Symposium on Electronic Imaging, San Francisco, CA, USA, 19 January 2023; pp. 122-1–122-7. [Google Scholar]
Figure 1. Samples from nuScenes [17] with lidar and different radar formats: (a) an image with 1 sweep of sparse lidar projection, (b) 5 sweeps of raw sparse radar projection, (c) height-extended radar [14], (d) S 3 radar (ad hoc) [18], (e) MER with RC-PDA 0.5 [15], (f) proposed joint bilateral filter expansion. All the point sizes are dilated for better visualization. The color of lidar and radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 1. Samples from nuScenes [17] with lidar and different radar formats: (a) an image with 1 sweep of sparse lidar projection, (b) 5 sweeps of raw sparse radar projection, (c) height-extended radar [14], (d) S 3 radar (ad hoc) [18], (e) MER with RC-PDA 0.5 [15], (f) proposed joint bilateral filter expansion. All the point sizes are dilated for better visualization. The color of lidar and radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g001
Figure 2. Illustration of the proposed joint bilateral filter expansion process. The expansion window for each radar point is initially determined by a predefined width and height, alongside the distance of the radar point under consideration, highlighted with red frames. Subsequently, both spatial and range kernels are employed to determine the expansion confidence score for every point within the window. The final radar expansion is determined by considering the bilateral confidence alongside a predefined threshold. The details of the proposed joint bilateral expansion method are summarized in Algorithm 1. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 2. Illustration of the proposed joint bilateral filter expansion process. The expansion window for each radar point is initially determined by a predefined width and height, alongside the distance of the radar point under consideration, highlighted with red frames. Subsequently, both spatial and range kernels are employed to determine the expansion confidence score for every point within the window. The final radar expansion is determined by considering the bilateral confidence alongside a predefined threshold. The details of the proposed joint bilateral expansion method are summarized in Algorithm 1. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g002
Figure 3. Schematic diagram illustrating the proposed radar expansion method. The sparse radar depth and color intensity from camera images are given features. Following the computation of the expansion window for each sparse radar point, range and spatial confidence maps are calculated based on color and distance differences. The JBF confidence map is obtained by multiplying the range and spatial confidence maps, and the expansion map is generated after applying a threshold on the JBF confidence map. Finally, the expanded radar depth is obtained by combining the raw sparse radar depth with the expansion map.
Figure 3. Schematic diagram illustrating the proposed radar expansion method. The sparse radar depth and color intensity from camera images are given features. Following the computation of the expansion window for each sparse radar point, range and spatial confidence maps are calculated based on color and distance differences. The JBF confidence map is obtained by multiplying the range and spatial confidence maps, and the expansion map is generated after applying a threshold on the JBF confidence map. Finally, the expanded radar depth is obtained by combining the raw sparse radar depth with the expansion map.
Sensors 24 01864 g003
Figure 4. Samples of the proposed radar expansion. Top row: RGB image with the 5-frame raw radar. Bottom row: RGB image with the proposed JBF radar with σ s = 25 and σ r = 10 . All the point sizes are dilated for better visualization and better viewing in color. The color of expanded radar indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 4. Samples of the proposed radar expansion. Top row: RGB image with the 5-frame raw radar. Bottom row: RGB image with the proposed JBF radar with σ s = 25 and σ r = 10 . All the point sizes are dilated for better visualization and better viewing in color. The color of expanded radar indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g004
Figure 5. Qualitative comparison of results for radar-guided depth estimation experiments. From top to bottom: input monocular image, DORNradar [14], RC-PDA [15], Lin [13], RCDPT [34], our proposed radar with RCDPT. The color of the estimated depth indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 5. Qualitative comparison of results for radar-guided depth estimation experiments. From top to bottom: input monocular image, DORNradar [14], RC-PDA [15], Lin [13], RCDPT [34], our proposed radar with RCDPT. The color of the estimated depth indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g005
Figure 6. Samples of the proposed radar expansion with different σ s and σ r in spatial and range kernels. The columns from left to right show the RGB image with 5-frame raw radar. The proposed JBF radar with σ s = 10 and σ r = 5 ; σ s = 25 and σ r = 10 ; σ s = 50 and σ r = 20 . All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 6. Samples of the proposed radar expansion with different σ s and σ r in spatial and range kernels. The columns from left to right show the RGB image with 5-frame raw radar. The proposed JBF radar with σ s = 10 and σ r = 5 ; σ s = 25 and σ r = 10 ; σ s = 50 and σ r = 20 . All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g006
Figure 7. Samples of expanded radar by either a single kernel or both kernels. The columns from left to right show the RGB image with 5-frame raw radar, the proposed JBF radar with σ s = 10 and σ r = 5 , range kernel only with σ r = 10 , and spatial kernel only with σ s = 25 . All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Figure 7. Samples of expanded radar by either a single kernel or both kernels. The columns from left to right show the RGB image with 5-frame raw radar, the proposed JBF radar with σ s = 10 and σ r = 5 , range kernel only with σ r = 10 , and spatial kernel only with σ s = 25 . All the point sizes are dilated for better visualization and better viewing in color. The color of radar data indicates the distance, ranging from 0 m (blue) to 80 m (dark red).
Sensors 24 01864 g007
Table 1. Intrinsic errors comparing our proposed expansion with state of the art on the nuScenes dataset. Errors are calculated between each radar format and ground truth sparse lidar at the locations where radar and lidar have values. Note that the depth map shape is 450 × 800 , and the maximum evaluated distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively, where the “# points” refers to the number of points in the depth map. Numbers in bold denote the best result.
Table 1. Intrinsic errors comparing our proposed expansion with state of the art on the nuScenes dataset. Errors are calculated between each radar format and ground truth sparse lidar at the locations where radar and lidar have values. Note that the depth map shape is 450 × 800 , and the maximum evaluated distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively, where the “# points” refers to the number of points in the depth map. Numbers in bold denote the best result.
MethodLidar δ 1  ↑ δ 2  ↑RMSE ↓# Points ↑Density (%) ↑
Raw-0.410.6129.9339.010.01
Height-extend [14]-0.460.6721.5291872.55
S 3 (ad hoc) [18]-0.430.6328.0465181.81
MER [15]0.730.8511.2925,3707.05
JBF-0.590.7714.64103,24928.68
Table 2. Quantitative results comparing the proposed method with existing models on the nuScenes dataset. Top rows: baseline models using monocular images only. Middle rows: existing radar-guided depth estimation methods. Bottom rows: DORNradar and RCDPT trained from scratch with our proposed JBF radar. The maximum evaluation distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. Numbers in bold indicate the best result.
Table 2. Quantitative results comparing the proposed method with existing models on the nuScenes dataset. Top rows: baseline models using monocular images only. Middle rows: existing radar-guided depth estimation methods. Bottom rows: DORNradar and RCDPT trained from scratch with our proposed JBF radar. The maximum evaluation distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. Numbers in bold indicate the best result.
MethodRadar Format δ 1 δ 2 δ 3 RMSE ↓AbsRel ↓
DORN [5]-0.8720.9520.9785.3820.117
S2D [9]-0.8620.9480.9765.6130.126
DPT [26]-0.8860.9570.9805.2440.106
S2D [9]Raw0.8760.9490.9745.6280.115
DORNradar (1-stage) [14]Height-extend0.8890.9610.9845.1910.109
DORNradar (2-stage) [14]Height-extend0.8900.9600.9835.2600.108
Lin (1-stage) [13]Raw0.8840.9530.9775.4090.112
Lin (2-stage) [13]Raw0.9010.9580.9785.1800.100
Lee [16]Raw0.8950.9580.9785.2090.104
S 3 [18] S 3 0.7980.9210.9626.770.161
S 3 (GDC) [18] S 3 0.7990.9210.9626.760.160
FusionNet [19] RadarNet0.870.950.985.790.12
RC-PDA [15]MER0.8300.9170.9566.9420.128
DPT-Early [34]MER0.8920.9560.9785.4010.099
DPT-Late [34]MER0.8880.9580.9815.2070.104
RCDPT [34]MER0.9010.9610.9815.1650.095
DORNradar (1-stage) [14]JBF0.9010.9620.9815.2280.104
RCDPT [34]JBF0.9090.9640.9854.8730.089
Table 3. Evaluation results for radar inference experiments with different methods and input radar. Note that this experiment uses the ground truth sparse lidar as the supervision signal. We used radar with RC-PDA 0.5 in MER. CAP refers to the maximum depth range in meters. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. The bold formatting used for numbers denotes the best result.
Table 3. Evaluation results for radar inference experiments with different methods and input radar. Note that this experiment uses the ground truth sparse lidar as the supervision signal. We used radar with RC-PDA 0.5 in MER. CAP refers to the maximum depth range in meters. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. The bold formatting used for numbers denotes the best result.
ModelInput RadarCAP (m) δ 1 δ 2 RMSE ↓AbsRel ↓
Raw radar800.7160.7747.8170.260
Lo [14]Height-extend [14]800.7630.8446.5820.232
MER [15]800.7360.9027.7810.227
JBF800.7860.9027.6840.196
Raw radar800.7140.7688.1510.247
Lin [13]Height-extend [14]800.7830.8656.4040.220
MER [15]800.8010.8907.2900.155
JBF800.7850.9017.6980.179
Table 4. Selection of spatial sigma ( σ s ) and range sigma ( σ r ) in our proposed expansion method. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics respectively. The bold formatting used for numbers denotes the best result.
Table 4. Selection of spatial sigma ( σ s ) and range sigma ( σ r ) in our proposed expansion method. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics respectively. The bold formatting used for numbers denotes the best result.
( σ s , σ r ) δ 1 δ 2 δ 3 RMSE ↓AbsRel ↓
(10, 5)0.9010.9610.9815.1750.093
(25, 10)0.9090.9640.9854.8730.089
(50, 20)0.8910.9590.9805.3170.102
Table 5. Intrinsic errors comparing using either a single kernel or both kernels. Errors are calculated between each radar format and ground truth sparse lidar at the locations where radar and lidar have values. Note that the expansion threshold is 0.05, the depth map shape is 450 × 800 , and the maximum evaluated distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively whereas “# points” refers to the number of points in the depth map. Numbers in bold highlight the best result.
Table 5. Intrinsic errors comparing using either a single kernel or both kernels. Errors are calculated between each radar format and ground truth sparse lidar at the locations where radar and lidar have values. Note that the expansion threshold is 0.05, the depth map shape is 450 × 800 , and the maximum evaluated distance is 80 m. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively whereas “# points” refers to the number of points in the depth map. Numbers in bold highlight the best result.
Kernel( σ s , σ r ) δ 1 δ 2 RMSE ↓# Points ↑Density (%) ↑
JBF(25, 10)0.590.7714.64103,24928.68
Range(-, 10)0.540.6919.62181,60950.44
Spatial(25, -)0.520.6622.18166,66846.29
Table 6. The impact of using the confidence map in the RCDPT model. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. The bold formatting used for numbers denotes the best result.
Table 6. The impact of using the confidence map in the RCDPT model. The symbols ↑ and ↓ indicate that higher and lower values are better for the metrics, respectively. The bold formatting used for numbers denotes the best result.
Confidence Map δ 1 δ 2 δ 3 RMSE ↓AbsRel ↓
No0.9090.9640.9854.8730.089
Yes0.9110.9670.9864.7350.087
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lo, C.-C.; Vandewalle, P. Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation. Sensors 2024, 24, 1864. https://doi.org/10.3390/s24061864

AMA Style

Lo C-C, Vandewalle P. Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation. Sensors. 2024; 24(6):1864. https://doi.org/10.3390/s24061864

Chicago/Turabian Style

Lo, Chen-Chou, and Patrick Vandewalle. 2024. "Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation" Sensors 24, no. 6: 1864. https://doi.org/10.3390/s24061864

APA Style

Lo, C. -C., & Vandewalle, P. (2024). Expanding Sparse Radar Depth Based on Joint Bilateral Filter for Radar-Guided Monocular Depth Estimation. Sensors, 24(6), 1864. https://doi.org/10.3390/s24061864

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop