Next Article in Journal
Study on the Sand Reduction Effect of Slope Vegetation Combination in Loess Areas
Previous Article in Journal
A Study on the Charring Properties of Glued Laminated Korean Larch Timber Columns
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Channel Interaction and Transformer Depth Estimation Network: Robust Self-Supervised Depth Estimation Under Varied Weather Conditions

1
School of Information Science and Technology, Nantong University, Nantong 226001, China
2
School of Future Technology, South China University of Technology, Guangzhou 510641, China
3
School of Transportation and Civil Engineering, Nantong University, Nantong 226001, China
4
Henan Airport Group, Zhengzhou 450002, China
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(20), 9131; https://doi.org/10.3390/su16209131
Submission received: 19 September 2024 / Revised: 9 October 2024 / Accepted: 17 October 2024 / Published: 21 October 2024
(This article belongs to the Section Sustainable Transportation)

Abstract

:
Monocular depth estimation provides low-cost environmental information for intelligent systems such as autonomous vehicles and robots, supporting sustainable development by reducing reliance on expensive, energy-intensive sensors and making technology more accessible and efficient. However, in practical applications, monocular vision is highly susceptible to adverse weather conditions, significantly reducing depth perception accuracy and limiting its ability to deliver reliable environmental information. To improve the robustness of monocular depth estimation in challenging weather, this paper first utilizes generative models to adjust image exposure and generate synthetic images of rainy, foggy, and nighttime scenes, enriching the diversity of the training data. Next, a channel interaction module and Multi-Scale Fusion Module are introduced. The former enhances information exchange between channels, while the latter effectively integrates multi-level feature information. Finally, an enhanced consistency loss is added to the loss function to prevent the depth estimation bias caused by data augmentation. Experiments on datasets such as DrivingStereo, Foggy CityScapes, and NuScenes-Night demonstrate that our method, CIT-Depth, exhibits superior generalization across various complex conditions.

1. Introduction

Depth estimation is a critical environmental feature acquisition method in various domains, including autonomous driving [1], robotics [2], and 3D reconstruction [3]. In autonomous driving, depth estimation enables vehicles to interpret the 3D environment, thereby enhancing decision making and navigation. This enhances road safety while simultaneously promoting sustainable transportation by optimizing vehicle routes, reducing fuel consumption, and minimizing emissions. Similarly, in robotics, depth estimation is utilized to detect terrains, stairs, and room layouts, contributing to more efficient and sustainable automation in various industries. Compared to other depth estimation methods, monocular depth estimation is advantageous due to its low hardware cost, ease of integration, and wide range of application scenarios. Monocular depth estimation is a challenging task in computer vision, requiring the prediction of depth information for every pixel in an image. The approaches to this task can be categorized into supervised and self-supervised methods. Supervised depth estimation has achieved significant success; however, supervised depth estimation requires a large amount of labeled data with precise depth information for training. High-quality deeply labeled data are expensive to obtain, and the collected data lack the diversity needed to cover different scenarios and conditions. To mitigate the reliance on labeled data, self-supervised methods have received widespread attention. Self-supervised depth estimation leverages the characteristics of the input data themselves to learn depth information, without requiring explicit depth annotations. It predicts the relationship between adjacent frames using a depth estimation network and a camera pose estimation network and then employs an image reconstruction loss to measure the difference between the reconstructed and original images, thereby forming a self-supervised signal. This approach utilizes image reconstruction loss as a self-supervised signal, enabling deep learning without labeled data and thereby effectively reducing annotation costs.
Most existing self-supervised depth estimation models rely heavily on CNN-based encoding–decoding structures. Although such structures can effectively extract image features to serve depth estimation, the CNN-based structure lacks the allocation of attention to effective features, and it is difficult to construct the correlation between pixels in different regions, thus losing more depth estimation information. With the sparse feature distribution of severe weather images, this shortcoming can greatly reduce the accuracy of CNN-based depth estimation methods. The Transformer [4] employs self-attention mechanisms to model long-range dependencies between different regions in an image. The self-attention mechanism enables the model to consider the entire image context when processing each position, effectively capturing relationships between distant regions and complex global contextual information. However, with the large resolution inputs commonly used in self-supervised depth estimation, the direct use of Transformers significantly increases computational resource requirements, making the training and inference processes much slower.
The monocular depth estimation performance often degrades under adverse weather conditions such as rain or fog. The training datasets used in existing self-supervised methods focus on good weather or stable lighting conditions, leading to the weak generalization of the models. Monocular depth estimation relies on texture information in images to infer depth, which can be significantly blurred in harsh weather conditions. On rainy days, raindrops cause image blurring, increasing the uncertainty of the depth estimation. This issue is especially critical in fields such as autonomous vehicles and robotics, which are deployed in a wide range of complex environments. The Generative Adversarial Network (GAN) [5] is a powerful generative model capable of producing highly realistic data, and it is widely used in domains such as image generation and image restoration. Due to the differences between generated and real data, directly using generated data for model training may reduce the effectiveness of the enhancement or even introduce inaccurate training signals.
To maintain high-precision depth estimation under complex weather conditions and enhance the understanding of the overall scene, this paper proposes the Channel Interaction and Transformer depth estimation network (CIT-Depth). Both the original images and simulated weather images are utilized as training data. The network backbone consists of convolutional modules and Transformer blocks. CNNs handle local features, while the Transformer models long-range relationships within the image. The model can simultaneously capture both local and global information. A channel interaction module is added between the encoder and decoder to provide a global understanding of the entire environment. During the decoding phase, multi-scale fusion decoding is employed to better leverage feature information and details at different levels. An enhanced consistency loss is incorporated into the loss function. The main contributions of this paper are summarized as follows:
  • This paper utilizes a combination of original, exposure-adjusted, and enhanced images to strengthen the model’s ability to generalize across different scenarios.
  • Two key modules are integrated: the channel interaction module improves scene comprehension by integrating information across the preceding five frames, while the Multi-Scale Fusion Module optimizes the network’s use of features at multiple levels.
  • This paper introduces a mask when calculating the reconstruction loss of the enhanced images to reduce depth estimation bias caused by data augmentation. An enhanced consistency loss is proposed to ensure consistency in depth predictions.
The rest of this paper is organized as follows: Section 2 presents the related works. Section 3 describes the methods used in this paper. Section 4 presents the experimental results. We conclude this paper in Section 5.

2. Related Work

2.1. Supervised Depth Estimation

Initially, depth estimation was treated as a simple regression task using real depth data collected from one or more sensors. With the development of deep learning, Eigen et al. were the first to use CNNs for depth estimation tasks [6]. Subsequent research has built upon this foundation by introducing more complex architectures and loss functions to enhance accuracy and robustness. A fully convolutional residual network architecture was employed, along with a proposed network upsampling method and the introduction of inverse Huber loss for optimization, achieving end-to-end training [7]. Supervised methods convert regression tasks into categorical regression tasks to improve model stability [8,9]. These networks rely on highly accurate labeled data and are difficult to generalize to other datasets.

2.2. Self-Supervised Depth Estimation

Self-supervised methods provide alternatives that do not require large amounts of labeled data. SfM-Learner [10] was the first to achieve self-supervised depth estimation using view synthesis. This method simultaneously trains a Depth Network and a Pose Network, learning both the depth and camera pose by using view synthesis as the network’s supervisory signal. The presence of dynamic objects in an image affects the accuracy of depth estimation. Monodepth2 (MD2) [11] used the pixel-by-pixel minimization of photometric errors to address occlusion issues and introduces auto-masking to handle untextured areas and dynamic objects. A novel self-supervised semantics-guided depth estimation method was proposed by Klingner et al., designed to handle dynamic objects such as moving cars and pedestrians [12]. In addition to monocular videos, stereo pairs can also be used to train depth estimation networks [13]. Garg et al. utilized the known camera displacement and the predicted left image depth map during training, minimizing the difference between the reconstructed left image and the actual left image [14]. To maintain consistency between frames, MonoRec [15] and Manydepth [16] use time-series images to warp reference frame features to target image frames based on multiple predefined depth assumptions.
Many methods aim to improve the generalization ability of self-supervised depth estimation. Zhao et al. proposed an image transfer-based domain adaptation framework to address the performance challenges of self-supervised monocular depth estimation at nighttime [17]. The md4all [18] network is trained on raw images by generating complex samples and combining them with a standard loss function, guiding the model to perform depth estimation under diverse conditions. Spencer et al. proposed a framework that concurrently learns cross-domain dense feature representations and feature consistency-based depth estimation, performing well in nighttime driving scenarios [19]. SAFENet [20] integrated semantic awareness into the representation of deep features by combining semantic and geometric information. Wang et al. used a mapping consistency image enhancement module to boost image visibility and contrast, leading to a better performance on nighttime datasets [21]. While these networks generally perform well under single weather conditions, their performance deteriorates in complex real-world environments with changing weather conditions and varying lighting.
There are also several research efforts aimed at improving the structure of depth estimation networks. Yin et al. replaced VGG networks with ResNet-based encoder–decoder architectures [22]. PackNet [23] used 3D convolution to compress and decompress features through a symmetric packing and unpacking module. With the rise of Transformers, many studies have applied them to computer vision tasks. Dosovitskiy et al. integrated the Transformer architecture into computer vision [24]. MT-SfMLearner [25] combined the Dense Prediction Transformer (DPT) [26] with Monodepth2 modules to enhance the depth prediction performance. Li et al. proposed a parallel encoder architecture that combines Transformers and convolutional networks to separately handle long-range and short-range depth estimation [27]. Hwang et al. used CNNs in the early stages for local feature extraction and ViT in the later stages for global feature extraction, leading to a better model performance and reduced computational complexity [28]. The current research challenge lies in effectively combining CNNs and Transformers to leverage their respective strengths while balancing computational complexity and performance gains.

3. Method

3.1. Data Augmentation for Simulating Weather Variability

In depth estimation tasks, the model must accurately infer structural and depth features within an image. However, since the KITTI dataset primarily contains images captured under sunny and uniformly lit conditions, the model tends to overfit to these specific scenarios. Therefore, enriching the diversity of the dataset is crucial. This paper utilizes GANs to generate diverse weather and lighting effects, thereby augmenting the KITTI dataset. Training on these augmented datasets enables the model to maintain consistent geometric understanding under diverse weather conditions. For example, the model can learn to recognize slight blurriness caused by raindrops or low contrast in scenes affected by fog, which are common in real-world scenarios. By using both the original data and GAN-generated augmented data, the model can learn the features of real-world scenes while also enhancing its understanding of scenes under varying weather and lighting conditions. To better simulate real-world lighting variations, this paper introduces overexposure and underexposure effects into the images. Overexposure is applied to the bright regions, while underexposure is introduced to the dark regions, generating a series of images with varying exposures to mimic different lighting conditions. This method of localized exposure adjustment enhances the overall perceptual quality of the image while carefully preserving the original texture and visual cues, ensuring that the enhancement process does not cause the loss or distortion of key information. To further enhance the depth estimation model’s adaptability to various lighting conditions, this paper uses different overexposure enhancement factors E bright and underexposure reduction factors E dark to generate a series of images with varying exposure levels.
To generate a KITTI dataset with rainy weather effects, this paper adopts the CycleGAN [29] framework. The CycleGAN network structure includes two generators and two discriminators. Generator G is responsible for mapping the source domain (sunny KITTI images) to the target domain (rainy or foggy images), while Generator F maps the target domain back to the source domain. The architecture of the generator is based on CNNs, comprising convolutional layers, residual blocks, and transposed convolutional layers. Discriminator DX is used to distinguish between real sunny KITTI images and fake sunny images generated by F. Discriminator DY is used to distinguish between real rainy images and fake rainy images generated by G. The cycle consistency loss ensures that the generator retains the essential features and structure of the input image while producing the target domain image. Using the trained CycleGAN model, this paper inputs sunny KITTI images into Generator G to produce rainy and foggy images. This paper uses CoMoGAN [30] to generate scene data under complex lighting conditions such as night and dusk. CoMoGAN is capable of learning nonlinear continuous image transformations, making it particularly suitable for scenarios involving complex lighting variations. CoMoGAN employs functional instance normalization layers. By adjusting the ϕ value, the network can smoothly transition from day to dusk and then to night, allowing for precise control over lighting changes. Decoupled residual blocks are used to separate image content from lighting features, making the generated images visually closer to real night and dusk scenes. This paper combines various weather conditions, such as rain + fog and rain + night. To further enhance the model’s robustness, Gaussian noise, snow, and motion blur effects are added to the KITTI dataset.

3.2. Self-Supervised Depth Estimation Network

The proposed network framework consists of a depth estimation network and a pose estimation network, as shown in Figure 1. The depth estimation network predicts scene depth, while the pose estimation network estimates the camera’s relative pose between consecutive frames. The depth estimation network adopts an encoder–decoder architecture. The PoseNet in this paper uses ResNet18 [31] as its backbone network. The network input is created by concatenating the target view with the source view. The network output represents the relative pose between the target view and each source view. If there are N views, the network outputs N-1 relative poses corresponding to the source views. The relative pose is represented by three Euler angles and three translation components.

3.2.1. Channel Interaction Enhanced Depth Encoder

As shown in Figure 1, the deep encoder consists of a Conv-stem block, four CNN and Transformer modules, and a channel interaction module. The input image is processed by a Conv-stem block, which consists of two 3 × 3 convolutional layers, producing features at a resolution of H 2 × W 2 .
Image features are progressively extracted through four tandem CNN and Transformer blocks. As illustrated in Figure 2, the Multi-Scale Feature Embedding module stacks varying numbers of 3 × 3 convolutional layers to capture features with different receptive fields. Features with 3 × 3 receptive fields are processed by the convolution module, while features with 3 × 3 , 5 × 5 , and 7 × 7 receptive fields are fed into three parallel Transformer modules. The convolution block comprises 1 × 1 convolution, 3 × 3 depthwise convolution, and another 1 × 1 convolution, utilizing residual concatenation to improve information flow and extract local features X. Using the Transformer module proposed by MonoViT [32], the Multi-Scale Feature Embedding segments the image into visual tokens and applies factorized attention [33] to obtain global weighted features:
FactorAtt ( Q , K , V ) = Q C · ( softmax ( K ) V )
The Hybrid Feature Fusion Module combines the local feature X with the global features Z 1 , Z 2 , and Z 3 and then applies 1 × 1 convolutions to generate features for the next stage:
Y n + 1 = Conv ( Concat ( X n , Z n , 1 , Z n , 2 , Z n , 3 )
The encoder in the existing network outputs features from the final layer and directly forwards them to the decoder for further processing. This makes it difficult to model long-range relationships between different regions and to effectively capture global information, resulting in the decreased accuracy of depth estimation under adverse weather conditions. To address the above issues, this paper proposes a channel interaction module as shown in Figure 3 between the encoder and decoder. The feature map Y R H × W × C generated by the encoder is reshaped into a two-dimensional matrix Y R N × C . Matrix multiplication is applied between Y and its transposed feature map Y T to compute the channel similarity matrix C S R C × C :
C S m , n = Y m · Y n T
A low channel similarity indicates that the channels capture distinct features. To reduce the potential negative impact of noise from a single frame on channel similarity, this paper averages the channel similarities over the previous five frames:
G S = 1 T t = 1 T C S t
where T = 5 indicates that the average is taken over the preceding five frames. By integrating information from multiple frames, the method more effectively handles dynamic objects and improves global comprehension. Feature maps from different frames capture diverse details of the same scene. In foggy conditions, different frames capture variations in scene depth information due to changes in the fog density. In rainy conditions, the dynamic changes in raindrops can impact scene clarity and the stability of depth estimation. By integrating information from these frames, the model can better adapt to weather changes, thereby improving the accuracy and robustness of depth estimation. Using the softmax function, we obtain the attention matrix Att R C × C :
Att m , n = exp ( GS m , n ) n = 1 C exp ( GS m , n )
The reshaped feature map and the original feature map are added element-wise to obtain the final output F R H × W × C :
F m = n = 1 C Att m , n Y n + Y m
The final output features are derived from a weighted fusion of multi-frame channel features, facilitating the improved capture of global contextual information.

3.2.2. Multi-Scale Feature Fusion Depth Decoder

Skip connections directly convey shallow, high-resolution features to the upsampling stage of the decoder. Without proper selection, concatenation causes unimportant information to receive disproportionate attention. By incorporating the Global Channel Pooling (GC) module into the skip connections, as shown in Figure 4, the model enhances the features of important channels. Global average pooling is applied to the feature map F R H × W × C to obtain a global feature vector:
g c = 1 H × W h = 1 H w = 1 W F h , w , c
A two-layer fully connected network generates a weight s c R 1 × 1 × C for each channel:
s c = σ ( W 2 · ReLU ( W 1 · g ) )
where σ represents the sigmoid function. We apply the channel weight s c to the corresponding channel of the original feature map:
F rew = s c · F
High-resolution features obtained via skip connections primarily capture low-level information, containing a substantial amount of local texture features. Low-resolution features extracted from deeper layers capture more global contextual information. Most existing networks merely concatenate high-resolution features from skip connections with upsampled low-resolution feature maps. The direct concatenation of features hinders the model’s ability to balance global and local information, adversely affecting its performance in complex weather conditions. To more effectively fuse features from different levels, this paper proposes the Multi-Scale Fusion Module, as illustrated in Figure 5. First, concatenate the high-resolution features F skip from the skip connections with the low-resolution features F up after upsampling. Then, apply a 3 × 3 convolution followed by batch normalization and the ReLU activation function to obtain F con :
F con = ReLU ( BN ( Conv ( Concat ( F skip , F up ) ) ) )
The concatenated features F con are processed by a channel attention module to compute the weight W c R 1 × 1 × 2 C for each channel:
W c = ChannelAtt ( F con )
The channel weights W c are multiplied by the features F con to generate the weighted feature map. The introduction of the channel attention module enables the model to emphasize key features based on varying weather conditions. In rainy conditions, the model can emphasize depth differences between the background and foreground, thereby more effectively handling visual disturbances caused by raindrops. In foggy conditions, the model can enhance the recognition of important scene structures, ensuring that key depth information is accurately captured even in low-contrast settings. To highlight key features while maintaining the overall feature, the weighted features are added to the original features F con :
F fin = W c F con + F con
The processed features retain the structural information of the original scene while enhancing the model’s understanding of both global and local features under complex weather conditions.

3.3. Loss Function with Enhanced Consistency Loss

Based on the method proposed by Zhou et al. [10], this paper trains a depth estimation network and a Pose Network simultaneously for image reconstruction. Given an input image I t , the depth network predicts its depth D t . The pose estimation network uses neighboring frames of the input image to predict the relative pose T t t between I t and I t , where t { t 1 , t + 1 } . We obtain the reconstructed image I t :
I t t = I t Proj ( D t , P t t K )
where Proj ( ) projects the depth map D t of the target image I t into the 2D coordinate system of I t , denotes the sampling operator, and K is the camera’s intrinsic matrix. The Depth and Pose Networks are trained by minimizing the pixel photometric reprojection loss:
L P = min t pe ( I t , I t t )
Here, pe represents the photometric loss, which is the sum of the Structural Similarity Index (SSIM) [34] and L1 loss:
pe = α 2 ( 1 SSIM ( I t , I t ) ) + ( 1 α ) I t I t 1
where α = 0.85 . The GAN generates individual images without accounting for temporal consistency within the original image sequences, leading to unpredictable variations when enhanced image sequences are directly used for pose estimation. To resolve this issue, this paper inputs the enhanced images only into the depth estimation network, while the pose estimation network continues to use the original images. The enhanced images are reconstructed by
I t t aug = I t aug Proj ( D t aug , P t t , K )
Artifacts or noise in certain enhanced images lead to a high reprojection loss, which affects the overall loss calculation. To mitigate excessive errors from enhanced images, an element-wise mask M is incorporated into the loss function to ignore pixels with significant errors in the enhanced images. The element-wise mask independently evaluates each pixel, ensuring that high-error pixels caused by noise or artifacts are effectively disregarded during loss calculation. The model can focus on more reliable pixel regions, thereby mitigating the negative impact of unreliable data on the training process. In adverse weather or complex scenes, this helps the model learn more stably, effectively avoiding biases introduced by artifacts and noise. The reprojection loss of the enhanced images is incorporated into L P :
L P = min t p e ( I t , I t t ) + M · min t p e ( I t , I t t aug )
The depth information from the original and enhanced images, combined with the same pose estimation, is used to reconstruct the target image. The resulting reconstructed images should be identical. Based on this idea, this paper incorporates the enhanced consistency loss into the final loss function:
L C = min t p e ( I t t , I t t aug )
where I t t and I t t aug represent the reconstructed images generated from the original image and the augmented image. The enhanced consistency loss helps the model better maintain an understanding of scene structure and consistency by minimizing the error between the augmented reconstructed image and the original reconstructed image. Particularly under complex weather conditions, this loss ensures that the model consistently captures the geometric features of the scene when handling both original and augmented images, thereby reducing uncertainties introduced by data augmentation and enhancing the robustness and accuracy of depth estimation. To maintain continuity in textureless regions of the depth map while preserving details in edge details, this paper uses an edge-aware smoothness loss inspired by previous work [11]:
L S = x d t * e x I t + y d t * e y I t
where d t * = d t d ^ t represents the mean-normalized inverse depth. The final loss function combines the photometric consistency loss, edge-aware smoothness loss, and enhanced consistency loss:
L = μ L P + ω L S + γ L C
The complete process of our method is shown in Algorithm 1.
Algorithm 1 CIT-Depth: Depth Estimation Network
  • Components: DepthNet, PoseNet
  • Loss function: Photometric reprojection loss, Enhanced consistency loss, Edge-aware smoothness loss
  • if Training Phase then
  •     Input: Image sequence ( I t , I t ) , Augmented image sequence ( I t aug , I t aug )
  •     while not converged and not reached max iterations do
  •          I t DepthNet ⇒ Depth map ( D t )
  •          I t aug DepthNet ⇒ Depth map ( D t aug )
  •          ( I t , I t ) PoseNet ⇒ Relative pose ( P t t )
  •         Perform reprojection: I t Reproject ( D t , P t t , K ) I t t
  •         Perform reprojection: I t aug Reproject ( D t aug , P t t , K ) I t t aug
  •         Apply mask to augmented reprojection loss
  •         Compute photometric reprojection loss ( L P )
  •         Compute edge-aware smoothness loss ( L S )
  •         Compute enhanced consistency loss ( L C )
  •         Total loss: L = μ L P + ω L S + γ L C
  •         Update model parameters via backpropagation
  •     end while
  • else
  •     Inference Phase:
  •     Input: Single input image ( I )
  •     I ⇒ Predicted depth map ( D )
  •     Apply post-processing to convert predicted depth map to final depth map ( D final )
  •     Output: Final depth map ( D final )
  • end if

4. Experiments

4.1. Implementation Details

We implement our CIT-Depth model in PyTorch 1.8.0. The model is trained for 30 epochs on the KITTI dataset, with both the pose encoder and depth encoder initialized using ImageNet [35] pretrained weights. The model uses AdamW [36] as the optimizer, with a batch size of 12, an input image size of 640 × 192 , and a learning rate set to 10 4 . The experiments are conducted using a single RTX 4090 GPU (Nvidia: Santa Clara, CA, USA). For evaluation, this paper employs the seven standard metrics proposed in previous work [6].

4.2. Data Augmentation with GAN-Generated Images

To further enhance the model’s generalization ability, this paper uses GANs to generate image data under various weather and time conditions, including rainy days, dusk, and dawn. These synthetic images closely resemble real images and significantly broaden the diversity of the training dataset. By incorporating these complex scenarios into the original dataset, the model can learn a wider range of environmental features during training, resulting in greater robustness in diverse and complex real-world applications. Figure 6 shows some of the enhanced effects. The experimental results indicate that enriching the training data with GAN-generated images improves the model’s depth estimation performance in these specific scenarios.

4.3. Results on KITTI

The KITTI dataset [37] is a widely used benchmark in autonomous driving and computer vision research. It contains typical road scenes, including roads, buildings, trees, cars, and more. This paper uses the Eigen split [6], which includes 4424 images for evaluation and 39,810 images for training. For testing, 697 images containing real LiDAR data are used.
Table 1 shows the performance of existing self-supervised frameworks at a resolution of 640 × 192, comparing our model with other self-supervised methods under monocular training. Our method outperforms most comparable methods. Figure 7 shows the performance results of our method and other methods on the KITTI dataset, demonstrating that CIT-Depth performs better in various scenes. CIT-Depth demonstrates a more balanced performance across various scenes, particularly exhibiting smoother depth transitions in complex edges and occlusion areas. It also shows better continuity when handling distant backgrounds, avoiding abrupt depth changes or discontinuities. This indicates that CIT-Depth is well adapted to a variety of scenes.
Table 1. Results on the KITTI dataset using the Eigen split. All methods were pretrained on ImageNet. “M” denotes training with monocular video, while “M + Se” denotes training with stereo video. The optimal value for each metric is indicated in bold font.
Table 1. Results on the KITTI dataset using the Eigen split. All methods were pretrained on ImageNet. “M” denotes training with monocular video, while “M + Se” denotes training with stereo video. The optimal value for each metric is indicated in bold font.
MethodDataResolutionAbs RelSq RelRMSERMSE log δ 1 < 1.25 δ 2 < 1.25 2 δ 3 < 1.25 3
SfMLearner [10]M640 × 1920.1831.5956.7090.2700.7340.9020.959
Monodepth2-Res18 [11]M640 × 1920.1150.9034.8630.1930.8770.9590.981
Monodepth2-Res50 [11]M640 × 1920.1100.8304.6420.1890.8820.9610.981
SGDepth [12]M + Se640 × 1920.1130.8324.6910.1900.8800.9610.981
PackNet-SfM [23]M640 × 1920.1110.7844.6010.1890.8770.9600.982
HR-Depth [38]M640 × 1920.1090.7914.6330.1860.8840.9610.983
ADAADepth [39]M640 × 1920.1110.8154.6840.1870.8830.9610.982
BRNet [40]M640 × 1920.1050.6994.4650.1800.8880.9630.984
CADepth [41]M640 × 1920.1050.7654.5350.1810.8910.9640.983
DIFFNet [42]M640 × 1920.1020.7504.4470.1790.8960.9650.983
MonoViT [32]M640 × 1920.0990.7084.3740.1750.9000.9670.984
CIT-Depth (Ours)M640 × 1920.0970.6554.2140.1720.9020.9680.984
Figure 7. Qualitative results on the KITTI dataset. The first row shows the input image, followed by the predicted results of Monodepth2 [11], HR-Depth [38], DIFFNet [42], MonoViT [32], and CIT-Depth (ours).
Figure 7. Qualitative results on the KITTI dataset. The first row shows the input image, followed by the predicted results of Monodepth2 [11], HR-Depth [38], DIFFNet [42], MonoViT [32], and CIT-Depth (ours).
Sustainability 16 09131 g007

4.4. Results on Make3D

To evaluate the model’s generalization capability, we perform tests on the Make3D dataset [43], which includes 134 test images of different outdoor scenes. Compared to the KITTI dataset, the Make3D dataset features diverse structures and terrain variations, such as building facades, trees, and lawns. These scenes typically exhibit significant depth gradient changes, making them suitable for testing a depth estimation model’s ability to handle large-scale depth variations.
The model trained on the KITTI dataset is evaluated on the Make3D dataset without any further modifications. As shown in Table 2, the CIT-Depth model exhibits better adaptability to various scenes compared to other methods. Figure 8 presents the qualitative results, showing that our model produces clearer results in outdoor scenes. Other models perform poorly in certain areas, particularly around the edges of distant buildings or trees, where depth predictions exhibit discontinuities or blurriness. CIT-Depth provides more continuous and smooth depth predictions at the edges of distant buildings and trees, with more accurate edge handling. This indicates that the method has a stronger ability to handle scene details and large-scale depth variations.

4.5. Results on DrivingStereo

The DrivingStereo dataset [44] is a large-scale stereo vision dataset that includes images captured under various weather conditions, such as sunny, rainy, cloudy, and foggy weather. The dataset features a diverse range of scenes, including city roads, highways, and rural roads, making it highly valuable for addressing the diversity and complexity of real-world driving scenarios.
To verify the model’s generalization capability under various weather conditions, the model is tested on the DrivingStereo dataset. Table 3 summarizes the evaluation results of various state-of-the-art methods. All the models are trained solely on the KITTI dataset and tested directly under different weather conditions. CIT-Depth outperforms other methods in rainy, foggy, and cloudy conditions, demonstrating the robustness of the method. Figure 9 shows that while other models perform well under sunny and cloudy conditions, they exhibit significant depth-blurring issues in rainy and foggy conditions. In contrast, CIT-Depth effectively captures edge details and depth information in distant scenes, particularly under complex conditions such as rain and fog. It performs well across most weather conditions and demonstrates a strong performance in handling low contrast and visual noise, providing more accurate and consistent depth predictions for distant objects.

4.6. Results on Foggy CityScape

The Foggy CityScapes dataset [45], derived from the original CityScapes dataset [46], is specifically designed for researching the performance of autonomous driving systems in foggy weather conditions. By overlaying fog effects on the original CityScapes images, this dataset simulates varying levels of foggy weather, creating realistic foggy scenes. These images retain the complex structure of urban streets, including roads, buildings, pedestrians, and vehicles, but the contrast and clarity are significantly reduced due to the fog.
Table 4 shows that CIT-Depth also performs well on the foggy dataset. Figure 10 illustrates that, due to the effect of fog, models like Monodepth2 exhibit noticeable blurriness and discontinuities in depth estimation for distant objects. This is particularly evident when dealing with trees and buildings at a distance, where the model struggles to accurately predict depth. CIT-Depth outperforms other methods in foggy conditions, maintaining higher accuracy in depth estimation for distant objects, with the depth map displaying better smoothness and consistency. In regions with distant trees and buildings, edge handling is notably smoother. However, under extreme foggy conditions, the model still shows slight blurriness in depth prediction for distant objects.

4.7. Results on NuScene-Night

The NuScenes-Night dataset [47], derived from the NuScenes dataset, is specifically intended for researching autonomous driving perception tasks in nighttime conditions. The dataset includes a variety of nighttime driving scenes, covering city streets, suburban roads, and highways. The images were primarily collected under nighttime or dim lighting conditions, characterized by extremely low light and high-contrast areas.
Table 5 shows that our model maintains a strong performance even at night, outperforming some models specifically designed for nighttime data. As shown in Figure 11, Monodepth2 exhibits noticeable blurriness and discontinuities in depth predictions in poorly lit areas, especially at the edges of roads and for distant objects. Other models struggle to effectively handle the high noise and low contrast caused by insufficient lighting, leading to a loss of detail in the depth map. CIT-Depth exhibits better smoothness and continuity in low-light and high-contrast areas. Under extreme low-light conditions, the model encounters challenges in predicting the depth of distant objects, often leading to considerable blurring. This blurring is primarily caused by the low signal-to-noise ratio in images captured by the camera at night, where useful information is overwhelmed by noise. Although the proposed method demonstrates an improved performance at night, there remains an issue of insufficient information when handling fine-grained depth prediction for distant objects.
Figure 11. Qualitative results on the NuScene-Night dataset. The depth prediction results of Monodepth2 [11], ADDS-Depth [48], MonoViT [32], and CIT-Depth under nighttime conditions.
Figure 11. Qualitative results on the NuScene-Night dataset. The depth prediction results of Monodepth2 [11], ADDS-Depth [48], MonoViT [32], and CIT-Depth under nighttime conditions.
Sustainability 16 09131 g011
Table 5. Results on the NuScene-Night dataset. ADDS-DepthNet was trained on the RobotCar dataset [49]. The optimal value for each metric is indicated in bold font.
Table 5. Results on the NuScene-Night dataset. ADDS-DepthNet was trained on the RobotCar dataset [49]. The optimal value for each metric is indicated in bold font.
MethodAbs RelSq RelRMSERMSE log δ 1 < 1.25 δ 2 < 1.25 2 δ 3 < 1.25 3
Monodepth2 [11]0.3986.21014.5710.5680.3780.6500.794
HR-Depth [38]0.4606.63515.0280.6220.3050.5700.749
CADepth [41]0.4215.95014.5040.5930.3110.6130.776
DIFFNet [42]0.3444.85313.1540.4910.4400.7100.838
ADDS-Depth [48]0.3214.59412.9090.4790.4660.7110.840
MonoViT [32]0.3134.14412.2550.4560.4840.7360.858
CIT-Depth (Ours)0.3074.08011.5910.4310.5390.7810.863

4.8. Ablation Study

To evaluate the impact of the channel interaction module (CIM) and the Multi-Scale Fusion Module (MSFM) on model performance, we conducted ablation studies to verify the contributions of these two modules to the depth estimation task under complex weather conditions. First, we used a baseline model, then incrementally added the CIM and MSFM, and finally combined both modules. We tested different model configurations on multiple datasets, encompassing various weather conditions such as sunny, rainy, and foggy scenarios. The results in Table 6 indicate that the addition of these two modules yields limited improvement on the KITTI dataset but demonstrates noticeable enhancement on other datasets. The channel interaction module enhances interaction between feature channels, enabling the model to more effectively identify key features in the scene. The Multi-Scale Fusion Module plays a crucial role in fusing multi-scale features, enhancing the model’s ability to capture fine details. When both the CIM and MSFM are incorporated, the overall model performance reaches its optimum, significantly reducing errors and improving accuracy metrics, particularly under extreme weather conditions. This indicates that the two modules exhibit strong synergy in handling diverse weather variations, jointly enhancing the model’s depth estimation accuracy and robustness.

5. Conclusions

This paper proposed CIT-Depth, a new self-supervised monocular depth estimation network suitable for various weather conditions. The network backbone combines convolutional modules and Transformer modules and incorporates additional weather images during training. Channel fusion modules are added during the encoding and decoding stages to enhance the model’s adaptability to depth estimation in different scenarios. Our method achieves excellent results on the KITTI dataset. Experiments on the Foggy CityScape, DrivingStereo, and NuScene-Night datasets demonstrate that our model outperforms state-of-the-art architectures under various conditions (rain, fog, cloudy, and night). The proposed method for depth prediction of distant objects under extreme low-light conditions has room for improvement. In the future, integrating light enhancement techniques or specifically optimizing the feature extraction module will be necessary to address the challenge of insufficient information in complex environments. Although the robustness of the CIT-Depth model was validated under various weather conditions, future work will focus on conducting additional tests on datasets from diverse geographic regions and more extreme environmental variations to provide a more comprehensive evaluation of the model’s performance. To provide a more comprehensive evaluation of the model’s performance, future work will focus on conducting additional tests on datasets from diverse geographic regions and extreme environmental conditions. We will further optimize computational efficiency through techniques such as pruning and quantization to reduce model complexity while maintaining performance, making it more suitable for deployment on embedded devices.

Author Contributions

J.L.: draft preparation, model development, and manuscript revision; Z.G.: editing, software, and data curation; P.P.: overall planning, draft preparation, and manuscript review; H.Z.: supervision, project administration, and conceptualization; Q.S.: supervision, project administration, investigation, and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grant 52202496 and 62476145, the Natural Science Foundation of the Higher Education Institutions of Jiangsu Province under grant 22KJB510040, and the Nantong social livelihood science and technology project under grant MS12022015.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The KITTI public dataset in this paper is available at http://www.cvlibs.net/datasets/kitti/ (accessed on 18 September 2024). The Make3D public dataset in this paper is available at http://make3d.cs.cornell.edu (accessed on 18 September 2024). The DrivingStereo public dataset in this paper is available at http://drivingstereo-dataset.github.io (accessed on 18 September 2024). The Foggy CityScape public dataset in this paper is available at https://www.cityscapes-dataset.com/benchmarks/#foggy-cityscapes (accessed on 18 September 2024). The NuScene-Night public dataset in this paper is available at https://www.nuscenes.org/nuscenes (accessed on 18 September 2024).

Conflicts of Interest

Author Hao Zhang was employed by Henan Airport Group. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Afshar, M.F.; Shirmohammadi, Z.; Ghahramani, S.A.A.G.; Noorparvar, A.; Hemmatyar, A.M.A. An Efficient Approach to Monocular Depth Estimation for Autonomous Vehicle Perception Systems. Sustainability 2023, 15, 8897. [Google Scholar] [CrossRef]
  2. Ebner, L.; Billings, G.; Williams, S. Metrically scaled monocular depth estimation through sparse priors for underwater robots. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024; pp. 3751–3757. [Google Scholar]
  3. Jia, Q.; Chang, L.; Qiang, B.; Zhang, S.; Xie, W.; Yang, X.; Sun, Y.; Yang, M. Real-time 3D reconstruction method based on monocular vision. Sensors 2021, 21, 5909. [Google Scholar] [CrossRef] [PubMed]
  4. Vaswani, A. Attention is all you need. In Proceedings of the International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
  5. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  6. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. In Proceedings of the International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 2014; pp. 2366–2374. [Google Scholar]
  7. Laina, I.; Rupprecht, C.; Belagiannis, V.; Tombari, F.; Navab, N. Deeper depth prediction with fully convolutional residual networks. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 239–248. [Google Scholar]
  8. Fu, H.; Gong, M.; Wang, C.; Batmanghelich, K.; Tao, D. Deep ordinal regression network for monocular depth estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2002–2011. [Google Scholar]
  9. Bhat, S.F.; Alhashim, I.; Wonka, P. Adabins: Depth estimation using adaptive bins. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 4009–4018. [Google Scholar]
  10. Zhou, T.; Brown, M.; Snavely, N.; Lowe, D.G. Unsupervised learning of depth and ego-motion from video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1851–1858. [Google Scholar]
  11. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3828–3838. [Google Scholar]
  12. Klingner, M.; Termöhlen, J.-A.; Mikolajczyk, J.; Fingscheidt, T. Self-supervised monocular depth estimation: Solving the dynamic object problem by semantic guidance. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 582–600. [Google Scholar]
  13. Godard, C.; Mac Aodha, O.; Brostow, G.J. Unsupervised monocular depth estimation with left-right consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 270–279. [Google Scholar]
  14. Garg, R.; Vijay Kumar, B.G.; Carneiro, G.; Reid, I. Unsupervised CNN for single view depth estimation: Geometry to the rescue. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 740–756. [Google Scholar]
  15. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, H.; Zhu, Y.; Adam, H.; Yuille, A.; Chen, L.-C. Max-deeplab: End-to-end panoptic segmentation with mask transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 5463–5474. [Google Scholar]
  17. Zhao, C.; Tang, Y.; Sun, Q. Unsupervised monocular depth estimation in highly complex environments. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 6, 1237–1246. [Google Scholar] [CrossRef]
  18. Gasperini, S.; Morbitzer, N.; Jung, H.; Navab, N.; Tombari, F. Robust monocular depth estimation under challenging conditions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 8177–8186. [Google Scholar]
  19. Spencer, J.; Bowden, R.; Hadfield, S. Defeat-net: General monocular depth via simultaneous unsupervised representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 14402–14413. [Google Scholar]
  20. Choi, J.; Jung, D.; Lee, D.; Kim, C. Safenet: Self-supervised monocular depth estimation with semantic-aware feature extraction. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
  21. Wang, K.; Zhang, Z.; Yan, Z.; Li, X.; Xu, B.; Li, J.; Yang, J. Regularizing nighttime weirdness: Efficient self-supervised monocular depth estimation in the dark. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 16055–16064. [Google Scholar]
  22. Yin, Z.; Shi, J. Geonet: Unsupervised learning of dense depth, optical flow and camera pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1983–1992. [Google Scholar]
  23. Guizilini, V.; Ambrus, R.; Pillai, S.; Raventos, A.; Gaidon, A. 3D packing for self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2485–2494. [Google Scholar]
  24. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  25. Varma, A.; Chawla, H.; Zonooz, B.; Arani, E. Transformers in self-supervised monocular depth estimation with unknown camera intrinsics. arXiv 2022, arXiv:2202.03131. [Google Scholar]
  26. Lasinger, K.; Ranftl, R.; Schindler, K.; Koltun, V. Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. arXiv 2019, arXiv:1907.01341. [Google Scholar]
  27. Li, Z.; Chen, Z.; Liu, X.; Jiang, J. Depthformer: Exploiting long-range correlation and local information for accurate monocular depth estimation. Mach. Intell. Res. 2023, 20, 837–854. [Google Scholar] [CrossRef]
  28. Hwang, S.-J.; Park, S.-J.; Baek, J.-H.; Kim, B. Self-supervised monocular depth estimation using hybrid transformer encoder. IEEE Sens. J. 2022, 22, 18762–18770. [Google Scholar] [CrossRef]
  29. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  30. Pizzati, F.; Cerri, P.; De Charette, R. CoMoGAN: Continuous model-guided image-to-image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 14288–14298. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Zhao, C.; Zhang, Y.; Poggi, M.; Tosi, F.; Guo, X.; Zhu, Z.; Huang, G.; Tang, Y.; Mattoccia, S. Monovit: Self-supervised monocular depth estimation with a vision transformer. In Proceedings of the 2022 International Conference on 3D Vision (3DV), Prague, Czech Republic, 17–19 October 2022; pp. 668–678. [Google Scholar]
  33. Lee, Y.; Kim, J.; Willette, J.; Hwang, S.J. MPViT: Multi-path vision transformer for dense prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 7287–7296. [Google Scholar]
  34. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  35. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  36. Loshchilov, I. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  37. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The KITTI dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  38. Lyu, X.; Liu, L.; Wang, M.; Kong, X.; Liu, L.; Liu, Y.; Chen, X.; Yuan, Y. HR-Depth: High resolution self-supervised monocular depth estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, Number 3. pp. 2294–2301. [Google Scholar]
  39. Kaushik, V.; Jindgar, K.; Lall, B. ADAADepth: Adapting data augmentation and attention for self-supervised monocular depth estimation. IEEE Robot. Autom. Lett. 2021, 6, 7791–7798. [Google Scholar] [CrossRef]
  40. Han, W.; Yin, J.; Jin, X.; Dai, X.; Shen, J. Brnet: Exploring comprehensive features for monocular depth estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–28 August 2022; pp. 586–602. [Google Scholar]
  41. Yan, J.; Zhao, H.; Bu, P.; Jin, Y. Channel-wise attention-based network for self-supervised monocular depth estimation. In Proceedings of the 2021 International Conference on 3D Vision (3DV), Virtual, 18–21 October 2021; pp. 464–473. [Google Scholar]
  42. Zhou, H.; Greenwood, D.; Taylor, S. Self-supervised monocular depth estimation with internal feature fusion. In Proceedings of the British Machine Vision Conference (BMVC), London, UK, 22–25 November 2021. [Google Scholar]
  43. Saxena, A.; Sun, M.; Ng, A.Y. Make3d: Learning 3d scene structure from a single still image. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 31, 824–840. [Google Scholar] [CrossRef] [PubMed]
  44. Yang, G.; Song, X.; Huang, C.; Deng, Z.; Shi, J.; Zhou, B. DrivingStereo: A large-scale dataset for stereo matching in autonomous driving scenarios. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 899–908. [Google Scholar]
  45. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef]
  46. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  47. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11621–11631. [Google Scholar]
  48. Liu, L.; Song, X.; Wang, M.; Liu, Y.; Zhang, L. Self-supervised monocular depth estimation for all day images using domain separation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12737–12746. [Google Scholar]
  49. Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 year, 1000 km: The Oxford RobotCar dataset. Int. J. Robot. Res. 2017, 36, 3–15. [Google Scholar] [CrossRef]
Figure 1. Overview of our CIT-Depth architecture. Our CIT-Depth is composed of two parts: Depth Network and Pose Network [11]. The Depth Network utilizes both convolutional layers and Transformer architecture.
Figure 1. Overview of our CIT-Depth architecture. Our CIT-Depth is composed of two parts: Depth Network and Pose Network [11]. The Depth Network utilizes both convolutional layers and Transformer architecture.
Sustainability 16 09131 g001
Figure 2. The depth encoder employs CNN and Transformer module. Each module consists of a convolutional module and three Transformer modules.
Figure 2. The depth encoder employs CNN and Transformer module. Each module consists of a convolutional module and three Transformer modules.
Sustainability 16 09131 g002
Figure 3. Structures of channel interaction module.
Figure 3. Structures of channel interaction module.
Sustainability 16 09131 g003
Figure 4. Structures of Global Channel Pooling module.
Figure 4. Structures of Global Channel Pooling module.
Sustainability 16 09131 g004
Figure 5. Structure of Multi-Scale Fusion Module.
Figure 5. Structure of Multi-Scale Fusion Module.
Sustainability 16 09131 g005
Figure 6. The enhanced images generated by GAN under various weather and time conditions are displayed in the following order from top to bottom: original input image, mixed exposure image, rainy image, foggy image, and nighttime image.
Figure 6. The enhanced images generated by GAN under various weather and time conditions are displayed in the following order from top to bottom: original input image, mixed exposure image, rainy image, foggy image, and nighttime image.
Sustainability 16 09131 g006
Figure 8. Qualitative results on the Make3D dataset. CIT-Depth is compared with the predicted results of Monodepth2 [11] and MonoViT [32].
Figure 8. Qualitative results on the Make3D dataset. CIT-Depth is compared with the predicted results of Monodepth2 [11] and MonoViT [32].
Sustainability 16 09131 g008
Figure 9. A demonstration of the qualitative results from the DrivingStereo dataset. From left to right are images from the dataset representing sunny, rainy, foggy, and cloudy conditions.
Figure 9. A demonstration of the qualitative results from the DrivingStereo dataset. From left to right are images from the dataset representing sunny, rainy, foggy, and cloudy conditions.
Sustainability 16 09131 g009
Figure 10. Qualitative results on the Foggy CityScape dataset. The depth prediction results of Monodepth2 [11], DIFFNet [42], MonoViT [32], and CIT-Depth under foggy conditions.
Figure 10. Qualitative results on the Foggy CityScape dataset. The depth prediction results of Monodepth2 [11], DIFFNet [42], MonoViT [32], and CIT-Depth under foggy conditions.
Sustainability 16 09131 g010
Table 2. Results on the Make3D dataset. Models trained on KITTI with 640 × 192 images. The optimal value for each metric is indicated in bold font.
Table 2. Results on the Make3D dataset. Models trained on KITTI with 640 × 192 images. The optimal value for each metric is indicated in bold font.
MethodAbs RelSq RelRMSERMSE log
Monodepth2 [11]0.3213.3777.4170.164
HR-Depth [38]0.3153.2087.0310.155
CADepth [41]0.3183.2237.1510.158
DIFFNet [42]0.2992.9106.7600.153
MonoViT [32]0.2862.7596.6250.147
CIT-Depth (Ours)0.2752.6396.4020.144
Table 3. All models were trained on the KITTI dataset and tested under four weather conditions (foggy, cloudy, rainy, and sunny). The optimal value for each metric is indicated in bold font.
Table 3. All models were trained on the KITTI dataset and tested under four weather conditions (foggy, cloudy, rainy, and sunny). The optimal value for each metric is indicated in bold font.
DomainMethodAbs RelSq RelRMSERMSE log δ 1 < 1.25 δ 2 < 1 . 25 2 δ 3 < 1 . 25 3
foggyMonodepth2 [11]0.1431.9549.8180.2180.8120.9360.974
HR-Depth [38]0.1321.8269.5870.1980.8260.9490.982
CADepth [41]0.1411.7799.4500.2080.8110.9450.981
DIFFNet [42]0.1261.5628.7240.1890.8390.9560.985
MonoViT [32]0.1061.1557.2560.1610.8710.9700.990
CIT-Depth (Ours)0.1041.1477.2130.1590.8800.9750.992
cloudyMonodepth2 [11]0.1551.9026.9770.2090.8120.9430.979
HR-Depth [38]0.1481.6576.6590.2040.8150.9450.981
CADepth [41]0.1481.8056.7120.2050.8290.9470.981
DIFFNet [42]0.1401.5726.2980.1920.8370.9500.983
MonoViT [32]0.1351.4696.0950.1830.8570.9550.985
CIT-Depth (Ours)0.1331.4565.9120.1800.8600.9570.986
rainyMonodepth2 [11]0.2403.33911.0420.3010.5900.5870.953
HR-Depth [38]0.2222.96210.4950.2810.6310.8690.959
CADepth [41]0.2263.01510.8250.2870.6290.8510.956
DIFFNet [42]0.1922.4119.6260.2460.6770.9140.969
MonoViT [32]0.1742.1329.4900.2310.7280.9280.976
CIT-Depth (Ours)0.1702.0159.0230.2200.7360.9350.979
sunnyMonodepth2 [11]0.1782.1058.2090.2400.7820.9250.968
HR-Depth [38]0.1641.8397.8900.2270.7940.9360.975
CADepth [41]0.1621.7557.6890.2210.8010.9360.974
DIFFNet [42]0.1501.6167.5800.2100.8120.9400.978
MonoViT [32]0.1421.4577.0070.1990.8320.9480.981
CIT-Depth (Ours)0.1431.4597.0090.1990.8330.9490.981
Table 4. Results on the Foggy CityScape dataset. All models were trained on the KITTI dataset. The optimal value for each metric is indicated in bold font.
Table 4. Results on the Foggy CityScape dataset. All models were trained on the KITTI dataset. The optimal value for each metric is indicated in bold font.
MethodAbs RelSq RelRMSERMSE log δ 1 < 1.25 δ 2 < 1.25 2 δ 3 < 1.25 3
Monodepth2 [11]0.2083.09512.4490.3370.6560.8420.917
HR-Depth [38]0.2133.01512.2670.3360.6420.8410.920
CADepth [41]0.2072.73811.5500.3180.6500.8560.933
DIFFNet [42]0.1872.58311.3370.3040.6890.8670.937
MonoViT [32]0.1551.8739.5850.2440.7710.9100.967
CIT-Depth (Ours)0.1511.6898.0920.2290.7980.9390.970
Table 6. Performance evaluation of different model configurations across various datasets. The baseline model is compared with configurations using CIM, MSFM, and their combination.
Table 6. Performance evaluation of different model configurations across various datasets. The baseline model is compared with configurations using CIM, MSFM, and their combination.
DatasetModel ConfigurationAbs RelSq RelRMSERMSE log δ 1 < 1.25 δ 2 < 1 . 25 2 δ 3 < 1 . 25 3
KITTIBaseline0.0990.7014.4090.1770.8960.9650.983
+CIM0.0980.6894.2980.1740.8980.9670.984
+MSFM0.0980.6984.3410.1750.8990.9670.984
+CIM+MSFM0.0970.6554.2140.1720.9020.9680.984
DrivingStereo (foggy)Baseline0.1081.2017.3100.1650.8650.9670.990
+CIM0.1051.1557.2590.1630.8690.9690.991
+MSFM0.1051.1507.2550.1620.8710.9700.991
+CIM+MSFM0.1041.1477.2130.1590.8800.9750.992
DrivingStereo (rainy)Baseline0.1752.1599.4920.2320.7250.9260.976
+CIM0.1722.1229.2510.2250.7320.9310.978
+MSFM0.1732.1389.3190.2300.7290.9290.978
+CIM+MSFM0.1702.0159.0230.2200.7360.9350.979
Foggy CityScapeBaseline0.1561.8779.5920.2470.7700.9090.966
+CIM0.1521.7118.3010.2350.7850.9280.968
+MSFM0.1541.8519.2040.2400.7780.9160.968
+CIM+MSFM0.1511.6898.0920.2290.7980.9390.970
NuScene-NightBaseline0.3154.14812.2600.4570.4820.7390.859
+CIM0.3104.09811.8030.4400.5190.7690.862
+MSFM0.3124.12112.0090.4540.5010.7520.860
+CIM+MSFM0.3074.08011.5910.4310.5390.7810.863
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Guo, Z.; Ping, P.; Zhang, H.; Shi, Q. Channel Interaction and Transformer Depth Estimation Network: Robust Self-Supervised Depth Estimation Under Varied Weather Conditions. Sustainability 2024, 16, 9131. https://doi.org/10.3390/su16209131

AMA Style

Liu J, Guo Z, Ping P, Zhang H, Shi Q. Channel Interaction and Transformer Depth Estimation Network: Robust Self-Supervised Depth Estimation Under Varied Weather Conditions. Sustainability. 2024; 16(20):9131. https://doi.org/10.3390/su16209131

Chicago/Turabian Style

Liu, Jianqiang, Zhengyu Guo, Peng Ping, Hao Zhang, and Quan Shi. 2024. "Channel Interaction and Transformer Depth Estimation Network: Robust Self-Supervised Depth Estimation Under Varied Weather Conditions" Sustainability 16, no. 20: 9131. https://doi.org/10.3390/su16209131

APA Style

Liu, J., Guo, Z., Ping, P., Zhang, H., & Shi, Q. (2024). Channel Interaction and Transformer Depth Estimation Network: Robust Self-Supervised Depth Estimation Under Varied Weather Conditions. Sustainability, 16(20), 9131. https://doi.org/10.3390/su16209131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop