4.1. Settings
Experimental condition: The experimental core software contains PyTorch 1.71, and CUDA10.2, and the core hardware contains Nvidia Titan V, whose performance is comparable to that of the NVIDIA Quadro P6000 GPU.
Datasets: Rain1400 [
34], Rain100H [
21], Real-world SPA dataset [
14], and Raindrop dataset [
42] are used to train and test related models, while each high real-time model trained in one dataset above is retested in the other three datasets, whose results directly demonstrate the generalization ability.
Training tricks: We train our model proposed 5000 epochs with a batch size of 16 in Rain100H, 1200 epochs with a batch size of 16 in Rain1400, 5000 epochs with a batch size of 16 in Raindrops, 15 epochs with a batch size of 12 in Real-world dataset, respectively. The optimiser is Adam, and the learning rate is controlled by an adjusting function, in which no changes happen before the half-training. Then the learning rate slowly decreases in a straight line to zero at the end. There is no data augmentation and fancy training tricks, and the training tricks are simple.
Metrics: The metrics contain objective and subjective evaluations. The objective metrics are
and
, representing the performance of deraining models in specific numerical quantification.
and
are popular and accepted, and the larger the two values are, the better the comparison results. The
calculation is shown in Equation (
4). The subjective metrics [
43], taken as an assistant way to further distinguish the results by human eyes, are evaluations of colour changes, preserved detailed features, and deformations, which should be considered because adverse effects should be carefully restrained accompanying the deraining processing.
where
is the maximum value of the pixels.
4.2. Comparison on Public Datasets
We have compared DAMNet with several classic deraining methods on Rain1400 [
34], Rain100H [
21], and Real-world SPA [
14]. All testing results are obtained after acceleration from a single GPU with comparable performance. We divide these algorithms into three categories, high real-time (1000/fps
10 ms), real-time (10 < 1000/fps
40 ms), and low real-time (1000/fps > 40 ms) kinds, based on the average time used of all derained images.
The results comparison on Rain100H is shown in
Table 1. Although our deraining model is a relatively simple and model-free network, DAMNet gains parallel performance compared with the RCDNet [
10] which is an elaborate and intricate model built on inner dictionaries of the rain traces. Owing to avoiding the repetitive multi-stages and utilising the pixelshuffle structure of head networks, DAMNet achieves a high real-time inference, far faster than RCDNet, while maintaining impressive deraining effects. In contrast, EfDeRain [
23] has a comparable inference speed compared with our model, while DAMNet achieves better performance because of the adoption of joint training based on multi-loss functions. Although the five deraining layers lead to more inference burdens than the one deraining layer of EfDeRain, our deraining models maintain a high real-time speed by taking advantage of the dual adjacent method. In addition, we confirm that deraining models for a single image have a low inference speed even after GPU prompts. Because deraining methods utilised for a single image belong to the low real-time series, when the specific inference speed is not confirmed, the spent time per image is approximately assessed by “>40 ms”. Compared with the deraining models for a single image [
11,
12,
14,
21], DAMNet achieves the fastest inference speed while gaining impressive results, suggesting our muti-deraining method has practical effects despite the adoption of relatively simple networks. The visible comparison is shown in
Figure 5, DAMNet achieves sufficient deraining effects while persevering visible detailed features. Overall, our deraining model better balances the inference speed and performance on Rain100H [
21], whose rain backgrounds are more complex than those of Rain100L [
21].
The performances on Rain1400H are shown in
Table 2. MHEB [
13] achieves the best performance in
Table 2, while the corresponding network neglects the computational burden and only considers the deraining effect for a single image, leading to a low inference speed. Conversely, DAMNet better balances deraining effects and inference speed with a model-free structure. Although the performance of our model outperforms EfDeRain by a narrow margin on the Rain1400 [
34], the practical deraining effect of DAMNet is more visible and the restriction for adverse effects is recognisable, as shown in
Figure 6. Ears colour changes and deformation appear in the results of most deraining models, but the deraining effect of DAMNet is satisfactory while the adverse effects are restrained very well.
The demonstrations on the Realworld dataset are shown in
Table 3. Although achieving the best deraining effect, RCDNet [
10] removes the rain streaks with a low inference speed even after the acceleration of the GPU. DAMNet achieves a parallel performance compared with MHEB [
13] and balances the performance and inference speed, demonstrating the advantage of the structure of DAMNet. Although outperforming EfDeRain [
23] by a narrow margin, only considering the values of
and
, DAMNet performs better in the constraint of adverse deraining effects, as shown in
Figure 7. Although relatively thicker greens are visually better, whether to add a colour attribute that does not exist is another manifestation of the algorithm’s ability. From
Figure 7, DAMNet best maintains the original colour.
The practical effects of raindrop removal are shown in
Table 4. Raindrops are different from Rain100H, Rain1400, and Realworld. Except for the DAMNet we proposed and EfDeRain [
23], the models in
Table 4 are specific algorithms for removing raindrops. DAMNet and DfDeRain are directly applied and validated in Raindrops while achieving satisfactory effects, even better than the deraining effects of JORDERE [
21] and EIGEN [
44]. Although the specific inference speed in our platform can not be confirmed, we assign them to “>40 ms” considering their complex structures and the time consumed in public papers. Compared with EfDeRain, one of the high real-time kinds of
Table 4, DAMNet achieves a better balance between performance and inference speed. The removal raindrop effects of DAMNet and EfDeRain [
23] can demonstrate the model-free structures, avoiding multi-strategies for elaborate assumptions, have a strong generalisation ability and enough free adjustable space to fit the degraded pattern. Although belonging to the model-free method, DAMNet achieves higher values of
and
than EfDeRain [
23] which adopts the “one-off” loss method. Besides, DAMNet curbs the side effects at full steam and demonstrates satisfactory generalisation ability, as shown in
Figure 8.
Comprehensively, DAMNet reaches the degree of high real-time speed in all four datasets while maintaining a stable performance. Although five deraining layers acquire more computational cost than EfDeRain, which only adopts one deraining layer, DAMNet utilises the pixelshuffle operation to form a fast U-Net to maintain the parallel inference speed with EfDeRain. Our deraining model reaches one of the most efficient models in single datasets in which the training and testing images are different but stem from the same datasets. The comparison of visualisation on single datasets is also given in
Figure 9. The deraining effects of DAMNet outperform EfDeRain, while the adverse deraining effects of DAMNet are better restrained.
Different datasets have different properties, and the method of training and testing in different or crossing datasets is a practical way to demonstrate the generalisation of the high real-time deraining models. Therefore, we compare DAMNet with EfDeRain in the crossing datasets in which the training and testing images are derived from different datasets. The generalisation of our model outperforms EfDeRain [
23] by an impressive margin, as shown in
Table 5. The models’ suffixes are datasets for training themselves. The values highlighted in black represent that DAMNet outperforms EfDeRain under correspondingly identical training conditions. DAMNet achieves the higher mean value of
in all crossing datasets while keeping an overall advantage by measuring
. Comprehensively, our deraining model performs relatively stably, which is why we adopt the dual adjacent indexing and the joint training methods. Although the multi-deraining layers lead to affordable computing burdens compared with one deraining layer of EfDeRain, the generalisation ability of DAMNet has practically improved, demonstrating our deraining method is a more efficient model-free algorithm. The visualisation of the results on crossing datasets is shown in
Figure 10. Although the derained images have visible rain streaks, the more obvious tendency to remove rain tracks is visible.
4.3. Ablation Study
Rain100H is the dataset most severely disturbed by rain streaks, compared with Rain1400 and Realworld, which can be reflected in
Figure 5,
Figure 6 and
Figure 7. The ablations of DAMNet on Rain100H can evaluate the relationship between the dual adjacent method and the joint training method of computing the loss of each deraining layer. Experimental ablations on Rain100H are shown in
Table 6. The results of DAMNetfinalLoss are lower than DAMNet, which represents the “one-off” loss computation that definitely weakens the deraining effects and is also why DAMNet outperforms EfDeRain in single and crossing datasets. Comparing the results of DAMNet_noDual and DAMNet, we can confirm that the dual adjacent method can offer more useful information than the direct reshaped operations. The only participation of all deraining layers rarely maintains DAMNet performance when the dual adjacent method is absent. Compared with DAMNet04 saving both the first and final deraining layers, the models just maintain first or final deraining layers have an impressive decrease in deraining effects, dropping approximately 38.2% for
, which denotes the combination of first and final deraining layers is critically important for the deraining effect of DAMNet. Considering the comparison of DAMNet04 and DAMNet, the abandonment of middle deraining layers weakens the performance, whose phenomenon occurs in DAMNet_noMSE and DAMNet_noSSIMtotal. The
gap between DAMNet_noDual and DAMNet reaches more than 1%, which is larger than the comparison in
Table 1,
Table 2,
Table 3 and
Table 4, denoting the dual adjacent method is also nonnegligible for improving the performance. Comprehensively, DAMNet has the best performance, indicating that each deraining layer and the dual adjacent method are indispensable to gaining the efficiency and robustness of DAMNet, and suggesting the joint training method of multi-loss on the combination of multi-deraining layers is critical to prompt the performance.