Next Article in Journal
K-Band Low Phase Noise VCO Based on Q-Boosted Switched Inductor
Next Article in Special Issue
Real-Time Object Detection in Remote Sensing Images Based on Visual Perception and Memory Reasoning
Previous Article in Journal
A New Quadratic Binary Harris Hawk Optimization for Feature Selection
Previous Article in Special Issue
Efficient Implementation of 2D and 3D Sparse Deconvolutional Neural Networks with a Uniform Architecture on FPGAs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Wildfire Smoke Images Based on a Densely Dilated Convolutional Network

1
School of Technology, Beijing Forestry University, Beijing 100083, China
2
Key Lab of State Forestry and Grassland Administration for Forestry Equipment and Automation, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
Electronics 2019, 8(10), 1131; https://doi.org/10.3390/electronics8101131
Submission received: 29 August 2019 / Revised: 3 October 2019 / Accepted: 5 October 2019 / Published: 7 October 2019

Abstract

:
Recently, many researchers have attempted to use convolutional neural networks (CNNs) for wildfire smoke detection. However, the application of CNNs in wildfire smoke detection still faces several issues, e.g., the high false-alarm rate of detection and the imbalance of training data. To address these issues, we propose a novel framework integrating conventional methods into CNN for wildfire smoke detection, which consisted of a candidate smoke region segmentation strategy and an advanced network architecture, namely wildfire smoke dilated DenseNet (WSDD-Net). Candidate smoke region segmentation removed the complex backgrounds of the wildfire smoke images. The proposed WSDD-Net achieved multi-scale feature extraction by combining dilated convolutions with dense block. In order to solve the problem of the dataset imbalance, an improved cross entropy loss function, namely balanced cross entropy (BCE), was used instead of the original cross entropy loss function in the training process. The proposed WSDD-Net was evaluated according to two smoke datasets, i.e., WS and Yuan, and achieved a high AR (99.20%) and a low FAR (0.24%). The experimental results demonstrated that the proposed framework had better detection capabilities under different negative sample interferences.

1. Introduction

Wildfires not only destroy the natural ecological environment, but also threaten human safety and property [1]. Since image-based fire detection effectively reduces outside interference compared to the currently available sensors, image-based fire detection has become a hot topic in modern wildfire alarm systems [2]. Fire is often accompanied by smoke, which is emitted faster than flames. Therefore, smoke detection is an effective way to recognize potential fire disasters at the beginning of a breakout.
Existing methods regarding automatic smoke detection can be itemized into two categories, namely conventional detection approaches based on shallow machine learning [3,4,5], and deep learning methods based on deep neural networks [6]. Conventional detection approaches generally use handcrafted smoke features to train the classifiers (e.g., K-nearest-neighbor (KNN), support vector machine (SVM), AdaBoost) in the training and test scenes. Deep learning-based methods have been widely applied to visual detection tasks, such as wildlife identification [7], disease detection [8], pedestrian detection [9], etc. Extensive studies on smoke detection using convolutional neural networks (CNNs) have been shown to be capable of learning representative and essential features of smoke images and have presented ideal detection accuracy [10]. Due to the complex backgrounds of wildfire smoke images, the accuracy of many existing image-based smoke detection techniques is still inconsistent. In previous studies that used tailored smoke images for training, there would have been errors when testing the entire image.
Recent progress regarding smoke image detection, such as background processing [11], demonstrated that CNNs could deal with complex background interference and achieve high accuracy. However, it is still infeasible to apply these technologies directly to wildfire smoke detection. In order to analyze the interference of smoke-colored regions like cloud, fog, and haze on wildfire smoke detection, images with pedestrians and vehicles are not used as non-smoke samples. Smoke-colored regions resulting from fog, cloud, or haze are similar to smoke in color, texture, and shape features. Therefore, it is difficult to discriminate smoke from smoke-colored regions, thereby causing a high false-alarm rate. Another crucial issue is the lack of wildfire smoke images. There are not many wildfires and so wildfire smoke images cannot be collected often. The lack of wildfire smoke images leads to a serious training dataset imbalance, which results in over-fitting.
To address the above challenges, we propose a novel framework integrating conventional methods into CNN for wildfire smoke detection, consisting of a candidate smoke region segmentation strategy and advanced network architecture, namely wildfire smoke dilated DenseNet (WSDD-Net). In order to ensure that color information was not affected by illumination changes, RGB images were first converted into the YUV color space. Then, candidate smoke regions were segmented from wildfire smoke images using a segmentation strategy, which is defined in Section 3, and the segmented images were directly trained in WSDD-Net. A balanced cross entropy (BCE) loss function was also employed during training according to the proportion of smoke images in the training data, in order to weaken the imbalanced classes problem.
The rest of the paper is organized as follows. Section 2 gives a short summary of fire detection and smoke detection. The detailed segmentation strategy and WSDD-Net architecture is described in Section 3. Section 4 presents our evaluation methods and results compared with some other CNN architectures. Section 5 presents the conclusions of this paper.

2. Related Work

Fire detection-related studies have aroused much more attention in recent years. Yang et al. [12] established a model based on the Levenberg–Marquardt back propagation neutral network to recognize fire status. In Zhang, Shen, and Zou [13], a video-based fire recognition probability method based on color and motion features was presented. Smoke detection is another method of fire detection, which aims to improve the detection rate and reliability. Tung et al. [14] employed a median method combined with a fuzzy c-means method to segment moving regions and cluster candidate smoke regions from moving regions. Jia et al. [15] proposed a method of segmenting smoke regions based on saliency detection. Yuan [16] used a histogram sequence of local binary pattern (LBP) and local binary pattern variance (LBPV) pyramids to detect smoke.
All of the above detection algorithms mainly used handcrafted features and employed shallow machine-learning methods. A cost-effective smoke detection framework based on deep learning was proposed by Muhammad et al. [17] and focused on detection accuracy in complex backgrounds. In Mao, Wang, Dou, and Li [18], a smoke detection method based on multi-channel CNNs with an accuracy of 98% was introduced. Sharma et al. [19] used two pre-trained CNNs, namely VGG-16 and ResNet-50, to train an imbalanced smoke image dataset for fire detection. In order to avoid over-fitting due to lack of smoke images and imbalanced training data, Xu et al. [20] used synthetic smoke images to expand the smoke dataset and domain adaption to train the CNN structures. Zhang et al. [21] used synthetic wildfire smoke images to train an end-to-end object detection framework and a faster RCNN, which demonstrated the feasibility of synthetic smoke images to expand the dataset. However, this method of expanding datasets with synthetic wildfire smoke images consumes manpower, and it takes an excessive amount of time for the data size to meet the training requirements.
Recent works have shown that if CNNs contain connections between the previous output and input layers, the CNNs can be deeper, more accurate, and more efficient for training. Huang et al. [22] solved this problem and proposed the DenseNet, which connected each layer to every other layer in a feed-forward manner. Compared with classical deep learning methods, DenseNet not only performs better regarding wildfire smoke detection, but also achieves better results in false-alarm rate and accuracy. It alleviates the problem of the vanishing-gradient, enhances feature propagation, and reduces the number of parameters and training time by feature reuse.
Almost all of the models used the cross entropy loss function, but it is easy to over-fit when there is a dataset class imbalance. The focal loss function was proposed to solve this issue [23]. Abulnaga and Rubin [24] developed fully convolutional neural network models for the segmentation of ischemic stroke lesions, and made use of the focal loss function, which demonstrated the ability to identify finer features by focusing on hard-to-classify examples.
Inspired by recent use of focal loss in image-based object detection, we improved the cross entropy loss function to be suitable for wildfire smoke detection. We also added negative samples, such as cloud, fog, and haze, as interference to our dataset. Unlike the aforementioned works, our network directly trained candidate smoke regions to eliminate the interference of complex backgrounds in the feature extraction process. In addition, we attempted to address the problems of high false-alarm rates and imbalanced training sets simultaneously.

3. Methods

Our goal was to reduce the false-alarm rate in detecting wildfire smoke, alleviate the imbalance of the wildfire smoke dataset, and undertake a novel framework integrating conventional methods into CNN for wildfire smoke detection. The framework was divided into two major phases. First, a segmentation strategy of candidate smoke regions was introduced as the pre-processing method of separating smoke and non-smoke images. Second, the candidate smoke regions were detected by our proposed WSDD-Net.

3.1. Candidate Smoke Region Segmentation

Smoke images show different results in each color space due to the change in illumination and environment [25]. Past research discussed smoke images in different color spaces [26,27,28]. The results showed that color information is least affected by illumination changes in the YUV color space. According to the probability statistics of thousands of real smoke images, Prema, Vinsley, and Suresh [25] demonstrated that the U-V of non-smoke regions was mainly distributed from 0 to 40, while candidate smoke regions were mainly distributed between 40 and 130. The segmentation strategy is shown in the following formula.
S 1 ( x , y ) = { 1 i f | U ( x , y ) 128 | > T U 0 O t h e r w i s e ,
S 2 ( x , y ) = { 1 i f | U ( x , y ) V ( x , y ) | > T U V 0 O t h e r w i s e
where U ( x , y ) and V ( x , y ) are images of the U and V components, T U and T U V are the threshold values of Equation (1) and Equation (2), which were derived from several experiments, and S 1 ( x , y ) and S 2 ( x , y ) are the pixels’ satisfying rules. Finally, candidate smoke regions were determined by using Equation (3).
S 0 ( x , y ) = { I i ( x , y ) i f S 1 ( x , y ) = 1 ( o r ) S 2 ( x , y ) = 1 0 O t h e r w i s e ,
where I i ( x , y ) is the input RGB image and S 0 ( x , y ) is the segmented candidate smoke region.
Figure 1 shows the segmentation results in the YUV color space. It was obvious that the segmentation strategy removed the complex background, except for smoke-colored regions including elements like cloud, fog, or haze. For the sake of eliminating the interference of a non-smoke region such as sky, cloud, fog, and haze, further classification of the candidate smoke region is needed.

3.2. Proposed WSDD-Net

In order to further classify the candidate smoke regions, we proposed a novel deep learning architecture, i.e., dilated DenseNet, for wildfire smoke detection. Below, we introduce the major structure of DenseNet, followed by a detailed description of our proposed wildfire smoke dilated DenseNet (WSDD-Net) architecture. The WSDD-Net architecture is presented in Figure 2. The dilated dense block in Figure 2 replaced the common convolutions in the original DenseNet, resulting in dilated convolutions.

3.2.1. DenseNet

The WSDD-Net was inspired by the state-of-the-art network architecture, i.e., DenseNet [22]. DenseNet is made up of many dense blocks. Each layer takes all previous outputs as input, each dense block is connected to a transition layer, and the last layer is connected to the global pooling and the fully connected layer directly. Figure 3 illustrates the architecture of the original dense block. Batch normalization (BN), activation function, pooling, and convolution are defined as the composite function F l ( ) . We denote the output of the l t h layer as x l . The l t h layer receives the outputs of all previous layers as input
x l = H l ( [ x 0 , x 1 , , x l 1 ] ) ,
where x 0 , x 1 , , x l 1 denote the outputs of 0 , 1 , , l 1 layers.

3.2.2. Dilated Dense Connection (DDC)

Using the dilated convolutions, multi-scale features were extracted without extra computational cost compared to that of approaches using common convolution kernels [29]. Taking the advantages into account, the DDC replaced the common convolutions in the original DenseNet, resulting in dilated convolutions. The original dense block achieved multi-scale feature extraction by stacking 3 × 3 convolutions. As the dilated convolutional had a larger receptive field compared to the common convolution [30], the proposed DDC block adopted the dilated convolutions to achieve multi-scale feature extraction. As shown in Figure 4, dilated convolutions with two dilation rates, i.e., 2 and 3, were involved in the proposed DDC block. The common 3 × 3 convolution was placed after each dilated convolution to fuse the previous outputs.
The proposed DDC resolved two crucial shortcomings of the existing structure. First, compared to that of the proposed DDC block, the dilation rates of existing structures are usually 4, 8, 16, etc. Because of these large numbers, the receptive field of existing structures normally exceeded the image size. As a result, convolution computation required a large number padding zeros. Second, the architecture of the prior framework had no short-cut connections, which means it was unable to perform multi-scale feature extraction.

3.3. Loss Function

We trained the network using an improved cross entropy loss function, namely balanced cross entropy (BCE), which was inspired by the focal loss function [23]. It was designed for training to account for the imbalance in wildfire smoke datasets by down-weighting the non-smoke part of the dataset. The improved loss function was introduced from the cross entropy loss for binary classification:
C E ( p , y ) = { log p y = 1 log ( 1 p ) y = 0 ,
where y = { 0 , 1 } is the real image label and p [ 0 , 1 ] is the predicted class probability for the class with the label y = 1 . In order to address class imbalance, a weighting factor α [ 0 , 1 ] for class 1 and 1 α for class 0 was introduced. α was treated as a hyper-parameter to set cross validation. We wrote the balanced cross entropy loss as
B C E ( p , y ) = { α log p y = 1 ( 1 α ) log ( 1 p ) y = 0 .
This loss function alleviated the problem of the wildfire smoke dataset imbalance while guaranteeing a high detection accuracy rate and low false-alarm rate.

4. Experimental Results

The proposed WSDD-Net was established using the Pytorch toolbox. The experiment system environment was Ubuntu 16.04 and the programming language was python. The hardware configuration consisted of an E5-2620 CPU (Intel Corporation, Santa Clara, CA, USA) and a GeForce GTX 1080ti GPU (Nvidia Corporation, Santa Clara, CA, USA). The network was trained with a mini-batch size of 32. In order to achieve faster network convergence, the initial learning rate was set to 0.1. After that, the learning rate decreased by 0.1 times per 20 epochs. The network input materials were RGB color images, which were resized to 224 × 224 .

4.1. Wildfire Smoke Dataset

The wildfire smoke (WS) dataset consisted of 4595 images (1685 wildfire smoke images and 2910 cloud, fog, and haze images), which was collected from the literature and websites by the authors. From the smoke dataset and the non-smoke dataset, 80% of images were selected, respectively. These smoke images and non-smoke images were randomly separated according to the ratio of 3:1 for training and validation. The remaining 20% of the images were prepared as the testing set for the performance comparison between the benchmark algorithms and our framework. Rotation and mirroring were used to expand the size of the training dataset. Smoke and non-smoke images in the training dataset were rotated by 180° and then reflected vertically to generate additional images, resulting in an augmented training dataset with 11,028 images. Examples of different categories of the WS dataset are shown in Figure 5. The first two rows show the segmented patches of wildfire smoke image patches. The third and fourth rows show fog and cloud images patches, respectively. It can be seen that illumination, shooting distance, smoke color and background were different due to the different sources of these images.

4.2. Evaluation Criterion

To evaluate the performance of the proposed method, we uses a train–validation–test scheme. The training set was used for the actual training of the method, while fine-tuning the hyper-parameters was carried out using the validation set. The total performance of proposed framework was evaluated using the wildfire smoke testing set. When evaluating models for binary classification on a given dataset of positives and negatives, four different types of data are usually defined: true positive (TP), true negative (TN), false positive (FP), and false negative (FN). TP is the number of true positives, i.e., the number of smoke patches classified as smoke, TN is the number of true negatives, i.e., the number of non-smoke patches classified as non-smoke, FN is the number of false negatives, i.e., the number of non-smoke patches classified as smoke, and FP is the number of false positives, i.e., the number of smoke patches classified as non-smoke. The accuracy rate (AR), false-alarm rate (FAR), and recall rate (RR, defined below) were adopted as the criteria for the performance evaluation. In some extreme cases, RR and AR were conflicting, which required the use of an F1-score (F1) to evaluate the performance of the algorithm [18]. The definitions are as follows.
A R = T P + T N T P + T N + F P + F N ,
F A R = F N T N + F N ,
R R = T P T P + F N ,
F 1 = 2 ( R R × D R ) R R + D R ,
where
D R = T P T P + F P .

4.3. Results of the WS Dataset

To demonstrate the advantages produced by proposed WSDD-Net, several well-known deep learning networks like ResNet-50/101 [31], Incveption-V3 [32], and DenseNet-121/169 [22], were also used for the performance evaluation.
The training processes of these networks are illustrated in Figure 6. Figure 6a shows the changing accuracy of the system with training over different epochs. It was clear that the AR of the proposed WSDD-Net rapidly rose from 0 to 5 epochs and gradually rose from 5 to 10 epochs. Then, the AR reached 99% and fluctuated around 99%. Figure 6b shows the changing loss of the training set. It was obvious that the loss of WSDD-Net sharply dropped from 0 to 5 epochs and slightly descended from 5 to 10 epochs. Following this, the loss varied gradually and fluctuated between 0.05 and 0.09. Although the WSDD-Net had a better performance than other classical networks in training, the critical performance of the network for wildfire smoke detection depended on the results of the testing phase.
The results for the wildfire smoke test set are listed in Table 1. It was clear that the proposed WSDD-Net achieved the highest AR (99.20%) and F1 (99.25) among the models trained on our dataset, which was 1.91% and 1.78 higher than the runner-up, DenseNet-169. Furthermore, we noticed that almost all of the listed classical algorithms performed poorly when dealing with the images of non-smoke dataset, i.e., the lowest FAR produced by the ResNet-101 was 2.41%, which was seven times higher than the FAR of WSDD-Net. The test time of our WSDD-Net was 9.93 ms, which was competitive compared to that of the best performance. The DenseNet-121 and WSDD-Net have much smaller model sizes than other networks, and have the ability to adopt small datasets to prevent overfitting. Compared to the DenseNet-121, the proposed WSDD-Net uses multiple dilated convolutions to extract multi-scale features. As shown in Table 1, the proposed WSDD-Net outperformed the DenseNet-121 and produced the best F1. The experimental results showed that the proposed novel framework integrating conventional method into CNN maintained a good balance between network size and feature learning capacity, which is extremely effective for small-scale wildfire smoke datasets.
To verify the availability of our proposed WSDD-Net, statistical measures were also used to evaluate model performance, which was inspired by the paired t-test used in Ghaffari et al. [33]. Detailed results of the experiments are shown in Figure 7 and Table 2. It was observed that the proposed method exhibited excellent performance in both AR and FAR, showing large advantages compared to other algorithms. These results indicated the stability of our model. Table 2 represents the statistical significance from comparing the average AR of each model using an independent t-test. A significance of <0.05 was considered to be statistically significant. As shown in Table 2, the greater the t-value, the more significant the difference between the two models. The results demonstrated that WSDD-Net had a higher average AR and was more suitable for wildfire smoke detection than other classical networks from a statistical perspective.

4.4. Results on the Yuan Dataset

To further illustrate the excellent performance of the proposed framework, we evaluated the framework using a public smoke dataset called the Yuan dataset [34]. The algorithms for the comparisons included MCCNN [1], DNCNN [34], ZF-Net [35], and HLTPMC [36].
The experimental results are presented in Figure 8. The proposed WSDD-Net achieved the highest AR (99.71%) and RR (99.64%) among the five networks using the Yuan dataset, and obtained a higher F1 than the other methods, except for MCCNN. Although our method obtained a lower F1 than MCCNN, the FAR of our network was the lowest. In conclusion, the DDC structure produced significant improvements in both AR and FAR for our WSDD-Net. Better experimental results were obtained using the WS and Yuan datasets, indicating that the framework integrating conventional methods into CNN for wildfire smoke detection had better smoke detection capabilities under different negative sample interferences and achieved a high accuracy rate and a low false-alarm rate.

5. Conclusions

Many researchers have attempted to use CNN for wildfire smoke detection. However, the application of CNNs in wildfire smoke detection still faces several issues. For example, almost all smoke datasets have a serious class imbalance. It is difficult for deep neural networks to directly detect wildfire smoke images due to the interference of candidate smoke regions. To address these issues, we proposed a novel framework integrating conventional methods into CNN for wildfire smoke detection. The framework consisted of a candidate smoke region segmentation strategy and advanced network architecture, namely wildfire smoke dilated DenseNet (WSDD-Net). The candidate smoke region segmentation removed the complex backgrounds of wildfire smoke images. The proposed WSDD-Net achieved multi-scale feature extraction by combining dilated convolutions with dense block. In order to solve the problem of the dataset imbalance, an improved cross entropy loss function, namely balanced cross entropy (BCE), was used instead of the original cross entropy loss function in the training process. The proposed WSDD-Net was evaluated using the WS and Yuan smoke datasets, and achieved a high AR (99.20%) and a low FAR (0.24%). The experimental results demonstrated that the proposed framework had better detection capabilities under different negative sample interferences.

Author Contributions

Data curation, E.Z.; project administration, J.Z.; writing—original draft, T.L. and C.H.; writing—review and editing, T.L.

Funding

This study was financially supported by the Fundamental Research Funds for the Central Universities (Grant No.2016ZCQ08 and Grant No.2019SHFWLC01).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Luo, Y.M.; Zhao, L.; Liu, P.Z.; Huang, D.T. Fire smoke detection algorithm based on motion characteristic and convolutional neural networks. Multimedia Tools Appl. 2018, 77, 15075–15092. [Google Scholar] [CrossRef]
  2. Lin, G.H.; Zhang, Y.M.; Xu, G.; Zhang, Q.X. Smoke detection on video sequences using 3D convolutional neural networks. Fire Technol. 2019, 55, 1827–1847. [Google Scholar] [CrossRef]
  3. Yuan, F.N. A double mapping framework for extraction of shape-invariant features based on multi-scale partitions with AdaBoost for video smoke detection. Pattern Recognit. 2012, 45, 4326–4336. [Google Scholar] [CrossRef]
  4. Ko, B.C.; Cheong, K.; Nam, J. Fire detection based on vision sensor and support vector machines. Fire Saf. J. 2009, 44, 322–329. [Google Scholar] [CrossRef]
  5. Yuan, W.; Li, J.; Fang, J.; Zhang, Y.M. Color model and method for video fire flame and smoke detection using Fisher linear discriminant. Opt. Eng. 2013, 52, 027205. [Google Scholar]
  6. Yin, M.X.; Lang, C.Y.; Li, Z.; Feng, S.H.; Wang, T. Recurrent convolutional network for video-based smoke detection. Multimedia Tools Appl. 2019, 78, 237–256. [Google Scholar] [CrossRef]
  7. Xie, J.J.; Li, A.Q.; Zhang, J.G.; Cheng, Z.A. An Integrated wildlife recognition model on multi-branch aggregation and squeeze-and-excitation network. Appl. Sci. 2019, 9, 2794. [Google Scholar] [CrossRef]
  8. Gil-Martín, M.; Montero, J.M.; San-Segundo, R. Parkinson’s disease detection from drawing movements using convolutional neural networks. Electronics. 2019, 8, 907. [Google Scholar] [CrossRef]
  9. Li, Z.Q.; Chen, Z.X.; Wu, Q.M.J.; Liu, C.Y. Real-time pedestrian detection with deep supervision in the wild. Signal Image Video Process. 2019, 13, 761–769. [Google Scholar] [CrossRef]
  10. Xu, G.; Zhang, Y.M.; Zhang, Q.X.; Lin, G.H. Video smoke detection based on deep saliency network. arXiv 2018, arXiv:1809.02802. [Google Scholar] [CrossRef]
  11. Matlani, P.; Shrivastava, M. Hybrid deep VGG-NET convolutional classifier for video smoke detection. CMES Comput. Model. Eng. Sci. 2019, 119, 427–458. [Google Scholar] [CrossRef]
  12. Yang, B.H.; Zheng, D.; Zhang, Y.H.; Zheng, X.M. Recognition of fire detection based on neural network. In Life System Modeling and Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2010; pp. 250–258. [Google Scholar]
  13. Zhang, Z.J.; Shen, T.; Zou, J.H. An improved probabilistic approach for fire detection in videos. Fire Technol. 2014, 50, 745–752. [Google Scholar] [CrossRef]
  14. Tung, T.X.; Kim, J. An effective four-stage smoke-detection algorithm using video image for early fire-alarm systems. Fire Saf. J. 2011, 5, 276–282. [Google Scholar] [CrossRef]
  15. Jia, Y.; Yuan, J.; Wang, J.J.; Fang, J.; Zhang, Q.X.; Zhang, Y.M. A saliency- based method for early smoke detection in video sequences. Fire Technol. 2016, 52, 1271–1292. [Google Scholar] [CrossRef]
  16. Yuan, F.N. Video-based smoke detection with histogram sequence of LBP and LBPV pyramids. Fire Saf. J. 2011, 46, 132–139. [Google Scholar] [CrossRef]
  17. Muhammad, K.; Ahmad, J.; Mehmood, I.; Rho, S. Convolutional Neural Networks based Fire Detection in Surveillance Videos. IEEE Access 2018, 6, 18174–18183. [Google Scholar] [CrossRef]
  18. Mao, W.T.; Wang, W.P.; Dou, Z.; Li, Y. Fire recognition based on multi-channel convolutional neural network. Fire Technol. 2018, 54, 531–554. [Google Scholar] [CrossRef]
  19. Sharma, J.; Granmo, O.; Goodwin, M.; Fidje, J.T. Deep Convolutional Neural Networks for Fire Detection in Images. In Proceedings of the International Conference on Engineering Applications of Neural Networks, EANN2017, Athens, Greece, 25–27 August 2017. [Google Scholar]
  20. Xu, G.; Zhang, Y.M.; Zhang, Q.X.; Lin, G.H.; Wang, J.J. Domain adaptation from synthesis to reality in single-model detector for video smoke detection. arXiv 2017, arXiv:1709.08142. [Google Scholar]
  21. Zhang, Q.X.; Lin, G.H.; Zhang, Y.M.; Xu, G.; Wang, J.J. Wildland forest fire smoke detection based on faster r-cnn using synthetic smoke images. Procedia Eng. 2018, 211, 411–466. [Google Scholar] [CrossRef]
  22. Huang, G.; Liu, Z.; Maaten, L.V.; Weinberger, K.Q. Densely connected convolutional networks. arXiv 2016, arXiv:1608.06993. [Google Scholar]
  23. Lin, T.; Goyal, P.; Grishick, R.; He, K.; Dollár, P. Focal Loss for dense object detection. arXiv 2017, arXiv:1708.02002. [Google Scholar]
  24. Abulnaga, S.M.; Rubin, J. Ischemic stroke lesion segmentation in CT perfusion scans using pyramid pooling and focal loss. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Springer: Cham, Switzerland, 2018; pp. 352–363. [Google Scholar]
  25. Prema, C.E.; Vinsley, S.S.; Suresh, S. Multi feature analysis of smoke in YUV color space for early forest fire detection. Fire Technol. 2016, 52, 1319–1342. [Google Scholar] [CrossRef]
  26. Kim, D.K.; Wang, Y.F. Smoke detection in video. In Proceedings of the 2009 WRI World Congress on Computer Science and Information Engineering, Los Angeles, CA, USA, 31 March–2 April 2009; pp. 759–763. [Google Scholar]
  27. Ye, W.; Zhao, J.H.; Wang, S.; Wang, Y.; Zhang, D.Y.; Yuan, Z.Y. Dynamic texture based smoke detection using Surfacelet transform and HMT model. Fire Saf. J. 2015, 73, 91–101. [Google Scholar] [CrossRef]
  28. Benazza-Benyahia, A.; Hamouda, N.; Tlili, F.; Ouerghi, S. Early smoke detection in forest areas from DCT based compressed video. In Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 2752–2756. [Google Scholar]
  29. Szegedy, C.; Liu, W.; Jia, Y.Q.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. arXiv 2014, arXiv:1409.4842. [Google Scholar]
  30. Li, Y.X.; Xie, X.P.; Shen, L.L.; Liu, S.X. Reversed active learning based Atrous DenseNet for pathological image classification. arXiv 2018, arXiv:1807.02420. [Google Scholar]
  31. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  32. Szegedy, C.; Vanhoucke, V.; Loffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. arXiv 2015, arXiv:1512.00567. [Google Scholar]
  33. Ghaffari, A.; Abdollahi, H.; Khoshayand, M.R.; Bozchalooi, I.S.; Dadgar, A.; Rafiee-Tehrani, M. Performance comparison of neural network training algorithms in modeling of bimodal drug delivery. Int. J. Pharm. 2006, 327, 126–138. [Google Scholar] [CrossRef]
  34. Yin, Z.J.; Wan, B.Y.; Yuan, F.N.; Xia, X.; Shi, J.T. A deep normalization and convolutional neural network for image smoke detection. IEEE Access 2017, 5, 18429–18438. [Google Scholar] [CrossRef]
  35. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. arXiv 2013, arXiv:1311.2901. [Google Scholar]
  36. Yuan, F.N.; Shi, J.T.; Xia, X.; Fang, Y.M.; Fang, Z.J.; Mei, T. High-order local ternary patterns with locality preserving projection for smoke detection and image classification. Inf. Sci. 2016, 372, 225–240. [Google Scholar] [CrossRef]
Figure 1. Segmentation results in the YUV color space. (a) RGB image, (b) Y component, (c) U component, (d) V component, and (e) candidate smoke region.
Figure 1. Segmentation results in the YUV color space. (a) RGB image, (b) Y component, (c) U component, (d) V component, and (e) candidate smoke region.
Electronics 08 01131 g001
Figure 2. Network architecture of proposed wildfire smoke dilated DenseNet (WSDD-Net).
Figure 2. Network architecture of proposed wildfire smoke dilated DenseNet (WSDD-Net).
Electronics 08 01131 g002
Figure 3. A five-layer dense block.
Figure 3. A five-layer dense block.
Electronics 08 01131 g003
Figure 4. Framework of the proposed dilated dense connection (DDC).
Figure 4. Framework of the proposed dilated dense connection (DDC).
Electronics 08 01131 g004
Figure 5. Sample images from the training and testing sets. (a) Smoke image, (b) smoke image, (c) fog image, and (d) cloud image.
Figure 5. Sample images from the training and testing sets. (a) Smoke image, (b) smoke image, (c) fog image, and (d) cloud image.
Electronics 08 01131 g005
Figure 6. Training process curves of ResNet-50/101, Inception-v3, DenseNet-121/169, and WSDD-Net. (a) Accuracy curve and (b) loss curve.
Figure 6. Training process curves of ResNet-50/101, Inception-v3, DenseNet-121/169, and WSDD-Net. (a) Accuracy curve and (b) loss curve.
Electronics 08 01131 g006
Figure 7. Test results from the wildfire smoke (WS) dataset. (a) Box-plot for 10 AR values of each method and (b) box-plot for 10 FAR values of each method.
Figure 7. Test results from the wildfire smoke (WS) dataset. (a) Box-plot for 10 AR values of each method and (b) box-plot for 10 FAR values of each method.
Electronics 08 01131 g007
Figure 8. Experiment results using the Yuan dataset.
Figure 8. Experiment results using the Yuan dataset.
Electronics 08 01131 g008
Table 1. Comparisons with classical deep convolutional neural networks (CNNs).
Table 1. Comparisons with classical deep convolutional neural networks (CNNs).
ModelSize (MB)AR (%)FAR (%)F1 RR (%)DR (%)Test time (ms)
ResNet-50[30]89.6996.022.4196.2397.8594.6610.31
ResNet-101[30]162.1495.547.5695.9493.7798.2212.89
Inception-V3[31]83.1291.7215.1292.6888.2097.6310.70
DenseNet-121[22]26.5396.344.4796.6096.1897.038.17
DenseNet-169[22]47.6497.292.7597.4797.6297.3310.07
WSDD-Net39.7299.200.3499.2599.7098.819.93
Table 2. Statistical significance of performance comparisons.
Table 2. Statistical significance of performance comparisons.
Null Hypothesis(H0)T-valueSig
ResNet-50[30] = ResNet-101[30]1.9880.063
DenseNet-121[22] = DenseNet-169[22]1.3490.194
WSDD-Net = Inception-V3[31]41.5040.000
WSDD-Net = DenseNet-121[22]22.3380.000
WSDD-Net = DenseNet-169[22]20.8630.000
WSDD-Net = ResNet-50[30]23.1860.000
WSDD-Net = ResNet-101[30]27.1160.000

Share and Cite

MDPI and ACS Style

Li, T.; Zhao, E.; Zhang, J.; Hu, C. Detection of Wildfire Smoke Images Based on a Densely Dilated Convolutional Network. Electronics 2019, 8, 1131. https://doi.org/10.3390/electronics8101131

AMA Style

Li T, Zhao E, Zhang J, Hu C. Detection of Wildfire Smoke Images Based on a Densely Dilated Convolutional Network. Electronics. 2019; 8(10):1131. https://doi.org/10.3390/electronics8101131

Chicago/Turabian Style

Li, Tingting, Enting Zhao, Junguo Zhang, and Chunhe Hu. 2019. "Detection of Wildfire Smoke Images Based on a Densely Dilated Convolutional Network" Electronics 8, no. 10: 1131. https://doi.org/10.3390/electronics8101131

APA Style

Li, T., Zhao, E., Zhang, J., & Hu, C. (2019). Detection of Wildfire Smoke Images Based on a Densely Dilated Convolutional Network. Electronics, 8(10), 1131. https://doi.org/10.3390/electronics8101131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop