Next Article in Journal
Research on a Defecation Pre-Warning Algorithm for the Disabled Elderly Based on a Semi-Supervised Generative Adversarial Network
Previous Article in Journal
Investigating Stroke Effects on Respiratory Parameters Using a Wearable Device: A Pilot Study on Hemiplegic Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Thermal Infrared Pedestrian-Detection Method for Edge Computing Devices

1
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
2
Institute of High Performance Computing and Bigdata, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
3
School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
4
Nanjing Center of HPC China, Nanjing 210023, China
5
Jiangsu HPC and Intelligent Processing Engineer Research Center, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6710; https://doi.org/10.3390/s22176710
Submission received: 12 July 2022 / Revised: 30 August 2022 / Accepted: 1 September 2022 / Published: 5 September 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
The thermal imaging pedestrian-detection system has excellent performance in different lighting scenarios, but there are problems regarding weak texture, object occlusion, and small objects. Meanwhile, large high-performance models have higher latency on edge devices with limited computing power. To solve the above problems, in this paper, we propose a real-time thermal imaging pedestrian-detection method for edge computing devices. Firstly, we utilize multi-scale mosaic data augmentation to enhance the diversity and texture of objects, which alleviates the impact of complex environments. Then, the parameter-free attention mechanism is introduced into the network to enhance features, which barely increases the computing cost of the network. Finally, we accelerate multi-channel video detection through quantization and multi-threading techniques on edge computing devices. Additionally, we create a high-quality thermal infrared dataset to facilitate the research. The comparative experiments on the self-built dataset, YDTIP, and three public datasets, with other methods show that our method also has certain advantages.

1. Introduction

Pedestrian object detection is an important research direction in computer vision, and consists of feature extraction, object classification, and object localization. Most of the existing methods are mainly concerned with object detection under visible light. However, due to complex lighting conditions, e.g., cloud, rain, fog, etc., visible light methods are insufficient for extracting features. Hence, infrared cameras can be used, which can sense the radiation of the object and can effectively avoid the interference of complex scenes during detection [1]. However, thermal infrared images have the disadvantages of low contrast, weak texture, and loud noise. These physical properties still lead to challenging object detection in thermal infrared scenes [2,3].
Infrared pedestrian-detection methods can be divided into traditional methods and convolutional neural network (ConvNet) methods [4]. Feature extraction and classification are performed separately in traditional object detection. Firstly, the pedestrian features (e.g., color, edge, texture, etc.) are extracted by handcraft; then, the classifier is trained with these features; finally, sliding windows are utilized for object localization [5]. Compared with traditional detection methods, ConvNets can implement end-to-end pedestrian detection, which is independent of expert experience [6]. The ConvNet methods consist of a feature extraction network, bounding-box regression, and object classification [7].
The traditional methods are based on handcraft features [8]. Zhang et al. [9] designed rectangular templates of the human body with multi-modal and multi-channel Haar-like features. Watanabe et al. [10] utilized two co-occurrence histogram features (i.e., CoHOG and CoHLBP), which describe different aspects of object shapes to improve the accuracy of pedestrian detection. Felzenszwalb et al. [11] proposed latent SVM, which combines a margin-sensitive method for data mining hard negative examples. Zhang et al. [12] extended two feature classes (edgelets and HOG features) and two classification models (AdaBoost and SVM cascade) to infrared images. Lowe et al. [13] introduced a scale-invariant feature transform for object recognition. Similarly, Dalal et al. [14] utilized the histogram of an oriented gradient for feature extraction. Professor Maldague’s team utilized visible light and thermal infrared image pairs to extract regions of interest [15]. However, these methods have insufficient performance in complex scenarios. Moreover, the strategy of sliding windows generates a massive number of redundant boxes. Additionally, the strategy of handcrafted features lacks high-level semantic information. These two strategies lead to low object detection efficiency [16].
The ConvNet methods involve the Yolo series [17] and the SSD series [18], which have huge parameters (i.e., the weight parameters of the network) and high latency on edge computing devices. The edge computing devices for these series are only Internet-of-Things terminal devices, which have low computing power and power consumption [19]. Hence, the MobileNet [20] series and the ShuffleNet [21] series were presented, which are represented by a lightweight base model [22]. Zheng et al. [23] utilized MobileNetv3 and depthwise separable convolution to achieve parameter compression and inference acceleration of Yolov4 (i.e., Yolov4-lite). Wang et al. [24] achieved SSD parameter compression using MoblieNetV2 in infrared image pedestrian detection. There is also Yolov4-tiny [25], which performs lightweight reconstruction of Yolov4. According to the scenario’s needs, EfficientDet [26], and Yolov5 [27] utilized hyperparameters to control the network parameters. For example, the network structure of EfficentDet (i.e., D0 to D7) is, respectively, associated with the input size, where D0 corresponds to the input size of 512 × 512. Rep-Yolov4 [28] utilized network pruning to compress network parameters, speed up inference, etc.
The above ConvNet methods (i.e., anchor-based object detectors) are less effective for some specific datasets, which require the reclustering of anchors [29]. Many studies have shown that anchor-free detectors can perform comparably to anchor-based detectors in recent years. The Yolox of anchor-free detectors introduces a decoupled head, which reduces the conflict between classification and regression tasks [30]. Liu et al. [4] proposed a thermal infrared pedestrian and vehicle detector based on Yolov3 by expanding the parameters of YoloHead, which consist of anchor-free and anchor-based. To enhance the features, the attention mechanism was introduced to improve detection performance, e.g., SENet [31] and CBAM [32]. Xue et al. [33] proposed a spatio-temporal sharpening attention mechanism, which combined the channel attention module and the sharpening spatial attention module. Xue et al. [34] proposed novel multi-modal attention fusion based on Yolo for real-time pedestrian detection.
However, due to the small coverage area, low resolution, and reduced feature information of small objects in the thermal infrared images, the performance of the above methods is unsatisfactory. Gao et al. [35] proposed a high-precision detection method based on the feature mapping of deep neural networks with a spindle network structure. Lu et al. [36] proposed a novel small-object detection method via multidirectional derivative-based weighted contrast measures. Zhu et al. [37] introduced an attention mechanism based on RefineDet for the infrared detection of small pedestrians. Li et al. [38] utilized the weighted fusion of visible images and infrared images, which enhanced small-feature information. In this paper, we proposed an efficient thermal infrared detection method based on Yolox-S in complex pedestrian scenes. Our method can effectively solve the problem of small objects and occlusion objects, without adding any parameters. To sum up, our main contributions are:
(1)
We adjust the input size of the network, which reduces the calculation redundancy and alleviates the imbalanced number of small objects based on the self-built dataset.
(2)
Multi-scale mosaic data augmentation is proposed to extend object diversity in training, which simulates multiple objects, small objects, and occlusion objects scenarios.
(3)
To achieve effective feature fusion in the neck, we introduce the parameter-free attention mechanism to implement feature enhancement.
(4)
We accelerate the network inference using quantization technology on the edge device. To fully exploit hardware resources, multi-threading technology is utilized to realize multi-channel video detection tasks.
The rest of this paper is organized as follows. Section 2 details the proposed method. Section 3 conducts the experiments based on the infrared image datasets (YDTIP, UTMD [39], KAIST [40] and CVC-14 [41]) to verify the effectiveness of the proposed method. Section 4 summarizes this paper.

2. Proposed Method

In recent years, the ConvNet structure has been iteratively updated, and the capability of feature extraction has become stronger. In this paper, the main improvements based on Yolox-S in thermal imaging pedestrian-detection (TIPYOLO) are as follows: we modify the size of the input image by analyzing the characteristics of the dataset, and increase the diversity of the training set through a multi-scale mosaic; the parameter-free attention mechanism is introduced into Yolox-S; we utilize quantization and multi-threading technology to meet business timeliness requirements.

2.1. Multi-Scale Mosaic

In the object detection network, the input of the network has a uniform size, and generally has a square structure. The customary methods obtain input sizes of the standard by cutting, stretching, scaling, etc. However, these operations are prone to problems such as missing objects, a loss of resolution, and reduced accuracy. In the testing stage of Yolox-S, the sizes of the input images are 416 × 416 or 640 × 640. Hence, we utilize gray filling to achieve image scaling in the model inference phase.
Here, we take an image of 416 × 416 as an example; in Figure 1, A is the original input image. After calculating the scaling ratio of the A image, image B is obtained using the BICUBIC interpolation algorithm. Image B is pasted into a 416 × 416 gray image to obtain C1. This method of image scaling reduces the image resolution without distortion, but the filled gray pixel area in C1 is too large (i.e., more than half of the gray-filled area, causing redundant computation), which increases the inference time of the model. Meanwhile, considering the limited resources of edge computing devices, the maximum downsampling rate of the ConvNet is 32 times, which results in an optimal gray filling scale of 416 × 192, as shown by C2 in Figure 1.
For the model training phase, we introduce mosaic data augmentation [17]. Mosaic data augmentation mainly crops four pictures in a central position, and then, merges them into one picture after cropping. As shown in Figure 2a,b, the mosaic can enrich the scene, which involves the image background, small objects, and pedestrian occlusion. Moreover, a larger batchsize will increase the GPU memory burden. The Mosaic operation can reduce the dimension of the input data, that is, channel × height × width (CHW) × batchsize is reduced to CHW × batchsize × 1/4, thereby reducing the memory usage. This is due to the mosaic data augmentation being away from the distribution of natural images. Therefore, the code automatically turns off the mosaic during the training of the last 30 epochs, and the random probability of training with mosaic is 0.5.
On the basis of the above strategies, we utilize a random multi-scale training strategy. Since the maximum downsampling rate of the network is 32 times, the size sequence of the input image is also selected with an integer multiple of 32, that is, 288 × 64, 320 × 96, 352 × 128, 384 × 160, 416 × 192, 480 × 224, 512 × 256, 544 × 288, 576 × 320, 608 × 352, and 640 × 384. A scale is randomly selected in the sequence as the input size. This strategy can not only improve the adaptability of the network to different scales, but can also realize the scale perturbation of objects.

2.2. TIPYOLO Network Details

TIPYOLO is an improvement on Yolox-S in the thermal imaging of pedestrian detection. The input of model is a three-channel RGB image (i.e., 416 × 192 × 3). The backbone consists of CSPDarknet, which integrates general residuals and cross-stage partial residuals. Hence, CSPDarknet achieves efficient feature extraction with few parameters. As shown in Figure 3, SPPBottleneck is added to CSPDarknet, which mainly plays the role of multi-receptive field fusion. Between the backbone and neck, we introduce the parameter-free attention module (i.e., simAM) [42]. We utilize simAM to filter more focused feature information for the neck.
The theory of simAM is based on visual neuroscience. In visual processing, neurons with clear spatial suppression effects should be given higher priority. To better implement the attention mechanism, the simAM algorithm evaluates the importance of each neuron using the minimum energy equation e l * . The lower the energy, the more distinctive the target neuron from the surrounding neurons, and the more important it is for visual processing. The importance of each neuron can be obtained using 1 / e l * , as shown in Equation (1).
1 / e l * = ( t u ^ ) 2 / 4 ( σ ^ 2 + λ ) + 0.5
where u ^ = 1 / M i = 1 M x i and σ ^ 2 = 1 / M i = 1 M ( x i u ^ ) 2 ; t and x i are the target neuron and other neurons in a single channel of the input feature X C × H × W ; i is index over spatial dimension; and M = H × W is the number of neurons on that channel. The hyper-parameter λ is set to 0.0001
According to the definition of an attention mechanism, we need to enhance the feature, as shown in Equation (2).
X ˜ = s i g m o i d ( 1 / E ) X
where E groups all of e l * across channel and spatial dimensions.
PANet performs the fusion features of bottom-up and top-down; these features are filtered by the attention module. Here, feature fusion adopts Concat in the channel dimension, which has rich semantic information, that is, it includes both network deep features and shallow features. The fused features are passed to YoloHead for object classification and location regression. The first Conv layer of YoloHead mainly implements channel adjustment, which reduces the channel size and the amount of branch computation. The two parallel branches (i.e., localization and classification) have two 3 × 3 Conv layers, respectively, which further improves the nonlinear fitting ability of the network.

2.3. Video Processing Acceleration

During the deployment phase of the method, we utilize post-training quantization techniques to achieve network inference acceleration on edge computing devices. The model quantization is the process of mapping float values to int values. We take int8 as an example, to map the floating-point input to the fixed point. First, we need to count the numerical range of the input, which provides the maximum value of its absolute value (i.e., | max | ). Then, | max | is mapped to 127 (i.e., the maximum value under int8), obtaining the mapped ratio s 1 using Equation (3).
s 1 = 127 | max |
s 1 is the scaling ratio, which is the floating-point input mapped to the fixed point. Similarly, the weight scaling ratio s 2 can be calculated. Finally, s i 1 and s i 2 are saved in the weight parameter for subsequent fixed-point calculations, where the index i is the current network layer.
Here, we take the convolution operator as an example, as shown in Figure 4. First, the floating-point input is quantized to obtain an integer input using the scaling ratio (i.e., s i 1 and s i 2 ), as shown in Equations (4)–(6).
x = r o u n d ( x s i 1 )
c l a m p ( x , | T | ) = { x , | x | | T | s i g n ( x ) × | T | , | x | > | T |
q x = c l a m p ( x , | T | )
where x is the input, and q x is the quantized result. T is the threshold, and its value range is [−127, 127]. Similarly, the weight of ConvNet is quantized to obtain an integer weight. Then, an integer Conv operation is performed to obtain an output for the integer result. Finally, the floating-point output is obtained through the dequantization Equation (7), where q y is the result of the Conv operation, and y is the dequantized result [43].
y = q y s i 1 × s i 2
The above quantification algorithm is mainly utilized for network inference acceleration, which neglects the calculation acceleration of pre-processing and post-processing. The pre-processing, inference, and post-processing of the model are serial executions. Inevitably, some units of computation will be waiting. To alleviate the wait of series execution, we convert the three computational units of the model into pipeline executions, as shown in Figure 5b, that is, each computing unit is assigned a thread. As shown in Figure 5c, unlike b, two threads are added to speed up the pre-processing. However, pipelineV2 only works in single-channel video processing. For multi-channel video processing, we take two-channel video as an example, and the specific strategy is shown in Figure 5d,e. As shown in Figure 5d, we utilize pipelineV2 to implement parallelism processing of the two-channel video. However, considering that the two models are running on edge computing devices at the same time, which inevitably increases the hardware burden, we implement two-channel video processing based on the improved pipelineV2. The specific implementation details are: First, we utilize two buffer queues to temporarily store the pre-processing data. Then, the model reads the data alternately from the two queues. Once model inference is completed, the video data of post-processing are stored alternately in two buffer queues using the Flag operation. Finally, the visualization thread utilizes two queue datapoints to visualize.

3. Experiments and Results

3.1. Data Set and Experiment Setting

Currently, most of the public thermal infrared pedestrian datasets have relatively simple meteorological environments and low resolution. Hence, we generated a high-resolution thermal infrared pedestrian dataset (i.e., YDTIP), which was collected at a school. Additionally, YDTIP contains a rich meteorological environment, which involves cloud, rain, day-time, and night-time. The resolution of each image is 2624 × 1080 in the dataset. The dataset was collected using the automatic thermal imaging temperature measurement system TIC400B, the view field of system was about 25° × 19°, and the lens focal length was 18 mm.
During the COVID-19 epidemic, the system has been widely utilized for temperature measurement in crowded places. The thermal infrared images of pedestrians in different environmental scenarios are shown in Figure 6. As can be seen from the figure, infrared imaging is insensitive to interference by environmental factors.
First, we extracted interval frames from the video, then we cleaned them, and finally, we obtained 3193 pictures. We utilized LabelImg to manually label these images, and only labeled the objects whose occlusion was less than 80%, that is, heavy occlusion (35~80%), partial occlusion (1~35%), no occlusion (0%) [44]. The label data were saved in VOC format (i.e., XML). According to coco API’s definition of large, medium, and small objects, a small object is defined as an object with pixels less than 32 × 32, a medium object is defined as an object with pixels greater than 32 × 32 and less than 96 × 96, and a large object is defined as an object with pixels greater than 96 × 96 [45]. Here, the definition is an absolute concept. Hence, we scaled the long edge of the images to 416; the generated data distribution of large, medium, and small objects is shown in Figure 7a, of which there are 7847 small objects, 7133 medium objects, and 977 large objects.
At the University of Tokyo, Japan, researchers generated a multispectral dataset for autonomous vehicles. Multispectral images are composed of RGB images, near-infrared images, middle-infrared images, and far-infrared images, corresponding to 7512 images, respectively. Since the multispectral dataset contained five classes (i.e., bike, car, car_stop, color_cone, and person), the experiments only utilized the person class. According to the person labels, 2537 images were obtained from middle-infrared images. The data distribution is shown in Figure 7b, of which there are 2719 small objects, 1149 medium objects, and 232 large objects, namely UTMD. It should be noted that each image has a size of 320 × 256 in the UTMD dataset [39]. For the size characteristics of this dataset, during training, since the maximum downsampling rate of the network was 32, the input image size sequence was also selected with an integer multiple of 32, that is, 224 × 160, 256 × 192, 288 × 224, 320 × 256, 352 × 288, and 384 × 320.
We utilized the Pytorch deep learning framework for experiments under the Ubuntu 18.04 system. The machine configuration utilized in this experiment was as follows: the processor was AMD Ryzen 5 3500X 6-Core; the memory was 16GB; and the graphics card was Nvidia 1080Ti (11GB). The software environment utilized was as follows: Pytorch 1.8.1 + cu102, Python 3.6.13.
The model training adopts the transfer learning method. The model loads pre-trained weights on the VOC dataset, which can speed up the convergence of model training. Hence, this strategy is widely adopted. The thermal imaging dataset has a drab object color. The model is prone to overfitting due to only using the original image during the training phase. Hence, we added image horizontal flipping (IHF) and length and width scaling (LWS), as shown in Figure 8b. Then, we added color gamut transformation on the HSV color space (CGT-HSV), as shown in Figure 8c. Moreover, according to the distribution of the RGB three-channel pixel histogram in Figure 8d–f, it can be seen that the pixel difference between the three images is obvious. This is especially true for the augmentation operation in Figure 8f, which can further highlight the object contour information. This data augmentation strategy was utilized to increase the diversity of the dataset and reduce the risk of model overfitting. When the model was tested, only the gray pixels were filled. The YDTIP dataset is divided into 1564 images for training, 671 images for validation, and 958 images for testing. The UTMD dataset is divided into 1242 images for training, 533 images for validation, and 762 images for testing. Meanwhile, we also trained and tested on the commonly used KAIST [40] and CVC-14 [41] datasets, corresponding to 3050 and 2936 images, which performed similar split operations of the dataset. The image size of the KAIST dataset is 640 × 512, and the image size of the CVC-14 dataset is 640 × 480.
During training, we utilized BCE loss for the cls and obj branch, and IoU loss for the reg branch [24]. Model training was completed in two stages. In the first stage, we froze the backbone network for training. In this phase, we set the batchsize to 16, the learning rate to 0.001, the momentum to 0.937, and the weight decay to 0.0005, and a total of 50 epochs were trained. In the second stage, we unfroze the backbone network for training, set the batchsize to 8, and also trained for 50 epochs. SGD (stochastic gradient descent) was utilized as the optimizer of the whole training process. The learning rate was adjusted utilizing a cosine schedule. In comparison experiments, these training hyper-parameters were the same as the settings of each detector.

3.2. Evaluation Indexes

To accurately evaluate the performance of the proposed method, we selected four generally utilized evaluation indexes: frames per second (FPS), mean average precision (mAP), log-average miss rate, and parameters. FPS was utilized as the evaluation index of the model inference speed, the mAP was utilized as the evaluation index of the model accuracy, and the parameters were utilized as the evaluation index of the model memory occupation. Precision represents the proportion of correctly detected persons in all positive detection results, and Recall represents the proportion of correctly detected persons in the ground-truth box. The definitions are expressed as Equations (8) and (9), respectively.
Precision = T P T P + F P
Recall = T P T P + F N
T P (true positive) represents the number of correct predictions for all positive samples, F P (false positive) represents the number of incorrect predictions for all negative samples, and F N (false negative) represents the number of incorrect predictions for all positive samples. The average precision (AP) was calculated by utilizing the integral area of the Precision–Recall curve; Recall and Precision are abscissa and ordinate, respectively [46]. The area enclosed by the curve is the value of AP, as in Equation (10). The larger the area, the higher the detection accuracy. AP50 is the average precision calculated when the IoU is 0.5, and AP75 is similar. AP is the average of the precision, calculated by IoU at intervals of 0.05 from 0.5 to 0.95. APS represents the evaluation accuracy of small objects by IoU at intervals of 0.05 from 0.5 to 0.95. APM and APL are similar [47].
AP = 0 1 P ( R ) d R
The log-average miss rate is the evaluation index of detectors corresponding to AP [48]. The miss rate (MR) can be calculated using the number of T P s and the number of ground-truths, as in Equation (11). Additionally, the false positives per image (FPPI) can be calculated by dividing F P by the number of images, as in Equation (12). The log-average MR is calculated by averaging the MRs under 9 FPPI equally spaced in the range of [0.01: 1] (i.e., MR−2). Conceptually, the log-average MR is similar to AP for object detection, in that they both use a reference value to represent the entire curve, that is, the Precision-Recall curve and the MR–FPPI curve. When evaluating detectors using MR−2, a lower MR−2 indicates better performance [44].
MR = F N T P + F N = 1 Recall
FPPI = F P N
where N is the number of images.

3.3. Analysis of Experimental Results

3.3.1. Ablation Study on the YDTIP Dataset

Here, we explain the reason for choosing the 416 × 416 input size. According to the definition of COCO API, we scaled the image’s long side to 640. At this size, the number of small objects was 3221, the number of medium objects was 9907, and the number of large objects was 2829, leading to the number imbalance problem. Small-object detection is a difficult task in the object detection field. Hence, we scaled the image’s long side to 416, the number of small objects was 7847, the number of medium objects was 7133, and the number of large objects was 977. Compared with 640, the imbalance problem of small and medium objects under the ratio of 416 was alleviated. Moreover, it can be seen from Table 1 that the detection of large objects at different scales has good performance. Meanwhile, from Table 1, we can see that the quantitative analysis of detector performance utilizing AP, AP50, and AP75 yields good results under 640 × 640 conditions. However, the performance is poor under the APS index, which may be caused by insufficient training data for small objects. Additionally, it can be seen from Table 1 that the overall performance of the detector improves slightly compared to 416 × 416 when the input is 416 × 192, which further shows that less gray redundant calculation is necessary.
To prove the effectiveness of each improvement strategy, we conducted ablation experiments on the YDTIP dataset, and the results are shown in Table 2. The operators were: spatial (GAP) or channel average pooling (CAP), spatial (GMP) or channel max pooling (CMP), standard convolution (C2D) or channel-based convolution (C1D), standard fully connected layers (FC), batch normalization (BN), and ReLU. k and r are convolutional filter numbers and the reduction ratio, respectively. C is the current feature channels.
We compared the simAM module with three representative attention modules, SE [31], CBAM [32], and ECA [49], respectively. The specific evaluation results are shown in Table 2 (rows 2, 3, 4, and 5). From Table 2, we can see that the performance of each attention is slightly degraded under some indexes, but the overall increase is still relatively large. It is worth noting that the simAM module does not add any parameters to the existing network, which has a great advantage over other modules.
To more intuitively view the impact of each attention module, we utilized Grad-CAM [50] to visualize the focus heatmap of each detector. As shown in Figure 9, the brighter the color of the heatmap, the more the detector pays attention to the feature. By observation, it is found that the attention area of simAM is larger than other attention modules on the object.
For the data augmentation ablation experiment, it can be seen in Table 2 (rows 1, 6, and 7) that the multi-scale mosaic has a significant increase compared with the baseline (i.e., Yolox-S). Additionally, the multi-scale mosaic, compared with the mosaic, AP, and AP75, increased by 0.1% and 0.2%, respectively, and APM and APL increased by 0.4% and 1.7%. This further proves that the multi-scale mosaic strategy can effectively improve dataset diversity and model performance.
The simAM and multi-scale mosaic strategies were effectively combined, and the performance evaluation results are shown in Table 2 (last row). Compared with the baseline, the combined performance is significantly improved; AP and AP75 are improved by 0.6% and 1%, and APS is improved by 0.8%. The activation heatmap of TIPYOLO is shown in Figure 8. From the figure, we can see that our method has a larger receptive field for the object and a higher focus on the global features.
In summary, the results of the ablation experiments show that our strategies significantly improve the performance of the detector without introducing any parameters.

3.3.2. Comparison Experiments with Other Methods on the YDTIP Dataset

To further verify the performance of the improved method, TIPYOLO was compared with other methods on the YDTIP dataset, and the comparison results are shown in Table 3.
The experimental results show that TIPYOLO is better than Yolov5-S in terms of accuracy, but the speed is slightly slower by about 5 FPS. The FPS of TIPYOLO is nearly two times worse than the lightweight method Yolov4-tiny, but in terms of accuracy, TIPYOLO is better overall than Yolov4-tiny. Although the TIPYOLO parameter is nearly 2.3 times that of the lightweight method EfficientDet-d0, both AP and FPS are better than EfficientDet-d0. The reason EfficientDet-d0 has few parameters and low speed may be due to more feature fusion and more network branches, and the current deep learning framework has not further accelerated its inference. Compared with Yolov4, the performance of TIPYOLO is 0.1% less than Yolov4 on AP50 and APS. Compared with Yolov3, TIPYOLO is 0.2% less than Yolov3 on AP50. But in FPS and parameters indexes, our method is better than Yolo3 and Yolov4. Meanwhile, the performance of our method is overall better than SSD.Yolov4-lite utilizes depthwise separable convolution to replace the standard 3 × 3 convolution, and realizes the parameter compression and inference speed improvement of Yolov4, but the accuracy also decreases. Compared with Yolov4-lite, TIPYOLO achieves the best performance overall. In terms of AP indicators, TIPYOLO is lower than Faster-RCNN, but it still has advantages in terms of FPS and parameters. On the MR−2 indicator, TIPYOLO and the three large models (i.e., Faster R-CNN, YOLOv3, and YOLOv4) have a small gap.
In the actual detection results of TIPYOLO and other methods on the YDTIP dataset, as we can see from the comparison Figure 10, compared with other methods, TIPYOLO can be applied to complex scenes such as those including small objects and occluded objects. In the dense pedestrian scene, other methods have missed detection to different degrees, as shown in the blue dotted box marked in Figure 10. Additionally, EfficientDet-D0 has false positives (i.e., two detection boxes appear for one object), as shown in the black dotted box marked in Figure 10. Since the two-stage method [48] is difficult to deploy on edge computing devices, only the results of one-stage object detection methods [51] are visualized here.

3.3.3. Comparison Experiments with Other Methods on the Public Dataset

To further verify the performance of TIPYOLO, a comparative experiment was conducted with other advanced methods on the UTMD [39] dataset. The comparison results between TIPYOLO and other object detection methods are shown in Table 4.
Compared with Yolox-S, the parameters of TIPYOLO are unchanged; it has a greater advantage in terms of detection accuracy, and the FPS performance is comparable. Compared with the lightweight Yolov4-tiny, although TIPYOLO does not have an advantage in terms of parameters and FPS, it has obvious advantages in terms of detection accuracy. Compared with the lightweight EfficientDet-d0, although the TIPYOLO parameter is nearly 2.3 times larger, it has obvious advantages in terms of accuracy and FPS. TIPYOLO is still comparable to YOLOv3 in MR−2 metrics, and slightly higher than YOLOv4 and Faster R-CNN.
In the actual detection results of TIPYOLO and other methods on the UTMD dataset, as we can see from Figure 11, TIPYOLO has better performance for small object detection, while other methods have missed detection to varying degrees.
To further validate the generalization performance of TIPYOLO, we also conducted performance tests on the commonly used KAIST and CVC-14 datasets, which are shown in Table 5. Since the value of FPS changes insignificantly, the statistics of the FPS are not shown in Table 5. As can be seen from the table, our method also achieves decent performance on these two datasets.
To sum up, it can be seen that the improvement in Yolox-S is meaningful, and TIPYOLO balances the detection accuracy, inference speed, and model complexity. Meanwhile, our method has certain generalization performance for different datasets. It is more applicable in thermal imaging pedestrian-detection scenarios.

3.3.4. Evaluation of Method Performance on Edge Computing Devices

To verify the effectiveness of TIPYOLO on edge computing devices, we utilized a dense pedestrian video with 5400 frames for performance testing. The experimental verification platform adopts Cambrian Extrobox202. The machine configuration utilized in this experiment was as follows: the processor was Intel i3-6100U 2-Core; the memory was 4GB; 240GB SSD; the power supply was 12V/4A; and there was active cooling; the NPU adopted an MLU220-M.2 card with support for int16 and int8; the memory was LPDDR4x and 4G 3200MHz; and the power consumption was 8.25w.
After quantification, the detailed test results of the method accuracy are shown in Table 6; the weight parameters are almost unchanged (i.e., only some scale parameters are added), so weight statistics were not performed here. From the table, we can see that the quantization operation can improve the generalization ability of the model to a certain extent, that is, the AP75 increases to different degrees. However, the APM and APL of “(float32, float16) to (int8, int8)” and “(float32, float32) to (int8, int8)” have decreased. The changes in MR−2 are insignificant.
Next, we analyzed the computing speed of the method. TIPYOLO consists of pre-processing, inference, post-processing, and visualization, in which pre-processing, post-processing, and visualization are performed on the CPU and model inference on the NPU. As can be seen from Table 7, the processing speed of Structure a does not meet the real-time requirements (i.e., above 30 FPS), with a maximum of 20.9 FPS. Moreover, we can see from the table the effect of different degrees of quantification on accelerated processing. In the same case, int8 quantization is faster than int16 quantization, which can be seen in Table 7 (rows 1 and 2 or rows 3 and 4). Similarly, float16 is faster than float32 (rows 1 and 3 or rows 2 and 4). The experimental analysis shows that the speed of Structure a is slower than that of Structure b due to the existence of operator serial dependence (columns 2 and 3). In addition, it can be seen from the experimental analysis that the pre-processing operation limits the overall calculation speed of the method, and the multi-threading strategy can effectively alleviate this problem (columns 1, 2, and 3).
For single-channel video processing, our proposed Structure c meets the requirement of scene timeliness. From Figure 12a, we can see that multi-thread stacking increases CPU utilization and also increases the calculation speed of the method. For two-channel video processing, it can be seen from Figure 1a,c that under the int8 + float16 combination strategy, the multi-channel concurrency method is slightly lower than multi-channel parallelism in terms of CPU utilization rate (the multi-channel parallelism CPU is almost fully loaded, and parallelism patterns have one more visualization thread and one more post-processing thread than concurrency), while the MLU core utilization rate is comparable. In general, the parallelism calculation speed is slightly slower, which may be due to an excessive CPU performance load in parallelism mode. Meanwhile, it can be seen from Figure 12b,c that parallelism is faster than concurrency when the MLU core utilization rate is close to the full load state, and it also means high power consumption.
From Figure 12b, we find that a greater degree of quantization has a mitigation effect on the NPU power consumption, and has a speed-up effect. Additionally, from the red and green lines in Figure 12c, we find that a greater degree of quantization can reduce the computational burden on the NPU core.

4. Conclusions

In this paper, we propose a lightweight and efficient object detection method to solve the problem of thermal infrared pedestrian detection. Moreover, we verify the effectiveness and robustness of our method through a comparative study. Multi-scale mosaic data augmentation improves the diversity of data and reduces the risk of model overfitting, which simulates body occlusion and multi-scale objects. The parameter-free attention module is introduced between the backbone and neck to enhance feature extraction. To meet the requirements of the single-channel and two-channel video processing business, we combine quantization and multi-threading technology to accelerate method calculation. The experimental results show that our method has certain advantages over other methods in the thermal imaging pedestrian-detection scene, which is effective for false positives and missed detections, and has a high computing speed on edge computing devices.
In future work, we will further focus on model parameter compression and precision tuning. Furthermore, our method will be applied to auxiliary temperature measurement during the COVID-19 epidemic, which achieves automated abnormal body temperature detection and tracking.

Author Contributions

Conceptualization, S.Y., Y.J., and S.L.; methodology, S.Y., Y.J., and S.L.; validation, S.Y., X.Y., and C.M.; investigation, S.Y., C.M., and X.Y.; visualization, S.Y.; writing—original draft preparation, S.Y. and Y.F.; writing—review and editing, S.Y., Y.J., and Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported, in part, by the National Key R&D Program of China (2018AAA0103300, 2018AAA0103302), Postgraduate Research & Practice Innovation Program of Jiangsu Province (KYCX22_1024), the Natural Science Foundation of Jiangsu Province (Higher Education Institutions) (BK20170900, 19KJB520046, and 20KJA520001), the Innovative and Entrepreneurial Talents Projects of Jiangsu Province, the Jiangsu Planned Projects for Postdoctoral Research Funds (No. 2019K024), the Six Talent Peaks Project in Jiangsu Province (JY02), the Postgraduate Research & Practice Innovation Program of Jiangsu Province (KYCX19_0921, KYCX19_0906), the Open Research Project of Zhejiang Lab (2021KF0AB05), and the NUPT DingShan Scholar Project and NUPTSF (NY219132).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Maldague, X.; Galmiche, F.; Ziadi, A. Advances in pulsed phase thermography. Infrared Phys. Technol. 2002, 43, 175–181. [Google Scholar] [CrossRef]
  2. Bhadoriya, A.S.; Vegamoor, V.; Rathinam, S. Vehicle Detection and Tracking Using Thermal Cameras in Adverse Visibility Conditions. Sensors 2022, 22, 4567. [Google Scholar] [CrossRef]
  3. Pozzer, S.; De Souza, M.P.V.; Hena, B.; Hesam, S.; Rezayiye, R.K.; Azar, E.R.; Lopez, F.; Maldague, X. Effect of different imaging modalities on the performance of a CNN: An experimental study on damage segmentation in infrared, visible, and fused images of concrete structures. NDT E Int. 2022, 132, 102709. [Google Scholar] [CrossRef]
  4. Liu, Y.; Su, H.; Zeng, C.; Li, X. A robust thermal infrared vehicle and pedestrian detection method in complex scenes. Sensors 2021, 21, 1240. [Google Scholar] [CrossRef]
  5. Oluyide, O.M.; Tapamo, J.R.; Walingo, T.M. Automatic dynamic range adjustment for pedestrian detection in thermal (infrared) surveillance videos. Sensors 2022, 22, 1728. [Google Scholar] [CrossRef]
  6. Fang, Q.; Ibarra-Castanedo, C.; Maldague, X. Automatic defects segmentation and identification by deep learning algorithm with pulsed thermography: Synthetic and experimental data. Big Data Cogn. Comput. 2021, 5, 9. [Google Scholar] [CrossRef]
  7. Altay, F.; Velipasalar, S. The Use of Thermal Cameras for Pedestrian Detection. IEEE Sens. J. 2022, 22, 11489–11498. [Google Scholar] [CrossRef]
  8. El Maadi, A.; Maldague, X. Outdoor infrared video surveillance: A novel dynamic technique for the subtraction of a changing background of IR images. Infrared Phys. Technol. 2007, 49, 261–265. [Google Scholar] [CrossRef]
  9. Zhang, S.; Bauckhage, C.; Cremers, A.B. Informed haar-like features improve pedestrian detection. In Proceedings of the Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 947–954. [Google Scholar]
  10. Watanabe, T.; Ito, S. Two co-occurrence histogram features using gradient orientations and local binary patterns for pedestrian detection. In Proceedings of the Asian Conference on Pattern Recognition, Okinawa, Japan, 5–8 November 2013; pp. 415–419. [Google Scholar]
  11. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1627–1645. [Google Scholar] [CrossRef]
  12. Zhang, L.; Wu, B.; Nevatia, R. Pedestrian detection in infrared images based on local shape features. In Proceedings of the Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  13. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  14. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. Comput. Vis. Pattern Recognit. 2005, 1, 886–893. [Google Scholar]
  15. Torresan, H.; Turgeon, B.; Ibarra-Castanedo, C.; Hébert, P.; Maldague, X.P. Advanced surveillance systems: Combining video and thermal imagery for pedestrian detection. Thermosense XXVI 2004, 5405, 506–515. [Google Scholar]
  16. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared patch-image model for small target detection in a single image. IEEE Trans. Image Processing 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  17. Bochkovskiy, A.; Wang, C.-Y.; Liao, H. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  19. Chen, J.; Ran, X. Deep learning with edge computing: A review. Proc. IEEE 2019, 107, 1655–1674. [Google Scholar] [CrossRef]
  20. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 1314–1324. [Google Scholar]
  21. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  22. Han, J.; Yang, Y. L-Net: Lightweight and fast object detector-based ShuffleNetV2. J. Real-Time Image Processing 2021, 18, 2527–2538. [Google Scholar] [CrossRef]
  23. Zheng, Y.; Wu, G. YOLOv4-Lite–Based Urban Plantation Tree Detection and Positioning with High-Resolution Remote Sensing Imagery. Front. Environ. Sci. 2022, 641, 756227. [Google Scholar] [CrossRef]
  24. Wang, Z.; Feng, J.; Zhang, Y. Pedestrian detection in infrared image based on depth transfer learning. Multimed. Tools Appl. 2022, 81, 1–20. [Google Scholar] [CrossRef]
  25. Hou, X.; Ma, J.; Zang, S. Airborne infrared aircraft target detection algorithm based on YOLOv4-tiny. J. Phys. Conf. Ser. 2021, 1865, 042007. [Google Scholar] [CrossRef]
  26. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  27. Yang, G.; Feng, W.; Jin, J.; Lei, Q.; Li, X.; Gui, G.; Wang, W. Face mask recognition system with YOLOV5 based on image recognition. In Proceedings of the International Conference on Computer and Communications, Ho Chi Minh City, Vietnam, 1–3 August 2020; pp. 1398–1404. [Google Scholar]
  28. Wu, Z.; Wang, X.; Chen, C. Research on lightweight infrared pedestrian detection model algorithm for embedded Platform. Secur. Commun. Netw. 2021, 2021, 1549772. [Google Scholar] [CrossRef]
  29. Li, X.; Wang, S.; Liu, B.; Chen, W.; Fan, W.; Tian, Z. Improved YOLOv4 network using infrared images for personnel detection in coal mines. J. Electron. Imaging 2022, 31, 013017. [Google Scholar] [CrossRef]
  30. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  31. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  32. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  33. Xue, M.; Chen, M.; Peng, D.; Guo, Y.; Chen, H. One Spatio-Temporal Sharpening Attention Mechanism for Light-Weight YOLO Models Based on Sharpening Spatial Attention. Sensors 2021, 21, 7949. [Google Scholar] [CrossRef] [PubMed]
  34. Xue, Y.; Ju, Z.; Li, Y.; Zhang, W. MAF-YOLO: Multi-modal attention fusion based YOLO for pedestrian detection. Infrared Phys. Technol. 2021, 118, 103906. [Google Scholar] [CrossRef]
  35. Gao, Z.; Dai, J.; Xie, C. Dim and small target detection based on feature mapping neural networks. J. Vis. Commun. Image Represent. 2019, 62, 206–216. [Google Scholar] [CrossRef]
  36. Lu, R.; Yang, X.; Li, W.; Fan, J.; Li, D.; Jing, X. Robust infrared small target detection via multidirectional derivative-based weighted contrast measure. IEEE Geosci. Remote Sens. Lett. 2020, 19, 7000105. [Google Scholar] [CrossRef]
  37. Zhu, Y.; Yang, J.; Deng, X.; Xiao, C.; An, W. Infrared pedestrian detection based on attention mechanism. J. Phys. Conf. Ser. 2020, 1634, 012032. [Google Scholar] [CrossRef]
  38. Li, C.; Song, D.; Tong, R.; Tang, M. Illumination-aware faster R-CNN for robust multispectral pedestrian detection. Pattern Recognit. 2019, 85, 161–171. [Google Scholar] [CrossRef]
  39. Takumi, K.; Watanabe, K.; Ha, Q.; Tejero-De-Pablos, A.; Ushiku, Y.; Harada, T. Multispectral object detection for autonomous vehicles. Themat. Workshops ACM Multimed. 2017, 2017, 35–43. [Google Scholar]
  40. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; So Kweon, I. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar]
  41. González, A.; Fang, Z.; Socarras, Y.; Serrat, J.; Vázquez, D.; Xu, J.; López, A.M. Pedestrian detection at day/night time with visible and FIR cameras: A comparison. Sensors 2016, 16, 820. [Google Scholar] [CrossRef]
  42. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning 2021, Shenzhen, China, 26 February–1 March 2021; pp. 11863–11874. [Google Scholar]
  43. Jacob, B.; Kligys, S.; Chen, B.; Zhu, M.; Tang, M.; Howard, A.; Adam, H.; Kalenichenko, D. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In Proceedings of the Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2704–2713. [Google Scholar]
  44. Ning, C.; Menglu, L.; Hao, Y.; Xueping, S.; Yunhong, L. Survey of pedestrian detection with occlusion. Complex Intell. Syst. 2021, 7, 577–587. [Google Scholar] [CrossRef]
  45. He, X.; Cheng, R.; Zheng, Z.; Wang, Z. Small object detection in traffic scenes based on YOLO-MXANet. Sensors 2021, 21, 7422. [Google Scholar] [CrossRef] [PubMed]
  46. Jiang, C.; Ren, H.; Ye, X.; Zhu, J.; Zeng, H.; Nan, Y.; Sun, M.; Ren, X.; Huo, H. Object detection from UAV thermal infrared images and videos using YOLO models. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102912. [Google Scholar] [CrossRef]
  47. Roszyk, K.; Nowicki, M.R.; Skrzypczyński, P. Adopting the YOLOv4 architecture for low-latency multispectral pedestrian detection in autonomous driving. Sensors 2022, 22, 1082. [Google Scholar] [CrossRef] [PubMed]
  48. Cao, J.; Pang, Y.; Xie, J.; Khan, F.S.; Shao, L. From handcrafted to deep features for pedestrian detection: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4913–4934. [Google Scholar] [CrossRef] [PubMed]
  49. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. Supplementary material for ‘ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the Computer Vision and Pattern Recognition 2020, Nanjing, China, 16–18 October 2020; pp. 13–19. [Google Scholar]
  50. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  51. Li, W. Infrared Image Pedestrian Detection via YOLO-V3. Adv. Inf. Technol. Electron. Autom. Control Conf. 2021, 5, 1052–1055. [Google Scholar]
Figure 1. The input image is undistorted. (a) the original image; (b) the resized image; (c) the gray-filled image with a data dimension of 416 × 416 × 3; and (d) the gray-filled image with a data dimension of 416 × 192 × 3.
Figure 1. The input image is undistorted. (a) the original image; (b) the resized image; (c) the gray-filled image with a data dimension of 416 × 416 × 3; and (d) the gray-filled image with a data dimension of 416 × 192 × 3.
Sensors 22 06710 g001
Figure 2. Mosaic data augmentation. (a) contains four images, corresponding to 1–4, each of which is processed through the data augmentation strategy described in Section 3.1; (b) is an image composed of 1–4 in (a), where the white boxes in the figure represent the ground-truth label.
Figure 2. Mosaic data augmentation. (a) contains four images, corresponding to 1–4, each of which is processed through the data augmentation strategy described in Section 3.1; (b) is an image composed of 1–4 in (a), where the white boxes in the figure represent the ground-truth label.
Sensors 22 06710 g002
Figure 3. TIPYOLO network structure.
Figure 3. TIPYOLO network structure.
Sensors 22 06710 g003
Figure 4. Each network layer’s quantization process.
Figure 4. Each network layer’s quantization process.
Sensors 22 06710 g004
Figure 5. Single-channel and multi-channel video computing acceleration with multi-threading. (a) A single thread completes pre-processing, inference, and post-processing; (b) pre-processing, inference, and post-processing are completed by a single thread, respectively; (c) differs from (b) by adding two threads to the pre-processing; (d) video parallelism processing using pipelineV2; (e) video concurrency processing using the improved pipelineV2.
Figure 5. Single-channel and multi-channel video computing acceleration with multi-threading. (a) A single thread completes pre-processing, inference, and post-processing; (b) pre-processing, inference, and post-processing are completed by a single thread, respectively; (c) differs from (b) by adding two threads to the pre-processing; (d) video parallelism processing using pipelineV2; (e) video concurrency processing using the improved pipelineV2.
Sensors 22 06710 g005
Figure 6. YDTIP dataset visualization results.
Figure 6. YDTIP dataset visualization results.
Sensors 22 06710 g006
Figure 7. Distribution of large, medium, and small objects in the dataset. (a) YDTIP; (b) UTMD.
Figure 7. Distribution of large, medium, and small objects in the dataset. (a) YDTIP; (b) UTMD.
Sensors 22 06710 g007
Figure 8. The result of an image being processed through the above data augmentation strategy. (a) The original image with gray filling; (b) the result of IHF and LWS; (c) the result of CGT-HSV, where the data dimensions of the three images are all 416 × 192 × 3; and (df) the pixel histograms of the RGB channels, which correspond to (ac).
Figure 8. The result of an image being processed through the above data augmentation strategy. (a) The original image with gray filling; (b) the result of IHF and LWS; (c) the result of CGT-HSV, where the data dimensions of the three images are all 416 × 192 × 3; and (df) the pixel histograms of the RGB channels, which correspond to (ac).
Sensors 22 06710 g008
Figure 9. The activation heatmap for each detector generated using GradCAM.
Figure 9. The activation heatmap for each detector generated using GradCAM.
Sensors 22 06710 g009
Figure 10. The detection results of YIPYOLO and other methods on the YDTIP dataset. The blue dotted boxes represent missed detections; the black dotted boxes represent false-positive detections.
Figure 10. The detection results of YIPYOLO and other methods on the YDTIP dataset. The blue dotted boxes represent missed detections; the black dotted boxes represent false-positive detections.
Sensors 22 06710 g010
Figure 11. The detection results of YIPYOLO and other methods on the UTMD dataset.
Figure 11. The detection results of YIPYOLO and other methods on the UTMD dataset.
Sensors 22 06710 g011
Figure 12. Effect of differently structured methods on hardware utilization and power consumption. (a) The change in CPU utilization when the five structures in Figure 5, respectively, are running; (b) the effect of different strategies on NPU power consumption; and (c) the effect of different strategies on NPU core utilization.
Figure 12. Effect of differently structured methods on hardware utilization and power consumption. (a) The change in CPU utilization when the five structures in Figure 5, respectively, are running; (b) the effect of different strategies on NPU power consumption; and (c) the effect of different strategies on NPU core utilization.
Sensors 22 06710 g012
Table 1. Experiment results of different input sizes. “↑” indicates that the larger the value, the better the performance.
Table 1. Experiment results of different input sizes. “↑” indicates that the larger the value, the better the performance.
SizeAP ↑AP50AP75APSAPMAPL
640 × 6400.7010.9590.7850.4130.7220.869
416 × 4160.6540.9500.7210.4940.7730.895
416 × 1920.6570.9510.7260.5010.7720.895
Table 2. Results of ablation experiment. “↑” indicates that the larger the value, the better the performance.
Table 2. Results of ablation experiment. “↑” indicates that the larger the value, the better the performance.
MethodOperatorsParametersAP ↑AP50AP75APSAPMAPL
Baseline--0.6570.9510.7260.5010.7720.895
+ SEGAP, FC, ReLU 2 ( C 1 2 + C 2 2 + C 3 2 ) / r 0.6590.9510.7330.5050.7740.891
+ CBAMGAP, GMP, FC, ReLU, CAP, CMP, BN, C2D 2 ( C 1 2 + C 2 2 + C 3 2 ) / r + 2 ( k 1 2 + k 2 2 + k 3 2 ) 0.6590.9500.7270.5000.7760.900
+ ECAGAP, C1D k 1 + k 2 + k 3 0.6610.9490.7340.5040.7750.902
+ simAM GAP ,   / ,   , +00.6590.9510.7310.5020.7760.893
+ Mosaic-00.6630.9500.7360.5090.7750.886
+ Multi-scale mosaic-00.6640.9510.7380.5050.7790.903
+ Multi-scale mosaic
+ simAM
GAP ,   / ,   , +00.6630.9510.7360.5090.7780.897
Table 3. The comparison results of the methods on the YDTIP dataset. “↑” indicates that the larger the value, the better the performance; “” indicates that the smaller the value, the better the performance.
Table 3. The comparison results of the methods on the YDTIP dataset. “↑” indicates that the larger the value, the better the performance; “” indicates that the smaller the value, the better the performance.
MethodSizeParametersAP ↑AP50AP75APSAPMAPLMR−2FPS ↑
Faster R-CNN [48]416 × 416157.75 M0.6780.9490.7600.5490.7760.8510.14622.2
Yolov4-tiny [25]416 × 41622.57 M0.5660.9240.6140.4170.6830.7240.206160.3
Yolov4 [17]416 × 416243.9 M0.6450.9520.7240.510.7470.8060.14641.2
Yolov3 [51]416 × 416234.69 M0.6110.9530.680.4760.7140.8010.14554.4
Yolov4-lite [23]416 × 41646.79 M0.5760.9390.6290.4340.6780.7550.19356.1
SSD [18]416 × 41690.58 M0.5730.9170.6130.4050.6930.7340.27563.1
EfficientDet-D0 [26]512 × 51214.6 M0.5590.8910.6140.4130.6710.8080.31925.5
Yolov5-S [27]416 × 41626.95 M0.5430.90.5720.3690.6760.7930.26779.7
Yolox-S [30]416 × 41634.09 M0.6540.950.7210.4940.7730.8950.15376.6
TIPYOLO (ours)416 × 19234.09 M0.6630.9510.7360.5090.7780.8970.15675.0
Table 4. The comparison results of the methods on the UTMD dataset. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
Table 4. The comparison results of the methods on the UTMD dataset. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
MethodSizeParametersAP ↑AP50AP75APSAPMAPLMR−2FPS ↑
Faster R-CNN [48]320 × 256157.75 M0.5390.8840.5880.4520.6730.7330.2719.6
Yolov4-tiny [25]320 × 25622.57 M0.3520.6770.3460.2220.5730.6510.447162.5
Yolov4 [17]320 × 256243.9 M0.4390.8450.4030.3430.590.6580.28646.8
Yolov3 [51]320 × 256234.69 M0.420.8420.3750.3220.5740.6690.29665.8
Yolov4-lite [23]320 × 25646.79 M0.360.7570.3060.2390.5330.6620.42257.5
SSD [18]320 × 32090.58 M0.3950.7570.3450.2930.5470.6620.41785.5
EfficientDet-D0 [26]512 × 51214.6 M0.3660.6760.3570.2780.5280.590.53725.9
Yolov5-S [27]320 × 25626.95 M0.4010.7560.380.2990.5860.670.4678.3
Yolox-S [30]320 × 25634.09 M0.470.8370.4720.350.6460.7880.3275.0
TIPYOLO (ours)320 × 25634.09 M0.4840.8480.4810.3690.650.780.29870.6
Table 5. The comparison results of the methods on KAIST and CVC-14 datasets. In columns 2 to 8, the values to the left of “/” represent the test result on the KAIST data, and the values to the right of “/” represent the test result on the CVC-14 dataset. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
Table 5. The comparison results of the methods on KAIST and CVC-14 datasets. In columns 2 to 8, the values to the left of “/” represent the test result on the KAIST data, and the values to the right of “/” represent the test result on the CVC-14 dataset. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
MethodAP ↑AP50AP75APSAPMAPLMR−2
Faster R-CNN [48]0.73/0.6410.954/0.9670.843/0.7560.669/0.5120.698/0.6660.86/0.7440.078/0.101
Yolov4-tiny [25]0.464/0.450.92/0.830.395/0.4480.296/0.2670.439/0.5660.617/0.6330.217/0.298
Yolov4 [17]0.543/0.5630.963/0.9720.565/0.6050.417/0.450.536/0.5870.666/0.6550.063/0.097
Yolov3 [51]0.591/0.5670.971/0.9690.663/0.6190.437/0.4570.58/0.5920.706/0.6620.058/0.111
Yolov4-lite [23]0.529/0.5260.957/0.9540.514/0.5430.385/0.3660.499/0.5630.662/0.6240.103/0.148
SSD [18]0.631/0.5450.957/0.9390.719/0.5660.463/0.4050.608/0.5820.789/0.6200.101/0.215
EfficientDet-D0 [26]0.366/0.4030.799/0.7860.265/0.3620.21/0.1610.347/0.4680.51/0.6050.49/0.483
Yolov5-S [27]0.323/0.510.728/0.9110.21/0.5120.169/0.3220.311/0.5530.455/0.6620.538/0.283
Yolox-S [30]0.595/0.5520.924/0.920.655/0.5990.389/0.3910.646/0.6360.806/0.7470.162/0.206
TIPYOLO(Ours)0.695/0.6140.974/0.9670.808/0.7120.537/0.4610.676/0.6470.825/0.7270.05/0.112
Table 6. Effects of different quantization degrees on the method’s accuracy. In the first column, the first value in the first parentheses represents the numerical type of the weight, and the second value in the first parentheses represents the numerical type of the input. “to” represents converting the numeric type in the first parenthesis to int8 or int16. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
Table 6. Effects of different quantization degrees on the method’s accuracy. In the first column, the first value in the first parentheses represents the numerical type of the weight, and the second value in the first parentheses represents the numerical type of the input. “to” represents converting the numeric type in the first parenthesis to int8 or int16. “↑” indicates that the larger the value, the better the performance; “↓” indicates that the smaller the value, the better the performance.
StrategyAP ↑AP50AP75APSAPMAPLMR−2
unquantified0.6630.9510.7360.5090.7780.8970.156
(float32, float16) to (int16, int16)0.670 0.951 0.749 0.519 0.782 0.896 0.156
(float32, float16) to (int8, int8)0.664 0.950 0.745 0.514 0.774 0.888 0.158
(float32, float32) to (int16, int16)0.670 0.952 0.749 0.519 0.782 0.897 0.156
(float32, float32) to(int8, int8)0.662 0.952 0.741 0.514 0.771 0.888 0.152
Table 7. Effects of different quantization degrees on the method’s FPS. a, b, c, d, and e in the table, respectively, represent the five video processing methods in Figure 5, that is, the strategy corresponding to the five subfigures, where the two methods in strategy d are run in parallel.
Table 7. Effects of different quantization degrees on the method’s FPS. a, b, c, d, and e in the table, respectively, represent the five video processing methods in Figure 5, that is, the strategy corresponding to the five subfigures, where the two methods in strategy d are run in parallel.
Strategyabcde
(float32, float16) to (int16, int16)19.634.748.526.326.149.3
(float32, float16) to (int8, int8)20.935.451.226.626.553.4
(float32, float32) to (int16, int16)18.834.045.126.025.244.5
(float32, float32) to (int8, int8)19.234.247.026.226.046.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

You, S.; Ji, Y.; Liu, S.; Mei, C.; Yao, X.; Feng, Y. A Thermal Infrared Pedestrian-Detection Method for Edge Computing Devices. Sensors 2022, 22, 6710. https://doi.org/10.3390/s22176710

AMA Style

You S, Ji Y, Liu S, Mei C, Yao X, Feng Y. A Thermal Infrared Pedestrian-Detection Method for Edge Computing Devices. Sensors. 2022; 22(17):6710. https://doi.org/10.3390/s22176710

Chicago/Turabian Style

You, Shuai, Yimu Ji, Shangdong Liu, Chaojun Mei, Xiaoliang Yao, and Yujian Feng. 2022. "A Thermal Infrared Pedestrian-Detection Method for Edge Computing Devices" Sensors 22, no. 17: 6710. https://doi.org/10.3390/s22176710

APA Style

You, S., Ji, Y., Liu, S., Mei, C., Yao, X., & Feng, Y. (2022). A Thermal Infrared Pedestrian-Detection Method for Edge Computing Devices. Sensors, 22(17), 6710. https://doi.org/10.3390/s22176710

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop