Next Article in Journal
Blood Cell Attribute Classification Algorithm Based on Partial Label Learning
Next Article in Special Issue
Adaptive Dynamic Shuffle Convolutional Parallel Network for Image Super-Resolution
Previous Article in Journal
Design Techniques for Wideband CMOS Power Amplifiers for Wireless Communications
Previous Article in Special Issue
A Dynamic Network with Transformer for Image Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization

School of Electronic, Electrical, and Communication Engineering, University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2024, 13(9), 1697; https://doi.org/10.3390/electronics13091697
Submission received: 30 March 2024 / Revised: 17 April 2024 / Accepted: 22 April 2024 / Published: 27 April 2024
(This article belongs to the Special Issue Big Model Techniques for Image Processing)

Abstract

:
In recent years, the detection and localization of tiny persons have garnered significant attention due to their critical applications in various surveillance and security scenarios. Traditional multi-modal methods predominantly rely on well-registered image pairs, necessitating the use of sophisticated sensors and extensive manual effort for registration, which restricts their practical utility in dynamic, real-world environments. Addressing this gap, this paper introduces a novel non-registered multi-modal benchmark named NRPerson, specifically designed to advance the field of tiny person detection and localization by accommodating the complexities of real-world scenarios. The NRPerson dataset comprises 8548 RGB-IR image pairs, meticulously collected and filtered from 22 video sequences, enriched with 889,207 high-quality annotations that have been manually verified for accuracy. Utilizing NRPerson, we evaluate several leading detection and localization models across both mono-modal and non-registered multi-modal frameworks. Furthermore, we develop a comprehensive set of natural multi-modal baselines for the innovative non-registered track, aiming to enhance the detection and localization of unregistered multi-modal data using a cohesive and generalized approach. This benchmark is poised to facilitate significant strides in the practical deployment of detection and localization technologies by mitigating the reliance on stringent registration requirements.

1. Introduction

Person detection and localization is a critical research topic in computer vision, fundamental to numerous higher-level applications such as surveillance [1,2,3,4], tracking [5,6,7,8], and rapid rescue operations [9,10]. Despite longitudinal studies and significant progress, person detection continues to face challenges such as small object scale, variable lighting, and complex backgrounds, which severely hinder the accuracy and reliability of detection systems.
Detecting and localizing tiny-sized objects within an image has become a focus for many computer vision researchers due to the substantial technical challenges and significant application value it presents [9,10,11,12,13,14,15]. The detection of tiny persons is especially crucial in wide-area and long-range scenarios, where precise identification can aid in critical tasks such as monitoring crowded events or coordinating rescue missions in expansive environments.
Current studies predominantly utilize mono-modal RGB images for detection tasks [9,10,11,12,13,14,15,16,17,18,19,20]. However, this approach often fails to address all environmental challenges, as RGB data alone may not provide sufficient information under poor lighting or when objects blend into complex backgrounds.
Multi-modal data, integrating inputs from various sensor types such as RGB and infrared, have proven effective in enhancing mono-modal detection tasks [3,21,22,23,24,25,26,27,28,29]. A common practice in multi-modal person detection tasks assumes that the image pairs are well registered [21,30,31,32,33,34,35]. However, this requirement for precise sensor alignment is often impractical in dynamic real-world conditions and leads to significant challenges in deployment due to the complex registration processes required.
Recognizing these limitations, we introduce NRPerson, a novel benchmark for non-registered multi-modal tiny person detection and localization. This benchmark aims to explore the utility of multi-modal image pairs without the stringent necessity for exact registration, thus more closely mimicking real-world scenarios, where perfect alignment of sensors cannot be guaranteed. NRPerson comprises 8548 RGB-IR image pairs collected from 22 video sequences, supported by 889,207 high-quality annotations that are manually verified to ensure their accuracy and reliability (see Figure 1).
This approach not only simplifies the data collection process by reducing dependency on sophisticated registration techniques but also enhances the practical application of detection systems in diverse operational environments. We evaluate several leading detection models on both mono-modal and non-registered multi-modal tracks to assess their effectiveness in leveraging non-registered data for improved performance.
The contributions of this paper are significant in advancing the field of multi-modal person detection:
  • We establish a large-scale NRPerson benchmark, introducing the concept of non-registration in multi-modal data handling for tiny person detection and localization, thereby promoting tasks that align closer to real-world scenarios.
  • We perform a comprehensive evaluation of state-of-the-art detection and localization models on this new benchmark, demonstrating the potential and limitations of existing technologies when applied to non-registered multi-modal data.
  • We develop a set of diverse natural multi-modal baselines that effectively utilize non-registered data, laying a robust foundation for future research in multi-modal detection and localization.
In the subsequent sections, we detail the methodology for constructing the NRPerson dataset, describe our evaluation framework for detection models, and discuss the broader implications of our findings for the field.

2. Related Work

2.1. Multi-Modal Person Detection

Multi-modal person detection has become a popular topic in the research community, and many state-of-the-art multi-modal person detectors [21,30,31,32,33,34,35,36] have been proposed. ProbEn [37] fuses outputs from multimodal detectors such as RGB and thermal cameras. This approach leverages Bayes’ rule to calculate a combined score that more accurately reflects the likelihood of object presence, thereby improving detection performance even in poor illumination—a common challenge in person detection scenarios.
A common assumption for these multi-modal detectors is that the image pairs are well registered. However, in real-world scenarios, the multi-modal registration assumption is hard to come into existence, and research [36] has revealed that even manually calibrated multi-modal data still suffer from weak alignment. Weak alignment is also a relatively ideal situation that still requires complex operation, while the actual data collection situation is more complicated and changeable. Therefore, we establish a non-registered multi-modal benchmark to change the prerequisites for good alignment, thus getting rid of the reliance on manual calibration to some extent.

2.2. Tiny Object Detection

Recently, many researchers have devoted themselves to the study of tiny object detection [9,13,16,17,18,19,20], but research in the field remains limited. The low resolution of tiny objects makes feature representation insufficient, leading to poor performance compared to large objects. CEFP2N [38] introduce a feature pyramid network enhanced by a Context Enhancement Module (CEM) and a Feature Purification Module (FPM) for tiny object detection. Their approach augments context information and purifies features, significantly improving detection accuracy with a novel “copy-reduce-paste” data augmentation method. Accordingly, we are motivated to propose NRPerson, aiming at detection task in multi-modal and long-distance scenarios for tiny persons. Based on NRPerson, multi-modal data can be used to complement instance feature information to alleviate the lack of feature representation capability.

2.3. Point-Based Object Localization

Point-based object localization aims to train with point-level annotations, thereby learning the capability to predict the location coordinates of objects [10,12,39]. Unlike object detection, object localization is not interested in the object’s scale. Therefore, costly and time-consuming box-level annotations are unnecessary, and simple and time-saving point-level annotations are more promising and valuable. CPR [10] introduces a general object localization method based on coarse point supervision and provides pseudo boxes generated by refined point annotations. As a result, the point-to-point learning process can be converted to a box-to-box one. CPR++ [40] advances the field of point-based object localization by refining coarse point annotations through a dynamic, multi-stage process that progressively reduces semantic variance, improving localization accuracy. Point-based object localization is in the ascendant, and the proposed NRPerson provides a large-scale benchmark and extends to multi-modal data.

2.4. Other Public Datasets

Representative public datasets that are relevant to NRPerson are summarized in Table 1. They can be divided into three categories: (1) multi-modal datasets [6,22,23,24,25,41,42], (2) person detection/localization datasets [1,2,5,9,10,43,44,45] and (3) multi-modal person detection/localization datasets [3,8,21,46]. OSU [41] is a thermal-visible video fusion dataset for moving target tracking and pedestrian motion analysis and classification. LITIV [6] consists of videos for tracking scenarios captured through thermal and visual cameras with different zoom settings and positions. CityPersons [2] is a rich and diverse pedestrian detection dataset. It is recorded by a car traversing various cities and contains dense pedestrians. TinyPerson [9] is the first benchmark for long-distance tiny person detection. KAIST [21] provides well-aligned color-thermal image pairs collected in various traffic scenes day and night. LLVIP [3] is a visible-infrared paired dataset containing a large number of pedestrians with low brightness. NRPerson is a large-scale multi-modal tiny person detection and localization dataset. Unlike other datasets based on manual registration and provide modality-shared annotations, NRPerson first introduces the concept of non-registration, which is closer to real-world scenarios.

3. The NRPerson Dataset

3.1. Data Collection and Annotation

Data Collection. The equipment is a binocular camera platform consisting of an RGB camera and an IR camera. As we are committed to exploring the detection challenges in wild maritime quick rescue, images containing many person objects are captured from various seaside scenes. We sample images from 22 video sequences, and discard images with no objects or high homogeneity. After time alignment and manual filtering, time-synchronized and high-quality image pairs containing persons are selected. The resolutions of RGB and IR images are 1920 × 1080 and 960 × 576 pixels, respectively.
Data Annotation. We use three types of labels to annotate objects person, ignore, and uncertain. Distinguishable persons are annotated as person by accurate bounding boxes. Areas with dense crowds, reflections, or ambiguities are annotated as ignore. Some objects are difficult to recognize as persons, and thus we annotate them as uncertain. Among the filtered 8548 RGB-IR image pairs, a total of 889,207 persons with bounding boxes are manually annotated and carefully double-checked, including 471,924 persons in RGB images and 417,283 persons in IR images as shown in Table 1 and Table 2. In addition, we follow [10] to generate coarse point annotation in each labeled bounding box.
Training and Test Sets. The training, validation and test sets are constructed by randomly dividing the images into three subsets, and images from the same video sequence cannot be divided into different subsets. The results of the dataset division are summarized in Table 2. Specifically, the training set contains 8 video sequences with 4614 image pairs and 346,413 annotations, the validation set contains 1 video sequence with 375 image pairs and 33,385 annotations, and the test set contains 13 video sequences with 3559 image pairs and 509,409 annotations.

3.2. Dataset Properties

Time Alignment. RGB and IR video sequences are captured simultaneously using a binocular camera. The two video sequences are temporally aligned through manual calibration. Then, we sample the frames and ensure that the capturing time for each image pair is identical to guarantee strict time alignment.
Spatial Overlapping. Although time-aligned image pairs are not geometrically aligned due to the different physical properties of RGB and IR sensors, the shooting angles of the binocular sensors are very close. Therefore, the image pairs have spatial overlap. Manual filtering is adopted to remove anomalies and ensure that each image pair has spatial overlap.
Non-registration. Non-registration means that RGB-IR image pairs are time-aligned and spatial-overlapped but are not geometrically aligned. Non-registered image pairs have the following three manifestations:
(1)
Position shift: Sensor differences and external factors lead to position shifts between corresponding instances in image pairs. This is the most common phenomenon, and even manually registered data often cannot completely eliminate it.
(2)
Size difference: Differences in the field of view and image pixels captured by RGB and IR sensors lead to the size difference of the corresponding instances. This can basically be eliminated through manual calibration.
(3)
No correspondence: There is no correspondence between the pixel coordinates of non-registered image pairs. Hence, instances between different modalities cannot be matched and paired.
Tiny. Absolute size is defined as the number of pixels in the bounding box area, and relative size is the absolute size divided by the image pixels. As shown in Table 3 and Figure 2, the absolute size and relative size of persons in NRPerson are very small compared to other representative datasets [2,3,8,9,10,21]. Objects with small absolute scales are blurred and contain less semantic information, which increases the difficulty of the detector to identify the objects and greatly affects the performance [9]. Tiny relative scales result in the scenes having a large number of complex backgrounds, which exacerbates the issues of positive–negative imbalance and false positives, hence being more challenging and exploratory.
Aspect Ratio Diversity. The aspect ratio of persons in NRPerson has a large variance as shown in Table 3. It suggests that unlike conventional person detection/localization datasets with relatively single perspectives and poses, our dataset is an effective complement in terms of aspect ratio diversity, bringing more complexity and making detection and localization more challenging.
Privacy. Personal privacy is a sensitive matter and an important issue in person-centered research, such as person detection, face detection, and person re-identification. For the NRPerson benchmark, objects are tiny persons with low resolution at long distances (about 24 pixels per person on average, as shown in Table 3). Therefore, little personal information emerges through the data. Additionally, we will enforce dataset terms before releasing NRPerson:
(1)
Sharing dataset images on the Internet in any form without permission is prohibited;
(2)
Identifiable images of people cannot be used to make demos or promotions;
(3)
Use face obfuscation or shield face regions when applicable and study its impact on detection and localization [47].

4. Tracks and Metrics

4.1. Mono-Modal Track

The NRPerson dataset contains many tiny-sized objects, as shown in Figure 1, making it a favorable complement to the existing tiny object detection datasets. In addition, other datasets are basically based on RGB images but rarely on IR images. NRPerson bridges this gap and facilitates the development of research on IR tiny object detection and localization in the community.
We establish a mono-modal track on NRPerson that follows the task settings of generic object detection [48,49] and localization [10,12]. Object detection has been widely studied and is well known, so we do not describe it further. Object localization refers to training the model with point-level annotations to predict the objects’ locations. There are various settings of point-level annotations in previous studies [10,12,50]. In this paper, we choose a more generalized annotation, i.e., a single point annotation with a random position on each object.
In addition to benchmarking the performance of detectors and localizers under their respective tasks, we also try to explore the performance of the same model structure under multiple tasks. However, detectors and localizers have different learning supervision and inference results, so it is challenging to perform multi-task unification directly. According to [10], we convert the point-to-point problem into a box-to-box, enabling the detectors to achieve object localization under point supervision. Specifically, CPR [10] is adopted to refine the initial coarse point annotations, and then pseudo boxes are generated with the points as the center that serve as supervision for model training. During the inference, the center points of the predicted boxes are regarded as the localization results. In this way, we build a multi-task mono-modal track on NRPerson to report the detection and localization performance of representative models.
Detection metrics. We adopt average precision ( AP ) as the evaluation metric for object detection. Since many applications of tiny person detection (e.g., shipwreck search and rescue) focus more on finding the approximate location of a person rather than the exact bounding box coordinates, we adopt AP 50 as the evaluation metric, i.e., the Intersection over Union (IoU) threshold is set to 0.5. For more detailed experimental comparisons, bounding box sizes are divided into three intervals: t ( t i n y ) [ 2 , 20 ] , s ( s m a l l ) ( 20 , 32 ] , and  r ( r e a s o n a b l e ) ( 32 , ) .
Localization metrics. We follow the evaluation metric of object localization adopted in [10]. Specifically, the distance d between point p = ( x , y ) and bounding box b = ( x c , y c , w , h ) is calculated as:
d ( p , b ) = x x c w 2 + y y c h 2 ,
where ( x c , y c ) , w, h are the center point coordinates, width, and height of the bounding box. A point matches a ground-truth bounding box if the distance d is smaller than the threshold τ . If there are multiple matching points through a ground-truth, the point with the highest score is selected; if a point has multiple matching ground-truths, the ground-truth with the smallest distance d is selected. A point is counted as a true positive (TP) when it matches the ground-truth, and a false positive (FP) when it does not. If a point matches an object annotated as ignore, neither TP nor FP are counted, following [2,9]. Likewise, we adopt AP τ as the evaluation metric and set τ = 1.0 as the default.

4.2. Non-Registered Multi-Modal Track

We introduce the concept of non-registration on the basis of multi-modal tiny person detection and localization, thereby establishing a non-registered multi-modal track. The goal is to directly learn from non-registered data such that the model achieves robust detection and localization capabilities in each modality. A flowchart of non-registered multi-modal model is shown in Figure 3.
Task Definition. The RGB image set is denoted as X r g b = { x r 1 , x r 2 , , x r n } and the IR image set is denoted as X i r = { x i 1 , x i 2 , , x i n } , where x r k and x i k represent a non-registered RGB-IR image pair satisfying both time-alignment and spatial-overlapping. We require the model M to input non-registered images in pairs and output the results for each modality, i.e., ( y r k , y i k ) = M ( x r k , x i k ) , where y r k and y i k denote the detection/localization results of RGB and IR images, respectively. During the evaluation, the multi-modal test set is used to comprehensively evaluate the performance across all modalities.
Dual-Stream Input. The conventional detectors and localizers are based on a single image input. To meet the requirements of the non-registered multi-modal track, we adopt a dual-stream input and a paired sampling strategy in multi-modal baselines. Dependent on the time-alignment, the RGB set X r g b and the IR set X i r are pair-sampled to form an image pair set, i.e., X p a i r = { ( x r 1 , x i 1 ) , ( x r 2 , x i 2 ) , , ( x r n , x i n ) } . To guarantee training robustness and accelerate parameter convergence, random shuffling is adopted to disrupt the order of image pairs. The shuffled image pairs are fed into the multi-modal model as dual-stream input.
Evaluation Metrics. The evaluation of detectors and localizers on the non-registered multi-modal track is still based on the metrics described in Section 4.1. Unlike the mono-modal track, the model in this track takes multi-modal data as dual-stream input and outputs paired results. Therefore, three models are designed for evaluation: RGB-test, IR-test, and multi-test. RGB-test and IR-test represent the evaluation of the RGB and IR test sets, respectively. Multi-test is simply combining the results of the RGB and IR test sets and evaluating them together without using additional data. We report the multi-test in order to consider both modalities’ performance simultaneously and to facilitate comparisons between models.

4.3. Advantages and Challenges of Non-Registration

A common assumption of conventional multi-modal tasks is that the image pairs are well registered. However, performing the registration is a cumbersome operation that requires sophisticated sensors and heavy human efforts. Specifically, multi-modal image registration suffers from the following major problems:
(1)
Multi-modal images often have different resolutions and fields of view due to the different sensors in the cameras. Therefore, it is almost impossible to achieve image registration simply by calibrating the inner parameters. It is also difficult to ensure that the image acquisition process is free from external disturbances even after the parameter calibration.
(2)
(Automatic feature matching methods may not work well on data with complicated scenes and across modalities. A typical example is shown in Figure 4, where we use SURF [51] to achieve feature point extraction and matching. Our dataset cannot rely on such unsatisfactory matching results for automatic registration. Further, for tiny object analysis, matching key points for registration will be more difficult due to the low signal-to-noise ratio.
(3)
The above two problems illustrate the difficulty of achieving the automatic alignment of multi-modal images. Therefore, the existing multi-modal dataset [3,21,46] almost relies on manual registration that requires processing each image pair and relies on a large amount of manpower. In addition, research shows that in practice, even well-processed data still suffer from weak alignment problems [36] (i.e., position shifts), which degrades model performance.
(4)
Our dataset focuses on tiny objects at a distance. In this case, registration becomes more complicated since long-distance shots amplify small deviations. In scenes with dense and minute-scale objects, slight positional offsets may confuse the correspondence between objects in different modes.
In summary, multi-modal image registration is complex and not guaranteed to be completely accurate, and non-registration is an unavoidable problem.
The above problems undoubtedly limit the development and application of multi-modal technology. Meanwhile, they confirm the necessity of introducing non-registration issues. The non-registered multi-modal track attempts to address these shortcomings by omitting the registration operation and only requiring the multi-modal image pairs to be time-aligned and spatial-overlapped. This not only saves time and manpower for registration but also performs data post-processing and object detection/localization in an end-to-end manner, greatly facilitating the generalization of multi-modal tasks to practical applications. However, the omission of registration results in the lack of instance-level correspondence established between image pairs. Therefore, using the multi-modal information effectively is challenging to improve detection performance.

5. Experiments

5.1. Experimental Settings

Our codes are based on MMDetection [52], and we use 8 RTX 3090 GPUs for training and evaluation.
Image cutting and pairing. Due to the limited memory of GPUs, we crop the original images into sub-images with overlaps during training and inference. We resize images of different resolutions to the same size (i.e., 1920 × 1080 pixels) and cut out a series of 640 × 640 patches. For the non-registered multi-modal track, multi-modal images are input in pairs, and thus the cropped sub-images need to be paired one by one. Sub-images at the corresponding positions in the original multi-modal image pair are formed into a new image pair, and the new paired images can be fed into the multi-modal network. According to previous works [48,49,53,54], pure background images (i.e., images without objects) are not used for training. For an image pair, there may be a situation where one image has objects, and the other does not. In our experiment, we filter out the image pairs in this situation and keep the pairs that both contain objects.

5.1.1. Mono-Modal Tiny Person Detection

In this detection track, Faster R-CNN [48], (Adaptive) RetinaNet [55], (Adaptive) RepPoints [56], Sparse R-CNN [57], Deformable DETR [58] and Swin-T (Faster R-CNN) [59] are included as the detectors. Faster R-CNN is based on the feature pyramid network (FPN) [16] with a ResNet-50 [60] backbone network. The number of training epochs is set to 12, and the stochastic gradient descent (SGD) optimizer is used with a learning rate of 0.01, decayed by 0.1 at the 8th and 11th epochs, respectively. Anchor sizes are set to [ 8 , 16 , 32 , 64 , 128 ] , and aspect ratios are set to [ 0.5 , 1.0 , 2.0 ] . Since some images have dense objects in NRPerson, the maximum number of detection results per image is set to 1000 for evaluation.
To increase the data diversity and model robustness, we adopt various data augmentation methods, including random flipping, random cropping, and photometric distortion. There are eight kinds of transformations for photometric distortion: random brightness, random contrast, convert color from BGR to HSV, random saturation, random hue, convert color from HSV to BGR, random contrast, and randomly swap channels. In this paper, this series of data augmentation methods are identified as standard augmentation.
RetinaNet and RepPoints use the same experimental settings as Faster R-CNN and also adopt the standard augmentation.
Adaptive RetinaNet and Adaptive RepPoints denote the models with FPN adjustments for RetinaNet and RepPoints, respectively, which use the ( P 2 , P 3 , P 4 , P 5 , P 6 ) of the FPN instead of the default ( P 3 , P 4 , P 5 , P 6 , P 7 ) . The purpose of this adjustment is to use features at a lower layer, which is more friendly for detecting tiny objects. The rest of the experimental settings are kept the same as RetinaNet and RepPoints.
Sparse R-CNN sets the experimental super-parameters somewhat differently from the above models. The number of training epochs is set to 36; the Adam optimizer is used with a learning rate of 2.5 × 10−5, decayed by 0.1 at the 27th and 33rd epochs, respectively; and the number of learnable proposal boxes is set to 1000. In addition, data augmentation includes random scale jittering and random flipping.
Deformable DETR sets the number of training epochs to 50, uses the Adam optimizer with a learning rate of 1 × 10−4, decayed by 0.1 at the 40th epoch and sets the number of queries to 1000. According to the paper, random scale jittering, random flipping, and random cropping augmentation are used.
Swin-T is a strong backbone which performs well in a broad range of vision tasks. We conduct our experiments with it on Faster R-CNN and keep the same settings as Faster R-CNN above except for the Adam optimizer with a learning rate of 1 × 10−4 and the 50 training epochs.

5.1.2. Mono-Modal Tiny Person Localization

In this localization track, P2PNet [12] is included as the localizer, and CPR [10] acts as the refinement module to optimize coarse point annotations. In addition, we convert the point-to-point problem in tiny person localization into a box-to-box one as described in Section 4.1. In this way, the above-mentioned detectors can be benchmarked in the localization track to explore the performance on multiple tasks.
P2PNet is specifically designed for point-based object localization. We follow the improvements to P2PNet in [10]: (1) We adopt ResNet-50 as the backbone network with P 2 of FPN as the output features. (2) The focal loss [55] is used when optimizing classification to better deal with the problem of imbalance. (3) The smooth- L 1 loss [61] is used for regression. (4) We assign top-k positive samples for each ground-truth and regard the remaining samples as background in label assignment, and then perform non-maximum suppression (NMS) [62] on the points to obtain the final results. The number of training epochs of P2PNet is set as 12, and the learning rate is set as 1 × 10−4, decayed by 0.1 at the 8th and 11th epochs, respectively.
CPR is a refinement module used to reduce the semantic variance during annotation. Hence, it can be combined with any localizer to provide refined point annotations. For CPR, the number of training epochs is set as 12, and the learning rate is set as 1 × 10−3, decayed by 0.1 at the 8th and 11th epochs, respectively. The sampling radius R is set as 5.
Other localizers are converted from detectors in the detection track. They are trained with pseudo boxes refined and generated by CPR and adopt the same experimental settings as in the detection track.

5.1.3. Result Analysis

The detection and localization results of the mono-modal track are summarized in Table 4.
Detection. According to Figure 2, there are a large number of small-scale objects in NRPerson, but the detection performance for these tiny objects (in terms of AP 50 t ) is not satisfactory, indicating that our dataset is a challenging benchmark for the state-of-the-art person detection algorithms. For objects with small scales, the spatial information of the network may be more important than the deeper layers. Therefore, Adaptive RetinaNet and Adaptive RepPoints improve performance (+1.7 and +1.0 AP 50 , respectively) by adopting a lower layer of FPN to obtain more spatial information about objects. In particular, the detection performance for objects whose scales are tiny is significantly improved.
Localization. It can be seen that after the refinement coarse annotations by CPR, the models achieve highly competitive localization performance on our dataset. Different from the detection results, better localization performance is achieved on smaller objects, which indicates that small objects are not sensitive to the semantic variance [10] caused by coarse point annotations. Moreover, a reasonable object may be predicted with multiple results due to its large area, whereas a tiny object is an opposite. Consequently, there are fewer false positives on the small objects. The impressive performance on small objects demonstrates the significance and the application value of the localization task, especially in scenarios in which only the location of the object rather than the scale is concerned.

5.2. Non-Registered Multi-Modal Track

Multi-Modal Baseline. Since the non-registered multi-modal track is new, we construct a diverse set of natural multi-modal baselines. The multi-modal baselines employ the same models and experimental settings as in the mono-modal track and are simultaneously trained with RGB-modal and IR-modal data to explore adaptation to multi-modal detection and localization. Moreover, a comparison of multi-modal baselines with the mono-modal track can be used to explore the effect of adding additional data from other modalities on performance.
Result Analysis. The detection and localization results of the non-registered multi-modal track are summarized in Table 5. Although the correlation information between different modalities is not intentionally used in the multi-modal baselines, adding data from other modalities for training leads to the improved performance of many models (compared to those in the mono-modal track), especially in the RGB modality. Such performance improvement is reasonable and shows that if the complementary information between the multi-modal data is fully exploited, it will have a more positive impact on detection performance. The results provide sound empirical evidence on the superiority of multi-modal baselines in terms of their accuracy and applicability. However, there are also some models that show a degradation in performance, suggesting that simply adding multi-modal data does not always have a positive effect. It indicates the need to explore more effective fusion methods to fully utilize multi-modal information in future studies.
Visualization. We select some qualitative results for the visualization of Adap-RepPoints (with the best detection performance) and P2PNet (with the best localization performance) as shown in Figure 5. Since the size of objects in NRPerson is relatively small, we partially zoom in on the original images for display. The visualization results reflect that the models generally exhibit robust detection and localization capabilities. However, the models still have missing or wrong results for some areas where the objects are dense or blurred. This shows that our dataset is exceptionally challenging and needs to be continuously improved by follow-up research.

6. Conclusions

This paper contributes a non-registered multi-modal benchmark for tiny person detection and localization, called NRPerson. It introduces the concept of non-registration into the multi-modal task, eliminating the requirement of multi-modal image registration based on sophisticated sensors and heavy human efforts. We benchmark the representative models on the mono-modal track and the newly defined non-registered multi-modal track, benefiting from the empirical results to understand the detection/localization ability of different models and the impact of using multi-modal data. We construct a diverse set of natural multi-modal baselines, attempt to leverage non-registered multi-modal data for multi-task with a unified framework, and lay a robust foundation for facilitating future research.
We are committed to continuing our research using the NRPerson benchmark. Our future work will focus on exploring the detection of weak and small targets and addressing challenges of non-alignment in pretraining scenarios. Additionally, we plan to tackle the registration issues in non-aligned multimodal data, aiming to enhance the applicability and accuracy of detection methods under these complex conditions.
While our dataset is challenging, we believe that the baselines provided in this paper are far from reaching the upper limit of performance. Therefore, how to better utilize non-registered multi-modal data to deal with tiny objects is worth exploring. In the future, we will continue investigating more effective and robust models for the two tracks. We hope NRPerson can pave the way for follow-up studies in this area.

Author Contributions

Conceptualization, Y.Y. and X.H.; methodology, Y.Y. and X.H.; software, Y.Y., X.H. and K.W.; validation, Y.Y., X.H. and K.W.; formal analysis, Y.Y., X.H. and X.Y.; investigation, Y.Y. and X.Y.; resources, Y.Y., Z.H. and J.J.; data curation, Y.Y. and K.W.; writing—original draft preparation, Y.Y. and X.H.; writing—review and editing, Y.Y., Z.W. and W.Y.; visualization, Y.Y., Z.W. and W.Y.; supervision, Z.H., G.L. and J.J.; project administration, Z.H., G.L. and J.J.; funding acquisition, Z.H. and J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data supporting the results reported in this study are openly available in a public repository. The datasets analyzed during the current study are available at the following GitHub repository: https://github.com/NRPerson/NRPerson. This repository includes all the datasets that were analyzed or generated during the course of this study. Further details and access to the data can be found directly via the provided URL. Researchers who wish to use these datasets are encouraged to cite this article and reference the data source appropriately in their work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dollár, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian detection: A benchmark. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  2. Zhang, S.; Benenson, R.; Schiele, B. Citypersons: A diverse dataset for pedestrian detection. In Proceedings of the In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  3. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A Visible-infrared Paired Dataset for Low-light Vision. arXiv 2021, arXiv:2108.10831. [Google Scholar]
  4. Zhang, Y.; Bai, Y.; Ding, M.; Xu, S.; Ghanem, B. KGSNet: Key-Point-Guided Super-Resolution Network for Pedestrian Detection in the Wild. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2251–2265. [Google Scholar] [CrossRef] [PubMed]
  5. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
  6. Torabi, A.; Massé, G.; Bilodeau, G.A. An iterative integrated framework for thermal–visible image registration, sensor fusion, and people tracking for video surveillance applications. Comput. Vis. Image Underst. 2012, 116, 210–221. [Google Scholar] [CrossRef]
  7. Wu, Z.; Fuller, N.; Theriault, D.; Betke, M. A thermal infrared video benchmark for visual analysis. In Proceedings of the CVPRW, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  8. FLIR, T. Free Flir Thermal Dataset for Algorithm Training. 2018. Available online: https://www.flir.com/oem/adas/adas-dataset-form/ (accessed on 29 March 2024).
  9. Yu, X.; Gong, Y.; Jiang, N.; Ye, Q.; Han, Z. Scale match for tiny person detection. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  10. Yu, X.; Chen, P.; Wu, D.; Hassan, N.; Li, G.; Yan, J.; Shi, H.; Ye, Q.; Han, Z. Object Localization under Single Coarse Point Supervision. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  11. Yang, S.; Luo, P.; Loy, C.C.; Tang, X. Wider face: A face detection benchmark. In Proceedings of the CVPR, Las Vegas, NV, USA, 17–30 June 2016. [Google Scholar]
  12. Song, Q.; Wang, C.; Jiang, Z.; Wang, Y.; Tai, Y.; Wang, C.; Li, J.; Huang, F.; Wu, Y. Rethinking counting and localization in crowds: A purely point-based framework. In Proceedings of the ICCV, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  13. Gong, Y.; Yu, X.; Ding, Y.; Peng, X.; Zhao, J.; Han, Z. Effective fusion factor in FPN for tiny object detection. In Proceedings of the WACV, Waikoloa, HI, USA, 3–8 January 2021. [Google Scholar]
  14. Jiang, N.; Yu, X.; Peng, X.; Gong, Y.; Han, Z. SM+: Refined Scale Match for Tiny Person Detection. In Proceedings of the ICASSP, Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
  15. Yu, X.; Han, Z.; Gong, Y.; Jan, N.; Zhao, J.; Ye, Q.; Chen, J.; Feng, Y.; Zhang, B.; Wang, X.; et al. The 1st tiny object detection challenge: Methods and results. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  16. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  17. Singh, B.; Davis, L.S. An analysis of scale invariance in object detection snip. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  18. Cao, G.; Xie, X.; Yang, W.; Liao, Q.; Shi, G.; Wu, J. Feature-fused SSD: Fast detection for small objects. In Proceedings of the ICGIP, Chengdu, China, 12–24 December 2018. [Google Scholar]
  19. Duan, K.; Du, D.; Qi, H.; Huang, Q. Detecting small objects using a channel-aware deconvolutional network. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 1639–1652. [Google Scholar] [CrossRef]
  20. Hu, X.; Xu, X.; Xiao, Y.; Chen, H.; He, S.; Qin, J.; Heng, P.A. SINet: A scale-insensitive convolutional neural network for fast vehicle detection. IEEE Trans. Intell. Transp. Syst. 2018, 20, 1010–1019. [Google Scholar] [CrossRef]
  21. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; So Kweon, I. Multispectral pedestrian detection: Benchmark dataset and baseline. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  22. Li, C.; Zhao, N.; Lu, Y.; Zhu, C.; Tang, J. Weighted sparse representation regularized graph learning for RGB-T object tracking. In Proceedings of the ACM MM, Mountain View, CA, USA, 23–27 October 2017. [Google Scholar]
  23. Li, C.; Liang, X.; Lu, Y.; Zhao, N.; Tang, J. RGB-T object tracking: Benchmark and baseline. Pattern Recognit. 2019, 96, 106977. [Google Scholar] [CrossRef]
  24. Nguyen, D.T.; Hong, H.G.; Kim, K.W.; Park, K.R. Person recognition system based on a combination of body images from visible light and thermal cameras. Sensors 2017, 17, 605. [Google Scholar] [CrossRef] [PubMed]
  25. Wu, A.; Zheng, W.S.; Gong, S.; Lai, J. RGB-IR person re-identification by cross-modality similarity preservation. Int. J. Comput. Vis. 2020, 128, 1765–1785. [Google Scholar] [CrossRef]
  26. Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Zhao, J.; Guo, G.; Han, Z. Anti-UAV: A large multi-modal benchmark for UAV tracking. arXiv 2021, arXiv:2101.08466. [Google Scholar]
  27. Sun, Z.; Zhao, F. Counterfactual attention alignment for visible-infrared cross-modality person re-identification. Pattern Recognit. Lett. 2023, 168, 79–85. [Google Scholar] [CrossRef]
  28. Luo, X.; Jiang, Y.; Wang, A.; Wang, J.; Zhang, Z.; Wu, X. Infrared and visible image fusion based on Multi-State contextual hidden Markov Model. Pattern Recognit. 2023, 138, 109431. [Google Scholar] [CrossRef]
  29. Xu, M.; Tang, L.; Zhang, H.; Ma, J. Infrared and visible image fusion via parallel scene and texture learning. Pattern Recognit. 2022, 132, 108929. [Google Scholar] [CrossRef]
  30. Dollár, P.; Appel, R.; Belongie, S.J.; Perona, P. Fast Feature Pyramids for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1532–1545. [Google Scholar] [CrossRef] [PubMed]
  31. Li, C.; Song, D.; Tong, R.; Tang, M. Illumination-aware faster R-CNN for robust multispectral pedestrian detection. Pattern Recognit. 2019, 85, 161–171. [Google Scholar] [CrossRef]
  32. Liu, J.; Zhang, S.; Wang, S.; Metaxas, D.N. Multispectral deep neural networks for pedestrian detection. In Proceedings of the BMVC, York, UK, 19–22 September 2016. [Google Scholar]
  33. Konig, D.; Adam, M.; Jarvers, C.; Layher, G.; Neumann, H.; Teutsch, M. Fully convolutional region proposal networks for multispectral person detection. In Proceedings of the CVPRW, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  34. Xu, D.; Ouyang, W.; Ricci, E.; Wang, X.; Sebe, N. Learning cross-modal deep representations for robust pedestrian detection. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  35. Zhang, L.; Liu, Z.; Zhang, S.; Yang, X.; Qiao, H.; Huang, K.; Hussain, A. Cross-modality interactive attention network for multispectral pedestrian detection. Inf. Fusion 2019, 50, 20–29. [Google Scholar] [CrossRef]
  36. Zhang, L.; Zhu, X.; Chen, X.; Yang, X.; Lei, Z.; Liu, Z. Weakly aligned cross-modal learning for multispectral pedestrian detection. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  37. Chen, Y.T.; Shi, J.; Ye, Z.; Mertz, C.; Ramanan, D.; Kong, S. Multimodal object detection via probabilistic ensembling. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 139–158. [Google Scholar]
  38. Xiao, J.; Guo, H.; Zhou, J.; Zhao, T.; Yu, Q.; Chen, Y.; Wang, Z. Tiny object detection with context enhancement and feature purification. Expert Syst. Appl. 2023, 211, 118665. [Google Scholar] [CrossRef]
  39. Ribera, J.; Guera, D.; Chen, Y.; Delp, E.J. Locating objects without bounding boxes. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  40. Yu, X.; Chen, P.; Wang, K.; Han, X.; Li, G.; Han, Z.; Ye, Q.; Jiao, J. CPR++: Object Localization via Single Coarse Point Supervision. IEEE Trans. Pattern Anal. Mach. Intell. 2024. [Google Scholar] [CrossRef] [PubMed]
  41. Leykin, A.; Ran, Y.; Hammoud, R. Thermal-visible video fusion for moving target tracking and pedestrian classification. In Proceedings of the CVPR, Virtual, 19–25 June 2007. [Google Scholar]
  42. Toet, A. TNO Image Fusion Dataset. 2014. Available online: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029/1 (accessed on 29 March 2024).
  43. Shao, S.; Zhao, Z.; Li, B.; Xiao, T.; Yu, G.; Zhang, X.; Sun, J. Crowdhuman: A benchmark for detecting human in a crowd. arXiv 2018, arXiv:1805.00123. [Google Scholar]
  44. Xu, Z.; Zhuang, J.; Liu, Q.; Zhou, J.; Peng, S. Benchmarking a large-scale FIR dataset for on-road pedestrian detection. Infrared Phys. Technol. 2019, 96, 199–208. [Google Scholar] [CrossRef]
  45. Lin, T.; Maire, M.; Belongie, S.J.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the ECCV, Zurich, Switzerland, 6–12 September 2014; Volume 8693, pp. 740–755. [Google Scholar]
  46. González, A.; Fang, Z.; Socarras, Y.; Serrat, J.; Vázquez, D.; Xu, J.; López, A.M. Pedestrian detection at day/night time with visible and FIR cameras: A comparison. Sensors 2016, 16, 820. [Google Scholar] [CrossRef]
  47. Yang, K.; Yau, J.; Fei-Fei, L.; Deng, J.; Russakovsky, O. A study of face obfuscation in imagenet. arXiv 2021, arXiv:2103.06191. [Google Scholar]
  48. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef] [PubMed]
  49. Zhao, Z.; Zheng, P.; Xu, S.; Wu, X. Object Detection With Deep Learning: A Review. IEEE Trans. Neural Networks Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef]
  50. Chen, P.; Yu, X.; Han, X.; Hassan, N.; Wang, K.; Li, J.; Zhao, J.; Shi, H.; Han, Z.; Ye, Q. Point-to-Box Network for Accurate Object Detection via Single Point Supervision. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  51. Bay, H.; Tuytelaars, T.; Gool, L.V. Surf: Speeded up robust features. In Proceedings of the ECCV, Crete, Greece, 5–11 September 2006. [Google Scholar]
  52. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  53. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the CVPR, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  54. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully Convolutional One-Stage Object Detection. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  55. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the ICCV, Venice, Italy, 22–29 October 2017. [Google Scholar]
  56. Yang, Z.; Liu, S.; Hu, H.; Wang, L.; Lin, S. Reppoints: Point set representation for object detection. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  57. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  58. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. In Proceedings of the ICLR, Virtual Event, 3–7 May 2021. [Google Scholar]
  59. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the ICCV, Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
  60. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the CVPR, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  61. Girshick, R. Fast r-cnn. In Proceedings of the 2015 ICCV, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  62. Neubeck, A.; Gool, L.V. Efficient Non-Maximum Suppression. In Proceedings of the ICPR, Hong Kong, China, 20–24 August 2006. [Google Scholar]
Figure 1. Examples of NRPerson. They are maritime aerial scenes containing a large number of tiny persons. Each column represents a time-aligned RGB-IR image pair. The dashed box on each image pair represents a consistent field of view, showing that the image pairs are spatial-overlapped and non-registered.
Figure 1. Examples of NRPerson. They are maritime aerial scenes containing a large number of tiny persons. Each column represents a time-aligned RGB-IR image pair. The dashed box on each image pair represents a consistent field of view, showing that the image pairs are spatial-overlapped and non-registered.
Electronics 13 01697 g001
Figure 2. The histogram of the absolute size of person annotations in NRPerson. The statistical results reflect that the object size is concentrated in areas with small values, showing the tiny property of our dataset.
Figure 2. The histogram of the absolute size of person annotations in NRPerson. The statistical results reflect that the object size is concentrated in areas with small values, showing the tiny property of our dataset.
Electronics 13 01697 g002
Figure 3. A flowchart of multi-modal model on the non-registered multi-modal track. Multi-modal image pairs are simultaneously input into the model to obtain detection or localization results for each modality. Orange represents RGB features and Blue represents the IR features.
Figure 3. A flowchart of multi-modal model on the non-registered multi-modal track. Multi-modal image pairs are simultaneously input into the model to obtain detection or localization results for each modality. Orange represents RGB features and Blue represents the IR features.
Electronics 13 01697 g003
Figure 4. The poor feature-matching results suggest that, using our NRPerson dataset, it is difficult to achieve automatic registration.The red color represents the target in the visible image, the green color represents the target in the infrared image, and the yellow line indicates the matching between the two.
Figure 4. The poor feature-matching results suggest that, using our NRPerson dataset, it is difficult to achieve automatic registration.The red color represents the target in the visible image, the green color represents the target in the infrared image, and the yellow line indicates the matching between the two.
Electronics 13 01697 g004
Figure 5. Some qualitative results of multi-modal baselines on the non-registered multi-modal track. The first two rows are the detection results of Adap-RepPoints (orange); the middle two rows are the localization results of P2PNet with CPR (blue); and the last two rows are the ground-truth (green). For better visualization results, we zoom in on the original images and capture the sub-images in which each pair has a roughly consistent field of view.
Figure 5. Some qualitative results of multi-modal baselines on the non-registered multi-modal track. The first two rows are the detection results of Adap-RepPoints (orange); the middle two rows are the localization results of P2PNet with CPR (blue); and the last two rows are the ground-truth (green). For better visualization results, we zoom in on the original images and capture the sub-images in which each pair has a roughly consistent field of view.
Electronics 13 01697 g005
Table 1. Comparisons of the scale and properties of NRPerson with several related datasets. The description of the properties is presented in Section 3.2 in which non-registration refers to the data property and annotations based on non-registration image pairs rather than shared across modalities. TA denotes time-aligned, SO denotes spatial-overlapped, and NR denotes non-registered. PD is person detection, PL is person localization, OD is object detection, and OT is object tracking. – : the property is not applicable to this dataset. : the visible images of CVC-14 are in grayscale. : some image pairs are not strictly time-aligned.
Table 1. Comparisons of the scale and properties of NRPerson with several related datasets. The description of the properties is presented in Section 3.2 in which non-registration refers to the data property and annotations based on non-registration image pairs rather than shared across modalities. TA denotes time-aligned, SO denotes spatial-overlapped, and NR denotes non-registered. PD is person detection, PL is person localization, OD is object detection, and OT is object tracking. – : the property is not applicable to this dataset. : the visible images of CVC-14 are in grayscale. : some image pairs are not strictly time-aligned.
DatasetImagesAnnotationsRGBIRTASONRTaskYear
Multi-modal dataset
OSU-CT [41]17,088×Fusion2007
LITIV [6]12,650×Fusion2012
TNO [42]480×Fusion2014
RGB-T210 [22]210,000210,000×OT2017
RGB-T234 [23]233,800233,800×OT2019
RegDB [24]82408240Re-ID2017
SYSU-MM01 [25]45,86345,863 Re-ID2020
Person detection/localization dataset
Caltech [1]249,884346,621×PD2009
KITTI [5]14,99980,256×OD2012
COCOPersons [45]64,115273,469×PD2014
CityPersons [2]500035,016×PD2017
CrowdHuman [43]24,370456,098×PD2018
SCUT-FIR [44]211,011477,907×PD2019
TinyPerson [9]161072,651×PD2019
SeaPerson [10]12,032619,627 ×PL2022
Multi-modal person detection/localization dataset
KAIST [21]190,656103,128×PD2015
CVC-14 [46]17,00217,929✓ ✓ ×PD2016
FLIR [8]28,267119,491✓ ×OD2018
LLVIP [3]30,97641,579×PD2021
NRPerson 17,096889,207 PD & PL2022
Table 2. Detailed statistical information for NRPerson.
Table 2. Detailed statistical information for NRPerson.
NRPerson TrainValidTestTotal
image pairs461437535598548
annos346,41333,385509,408889,207
RGB annos174,31216,943280,669471,924
IR annos172,10116,442228,739417,282
Table 3. Mean and standard deviation statistics of absolute size, relative size and aspect ratio of person annotations in NRPerson and several representative datasets.
Table 3. Mean and standard deviation statistics of absolute size, relative size and aspect ratio of person annotations in NRPerson and several representative datasets.
DatasetAbsolute SizeRelative SizeAspect Ratio
KAIST [21]54.2 ± 24.80.10 ± 0.050.45 ± 0.10
FLIR [8]26.6 ± 23.10.05 ± 0.040.43 ± 0.19
LLVIP [3]146.1 ± 32.40.13 ± 0.030.47 ± 0.19
CityPersons [2]79.8 ± 67.50.56 ± 0.050.41 ± 0.01
TinyPerson [9]18.0 ± 17.40.01 ± 0.010.68 ± 0.42
SeaPerson [10]22.6 ± 10.80.02 ± 0.010.72 ± 0.42
NRPerson 23.7 ± 9.60.02 ± 0.010.81 ± 0.45
Table 4. The detection and localization results of representative models on the mono-modal track. For localization, the models adopt the annotations refined by CPR except for the explicit description of P2PNet. The first, second and third best results are marked in red, blue and black bold respectively.
Table 4. The detection and localization results of representative models on the mono-modal track. For localization, the models adopt the annotations refined by CPR except for the explicit description of P2PNet. The first, second and third best results are marked in red, blue and black bold respectively.
DetectionRGB-ModalIR-Modal
AP 50 AP 50 r AP 50 s AP 50 t AP 50 AP 50 r AP 50 s AP 50 t
Faster R-CNN [48]64.775.874.357.151.166.760.644.5
RetinaNet [55]64.474.974.757.549.366.062.339.3
Adaptive RetinaNet [55]65.574.174.361.251.466.562.544.6
RepPoints [56]66.874.576.760.351.467.964.240.6
Adaptive RepPoints [56]67.176.176.363.551.967.264.842.9
Sparse R-CNN [57]66.477.376.153.748.163.259.343.9
Deformable DETR [58]65.975.177.955.248.563.658.842.8
Swin-T (Faster R-CNN) [59]67.275.877.157.749.267.958.842.6
LocalizationRGB-ModalIR-Modal
AP 1 . 0 AP 1 . 0 r AP 1 . 0 s AP 1 . 0 t AP 1 . 0 AP 1 . 0 r AP 1 . 0 s AP 1 . 0 t
Faster R-CNN [48]71.521.156.277.574.833.866.972.3
RetinaNet [55]76.027.164.480.375.537.969.871.8
Adaptive RetinaNet [55]74.525.661.579.574.735.367.371.8
RepPoints [56]76.226.164.481.276.535.169.873.7
Adaptive RepPoints [56]75.125.963.280.474.533.066.871.8
Sparse R-CNN [57]76.427.865.380.274.234.068.170.1
Deformable DETR [58]78.029.768.283.277.740.973.175.8
Swin-T (Faster R-CNN) [59]72.723.258.476.976.341.671.172.6
P2PNet [12] w/o CPR [10]71.425.951.982.166.426.354.272.0
P2PNet [12] w/CPR [10]80.133.671.477.981.546.577.672.6
Table 5. The detection and localization results of the representative models on the non-registered multi-modal track. For localization, the models adopt the annotations refined by CPR except for the explicit description of P2PNet. The first, second and third best results are marked in red, blue and black bold, respectively.
Table 5. The detection and localization results of the representative models on the non-registered multi-modal track. For localization, the models adopt the annotations refined by CPR except for the explicit description of P2PNet. The first, second and third best results are marked in red, blue and black bold, respectively.
DetectionMulti-TestRGB-TestIR-Test
AP 50 AP 50 r AP 50 s AP 50 t AP 50 AP 50 r AP 50 s AP 50 t AP 50 AP 50 r AP 50 s AP 50 t
Faster R-CNN [48]60.171.870.452.066.577.377.155.150.966.257.635.9
RetinaNet [55]58.970.372.049.665.474.676.953.049.465.058.532.0
Adaptive RetinaNet [55]60.670.671.553.867.074.876.656.851.164.858.136.7
RepPoints [56]61.673.073.352.766.472.774.260.152.465.160.236.0
Adaptive RepPoints [56]62.672.372.050.469.679.479.458.852.767.461.135.7
Sparse R-CNN [57]59.870.571.051.467.178.477.453.848.861.357.037.7
Deformable DETR [58]60.171.371.451.767.978.378.554.449.363.957.836.9
Swin-T (Faster R-CNN) [59]59.169.670.149.768.677.277.956.047.164.258.938.5
LocalizationMulti-TestRGB-TestIR-Test
AP 1 . 0 AP 1 . 0 r AP 1 . 0 s AP 1 . 0 t AP 1 . 0 AP 1 . 0 r AP 1 . 0 s AP 1 . 0 t AP 1 . 0 AP 1 . 0 r AP 1 . 0 s AP 1 . 0 t
Faster R-CNN [48]73.930.663.376.073.526.360.879.274.433.766.571.8
RetinaNet [55]76.334.666.977.877.031.866.181.375.537.569.472.2
Adaptive RetinaNet [55]75.533.265.577.175.929.964.180.775.036.468.671.5
RepPoints [56]75.629.865.177.076.228.564.880.875.933.667.573.3
Adaptive RepPoints [56]74.530.363.877.074.727.062.280.574.533.567.071.5
Sparse R-CNN [57]78.434.470.178.880.132.970.983.276.635.970.172.7
Deformable DETR [58]77.134.569.580.577.530.068.383.277.739.673.076.1
Swin-T (Faster R-CNN) [59]72.633.363.075.072.024.058.277.273.342.468.371.3
P2PNet [12] w/o CPR [10]66.224.048.975.068.925.849.480.463.422.949.267.2
P2PNet [12] w/CPR [10]80.339.474.476.080.835.373.879.678.643.874.369.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Han, X.; Wang, K.; Yu, X.; Yu, W.; Wang, Z.; Li, G.; Han, Z.; Jiao, J. NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization. Electronics 2024, 13, 1697. https://doi.org/10.3390/electronics13091697

AMA Style

Yang Y, Han X, Wang K, Yu X, Yu W, Wang Z, Li G, Han Z, Jiao J. NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization. Electronics. 2024; 13(9):1697. https://doi.org/10.3390/electronics13091697

Chicago/Turabian Style

Yang, Yi, Xumeng Han, Kuiran Wang, Xuehui Yu, Wenwen Yu, Zipeng Wang, Guorong Li, Zhenjun Han, and Jianbin Jiao. 2024. "NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization" Electronics 13, no. 9: 1697. https://doi.org/10.3390/electronics13091697

APA Style

Yang, Y., Han, X., Wang, K., Yu, X., Yu, W., Wang, Z., Li, G., Han, Z., & Jiao, J. (2024). NRPerson: A Non-Registered Multi-Modal Benchmark for Tiny Person Detection and Localization. Electronics, 13(9), 1697. https://doi.org/10.3390/electronics13091697

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop