Recent Research Progress on Ground-to-Air Vision-Based Anti-UAV Detection and Tracking Methodologies: A Review
Abstract
:1. Introduction
- A critical analysis of the most recent vision-based ground-to-air UAV detection and tracking techniques, highlighting their strengths, limitations, and trade-offs.
- A detailed overview of publicly available UAV datasets, presenting their characteristics and how they address (or fail to address) specific research challenges.
- Identification of existing gaps in the literature and areas where further research is needed.
2. UAV Detection
2.1. Two-Stage vs. One-Stage
2.2. Anchor-Based vs. Anchor-Free
2.3. Detection Network Architecture
2.3.1. Anchor Modification
2.3.2. Backbone Modification
2.3.3. Neck Modification
2.3.4. Head Modification
2.4. Critical Analysis
3. UAV Tracking
- SOT focuses on tracking the trajectory of a single UAV over time, even as it moves through varying backgrounds or experiences occlusion.
- MOT is concerned with tracking multiple UAVs simultaneously, which adds complexity due to interactions between UAVs and occlusion.
3.1. Single-Object Tracking
3.2. Multi-Object Tracking
3.3. Tracking Methodologies
4. Datasets
4.1. Detailed Analysis of Popular Datasets
4.1.1. TIB-Net
4.1.2. MAV-VID
4.1.3. Anti-UAV
4.1.4. Drone-vs.-Bird
4.1.5. UAVSwarm
4.1.6. DUT Anti-UAV
5. Challenges and Future Works
- Detection and Tracking of Small and Fast-moving UAVs: Detecting and tracking small, fast-moving UAVs against cluttered and dynamic backgrounds is a major challenge for vision-based systems. These UAVs appear as only a few pixels in high-resolution images, making it difficult for existing object detection models to identify them accurately. As a result, UAVs are frequently mistaken for birds, clouds, or background noise, leading to false positives and missed detections.
- Real-time Processing Constraints: Real-time detection and tracking are crucial for Anti-UAV systems in scenarios where UAVs pose immediate threats, such as near airports or critical infrastructure. However, achieving high accuracy in real-time conditions is challenging due to the computational complexity of current detection algorithms. Models have made progress in optimizing inference speed but often sacrifice accuracy, especially when detecting small, fast-moving UAVs. Additionally, tracking methods prioritizing high precision may introduce unacceptable latency in time-sensitive applications.
- Multi-UAV Detection and Tracking: The current research primarily focuses on single UAV detection and tracking, which simplifies the problem in controlled or isolated environments. However, real-world commercial scenarios would involve multiple UAVs operating simultaneously in a 3D air volume, introducing complexity due to occlusion, interactions, and merging with background elements. The limitations of existing methods become apparent when faced with the dynamics of UAV swarms or the congested airspace characteristic of urban areas. Existing methods may not be suitable to handle overlapping trajectories and the rapid movements of UAVs of varying sizes and also intent. Little to no work has been performed to identify and track multiple rogues from normal UAV traffic. Consequently, Multi-UAV tracking remains under-explored, offering opportunities for future research. Addressing these challenges will be essential for creating robust Anti-UAV systems with the capability of detecting and tracking multiple rogues or swarms of rogues concurrently.
- Environmental Variability and Robustness: Vision-based systems are impacted by environmental factors such as lighting, weather, and background complexity. Existing models often struggle in low-light conditions, fog, or rain. While efforts have been made to address some limitations, ensuring reliability in diverse conditions remains a challenge for Anti-UAV systems. There is a lack of representative benchmark datasets that truly reflect a commercialized UAV traffic scenario in variable environmental conditions. A potential research direction could be to either collect such a dataset or to artificially generate a dataset using data augmentation techniques, generative AI, etc.
- Integration into the UTM: While Anti-UAV systems are being developed to detect rogue or unauthorized UAVs, there is a significant gap in integrating these systems into broader UTM frameworks. UTM systems are designed to manage authorized UAV traffic in uncontrolled airspace; Anti-UAV systems must operate alongside these frameworks to ensure that only non-compliant UAVs are targeted. The Federal Aviation Administration’s (FAA) proposed UTM system and the European Union’s U-space system are both designed with a service-oriented architecture in mind [95,96]. This approach allows for flexibility and innovation, as various components of these frameworks can be developed, managed, and maintained by independent service providers. This service-oriented model not only promotes specialization but also ensures that advancements can be integrated seamlessly, enhancing the overall efficiency and safety of airspace management for Unmanned Aerial Vehicles across both regions. The current literature lacks a holistic approach to seamlessly integrating detection and tracking methodologies within the UTM framework, including real-time coordination between ground stations and flight management systems. This requires more comprehensive solutions as well as better datasets to distinguish between regular and non-conforming UAVs in complex airspace environments.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
DURC Statement
Conflicts of Interest
Abbreviations
AGL | Above Ground Level |
ASFF | Adaptive Spatial Feature Fusion |
ATC | Air Traffic Control |
ATOM | Accurate Tracking by Overlap Maximization |
ATM | Air Traffic Management |
ATSS | Adaptive Training Sample Selection |
AUC | Area Under the Curve |
C2f | Convolution to Feature |
CA | Coordinate Attention |
CBAM | Convolutional Block Attention Module |
CSM | Class-Level Semantic Modulation |
CSPDarkNet | Cross-Stage-Partial-connections DarkNet |
DeepSORT | Simple Online and Real-time Tracking with a Deep Association Metric |
DETR | Detection Transformer |
DiMP | Discriminative Model Prediction |
DSFC | Dual-Flow Semantic Consistency |
ECO | Efficient Convolution Operators |
EMA | Efficient Multi-scale Attention |
EXTD | Extremely Tiny Face Detector |
FAA | Federal Aviation Administration |
FCOS | Fully Convolutional One-Stage Object Detector |
FPN | Feature Pyramid Network |
FPS | Frames Per Second |
GNMOT | Graph Networks for Multi-Object Tracking |
ISM | Instance-Level Semantic Modulation |
JDE | Joint Detection and Embedding |
LSTM | Long Short-Term Memory |
LTMU | Long-term Tracking with Meta-Updater |
mAP | Mean Average Precision |
MEGA | Memory Enhanced Global-Local Aggregation |
MobileViT | Mobile Vision Transformer |
MOT | Multi-Object Tracking |
MOTA | Multiple Object Tracking Accuracy |
PANet | Path Aggregation Network |
QDTrack | Quasi-dense Tracking |
R-CNN | Region-based Convolutional Neural Network |
RF | Radio Frequency |
SAM | Spatial Attention Module |
SE | Squeeze and Excitation |
SimAM | Simple parameter-free Attention Module |
SORT | Simple Online Real-time Tracking |
SOT | Single-Object Tracking |
SPP | Spatial Pyramid Pooling |
SPLT | Skimming-Perusal Tracking |
SSD | Single Shot Detector |
TransT | Transformer Tracking |
UAS | Unmanned Aircraft System |
UAV | Unmanned Aerial Vehicle |
UTM | Unmanned Aircraft System Traffic Management |
VGG | Visual Geometry Group |
YOLO | You Only Look Once |
Appendix A
Reference | Year | Techniques Used | Dataset(s) Used | Accuracy Metrics Reported | Lightweight |
---|---|---|---|---|---|
[39] | 2019 | 1. The last four scales of feature maps are adopted instead of the last three in YOLOv3 to predict bounding boxes of objects, which can obtain more texture and contour information 2. K-means clustering is used on the training set to decide the number of the scales 3. The number and size of the anchor boxes are also adjusted using k-means clustering | Not public. Self-collected. | [email protected] 37.41% 56.3 FPS | No |
[40] | 2020 | 1. Cyclic pathway is added to EXTD, which enhances the capability to extract the effective features of small objects but does not increase the model size much 2. Spatial Attention Module is added to the network backbone to emphasize information of small objects | TIBNet | mAP 89.25% | Yes |
[41] | 2020 | 1. Standard YOLOv4 is implemented 2. Dataset is augmented by rotating and flipping collected images | Not public. Self-collected. | [email protected] 89.32% 39.64 FPS | No |
[42] | 2021 | Standard Faster R-CNN is implemented with multiple backbones (VGG-16, ResNet50, DarkNet-53, DenseNet-201) | UAVData | VGG-16 mAP 90.6%, 11 FPS ResNet50 mAP 90.4%, 10 FPS DarkNet-53 mAP 86.3%, 10 FPS DenseNet-201 mAP Failed | No |
[42] | 2021 | Standard YOLOv3 is implemented with multiple backbones (VGG-16, ResNet50, DarkNet-53, DenseNet-201) | UAVData | VGG-16 mAP 90.8%, 70 FPS ResNet50 mAP 90.6%, 86 FPS DarkNet-53 mAP 90.8%, 72 FPS DenseNet-201 mAP 90.7%, 49 FPS | No |
[42] | 2021 | Standard SSD is implemented with multiple backbones (VGG-16, ResNet50, DarkNet-53, DenseNet-201) | UAVData | VGG-16 mAP 74.2%, 42 FPS ResNet50 mAP 75.3%, 22 FPS DarkNet-53 mAP 74.8%, 24 FPS DenseNet-201 mAP 73.5, 14 FPS | Yes (VGG-16) No for rest |
[13] | 2021 | Standard Faster R-CNN is implemented with ResNet-50 backbone and an FPN at the end of each convolutional block | MAV-VID Drone-vs.-Bird Anti-UAV RGB | MAV-VID mAP 97.8%, 18 FPS Drone-vs.-Bird mAP 63.2%, 18 FPS Anti-UAV RGB mAP 98.2%, 18 FPS | No |
[13] | 2021 | Standard YOLOv3 is implemented with DarkNet-53 backbone | MAV-VID Drone-vs.-Bird Anti-UAV RGB | MAV-VID mAP 96.3%, 36 FPS Drone-vs.-Bird mAP 54.6%, 36 FPS Anti-UAV RGB mAP 98.6%, 36 FPS | No |
[13] | 2021 | Standard SSD512 is implemented | MAV-VID Drone-vs.-Bird Anti-UAV RGB | MAV-VID mAP 96.7%, 32.4 FPS Drone-vs.-Bird mAP 62.9%, 32.4 FPS Anti-UAV RGB mAP 97.9%, 32.4 FPS | Yes |
[13] | 2021 | Standard DETR with ResNet-50 backbone | MAV-VID Drone-vs.-Bird Anti-UAV RGB | MAV-VID mAP 97.1%, 21.4 FPS Drone-vs.-Bird mAP 66.7%, 21.4 FPS Anti-UAV RGB mAP 97.8%, 21.4 FPS | No |
[43] | 2021 | 1. Standard YOLOv4 is implemented 2. Mosaic data augmentation is applied | Not public. Self-collected. | mAP 74.36% 19.75 FPS | No |
[44] | 2021 | 1. Convolutional channel and shortcut layer of YOLOv4 are pruned to make the model thinner and shallower 2. Dataset is augmented by copy and pasting small drones | Not public. Self-collected. | mAP 92.7% 69 FPS | Yes |
[45] | 2022 | Standard Faster R-CNN is implemented with different backbones | DUT Anti-UAV | ResNet-50 mAP 65.3%, 12.8 FPS ResNet-18 mAP 60.5%, 19.4 FPS VGG-16 mAP 63.3%, 9.3 FPS | No |
[45] | 2022 | Standard Cascade R-CNN is implemented with different backbones | DUT Anti-UAV | ResNet-50 mAP 68.3%, 10.7 FPS ResNet-18 mAP 65.2%, 14.7 FPS VGG-16 mAP 66.7%, 8 FPS | No |
[45] | 2022 | Standard ATSS method is implemented with different backbones | DUT Anti-UAV | ResNet-50 mAP 64.2%, 13.3 FPS ResNet-18 mAP 61%, 20.5 FPS VGG-16 mAP 64.1%, 9.5 FPS | N/A |
[45] | 2022 | Standard YOLOX is implemented with different backbones | DUT Anti-UAV | ResNet-50 mAP 42.7%, 21.7 FPS ResNet-18 mAP 40%, 53.7 FPS VGG-16 mAP 55.1%, 23 FPS | No |
[45] | 2022 | Standard SSD is implemented with different backbones | DUT Anti-UAV | VGG-16 mAP 63.2%, 33.2 FPS | Yes |
[46] | 2022 | 1. Background difference is used to extract potential drone targets in high-resolution images to reduce computational overhead 2. Ghost module and SimAM attention mechanism are introduced to reduce the total number of model parameters and improve feature extraction 3. -DIoU loss is used instead of DIoU loss to improve the accuracy of bounding box regression | Drone-vs.-Bird | mAP 97.6% 13.2 FPS | Yes |
[47] | 2022 | 1. The last four scales of feature maps are adopted instead of the last three in YOLOv3 to predict bounding boxes of objects, which can obtain more texture and contour information 2. Data augmentation is performed through changing brightness, and contrast of the images and rotating and flipping the images | Not public. Self-collected. | mAP 25.12% 21 FPS | No |
[48] | 2022 | 1. YOLOv5 backbone is replaced with Efficientlite, to reduce the number of parameters 2. Adaptive spatial feature fusion is injected into the head to improve the accuracy loss caused by the lightweight of the model backbone 3. A constraint of angle is introduced into the original regression loss function to improve the speed of convergence 4. Data augmentation by adding random noise points and binarization | Kaggle Dataset | mAP 94.82% | Yes |
[49] | 2023 | 1. Spatial Attention module added to the backbone 2. SPPS and ResNeck modules added to the neck 3. Data are augmented using Mosaic Augmentation | TIBNet | mAP 89.7% | Yes |
[50] | 2023 | 1. MobileViT is used as the backbone to reduce network complexity 2. Added Coordinate Attention to PANet of YOLOv4 to obtain better positional information and improve information fusion of high and low-dimensional features 3. K-means++ is used to adjust anchor boxes 4. Data are augmented using Mosaic Augmentation | Not public. Self-collected. | mAP 92.8% 40 FPS | Yes |
[51] | 2023 | 1. Depthwise separable convolution is used to simplify and optimize the network 2. Squeeze-and-Excitation (SE) module is introduced into the backbone to improve the model’s ability to extract features 3. Convolutional Block Attention Module (CBAM) is added in the feature fusion network to make the network pay more attention to important features and suppress unnecessary features 4. Distance-IoU (DIoU) is used to replace Intersection over Union (IoU) to calculate the regression loss for model optimization 5. Data are augmented using MixUp and Mosaic Augmentation | UAVSwarm | mAP 82.32% 14 FPS | Yes |
[52] | 2024 | 1. Ghost convolution is included in the neck to reduce model size 2. Efficient multi-scale attention (EMA) is added to preserve pixel-level attributes and spatial information on the feature map 3. Deformable Convolutional Net v2 (DCNv2) is used in the detection head to improve model robustness | DUT Anti-UAV | mAP 97.1 56.2 FPS | Yes |
[53] | 2024 | 1. Anchor boxes are adjusted using k-means clustering 2. InceptionNeXT module is added to neck to capture more global semantic information 3. SPPFCSPC-SR module is added to the backbone to reduce feature loss, suppress confusion, and make the model pay more attention to small target areas 4. FPN is replaced with Get-and-Send module to improve model’s capability to fuse information across different levels | DUT Anti-UAV and Amateur UAV combined | mAP 93.2% 104 FPS | Yes |
References
- Houthi Drone Attacks on 2 Saudi Aramco Oil Facilities Spark Fires. Available online: https://tinyurl.com/UAVattack (accessed on 4 December 2024).
- Drones Market Size & Share Analysis—Growth Trends & Forecasts (2024–2029). Available online: https://tinyurl.com/dronesfore (accessed on 4 December 2024).
- Forecast Highlights (2023–2043). Available online: https://tinyurl.com/FAAForecasts (accessed on 4 December 2024).
- Davies, L.; Vagapov, Y.; Grout, V.; Cunningham, S.; Anuchin, A. Review of air traffic management systems for UAV integration into urban airspace. In Proceedings of the 2021 28th International Workshop on Electric Drives: Improving Reliability of Electric Drives (IWED), Moscow, Russia, 27–29 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
- Chamola, V.; Kotesh, P.; Agarwal, A.; Gupta, N.; Guizani, M. A comprehensive review of unmanned aerial vehicle attacks and neutralization techniques. Ad Hoc Netw. 2021, 111, 102324. [Google Scholar] [CrossRef] [PubMed]
- Klare, J.; Biallawons, O.; Cerutti-Maori, D. UAV detection with MIMO radar. In Proceedings of the 2017 18th International Radar Symposium (IRS), Prague, Czech Republic, 28–30 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–8. [Google Scholar]
- Mohajerin, N.; Histon, J.; Dizaji, R.; Waslander, S.L. Feature extraction and radar track classification for detecting UAVs in civillian airspace. In Proceedings of the 2014 IEEE Radar Conference, Cincinnati, OH, USA, 19–23 May 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 674–679. [Google Scholar]
- Xiao, Y.; Zhang, X. Micro-UAV detection and identification based on radio frequency signature. In Proceedings of the 2019 6th International Conference on Systems and Informatics (ICSAI), Shanghai, China, 2–4 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1056–1062. [Google Scholar]
- Al-Emadi, S.; Al-Senaid, F. Drone detection approach based on radio-frequency using convolutional neural network. In Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies (ICIoT), Doha, Qatar, 2–5 February 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 29–34. [Google Scholar]
- Yang, B.; Matson, E.T.; Smith, A.H.; Dietz, J.E.; Gallagher, J.C. UAV detection system with multiple acoustic nodes using machine learning models. In Proceedings of the 2019 Third IEEE International Conference on Robotic Computing (IRC), Naples, Italy, 25–27 February 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 493–498. [Google Scholar]
- Hauzenberger, L.; Holmberg Ohlsson, E. Drone Detection Using Audio Analysis. Master’s Thesis, Lund University, Lund, Sweden, 2015. [Google Scholar]
- Behera, D.K.; Raj, A.B. Drone detection and classification using deep learning. In Proceedings of the 2020 4th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 13–15 May 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1012–1016. [Google Scholar]
- Isaac-Medina, B.K.; Poyser, M.; Organisciak, D.; Willcocks, C.G.; Breckon, T.P.; Shum, H.P. Unmanned aerial vehicle visual detection and tracking using deep neural networks: A performance benchmark. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, QC, Canada, 10–17 October 2021; pp. 1223–1232. [Google Scholar]
- Mendis, G.J.; Randeny, T.; Wei, J.; Madanayake, A. Deep learning based doppler radar for micro UAS detection and classification. In Proceedings of the MILCOM 2016—2016 IEEE Military Communications Conference, Baltimore, MD, USA, 1–3 November 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 924–929. [Google Scholar]
- Bisio, I.; Garibotto, C.; Lavagetto, F.; Sciarrone, A.; Zappatore, S. Unauthorized amateur UAV detection based on WiFi statistical fingerprint analysis. IEEE Commun. Mag. 2018, 56, 106–111. [Google Scholar] [CrossRef]
- Chu, Z.; Song, T.; Jin, R.; Jiang, T. An Experimental Evaluation Based on New Air-to-Air Multi-UAV Tracking Dataset. In Proceedings of the 2023 IEEE International Conference on Unmanned Systems (ICUS), Hefei, China, 13–15 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 671–676. [Google Scholar]
- Muna, S.I.; Mukherjee, S.; Namuduri, K.; Compere, M.; Akbas, M.I.; Molnár, P.; Subramanian, R. Air Corridors: Concept, Design, Simulation, and Rules of Engagement. Sensors 2021, 21, 7536. [Google Scholar] [CrossRef] [PubMed]
- Liu, Z.; An, P.; Yang, Y.; Qiu, S.; Liu, Q.; Xu, X. Vision-Based Drone Detection in Complex Environments: A Survey. Drones 2024, 8, 643. [Google Scholar] [CrossRef]
- Seidaliyeva, U.; Ilipbayeva, L.; Taissariyeva, K.; Smailov, N.; Matson, E.T. Advances and Challenges in Drone Detection and Classification Techniques: A State-of-the-Art Review. Sensors 2024, 24, 125. [Google Scholar] [CrossRef]
- Wang, B.; Li, Q.; Mao, Q.; Wang, J.; Chen, C.L.P.; Shangguan, A.; Zhang, H. A Survey on Vision-Based Anti Unmanned Aerial Vehicles Methods. Drones 2024, 8, 518. [Google Scholar] [CrossRef]
- Al-lQubaydhi, N.; Alenezi, A.; Alanazi, T.; Senyor, A.; Alanezi, N.; Alotaibi, B.; Alotaibi, M.; Razaque, A.; Hariri, S. Deep learning for unmanned aerial vehicles detection: A review. Comput. Sci. Rev. 2024, 51, 100614. [Google Scholar] [CrossRef]
- Coluccia, A.; Fascista, A.; Schumann, A.; Sommer, L.; Dimou, A.; Zarpalas, D.; Akyon, F.C.; Eryuksel, O.; Ozfuttu, K.A.; Altinuc, S.O.; et al. Drone-vs.-bird detection challenge at IEEE AVSS2021. In Proceedings of the 2021 17th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Washington, DC, USA, 16–19 November 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
- Abdullah, Q.A. Classification of the Unmanned Aerial Systems. Pennsylvania State. 2014. Available online: https://www.e-education.psu.edu/geog892/node/5 (accessed on 13 January 2025).
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
- Kaiming, H.; Georgia, G.; Piotr, D.; Ross, G.S. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision 2017, Venice, Italy, 22–29 October 2017; Volume 2017, pp. 2961–2969. [Google Scholar]
- Lu, X.; Li, B.; Yue, Y.; Li, Q.; Yan, J. Grid r-cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7363–7372. [Google Scholar]
- Farhadi, A.; Redmon, J. Yolov3: An incremental improvement. In Computer Vision and Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2018; Volume 1804, pp. 1–6. [Google Scholar]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
- Ross, T.Y.; Dollár, G.K.H.P. Focal loss for dense object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
- Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-shot refinement neural network for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4203–4212. [Google Scholar]
- Li, S.; Sultonov, F.; Tursunboev, J.; Park, J.H.; Yun, S.; Kang, J.M. Ghostformer: A GhostNet-based two-stage transformer for small object detection. Sensors 2022, 22, 6939. [Google Scholar] [CrossRef]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
- Liu, S.; Zhou, H.; Li, C.; Wang, S. Analysis of anchor-based and anchor-free object detection methods based on deep learning. In Proceedings of the 2020 IEEE International Conference on Mechatronics and Automation (ICMA), Beijing, China, 2–5 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1058–1065. [Google Scholar]
- Zhou, X.; Wang, D.; Krähenbühl, P. Objects as points. arXiv 2019, arXiv:1904.07850. [Google Scholar]
- Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
- Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
- Hu, Y.; Wu, X.; Zheng, G.; Liu, X. Object detection of UAV for anti-UAV based on improved YOLO v3. In Proceedings of the 2019 Chinese Control Conference (CCC), Guangzhou, China, 27–30 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 8386–8390. [Google Scholar]
- Sun, H.; Yang, J.; Shen, J.; Liang, D.; Ning-Zhong, L.; Zhou, H. TIB-Net: Drone detection network with tiny iterative backbone. IEEE Access 2020, 8, 130697–130707. [Google Scholar] [CrossRef]
- Shi, Q.; Li, J. Objects detection of UAV for anti-UAV based on YOLOv4. In Proceedings of the 2020 IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT), Weihai, China, 14–16 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1048–1052. [Google Scholar]
- Zeng, Y.; Duan, Q.; Chen, X.; Peng, D.; Mao, Y.; Yang, K. UAVData: A dataset for unmanned aerial vehicle detection. Soft Comput. 2021, 25, 5385–5393. [Google Scholar] [CrossRef]
- Singha, S.; Aydin, B. Automated drone detection using YOLOv4. Drones 2021, 5, 95. [Google Scholar] [CrossRef]
- Liu, H.; Fan, K.; Ouyang, Q.; Li, N. Real-time small drones detection based on pruned yolov4. Sensors 2021, 21, 3374. [Google Scholar] [CrossRef]
- Zhao, J.; Zhang, J.; Li, D.; Wang, D. Vision-based anti-uav detection and tracking. IEEE Trans. Intell. Transp. Syst. 2022, 23, 25323–25334. [Google Scholar] [CrossRef]
- Lv, Y.; Ai, Z.; Chen, M.; Gong, X.; Wang, Y.; Lu, Z. High-resolution drone detection based on background difference and SAG-Yolov5s. Sensors 2022, 22, 5825. [Google Scholar] [CrossRef]
- Zhai, H.; Zhang, Y. Target Detection of Low-Altitude UAV Based on Improved YOLOv3 Network. J. Robot. 2022, 2022, 4065734. [Google Scholar] [CrossRef]
- Liu, B.; Luo, H. An improved Yolov5 for multi-rotor UAV detection. Electronics 2022, 11, 2330. [Google Scholar] [CrossRef]
- Zhou, X.; Yang, G.; Chen, Y.; Gao, C.; Zhao, B.; Li, L.; Chen, B.M. ADMNet: Anti-Drone Real-Time Detection and Monitoring. In Proceedings of the 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Detroit, MI, USA, 1–5 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 3009–3016. [Google Scholar]
- Cheng, Q.; Li, X.; Zhu, B.; Shi, Y.; Xie, B. Drone detection method based on MobileViT and CA-PANet. Electronics 2023, 12, 223. [Google Scholar] [CrossRef]
- Wang, C.; Meng, L.; Gao, Q.; Wang, J.; Wang, T.; Liu, X.; Du, F.; Wang, L.; Wang, E. A lightweight UAV swarm detection method integrated attention mechanism. Drones 2022, 7, 13. [Google Scholar] [CrossRef]
- Huang, M.; Mi, W.; Wang, Y. EDGS-YOLOv8: An Improved YOLOv8 Lightweight UAV Detection Model. Drones 2024, 8, 337. [Google Scholar] [CrossRef]
- Bo, C.; Wei, Y.; Wang, X.; Shi, Z.; Xiao, Y. Vision-based anti-UAV detection based on YOLOv7-GS in complex backgrounds. Drones 2024, 8, 331. [Google Scholar] [CrossRef]
- Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
- Redmon, J. Darknet: Open Source Neural Networks in C. 2018. Available online: http://pjreddie.com/darknet/ (accessed on 4 December 2024).
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Howard, A.G. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Yoo, Y.; Han, D.; Yun, S. Extd: Extremely tiny face detector via iterative filter reuse. arXiv 2019, arXiv:1906.06579. [Google Scholar]
- Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
- Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
- Zhang, H. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
- Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
- Christiansen, A. Anchor Boxes—The Key to Quality Object Detection; Towards Data Science: San Francisco, CA, USA, 2018. [Google Scholar]
- Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
- Guo, J.M.; Yang, J.S.; Seshathiri, S.; Wu, H.W. A light-weight CNN for object detection with sparse model and knowledge distillation. Electronics 2022, 11, 575. [Google Scholar] [CrossRef]
- Wang, C.; Su, Y.; Wang, J.; Wang, T.; Gao, Q. UAVSwarm dataset: An unmanned aerial vehicle swarm dataset for multiple object tracking. Remote Sens. 2022, 14, 2601. [Google Scholar] [CrossRef]
- Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
- Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient multi-scale attention module with cross-spatial learning. In Proceedings of the ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
- Yu, W.; Zhou, P.; Yan, S.; Wang, X. Inceptionnext: When inception meets convnext. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 5672–5683. [Google Scholar]
- Liu, S.; Huang, D.; Wang, Y. Learning spatial fusion for single-shot object detection. arXiv 2019, arXiv:1911.09516. [Google Scholar]
- Wang, R.; Shivanna, R.; Cheng, D.; Jain, S.; Lin, D.; Hong, L.; Chi, E. Dcn v2: Improved deep & cross network and practical lessons for web-scale learning to rank systems. In Proceedings of the Web Conference, Ljubljana, Slovenia, 19–23 April 2021; pp. 1785–1797. [Google Scholar]
- Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
- Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef]
- Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the Computer Vision—ECCV 2016 Workshops, Amsterdam, The Netherlands, 8–10, 15–16 October 2016; Proceedings, Part II 14. Springer International Publishing: Cham, Switzerland, 2016; pp. 850–865. [Google Scholar]
- Cheng, F.; Liang, Z.; Peng, G.; Liu, S.; Li, S.; Ji, M. An anti-UAV long-term tracking method with hybrid attention mechanism and hierarchical discriminator. Sensors 2022, 22, 3701. [Google Scholar] [CrossRef]
- Wang, C.; Shi, Z.; Meng, L.; Wang, J.; Wang, T.; Gao, Q.; Wang, E. Anti-occlusion UAV tracking algorithm with a low-altitude complex background by integrating attention mechanism. Drones 2022, 6, 149. [Google Scholar] [CrossRef]
- Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; Yan, J. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4282–4291. [Google Scholar]
- Danelljan, M.; Bhat, G.; Khan, F.S.; Felsberg, M. Atom: Accurate tracking by overlap maximization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4660–4669. [Google Scholar]
- Yan, B.; Zhao, H.; Wang, D.; Lu, H.; Yang, X. ‘skimming-perusal’ tracking: A framework for real-time and robust long-term tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2385–2393. [Google Scholar]
- Dai, K.; Zhang, Y.; Wang, D.; Li, J.; Lu, H.; Yang, X. High-performance long-term tracking with meta-updater. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6298–6307. [Google Scholar]
- Bhat, G.; Danelljan, M.; Gool, L.V.; Timofte, R. Learning discriminative model prediction for tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6182–6191. [Google Scholar]
- Chen, X.; Yan, B.; Zhu, J.; Wang, D.; Yang, X.; Lu, H. Transformer tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8126–8135. [Google Scholar]
- Huang, L.; Zhao, X.; Huang, K. Globaltrack: A simple and strong baseline for long-term tracking. Proc. AAAI Conf. Artif. Intell. 2020, 34, 11037–11044. [Google Scholar] [CrossRef]
- Jiang, N.; Wang, K.; Peng, X.; Yu, X.; Wang, Q.; Xing, J.; Li, G.; Zhao, J.; Guo, G.; Han, Z. Anti-UAV: A large multi-modal benchmark for UAV tracking. arXiv 2021, arXiv:2101.08466. [Google Scholar]
- Wu, Y.; Lim, J.; Yang, M.H. Online object tracking: A benchmark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2411–2418. [Google Scholar]
- Bernardin, K.; Elbs, A.; Stiefelhagen, R. Multiple object tracking performance metrics and evaluation in a smart room environment. In Proceedings of the Sixth IEEE International Workshop on Visual Surveillance, Graz, Austria, 13 May 2006; Conjunction with ECCV. Citeseer: University Park, PA, USA, 2006; Volume 90. [Google Scholar]
- Rodriguez-Ramos, A.; Rodriguez-Vazquez, J.; Sampedro, C.; Campoy, P. Adaptive inattentional framework for video object detection with reward-conditional training. IEEE Access 2020, 8, 124451–124466. [Google Scholar] [CrossRef]
- Coluccia, A.; Fascista, A.; Schumann, A.; Sommer, L.; Dimou, A.; Zarpalas, D.; Méndez, M.; De la Iglesia, D.; González, I.; Mercier, J.P.; et al. Drone vs. bird detection: Deep learning algorithms and results from a grand challenge. Sensors 2021, 21, 2824. [Google Scholar] [CrossRef]
- Faa.gov. UTM Concept of Operations Version 2.0 (UTM ConOps v2.0) | Federal Aviation Administration. 2022. Available online: https://www.faa.gov/researchdevelopment/trafficmanagement/utm-concept-operations-version-20-utm-conops-v20 (accessed on 13 January 2025).
- Barrado, C.; Boyero, M.; Brucculeri, L.; Ferrara, G.; Hately, A.; Hullah, P.; Martin-Marrero, D.; Pastor, E.; Rushton, A.P.; Volkert, A. U-space concept of operations: A key enabler for opening airspace to emerging low-altitude operations. Aerospace 2020, 7, 24. [Google Scholar] [CrossRef]
- Wei, L.; Ma, Z.; Yang, C.; Yao, Q. Advances in the Neural Network Quantization: A Comprehensive Review. Appl. Sci. 2024, 14, 7445. [Google Scholar] [CrossRef]
Ref. | Year | Base Model | Stage Type | Anchor Type |
---|---|---|---|---|
[39] | 2019 | YOLOv3 | One-Stage | Anchor-based |
[40] | 2020 | EXTD | One-Stage | Anchor-free |
[41] | 2020 | YOLOv4 | One-Stage | Anchor-based |
[42] | 2021 | Faster R-CNN | Two-Stage | Anchor-based |
[42] | 2021 | YOLOv3 | One-Stage | Anchor-based |
[42] | 2021 | SSD | One-Stage | Anchor-based |
[13] | 2021 | Faster R-CNN | Two-Stage | Anchor-based |
[13] | 2021 | YOLOv3 | One-Stage | Anchor-based |
[13] | 2021 | SSD512 | One-Stage | Anchor-based |
[13] | 2021 | DETR | One-Stage | Anchor-free |
[43] | 2021 | YOLOv4 | One-Stage | Anchor-based |
[44] | 2021 | YOLOv4 | One-Stage | Anchor-based |
[45] | 2022 | Faster R-CNN | Two-Stage | Anchor-based |
[45] | 2022 | Cascade R-CNN | Two-Stage | Anchor-based |
[45] | 2022 | ATSS | N/A | N/A |
[45] | 2022 | YOLOX | One-Stage | Anchor-free |
[45] | 2022 | SSD | One-Stage | Anchor-based |
[46] | 2022 | YOLOv5s | One-Stage | Anchor-based |
[47] | 2022 | YOLOv3 | One-Stage | Anchor-based |
[48] | 2022 | YOLOv5 | One-Stage | Anchor-based |
[49] | 2023 | YOLOv4 | One-Stage | Anchor-based |
[50] | 2023 | YOLOv4 | One-Stage | Anchor-based |
[51] | 2023 | YOLOX-nano | One-Stage | Anchor-free |
[52] | 2024 | YOLOv8 | One-Stage | Anchor-free |
[53] | 2024 | YOLOv7-tiny | One-Stage | Anchor-based |
Ref. | Techniques Used | Dataset | Best Accuracy Metrics Reported | Comments |
---|---|---|---|---|
[81] | 1. Uses SiamRPN++ 2. Includes a re-detection module based on YOLOv5 that uses a hybrid attention mechanism and hierarchical discriminator | Youtube- BoundingBoxes, ImageNet VID, ImageNet, COCO, Anti-UAV | 67.7% AUC 88.4% Precision | This method is good for long-term tracking |
[82] | 1. Uses ATOM 2. Includes an SE attention mechanism 3. Includes an occlusion sensing module | OTB-100 | 67.7% AUC 87.9% Precision | This method is good for occluded environments. |
[82] | GOT-10k | 73.1% AUC 59.9% Precision | ||
[82] | Drones-vs.-bird + LaSOT | 50.5% AUC 79% Precision | ||
[70] | Uses ByteTrack | UAVSwarm | UAVSwarm-06 88.3% MOTA UAVSwarm-28 32.5% MOTA UAVSwarm-30 87.2% MOTA UAVSwarm-46 −12% MOTA | The performance of ByteTrack is inconsistent while GNMOT seems to give promising results |
[70] | Uses GNMOT | UAVSwarm | UAVSwarm-06 100% MOTA UAVSwarm-28 98.4% MOTA UAVSwarm-30 100% MOTA UAVSwarm-46 99.75 MOTA | |
[45] | 1. Uses SiamFC with Cascade-RCNN as the built-in detector 2. Cascade-RCNN is used with a ResNet50 backbone | DUT Anti-UAV | 61.7% AUC 93.3% Precision | LTMU+Faster-RCNN +VGG16 tracking by detection model gives the best results for DUT Anti-UAV |
[45] | 1. Uses SiamFC with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a VGG16 backbone | DUT Anti-UAV | 61.5% AUC 94.3% Precision | |
[45] | 1. Uses ECO with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a VGG16 backbone | DUT Anti-UAV | 62% AUC 95.4% Precision | |
[45] | 1. Uses SPLT with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a VGG16 backbone | DUT Anti-UAV | 55.3% AUC 87.5% Precision | |
[45] | 1. Uses ATOM with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a ResNet18 backbone | DUT Anti-UAV | 63.5% AUC 93.6% Precision | LTMU+Faster-RCNN +VGG16 tracking by detection model gives the best results for DUT Anti-UAV |
[45] | 1. Uses SiamRPN++ with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a VGG16 backbone | DUT Anti-UAV | 61.2% AUC 88.1% Precision | |
[45] | 1. Uses DiMP with Faster-RCNN as the built-in detector 2. Cascade-RCNN is used with a ResNet50 backbone | DUT Anti-UAV | 65.7% AUC 94.9% Precision | |
[45] | 1. Uses TransT with Cascade-RCNN as the built-in detector 2. Cascade-RCNN is used with a ResNet50 backbone | DUT Anti-UAV | 62.4% AUC 88.8% Precision | |
[45] | 1. Uses LTMU with Faster-RCNN as the built-in detector 2. Fascade-RCNN is used with a VGG16 backbone | DUT Anti-UAV | 66.4% AUC 96.1% Precision | |
[90] | Uses SiamRCNN with DSFC training strategy that CSM and ISM modules | Anti-UAV RGB | 67.04% AUC 90.71% Precision | SiamRCNN and GlobalTrack are the best methods. |
[90] | Uses GlobalTrack with DSFC training strategy that CSM and ISM modules | Anti-UAV RGB | 62.36% AUC 87.65% Precision | |
[90] | Uses LTDSE with DSFC training strategy that CSM and ISM modules | Anti-UAV RGB | 58.58% AUC 82.56% Precision | |
[90] | Uses SiamRPN++LT with DSFC training strategy that CSM and ISM modules | Anti-UAV RGB | 57.28% AUC 77.93% Precision | |
[90] | Uses Super-DiMP with DSFC training strategy that CSM and ISM modules | Anti-UAV RGB | 53.77% AUC 75.29% Precision | |
[13] | Uses Tracktor with Faster R-CNN detection module | MAV-VID | 95.5% MOTA | Tracktor performs the best consistently but depending on the difficulty of the dataset, different detection modules give different results. |
[13] | Uses Tracktor with SSD512 detection module | Drone-vs.-Bird | 52.5% MOTA | |
[13] | Uses Tracktor with DETR detection module | Anti-UAV RGB | 94.0% MOTA |
Dataset | Objective | Link |
---|---|---|
TIB-Net [40] | Detection | https://github.com/kyn0v/TIB-Net/tree/master (accessed on 4 December 2024) |
MAV-VID [93] | Detection, MOT | https://bitbucket.org/alejodosr/mav-vid-dataset/src/master/ (accessed on 4 December 2024) |
Anti-UAV [90] | Detection, MOT | https://github.com/ucas-vg/Anti-UAV (accessed on 4 December 2024) |
Drone-vs.-Bird [94] | Detection | https://github.com/wosdetc/challenge (accessed on 4 December 2024) |
UAVSwarm [70] | Detection, MOT | https://github.com/UAVSwarm/UAVSwarm-dataset (accessed on 4 December 2024) |
DUT Anti-UAV [45] | Detection, SOT | https://github.com/wangdongdut/DUT-Anti-UAV (accessed on 4 December 2024) |
Dataset | Size | UAV Type | Resolution | Environment | Light Conditions |
---|---|---|---|---|---|
TIB-Net [40] | 2860 images, 694 MB | multi-rotor, fixed-wing | 1920 × 1080 | Homogeneous, simple backgrounds | Day, nightfall and night |
MAV-VID [93] | 40,232 images, 64 videos, 11.3 GB | multi-rotor | 1920 × 1080 | Heterogeneous, Varying background conditions | Day |
Anti-UAV (RGB) [90] | 93,247 images, 100 videos, 5.25 GB | multi-rotor | 1920 × 1080 | Heterogeneous, Varying weather conditions | Day, nightfall and night |
Drone-vs.-Bird [94] | 104,760 images, 77 videos, 7.1 GB | multi-rotor, fixed-wing | varies from 720 × 576 to 3840 × 2160 | Heterogeneous, Varying weather and background conditions | Day |
UAVSwarm [70] | 12,598 images, 72 videos, 1.87 GB | multi-rotor, fixed-wing | varies from 446 × 270 to 1919 × 1079 | Heterogeneous, Varying light and background conditions | Day, nightfall and night |
DUT Anti-UAV [45] | 10,000 images, 20 videos, 8.76 GB | multi-rotor | varies extremely | Heterogeneous, Varying light and background conditions | Day, nightfall and night |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yasmeen, A.; Daescu, O. Recent Research Progress on Ground-to-Air Vision-Based Anti-UAV Detection and Tracking Methodologies: A Review. Drones 2025, 9, 58. https://doi.org/10.3390/drones9010058
Yasmeen A, Daescu O. Recent Research Progress on Ground-to-Air Vision-Based Anti-UAV Detection and Tracking Methodologies: A Review. Drones. 2025; 9(1):58. https://doi.org/10.3390/drones9010058
Chicago/Turabian StyleYasmeen, Arowa, and Ovidiu Daescu. 2025. "Recent Research Progress on Ground-to-Air Vision-Based Anti-UAV Detection and Tracking Methodologies: A Review" Drones 9, no. 1: 58. https://doi.org/10.3390/drones9010058
APA StyleYasmeen, A., & Daescu, O. (2025). Recent Research Progress on Ground-to-Air Vision-Based Anti-UAV Detection and Tracking Methodologies: A Review. Drones, 9(1), 58. https://doi.org/10.3390/drones9010058