Next Article in Journal
A Systematic Evaluation of Recurrent Neural Network Models for Edge Intelligence and Human Activity Recognition Applications
Next Article in Special Issue
Automatic Optimization of Deep Learning Training through Feature-Aware-Based Dataset Splitting
Previous Article in Journal
Root Cause Tracing Using Equipment Process Accuracy Evaluation for Looper in Hot Rolling
Previous Article in Special Issue
Comparison of Different Radial Basis Function Networks for the Electrical Impedance Tomography (EIT) Inverse Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches

by
Noor Ul Ain Tahir
1,†,
Zuping Zhang
1,*,†,
Muhammad Asim
2,3,*,†,
Junhong Chen
3,4,† and
Mohammed ELAffendi
2
1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
EIAS Data Science and Blockchain Laboratory, College of Computer and Information Sciences, Prince Sultan University, Riyadh 11586, Saudi Arabia
3
School of Computer Science and Technology, Guangdong University of Technology, Guangzhou 510006, China
4
Expertise Centre for Digital Media, Flanders Make, Hasselt University, 3500 Hasselt, Belgium
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Algorithms 2024, 17(3), 103; https://doi.org/10.3390/a17030103
Submission received: 14 January 2024 / Revised: 16 February 2024 / Accepted: 19 February 2024 / Published: 26 February 2024

Abstract

:
Enhancing the environmental perception of autonomous vehicles (AVs) in intelligent transportation systems requires computer vision technology to be effective in detecting objects and obstacles, particularly in adverse weather conditions. Adverse weather circumstances present serious difficulties for object-detecting systems, which are essential to contemporary safety procedures, infrastructure for monitoring, and intelligent transportation. AVs primarily depend on image processing algorithms that utilize a wide range of onboard visual sensors for guidance and decisionmaking. Ensuring the consistent identification of critical elements such as vehicles, pedestrians, and road lanes, even in adverse weather, is a paramount objective. This paper not only provides a comprehensive review of the literature on object detection (OD) under adverse weather conditions but also delves into the ever-evolving realm of the architecture of AVs, challenges for automated vehicles in adverse weather, the basic structure of OD, and explores the landscape of traditional and deep learning (DL) approaches for OD within the realm of AVs. These approaches are essential for advancing the capabilities of AVs in recognizing and responding to objects in their surroundings. This paper further investigates previous research that has employed both traditional and DL methodologies for the detection of vehicles, pedestrians, and road lanes, effectively linking these approaches with the evolving field of AVs. Moreover, this paper offers an in-depth analysis of the datasets commonly employed in AV research, with a specific focus on the detection of key elements in various environmental conditions, and then summarizes the evaluation matrix. We expect that this review paper will help scholars to gain a better understanding of this area of research.

1. Introduction

The World Health Organization (WHO) releases statistics each year on the number of people who have been hurt in traffic accidents. Around 1.3 million people worldwide die each year, and 20 to 50 million suffer serious injuries [1], with young males under 25 constituting the majority, 73%, of fatal traffic accidents. Extensive research on the worrying prevalence of traffic fatalities has led to the development of upgraded intelligent transportation systems (ITS) with expanded capabilities. The cutting-edge of contemporary automotive technology is represented by automated driving systems (ADS) and autonomous vehicles (AVs). AVs, especially self-driving cars, have generated a great deal of interest in the fields of intelligent transportation, computer vision, and artificial intelligence [2]. No other invention since the creation of vehicles has had such a dramatic impact on the automotive industry as self-driving vehicles. As per the forecasts presented in Figure 1 [3], the market for self-driving cars is expected to grow from 23.80 million units in 2022 to 80.43 million units by 2032. These vehicles primarily rely on cameras and sensors, and examples include LiDAR (light detection and ranging), radar, ultrasonic, etc., which have a number of autonomous features, including reduced use of fuel, cutting carbon dioxide emissions by up to 10%, decreased traffic congestion [4], decreased road accidents, and utilizing roads more efficiently to optimize the transportation system. AVs can assist in managing and studying traffic by gathering effective and valuable traffic data. Along with other road-related capabilities, they are capable of recognizing static and non-static objects such as vehicles, pedestrians, and road lanes. Object detection (OD) is critical for AVs as it ensures safety by identifying and avoiding moving obstacles in their path. Numerous research works have investigated methods for identifying objects in AVs, broadly divided into manual, semi-automated, and fully automated systems [5], as shown in Figure 2. Roadside artifacts are visually assessed by human inspectors using manual procedures; nonetheless, subjectivity and labor intensity are drawbacks. While semi-automated methods gather data from moving cars and process them afterward, they are still labor-intensive [6]. High-resolution cameras and sensors on cars record pictures and videos in completely autonomous systems. Software-based models that have been trained to detect objects evaluate the data either after they have been collected or in real time. These techniques seek to solve safety problems while improving the precision and effectiveness of item identification on roadways and streets. Vehicle-based surveillance is a widely used technique since it is a quick and efficient way to examine objects. However, it has always been difficult for autonomous programs to identify objects under adverse weather conditions such as rain, fog, snow, haze, storms, and low lighting effects.
The weather has a number of negative consequences for mobility and traffic. Precipitation is recorded roughly 11.0% of the time worldwide on average [7]. Studies have unequivocally shown that, in comparison to typical weather, rainfall can cause a 70% increase in the likelihood of accidents [8]. Moreover, snowfall occurs in 77% of the countries worldwide. For example, according to US national data, annually, 24% of vehicle accidents related to weather conditions take place on slippery, snowy, slushy, or icy roads, while 15% occur when snow is actively falling or mixed with sleet [9], highlighting the real risks associated with snowy weather. Visibility is greatly reduced by environmental variables such as fog, haze, sandstorms, and strong sunlight, which presents real difficulties for drivers [10].
Traditional and deep learning (DL) algorithms are crucial for enhancing OD in autonomous vehicles, especially in adverse weather. While traditional AI, based on statistical learning, has been foundational, it faces limitations due to manual feature engineering and limited adaptability to dynamic environments, requiring extensive retraining and hindering its effectiveness in real-time applications.
DL offers a powerful alternative to traditional machine learning methods for OD in AVs, addressing the limitations of manual feature engineering and adaptability in dynamic environments. DL’s multi-layer neural networks automatically extract features and learn from data, providing a more flexible and adaptable solution that enhances the performance of surveillance systems, self-driving vehicles, and smart city applications [11]. DL’s neural network-based approach is particularly adept at handling complex models that traditional techniques cannot [12]. In AV object detection, hybrid models combining one-stage and two-stage detectors are effective, with two-stage models focusing on accuracy and one-stage models on speed, although the reasons for the latter’s lower accuracy are not fully understood.
The research significance of this paper lies in its comprehensive analysis of the challenges and advancements in OD for AVs in adverse weather conditions. It provides a critical review of both traditional and DL approaches, offering insights into the limitations and potential improvements of the current detection algorithms. The paper contributes to the broader field of intelligent transportation systems by emphasizing the need for robust and reliable detection systems that can operate effectively in a variety of weather scenarios, which is crucial for the safe deployment of AVs in real-world conditions. Building on this significance, this paper also contributes to the discussion regarding AV architecture and challenges in adverse weather, and reviews the literature on detecting pedestrians, vehicles, and road lanes using traditional and DL methods. It also summarizes common evaluation metrics for OD. In this paper, we contribute to the field by examining the fundamental architecture of AVs and the specific challenges they face in adverse weather conditions. We have compiled comprehensive datasets, leveraging real-world statistics from LiDAR and camera sensors, to provide a robust foundation for our analysis. We detail the core structure of OD systems and elucidate both traditional and DL methodologies for AVs. Building upon these approaches, we provide a critical review of the existing literature, focusing on the detection of three primary objects—pedestrians, vehicles, and road lanes—under challenging weather conditions. The structure of this paper is shown in Figure 3.

2. Related Survey Papers

This section critically evaluates the existing literature on autonomous vehicle (AV) sensor performance under challenging weather conditions, with a focus on the integration of traditional and DL detection techniques. One study [13] presents an in-depth analysis of the impact of inclement weather on AV sensors, particularly the attenuation and backscatter effects on millimeter-wave radar in heavy rain. Despite the study’s novel modeling approach, it does not comprehensively address the multifaceted challenges posed by other weather phenomena, such as fog and snowfall, which is crucial for a holistic sensor performance evaluation. Research [14] offers a comprehensive review of sensor performance in adverse weather conditions, emphasizing the importance of sensor fusion for improved vehicle perception. However, the study’s focus on range and resolution does not fully explore the nuanced challenges of object classification and sensor reliability in cold weather conditions, which is critical for AV operation. The study in [15] aims to develop a robust operational system for AVs in various weather conditions, advocating for sensor fusion and infrastructure support. Yet, the study’s discussion on sensor reliability does not adequately address the practical considerations of sensor installation costs and benefits, which are significant factors in the practical implementation of AV technology. The study in [16] investigates the effects of adverse weather on LiDAR and camera lenses in AVs, suggesting improvements in sensor hardware and the application of sophisticated machine learning methods. While the study’s focus on sensor fusion and network infrastructure is commendable, it does not provide a comprehensive evaluation of these technologies across different weather conditions. The research in [17] emphasizes the role of weather recognition and categorization systems in AV decisionmaking, focusing on the detection of road lanes, vehicles, and pedestrians. While the study provides valuable insights, it acknowledges that detecting smaller, obscured, or partially visible objects remains a challenge, indicating a need for further research in this area. The study in [18] serves as a tutorial on advanced techniques addressing the impact of rainy conditions on AV object detection. The study evaluates the performance of detection methods under clear and rainy scenarios, highlighting the efficacy of deraining methods, DL-based domain adaptation, and image translation frameworks. However, it does not comprehensively address the limitations of these methods in other adverse weather conditions. The study in [19] examines vehicle detection approaches in video-based traffic surveillance, highlighting the challenges posed by dim illumination, variable weather, occlusion, and shadows. The study’s exploration of traditional approaches is valuable, but it does not fully integrate the advancements in DL methods for improved detection. Other [20] research focuses on DL methods for detecting on-road vehicles, providing a valuable list of works. However, the study’s limited scope does not encompass the full range of challenges faced by AVs in adverse weather conditions. Another paper [21] exclusively examines 3D methods for vehicle detection, offering insights into a specific detection approach. Yet, the study’s narrow focus does not address the broader context of AVs and their performance under various weather conditions.
In summary, while the surveyed literature provides a valuable foundation for understanding AV sensor technologies and detection techniques, there is a critical gap in the holistic examination of both traditional and DL approaches under adverse weather conditions. This survey paper focuses on traditional approaches, deep learning approaches, and their applications in object detection in AVs under adverse weather conditions. Furthermore, we have detailed our literature review methodology, which involves a systematic search across reputable digital repositories such as Springer Link, Elsevier, IEEE, MDPI, and Wiley Online Library. We utilized targeted keywords to identify relevant research articles and focused on studies published in peer-reviewed journals and international conferences from 2001 to 2023. This approach ensured a comprehensive and high-quality selection of literature for our review.

3. An overview of AVs

The Society of Automotive Engineers (SAE) has established a classification system that divides Autonomous Driving Systems (ADS) into six levels of automation, ranging from Level 0 (No Automation) to Level 5 (Full Automation), as illustrated in Figure 4 [22]. This system is pivotal in understanding the evolution of autonomous vehicles (AVs), which are equipped with advanced sensors and software to navigate independently, thereby enhancing safety technologies and autonomy. Such a progression necessitates a collaborative effort among scientists and engineers across various disciplines to address the complex challenges associated with AV development [23]. The SAE automation levels provide a standardized framework that is critical for evaluating the capabilities and limitations of AVs at different stages of their development. By categorizing AVs into these distinct levels, it becomes possible to systematically assess the technological advancements and hurdles encountered at each level. This structured approach is instrumental in setting realistic expectations about AV performance, shaping regulatory frameworks, and educating the public on the operational capabilities of AVs under various scenarios. Thus, the classification of AVs according to SAE levels is essential for advancing the field of autonomous driving, guiding its regulatory landscape, and informing societal understanding of these emerging technologies.
Initially, in the 1920s, AVs were referred to as “phantom autos” due to their lack of a human driver, relying on remote control [24]. Significant advancements in AV technology emerged in the 1980s, notably with Pomerleau’s “Autonomous Land Vehicle in a Neural Network” project [25]. This project demonstrated the potential of neural networks to significantly enhance autonomous navigation systems. The Defense Advanced Research Projects Agency (DARPA) initiated the inaugural Grand Challenge in 2004 with the primary aim of stimulating research and the advancement of AVs. Subsequent to the 2004 challenge, DARPA hosted additional grand challenge events in 2005 and 2007, famously recognized as the Urban Challenge [25]. Developments in recent times have resulted in the global implementation of AVs. Some manufacturers have achieved Level 4 SAE standard [26] autonomy through the use of LiDAR technology. Google started secretly building its first AV in 2009, and, on May 1, 2012, in Las Vegas, they successfully completed their first autonomous driving test [27]. The UK government launched a competition in 2014 with the goal of encouraging and developing AVs [28]. The Mcity driverless shuttle project was started recently by the University of Michigan. It is the first Level 4 automated shuttle project in the US [29]. However, due to safety considerations, AVs cannot operate in severe rain or snow. Although testing and research have been carried out in inclement weather, problems such as the persistent usage of windshield wipers may result in an operational stoppage. Low temperatures impacted charging during the night and limited daily operating hours for the Sohjoa Baltic Project [30]. Tesla’s autopilot works well in mild rain and snow and has obvious lane markers, but it struggles in inclement weather [31]. Regulation (EU) 2019/2144 was introduced by the European Parliament and Council in 2019. This was the first time that guidelines pertaining to automated and fully autonomous vehicles had been established [32]. When Waymo (formerly Google’s self-driving division, rebranded as Waymo in 2016) announced that their “Waymo Driver” had successfully completed 20 million self-driven miles and carried out simulations that equated to 15 billion miles, it marked an important turning point in the growth of AVs [33]. While researchers are striving to develop sensors for different weather scenarios, thorough studies that address every facet are still absent. The AVs use environment-sensing skills to function with the least amount of human interaction [34]. Using an autonomous driving framework, the weather has a significant impact on the data obtained from car sensors and the general state of the surroundings. Adverse weather conditions create obstacles to object identification, monitoring, and localization, necessitating modifications to control and planning techniques. Furthermore, the efficiency of the vehicle can be directly impacted by meteorological factors like wind, general weather, and road conditions.
This interaction results in an ongoing cycle whereby changes in the environment and vehicle states impact one another, highlighting the weather’s crucial role in autonomous driving. Figure 5 shows the architecture of AVs.

3.1. Sensor Technologies

In the pursuit of improving OD in inclement weather, selecting the right sensor technologies is essential to guaranteeing precise and dependable results. Unpredictable natural events, like adverse weather, might have an impact on the operating environment of AVs. These changes in the surrounding environment bring about differences in how AV sensor systems, which are the fundamental parts of ADS, operate. The main sensory elements used in AVs for perception are described in the section that follows, and the comparison of sensors is shown in Table 1.

3.1.1. LiDAR

LiDAR is considered the primary perceptive sensor in the self-driving car industry. Even though 3D-LiDAR has only been used in automobiles for slightly over ten years, it has already been shown to be crucial to the development of AVs and ADS [36]. Its exceptional ability to measure things precisely and its resilience to changing lighting conditions highlight how vital it is. Critical features of 3D scanning using lasers include its form aspect, affordability, tolerance to external variables, measuring range, precision, pinpoint density, scanning acceleration, configurational versatility, and spectral qualities [37]. These sensors work well in a range of weather conditions by using laser beams to measure distances and produce accurate 3D point clouds. However, there could be problems when there is a great deal of rain, snow, or fog since laser beams can disperse and potentially skew the data. Older types, such as the Velodyne HDL64-S2 [38], continue to function better in dense fog, while the performance of more modern laser scanners decreases. While making manual modifications can increase performance, it is still difficult to perceive reliably in deep fog. For this reason, alternative technologies, including gated imaging, should be considered for better performance in adverse conditions [39].

3.1.2. Radar

The fundamental components of ADS and self-driving automobiles are automotive radars and a variety of additional sensors. The successful use of these technical advances depends on complex systems with long signal-processing chains that link the radars and sensors to the controller. The vital tasks of identifying objects and obstructions, pinpointing their exact location, and evaluating their relative speed in relation to the vehicle fall to automotive radar equipment. Advances in millimeter-wave (mm-wave) semiconductor technology, along with the development of signal processing techniques, form the basis for the progress of car radar systems. A variety of signal processing methods have been developed to improve resolution and reliability in a range of measuring parameters, including target velocities in the vicinity, azimuth elevation perspectives, and the radius of the vehicle [40]. Frequency Modulated Continuous Wave radar, or FMCW radar, modifies the broadcast signal’s frequency continually at a predetermined pace. By measuring the gap between the sent and reflecting signals, which is inversely related to the amount of time required for the electrical signal to make its way to and from the object being targeted, we can determine the range due to this modulation of frequency. FMCW radar has excellent accuracy and range resolution, as well as speed measurement features, which make it especially useful for autonomous driving applications [41].

3.1.3. Ultrasonic

One typical vehicle sensor that is frequently left out of discussions regarding ADS is the ultrasonic sensor. Despite this error, it is positioned strategically on the vehicle’s bumpers and throughout its body, making it indispensable for parking assistance and blind spot surveillance. Surprisingly, ultrasonic sensors have proven to be dependable and reasonably priced over a long period of time, which makes them a useful sensor choice [42]. The frequency range of ultrasound sounds, which are audible only to humans, is normally between 30 and 480 kHz. In the field of ultrasonic sensing, the frequency range that is most frequently utilized is between 40 and 70 kHz. The resolution and sensor range are highly dependent on the selected frequency. Longer sensing ranges are correlated with the lower ones. For example, the measuring range is up to 11 m, and the accuracy is one centimeter (cm) at the commonly used frequency of 58 kHz. Conversely, higher frequencies such as 300 kHz provide amazing resolution, maybe down to one millimeter, but at the penalty of a shorter range, capped at about 30 cm [43]. Ultrasonic sensors are useful for close-quarters applications like parking because their normal operating range is 11 m [44]. However, they can be used in autonomous driving; for example, Tesla’s “summon” capability can be used to navigate garage doors and parking lots [45].

3.1.4. Camera

As the eyes and storytellers of automobiles, cameras are essential components of ADS. They are a crucial component of ADS, capturing the dynamic story of the surroundings, even though they are technologically more advanced than LiDAR. Installed on Windows, dashcams record continuously and have made a substantial contribution to the initial ADS datasets. Fisheye-lens professional camera setups nowadays increase the volume of data that may be collected. On the other hand, adverse weather can cause visual problems for cameras due to rain, snow, and fog. In low light, specialized cameras improve visibility, such as night vision versions. The potential to overcome weather-related constraints in sensor fusion methods through the integration of these cameras is encouraging. The function of cameras will change further as autonomous driving technology develops, influencing sensor technologies and how they interact with the outside world. In summary, cameras are the storytellers of autonomous cars; they are sensitive to bad weather yet versatile enough to be essential to safer autonomous systems.

3.2. Actuators

AVs use data from the planning stage to carry out the vehicle’s motions in the control stage. This entails giving actuators instructions for steering, braking, accelerating, and signaling. A drive-by-wire method is frequently used to transmit commands effectively. The control system is in charge of creating and monitoring trajectories, making sure that intended routes are taken. Methods for generating trajectory include sensor-based, which is appropriate for robotics, and dynamics-based, which is more appropriate for vehicles. The majority of the time, trajectory tracking uses geometrical or model-based techniques. Although they have limits and only react to errors as they happen [46], feedback controllers like the proportional-integral-derivative (PID) are employed to prevent deviations from intended trajectories. Two-degree independent controllers combine feedforward and feedback controllers to circumvent these restrictions.

3.3. Perception

To highlight the importance of perception for AVs, it is important to emphasize how it helps the AV determine its location and its surroundings. In order to extract road features, detect objects, anticipate their behavior, and carry out Simultaneous Localization and Mapping (SLAM) activities, the perception module uses data from sensors and communication equipment [47]. For detecting the surroundings, a variety of active and passive sensors, such as cameras, LIDAR, and RADAR, can be employed; each has advantages and disadvantages of its own, as shown in Table 1. Emphasized are the benefits and drawbacks of various sensors, illustrating the compromises made in terms of resolution, weather sensitivity, and accuracy. One way to overcome the limits of individual sensors is to introduce sensor fusion, a technique that combines data from several sensors. There are two main types of AV systems discussed: one is sensor fusion, which uses various sensors such as LIDAR, RADAR, and cameras, and the other is pure vision, which uses cameras and computer vision algorithms. It is mentioned that a pure vision-based system has advantages like affordability and simplicity of implementation. In the parts that follow, this research concentrates on the benefits of a pure vision-based method [48].

3.4. Planning and Control

Global route planning and local path planning are the two main planning challenges in autonomous driving. When given an origin and destination, the global planner’s job is to come up with possible road options and routes. Local path planning, on the other hand, concentrates on carrying out the chosen path without running into obstructions or breaking any traffic laws. Other modules, such as OD, which guarantees safe navigation, frequently supplement this role. Further details regarding motion planning can be found in [49] for a more thorough explanation. Like traditional cars, steering, throttle, and brake inputs are the main controls for AVs [50]. These inputs control the AV’s direction, speed, and acceleration, which in turn control its driving, under the supervision of the ADS mainframe’s judgments based on sensor data and analysis. Low-level safety precautions can also be put in place using sensors like sonar and electrical control units. ADS sensors are challenged by weather conditions, which interfere with perception functions and have an impact on detection, planning, and control, among other factors. Weather factors like wind and variations in the road surface have an immediate effect on the vehicle and its state, in addition to sensor-related problems. Environmental state changes result from these state changes, which also have an impact on other nearby cars and sensor operations.

3.5. Challenges for AVs in Adverse Weather

Automation parts up to Level 3 have been added by automakers, and they rely on sensors and cameras. These devices function best in optimal conditions; thus, unfavorable weather presents obstacles. Weather has a major impact on AV performance, regardless of whether it is road-related or generic. A few of these effects are shown in Figure 6, and these are explained in more detail below.

3.5.1. Performance of Sensors

Adverse weather conditions on roads provide serious difficulties for sensors in AVs. The current sensing technologies, including LiDAR and cameras, perform well in clear weather but have trouble when the roadway is covered with snow and may contain erroneous reflections caused by rain or snow, making it difficult to discern between different objects [13]. Even though forward radar is good at spotting large objects and piercing through weather, it cannot improve safety applications on its own. In snowy circumstances, ultrasonic detectors may set off false alarms, and glare and surface reflection are just two examples of weather-related variables that might affect sensor accuracy.

3.5.2. Performance of Cameras

While camera systems are crucial for applications involving driverless vehicles, unfavorable weather conditions such as fog, intense rain, snow, and dust reduce their effectiveness. Although sophisticated techniques for image processing try to solve these problems, accuracy and dependability issues still exist, which presents significant obstacles to the advancement of autonomous driving [51]. In the right circumstances, glare can potentially affect functionality. However, these systems may be able to identify hazardous weather conditions immediately.

3.5.3. Technical Parameters of the Vehicle

Adverse weather affects a vehicle’s functioning factors. Slow speeds and longer following distances are necessary for safety due to decreased road friction and slick conditions. As was previously mentioned, during bad weather, driver behavior can also have a parallel impact on vehicle operating parameters.

3.5.4. Behavior of Driver

Driver conduct is still crucial once AVs are deployed, particularly in situations with mixed traffic. In order to ensure operational safety, AV applications need to take into account the existence of other drivers and environmental elements. These factors include driver-state surveillance and the ability to react to different weather conditions.

3.5.5. V2V, V2I, and VLC Communications

The influence of meteorological conditions on network latency can sometimes prevent vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I) communications from connecting in a direct line of sight. For safety applications, minimal latency is essential. While radio-based communications are less impacted by snow, they are nevertheless susceptible to problems like solar flares and thunderstorms. Communication interruptions can result from bad weather such as fog, heavy rain, snow, and hail obstructing the visible light communication (VLC) line of sight.
The process of OD in AVs unfolds in several key steps, beginning with the collection of image data by the sensor system. Following this, a Region of Interest (ROI) or regional proposal strategy is employed, marking a critical phase in object tracking. This approach is commonly utilized in devices such as cameras and stereoscopic systems, where ROIs serve as the foundational elements for tracking objects. The subsequent step involves the use of classifiers to meticulously analyze and refine shapes, edges, and lines. This analysis aids in the classification of targets and their identification based on human categories, embedding these ROIs within the image for object recognition.
Various techniques, including sliding windows, Locally De-correlated Channel Features (LDCF), and constrained searches, are applied to locate ROIs efficiently. The extraction of features from these ROIs represents the third stage of the process. At this juncture, either DL-based methods or traditional manual OD algorithms can be employed for the extraction and classification of features. While manual feature extraction techniques can be advantageous for models that rely on lower-level features by facilitating manual ROI suggestions [52], their efficacy diminishes when dealing with complex features, rendering them less reliable. However, to derive useful and relevant data from video or photographs using a computing device, a great deal of computation has to be completed [53]. In contrast, DL methodologies enable systems to autonomously learn feature representations, thereby enhancing the accuracy and adaptability of feature extraction in dynamic and complex driving environments.
The evolution of research on OD in bad weather depends heavily on datasets. They act as the building blocks for creating, refining, and testing algorithms and models. In this part, we are going to discuss a few of the datasets that are frequently utilized in this field of study. One of the most often used benchmarks for autonomous driving research is the KITTI [54] dataset. It features a range of various metropolitan settings. The collection offers LiDAR data, excellent-quality pictures, and annotations for cars. A2D2 [55] is a sizable dataset and delivers extensive annotations for several items, including vehicles, together with high-resolution camera and LiDAR data. Waymo’s dataset [56], which is renowned for its comprehensive and varied gathering of data from AVs and includes information from a range of weather and lighting circumstances, is a useful tool for research on inclement weather. Among the noteworthy datasets, NuScenes [57], BDD100K [58], ApolloScape [59], and SYNTHIA [60] datasets offer diverse and extensive resources for exploring vehicle detection in challenging conditions. A quick overview of the weather conditions considered and the particular sensors utilized for gathering data for each dataset can be found in Table 2. The University of Michigan used a Segway robot to collect four seasons of LIDAR data for the NCLT [61] on campus. The first AV dataset, called the Canadian Adverse Driving Conditions (CADC) [62] dataset, focuses on snow conditions. The Challenging Conditions Dataset with Correspondences (ACDC) [54], created for training and testing lexical techniques for segmentation in difficult visual conditions, was later introduced by the same team. ACDC includes a high-quality pixel-level resolution collection of well-labeled photos encompassing various categories of mist, at night, rain, and snowfall. The public’s Digital Video Collection now includes the IUPUI [63] Driving Video/Image Benchmark. Example views of varied lighting and roadway conditions captured by in-car webcams are included in this baseline. In these movies, patrolling and disaster response situations involving safety automobiles are shown. Together, these datasets provide priceless resources for developing the discipline and promoting creativity in the creation of reliable and weather-adaptive vehicle detection systems.

4. Overview of Object Detection in AVs

Most OD algorithms primarily adhere to a common framework, as shown in Figure 7.
In OD, we have considered the following three issues:
  • Vehicle Detection: Vehicle detection is the process by which AVs identify and locate other vehicles on the road. This capability is crucial for AVs to make informed decisions about their own movement, such as maintaining safe distances, changing lanes, or responding to traffic situations.
  • Pedestrian Detection: Pedestrian detection involves the recognition and tracking of people walking near or crossing the road. This is a vital safety feature for AVs as it enables the vehicle to anticipate and react to the presence of pedestrians, preventing collisions and ensuring the safety of both the vehicle’s occupants and those outside.
  • Road Lane Detection: Road lane detection is the ability of AVs to identify and understand the position and orientation of road lanes. This information is essential for the vehicle to navigate within its designated lane, follow traffic rules, and make correct turns, ensuring a smooth and safe driving experience.

5. Overview of Traditional and DL Approaches for Object Detection in AVs

The detection techniques are composed of three parts: the DL approach, the traditional technique, and a hybrid approach that utilizes both. The DL and traditional techniques are explained in detail in the section that follows. The performance graph of both approaches is shown in Figure 8.

5.1. Traditional Approach

Various algorithms were developed for the purpose of OD. Prominent traditional feature extraction methods include Scale Invariant Feature Transform (SIFT) [83], Viola–Jones rectangles [84], Haar-Like-wavelets and Histogram of Oriented Gradient (HOG) [85], Adaptive Boosting (AdaBoost), non-linear Support Vector Machines (SVMs), and linear SVMs, which are common classical approaches for object classification. A number of steps are involved in the SIFT algorithm [83], which is well known for its ability to determine scale and rotation-invariant features, making it resistant to partial occlusion, clutter, noise, and lighting variations. These include the detection of scale-space extrema, the localization of key points, the assignment of orientations, the creation of keypoint descriptors, and the final step of using key point matching to recognize objects. Descriptors are used to compare and identify items in a database using methods like nearest neighbor indexing and the Hough transform. Nevertheless, SIFT has some drawbacks, including the absence of ongoing key point consistency in dynamic objects, the significant dimensions of feature descriptors that can affect matchmaking, and some restrictions related to patents.
The Viola–Jones approach (Viola and Jones, [84] was developed for human face identification using Haar-like feature extractors. It consists of four primary parts; the primary image algorithm for effective image participation is AdaBoost. Haar-like feature extraction is used to uncover biased features. Cascade classifier implementation is used to filter out background data and concentrate on regions that are more likely to contain the object of interest. The algorithm performs remarkably well, taking only 0.067 s to process a 384 × 288 pixel image. Its detection rate is 91.4%, and it produces 50 false positives. Nevertheless, it is not well adapted for the generic recognition of objects and has slow periods for training. The HOG method was created for identifying humans in digital photos and was first presented in [85]. In order to determine the gradient’s length and direction, this method examines input images employing gradient filters, both horizontal and vertical. The filtered images are divided into 8 × 8 pixel cells, which are then further divided into 2 × 2 cell blocks that have 50% overlap. For every cell, orientation histograms are produced, quantizing gradient directions into nine bins ranging from 0 ° to 180 ° . The magnitudes of these bins are used as votes. A vector representing the HOG is produced by concatenating these histograms for every cell in a block. To take into consideration changes in illumination and contrast, these vectors are standardized. The final HOG descriptor, which consists of all the normalized blocks, is fed into a Support Vector Machine (SVM) to determine if an image is human or not. Although the HOG algorithm’s dense grid method results in a larger processing burden and less efficient performance in cases involving obscured objects, it is effective in decreasing false positives when compared with Haar wavelets.
These AI techniques, which have their roots in statistical principles, have traditionally formed the basis of machine learning models. These techniques enable computers to recognize trends and forecast outcomes based on past data. However, even with their historical importance, traditional methods have clear drawbacks. Reliance on manual feature engineering, as shown in Figure 9, in which experts carefully craft features for the model necessitates domain knowledge and frequently fails to capture the nuances of intricate datasets. Furthermore, these models’ ability to adjust to unfavorable weather or dynamic surroundings is hindered by their inability to quickly incorporate new data without requiring significant retraining, which limits their usefulness in situations that change quickly. The development of more sophisticated methods, or DL approaches, to overcome these obstacles and improve the power of AI systems has been made possible by this realization of their limitations.

5.2. DL Approaches

DL emerged as a subfield of machine learning and artificial intelligence in the 1990s [86]. Unlike traditional methods, DL offers distinct advantages, including the ability to achieve higher levels of abstraction, improved precision, and faster execution. These advantages make DL a valuable choice for OD. The OD algorithm employing DL typically consists of three core components: a recurrent neural network (RNN), a depth belief network (DBN), and a convolutional neural network (CNN). Overall, object detectors based on CNN [87] can be categorized into two primary types.
1. Single or one-stage detectors are known as the “non-regional proposal method” and “dense prediction”.
2. A two-stage detector is known as the “regional proposal method” and “sparse prediction,” as shown in Figure 10. In a single-stage detector, all the tasks are integrated into a unified system structure. Conversely, a two-stage detector separates the system into distinct stages for region selection, classification, and localization. Some regional proposal techniques consist of region-based CNN (R-CNN) [88], Fast R-CNN [89], Faster R-CNN [90], Spatial Pyramid Pooling Networks (SPPNet) [91], and Feature Pyramid Network (FPN) [92]. On the other side, the non-regional category includes Single Shot Multibox Detector (SSD) [93], You Only Look Once (YOLO 1-8) [94], EfficientDet [95], DEtection TRansformer (DETR) [96], and Fully Convolutional One-Stage (FCOS) [97]. Since these methods constitute the foundation of CNN, they have emerged as the standard for OD. The amalgamation of one-stage and two-stage OD techniques has gained prominence in the field of OD within AVs. These hybrid approaches have demonstrated effectiveness across diverse scenarios, achieving promising results in precise OD and localization. Two-stage algorithms tend to deliver superior accuracy, while one-stage algorithms offer faster processing speeds. Notably, the reasons behind the lower accuracy of one-stage algorithms remain unclear. A study examining the drawbacks of one-stage algorithms, especially those with dense sampling, was conducted in Ref. [98]. The study found that there was a major problem with the unbalanced background samples (negative instances) and foreground values (positive examples).
To solve this problem, the authors created Focal Loss, which alters the traditional cross-entropy loss function to give problematic cases more weight and thereby puts the emphasis on positive samples. After this adjustment, the RetinaNet network was created, showing better accuracy and quicker detection times than one- and two-stage methods. It should be mentioned too that the approach necessitates adjusting the focusing value, an extra hyperparameter [99]. However, the two primary state-of-the-art methods for DL-based detection these days are the YOLOv8 from the YOLO series (YOLOv1–YOLOv8) and the Faster R-CNN from the R-CNN family.
In order to speed up the region proposal step and allow the system to acquire knowledge of optimal region proposals, Ren et al. [90] proposed Faster R-CNN as an improvement to Fast R-CNN. The architecture of Faster R-CNN is described in Figure 11. This was achieved mainly by integrating the region proposal network (RPN). RPN’s main objective is to produce proposals with various ratios of aspects and sizes. Similar to Fast R-CNN, it employs high-resolution feature maps as input for RPN to identify the regions of interest in an image rather than depending on selective search for region recommendations. While Faster R-CNN improves the precision of detection and reaches an evaluation time of 0.2 s (five frames per second), it still requires a large amount of processing power and cannot fully meet the real-time system demands. It is also more complex because it requires training two interconnected neural networks, which makes system optimization difficult [100]. While Faster R-CNN algorithms demonstrate satisfactory accuracy on COCO and ImageNet datasets, they encounter difficulties when it comes to detecting small objects and objects at varying scales. This approach, while being more memory-efficient and faster, faces challenges in achieving consistent performance during both training and test-time inferences.
Jocher [100] implemented YOLOv8; however, that work has not yet been published in a journal. The structure of YOLOv8 is shown in Figure 12. Different versions are available with different scales of YOLO (N, S, M, L, and X) according to the scaling coefficient. The two main differences are the variety of kernels at different locations in the network and the number of feature extraction modules. These modifications are intended to meet the unique requirements of various situations. The backbone, neck, and head are the three primary modules that make up the YOLOv8 networks. First, the backbone module extracts features from the input photos. The neck component then performs feature fusion to create features in three different sizes: large, medium, and small. The final output is the detection results. Eventually, these characteristics of various sizes are sent to the head section for identification of the object. To improve deep image extraction features, the YOLOv8 system uses a layered C2f technique inside the backbone section. The main function of the SPPF layer, which is the tenth layer in the backbone module, is to transform input feature maps—regardless of their original size—into a fixed-size feature vector. The goal of this change is to increase the responsiveness of the network. YOLOv8 uses the traditional top-down as well as bottom-up feature fusion technique referred to as PAN (Path Aggregation Network) in the neck component. This tactic successfully combines elements from several scales. YOLOv8 adopts the widely used decoupled head architecture in the head section, which divides the categorization head from the identification head. Furthermore, there is a noticeable change in the network’s OD technique: the anchor-free strategy replaces the anchor-based method.
The distinction between traditional and DL approaches is justified by their fundamentally different methodologies and performance in various weather conditions. Traditional methods, while effective in ideal conditions, struggle with adverse weather due to their sensitivity to lighting, occlusion, and the need for precise feature engineering. DL methods, on the other hand, have shown superior performance in handling these challenges by learning robust representations from large datasets, which include diverse weather conditions. This allows DL-based systems to generalize better and maintain high detection accuracy even under challenging circumstances.

6. Traditional and DL Approaches for Object Detection in AVs

AVs need to be able to identify moving items like cyclists, livestock, pedestrians, and different kinds of automobiles, as well as stationary elements like traffic signals, road signs, and parked vehicles because pedestrians are especially vulnerable and because AVs frequently and critically engage with other types of vehicles. This section focuses only on the previous literature using traditional and DL approaches for vehicles, pedestrians, and road lane detection.

6.1. Traditional Approaches for Object Detection in AVs

In this section, we will summarize all the works on traditional approaches for OD in AVs.

6.1.1. Traditional Approaches for Vehicle Detection

Traditional vehicle detection techniques, which rely on appearance- and motion-based algorithms, have drawbacks, including being sensitive to changes in the camera’s position, having trouble telling moving objects apart, and requiring a large number of photos to accurately identify motion. Appearance-based algorithms that make use of features, such as HOG, Haar-Like, and SIFT, can improve identification, but they are vulnerable to exterior objects, adverse weather, and lighting variations. Even though they help to identify objects, biased classifiers like SVM and ANN might perform poorly in scenarios involving complicated traffic, which can result in false positives and make them susceptible to changes in the background, obstacles, lighting, and dimensions.
One study [101] investigated issues with physical detection techniques and equipment maintenance by examining the application of SVM and Pyramid Histogram of Words (PHOW) for traffic and road condition detection. The experimental findings show an accuracy of more than 82.5% on average under different weather circumstances. The suggested technique efficiently divides photos into four categories based on the type of traffic and road conditions. However, the paper does not detail the dataset’s diversity, which could impact the model’s robustness and generalizability. The research in [102] tackled the requirement for a reliable traffic surveillance system that takes into consideration variables like headlights, reflections, and shadows and operates well in a variety of weather and lighting circumstances. The suggested method is based on pattern classification and uses a distinctive image’s histogram as a distinguishing characteristic for classification. Because of its excellent generalization skills, SVM is used as a classifier. In tests, the system can discriminate between headlights and shadows with good accuracy and recognize moving cars in a variety of weather and lighting scenarios. However, this paper does not address the system’s performance under high-density traffic or its computational efficiency, which could be critical limitations for real-time applications. Using webcams as weather sensors, researchers [85] came up with a novel technique that allowed them to identify over 180,000 annotated photographs depending on indoor or outdoor circumstances with a 98% accuracy rate using SVM. The results of the research [103] indicate that DPM effectively addresses occlusion difficulties, even though it might not be required for ordinary detection settings. Accuracy is improved by adding image context information, and system speed is increased by adding other features. On the other hand, the multi-scale system only slightly improves accuracy, and the classifier has little effect on detector quality [21,104,105].
The vehicle detection method in [106], which utilizes pseudo-visual pursuit and HOG-LBP feature fusion, showcases the superiority of traditional machine learning techniques in achieving high detection accuracy and speed in severe weather conditions. This approach, with its reliance on radar information, demonstrates a clear advantage over methods that do not incorporate such advanced sensor data, particularly in scenarios where visibility is compromised. In contrast, the work by Abdullah et al. [107], while also effective, highlights the limitations of traditional methods when solely relying on MOG 2 background subtraction and histogram equalization. Their system’s high effectiveness rate of 96.4% across various weather conditions suggests that traditional techniques can still perform well. The vehicle detection system by Wu and Juang [108], which uses histogram extension and the Gray-level Differential Value Method (GDVM), shows that traditional methods can be effective in complex traffic conditions. However, the system’s reduced accuracy in rainy and dark conditions suggests a limitation of traditional approaches when dealing with significant environmental variations. The authors in [109] developed a system for concurrent vehicle and weather detection using optical flow and color histogram approaches. While this system demonstrates the utility of traditional methods in distinguishing between raindrops and normal conditions, it also indicates that these methods may struggle with the nuanced recognition tasks required in adverse weather.
Tian et al. [110] proposed a method for vehicle detection and tracking at night using HOG features and SVM for object recognition, effectively leveraging traditional machine learning techniques to enhance video surveillance in low-light conditions. Kuang et al. [111] achieved a high detection rate of 95.82% with minimal false positives in their multiclass system for nighttime conditions, employing tensor decomposition and object proposal methods. Furthermore, the research in [112] presents a system that integrates rule-based algorithms with a shallow neural network for nighttime vehicle detection, demonstrating significant improvements over traditional methods in terms of detection speed and proactive capabilities. These studies highlight the strengths of both traditional and hybrid approaches in the realm of nighttime vehicle detection, providing a foundation for future advancements in this field.
A vehicle tracking and counting system with camera shaking was presented in [113] for difficult sandy weather conditions. The method used particle filter tracking for vehicle detection, integrated headlight analysis, and background subtraction. While the method worked well in difficult situations, such as evenings and sandy weather, with a vibrating camera, it had trouble identifying white automobiles in snowy settings. Similarly, the background subtraction method encountered difficulties when there was movement in the backdrop or vibrations from the camera. A vehicle detection and tracking system for complex traffic was described in [114]. It used a background reduction technique with “low rank + sparse” disassembly and performed well in both qualitative and quantitative evaluations. Table 3 tabulates these works.

6.1.2. Traditional Approaches for Pedestrian Detection

This section delves into traditional techniques for pedestrian detection, comparing their effectiveness and limitations in various conditions. A notable study [115] presents a method that leverages an infrared camera for nighttime pedestrian detection, combining Haar-Cascade and Histogram of Oriented Gradients-Support Vector Machines (HOG-SVM) to reduce false alerts, demonstrating the superiority of this approach in low-light conditions. Another study [116] employed LiDAR point clouds and SVM for object clustering and classification in snowy environments, showcasing the robustness of this method for sidewalk snowplow machines at night. The real-world validation of this approach underscores its practical advantages over traditional vision-based systems in adverse weather. The work in [117] stands out for its real-time pedestrian detection system (PDS) that uses stereo-based segmentation and SVM to distinguish pedestrians from objects with vertical edge features. The high speed and detection rate of this method highlight its effectiveness, particularly in dynamic scenarios where camera motion is a factor. Research [118] addresses the challenge of detecting pedestrians from a moving vehicle in inclement weather, using SURF key points to compensate for camera motion and SVM for classification. The impressive performance of this approach, as evidenced by its speed and detection rate, underscores the potential of hybrid feature extraction methods in complex environments. Another study [119] combines Haar and HOG features to improve multi-person recognition and tracking, particularly in occluded and congested settings. The success of approaches in detecting both the side and frontal faces of pedestrians, as demonstrated through experiments on various datasets, highlights the limitations of single-feature approaches. The work in [120] introduces a novel method for pedestrian recognition in hazy environments by integrating a dark channel prior haze removal model with HOG descriptors. The improved performance of this method over conventional HOG-based techniques on the ’INRIA’ dataset indicates the benefits of adaptive preprocessing for robust detection in challenging conditions. With an emphasis on various weather situations, the study in [121] presents a robust vision-based pedestrian detection system for intelligent transportation systems. It incorporates feature extraction and SVM-based classification and addresses issues like variability and occlusions in real traffic photos. In a more constrained environment, the work in [122] demonstrates a pedestrian recognition method using SVM and HOG features, achieving an impressive 98.00% accuracy in counting individuals in a limited field of view. This high accuracy, achieved through training on a dataset of 2000 road photos, underscores the effectiveness of the model in controlled settings, potentially outperforming other approaches that do not account for such precision. Another study [123] provides a two-phase video recording pedestrian recognition method that combines local and global data with HOG features. The system’s efficacy in pedestrian detection is demonstrated by increased detection rates and decreased false positives obtained through testing with the PETS 2009 dataset.

6.1.3. Traditional Approaches for Road Lane Detection

In the study in [124], lane detection and sliding-mode control (SMC) were employed to demonstrate a novel approach for autonomous tracking control in intelligent electric vehicles (IEVs). This method, incorporating an optimal preview linear quadratic regulator (OP-LQR) with SMC and a 2-DOF vehicle model, outperforms traditional controllers in route tracking and lane marking recognition. Additionally, it enables efficient allocation of unequal braking force to all four wheels, showcasing superiority over conventional methods. Addressing the challenge of adverse weather conditions, Ref. [125] proposed a weather-resistant lane recognition method using horizontal optical flow, highlighting its capability for reliable obstacle assignment and lateral control essential for driver assistance and autonomous driving. Moreover, Ref. [126] presents an adaptive lane marking identification method for low-light conditions, demonstrating its accuracy and robustness compared to existing techniques. Similarly, Ref. [127] developed an advanced driver assistance system (ADAS) focusing on nighttime conditions, showcasing its effectiveness through successful highway tests. Lastly, Ref. [128] introduced a real-time lane detection technique for intelligent vehicles, highlighting its performance in complex road conditions and its potential for improving driving safety.

6.2. DL Approaches for Object Detection in AVs

DL has become much more popular and prominent than regular algorithms in recent times. This change in direction is attributed to DL’s exceptional work, reliable outcomes, and breadth of industry experience. Jones and Viola introduced the well-known VJ infrared technology, which improved immediate detection efficacy and capabilities [84]. In their research [129], Romero and Antonio mainly concentrated on defining DL methods. Previous relevant research on OD, e.g., vehicles, pedestrians, and road lanes, is examined in this section under adverse weather.

6.2.1. DL Approaches for Vehicle Detection

A key element of ITS is vehicle identification, which is used in many applications such as traffic monitoring, ADAS, AVs, and traffic data. Robustness, acceleration, precision, and cost-effectiveness are the hallmarks of an intelligent system. For this, a variety of imaging sources are used, including satellite imagery, in-car cameras, UAV cameras, and traffic monitoring equipment. This study is mostly focused on on-road vehicle surveillance, wherein in-car cameras are essential. Nonetheless, there are a variety of difficulties in the field of vehicle detection. There are several different classifications of vehicles, each having a distinct size, color, structure, and design. Additionally, they can be viewed at a variety of scales and positions, frequently in intricate traffic scenarios. Adverse weather conditions create extra challenges since they distort visibility and add noise to the sensors, which can result in missed objects and false alarms. This section will examine the major research works and methodologies that have influenced DL approaches for vehicle detection under adverse weather conditions, offering a basis for the discussion of novel strategies and developments in this crucial domain. Table 4 summarizes the previous work on vehicle detection under adverse weather conditions.
One paper [130] underscores the efficacy of CNNs in image processing through a tailored methodology for multiclass weather image categorization. Experimentation with various batch sizes (128, 256, 512, and 1024) is recommended to enhance accuracy and generalization. Testing on larger datasets enhances classifier accuracy and widens learning scope. In the realm of two-stage detectors, Ref. [131] introduces a practical approach for vehicle detection in challenging conditions like fog, low light, and sun glare, leveraging three trained CNNs (ResNet, GoogleNet, and SqueezeNet) with transfer learning. Notably, ResNet achieves a validation accuracy of 100%, while GoogleNet and SqueezeNet achieve 65.50% and 90.0%, respectively. Additionally, Ref. [132] employs Fast R-CNN for day and night vehicle detection, yielding impressive results under adverse conditions, validated by high recall (98.44%), accuracy (94.20%), and precision (90%). Furthermore, the on-road vehicle detection method proposed in [133] utilizes multiple region proposal networks (RPNs) in Faster R-CNN to identify vehicles across diverse sizes and weather scenarios, outperforming the existing techniques with high average precision rates on various datasets, including DAWN (89.48%), CDNet 2014 (91.20%), and LISA (95.16%). On the other hand, regarding one-stage detectors, another study [134] introduces an improved SSD-based front vehicle recognition algorithm for smart automobiles. Tested on the KITTI dataset, it achieves a high mAP of 92.18% and processes frames quickly at 15 ms per frame. The system enhances smart car safety in congested traffic and adverse weather conditions, prioritizing both accuracy and real-time performance. Another study [135] enhances a YOLO-based algorithm for precise vehicle detection in foggy weather by integrating a dehazing module with multi-scale retinex. This enhanced model, trained with augmented data, surpasses traditional YOLO in foggy conditions. Additionally, Ref. [136] proposes a modified YOLO model for on-road vehicle detection and tracking across various weather conditions. Utilizing a single CNN, it exhibits robustness and outperforms the existing techniques in intelligent transportation systems. Furthermore, Miao et al. [137] developed a nighttime vehicle detection method using fine-tuned YOLOv3, outperforming Faster R-CNN and SSD in terms of accuracy and efficiency. They achieved an average precision of 93.66%. Another study [138] utilized YOLOv4 with SPP-NET layers for vehicle detection, achieving an 81% mAP. In contrast, the study in [139] focused on harsh weather conditions, introducing YOLOv4 with an anchor-free and decoupled head, albeit achieving a 60.3% mAP and focusing exclusively on a single class. Moreover, the goal of [140] was to enhance self-driving vehicle detection in adverse weather using YOLOv5 with Transformer and CBAM modules, achieving an impressive mAP of 94.7% and FPS of 199.86. The DL approach proposed in [141] for nighttime vehicle detection in autonomous cars, combining a Generative Adversarial Network for image translation and YOLOv5 for detection, achieved a high accuracy of 96.75%, significantly enhancing the reliability of AV recognition models for night conditions. This study in [12] presents a DL-based intelligent AV weather-detecting system. Using a combined dataset from MCWDS2018 and DAWN2020, the performance of three deep convolutional neural networks was evaluated by categorizing six weather categories: overcast, rainy, snowy, sandy, sunny, and sunrise. The CNNs are SqueezeNet, ResNet-50, and EfficientNet-b0. The ResNet-50, EfficientNet-b0, and SqueezeNet models achieved 98.48%, 97.78%, and 96.05% accuracy, respectively, in the experiments, demonstrating remarkable accuracy rates while preserving quick inference times using GPU acceleration. By combining previously disparate datasets, the study’s novel methodology makes a significant addition to DL applications for autonomous systems’ weather detection.
In simple traffic situations, the aforementioned extraction and classification algorithms have shown themselves to be successful at detecting vehicles. However, issues including scale sensitivity, occlusion, and a high number of false positive results limit their efficacy in more complex traffic conditions [142]. These algorithms perform well in simple traffic scenarios, but they are limited in complex traffic conditions due to a significant number of false positive detections. As a result, an explanation connected with vehicle detection is intended for this part, and the difficulties associated are also highlighted.
Table 4. Vehicle detection previous work with DL.
Table 4. Vehicle detection previous work with DL.
Ref.YearTechniqueWeather ClassesAccuracy
[12]2022SqueezeNet, ResNet-50, EfficientNet-b0cloudy, rainy, snowy, sandy, shine,
and sunrise
98.48% 97.78% 96.05%
[137]2020YOLOv3night93.66%
[140]2022YOLOv5Fog, rain, snow, sand94.7%
[130]2023CNNCloudy, Fog, Rain, Shine Sunrise98%
[131]2022ResNet, GoogleNet, SqueezeNetFog, Low light, sun glare100% 65.50% 90.0%
[133]2021Faster R-CNNRain, Fog, Snow, Sand95.16%
[134]2020SSDSunny, overcast, Night, rain92.18%
[141]2022YOLOv5night96.75%

6.2.2. DL Approaches for Pedestrian Detection

A useful deformation model algorithm for managing occlusions is the Deformable Part Model (DPM), which scores blocks or sections [143]. It prevents objects from obscuring pedestrians. Classification algorithms use object features to determine which objects are pedestrians and which are not. A DL- and feature-selection-driven pedestrian detection system (PDS) is presented in [144]. The way the technology works is that a camera installed on the car’s dashboard records the surroundings as a video. After that, this video data are put through a number of preprocessing stages until pedestrians are finally classified and localized. One of the paper’s key findings is that the composition and quality of the datasets for training and testing that are used in the detection phase have a direct impact on the system’s efficacy.
The research presented in [145] demonstrates a significant advancement in nighttime human detection using visible light cameras, particularly through the application of convolutional neural networks (CNNs). The method’s effectiveness is evident in its superior performance across various contexts, including open and self-constructed databases such as DNHD-DB1, KAIST, and CVC. This achievement is a notable improvement over previous techniques and is crucial for the development of intelligent surveillance systems. Another study [146] addresses the challenges of pedestrian identification in adverse weather conditions by introducing a novel DL network framework. This framework’s ability to adjust to different lighting conditions through multispectral fusion and illumination perception is a clear advantage. The integration of cloud computing and the Internet of Vehicles (IoV) ecosystem further expands its potential for real-time applications in autonomous vehicles, offering a versatile and adaptable solution. Furthermore, Ref. [147] introduces an image fusion module (MSRCR-IF) that significantly enhances the Mask R-CNN-based pedestrian identification in low-light conditions. The 4.66% improvement in detection accuracy, reaching 85.05% on a self-built dataset, is a substantial gain over the existing mainstream techniques. This advancement is particularly valuable for surveillance systems operating in challenging lighting conditions. Table 5 summarizes the work. Another study [148] provides valuable insights into the challenges of data collection across diverse meteorological conditions and introduces the ZUT dataset, which boasts over 122k annotations. The research emphasizes the potential benefits of integrating Automobile CAN data with ADAS systems and suggests temperature-based normalization as a strategy for improvement. The recommendation to use 16-bit images for a 10.67% increase in detection accuracy and the consideration of car speed for parameter adjustments are noteworthy. The study also acknowledges the need for onboard precipitation sensors and a more extensive dataset that spans multiple seasons to enhance pedestrian detection accuracy, as demonstrated by the comparison with the SCUT dataset. The study in [149] focuses on metropolitan areas with a speed restriction of 48–70 km/h and showcases impressive performance on both rainy and dry roads. The method utilizes the SMP technique for depth calculation and stereo-vision technology with two cameras. The integration of data from various sensors, such as radar or LIDAR, is proposed to enhance autonomous driving performance. The research also advocates for the application of DL methods to refine feature extraction and categorization, achieving an AP of 89.85% and a mAP of 90% on the CVC-09 and LSIFIR databases. In the realm of nighttime pedestrian recognition, Ref. [150] introduces an enhanced Faster R-CNN architecture, which delivers significant improvements, particularly for distant pedestrians. The study in [151] tackles the issue of optical camera-based pedestrian recognition under adverse weather conditions by employing a cascaded classification technique that incorporates both local and global features, resulting in increased detection rates. The effectiveness of this strategy is demonstrated in challenging weather conditions such as snow or rain. The use of YOLO-v5 in [152] introduces a novel multispectral method for pedestrian detection, combining independent subnetworks for RGB and IR images to create an effective detector. The method’s performance on open databases, with high mAP scores and IoU of 0.5 across various datasets, is encouraging. and another study [153] evaluates YOLOv7, Faster R-CNN, SSD, and HoG for both vehicle and pedestrian detection in different weather conditions. YOLOv7’s superior performance in terms of accuracy, precision, recall, and mAP underscores the critical role of weather conditions in selecting appropriate DL methods for autonomous driving.

6.2.3. DL Approaches for Road Lane Detection

A color-based road lane-detecting system was used in research [154,155,156,157]. In favorable weather, Refs. [154,155] achieved above 90% accuracy. However, in adverse weather conditions, the accuracy considerably decreased. Some researchers used LiDAR technology for lane detection in order to reduce the impact of ambient influences on the detection algorithm. With the use of both images and LiDAR data, Shin et al. [158] developed a lane detection algorithm that combines camera and LiDAR information, improving the capacity to detect lanes in difficult environmental situations. By combining data from cameras, GPS, and LiDAR, the authors in [159] were able to develop a precise distance estimation between the vehicle and lateral lane that was accurate to the centimeter. However, the high cost of LiDAR and its limited sensitivity to unfavorable weather conditions make it difficult to apply to large-scale lane detection systems.
A different method [160] uses the HSV color space to achieve an accuracy of 86.21% regarding the rate of identification. This method works best in daylight or with white light because it takes advantage of the near-color correspondence between white and yellow lane markers. Consequently, global thresholding is necessary for precisely dividing color planes and obtaining lane markers in both the YCbCr and HSV color models. In contrast, in tunnel environments with distinct color illumination, the lane markers diverge from their real hues, making reliable detection difficult. Sattar et al. [161] developed a different technique called “SafeDrive,” which identifies the lane marking in areas with poor lighting. They located other distinct images of highways at the same place using the vehicle’s location information, and they used those images to identify the lanes. Other researchers [162] also carried out an additional study that, because of the solid and easily visible lane markings, was able to identify lanes in several types of weather conditions.
Using already installed roadside webcams, real-time road weather and road state detection algorithms were developed in [163]. Transfer learning was utilized to train detection models utilizing three previously learned CNN architectures. When it came to accuracy, ResNet18 outperformed the other models, scoring about 97% for weather and 99% for surface conditions. These models can be useful in improving road safety since they can automatically recognize and relate real-time circumstances to road networks without the need for human intervention. They can be incorporated into advanced traveler information systems (ATIS) for drivers and utilized to optimize winter maintenance routes. Additionally, the models could replace manual reporting in snowplows, increasing driver safety and accuracy. Research [164] presents a novel approach that combines computer vision and DL to extract meteorological data from street-level photos without any image restrictions. With recognition accuracy ranging from 91% to 95.6% across several categories, the study presents four deep CNN models designed to identify a variety of visibility circumstances, including dawn and dusk, daytime, evening, glare, and weather elements like rain and snow. A study [165] presents a DL-based method for lane marker detection in order to address the difficulties associated with classifying road markers during rainy weather. Specifically designed for bad weather, the approach gives priority to the best feature selection from video frames in order to counteract rain-induced blurriness, leading to adequate classification precision even in difficult weather circumstances. In other research [166], two different models for AV lane prediction in highway traffic situations are presented. The study’s main focus is on the use of AI for lane prediction using a sizable dataset from NGSIM. To showcase the effectiveness of these models, two distinct subsets with 5000 car samples each were used. The strategy employed a range of classifiers in the Identification Learner and different methods in Neural Net Fitting, emphasizing the methodology’s importance in accomplishing good lane prediction without diving into particular accuracy measures.
Due to different illumination circumstances, global thresholds for lane marker identification frequently produce disappointing outcomes. An efficient lane departure (LD) technique is presented, enabling lane marker identification that may be used in daylight highway scenarios as well as tunnel situations with artificial lighting. In order to obtain lane features, the usual LD technique begins by performing pre-processing to eradicate perspective-related distortion and isolate an ROI. In order to identify suitable lane markers using color, shape, alignment, or geometrical measurements using the road scenario data, two categorization methods, model-based and feature-based, are used [17]. To minimize false positives, prospective lane markers are further refined and validated using curved or linear lane fitting. This helps with features like adaptive speed control, automatic lane focusing, and lane departure alert. LD devices may not be as successful due to a number of issues, such as road obstructions, ambient lighting sources, and noisy weather. To address these issues, a shadow-resistant methodology that makes use of Fourier transformation and the maximally stable extreme region (MSER) approach in the blue color stream is used as a reference [167]. The research findings [168] suggest that the minimal safe separation between the self-absorbed car and the vehicle should be used to estimate the extraction of ROI. According to the study, a car traveling at 110 km/h may cover a distance of 35 m with a height of just 150 pixels. Usually, a predetermined model or set of features is used to construct lane markers using the ROI.
In [169], the authors modified YOLOv3 and presented a brand-new BGRU-Lane (BGRU-L) model; the method integrates spatial distribution with visual data. High accuracy (90.28 mAP) and real-time detection speed (40.20 fps) are achieved in difficult settings through integration utilizing the Dempster–Shafer algorithm, as demonstrated by datasets from KIT, Toyota Technological Institute, and Euro Truck Simulator 2.
The dynamic environment of research and innovation in the field of road lane detection for autonomous cars is shown by our thorough analysis of the literature. The abundance of information and developing techniques indicates the ongoing commitment to increasing the potential of self-governing systems.

6.3. Practical Implications

The practical implementation of DL models for pedestrian and vehicle detection in AVs is a critical area of research as it directly impacts the safety and reliability of these systems. Studies such as Kim et al. [145] have demonstrated the effectiveness of CNNs in low-light conditions, a significant advancement for nighttime AV navigation. Lai et al. [147] have further optimized the Mask R-CNN algorithm for low-light environments, enhancing detection rates and computational efficiency. Montenegro et al. [152] have shown that YOLO-v5 can be fine-tuned to maintain high detection accuracy across varying lighting conditions, ensuring consistent performance for AVs. Zaman et al. [153] have addressed the challenge of adverse weather conditions, a common obstacle for AVs, by developing DL models that can adapt to weather-induced distortions, thereby improving system reliability. These studies have identified key performance indicators (KPIs) such as detection accuracy, false negative rate, computational efficiency, and real-time processing capabilities, which are essential for evaluating the practicality of DL models in AVs, as shown in Table 6. The challenges faced, including limited visibility and dynamic lighting conditions, have been mitigated through model optimization, data augmentation, and sensor fusion techniques. The insights from these studies are invaluable for AV developers as they provide a roadmap for creating systems that can operate effectively in a wide range of environmental conditions, bringing us closer to a future where autonomous vehicles are a safe and integral part of our transportation infrastructure.

7. Discussion, Limitations, and Future Research Trends

7.1. Discussion

Recent literature highlights the increasing use of DL approaches for OD, yielding promising results under typical conditions. However, there is a pressing need for further advancements, particularly in adverse weather and complex scenarios. Enhancing OD in such conditions is crucial, especially for AVs, to prevent accidents and ensure safety. In the following discussion, we delve into the key insights and implications drawn from our research on OD in adverse weather conditions for AVs. Our research has explored both traditional and DL methods for AV object detection, focusing on vehicles, pedestrians, and road lanes. Traditional methods, despite their foundational role, struggle with high computational demands, slow processing, and occasional misidentification. DL, on the other hand, excels by learning complex patterns from data, offering faster and more precise detection, especially in challenging weather conditions. This makes DL a more effective and adaptable solution for AV systems. Traditional vehicle detection techniques, which rely on appearance- and motion-based algorithms, face significant challenges in adverse weather conditions. Appearance-based algorithms like HOG, Haar-like, and SIFT are sensitive to exterior objects, adverse weather, and lighting variations, making them vulnerable in situations like heavy rain or fog. These algorithms often struggle to maintain accuracy due to the reduced visibility of vehicle features and the difficulty in differentiating vehicles from their surroundings. Motion-based detection methods, which track objects based on movement relative to the camera, also encounter issues in adverse weather. They can have trouble distinguishing between moving vehicles and other dynamic elements, such as pedestrians or debris, especially when motion cues are obscured by weather conditions. DL approaches are more accurate and perform better, as shown in Table 4 and Table 5. A comparison of traditional and DL approaches is shown in Table 7.
Figure 13 presents the overall percentage distribution of papers, while Figure 14 describes the papers of traditional and DL approaches related to the studied three object detection issues. From Figure 13, one can see that vehicle detection is frequently studied using traditional and DL approaches, followed by pedestrian and road lane detection. As shown in Figure 14, DL approaches were more frequently used for all three issues in AVs compared to traditional approaches. Prior to 2008, traditional feature extraction methods were prevalent for detection and classification but had limitations in adverse conditions. Manual feature extraction made them less suitable for complex applications. DL has become much more popular and prominent than regular algorithms in recent times. This change in direction is attributed to DL’s exceptional work, reliable outcomes, and breadth of industry experience. Romero and Antonio [129] also mainly concentrated on defining DL methods. Peer review has been applied to DL one-stage and two-stage detectors from the investigations. Currently, the leading methods are YOLOv8 from the YOLO series and Faster R-CNN from the R-CNN family, renowned for their superior accuracy and performance. Figure 15 shows the number of summarized papers for one-stage and two-stage detectors, which shows that one-stage detectors are frequently used for vehicle detection, while, for pedestrian detection, both are equally applied. Two-stage detectors are frequently used for road lanes compared with single-stage detectors.
In order to improve OD, Faster R-CNN was integrated into the RPN. This resulted in shorter evaluation times, but it also required a significant amount of processing power and had difficulties maintaining consistent performance for small and variously scaled objects.YOLOv8 is designed to cater to specific requirements. The YOLOv8 structure comprises backbone, neck, and head modules, enhancing feature extraction with C2f and employing the PAN feature fusion technique. It also adopts a decoupled head architecture and introduces an anchor-free strategy for OD. Furthermore, we analyzed high-quality datasets related to AVs, and some standard evaluation metrics were also reviewed and summarized. Our discussion underscores the advancements in OD for autonomous vehicles under adverse weather, paving the way for enhanced safety and reliability in ADs.

7.2. Limitations

  • Our literature review primarily focuses on the detection of pedestrians, vehicles, and road lanes, which may not encompass all possible objects and scenarios relevant to AVs in adverse weather conditions. There may be other critical elements, such as traffic signals or animals, that warrant further investigation.
  • The detection algorithms discussed in our review may have inherent limitations, such as difficulties in detecting small or occluded objects, which could impact the accuracy and reliability of AVs in certain situations.
  • Our study primarily considers a range of adverse weather conditions, but there may be other environmental factors, such as dust storms or heavy fog, that were not extensively covered in the reviewed literature.
  • The field of AV technology is rapidly evolving, and new detection algorithms and sensor technologies may emerge that could significantly impact the findings of our review. Our study may not capture the most recent advancements in this area.
  • While our study includes datasets that simulate adverse weather conditions, the simulation environments may not perfectly replicate real-world driving scenarios. The complexity of real-world conditions, including unpredictable human behavior and dynamic traffic patterns, can introduce additional challenges not fully captured in simulated datasets.
  • The ethical considerations and societal acceptance of AVs, especially in challenging conditions, are not addressed in our study. Public trust and the ethical use of AV technology are essential factors for their successful integration into smart cities.

7.3. Future Research Trends

  • It has become more important to address the real-time requirements for OD in real-world applications. Deep neural networks, however, frequently require large amounts of computational power, which presents difficulties for embedded systems. To properly fulfill these objectives, resource-efficient technique development has become essential. To ensure the practical usefulness of the suggested methodologies, future research should focus heavily on their computational components, offering both quantitative and qualitative analysis.
  • The existing deep neural network techniques for difficult item detection mainly depend on large-scale annotated datasets. However, creating these databases is expensive and time-consuming. Consequently, there is an urgent need to create OD algorithms that can train with little to no labeled information.
  • The employment of various evaluation metrics and IoU criteria for OD in difficult situations has resulted in the absence of a clear benchmark standard, which makes comparing methods difficult. For future research in this area to be uniform and comparable, a global baseline must be established.
  • Creation of extensive simulation environments should occur that imitate inclement weather to thoroughly test and improve object identification algorithms.
  • To develop comprehensive solutions for adverse weather OD, researchers, engineers, and policymakers should collaborate more closely.
  • It is necessary to study the psychology and behavior of human drivers in adverse weather, with an emphasis on developing successful communication and trust with AVs.
  • Creation of novel tactics for the real-time modification of OD algorithms for AVs in response to changing environmental circumstances should be achieved.
  • Investigation into cutting-edge techniques for combining current weather data with weather forecasts to enable proactive decisionmaking during unfavorable weather conditions is necessary.
  • Improvements in sensor fusion methods should be attained, which integrate information from several sensor types to provide more accurate and dependable identification in adverse weather.
  • To develop behavior prediction models for AVs, leveraging machine learning and deep learning to forecast the actions of vehicles, pedestrians, and cyclists should occur. These models will operate effectively in adverse weather, improving AV decisionmaking for enhanced safety and efficiency.

8. Conclusions

This paper reviewed the traditional and DL approaches for vehicles, pedestrians, and road lane detection in AVs under adverse weather conditions. We first studied the architecture of AVs with sensor technologies and other components and also discussed the challenges for AVs in adverse weather. After an overview of almost all the datasets related to AVs covering different weather conditions, we explained the basic structure of OD and the evaluation matrices used for it. Then, we explained the traditional approaches and DL approaches and discussed the traditional feature extraction methods that were prevalent for detection and classification but had limitations in adverse conditions. Manual feature extraction made them less suitable for complex applications. DL has become much more popular and prominent than regular algorithms in more recent times. DL approaches explain the structure of YOLOV8 (one-stage detectors) and Faster R-CNN (two-stage detectors). Two-stage algorithms tend to deliver superior accuracy, while one-stage algorithms offer faster processing speeds. Notably, the reasons behind the lower accuracy of one-stage algorithms remain unclear. In addition, the statistics about the status quo of traditional and DL approaches for OD in AVs were provided based on the works collected in this survey paper. We found that DL was intensively used for vehicles, pedestrians, and road lane detection in AVs compared with the traditional approaches. Specifically, one-stage detectors were frequently used for vehicle detection compared with two-stage detectors. In addition, vehicle detection was frequently studied using both traditional and DL approaches followed by road lane and pedestrian detection. Finally, we presented a useful discussion along with some future research directions.

Author Contributions

All authors have equally contributed. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by EIAS Data Science Lab, CCIS, Prince Sultan University and also by the Special Research Fund (BOF) of Hasselt University (No. BOF23DOCBL11), and Chen Junhong was sponsored by the China Scholarship Council (No. 202208440309).

Acknowledgments

The authors would like to thank Prince Sultan University for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. World Health Organization. WHO: Global Status Report on Road Safety: Summary; Technical Report. Available online: https://www.who.int/health-topics/road-safety#tab=tab_1 (accessed on 18 February 2024).
  2. Iftikhar, S.; Asim, M.; Zhang, Z.; El-Latif, A.A.A. Advance generalization technique through 3D CNN to overcome the false positives pedestrian in autonomous vehicles. Telecommun. Syst. 2022, 80, 545–557. [Google Scholar] [CrossRef]
  3. Self-Driving Cars Global Market Size. Available online: https://precedenceresearch.com/self-driving-cars-market (accessed on 18 December 2023).
  4. Badue, C.; Guidolini, R.; Carneiro, R.V.; Azevedo, P.; Cardoso, V.B.; Forechi, A.; Jesus, L.; Berriel, R.; Paixao, T.M.; Mutz, F.; et al. Self-driving cars: A survey. Expert Syst. Appl. 2021, 165, 113816. [Google Scholar] [CrossRef]
  5. Bansal, P.; Kockelman, K.M. Are we ready to embrace connected and self-driving vehicles? A case study of Texans. Transportation 2018, 45, 641–675. [Google Scholar] [CrossRef]
  6. Chehri, A.; Sharma, T.; Debaque, B.; Duclos, N.; Fortier, P. Transport systems for smarter cities, a practical case applied to traffic management in the city of Montreal. In Sustainability in Energy and Buildings 2021; Springer: Singapore, 2022; pp. 255–266. [Google Scholar]
  7. Trenberth, K.E.; Zhang, Y. How often does it really rain? Bull. Am. Meteorol. Soc. 2018, 99, 289–298. [Google Scholar] [CrossRef]
  8. Andrey, J.; Yagar, S. A temporal analysis of rain-related crash risk. Accid. Anal. Prev. 1993, 25, 465–472. [Google Scholar] [CrossRef]
  9. National Oceanic and Atmospheric Administration, Getting Traction: Tips for Traveling in Winter Weather. Available online: https://www.weather.gov/wrn/getting_traction (accessed on 24 October 2023).
  10. Mehra, A.; Mandal, M.; Narang, P.; Chamola, V. ReViewNet: A fast and resource optimized network for enabling safe autonomous driving in hazy weather conditions. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4256–4266. [Google Scholar] [CrossRef]
  11. Abu Al-Haija, Q.; Krichen, M.; Abu Elhaija, W. Machine-learning-based darknet traffic detection system for IoT applications. Electronics 2022, 11, 556. [Google Scholar] [CrossRef]
  12. Al-Haija, Q.A.; Gharaibeh, M.; Odeh, A. Detection in adverse weather conditions for autonomous vehicles via deep learning. AI 2022, 3, 303–317. [Google Scholar] [CrossRef]
  13. Zang, S.; Ding, M.; Smith, D.; Tyler, P.; Rakotoarivelo, T.; Kaafar, M.A. The impact of adverse weather conditions on autonomous vehicles: How rain, snow, fog, and hail affect the performance of a self-driving car. IEEE Veh. Technol. Mag. 2019, 14, 103–111. [Google Scholar] [CrossRef]
  14. Mohammed, A.S.; Amamou, A.; Ayevide, F.K.; Kelouwani, S.; Agbossou, K.; Zioui, N. The perception system of intelligent ground vehicles in all weather conditions: A systematic literature review. Sensors 2020, 20, 6532. [Google Scholar] [CrossRef]
  15. Yoneda, K.; Suganuma, N.; Yanase, R.; Aldibaja, M. Automated driving recognition technologies for adverse weather conditions. IATSS Res. 2019, 43, 253–262. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Carballo, A.; Yang, H.; Takeda, K. Perception and sensing for autonomous vehicles under adverse weather conditions: A survey. ISPRS J. Photogramm. Remote Sens. 2023, 196, 146–177. [Google Scholar] [CrossRef]
  17. Arthi, V.; Murugeswari, R.; Nagaraj, P. Object Detection of Autonomous Vehicles under Adverse Weather Conditions. In Proceedings of the 2022 International Conference on Data Science, Agents & Artificial Intelligence (ICDSAAI), Chennai, India, 8–10 December 2022; Volume 1, pp. 1–8. [Google Scholar]
  18. Hnewa, M.; Radha, H. Object detection under rainy conditions for autonomous vehicles: A review of state-of-the-art and emerging techniques. IEEE Signal Process. Mag. 2020, 38, 53–67. [Google Scholar] [CrossRef]
  19. Abbas, A.F.; Sheikh, U.U.; Al-Dhief, F.T.; Mohd, M.N.H. A comprehensive review of vehicle detection using computer vision. TELKOMNIKA Telecommun. Comput. Electron. Control 2021, 19, 838–850. [Google Scholar] [CrossRef]
  20. Yang, Z.; Pun-Cheng, L.S. Vehicle detection in intelligent transportation systems and its applications under varying environments: A review. Image Vis. Comput. 2018, 69, 143–154. [Google Scholar] [CrossRef]
  21. Muneeswaran, V.; Nagaraj, P.; Rajasekaran, M.P.; Reddy, S.U.; Chaithanya, N.S.; Babajan, S. IoT based Multiple Vital Health Parameter Detection and Analyzer System. In Proceedings of the 2022 7th International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 22–24 June 2022; pp. 473–478. [Google Scholar] [CrossRef]
  22. SAE International. Available online: https://www.sae.org/standards/content/j3016_202104/ (accessed on 11 November 2023).
  23. Bimbraw, K. Autonomous cars: Past, present and future a review of the developments in the last century, the present scenario and the expected future of autonomous vehicle technology. In Proceedings of the 2015 12th International Conference on Informatics in Control, Automation and Robotics (ICINCO), Colmar, France, 21–23 July 2015; Volume 1, pp. 191–198. [Google Scholar]
  24. Lafrance, A. Your Grandmother’s Driverless Car. The Atlantic, 29 June 2016. Available online: https://www.theatlantic.com/technology/archive/2016/06/beep-beep/489029/ (accessed on 27 October 2023).
  25. Pomerleau, D.A. Alvinn: An autonomous land vehicle in a neural network. In NIPS’88: Proceedings of the 1st International Conference on Neural Information Processing Systems, 1 January 1988; MIT Press: Cambridge, MA, USA; pp. 305–313.
  26. SAE International. Surface Vehicle Recommended Practice. Joint SAE/RCCC Fuel Consumption Test Procedure (Short Term in-Service Vehicle) Type I. 1986. Available online: https://ca-times.brightspotcdn.com/54/02/2d5919914cfe9549e79721b12e66/j372016-202104.pdf (accessed on 1 May 2021).
  27. How Google’s Autonomous Car Passed the First US State Self-Driving Test. 2014. Available online: https://spectrum.ieee.org/how-googles-autonomous-car-passed-the-first-us-state-selfdriving-test (accessed on 31 December 2023).
  28. Claire, P. The Pathway to Driverless Cars: Summary Report and Action Plan; OCLC: London, UK, 2015. [Google Scholar]
  29. Briefs, U. Mcity Grand Opening. Res. Rev. 2015, 46, 1–3. [Google Scholar]
  30. Bellone, M.; Ismailogullari, A.; Müür, J.; Nissin, O.; Sell, R.; Soe, R.M. Autonomous driving in the real-world: The weather challenge in the Sohjoa Baltic project. In Towards Connected and Autonomous Vehicle Highways: Technical, Security and Social Challenges; Springer International Publishing: Cham, Switzerland, 2021; pp. 229–255. [Google Scholar]
  31. Lambert, F. Watch Tesla Autopilot Go through a Snowstorm. Available online: https://electrek.co/2019/01/28/tesla-autopilot-snow-storm/ (accessed on 2 January 2023).
  32. EUR-Lex. E.L. EUR-Lex-32019R2144-EN-EUR-Lex. 2014. Available online: https://eur-lex.europa.eu/legal-content/en/ALL/?uri=CELEX%3A32019R2144 (accessed on 13 October 2023).
  33. Waymo, W. Waymo Safety Report. 2020. Available online: https://waymo.com/safety/ (accessed on 18 August 2023).
  34. Gehrig, S.K.; Stein, F.J. Dead reckoning and cartography using stereo vision for an autonomous car. In Proceedings of the 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289), Kyongju, Republic of Korea, 17–21 October 1999; Volume 3, pp. 1507–1512. [Google Scholar] [CrossRef]
  35. Rawashdeh, N.A.; Bos, J.P.; Abu-Alrub, N.J. Drivable path detection using CNN sensor fusion for autonomous driving in the snow. In Proceedings of the SPIE 11748, Autonomous Systems: Sensors, Processing, and Security for Vehicles and Infrastructure 2021, Online, FL, USA, 12–17 April 2021; Volume 1174806. [Google Scholar] [CrossRef]
  36. Thrun, S.; Montemerlo, M.; Dahlkamp, H.; Stavens, D.; Aron, A.; Diebel, J.; Fong, P.; Gale, J.; Halpenny, M.; Hoffmann, G.; et al. Stanley: The robot that won the DARPA Grand Challenge. J. Field Robot. 2006, 23, 661–692. [Google Scholar] [CrossRef]
  37. Carballo, A.; Lambert, J.; Monrroy, A.; Wong, D.; Narksri, P.; Kitsukawa, Y.; Takeuchi, E.; Kato, S.; Takeda, K. LIBRE: The multiple 3D LiDAR dataset. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1094–1101. [Google Scholar] [CrossRef]
  38. Velodyne, HDL-64E Spec Sheet. Available online: https://velodynesupport.zendesk.com/hc/en-us/articles/115003632634-HDL-64E-Spec-Sheet (accessed on 13 October 2023).
  39. Bijelic, M.; Gruber, T.; Ritter, W. A benchmark for lidar sensors in fog: Is detection breaking down? In Proceedings of the Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 760–767. [Google Scholar] [CrossRef]
  40. Patole, S.M.; Torlak, M.; Wang, D.; Ali, M. Automotive radars: A review of signal processing techniques. IEEE Signal Process. Mag. 2017, 34, 22–35. [Google Scholar] [CrossRef]
  41. Navtech Radar, FMCW Radar. Available online: https://navtechradar.com/explore/fmcw-radar/ (accessed on 22 November 2021).
  42. Carullo, A.; Parvis, M. An ultrasonic sensor for distance measurement in automotive applications. IEEE Sens. J. 2001, 1, 143. [Google Scholar] [CrossRef]
  43. Frenzel, L. Ultrasonic Sensors: A Smart Choice for Shorter-Range Applications. 2018. Available online: https://www.electronicdesign.com/industrialautomation/article/21806202/ultrasonic-sensors-a-smart-choicefor-shorterrange-applications (accessed on 14 August 2023).
  44. Kamemura, T.; Takagi, H.; Pal, C.; Ohsumi, A. Development of a long-range ultrasonic sensor for automotive application. SAE Int. J. Passeng. Cars-Electron. Electr. Syst. 2008, 1, 301–306. [Google Scholar] [CrossRef]
  45. Tesla, Summon Your Tesla from Your Phone. Available online: https://www.tesla.com/blog/summon-your-tesla-your-phone (accessed on 2 November 2023).
  46. Pendleton, S.D.; Andersen, H.; Du, X.; Shen, X.; Meghjani, M.; Eng, Y.H.; Rus, D.; Ang, M.H., Jr. Perception, planning, control, and coordination for autonomous vehicles. Machines 2017, 5, 6. [Google Scholar] [CrossRef]
  47. Nasseri, A.; Shlomit, H. Autonomous Vehicle Technology Report. 2020. Available online: https://www.wevolver.com/article/ (accessed on 28 October 2023).
  48. Rosique, F.; Navarro, P.J.; Fernández, C.; Padilla, A. A systematic review of perception system and simulators for autonomous vehicles research. Sensors 2019, 19, 648. [Google Scholar] [CrossRef] [PubMed]
  49. Paden, B.; Čáp, M.; Yong, S.Z.; Yershov, D.; Frazzoli, E. A survey of motion planning and control techniques for self-driving urban vehicles. IEEE Trans. Intell. Veh. 2016, 1, 33–55. [Google Scholar] [CrossRef]
  50. Shladover, S.E.; Desoer, C.A.; Hedrick, J.K.; Tomizuka, M.; Walrand, J.; Zhang, W.B.; McMahon, D.H.; Peng, H.; Sheikholeslam, S.; McKeown, N. Automated vehicle control developments in the PATH program. IEEE Trans. Veh. Technol. 1991, 40, 114–130. [Google Scholar] [CrossRef]
  51. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  52. Sun, H.; Zhang, W.; Yu, R.; Zhang, Y. Motion planning for mobile robots—Focusing on deep reinforcement learning: A systematic review. IEEE Access 2021, 9, 69061–69081. [Google Scholar] [CrossRef]
  53. Rai, M.; Khosla, B.; Dhawan, Y.; Kharotia, H.; Kumar, N.; Bandi, A. CLEAR: An Efficient Traffic Sign Recognition Technique for Cyber-Physical Transportation Systems. In Proceedings of the 2022 IEEE/ACM 15th International Conference on Utility and Cloud Computing (UCC), Vancouver, WA, USA, 6 December 2022; pp. 418–423. [Google Scholar]
  54. Sakaridis, C.; Dai, D.; Van Gool, L. ACDC: The adverse conditions dataset with correspondences for semantic driving scene understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10765–10775. [Google Scholar]
  55. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S.; et al. A2d2: Audi autonomous driving dataset. arXiv 2020, arXiv:2004.06320. [Google Scholar]
  56. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in perception for autonomous driving: Waymo open dataset. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2443–2451. [Google Scholar] [CrossRef]
  57. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar] [CrossRef]
  58. Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2636–2645. [Google Scholar] [CrossRef]
  59. Huang, X.; Wang, P.; Cheng, X.; Zhou, D.; Geng, Q.; Yang, R. The apolloscape open dataset for autonomous driving and its application. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2702–2719. [Google Scholar] [CrossRef] [PubMed]
  60. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The synthia dataset: A large collection of synthetic images for semantic segmentation of urban scenes. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 3234–3243. [Google Scholar] [CrossRef]
  61. Carlevaris-Bianco, N.; Ushani, A.K.; Eustice, R.M. University of Michigan North Campus long-term vision and lidar dataset. Int. J. Robot. Res. 2016, 35, 1023–1035. [Google Scholar] [CrossRef]
  62. Pitropov, M.; Garcia, D.E.; Rebello, J.; Smart, M.; Wang, C.; Czarnecki, K.; Waslander, S. Canadian adverse driving conditions dataset. Int. J. Robot. Res. 2021, 40, 681–690. [Google Scholar] [CrossRef]
  63. Zheng, J.Y. IUPUI driving videos and images in all weather and illumination conditions. arXiv 2021, arXiv:2104.08657. [Google Scholar]
  64. Bos, J.P.; Chopp, D.; Kurup, A.; Spike, N. Autonomy at the end of the earth: An inclement weather autonomous driving data set. In Proceedings of the Autonomous Systems: Sensors, Processing, and Security for Vehicles and Infrastructure 2020, Online, CA, USA, 27 April–9 May 2020; Volume 11415, pp. 36–48. [Google Scholar] [CrossRef]
  65. Vora, S.; Lang, A.H.; Helou, B.; Beijbom, O. Pointpainting: Sequential fusion for 3d object detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4603–4611. [Google Scholar] [CrossRef]
  66. Chang, M.F.; Lambert, J.; Sangkloy, P.; Singh, J.; Bak, S.; Hartnett, A.; Wang, D.; Carr, P.; Lucey, S.; Ramanan, D.; et al. Argoverse: 3d tracking and forecasting with rich maps. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 8748–8757. [Google Scholar] [CrossRef]
  67. Meyer, M.; Kuschk, G. Automotive radar dataset for deep learning based 3d object detection. In Proceedings of the 2019 16th European Radar Conference (EuRAD), Paris, France, 2–4 October 2019; pp. 129–132. [Google Scholar]
  68. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 15–20 June 2020; pp. 11682–11692. [Google Scholar]
  69. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef]
  70. Kim, J.; Rohrbach, A.; Darrell, T.; Canny, J.; Akata, Z. Textual Explanations for Self-Driving Vehicles. In Computer Vision–ECCV 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 11206. [Google Scholar]
  71. Neuhold, G.; Ollmann, T.; Rota Bulo, S.; Kontschieder, P. The mapillary vistas dataset for semantic understanding of street scenes. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4990–4999. [Google Scholar]
  72. Braun, M.; Krebs, S.; Flohr, F.; Gavrila, D.M. Eurocity persons: A novel benchmark for person detection in traffic scenes. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1844–1861. [Google Scholar] [CrossRef]
  73. Maddern, W.; Pascoe, G.; Linegar, C.; Newman, P. 1 year, 1000 km: The oxford robotcar dataset. Int. J. Robot. Res. 2017, 36, 3–15. [Google Scholar] [CrossRef]
  74. Pham, Q.H.; Sevestre, P.; Pahwa, R.S.; Zhan, H.; Pang, C.H.; Chen, Y.; Mustafa, A.; Chandrasekhar, V.; Lin, J. A 3D dataset: Towards autonomous driving in challenging environments. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 2267–2273. [Google Scholar]
  75. Wenzel, P.; Wang, R.; Yang, N.; Cheng, Q.; Khan, Q.; von Stumberg, L.; Zeller, N.; Cremers, D. 4Seasons: A cross-season dataset for multi-weather SLAM in autonomous driving. In Pattern Recognition; Akata, Z., Geiger, A., Sattler, T., Eds.; DAGM GCPR 2020. Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2021; Volume 12544, pp. 404–417. [Google Scholar]
  76. Zendel, O.; Honauer, K.; Murschitz, M.; Steininger, D.; Dominguez, G.F. Wilddash-creating hazard-aware benchmarks. In Proceedings of the Computer Vision–ECCV 2018: 15th European Conference, Munich, Germany, 8–14 September 2018; Proceedings, Part VI. pp. 402–416. [Google Scholar]
  77. Choi, Y.; Kim, N.; Hwang, S.; Park, K.; Yoon, J.S.; An, K.; Kweon, I.S. KAIST multi-spectral day/night data set for autonomous and assisted driving. IEEE Trans. Intell. Transp. Syst. 2018, 19, 934–948. [Google Scholar] [CrossRef]
  78. Sheeny, M.; De Pellegrin, E.; Mukherjee, S.; Ahrabian, A.; Wang, S.; Wallace, A. RADIATE: A radar dataset for automotive perception in bad weather. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 1–7. [Google Scholar]
  79. Yan, Z.; Sun, L.; Krajník, T.; Ruichek, Y. EU long-term dataset with multiple sensors for autonomous driving. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 10697–10704. [Google Scholar]
  80. Burnett, K.; Yoon, D.J.; Wu, Y.; Li, A.Z.; Zhang, H.; Lu, S.; Qian, J.; Tseng, W.K.; Lambert, A.; Leung, K.Y.; et al. Boreas: A multi-season autonomous driving dataset. Int. J. Robot. Res. 2023, 42, 33–42. [Google Scholar] [CrossRef]
  81. Kenk, M.A.; Hassaballah, M. DAWN: Vehicle detection in adverse weather nature dataset. arXiv 2020, arXiv:2008.05402. [Google Scholar]
  82. Saralajew, S.; Ohnemus, L.; Ewecker, L.; Asan, E.; Isele, S.; Roos, S. A dataset for provident vehicle detection at night. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 9750–9757. [Google Scholar]
  83. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–25 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  84. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001, Kauai, HI, USA, 8–14 December 2001. [Google Scholar] [CrossRef]
  85. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–26 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  86. Asim, M.; Wang, Y.; Wang, K.; Huang, P.Q. A review on computational intelligence techniques in cloud and edge computing. IEEE Trans. Emerg. Top. Comput. Intell. 2020, 4, 742–763. [Google Scholar] [CrossRef]
  87. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  88. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  89. Girshick, R. Fast r-cnn. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar] [CrossRef]
  90. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  91. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  92. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef]
  93. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision–ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2016; Volume 9905, pp. 21–37. [Google Scholar]
  94. Terven, J.; Cordova-Esparza, D. A comprehensive review of YOLO: From YOLOv1 to YOLOv8 and beyond. arXiv 2023, arXiv:2304.00501. [Google Scholar]
  95. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
  96. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Computer Vision–ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12346, pp. 213–229. [Google Scholar]
  97. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9626–9635. [Google Scholar] [CrossRef]
  98. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  99. Galvao, L.G.; Abbod, M.; Kalganova, T.; Palade, V.; Huda, M.N. Pedestrian and vehicle detection in autonomous vehicle perception systems—A review. Sensors 2021, 21, 7267. [Google Scholar] [CrossRef]
  100. Jocher, G. Ultralytics/Yolov8, GitHub, GitHub Repository. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 15 December 2023).
  101. Jia, S.J.; Zhai, Y.T.; Jia, X.W. Detection of Traffic and Road Condition Based on SVM and PHOW. Appl. Mech. Mater. 2014, 513, 3651–3654. [Google Scholar] [CrossRef]
  102. Gao, D.; Zhou, J.; Xin, L. SVM-based detection of moving vehicles for automatic traffic monitoring. In Proceedings of the ITSC 2001. 2001 IEEE Intelligent Transportation Systems. Proceedings (Cat. No.01TH8585), Oakland, CA, USA, 25–29 August 2001; pp. 745–749. [Google Scholar] [CrossRef]
  103. Zheng, F.; Luo, S.; Song, K.; Yan, C.W.; Wang, M.C. Improved lane line detection algorithm based on Hough transform. Pattern Recognit. Image Anal. 2018, 28, 254–260. [Google Scholar] [CrossRef]
  104. Vamsi, A.M.; Deepalakshmi, P.; Nagaraj, P.; Awasthi, A.; Raj, A. IOT based autonomous inventory management for warehouses. In EAI International Conference on Big Data Innovation for Sustainable Cognitive Computing; Haldorai, A., Ramu, A., Mohanram, S., Onn, C., Eds.; EAI/Springer Innovations in Communication and Computing; Springer: Cham, Switzerland, 2020; pp. 371–376. [Google Scholar]
  105. Nagaraj, P.; Muneeswaran, V.; Sudar, K.M.; Ali, R.S.; Someshwara, A.; Kumar, T.S. Internet of Things based smart hospital saline monitoring system. In Proceedings of the 2021 5th International Conference on Computer, Communication and Signal Processing (ICCCSP), Chennai, India, 24–25 May 2021; pp. 53–58. [Google Scholar] [CrossRef]
  106. Wang, Z.; Zhan, J.; Duan, C.; Guan, X.; Yang, K. Vehicle detection in severe weather based on pseudo-visual search and HOG–LBP feature fusion. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2022, 236, 1607–1618. [Google Scholar] [CrossRef]
  107. Abdullah, M.N.; Ali, Y.H. Vehicles detection system at different weather conditions. Iraqi J. Sci. 2021, 62, 2040–2052. [Google Scholar] [CrossRef]
  108. Wu, B.F.; Juang, J.H. Adaptive vehicle detector approach for complex environments. IEEE Trans. Intell. Transp. Syst. 2012, 13, 817–827. [Google Scholar] [CrossRef]
  109. Padilla, D.A.; Villaverde, J.F.; Magdaraog, J.J.T.; Oconer, A.J.L.; Ranjo, J.P. Vehicle and Weather Detection Using Real Time Image Processing Using Optical Flow and Color Histogram. In Proceedings of the 2019 5th International Conference on Control, Automation and Robotics (ICCAR), Beijing, China, 19–22 April 2019; pp. 880–883. [Google Scholar] [CrossRef]
  110. Tian, Q.; Zhang, L.; Wei, Y.; Zhao, W.; Fei, W. Vehicle Detection and Tracking at Night in Video Surveillance. Int. J. Online Eng. 2013, 9, 60–64. [Google Scholar] [CrossRef]
  111. Kuang, H.; Chen, L.; Chan, L.L.H.; Cheung, R.C.; Yan, H. Feature selection based on tensor decomposition and object proposal for night-time multiclass vehicle detection. IEEE Trans. Syst. Man Cybern. Syst. 2018, 49, 71–80. [Google Scholar] [CrossRef]
  112. Ewecker, L.; Asan, E.; Ohnemus, L.; Saralajew, S. Provident vehicle detection at night for advanced driver assistance systems. Auton. Robot. 2023, 47, 313–335. [Google Scholar] [CrossRef]
  113. Yaghoobi Ershadi, N. Improving vehicle tracking rate and speed estimation in dusty and snowy weather conditions with a vibrating camera. PLoS ONE 2017, 12, e0189145. [Google Scholar] [CrossRef]
  114. Yang, H.; Qu, S. Real-time vehicle detection and counting in complex traffic scenes using background subtraction model with low-rank decomposition. IET Intell. Transp. Syst. 2018, 12, 75–85. [Google Scholar] [CrossRef]
  115. Govardhan, P. Night Time Pedestrian Detection for Advanced Driving Assistance Systems (ADAS) Using Near Infrared Images. Ph.D. Thesis, National Institute of Technology, Rourkela, India, 2014. [Google Scholar]
  116. Sasaki, Y.; Emaru, T.; Ravankar, A.A. SVM based pedestrian detection system for sidewalk snow removing machines. In Proceedings of the 2021 IEEE/SICE International Symposium on System Integration (SII), Iwaki, Japan, 11–14 January 2021; pp. 700–701. [Google Scholar]
  117. Kang, S.; Byun, H.; Lee, S.W. Real-time pedestrian detection using support vector machines. In Pattern Recognition with Support Vector Machines. SVM 2002; Lee, S.W., Verri, A., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2002; Volume 2388, pp. 268–277. [Google Scholar]
  118. Jegham, I.; Khalifa, A.B. Pedestrian detection in poor weather conditions using moving camera. In Proceedings of the 2017 IEEE/ACS 14th International Conference on Computer Systems and Applications (AICCSA), Hammamet, Tunisia, 30 October–3 November 2017; pp. 358–362. [Google Scholar]
  119. Fan, Y. Research of Pedestrian Tracking Based on HOG Feature and Haar Feature. Comput. Sci. 2013, 40, 199–203. [Google Scholar]
  120. Ding, B.; Liu, Z.; Sun, Y. Pedestrian detection in haze environments using dark channel prior and histogram of oriented gradient. In Proceedings of the 2018 Eighth International Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC), Harbin, China, 19–21 July 2018; pp. 1003–1008. [Google Scholar]
  121. Sotelo, M.; Parra, I.; Fernandez, D.; Naranjo, E. Pedestrian detection using SVM and multi-feature combination. In Proceedings of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 17–20 September 2006; pp. 103–108. [Google Scholar] [CrossRef]
  122. Satyawan, A.S.; Fuady, S.; Mitayani, A.; Sari, Y.W. HOG Based Pedestrian Detection System for Autonomous Vehicle Operated in Limited Area. In Proceedings of the 2021 International Conference on Radar, Antenna, Microwave, Electronics, and Telecommunications (ICRAMET), Bandung, Indonesia, 23–24 November 2021; pp. 147–152. [Google Scholar]
  123. Jang, G.; Park, J.; Kim, M. Cascade-Adaboost for Pedestrian Detection Using HOG and Combined Features. In Advances in Computer Science and Ubiquitous Computing; Springer: Singapore, 2016; pp. 430–435. [Google Scholar]
  124. Zhang, X.; Zhu, X. Autonomous path tracking control of intelligent electric vehicles based on lane detection and optimal preview method. Expert Syst. Appl. 2019, 121, 38–48. [Google Scholar] [CrossRef]
  125. Gern, A.; Moebus, R.; Franke, U. Vision-based lane recognition under adverse weather conditions using optical flow. In Proceedings of the Intelligent Vehicle Symposium 2002, Versailles, France, 17–21 June 2002; Volume 2, pp. 652–657. [Google Scholar]
  126. Tran, T.T.; Son, J.H.; Uk, B.J.; Lee, J.H.; Cho, H.M. An adaptive method for detecting lane boundary in night scene. In Advanced Intelligent Computing Theories and Applications. With Aspects of Artificial Intelligence; Huang, D.S., Zhang, X., Reyes García, C.A., Zhang, L., Eds.; ICIC 2010. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6216, pp. 301–308. [Google Scholar]
  127. Chen, T.Y.; Chen, C.H.; Luo, G.M.; Hu, W.C.; Chern, J.C. Vehicle detection in nighttime environment by locating road lane and taillights. In Proceedings of the 2015 International Conference on Intelligent Information Hiding and Multimedia Signal Processing (IIH-MSP), Adelaide, SA, Australia, 23–25 September 2015; pp. 60–63. [Google Scholar]
  128. Guo, J.; Wei, Z.; Miao, D. Lane detection method based on improved RANSAC algorithm. In Proceedings of the 2015 IEEE Twelfth International Symposium on Autonomous Decentralized Systems, Taichung, Taiwan, 25–27 March 2015; pp. 285–288. [Google Scholar]
  129. Antonio, J.A.; Romero, M. Pedestrians’ Detection Methods in Video Images: A Literature Review. In Proceedings of the 2018 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 12–14 December 2018; pp. 354–360. [Google Scholar]
  130. Papadimitriou, O.; Kanavos, A.; Mylonas, P.; Maragoudakis, M. Advancing Weather Image Classification using Deep Convolutional Neural Networks. In Proceedings of the 2023 18th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP) 18th International Workshop on Semantic and Social Media Adaptation & Personalization (SMAP 2023), Limassol, Cyprus, 25–26 September 2023; pp. 1–6. [Google Scholar]
  131. Alhammadi, S.A.; Alhameli, S.A.; Almaazmi, F.A.; Almazrouei, B.H.; Almessabi, H.A.; Abu-Kheil, Y. Thermal-Based Vehicle Detection System using Deep Transfer Learning under Extreme Weather Conditions. In Proceedings of the 2022 8th International Conference on Information Technology Trends (ITT), Dubai, United Arab Emirates, 25–26 May 2022; pp. 119–123. [Google Scholar]
  132. Arora, N.; Kumar, Y.; Karkra, R.; Kumar, M. Automatic vehicle detection system in different environment conditions using fast R-CNN. Multimed. Tools Appl. 2022, 81, 18715–18735. [Google Scholar] [CrossRef]
  133. Ghosh, R. On-road vehicle detection in varying weather conditions using faster R-CNN with several region proposal networks. Multimed. Tools Appl. 2021, 80, 25985–25999. [Google Scholar] [CrossRef]
  134. Cao, J.; Song, C.; Song, S.; Peng, S.; Wang, D.; Shao, Y.; Xiao, F. Front vehicle detection algorithm for smart car based on improved SSD model. Sensors 2020, 20, 4646. [Google Scholar] [CrossRef] [PubMed]
  135. Li, W. Vehicle detection in foggy weather based on an enhanced YOLO method. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2022; Volume 2284, p. 012015. [Google Scholar]
  136. Ghosh, R. A modified yolo model for on-road vehicle detection in varying weather conditions. In Intelligent Computing and Communication Systems; Springer: Singapore, 2021; pp. 45–54. [Google Scholar]
  137. Miao, Y.; Liu, F.; Hou, T.; Liu, L.; Liu, Y. A nighttime vehicle detection method based on YOLO v3. In Proceedings of the 2020 Chinese Automation Congress (CAC), Shanghai, China, 6–8 November 2020; pp. 6617–6621. [Google Scholar]
  138. Humayun, M.; Ashfaq, F.; Jhanjhi, N.Z.; Alsadun, M.K. Traffic management: Multi-scale vehicle detection in varying weather conditions using yolov4 and spatial pyramid pooling network. Electronics 2022, 11, 2748. [Google Scholar] [CrossRef]
  139. Wang, R.; Zhao, H.; Xu, Z.; Ding, Y.; Li, G.; Zhang, Y.; Li, H. Real-time vehicle target detection in inclement weather conditions based on YOLOv4. Front. Neurorobot. 2023, 17, 1058723. [Google Scholar] [CrossRef]
  140. Yao, J.; Fan, X.; Li, B.; Qin, W. Adverse Weather Target Detection Algorithm Based on Adaptive Color Levels and Improved YOLOv5. Sensors 2022, 22, 8577. [Google Scholar] [CrossRef] [PubMed]
  141. Razzok, M.; Badri, A.; Mourabit, I.E.; Ruichek, Y.; Sahel, A. Pedestrian detection under weather conditions using conditional generative adversarial network. Int. J. Artif. Intell. 2023, 12, 1557–1568. [Google Scholar] [CrossRef]
  142. Pawełczyk, M.; Wojtyra, M. Real world object detection dataset for quadcopter unmanned aerial vehicle detection. IEEE Access 2020, 8, 174394–174409. [Google Scholar] [CrossRef]
  143. Yahiaoui, M.; Rashed, H.; Mariotti, L.; Sistu, G.; Clancy, I.; Yahiaoui, L.; Kumar, V.R.; Yogamani, S. Fisheyemodnet: Moving object detection on surround-view cameras for autonomous driving. arXiv 2019, arXiv:1908.11789. [Google Scholar]
  144. Ragesh, N.; Rajesh, R. Pedestrian detection in automotive safety: Understanding state-of-the-art. IEEE Access 2019, 7, 47864–47890. [Google Scholar] [CrossRef]
  145. Kim, J.H.; Hong, H.G.; Park, K.R. Convolutional neural network-based human detection in nighttime images using visible light camera sensors. Sensors 2017, 17, 1065. [Google Scholar] [CrossRef] [PubMed]
  146. Hsia, C.H.; Peng, H.C.; Chan, H.T. All-Weather Pedestrian Detection Based on Double-Stream Multispectral Network. Electronics 2023, 12, 2312. [Google Scholar] [CrossRef]
  147. Lai, K.; Zhao, J.; Liu, D.; Huang, X.; Wang, L. Research on pedestrian detection using optimized mask R-CNN algorithm in low-light road environment. In Journal of Physics: Conference Series; IOP Publishing: Bristol, UK, 2021; Volume 1777, p. 012057. [Google Scholar]
  148. Tumas, P.; Nowosielski, A.; Serackis, A. Pedestrian detection in severe weather conditions. IEEE Access 2020, 8, 62775–62784. [Google Scholar] [CrossRef]
  149. Shakeri, A.; Moshiri, B.; Garakani, H.G. Pedestrian detection using image fusion and stereo vision in autonomous vehicles. In Proceedings of the 2018 9th International Symposium on Telecommunications (IST), Tehran, Iran, 17–19 December 2018; pp. 592–596. [Google Scholar] [CrossRef]
  150. Galarza-Bravo, M.A.; Flores-Calero, M.J. Pedestrian detection at night based on faster R-CNN and far infrared images. In Intelligent Robotics and Applications; Chen, Z., Mendes, A., Yan, Y., Chen, S., Eds.; ICIRA 2018. Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; Volume 10985, pp. 335–345. [Google Scholar]
  151. Xu, Y.; Cao, X.; Qiao, H. Optical camera based pedestrian detection in rainy or snowy weather. In Fuzzy Systems and Knowledge Discovery; Wang, L., Jiao, L., Shi, G., Li, X., Liu, J., Eds.; FSKD 2006. Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4223, pp. 1182–1191. [Google Scholar]
  152. Montenegro, B.; Flores-Calero, M. Pedestrian detection at daytime and nighttime conditions based on YOLO-v5. Ingenius. Rev. Cienc. Tecnol. 2022, 85–95. [Google Scholar] [CrossRef]
  153. Zaman, M.; Saha, S.; Zohrabi, N.; Abdelwahed, S. Deep Learning Approaches for Vehicle and Pedestrian Detection in Adverse Weather. In Proceedings of the 2023 IEEE Transportation Electrification Conference & Expo (ITEC), Detroit, MI, USA, 21–23 June 2023; pp. 1–6. [Google Scholar]
  154. Son, J.; Yoo, H.; Kim, S.; Sohn, K. Real-time illumination invariant lane detection for lane departure warning system. Expert Syst. Appl. 2015, 42, 1816–1824. [Google Scholar] [CrossRef]
  155. Kortli, Y.; Marzougui, M.; Atri, M. Efficient implementation of a real-time lane departure warning system. In Proceedings of the 2016 International Image Processing, Applications and Systems (IPAS), Hammamet, Tunisia, 5–7 November 2016; pp. 1–6. [Google Scholar]
  156. Sun, T.Y.; Tsai, S.J.; Chan, V. HSI color model based lane-marking detection. In Proceedings of the 2006 IEEE Intelligent Transportation Systems Conference, Toronto, ON, Canada, 7–12 September 2006; pp. 1168–1172. [Google Scholar]
  157. Chiu, K.Y.; Lin, S.F. Lane detection using color-based segmentation. In Proceedings of the IEEE Proceedings. Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; pp. 706–711. [Google Scholar]
  158. Shin, S.; Shim, I.; Kweon, I.S. Combinatorial approach for lane detection using image and LIDAR reflectance. In Proceedings of the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Goyangi, Republic of Korea, 28–30 October 2015; pp. 485–487. [Google Scholar]
  159. Rose, C.; Britt, J.; Allen, J.; Bevly, D. An integrated vehicle navigation system utilizing lane-detection and lateral position estimation systems in difficult environments for GPS. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2615–2629. [Google Scholar] [CrossRef]
  160. Hsieh, J.W.; Chen, L.C.; Chen, D.Y. Symmetrical SURF and its applications to vehicle detection and vehicle make and model recognition. IEEE Trans. Intell. Transp. Syst. 2014, 15, 6–20. [Google Scholar] [CrossRef]
  161. Yusuf, M.M.; Karim, T.; Saif, A.S. A robust method for lane detection under adverse weather and illumination conditions using convolutional neural network. In Proceedings of the ICCA 2020: International Conference on Computing Advancements, Dhaka, Bangladesh, 10–12 January 2020; pp. 1–8. [Google Scholar]
  162. Nguyen, V.; Kim, H.; Jun, S.; Boo, K. A study on real-time detection method of lane and vehicle for lane change assistant system using vision system on highway. Eng. Sci. Technol. Int. J. 2018, 21, 822–833. [Google Scholar] [CrossRef]
  163. Khan, M.N.; Ahmed, M.M. Weather and surface condition detection based on road-side webcams: Application of pre-trained convolutional neural network. Int. J. Transp. Sci. Technol. 2022, 11, 468–483. [Google Scholar] [CrossRef]
  164. Ding, Y.; Xu, Z.; Zhang, Y.; Sun, K. Fast lane detection based on bird’s eye view and improved random sample consensus algorithm. Multimed. Tools Appl. 2017, 76, 22979–22998. [Google Scholar] [CrossRef]
  165. Ab Ghani, H.; Daud, A.M.; Besar, R.; Sani, Z.M.; Kamaruddin, M.N.; Syahali, S. Lane Detection Using Deep Learning for Rainy Conditions. In Proceedings of the 2023 9th International Conference on Computer and Communication Engineering (ICCCE), Kuala Lumpur, Malaysia, 15–16 August 2023; pp. 373–376. [Google Scholar]
  166. Raj, N.; Dutta, K.K.; Kamath, A. Lane Prediction by Autonomous Vehicle in Highway Traffic using Artificial Neural Networks. In Proceedings of the 2021 Fourth International Conference on Electrical, Computer and Communication Technologies (ICECCT), Erode, India, 15–17 September 2021; pp. 1–6. [Google Scholar]
  167. Krishnaveni, P.; Sutha, J. Novel deep learning framework for broadcasting abnormal events obtained from surveillance applications. J. Ambient. Intell. Humaniz. Comput. 2020, 13, 4123. [Google Scholar] [CrossRef]
  168. Ouyang, W.; Wang, X. A discriminative deep model for pedestrian detection with occlusion handling. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3258–3265. [Google Scholar]
  169. Haris, M.; Hou, J.; Wang, X. Lane lines detection under complex environment by fusion of detection and prediction models. Transp. Res. Rec. 2022, 2676, 342–359. [Google Scholar] [CrossRef]
Figure 1. The global self-driving cars market size 2022–2032 in USD Million [3].
Figure 1. The global self-driving cars market size 2022–2032 in USD Million [3].
Algorithms 17 00103 g001
Figure 2. Object identification strategies.
Figure 2. Object identification strategies.
Algorithms 17 00103 g002
Figure 3. Paper organization, where OD = object detection.
Figure 3. Paper organization, where OD = object detection.
Algorithms 17 00103 g003
Figure 4. SAE automation levels for AVs.
Figure 4. SAE automation levels for AVs.
Algorithms 17 00103 g004
Figure 5. An architecture of AVs.
Figure 5. An architecture of AVs.
Algorithms 17 00103 g005
Figure 6. Weather impacts on AVs.
Figure 6. Weather impacts on AVs.
Algorithms 17 00103 g006
Figure 7. The basic framework for object detection systems.
Figure 7. The basic framework for object detection systems.
Algorithms 17 00103 g007
Figure 8. Performance graph of traditional and DL approaches.
Figure 8. Performance graph of traditional and DL approaches.
Algorithms 17 00103 g008
Figure 9. Primary structure of feature extraction for traditional and DL approaches.
Figure 9. Primary structure of feature extraction for traditional and DL approaches.
Algorithms 17 00103 g009
Figure 10. Traditional and DL approaches.
Figure 10. Traditional and DL approaches.
Algorithms 17 00103 g010
Figure 11. The architecture of Faster R-CNN [90].
Figure 11. The architecture of Faster R-CNN [90].
Algorithms 17 00103 g011
Figure 12. Structure of YOLOv8.
Figure 12. Structure of YOLOv8.
Algorithms 17 00103 g012
Figure 13. Overall percentage distribution of papers.
Figure 13. Overall percentage distribution of papers.
Algorithms 17 00103 g013
Figure 14. Distribution of papers for traditional and DL approaches.
Figure 14. Distribution of papers for traditional and DL approaches.
Algorithms 17 00103 g014
Figure 15. Paper distribution for one-stage and two-stage detectors.
Figure 15. Paper distribution for one-stage and two-stage detectors.
Algorithms 17 00103 g015
Table 1. Comparison of sensors [35].
Table 1. Comparison of sensors [35].
SensorsAdvantagesDisadvantages
LiDARHigh resolution Long range Wide FOVSensitive to weather Expensive
RadarLong range Detection of velocity Suitable for all types of weatherLow resolution Very sensitive
CameraHigh resolution Detection of colorsSensitive to weather Sensitive to lighting
UltrasonicNon-InvasiveReal-time feedbackLow resolution Expensive
Table 2. Publicly available AVs datasets.
Table 2. Publicly available AVs datasets.
Ref.YearDatasetLiDARRadarCameraAdverse Weather
[37]2020LIBRE-Rain, fog, Daytime
[51]2013Kitti--
[54]2021ACDC--Rain, fog, snow
[55]2020A2D2-Rain
[56]2020Waymo-Rain, night
[57]2020NuScenesRain, night
[58]2020BDD100K--Daytime, night
[59]2020ApolloScape--Rain, Night
[60]2016SYNTHIA---Snow
[61]2016NCLT-Snow
[62]2021CADCD-Snow
[64]2020End of the earth-Snow
[65]2020nuImages---
[66]2019Argoverse--
[67]2019Astyx-
[68]2020DENSERain, snow, fog, night
[69]2018Foggy Cityscape---Fog/haze
[70]2020Berkley DeepDrive--Rain, fog, snow, night
[71]2017Mapillary--Rain, fog, snow, night
[72]2019EuroCity--Rain, fog, snow, night
[73]2017Oxford RobotCarRain, snow, night
[74]2020A* 3D-Rain, night
[75]20214Seasons--Rain, night
[76]2018WildDash--Rain, fog, snow, night
[77]2018KAIST multispectral--Daytime, night
[78]2021RadiateRain, fog, snow, night
[79]2020EU-Snow, night
[80]2022BoreasRain, snow, night
[81]2020DAWN---Rain, snow, fog, sandstorm
[82]2021PVDN--night
Table 3. Vehicle detection previous work with traditional approaches.
Table 3. Vehicle detection previous work with traditional approaches.
Ref.YearTechniqueWeather ClassesAccuracy
[101]2014SVM+PHOWRain, Fog, Snow, Low light82.5%
[106]2022HOG–LBPseveral weather96.4%
[108]2012histogram GDVMRain, Night92.4%
[111]2019tensor decompositionNight95.82%
[113]2017Particle FilteringDust, Snow94.78%
Table 5. Pedestrian detection previous work with DL.
Table 5. Pedestrian detection previous work with DL.
Ref.YearTechniqueWeather ClassesAccuracy
[141]2023YOLOv3several weathers74.38% (AP)
[145]2017CNNnight time98.3%
[147]2021Mask R-CNNlowlight85.05%
[152]2022YOLOv5Day, Night96.6%
[153]2023YOLOv7, CNN, Faster R-CNN, SSD, HOGRain, Fog, Snow, Sand0.73, 0.883, 0.71, 0.657, 0.70
Table 6. Practical implications of DL approaches.
Table 6. Practical implications of DL approaches.
Ref.ApproachPractical ImplicationsKPIsChallengesAddressing Challenges
[145]CNN-based Human Detection at NighttimeEnhances pedestrian detection in low-light conditions, improving safety and navigation
for AVs.
Detection Accuracy, False Negative Rate, Computational EfficiencyLimited visibility, background noise interferenceUtilizes CNNs trained on nighttime images to recognize pedestrian features, potentially reducing false negatives.
[147]Optimized Mask R-CNN for Low-Light Pedestrian DetectionTailors the Mask R-CNN algorithm for low-light environments, maintaining high detection rates.Detection Accuracy, Precision, Recall, Real-time ProcessingOptimization for low-light conditions, computational complexityAdjusts model parameters, uses data augmentation, and employs hardware acceleration for real-time processing.
[152]YOLO-v5 for Pedestrian Detection in Daytime and NighttimeDemonstrates the versatility of YOLO-v5 across different lighting conditions, ensuring consistent performance.Detection Speed, Accuracy, RobustnessBalancing speed with accuracy in
varying lighting
Fine-tunes the YOLO-v5 model on diverse datasets, ensuring generalization across different lighting scenarios.
[153]DL Approaches for Adverse Weather DetectionImproves detection reliability in adverse weather, crucial for AV safety and functionality.Detection Accuracy, Robustness, System ReliabilityDegradation of performance due to weather conditionsUses DL models that learn to recognize patterns despite weather distortions and integrate multi-sensor data for enhanced detection.
Table 7. Comparison of traditional and DL approaches.
Table 7. Comparison of traditional and DL approaches.
Ref.ApproachesTechniqueSuperiorityLimitationsAdditional Considerations
[154,155]Color-based Lane DetectionTraditionalSimple to implement, low computational costLimited in adverse weather, reliance on good weather conditions for high accuracyResearch into adaptive color models or fusion with other sensor data could improve performance in challenging conditions.
[158,159]LiDAR IntegrationTraditionalReduces reliance on visual data, robust in various weatherHigh cost, sensitivity to unfavorable weather conditions.Cost reduction and miniaturization of LiDAR sensors could make this technology more accessible for widespread use.
[160]HSV Color SpaceTraditionalWorks well in daylight or with white light; uses color correspondenceRequires global thresholding, less effective in tunnel environments with distinct color illumination.Enhancing the color model with machine learning could improve its adaptability to different lighting conditions.
[161]SafeDrive TechniqueTraditionalUtilizes historical data and vehicle location for lane detection in low-light areasMay not generalize well to all environments, relies on available historical data.Incorporating real-time weather data and vehicle dynamics could improve the technique’s robustness.
[163]CNN-based Real-time Road Condition DetectionDLHigh accuracy with transfer learning automates real-time condition recognitionRequires pre-trained models, may not adapt quickly to new conditionsContinuous learning and model updating could help maintain high accuracy as conditions change.
[165]Lane Marker Detection in Rainy WeatherDLPrioritizes feature selection to counteract rain-induced blurrinessMay require extensive training data, complex model architectureUsing smaller, more specialized models could reduce computational demands and improve real-time performance.
[141]Generative Adversarial Networks (GANs)DLCan generate realistic data for training, improve feature extractionRequires significant computational resources, may struggle with certain types of adverse weather.Research into GANs for adverse weather simulation could provide more robust training data for DL models.
[152]Multispectral ImagingDLCombines different imaging modalities for improved detectionRequires specialized hardware; may be complex to integrateFurther development of multispectral imaging techniques could lead to more reliable detection systems.
[163]Transfer LearningDLAllows models to adapt to new tasks with fewer dataMay not perform as well as task-specific models if the transfer is not well-aligned.Fine-tuning transfer learning models for specific AV tasks could enhance their effectiveness.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tahir, N.U.A.; Zhang, Z.; Asim, M.; Chen, J.; ELAffendi, M. Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches. Algorithms 2024, 17, 103. https://doi.org/10.3390/a17030103

AMA Style

Tahir NUA, Zhang Z, Asim M, Chen J, ELAffendi M. Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches. Algorithms. 2024; 17(3):103. https://doi.org/10.3390/a17030103

Chicago/Turabian Style

Tahir, Noor Ul Ain, Zuping Zhang, Muhammad Asim, Junhong Chen, and Mohammed ELAffendi. 2024. "Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches" Algorithms 17, no. 3: 103. https://doi.org/10.3390/a17030103

APA Style

Tahir, N. U. A., Zhang, Z., Asim, M., Chen, J., & ELAffendi, M. (2024). Object Detection in Autonomous Vehicles under Adverse Weather: A Review of Traditional and Deep Learning Approaches. Algorithms, 17(3), 103. https://doi.org/10.3390/a17030103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop