In this section, we illustrate the details of our experimental setup and present the experimental results.
4.1. Experimental Setup
Training Settings: All models were trained for 20 epochs using the Adam optimizer with an initial learning rate of 0.02. The learning rate was scheduled to decay by a factor of 0.1 every five epochs. A batch size of 2 was used for training, and weight decay was set to 0.0005 to prevent overfitting. These hyperparameters were selected based on preliminary experiments to balance training efficiency and model performance.
Dataset: Unfortunately, in the field of autonomous vehicle detection, there is no comprehensive and open training dataset suitable for 3D adversarial attacks. Therefore, we chose a rendering-based generative dataset, which offers rich scenes and differentiable rendering. The Carla dataset is the basic dataset in this field, and to be consistent with the domain, the experimental dataset used in this study is the Carla dataset [
75]. Therefore, we use the Carla simulator, which is an open simulation platform to simulate the process of a vehicle driving in the city, to generate remote simulation pictures under different perspectives, distances, and environments during the driving process. In total, 15,000 pictures are generated, which are exported from the Carla simulator as the Carla dataset standard, as shown in
Figure 5.
Metric: To measure the effectiveness of the adversarial attack, we use two metrics: average precision (AP) and a custom metric called attack efficiency (AE). In this experiment, the AP is defined as the average precision for the class “car,” which is calculated using Equation (
17). The attack efficiency (AE) is calculated using Equation (
18):
where
is the smoothed precision-recall curve,
is the difference in average precision before and after the attack, and
is the proportion of faces modified during the attack. A reduction in AP indicates a decrease in the model’s ability to correctly detect and classify objects, thereby demonstrating the effectiveness of the adversarial attack. However, it is crucial to balance this reduction to avoid overly degrading the model’s performance, which could render the attack impractical or easily noticeable.
Experimental Schema: First, we train six detectors from the YOLOv3 and YOLOv5 series on the Carla dataset. We use three selection strategies (manual expert selection, random selection, and MARS) to identify different local regions. Full-body attacks are used as the baseline control group to compare and analyze the attack effects on these regions. The training was conducted under identical conditions for all models to ensure a fair comparison.
The experiments are divided into three parts:
- 1
Attack efficiency: the main objective is to demonstrate the superiority of MARS;
- 2
Transferability: this part assesses the transferability of the model by comparing the impacts of different texture optimization methods on the attack performance;
- 3
Parameter sensitivity analysis: this section presents an investigation of the significance and sensitivity of core parameters through variations and ablation experiments.
4.2. Local Adversarial Attacks with Different Region Selection Strategies
We use different selection strategies to identify various local regions and conduct local attacks using both fixed textures and optimized textures. Then, we compare their performance across different detection networks for the same dataset and training network conditions. YOLOv3 is selected as the training network, whereas YOLOv3, YOLOv5s, YOLOv5x, YOLOv5m, YOLOv5n, and YOLOv5l are selected as the detection networks. All detection networks are trained with identical settings using the Carla dataset.
Figure 6 and
Figure 7 show the experimental results.
The first row, which is labeled “fixed texture,” uses a fixed camouflage for local attacks with different region selection strategies. The first column contains sample images of the original unmodified targets. We aim to mitigate the impact of the perturbation on the detector to highlight the effects of different region selection strategies on the attack effectiveness. The second row, which is labeled “optimized texture,” employs optimized patches for local attacks with different region selection strategies. The first column contains sample images of full-body attacks, which serves as the baseline. We aim to demonstrate the superiority of local attacks using the MARS strategy. The comparison strategies for region selection are as follows: The second and third columns show fixed regions that are manually selected based on expert guidance and highlight the edge and center regions, respectively. The fourth column shows the randomly selected regions. The fifth column shows critical decision regions that are selected by MARS. We evaluate the AP for these attack scenarios, and
Table 1 shows the results.
Performance Variations Between YOLOv3 and YOLOv5: The differences in performance between YOLOv3 and YOLOv5 can be attributed to variations in their architectural complexity and depth. YOLOv5 models are generally deeper and incorporate more advanced features such as enhanced backbone networks and better anchor box strategies, which can influence their susceptibility to adversarial attacks. Specifically, YOLOv5l, being the largest variant, exhibits lower attack effectiveness due to its increased robustness and capacity to generalize from complex patterns, whereas smaller variants like YOLOv5s are more vulnerable due to their reduced complexity. These architectural differences explain why the attack performance varies across different YOLOv5 variants compared to YOLOv3.
Table 1 reveals significant differences between the Fixed(center) and Fixed(edge) strategies. Edge regions have lower visibility from different viewpoints than central regions, which leads to varied contributions during attacks. The AP after the Fixed(center) attacks is consistently lower than that after the Fixed(edge) attacks, which indicates that object edges are not equivalent to decision boundaries in neural networks. Random regions show inconsistent performance. With a fixed texture, the AP decreases to 0.967, which is better than the values of the fixed regions (0.980 and 0.978). However, with the optimized texture, the AP only decreases to 0.726, which is less effective than the values of the fixed regions (0.719 and 0.184). Thus, although random regions cover a broader area, their fragmented nature hampers the effective optimization and deception of neural networks. Additionally, YOLOv3 and YOLOv5 models respond differently to various attack strategies due to their distinct network architectures. YOLOv3 tends to be more sensitive to centralized attacks, whereas YOLOv5 models, especially larger variants, exhibit varied sensitivity based on their depth and feature extraction capabilities. Our proposed MARS strategy achieves AP decreases of 0.618–0.188 under optimized conditions, so AP surpasses all other strategies and closely approximates the full-body attack results.
Compared with the baseline, the attack efficiency (AE) for both fixed texture and optimized texture attacks is shown below, and the superior performance of the MARS strategy is highlighted.
Table 2 and
Table 3 show that the AE of our proposed MARS method consistently outperforms the baseline and other local region selection strategies. The variations in AE across different YOLOv5 variants indicate that deeper and more complex networks like YOLOv5l are more resilient to adversarial attacks, requiring more concentrated and effective perturbations to achieve significant AE. In YOLOv3, MARS achieved an AE of 2.615, which is more than double the baseline value (0.911). In the YOLOv5 series, MARS achieved at least 0.608, which significantly outperformed the control group (0.532). The average AE for MARS was 1.7235, i.e., a 0.986 (134%) improvement over the baseline. These experiments demonstrate that, compared with other methods, our method effectively balances the coverage and region completeness and significantly improves stability and transferability. This result confirms the feasibility of using local adversarial attacks with the Maximum Aggregated Region Sparseness (MARS) strategy on 3D objects to attack detectors.
A significant reduction in AP signifies that the adversarial attack effectively diminishes the model’s capability to accurately detect and classify objects. In practical terms, this could lead to scenarios where critical objects, such as vehicles in autonomous driving systems, go undetected or are misclassified, potentially causing safety hazards. However, the extent of AP reduction must be carefully managed to avoid rendering the system non-functional, which could be impractical or easily noticed by human operators.
However, due to differences in network depth and width, the robustness to adversarial examples varies. In particular, the YOLOv5l network consistently shows lower attack effectiveness. This variation highlights the need for tailored adversarial strategies that consider the architectural nuances of different detection models. Addressing the network complexity and enhancing adversarial robustness will be a focus of future work.
4.3. Attack Transferability
This section of the experiment has two main objectives: to determine whether MARS can be combined with different texture modification methods for attacks and to assess the attack effectiveness of models that are trained on YOLO detectors and applied to other detectors.
We select the full region, Fixed(center) region, and MARS region, which perform well and are representative of previous experiments, as variables. Using FCA [
13] and DAS [
60] for texture optimization, we compare the resulting adversarial outcomes across different detection networks. To ensure the independence of the results, the detection networks are selected outside the YOLO series and include Mask R-CNN, Cascade R-CNN, Faster R-CNN, SSD, and RetinaNet.
As shown in
Table 4, FCA and DAS exhibit varying performance across different region selection strategies. Full attacks consistently demonstrate stable results, where FCA and DAS achieve average AP decreases of 0.631 and 0.627, respectively. For the Fixed(center) region, Fixed(center) + FCA achieves an average AP decrease of 0.356, whereas Fixed(center) + DAS achieves an average AP decrease of 0.525. For the MARS region, MARS + FCA achieves an average AP decrease of 0.488, and MARS + DAS achieves an average AP decrease of 0.662. The MARS attack consistently outperforms other local attack methods, particularly with DAS, where it even outperforms the full attack. Thus, the critical decision regions identified by MARS align well with the decision boundaries of the model and demonstrate strong transferability with texture modification methods. Furthermore, the enhanced performance of MARS across different detection networks suggests that the aggregation and sparsity constraints effectively target universally critical regions, making the adversarial perturbations more versatile and robust against various model architectures. This adaptability is crucial for real-world applications where multiple detection systems may be in use.
Considering the attack effectiveness across different detection networks, the MARS attack outperforms other local attack methods on all networks. Specifically, MARS + DAS exceeds the performance of other local region selection strategies and surpasses the full attack in all networks except Cascade R-CNN. This superior performance is likely due to the MARS strategy’s ability to focus perturbations on regions that are consistently influential across different models, enhancing both the attack’s effectiveness and its transferability.
4.4. Attack Performance for Different Factors
In this section, we adjust various parameter coefficients to explore the significance of each loss parameter in this paper.
The training model is a YOLOv3 network trained for one epoch on the Carla dataset. Since we must only compare the effects of different coefficients, the selected detectors are also the YOLOv3 and YOLOv5s networks trained for one epoch on the Carla dataset.
First, we conduct experiments with different combinations of loss parameters. The adversarial loss, which ensures the fundamental effectiveness of the attack, is not adjusted. During training, the loss parameters are set as follows: (all), (single detection loss), (with aggregation regularization), and (with minimization regularization).
A comparison of
Figure 8 shows that only pursuing aggregation during training tends to result in a uniform distribution of mask weights. The resulting regions lack constraints on the mask area and mask weight range and consequently do not possess decisive characteristics.
Table 5 shows that with only the adversarial loss (
) set, the AP decreases to 0.686 and 0.704. When only aggregation regularization (
) is used, the AP only decreases to 0.77 and 0.655, which indicates poor attack performance. Conversely, when only sparseness regularization (
) is used, the AP decreases to 0.213 and 0.482, which significantly enhances the attack performance. This demonstrates that sparsity regularization is crucial for concentrating adversarial perturbations in critical regions, thereby increasing the attack’s effectiveness while maintaining stealthiness. The combined loss settings of
and
, which are ultimately adopted in this paper, achieve the best overall performance, identify universal decision regions, and successfully execute attacks.
In the second part of the experiment, we focus on adjusting the pre-parameters for the aggregation loss. As shown in
Table 6, adjustments to parameter
reveal that increasing the parameter gradually improves the attack performance. However, after reaching a certain balance, the efficiency begins to decline. The mid-range values exhibit relatively stable results across all networks. This indicates that there is an optimal range for the aggregation coefficient where the aggregation of adversarial regions is maximized without causing over-concentration that could potentially dilute the attack’s effectiveness. The analysis indicates that the
coefficient controls the degree of aggregation in local regions. Increasing the coefficient yields more complete regions, which provides more possibilities for the optimization process. This behavior enables the generation of various continuous patterns that can deceive deep neural networks, which significantly impacts the attack effectiveness. Coefficient
controls the network attack effect; to better enhance the attack efficiency, it is essential to consider the balance among the loss coefficients. This experiment briefly explores the impact of the parameter coefficients on the experimental results. In future research, we will more precisely define the significance of coefficient
and delve deeper into the relationships among various coefficients.
In this part of the experiment, we keep other coefficients constant while adjusting the pre-parameter for the sparsity coefficient to modify the number of masks generated during optimization. This process enables us to compare the impact of the mask size on the attack efficiency.
Table 7 shows the AP results. Since the variable here is the number of masks, we switch the metric to AE for easier comparison in
Table 8.
As shown in
Table 7, regarding the attack effectiveness, the number of masks has minimal effect when the count exceeds 2000. However, significant precision changes occur when the mask count fluctuates within the 0–2000 range. This result suggests that for this physical target, the optimal number of core critical regions detected by deep neural networks is less than 2000. This finding supports the main premise of this study: traditional adversarial attacks often optimize non-critical regions, which wastes computational resources. Our proposed method effectively identifies a sufficient number of local critical regions, reduces computational costs, enhances the visual effects, and achieves excellent attack performance across different detection networks under various mask counts.
As shown in
Table 8, reducing the number of masks leads to a noticeable decrease in precision impact but a slight increase in attack efficiency. We observe that reducing the number of masks consistently increases the attack efficiency, which further emphasizes the importance of studying critical decision regions. When the mask count decreases during training, the computational speed improves, which is negligible in small-scale training but highly significant in large-scale large-model training. The aim of our study was to ensure rapid training and excellent visual effects while achieving significant attack effectiveness. The improvement in attack efficiency positively reflects this goal. Overall, our method consistently achieves robust adversarial results across different networks under most mask count settings, which demonstrates its broad applicability.
Ethical Considerations: The development and deployment of adversarial attacks pose significant ethical concerns, particularly regarding their potential misuse in critical systems such as autonomous driving, surveillance, and security infrastructures. These attacks can undermine the reliability and safety of systems that people depend on daily, leading to severe societal and economic consequences. To mitigate these risks, it is essential to implement robust defense mechanisms, promote responsible research practices, and establish regulatory frameworks that govern the use of adversarial technologies. Additionally, raising awareness about the vulnerabilities of machine learning models can encourage the development of more resilient systems and ethical guidelines for deploying such technologies.
Societal Risks: Adversarial attacks against critical systems such as autonomous vehicles, security surveillance, and infrastructure monitoring pose significant societal risks. These attacks can lead to accidents, breaches of privacy, and disruptions of essential services, potentially causing widespread harm. The ability to deceive detection models undermines trust in automated systems and can have far-reaching implications for public safety and security. To address these risks, it is imperative to develop robust defense mechanisms, enforce strict ethical guidelines, and promote collaboration between researchers, policymakers, and industry stakeholders to ensure that advancements in adversarial attacks do not compromise societal well-being.