Figure 1.
Adversarial example that misleads an image classifier to predict the image as a cat.
Figure 1.
Adversarial example that misleads an image classifier to predict the image as a cat.
Figure 2.
Denoised autoencoder for preprocessing of an adversarial example to create a clean/denoised sample. The solid line is the process with the autoencoder, and the dashed line is the process without the autoencoder.
Figure 2.
Denoised autoencoder for preprocessing of an adversarial example to create a clean/denoised sample. The solid line is the process with the autoencoder, and the dashed line is the process without the autoencoder.
Figure 3.
Randomized smoothing method, where the most common predictions are picked as the output. In this example, four noises are generated by the noise generator.
Figure 3.
Randomized smoothing method, where the most common predictions are picked as the output. In this example, four noises are generated by the noise generator.
Figure 4.
Adversarial example detection technique where the detected samples are thrown away.
Figure 4.
Adversarial example detection technique where the detected samples are thrown away.
Figure 5.
An example of S-ReLU with a max value of 2.
Figure 5.
An example of S-ReLU with a max value of 2.
Figure 6.
Examples of the MNIST dataset.
Figure 6.
Examples of the MNIST dataset.
Figure 7.
Examples of the CIFAR10 dataset.
Figure 7.
Examples of the CIFAR10 dataset.
Figure 8.
Examples of the CIFAR100 dataset.
Figure 8.
Examples of the CIFAR100 dataset.
Figure 9.
Examples of the TinyImagenet dataset.
Figure 9.
Examples of the TinyImagenet dataset.
Figure 10.
Architecture of our approach with an added layer (in red) with D-ReLU before the output layer.
Figure 10.
Architecture of our approach with an added layer (in red) with D-ReLU before the output layer.
Figure 11.
Accuracy of two types of networks on clean MNIST and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 11.
Accuracy of two types of networks on clean MNIST and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 12.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 12.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 13.
Accuracy of several types of CNNs on clean CIFAR10 and adversarial examples when adding a convolutional layer with a D-ReLU function after the input layer.
Figure 13.
Accuracy of several types of CNNs on clean CIFAR10 and adversarial examples when adding a convolutional layer with a D-ReLU function after the input layer.
Figure 14.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 14.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 15.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 15.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples when adding a dense layer with a D-ReLU function before the output layer.
Figure 16.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 16.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 17.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 17.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 18.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 18.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer.
Figure 19.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 19.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 20.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 20.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 21.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 21.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 22.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 22.
Accuracy of several types of networks on clean CIFAR10 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 23.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 23.
Accuracy of several types of networks on clean CIFAR100 and adversarial examples generated by a black-box attack (i.e., square attack) when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 24.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples generated by black-box attacks when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 24.
Accuracy of several types of networks on clean TinyImagenet and adversarial examples generated by black-box attacks when adding a dense layer with a D-ReLU function before the output layer and training them with augmented data samples generated from the EDM.
Figure 25.
Accuracy of several approaches on the CIFAR10 dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Figure 25.
Accuracy of several approaches on the CIFAR10 dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Figure 26.
Accuracy of several approaches on the CIFAR100 dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Figure 26.
Accuracy of several approaches on the CIFAR100 dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Figure 27.
Accuracy of several approaches on the TinyImagenet dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Figure 27.
Accuracy of several approaches on the TinyImagenet dataset under an APGD_CE attack with various perturbation bounds, where mReLU is D-ReLU.
Table 1.
Accuracy metrics for dense networks and shallow CNNs under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different attacks on the MNIST dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Table 1.
Accuracy metrics for dense networks and shallow CNNs under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different attacks on the MNIST dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Model | Training | Clean | FGSM | PGD | APCE | APDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | AT | 98.10 (1) | 89.77 (4) | 87.70 (4) | 87.63 (4) | 87.47 (4) | 12.57 (4) |
TRADES-1 | 98.07 (2) | 93.03 (2) | 90.87 (2) | 90.97 (2) | 90.83 (2) | 16.50 (1) |
TRADES-6 | 96.20 (4) | 91.40 (3) | 89.53 (3) | 89.57 (3) | 89.13 (3) | 12.83 (2) |
D-ReLU- | 97.77 (3) | 97.47 (1) | 97.10 (1) | 96.93 (1) | 97.03 (1) | 12.63 (3) |
Shallow CNN | AT | 99.20 (2) | 96.77 (3) | 95.83 (3) | 95.70 (3) | 95.73 (3) | 16.47 (2) |
TRADES-1 | 98.90 (3) | 96.93 (2) | 96.77 (2) | 96.60 (2) | 96.67 (2) | 13.87 (4) |
TRADES-6 | 98.17 (4) | 96.47 (4) | 95.30 (4) | 95.03 (4) | 95.03 (4) | 16.23 (3) |
D-ReLU- | 99.40 (1) | 98.73 (1) | 99.00 (1) | 98.30 (1) | 98.10 (1) | 16.60 (1) |
Table 2.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the CIFAR10 dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Table 2.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the CIFAR10 dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Model | Training | Clean | FGSM | PGD | APCE | APDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | AT | 52.33 (1) | 34.20 (2) | 32.83 (2) | 32.73 (2) | 31.80 (2) | 40.10 (2) |
TRADES-1 | 52.32 (2) | 29.97 (3) | 29.23 (3) | 29.20 (3) | 28.37 (3) | 38.67 (3) |
TRADES-6 | 51.30 (4) | 37.00 (1) | 36.53 (1) | 36.50 (1) | 34.57 (1) | 42.30 (1) |
D-ReLU- | 51.87 (3) | 26.03 (4) | 23.87 (4) | 23.80 (4) | 23.77 (4) | 36.10 (4) |
Shallow CNN | AT | 67.13 (2) | 42.83 (2) | 40.07 (2) | 39.90 (2) | 38.37 (2) | 50.67 (2) |
TRADES-1 | 67.37 (1) | 38.83 (4) | 35.93 (4) | 35.97 (4) | 34.13 (4) | 48.60 (4) |
TRADES-6 | 63.47 (4) | 46.13 (3) | 44.80 (3) | 44.80 (3) | 42.67 (3) | 51.67 (3) |
D-ReLU- | 66.37 (3) | 65.60 (1) | 65.60 (1) | 64.60 (1) | 64.07 (1) | 65.83 (1) |
ResNet50 | AT | 78.20 (2) | 54.77 (2) | 49.37 (2) | 48.90 (2) | 49.97 (2) | 63.00 (2) |
TRADES-1 | 75.63 (4) | 52.12 (4) | 40.77 (4) | 39.87 (4) | 40.20 (4) | 56.43 (4) |
TRADES-6 | 71.63 (3) | 54.20 (3) | 50.90 (3) | 50.40 (3) | 48.23 (3) | 57.63 (3) |
D-ReLU- | 78.87 (1) | 78.83 (1) | 78.73 (1) | 78.20 (1) | 78.40 (1) | 78.87 (1) |
ResNet101 | AT | 68.90 (3) | 44.90 (4) | 40.33 (2) | 39.43 (2) | 38.27 (2) | 49.30 (3) |
TRADES-1 | 74.60 (1) | 47.07 (2) | 32.87 (4) | 31.17 (4) | 31.37 (4) | 51.40 (2) |
TRADES-6 | 66.67 (4) | 45.43 (3) | 39.80 (3) | 39.17 (3) | 35.93 (3) | 47.67 (4) |
D-ReLU- | 75.10 (2) | 75.03 (1) | 75.37 (1) | 74.73 (1) | 74.67 (1) | 75.10 (1) |
MobilenetV2 | AT | 77.97 (2) | 46.50 (2) | 32.93 (4) | 30.73 (4) | 32.10 (4) | 51.80 (2) |
TRADES-1 | 73.13 (4) | 46.23 (3) | 31.00 (3) | 28.87 (3) | 28.77 (3) | 49.37 (4) |
TRADES-6 | 68.40 (3) | 48.80 (2) | 43.23 (2) | 43.03 (2) | 40.80 (2) | 51.13 (3) |
D-ReLU- | 81.67 (1) | 81.57 (1) | 82.00 (1) | 80.87 (1) | 80.77 (1) | 81.67 (1) |
InceptionV3 | AT | 84.60 (2) | 64.27 (2) | 58.80 (2) | 58.30 (2) | 59.33 (2) | 66.47 (2) |
TRADES-1 | 82.53 (3) | 62.30 (3) | 52.67 (4) | 51.90 (4) | 51.87 (4) | 62.40 (4) |
TRADES-6 | 76.97 (4) | 61.97 (4) | 58.00 (3) | 57.80 (3) | 56.03 (3) | 62.10 (3) |
D-ReLU- | 87.17 (1) | 86.70 (1) | 86.57 (1) | 86.13 (1) | 86.23 (1) | 86.83 (1) |
Table 3.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the CIFAR100 dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Table 3.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the CIFAR100 dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Model | Training | Clean | FGSM | PGD | APCE | APDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | AT | 24.47 (1) | 14.80 (2) | 14.30 (2) | 14.20 (2) | 12.63 (2) | 17.53 (2) |
TRADES-1 | 22.97 (3) | 13.37 (4) | 13.23 (4) | 13.17 (4) | 11.30 (4) | 16.27 (4) |
TRADES-6 | 23.27 (2) | 13.87 (3) | 13.73 (3) | 13.60 (3) | 12.27 (3) | 16.60 (3) |
D-ReLU- | 21.47 (4) | 21.03 (1) | 21.00 (1) | 20.03 (1) | 19.77 (1) | 20.73 (1) |
Shallow CNN | AT | 37.03 (1) | 17.73 (3) | 16.47 (3) | 16.30 (3) | 14.43 (3) | 22.30 (3) |
TRADES-1 | 32.60 (3) | 12.87 (4) | 11.50 (4) | 11.43 (4) | 9.47 (4) | 18.50 (4) |
TRADES-6 | 34.80 (2) | 18.67 (2) | 17.87 (2) | 17.87 (2) | 15.40 (2) | 22.33 (2) |
D-ReLU-1 | 28.63 (4) | 27.53 (1) | 27.33 (1) | 24.87 (1) | 24.60 (1) | 27.23 (1) |
ResNet50 | AT | 48.67 (3) | 26.67 (3) | 21.83 (3) | 21.53 (3) | 23.13 (3) | 31.90 (2) |
TRADES-1 | 48.97 (2) | 26.57 (4) | 19.80 (4) | 19.27 (4) | 20.03 (4) | 30.50 (4) |
TRADES-6 | 43.97 (4) | 28.90 (2) | 26.03 (2) | 25.70 (2) | 24.03 (2) | 30.63 (3) |
D-ReLU- | 52.33 (1) | 51.53 (1) | 52.47 (1) | 50.20 (1) | 51.17 (1) | 51.63 (1) |
ResNet101 | AT | 44.97 (3) | 23.57 (4) | 18.67 (3) | 18.33 (3) | 18.77 (3) | 27.77 (4) |
TRADES-1 | 48.10 (1) | 24.17 (3) | 17.70 (4) | 16.87 (4) | 17.80 (4) | 28.10 (3) |
TRADES-6 | 45.20 (2) | 28.21 (2) | 20.53 (2) | 19.32 (2) | 19.44 (2) | 30.02 (2) |
D-ReLU-1 | 44.20 (4) | 39.03 (1) | 43.10 (1) | 37.33 (1) | 36.60 (1) | 40.63 (1) |
MobilenetV2 | AT | 51.37 (2) | 23.83 (3) | 15.30 (3) | 14.43 (3) | 15.73 (2) | 28.50 (2) |
TRADES-1 | 42.97 (3) | 19.50 (4) | 9.47 (4) | 8.20 (4) | 8.60 (4) | 20.70 (4) |
TRADES-6 | 40.13 (4) | 24.50 (2) | 20.73 (2) | 20.13 (2) | 18.87 (3) | 25.40 (3) |
D-ReLU-1 | 56.40 (1) | 54.90 (1) | 55.07 (1) | 53.80 (1) | 54.17 (1) | 54.97 (1) |
InceptionV3 | AT | 56.37 (3) | 32.57 (4) | 27.20 (3) | 26.60 (3) | 28.80 (3) | 34.33 (4) |
TRADES-1 | 60.63 (2) | 35.63 (2) | 26.80 (4) | 25.83 (4) | 26.50 (4) | 35.07 (2) |
TRADES-6 | 51.10 (4) | 34.43 (3) | 31.20 (2) | 30.90 (2) | 29.50 (2) | 34.47 (3) |
D-ReLU- | 67.07 (1) | 65.10 (1) | 64.43 (1) | 63.47 (1) | 63.70 (1) | 65.27 (1) |
Table 4.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the TinyImagenet dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Table 4.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluating them on both clean samples and adversarial examples generated by different adversarial attacks on the TinyImagenet dataset. Note that APCE is APGDCE, APDLR is APGDDLR, the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, TRADES-k indicates the TRADES approach with , and D-ReLU-k represents the D-ReLU approach with .
Model | Training | Clean | FGSM | PGD | APCE | APDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | AT | 8.63 (2) | 5.40 (2) | 5.13 (3) | 5.00 (3) | 4.27 (2) | 7.00 (4) |
TRADES-1 | 8.57 (3) | 4.80 (4) | 4.77 (4) | 4.73 (4) | 4.10 (3) | 7.47 (1) |
TRADES-6 | 8.70 (1) | 5.07 (3) | 5.13 (2) | 5.10 (2) | 3.93 (4) | 7.30 (2) |
D-ReLU- | 7.53 (4) | 7.30 (1) | 7.53 (1) | 6.87 (1) | 6.83 (1) | 7.30 (3) |
Shallow CNN | AT | 18.33 (1) | 4.80 (2) | 4.17 (2) | 4.10 (2) | 2.73 (2) | 10.60 (2) |
TRADES-1 | 14.93 (3) | 2.17 (4) | 1.60 (4) | 1.60 (4) | 0.97 (4) | 8.10 (3) |
TRADES-6 | 16.37 (2) | 4.57 (3) | 4.07 (3) | 3.97 (3) | 2.67 (3) | 10.63 (1) |
D-ReLU-1 | 8.40 (4) | 8.20 (1) | 7.97 (1) | 7.20 (1) | 6.93 (1) | 7.93 (4) |
ResNet50 | AT | 40.67 (3) | 17.57 (4) | 13.17 (4) | 12.93 (4) | 14.03 (4) | 30.87 (4) |
TRADES-1 | 48.10 (1) | 22.15 (3) | 16.10 (3) | 15.55 (3) | 14.95 (3) | 36.35 (1) |
TRADES-6 | 40.97 (2) | 23.93 (2) | 21.87 (2) | 21.57 (2) | 19.77 (2) | 31.57 (3) |
D-ReLU-1 | 38.53 (4) | 32.43 (1) | 36.93 (1) | 29.33 (1) | 30.83 (1) | 35.83 (2) |
ResNet101 | AT | 32.73 (3) | 15.43 (4) | 13.10 (4) | 12.63 (4) | 11.40 (4) | 24.17 (4) |
TRADES-1 | 47.57 (1) | 20.50 (3) | 15.07 (3) | 14.57 (3) | 14.43 (3) | 34.73 (1) |
TRADES-6 | 39.13 (2) | 22.30 (1) | 20.37 (2) | 20.03 (1) | 17.67 (2) | 30.63 (2) |
D-ReLU-1 | 27.83 (4) | 22.13 (2) | 25.77 (1) | 19.93 (2) | 21.10 (1) | 24.73 (3) |
MobilenetV2 | AT | 50.00 (2) | 23.13 (3) | 16.73 (3) | 16.30 (3) | 16.97 (3) | 37.73 (1) |
TRADES-1 | 48.87 (3) | 20.60 (4) | 13.57 (4) | 12.83 (4) | 12.00 (4) | 35.10 (3) |
TRADES-6 | 43.20 (4) | 23.70 (2) | 21.23 (2) | 20.87 (2) | 19.03 (2) | 33.73 (4) |
D-ReLU-1 | 51.10 (1) | 33.63 (1) | 38.00 (1) | 31.07 (1) | 34.63 (1) | 37.03 (2) |
InceptionV3 | AT | 39.07 (4) | 18.67 (4) | 14.63 (4) | 14.57 (4) | 15.20 (4) | 27.90 (4) |
TRADES-1 | 60.43 (1) | 32.53 (1) | 23.37 (3) | 22.67 (2) | 24.13 (2) | 46.17 (1) |
TRADES-6 | 50.43 (2) | 32.03 (2) | 29.23 (1) | 28.90 (1) | 28.30 (1) | 40.40 (2) |
D-ReLU-1 | 42.63 (3) | 22.13 (3) | 26.47 (2) | 19.83 (3) | 22.50 (3) | 27.97 (3) |
Table 5.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluated on both clean samples and adversarial examples generated by a black-box attach (i.e., Square) on the CIFAR10, CIFAR100, and TinyImagenet datasets. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, and TRADES-k indicates the TRADES approach with .
Table 5.
Accuracy metrics for multiple types of networks under various robust training schemes, evaluated on both clean samples and adversarial examples generated by a black-box attach (i.e., Square) on the CIFAR10, CIFAR100, and TinyImagenet datasets. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, the numbers in parentheses are the ranks for training methods under an architecture, and TRADES-k indicates the TRADES approach with .
Model | Training | CIFAR10 | CIFAR100 | TinyImagenet |
---|
Clean | Square | Clean | Square | Clean | Square |
---|
% | % | % | % | % | % |
---|
Dense | TRADES-1 | 52.33 (1) | 34.03 (2) | 22.97 (2) | 13.90 (2) | 8.57 (2) | 4.80 (2) |
TRADES-6 | 51.30 (2) | 38.47 (1) | 23.27 (1) | 14.13 (1) | 8.70 (1) | 4.87 (1) |
D-ReLU | 48.43 (3) | 33.43 (3) | 21.47 (3) | 11.33 (3) | 7.53 (3) | 3.07 (3) |
Shallow CNN | TRADES-1 | 67.37 (1) | 45.93 (3) | 32.60 (3) | 15.47 (2) | 14.93 (3) | 5.43 (2) |
TRADES-6 | 64.50 (3) | 49.30 (2) | 34.80 (1) | 19.70 (1) | 16.37 (1) | 7.13 (1) |
D-ReLU | 66.37 (2) | 51.33 (1) | 32.87 (2) | 13.53 (3) | 16.20 (2) | 5.40 (3) |
ResNet50 | TRADES-1 | 75.70 (2) | 50.70 (3) | 48.97 (2) | 25.10 (3) | 48.40 (1) | 26.53 (1) |
TRADES-6 | 71.63 (3) | 53.57 (2) | 43.97 (3) | 27.03 (2) | 40.97 (2) | 25.03 (2) |
D-ReLU | 78.53 (1) | 62.87 (1) | 52.33 (1) | 28.43 (1) | 38.53 (3) | 20.50 (3) |
ResNet101 | TRADES-1 | 74.60 (1) | 45.37 (2) | 48.10 (1) | 23.20 (2) | 47.57 (1) | 25.07 (1) |
TRADES-6 | 66.67 (3) | 43.63 (3) | 10.67 (3) | 1.67 (3) | 39.13 (2) | 24.00 (2) |
D-ReLU | 72.00 (2) | 53.03 (1) | 44.20 (2) | 28.07 (1) | 27.83 (3) | 12.43 (3) |
MobilenetV2 | TRADES-1 | 73.13 (2) | 43.13 (3) | 42.97 (2) | 15.40 (3) | 48.87 (2) | 25.00 (2) |
TRADES-6 | 68.60 (3) | 49.17 (2) | 40.13 (3) | 22.30 (2) | 43.20 (3) | 26.23 (1) |
D-ReLU | 82.90 (1) | 61.03 (1) | 56.40 (1) | 27.90 (1) | 51.10 (1) | 18.33 (3) |
InceptionV3 | TRADES-1 | 82.53 (2) | 64.17 (2) | 60.63 (2) | 34.50 (2) | 60.43 (1) | 39.60 (1) |
TRADES-6 | 76.97 (3) | 62.40 (3) | 51.10 (3) | 34.03 (3) | 50.43 (2) | 36.10 (2) |
D-ReLU | 87.17 (1) | 74.20 (1) | 67.07 (1) | 41.40 (1) | 42.63 (3) | 24.63 (3) |
Table 6.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the CIFAR10 dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Table 6.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the CIFAR10 dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Model | Training | Clean | FGSM | PGD | APGDCE | APGDDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | D-ReLU | 48.47 (2) | 46.87 (1) | 48.03 (1) | 45.57 (2) | 45.83 (1) | 47.33 (2) |
TRADES | 62.47 (1) | 46.67 (2) | 46.07 (2) | 46.13 (1) | 44.63 (2) | 52.8 (1) |
Shallow CNN | D-ReLU | 67.97 (2) | 66.57 (1) | 67.07 (1) | 65.4 (1) | 65.4 (1) | 66.97 (1) |
TRADES | 74.3 (1) | 59.03 (2) | 57.93 (2) | 57.93 (2) | 56.53 (2) | 63.6 (2) |
ResNet50 | D-ReLU | 79.1 (2) | 78.87 (1) | 78.67 (1) | 78.63 (1) | 78.57 (1) | 78.87 (1) |
TRADES | 80.6 (1) | 66.77 (2) | 65.97 (2) | 65.5 (2) | 64.03 (2) | 70.2 (2) |
ResNet101 | D-ReLU | 76.77 (2) | 76.37 (1) | 76.63 (1) | 76.43 (1) | 76.33 (1) | 76.43 (1) |
TRADES | 77.97 (1) | 63.43 (2) | 61.93 (2) | 61.87 (2) | 59.77 (2) | 67.33 (2) |
MobilenetV2 | D-ReLU | 81.8 (1) | 81.47 (1) | 81.6 (1) | 80.97 (1) | 80.97 (1) | 81.67 (1) |
TRADES | 79.33 (2) | 62.27 (2) | 61.1 (2) | 60.67 (2) | 58.4 (2) | 66.87 (2) |
InceptionV3 | D-ReLU | 87.4 (2) | 86.77 (1) | 86.23 (1) | 86.4 (1) | 86.33 (1) | 86.9 (1) |
TRADES | 87.73 (1) | 74.53 (2) | 73.17 (2) | 73.07 (2) | 72.1 (2) | 75.93 (2) |
Table 7.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the CIFAR100 dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Table 7.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the CIFAR100 dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Model | Training | Clean | FGSM | PGD | APGDCE | APGDDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | D-ReLU | 22.90 (2) | 22.13 (2) | 22.37 (2) | 21.17 (2) | 20.80 (2) | 22.23 (2) |
TRADES | 36.03 (1) | 23.93 (1) | 23.57 (1) | 23.47 (1) | 22.13 (1) | 26.97 (1) |
Shallow CNN | D-ReLU | 32.20 (2) | 31.50 (1) | 31.70 (1) | 28.57 (2) | 28.50 (1) | 31.03 (2) |
TRADES | 44.23 (1) | 29.90 (2) | 29.33 (2) | 29.30 (1) | 26.93 (2) | 33.90 (1) |
ResNet50 | D-ReLU | 53.83 (2) | 52.8 (1) | 53.03 (1) | 52.13 (1) | 52.50 (1) | 52.77 (1) |
TRADES | 55.33 (1) | 40.17 (2) | 38.03 (2) | 37.80 (2) | 37.27 (2) | 43.13 (2) |
ResNet101 | D-ReLU | 44.50 (2) | 43.90 (1) | 44.60 (1) | 43.47 (1) | 43.50 (1) | 44.20 (1) |
TRADES | 52.60 (1) | 37.73 (2) | 36.23 (2) | 36.03 (2) | 34.57 (2) | 41.27 (2) |
MobilenetV2 | D-ReLU | 56.57 (1) | 55.57 (1) | 55.77 (1) | 54.67 (1) | 54.87 (1) | 55.70 (1) |
TRADES | 51.27 (2) | 38.57 (2) | 37.10 (2) | 36.73 (2) | 35.50 (2) | 40.90 (2) |
InceptionV3 | D-ReLU | 63.47 (1) | 61.43 (1) | 61.07 (1) | 60.40 (1) | 60.70 (1) | 61.33 (1) |
TRADES | 62.90 (2) | 48.33 (2) | 46.5 (2) | 46.23 (2) | 45.67 (2) | 49.43 (2) |
Table 8.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the TinyImagenet dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Table 8.
Accuracy metrics for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluating them on both clean samples and adversarial examples generated by different white-box attacks on the TinyImagenet dataset. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Model | Training | Clean | FGSM | PGD | APGDCE | APGDDLR | CWL2 |
---|
% | % | % | % | % | % |
---|
Dense | D-ReLU | 1.3 (2) | 1.27 (1) | 1.37 (1) | 1.3 (1) | 1.3 (1) | 1.27 (2) |
TRADES | 2.4 (1) | 1.07 (2) | 1.07 (2) | 1.03 (2) | 0.8 (2) | 1.77 (1) |
Shallow CNN | D-ReLU | 1.87 (2) | 1.77 (2) | 1.77 (2) | 1.5 (2) | 1.53 (1) | 1.8 (2) |
TRADES | 7.33 (1) | 1.97 (1) | 1.87 (1) | 1.87 (1) | 1.13 (2) | 4.6 (1) |
ResNet50 | D-ReLU | 29.63 (1) | 24.43 (1) | 27.8 (1) | 21.47 (1) | 21.6 (1) | 26.43 (1) |
TRADES | 8.63 (2) | 4.13 (2) | 3.7 (2) | 3.57 (2) | 2.9 (2) | 5.97 (2) |
ResNet101 | D-ReLU | 17.6 (1) | 9.4 (1) | 12.6 (1) | 4.53 (1) | 5.13 (1) | 12.23 (1) |
TRADES | 7.3 (2) | 3.63 (2) | 3.37 (2) | 3.33 (2) | 2.87 (2) | 5.2 (2) |
MobilenetV2 | D-ReLU | 42.43 (1) | 24.43 (1) | 28.63 (1) | 20.93 (1) | 21.63 (1) | 29.2 (1) |
TRADES | 18.13 (2) | 8 (2) | 7.03 (2) | 6.63 (2) | 5.2 (2) | 12.63 (2) |
InceptionV3 | D-ReLU | 35.63 (1) | 10 (1) | 9.73 (1) | 3.33 (2) | 4.33 (1) | 16.9 (1) |
TRADES | 12.2 (2) | 5.57 (2) | 5.07 (2) | 5 (1) | 4.3 (2) | 7.63 (2) |
Table 9.
Accuracy for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluated on both clean samples and adversarial examples generated by a black-box attack (i.e., square) on the CIFAR10, CIFAR100, and TinyImagenet datasets. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Table 9.
Accuracy for multiple types of networks under various robust training schemes with generated samples from the EDM, evaluated on both clean samples and adversarial examples generated by a black-box attack (i.e., square) on the CIFAR10, CIFAR100, and TinyImagenet datasets. Note that the accuracy metrics in bold are the highest in a specific model among the different training methods, and the numbers in parentheses are the ranks for training methods under an architecture.
Model | Training | CIFAR10 | CIFAR100 | TinyImagenet |
---|
Clean | Square | Clean | Square | Clean | Square |
---|
% | % | % | % | % | % |
---|
Dense | D-ReLU | 52.6 (2) | 48.77 (1) | 22.9 (2) | 12.4 (2) | 1.3 (2) | 0.7 (2) |
TRADES | 62.47 (1) | 47.23 (2) | 36.03 (1) | 23.6 (1) | 2.4 (1) | 0.93 (1) |
Shallow CNN | D-ReLU | 67.97 (2) | 52.17 (2) | 35.3 (2) | 14.43 (2) | 2.67 (2) | 0.5 (2) |
TRADES | 74.3 (1) | 60.9 (1) | 44.23 (1) | 31.2 (1) | 7.33 (1) | 3.13 (1) |
ResNet50 | D-ReLU | 79.1 (2) | 64.93 (2) | 53.83 (2) | 33.03 (2) | 32.27 (1) | 14.37 (1) |
TRADES | 80.6 (1) | 67.9 (1) | 55.33 (1) | 40.07 (1) | 7.33 (2) | 4.7 (2) |
ResNet101 | D-ReLU | 76.77 (2) | 59.5 (2) | 47.43 (2) | 31.3 (2) | 17.6 (1) | 5.7 (1) |
TRADES | 77.97 (1) | 64.53 (1) | 52.6 (1) | 37.97 (1) | 7.3 (2) | 3.87 (2) |
MobilenetV2 | D-ReLU | 81.8 (1) | 62.33 (2) | 56.57 (1) | 31.27 (2) | 42.43 (1) | 18.93 (1) |
TRADES | 79.33 (2) | 64.53 (1) | 51.27 (2) | 37.97 (1) | 18.13 (2) | 9.9 (2) |
InceptionV3 | D-ReLU | 87.4 (2) | 74.73 (2) | 63.47 (1) | 42.37 (2) | 35.63 (1) | 14.93 (1) |
TRADES | 87.73 (1) | 76.63 (1) | 62.9 (2) | 48.8 (1) | 12.2 (2) | 6.63 (2) |