Figure 1.
Examples of samples we generated from a class of fractals. Note that in [
9], fractals belonging to the same class share similar geometric properties, as they are sampled by slightly perturbing on one of the parameters of the linear operator
A. Contrary to [
15], different fractals are grouped under the same class, lacking geometric continuity within samples from the same class.
Figure 1.
Examples of samples we generated from a class of fractals. Note that in [
9], fractals belonging to the same class share similar geometric properties, as they are sampled by slightly perturbing on one of the parameters of the linear operator
A. Contrary to [
15], different fractals are grouped under the same class, lacking geometric continuity within samples from the same class.
Figure 2.
Examples of samples we generated from a class of MandelbulbVAR-1k. We can observe that a class is composed of the same Mandelbulb taken from different perspectives and with various colour patterns, ensuring geometric continuity between objects of the same class.
Figure 2.
Examples of samples we generated from a class of MandelbulbVAR-1k. We can observe that a class is composed of the same Mandelbulb taken from different perspectives and with various colour patterns, ensuring geometric continuity between objects of the same class.
Figure 3.
Overview of the proposed “Multi-Formula” dataset. Fractals from different classes from the source dataset are grouped to be the features of new classes, where a variable number of fractals are present in a sample of a class.
Figure 3.
Overview of the proposed “Multi-Formula” dataset. Fractals from different classes from the source dataset are grouped to be the features of new classes, where a variable number of fractals are present in a sample of a class.
Figure 4.
The left-hand box (“Dataset Generation”) illustrates two distinct IFS, each defining unique codes obtained by sampling the parameters of the system which are used to generate both Fractal and Mandelbulb datasets. In the middle box ("Pre-Training") a computer vision model for multi-class classification is trained from the generated images, either with a single sample or multiple samples per image. Finally, in the last box (“Anomaly Detection”), the model is used as a feature extractor for unsupervised anomaly detection.
Figure 4.
The left-hand box (“Dataset Generation”) illustrates two distinct IFS, each defining unique codes obtained by sampling the parameters of the system which are used to generate both Fractal and Mandelbulb datasets. In the middle box ("Pre-Training") a computer vision model for multi-class classification is trained from the generated images, either with a single sample or multiple samples per image. Finally, in the last box (“Anomaly Detection”), the model is used as a feature extractor for unsupervised anomaly detection.
Figure 5.
Spider chart representing average image-level AUROC grouping MVTecAD and VisA classes into different object categories.
Figure 5.
Spider chart representing average image-level AUROC grouping MVTecAD and VisA classes into different object categories.
Figure 6.
Comparison between ImageNet and Fractals pre-training when using different feature hierarchies.
Figure 6.
Comparison between ImageNet and Fractals pre-training when using different feature hierarchies.
Figure 7.
Comparison between ImageNet, Fractals, and MandelbulbVAR-1k pre-training when using different feature hierarchies on PaDiM.
Figure 7.
Comparison between ImageNet, Fractals, and MandelbulbVAR-1k pre-training when using different feature hierarchies on PaDiM.
Figure 8.
Comparison of the filters from the first convolutional layer of WideResNet50 pre-trained with different datasets.
Figure 8.
Comparison of the filters from the first convolutional layer of WideResNet50 pre-trained with different datasets.
Figure 9.
Qualitative visualization for the MVTecAD’s classes: bottle, cable, carpet, hazelnut, and wood. In the first column, we have the original image and the ground-truth. In the blue box, we have the anomaly score and predicted segmentation mask for ImageNet pre-training, in the red box for Fractals, and the purple box for MandelbulbVAR-1k.
Figure 9.
Qualitative visualization for the MVTecAD’s classes: bottle, cable, carpet, hazelnut, and wood. In the first column, we have the original image and the ground-truth. In the blue box, we have the anomaly score and predicted segmentation mask for ImageNet pre-training, in the red box for Fractals, and the purple box for MandelbulbVAR-1k.
Figure 10.
Top-1 classification accuracy during training for different generated datasets.
Figure 10.
Top-1 classification accuracy during training for different generated datasets.
Figure 11.
Comparison of the filters from the first convolutional layer of WideResNet50 pre-trained with different dataset configurations.
Figure 11.
Comparison of the filters from the first convolutional layer of WideResNet50 pre-trained with different dataset configurations.
Figure 12.
The t-SNE plot of the CIFAR-10 validation set, using WideResNet-50 pre-trained on different datasets, is presented. We extracted feature vectors from the penultimate layers, prior to the final classification layers, without any fine-tuning. (Note: The legend in each t-SNE plot is intentionally small, as our focus is on illustrating the structure of the latent space rather than the classification of each individual point).
Figure 12.
The t-SNE plot of the CIFAR-10 validation set, using WideResNet-50 pre-trained on different datasets, is presented. We extracted feature vectors from the penultimate layers, prior to the final classification layers, without any fine-tuning. (Note: The legend in each t-SNE plot is intentionally small, as our focus is on illustrating the structure of the latent space rather than the classification of each individual point).
Figure 13.
Image- (left) and pixel-level (right) AUROC scores achieved with PatchCore at various epochs of the pre-training stage using different training configurations.
Figure 13.
Image- (left) and pixel-level (right) AUROC scores achieved with PatchCore at various epochs of the pre-training stage using different training configurations.
Figure 14.
Comparison of the filters from the first convolutional layer of WideResNet-50 that give the results reported in
Table 14. Some of the “dot-like” filters are framed in red.
Figure 14.
Comparison of the filters from the first convolutional layer of WideResNet-50 that give the results reported in
Table 14. Some of the “dot-like” filters are framed in red.
Table 1.
MVTecAD image-level AUROC. Each cell carries the results for ImageNet/Fractals.
Table 1.
MVTecAD image-level AUROC. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM | CutPaste | PANDA |
---|
carpet | 98.6/64.5 | 92.7/49.6 | 98.0/40.9 | 99.0/42.5 | 98.9/30.6 | 98.0/53.9 | 85.9/69.2 | 93.4/31.2 |
grid | 99.8/58.0 | 96.1/82.0 | 97.5/93.7 | 96.9/78.5 | 100.0/68.3 | 98.3/46.0 | 98.3/100.0 | 52.0/54.4 |
leather | 99.7/88.5 | 96.1/63.3 | 100.0/82.0 | 99.7/81.9 | 100.0/75.2 | 99.8/67.9 | 100.0/87.3 | 96.5/54.4 |
tile | 99.9/95.6 | 99.9/92.7 | 98.8/95.6 | 99.5/97.3 | 100.0/60.9 | 98.6/74.0 | 94.7/84.8 | 96.8/65.1 |
wood | 99.2/99.6 | 95.6/93.8 | 99.4/97.9 | 99.1/97.1 | 99.4/84.3 | 99.7/75.5 | 99.7/95.7 | 95.9/56.8 |
bottle | 100.0/97.6 | 100.0/56.7 | 100.0/88.2 | 99.8/95.9 | 99.9/93.2 | 100.0/54.9 | 99.8/97.9 | 96.8/65.1 |
cable | 92.9/55.6 | 92.0/45.9 | 98.8/52.2 | 93.2/61.4 | 96.2/58.6 | 91.3/43.9 | 90.6/85.8 | 84.5/54.9 |
capsule | 94.7/42.1 | 90.4/61.6 | 97.8/73.4 | 91.9/70.8 | 97.6/78.3 | 57.9/56.5 | 83.5/78.1 | 91.8/71.8 |
hazelnut | 97.9/97.6 | 99.6/85.7 | 100.0/92.0 | 94.1/93.9 | 100.0/89.5 | 100.0/90.8 | 97.2/71.3 | 88.5/61.3 |
metal_nut | 98.7/57.8 | 96.4/34.4 | 99.8/38.1 | 98.7/47.9 | 100.0/69.8 | 96.6/66.2 | 94.2/80.7 | 72.9/41.5 |
pill | 96.4/79.5 | 82.4/76.5 | 93.1/75.9 | 92.3/77.2 | 96.7/72.4 | 81.0/77.4 | 89.1/71.0 | 81.0/65.3 |
screw | 85.0/27.5 | 89.1/69.0 | 97.9/61.7 | 85.2/40.0 | 98.1/69.1 | 90.3/60.4 | 79.0/42.75 | 70.5/41.3 |
toothbrush | 77.5/60.8 | 71.4/78.3 | 100.0/99.2 | 87.2/98.6 | 93.9/96.7 | 85.0/79.2 | 87.8/97.8 | 88.1/68.9 |
transistor | 89.7/59.7 | 87.8/33.0 | 99.9/55.2 | 98.5/78.6 | 97.4/66.8 | 94.9/37.5 | 92.8/79.8 | 91.0/71.2 |
zipper | 89.3/74.4 | 91.6/44.6 | 99.3/81.2 | 88.3/76.8 | 98.3/83.2 | 81.5/46.5 | 99.8/70.9 | 97.0/57.6 |
Model Avg
| 94.6/70.6 | 92.1/64.5 | 98.7/75.1 | 94.9/75.9 | 98.4/73.1 | 91.5/62.0 | 92.8/80.9 | 86.4/57.4 |
Model STD
| 6.6/22.1 | 7.5/20.2 | 1.8/20.8 | 5.0/20.0 | 1.8/16.4 | 11.5/15.4 | 6.7/15.0 | 12.8/11.8 |
Table 2.
MVTecAD pixel-level AUROC. Each cell carries the results for ImageNet/Fractals.
Table 2.
MVTecAD pixel-level AUROC. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
carpet | 98.2/78.4 | 98.8/71.2 | 98.7/72.7 | 98.8/73.2 | 98.8/56.2 | 99.2/76.7 |
grid | 98.6/85.0 | 97.4/72.2 | 98.0/82.3 | 96.7/69.6 | 99.3/88.3 | 99.2/69.5 |
leather | 98.9/96.6 | 97.4/84.2 | 98.9/95.6 | 98.9/90.5 | 99.1/92.4 | 99.6/83.5 |
tile | 95.7/87.1 | 95.8/76.0 | 94.9/85.9 | 94.9/74.2 | 95.4/69.0 | 97.1/76.0 |
wood | 90.8/84.9 | 95.0/82.0 | 93.2/84.0 | 93.9/84.5 | 94.9/84.9 | 96.9/85.2 |
bottle | 97.8/92.3 | 98.5/59.3 | 98.0/84.4 | 98.3/92.2 | 98.3/76.4 | 98.7/59.9 |
cable | 93.8/78.2 | 95.6/68.3 | 98.0/84.3 | 97.2/89.0 | 96.4/53.9 | 94.9/73.8 |
capsule | 98.7/85.5 | 98.7/90.8 | 98.8/95.2 | 98.5/95.0 | 98.7/94.3 | 97.6/95.1 |
hazelnut | 95.3/95.9 | 98.2/95.7 | 98.4/97.1 | 98.6/97.9 | 98.8/96.5 | 99.1/95.2 |
metal_nut | 98.6/82.7 | 97.4/76.1 | 98.5/84.4 | 96.1/86.5 | 97.0/82.4 | 98.2/81.8 |
pill | 97.5/85.3 | 98.0/90.7 | 97.5/94.6 | 95.2/92.7 | 97.4/91.2 | 95.8/88.0 |
screw | 98.1/85.0 | 97.4/93.9 | 99.2/95.7 | 98.7/94.8 | 99.6/97.0 | 98.9/93.6 |
toothbrush | 95.2/72.6 | 98.2/88.2 | 98.7/97.1 | 99.0/97.6 | 98.9/93.2 | 99.0/91.9 |
transistor | 92.6/78.3 | 85.9/53.7 | 96.7/75.2 | 97.6/86.5 | 89.1/66.6 | 82.3/59.5 |
zipper | 95.9/74.3 | 96.3/70.7 | 98.1/86.6 | 97.2/88.0 | 98.5/78.0 | 98.1/78.6 |
Model AVG
| 96.4/84.1 | 96.6/78.2 | 97.7/87.7 | 97.3/87.5 | 97.3/81.4 | 97.0/80.6 |
Model STD
| 2.5/7.1 | 3.2/12.6 | 1.6/7.9 | 1.6/8.8 | 2.7/14.3 | 4.3/11.6 |
Table 3.
MVTecAD AUPRO. Each cell carries the results for ImageNet/Fractals.
Table 3.
MVTecAD AUPRO. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
carpet | –/51.3 | 93.8/33.1 | 92.7/31.4 | 95.3/39.6 | 94.8/24.8 | 97.0/51.9 |
grid | 95.1/63.2 | 90.8/40.3 | 90.1/60.7 | 89.0/41.1 | 97.3/70.2 | 97.0/31.6 |
leather | 98.3/89.8 | 90.8/47.9 | 96.3/76.7 | 98.0/68.9 | 97.9/69.0 | 99.0/51.6 |
tile | 87.4/72.1 | 90.2/63.3 | 79.6/69.0 | 86.3/64.3 | 87.5/45.1 | 92.4/49.5 |
wood | 89.3/75.0 | 88.6/50.7 | 84.6/54.9 | 91.6/65.5 | 91.3/70.3 | 95.7/62.7 |
bottle | 88.7/76.1 | 93.5/28.1 | 92.3/64.7 | 95.1/77.4 | 95.3/53.2 | 96.2/22.5 |
cable | 80.3/38.6 | 84.8/29.9 | 91.1/46.8 | 88.5/62.5 | 90.1/41.4 | 89.0/30.4 |
capsule | 92.4/59.3 | 91.0/73.9 | 92.3/75.1 | 91.1/77.6 | 93.0/81.8 | 91.1/81.9 |
hazelnut | 95.2/89.7 | 95.1/86.2 | 94.4/87.0 | 95.0/90.1 | 96.3/90.1 | 97.6/87.6 |
metal_nut | 92.8/47.5 | 87.2/27.4 | 91.9/49.4 | 91.9/54.1 | 93.8/40.0 | 95.4/36.8 |
pill | 91.3/68.9 | 93.4/65.0 | 93.8/83.8 | 94.4/85.6 | 96.2/82.2 | 95.1/72.7 |
screw | 91.2/59.9 | 89.2/80.3 | 95.5/84.0 | 94.7/83.6 | 97.7/88.5 | 95.0/78.8 |
toothbrush | 77.8/28.3 | 82.9/64.1 | 86.2/82.7 | 93.2/91.6 | 91.6/79.4 | 92.9/70.4 |
transistor | 79.1/44.4 | 73.8/21.8 | 94.0/42.3 | 94.0/62.4 | 79.2/41.1 | 69.4/16.0 |
zipper | 87.8/41.8 | 87.7/30.2 | 92.5/67.7 | 91.3/64.2 | 95.3/50.4 | 94.2/38.3 |
Model AVG
| 89.1/60.4 | 88.9/49.5 | 91.2/65.1 | 92.6/68.6 | 93.2/61.8 | 93.1/52.2 |
Model STD
| 23.8/18.4 | 5.4/21.3 | 4.5/17.1 | 3.1/16.0 | 4.9/20.7 | 7.1/22.7 |
Table 4.
VisA image-level AUROC. Each cell carries the results for ImageNet/Fractals.
Table 4.
VisA image-level AUROC. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM | CutPaste | PANDA |
---|
candle | 94.2/69.7 | 92.2/69.1 | 97.9/83.1 | 92.6/79.7 | 94.0/76.2 | 80.7/70.7 | 96.6/77.9 | 88.4/67.9 |
capsules | 85.6/49.8 | 79.4/69.1 | 68.4/79.6 | 65.6/62.7 | 84.6/62.7 | 88.4/68.4 | 83.7/71.4 | 57.1/68.2 |
cashew | 89.0/90.9 | 91.9/78.6 | 95.6/91.8 | 88.1/82.3 | 96.3/65.0 | 86.1/80.2 | 82.7/73.1 | 91.6/90.2 |
chewinggum | 95.8/91.6 | 98.4/80.1 | 99.4/81.9 | 98.3/71.7 | 99.4/67.8 | 98.2/73.5 | 96.6/86.0 | 92.2/69.0 |
fryum | 78.0/61.1 | 78.0/71.4 | 91.6/82.6 | 84.6/80.7 | 91.9/70.8 | 89.2/60.7 | 93.4/75.8 | 84.5/74.8 |
macaroni1 | 95.0/84.8 | 87.7/66.2 | 89.7/75.9 | 81.1/71.5 | 96.3/73.1 | 92.2/72.9 | 85.1/67.1 | 77.2/68.0 |
macaroni2 | 86.9/52.4 | 76.8/58.0 | 71.7/59.6 | 62.0/60.8 | 80.8/62.7 | 84.3/59.1 | 63.1/75.5 | 58.7/67.3 |
pcb1 | 95.2/72.4 | 90.9/54.6 | 95.1/89.8 | 83.2/83.3 | 97.0/62.9 | 87.6/36.0 | 89.4/92.7 | 87.0/59.5 |
pcb2 | 95.2/80.7 | 80.0/29.8 | 93.5/94.7 | 82.7/88.3 | 96.8/85.6 | 90.3/30.2 | 93.6/95.5 | 91.3/83.7 |
pcb3 | 94.4/50.5 | 85.6/56.6 | 91.9/71.1 | 78.9/76.5 | 96.5/93.2 | 90.0/64.0 | 89.7/72.6 | 78.1/64.3 |
pcb4 | 97.0/69.8 | 97.1/83.9 | 99.5/90.6 | 93.2/94.0 | 99.8/96.5 | 95.5/81.4 | 97.4/95.0 | 96.5/83.0 |
pipe_fryum | 99.5/64.8 | 94.8/64.5 | 98.5/64.4 | 96.7/66.1 | 97.3/74.6 | 92.6/64.3 | 76.3/67.3 | 80.1/59.8 |
Model AVG
| 92.1/69.9 | 87.7/65.2 | 91.1/80.4 | 83.9/76.5 | 94.2/74.3 | 89.6/63.4 | 87.3/79.2 | 81.9/71.3 |
Model STD
| 6.1/14.9 | 7.7/14.5 | 10.3/11.0 | 11.3/10.2 | 5.8/11.8 | 4.8/15.8 | 10.0/10.4 | 12.7/9.7 |
Table 5.
VisA pixel-level AUROC. Each cell carries the results for ImageNet/Fractals.
Table 5.
VisA pixel-level AUROC. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
candle | 99.2/80.7 | 98.7/74.6 | 98.9/82.6 | 98.7/77.4 | 99.0/85.9 | 98.9/86.5 |
capsules | 98.2/84.2 | 97.0/82.2 | 97.6/90.9 | 96.3/90.2 | 99.6/92.5 | 99.3/76.8 |
cashew | 98.2/89.6 | 99.1/91.8 | 99.0/75.1 | 98.6/74.3 | 95.1/41.4 | 97.0/92.8 |
chewinggum | 99.2/96.9 | 98.8/94.1 | 98.9/87.3 | 98.9/69.1 | 98.7/86.7 | 99.1/93.3 |
fryum | 89.0/88.5 | 96.5/89.0 | 94.9/94.2 | 95.5/94.1 | 96.3/92.1 | 95.4/87.0 |
macaroni1 | 96.3/98.0 | 98.6/91.3 | 98.2/95.2 | 97.4/93.8 | 99.5/98.6 | 99.4/97.3 |
macaroni2 | 98.7/94.9 | 97.5/90.9 | 96.9/91.8 | 94.9/91.0 | 99.2/96.2 | 99.6/95.5 |
pcb1 | 99.7/94.0 | 99.1/87.2 | 99.5/98.4 | 98.7/89.6 | 99.6/31.1 | 99.4/47.7 |
pcb2 | 98.7/91.0 | 96.1/84.0 | 97.8/92.8 | 97.3/94.3 | 98.5/89.5 | 97.3/76.8 |
pcb3 | 93.5/85.4 | 97.3/86.2 | 98.2/92.7 | 97.2/96.1 | 99.0/95.0 | 98.1/89.3 |
pcb4 | 98.4/77.0 | 97.8/81.9 | 97.7/83.2 | 96.5/88.4 | 98.1/94.3 | 98.2/89.6 |
pipe_fryum | 98.3/90.7 | 98.6/95.8 | 98.8/96.0 | 98.9/96.9 | 98.7/97.2 | 97.9/96.7 |
Model AVG
| 97.3/89.2 | 97.9/87.4 | 98.0/90.0 | 97.4/87.9 | 98.4/83.4 | 98.3/85.8 |
Model STD
| 3.0/6.5 | 1.0/6.0 | 1.2/6.7 | 1.4/0.2 | 1.4/22.5 | 1.3/13.8 |
Table 6.
VisA AUPRO. Each cell carries the results for ImageNet/Fractals.
Table 6.
VisA AUPRO. Each cell carries the results for ImageNet/Fractals.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
candle | 94.8/42.5 | 92.7/43.2 | 94.3/72.8 | 94.0/49.4 | 94.1/71.4 | 94.5/61.8 |
capsules | 90.6/45.9 | 75.3/51.3 | 67.8/61.9 | 68.7/56.8 | 93.1/51.7 | 95.3/44.6 |
cashew | 81.1/81.3 | 92.5/74.3 | 89.4/42.6 | 84.6/37.7 | 87.4/38.1 | 92.1/77.0 |
chewinggum | 84.4/62.7 | 88.9/53.7 | 84.7/43.0 | 86.5/29.8 | 80.5/48.0 | 83.0/68.6 |
fryum | 69.7/68.7 | 81.0/69.7 | 80.2/72.2 | 70.1/70.6 | 88.4/77.8 | 85.9/65.3 |
macaroni1 | 87.1/95.1 | 90.7/79.1 | 91.8/81.8 | 87.6/67.3 | 95.0/87.3 | 94.8/88.0 |
macaroni2 | 93.9/69.4 | 83.4/60.9 | 86.9/58.3 | 71.5/54.9 | 92.7/75.4 | 95.5/76.2 |
pcb1 | 92.5/64.9 | 88.1/49.7 | 89.9/77.8 | 87.5/74.4 | 95.6/18.0 | 92.3/14.4 |
pcb2 | 85.7/68.5 | 76.7/54.4 | 83.7/78.9 | 77.6/78.8 | 90.4/67.2 | 85.3/33.7 |
pcb3 | 79.6/42.1 | 73.5/64.9 | 80.4/78.5 | 70.6/80.7 | 91.0/88.4 | 89.6/77.1 |
pcb4 | 89.0/30.6 | 86.2/42.8 | 84.6/44.1 | 79.1/52.6 | 88.1/75.7 | 89.7/66.1 |
pipe_fryum | 86.1/78.0 | 92.9/87.0 | 93.4/78.5 | 90.5/79.2 | 95.0/88.9 | 93.7/88.9 |
Model AVG
| 86.2/62.5 | 85.2/60.9 | 85.6/65.9 | 80.7/61.0 | 90.9/65.7 | 91.0/63.5 |
Model STD
| 7.0/18.8 | 7.0/14.3 | 7.3/15.3 | 8.9/16.9 | 4.4/22.2 | 4.3/22.2 |
Table 7.
MVTecAD image-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 7.
MVTecAD image-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
carpet | 99.0/94.9 | 95.1/77.9 | 98.9/93.3 | 99.6/88.7 | 99.0/92.8 | 98.5/86.3 |
grid | 99.8/99.2 | 95.2/71.5 | 98.0/98.4 | 95.4/94.9 | 95.3/97.5 | 97.4/78.6 |
leather | 99.9/99.6 | 98.3/87.5 | 100.0/97.1 | 100.0/97.6 | 100.0/91.2 | 99.9/98.4 |
tile | 99.7/99.6 | 99.8/98.4 | 98.8/98.9 | 99.7/84.0 | 99.9/99.9 | 99.2/98.4 |
wood | 99.3/97.6 | 93.7/95.6 | 99.1/98.5 | 99.2/98.6 | 99.4/98.8 | 99.6/98.5 |
bottle | 99.7/99.6 | 100.0/97.7 | 100.0/99.6 | 100.0/100.0 | 100.0/100.0 | 97.5/95.4 |
cable | 95.8/90.1 | 84.7/77.8 | 98.8/98.5 | 89.5/92.1 | 95.5/83.7 | 81.5/63.5 |
capsule | 90.5/79.2 | 88.2/81.1 | 97.9/91.5 | 93.1/88.1 | 96.9/91.7 | 58.5/53.2 |
hazelnut | 95.6/83.3 | 96.7/84.8 | 100.0/98.1 | 92.3/71.3 | 100.0/94.9 | 98.2/93.5 |
metal_nut | 98.9/93.4 | 91.8/69.0 | 99.8/95.3 | 99.8/93.3 | 100.0/94.7 | 95.9/82.8 |
pill | 95.1/71.7 | 82.0/80.4 | 94.1/88.2 | 92.5/78.3 | 97.9/91.0 | 51.0/41.3 |
screw | 74.0/88.1 | 82.4/50.7 | 98.0/83.2 | 85.7/70.2 | 97.7/90.9 | 45.8/55.5 |
toothbrush | 85.2/65.5 | 85.8/90.8 | 99.7/99.4 | 90.2/96.6 | 93.6/99.9 | 81.6/68.3 |
transistor | 96.6/84.2 | 96.5/83.9 | 99.9/98.9 | 98.5/96.2 | 97.3/92.7 | 80.1/69.9 |
zipper | 92.3/94.3 | 93.1/93.9 | 99.1/99.1 | 88.6/79.1 | 97.5/95.7 | 79.2/55.6 |
Model AVG
| 94.8/89.4 | 92.2/82.7 | 98.8/95.9 | 94.9/88.6 | 98.0/94.4 | 84.3/75.9 |
Model STD
| 7.1/10.0 | 6.1/12.6 | 1.5/4.8 | 5.9/9.9 | 2.0/4.5 | 18.7/19.2 |
Table 8.
MVTecAD pixel-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 8.
MVTecAD pixel-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
carpet | 96.8/94.4 | 98.9/96.1 | 98.8/98.4 | 99.0/97.4 | 99.0/98.5 | 99.3/96.9 |
grid | 98.6/98.7 | 96.8/87.2 | 98.0/93.5 | 97.1/96.3 | 99.0/99.0 | 98.9/92.4 |
leather | 98.9/99.3 | 99.5/98.2 | 98.9/98.9 | 98.9/99.4 | 99.2/99.2 | 99.6/99.5 |
tile | 91.3/94.4 | 95.6/89.6 | 94.7/89.8 | 94.4/83.9 | 94.9/91.2 | 96.9/89.9 |
wood | 85.0/85.7 | 93.3/87.3 | 92.9/90.2 | 94.4/92.1 | 94.7/92.6 | 96.3/92.8 |
bottle | 97.4/98.2 | 98.2/97.7 | 98.1/98.4 | 98.4/98.8 | 98.5/98.2 | 94.9/88.4 |
cable | 94.2/93.8 | 94.4/88.8 | 98.0/96.3 | 97.0/96.1 | 96.7/91.8 | 92.0/88.1 |
capsule | 98.7/96.5 | 98.8/97.1 | 98.7/98.2 | 98.6/98.5 | 98.7/98.7 | 92.9/88.9 |
hazelnut | 95.5/97.5 | 98.6/97.9 | 98.4/98.7 | 98.0/98.5 | 98.7/99.1 | 97.3/97.9 |
metal_nut | 97.5/97.1 | 97.5/96.1 | 98.2/98.8 | 96.2/98.6 | 96.6/97.1 | 97.5/93.9 |
pill | 97.1/81.2 | 97.6/84.0 | 97.6/93.6 | 94.5/90.8 | 97.5/93.7 | 90.1/84.1 |
screw | 88.4/92.5 | 97.4/95.1 | 98.9/98.2 | 98.5/97.5 | 99.4/99.1 | 94.6/95.7 |
toothbrush | 94.8/91.2 | 98.2/97.9 | 98.6/98.3 | 99.0/98.7 | 99.0/99.0 | 98.4/85.2 |
transistor | 96.0/92.5 | 86.5/86.5 | 97.1/97.0 | 97.7/97.5 | 90.4/88.7 | 77.1/71.1 |
zipper | 92.0/98.1 | 96.8/96.2 | 97.9/98.7 | 97.1/98.3 | 98.2/98.8 | 96.0/73.5 |
Model AVG
| 94.8/94.1 | 96.5/93.0 | 97.7/96.5 | 97.3/96.2 | 97.4/96.3 | 94.8/89.2 |
Model STD
| 4.0/5.1 | 3.3/5.1 | 1.7/3.2 | 1.7/4.2 | 2.4/3.6 | 5.6/8.2 |
Table 9.
MVTecAD AUPRO. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 9.
MVTecAD AUPRO. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
carpet | 89.0/86.5 | 95.0/77.7 | 93.8/89.9 | 96.1/91.1 | 95.7/92.7 | 97.7/90.2 |
grid | 94.9/94.9 | 89.9/65.5 | 90.7/80.4 | 90.1/89.2 | 96.6/97.2 | 96.7/83.4 |
leather | 97.8/95.1 | 98.4/87.0 | 96.7/92.8 | 98.0/97.4 | 98.1/96.6 | 99.1/97.7 |
tile | 77.4/84.6 | 89.7/74.3 | 79.1/70.6 | 85.4/68.7 | 86.3/81.4 | 91.7/79.7 |
wood | 86.8/78.9 | 88.8/64.6 | 84.5/70.9 | 92.6/82.8 | 91.0/84.1 | 95.1/88.3 |
bottle | 89.1/90.7 | 92.4/85.6 | 92.8/90.6 | 95.2/95.0 | 95.9/93.9 | 85.4/71.8 |
cable | 75.6/85.9 | 79.4/66.8 | 91.2/88.9 | 86.4/88.7 | 90.7/78.0 | 80.4/61.7 |
capsule | 93.8/85.9 | 91.2/83.0 | 91.9/88.4 | 91.4/91.0 | 93.3/92.8 | 74.5/64.3 |
hazelnut | 95.3/92.7 | 95.3/85.6 | 93.9/92.0 | 93.4/93.7 | 96.0/94.6 | 95.3/93.4 |
metal_nut | 89.4/83.9 | 86.1/74.8 | 92.0/87.8 | 92.7/91.0 | 93.7/91.5 | 94.8/81.1 |
pill | 93.4/74.6 | 91.4/62.6 | 93.7/86.4 | 94.2/86.9 | 96.2/92.7 | 81.5/78.8 |
screw | 67.7/76.4 | 88.6/81.8 | 94.1/91.8 | 94.0/91.0 | 96.2/92.7 | 81.6/78.8 |
toothbrush | 73.6/68.6 | 84.4/78.8 | 85.5/85.1 | 93.0/92.6 | 92.5/93.1 | 88.2/46.5 |
transistor | 91.1/80.7 | 79.0/60.9 | 94.5/93.2 | 94.0/91.9 | 80.9/77.5 | 60.5/49.6 |
zipper | 77.2/94.0 | 88.7/86.0 | 92.0/94.2 | 91.3/93.9 | 95.0/95.6 | 89.2/27.4 |
Model AVG
| 86.1/84.9 | 89.2/75.6 | 91.1/86.9 | 92.5/89.7 | 93.2/90.5 | 87.4/73.2 |
Model STD
| 9.4/7.9 | 5.4/9.4 | 4.6/7.4 | 3.3/6.8 | 4.5/6.7 | 10.5/19.8 |
Table 10.
VisA image-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 10.
VisA image-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
candle | 94.0/89.6 | 90.2/76.1 | 98.3/87.0 | 91.7/76.1 | 94.7/89.1 | 76.4/66.0 |
capsules | 87.3/86.8 | 87.1/67.2 | 70.6/76.0 | 66.9/63.8 | 88.8/81.9 | 86.6/78.2 |
cashew | 92.7/90.0 | 92.7/88.5 | 96.8/96.4 | 89.1/84.4 | 95.8/93.9 | 87.3/65.8 |
chewinggum | 98.8/97.4 | 99.5/81.8 | 98.8/92.8 | 98.8/89.4 | 98.6/93.4 | 96.0/86.1 |
fryum | 96.5/95.2 | 65.1/85.1 | 95.0/95.2 | 88.1/89.2 | 88.6/94.6 | 80.4/89.9 |
macaroni1 | 94.2/87.2 | 77.1/70.7 | 87.0/83.1 | 79.9/74.7 | 96.4/92.1 | 88.3/86.3 |
macaroni2 | 87.6/79.5 | 71.2/53.3 | 69.7/63.4 | 61.4/66.3 | 82.6/82.2 | 75.0/54.1 |
pcb1 | 96.5/95.2 | 94.3/94.7 | 94.2/95.7 | 85.2/93.8 | 96.5/97.6 | 87.9/93.4 |
pcb2 | 96.5/93.6 | 84.2/83.3 | 93.9/97.0 | 82.7/85.9 | 96.3/95.5 | 86.1/82.6 |
pcb3 | 97.3/87.1 | 77.3/84.1 | 92.6/91.9 | 78.6/70.1 | 96.5/97.7 | 78.5/53.8 |
pcb4 | 99.6/94.2 | 97.1/94.5 | 99.2/98.7 | 92.9/93.9 | 99.7/99.3 | 94.0/65.8 |
pipe_fryum | 99.6/96.7 | 98.6/84.0 | 99.3/94.6 | 92.2/85.0 | 99.4/97.6 | 91.6/86.4 |
Model AVG
| 95.1/91.0 | 86.2/80.3 | 91.3/89.3 | 84.0/81.1 | 94.5/92.9 | 85.7/75.7 |
Model STD
| 4.2/5.3 | 11.3/11.9 | 10.5/10.5 | 11.0/10.5 | 5.2/5.8 | 6.8/13.9 |
Table 11.
VisA pixel-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 11.
VisA pixel-level AUROC. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
candle | 98.9/95.2 | 98.7/90.1 | 99.0/95.1 | 98.8/92.7 | 98.9/95.0 | 96.7/78.1 |
capsules | 99.0/98.1 | 97.2/79.0 | 97.6/96.9 | 95.7/93.8 | 99.6/99.0 | 99.1/96.0 |
cashew | 97.8/95.3 | 98.5/96.1 | 98.9/96.2 | 98.2/94.8 | 94.8/72.3 | 95.1/81.8 |
chewinggum | 98.6/97.6 | 98.8/93.5 | 98.8/96.0 | 99.0/94.0 | 98.6/91.8 | 98.4/94.0 |
fryum | 84.8/93.2 | 95.2/95.1 | 94.4/94.7 | 94.9/96.0 | 96.3/95.7 | 94.0/91.3 |
macaroni1 | 99.0/95.0 | 97.4/94.6 | 97.5/96.3 | 96.8/97.2 | 99.4/99.3 | 97.6/98.7 |
macaroni2 | 98.2/97.4 | 96.3/93.0 | 96.3/93.4 | 94.9/94.8 | 99.0/98.9 | 98.5/96.0 |
pcb1 | 99.5/99.3 | 99.3/99.1 | 99.5/99.3 | 99.0/99.3 | 99.7/99.6 | 99.3/99.1 |
pcb2 | 98.7/96.9 | 97.2/95.1 | 97.8/97.2 | 97.3/98.1 | 98.7/97.0 | 97.4/95.8 |
pcb3 | 98.9/97.5 | 96.7/97.2 | 98.0/98.2 | 97.2/98.2 | 99.1/99.0 | 95.9/95.7 |
pcb4 | 97.8/95.1 | 97.6/97.8 | 98.0/98.6 | 96.8/97.0 | 98.4/98.5 | 98.1/53.5 |
pipe_fryum | 96.5/98.7 | 98.5/99.0 | 98.9/99.0 | 99.0/99.1 | 98.9/99.0 | 97.3/96.9 |
Model AVG
| 97.3/96.6 | 97.6/94.1 | 97.9/96.7 | 97.3/96.3 | 98.5/95.4 | 97.3/89.7 |
Model STD
| 4.0/1.8 | 1.2/5.4 | 1.4/1.8 | 1.5/2.2 | 1.4/7.6 | 1.6/13.2 |
Table 12.
VisA AUPRO. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Table 12.
VisA AUPRO. Each cell carries the results for ImageNet/MandelbulbVAR-1k.
Class | FastFlow | C-Flow | PatchCore | PaDiM | RD | STFPM |
---|
candle | 95.1/91.6 | 93.4/78.4 | 95.1/86.7 | 94.8/83.6 | 94.3/90.9 | 91.4/61.3 |
capsules | 93.8/84.4 | 76.8/47.2 | 69.1/62.1 | 66.7/62.0 | 95.1/89.0 | 95.4/82.3 |
cashew | 84.1/80.5 | 92.8/63.5 | 90.4/60.5 | 82.0/64.8 | 89.2/58.7 | 91.7/66.9 |
chewinggum | 85.2/75.1 | 89.3/39.1 | 84.9/50.8 | 87.3/43.6 | 77.8/39.1 | 77.4/52.0 |
fryum | 74.7/70.2 | 69.5/54.3 | 78.4/66.1 | 70.3/71.5 | 88.7/84.7 | 85.2/79.9 |
macaroni1 | 94.3/84.9 | 87.7/75.1 | 89.9/84.3 | 86.5/82.9 | 92.8/92.9 | 82.8/88.3 |
macaroni2 | 94.2/89.2 | 71.9/72.1 | 87.3/77.1 | 71.4/71.4 | 91.7/91.2 | 86.4/79.8 |
pcb1 | 92.2/89.3 | 89.9/87.5 | 88.8/86.8 | 88.3/89.3 | 95.1/95.2 | 90.8/91.3 |
pcb2 | 89.1/80.4 | 81.5/74.2 | 82.6/79.6 | 77.5/85.2 | 89.1/86.7 | 81.3/82.6 |
pcb3 | 85.7/71.9 | 68.0/80.3 | 78.4/82.4 | 71.0/81.8 | 90.7/91.1 | 59.2/79.7 |
pcb4 | 85.5/66.5 | 86.8/85.7 | 86.3/85.9 | 80.5/81.0 | 89.1/88.7 | 89.6/2.5 |
pipe_fryum | 85.3/84.3 | 93.5/86.7 | 93.3/88.7 | 89.6/87.2 | 95.8/95.5 | 91.0/90.3 |
Model AVG
| 88.3/80.7 | 83.4/70.3 | 85.4/75.9 | 80.5/75.4 | 90.8/83.6 | 85.2/71.4 |
Model STD
| 6.0/8.1 | 9.6/16.0 | 7.3/12.7 | 9.1/13.3 | 4.9/17.0 | 9.7/24.8 |
Table 13.
Average image- and pixel-level AUROC express in % for PatchCore [
28] using the WideResNet-50 feature extractor on MVTec AD. The “Pre-training” column indicates which dataset has been used for pre-training, and in brackets, the number of classes in each dataset is indicated. Best and second-best scores are shown in underlined bold and bold, respectively.
Table 13.
Average image- and pixel-level AUROC express in % for PatchCore [
28] using the WideResNet-50 feature extractor on MVTec AD. The “Pre-training” column indicates which dataset has been used for pre-training, and in brackets, the number of classes in each dataset is indicated. Best and second-best scores are shown in underlined bold and bold, respectively.
Pre-Training | Image AUROC | Pixel AUROC |
---|
Random initialization | 0.772 | 0.860 |
ImangenNet (1000 cl.) | 0.991 | 0.981 |
Mandelbulbs (1000 cl.) | 0.678 | 0.784 |
MultiMandelbulbs (1000 cl.) | 0.809 | 0.919 |
MultiMandelbulbs-back (1000 cl.) | 0.719 | 0.833 |
Fractals (200 cl.) | 0.720 | 0.823 |
MultiFractals (200 cl.) | 0.771 | 0.900 |
Mandelbulbs (200 cl.) | 0.695 | 0.802 |
MultiMandlebulbs (200 cl.) | 0.817 | 0.921 |
MultiMandelbulbs-back (200 cl.) | 0.699 | 0.781 |
MultiMandelbulbs-transforms (200 cl.) | 0.791 | 0.912 |
MultiMandelbulbs-gray (200 cl.) | 0.793 | 0.908 |
Table 14.
Results of the best-performing model for each training configuration. The table shows the epoch at which the best model was selected, along with the corresponding validation accuracy and AUROC score achieved using PatchCore.
Table 14.
Results of the best-performing model for each training configuration. The table shows the epoch at which the best model was selected, along with the corresponding validation accuracy and AUROC score achieved using PatchCore.
Train Config. | Best Epoch | Best Val. Acc. | Image AUROC | Pixel AUROC |
---|
VAR1 | 83 | 9.33 | 0.857 | 0.941 |
VAR1-BATCH | 25 | 10.01 | 0.836 | 0.932 |
VAR1-noSCHEDULER | 97 | 16.69 | 0.844 | 0.931 |
Table 15.
MVTec LOCO AD image-level AUROC obtained via original code implementation of the different methods. Each cell carries the results for ImageNet/MultiMandelbulbs-small.
Table 15.
MVTec LOCO AD image-level AUROC obtained via original code implementation of the different methods. Each cell carries the results for ImageNet/MultiMandelbulbs-small.
Pre-Training | Efficient AD | PUAD | SINBAD |
---|
ImageNet | 0.898 | 0.925 | 0.841 |
Baseline | 0.773 | 0.818 | 0.733 |
VAR1 | 0.811 | 0.822 | 0.734 |
VAR1-BATCH | 0.789 | 0.837 | 0.780 |
VAR1-noSCHEDULER | 0.788 | 0.832 | 0.788 |
Table 16.
Pixel-level sPRO results on the MVTec LOCO AD dataset, obtained using the official dataset’s evaluation code. “Log.” and “Stru.” stand for logical and structural anomalies and “fpr” is the false positive rate used for the sPRO calculation.
Table 16.
Pixel-level sPRO results on the MVTec LOCO AD dataset, obtained using the official dataset’s evaluation code. “Log.” and “Stru.” stand for logical and structural anomalies and “fpr” is the false positive rate used for the sPRO calculation.
EfficientAD Pre-Training | Log. fpr = 0.05 | Stru. fpr = 0.05 | Log. fpr = 0.3 | Stru. fpr = 0.3 |
---|
ImageNet | 0.691 | 0.682 | 0.889 | 0.866 |
Baseline | 0.574 | 0.483 | 0.819 | 0.718 |
VAR1 | 0.574 | 0.518 | 0.832 | 0.756 |
VAR1-BATCH | 0.454 | 0.504 | 0.731 | 0.735 |
VAR1-noSCHEDULER | 0.484 | 0.493 | 0.752 | 0.721 |