Next Article in Journal
MM-EMOR: Multi-Modal Emotion Recognition of Social Media Using Concatenated Deep Learning Networks
Next Article in Special Issue
AI-Based User Empowerment for Empirical Social Research
Previous Article in Journal
An Empirical Study on Core Data Asset Identification in Data Governance
Previous Article in Special Issue
EnviroStream: A Stream Reasoning Benchmark for Environmental and Climate Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Contemporary Art Authentication with Large-Scale Classification

1
Department of Computer Science, University of North Carolina at Charlotte, Charlotte, NC 28223, USA
2
Department of Computer Science, Utah State University, Logan, UT 84322, USA
3
Department of Computer Science, Polish-Japanese Academy of Information Technology, 02-008 Warszawa, Poland
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2023, 7(4), 162; https://doi.org/10.3390/bdcc7040162
Submission received: 31 August 2023 / Revised: 24 September 2023 / Accepted: 7 October 2023 / Published: 9 October 2023
(This article belongs to the Special Issue Big Data and Cognitive Computing in 2023)

Abstract

:
Art authentication is the process of identifying the artist who created a piece of artwork and is manifested through events of provenance, such as art gallery exhibitions and financial transactions. Art authentication has visual influence via the uniqueness of the artist’s style in contrast to the style of another artist. The significance of this contrast is proportional to the number of artists involved and the degree of uniqueness of an artist’s collection. This visual uniqueness of style can be captured in a mathematical model produced by a machine learning (ML) algorithm on painting images. Art authentication is not always possible as provenance can be obscured or lost through anonymity, forgery, gifting, or theft of artwork. This paper presents an image-only art authentication attribute marker of contemporary art paintings for a very large number of artists. The experiments in this paper demonstrate that it is possible to use ML-generated models to authenticate contemporary art from 2368 to 100 artists with an accuracy of 48.97% to 91.23%, respectively. This is the largest effort for image-only art authentication to date, with respect to the number of artists involved and the accuracy of authentication.

1. Introduction

Proof of an artwork’s authenticity is imperative when buying or selling a piece of art. The authenticity of an artist and their work is normally accomplished through a process of documenting a certificate of authenticity, past ownership, artist signature, and other physical attributes, such as dimension, medium, and title. This process is known as the artwork’s provenance. As early as 2004, digital techniques for art authentication were conceived to augment physical authentication techniques by analyzing consistencies and inconsistencies in the first- and higher-order wavelet statistics collected from drawings or paintings. Results confirmed the proper authentication of 13 drawings and paintings, either by Pieter Bruegel the Elder or Perugino [1]. In 2017, hyperspectral imaging combined with advanced signal processing techniques correctly identified two Beltracchi forgeries by correctly classifying 78% of pigments in the forged paintings [2].
In the past five years, art authentication received increased attention due to artificial intelligence, digital image processing, forensic techniques, and legal cases. From an artificial intelligence perspective, supervised deep learning algorithms, when applied to images of paintings, have attained an accuracy of 67.78% in authenticating art for 90 artists using the WikiArt dataset [3], and an accuracy of 32.40% in authenticating art for 1199 artists using the Rijksmuseum dataset [4]. On the digital image processing front, an accuracy of 91.7% was achieved in authenticating art for two artists using the principal component analysis (PCA) and a custom van Gogh and Raphael dataset. These results involve fewer artists with the advantage of reduced resource costs [5]. An accuracy of 88% was achieved for authenticating art on an undisclosed number of artists using a decision tree on attribution markers and a custom dataset consisting of 43 authentic paintings and 12 forged paintings. It is important to note that the attribution markers consist of typical forensic metrics that are currently used by art historians for art authentication purposes in addition to markers from the painting image [6]. A similar concept to attribution markers involves the forensic technique of optical coherence tomography (OCT), which provides analysis of the cracks in paintings. Both the nature of painting cracks and the map of painting cracks for an authenticated artwork provide quick methods for determining art forgeries [7]. From a legal perspective, an art expert is used to authenticate art using methods of connoisseurship, provenance, and scientific analysis. Art experts are not legally regulated, and the methods are subject to human error. A look into the future indicates that companies, such as Art Recognition, and academic institutions, such as Rutgers University, have proprietary capabilities in detecting intentional forgeries with 80% accuracy, with respect to an undisclosed number of artists; this represents a step forward in eliminating human error with art authentication [8].
Art authentication is paramount for the value of artwork. Traditional art authentication attribute markers are expensive and time-consuming. This research is anticipated to serve as an additional attribute marker, solely from an artist’s painting images, to support art authentication, where traditional art authentication methods are inadequate or missing. This attribute marker is used for any artist in the model as a binary attribute. For the prediction of a piece of art for an artist in question, a successful prediction provides a favorable outcome for one artist and unfavorable outcomes for the remaining 2367 artists. Both the accuracy of the prediction and the number of artists being considered are important to indicate that the art in question is properly attributed to the artist in question and not another artist in the model.

2. Materials and Methods

Figure 1 demonstrates the process for the creation of an image-only art authentication attribute marker to model 2368 artists. When the process begins, images are partitioned into training, validation, and test sets. To learn the model, the process trains for up to 30 epochs. An epoch is a learning event that includes all paintings in the training set. In each epoch, training paintings are shuffled and mutated to prevent overfitting the model, and the artist’s style is gradually learned in batches. At regular intervals, the model is validated using validation paintings, and the results of learning validation make changes to the model, which are used in the next iteration. Once validation results meet a threshold or 30 epochs have passed, the process stops with the current state of the model. This model is used on test paintings to determine the artist. The results of this test produce a confusion matrix, which is a table showing true negatives/positives, false negatives, and false positives. True negatives/positives indicate that the model made a correct negative or positive prediction with respect to the artist and painting in question. False negatives indicate that the model predicted another artist instead of the actual artist. False positives indicate that the model predicted the actual artist, but it should have predicted another artist instead. There are a variety of metrics that can be calculated from this confusion matrix. Equation (3) is used for the primary metric due to the imbalanced nature of painting datasets.

2.1. Machine Learning Development and Evaluation

An ML model is developed to predict an artist by training a model using the state-of-the-art ResNet algorithm, to learn relationships between input painting images and corresponding artists that have been labeled manually by visual inspection. Hyperparameters listed in Table 1 are used to continually validate the model being generated until a desired result is achieved or max validation steps occur. The resulting model is evaluated on an unseen test dataset, which is the 20% test partition discussed in the next section.

2.2. Training, Validation and Testing Datasets

An image dataset, sourced from Artfinder, encompasses over 212,570 paintings and is utilized for training, validation, and test sets. Each image adheres to a minimum size of 1200 px × 1200 px and an sRGB color profile is used for training, validation, and test sets. Images are resized using bilinear interpolation into a 224 × 224 × 3 tensor. The training set is randomly generated from 70% of the images and the validation set is randomly generated from 10% of the images. The remaining 20% of images are set aside for testing after the model is trained. An epoch is defined as a single pass through the entire dataset, with images processed in batches of 130–140 images, up to 1050 times. Up to 30 epochs of training and validation occur in a cross-folded fashion every 50 iterations. To mitigate over-fitting concerns, images are shuffled, rotated between −90 and 90 degrees, randomly scaled between 1 and 2 times, and undergo random reflection on the x-axis with each epoch. The process of mutating images in this manner is known as data augmentation and it helps to prevent the network from overfitting and memorizing the exact details of the training images [9]. Transfer learning, given a model from the ImageNet project, is used for additional initialization of training parameters [10]. The ImageNet transfer learning model is used as a starting point for our model. Without this starting point, in past experiments, we found that results took longer to converge and the final accuracy suffered [4].

2.3. Model Selection

Existing work from the literature review is leveraged to select a model for training. Features extracted from images using SIFT, HOG, and other digital image processing algorithms consumed by basic ML algorithms, such as support vector machines (SVMs), decision trees, and the k-nearest neighbor algorithm (k-NN) train models quickly, but accuracy starts to suffer quickly as approximately fifty artist classes are approached. Deep neural networks remedy this limitation at the cost of the training time and the need for a high-performance cluster to generate the model. Of these networks, ResNet 101 outperforms earlier versions of ResNet as well as AlexNet, VGG, GoogLeNet, PigeoNet, and CaffeNet [11,12,13,14,15,16,17]. Moreover, there may be some performance improvements with SENet and deeper versions of ResNet. However, the scope of this work is not to perform a detailed model comparison or improve upon a model that is working well. The scope of this work is to leverage state-of-the-art models on a very large contemporary art dataset so we can compare these results with existing experiments for fitness and scalability. Therefore, ResNet 101—with an annealing process—is used to produce the models.

2.4. ResNet Architecture

The ResNet algorithm solves the exploding and vanishing gradient problem of deep neural networks with a deep residual learning framework, which allows for much deeper networks using the concept of skip connections [18]. A mathematical proof demonstrates how skip connections can largely circumvent the exploding and vanishing gradient problem [19]. ResNet works well with classifying art because the deep network enables multiple passes on an artist’s body of work at varying filter sizes, in a generic manner. This process produces a model that does a very good job of learning an artist’s style.

2.5. Artist Selection

Twenty-four experiments are performed to discover how classification metrics and artist style evolve as the number of artists is reduced. The first experiment is seeded with artists having 10 or more artworks. This provides 2368 artists for the first experiment. The next experiment consists of 2300 artists, and the process continues by reducing the artist count by 100, continuing until the final experiment, which involves 100 artists. The artist selection criteria used to determine which artists will be dropped are based on macro-balanced accuracy. Macro-balanced accuracy is used over micro-balanced accuracy because the metric provides a more granular selection for the fitness of an artist, which results in fewer ties [4].

2.6. Evaluation Using the Testing Set

In a final effort to determine the fitness of the model, it is tested on 20% of the paintings split out before training to assure the model has never seen these paintings. Since the number of paintings produced by artists is naturally unbalanced, and a true representation of the artist model without over- and under-sampling is desired, macro-balanced accuracy is calculated from the confusion matrix produced by the test [20,21]. The accuracy ranges from 48.97% for the largest experiment to 91.23% for our smallest experiment. The test accuracy is approximately equal to validation accuracy in all experiments. This indicates the model is not subject to overfitting concerns. Validation accuracy is calculated by Matlab with each validation iteration, which takes into consideration ROC analysis.

2.7. Limitations with Image-Based Art Authentication

Several limitations exist with performing art authentication with painting images alone. First and foremost, it is difficult to acquire data. Both physical and online art galleries protect image data because the image is the primary proprietary asset for sale. Access to the complete collection of an art gallery for research purposes requires a trusted relationship with the gallery or a legitimate method of crawling the gallery’s online website for image data. Second, there is a varying number of paintings produced by artists, which naturally leads to imbalanced data. The task of gathering more data samples is difficult because the time it takes for an artist to produce new works is nondeterministic. From a sampling perspective, undersampling is not desired because the model does not have the opportunity to learn more about an artist’s paintings, and oversampling is not desired because a true representation of the artist’s body of work is not obtained. Therefore, metric calculations are used to acquire meaningful multi-class metrics from the tests that assume input classes are not in balance [20,21]. Third, there are attribution markers other than a digital representation of the painting used when authenticating a painting. These markers have traditionally been used by art historians for art authentication. Over 30 attribution markers are discussed in state-of-the-art research dealing with art authentication. For example, there are markers corresponding to the UV, IR, and X-ray physical analysis of a painting. Markers characterizing the pigments and medium characteristics of the artist and time period are considered. Moreover, there are markers that have nothing to do with the actual image, such as signature and ownership documents, as well as history [6]. Fourth, there are no paintings representing true negatives on purpose in experiments.

Data Source

The data for experiments come from Artfinder, which is an online art marketplace. Raw data of artwork images were thoroughly reviewed to ensure sound data for experiments. Data from this website were collected over several years via automated web crawling technology [22]. Permission was given to use these data in this research and report aggregate results only. Specific artist names are hidden in this research (hence, the omission of the artist’s name and painting images). Upon request, data are available for research verification and extension.

3. Results

High-level results for experiments are listed in Table 2. This table represents all 24 experiments, starting with 2368 artists and ending with 100 artists. Validation accuracy (Val Acc) is the accuracy obtained during training. Test accuracy (Test Acc (M)) is the primary metric of interest and is the calculated macro-balanced accuracy of the test paintings that were not observed during training. Test accuracy is a bit higher than validation accuracy, which indicates the model did not encounter any overfitting issues during training. The number of paintings observed during the 70/10/20 split is represented by Train/Val/Test Cnt, respectively. The batch size of images used during each iteration of training is represented in the ‘Batch’ column and the total number of Iterations per epoch is represented by the ‘Iterations’ header. With each experiment, the average number of artworks per artist increases proportionally. It is important that this number is increasing to ensure that the model is not influenced by an artist with a more limited number of artworks. Specifically, the average starts at 18 paintings per artist for 2368 artists and ends with 41 paintings per artist for 200 artists. The average dips down to 28 for the last experiment of 100 artists. The first four metrics discussed in this section are listed in Table 2, and the remaining metrics are dropped due to importance and limited space.

3.1. Confusion Matrix

The confusion matrix in Figure 2 represents the largest experiment. Due to the large number of artist classes, a pixel-based confusion matrix is used. The intensity color of each pixel represents the strength of the metric. The diagonal from the upper left-hand corner to the lower right-hand corner in the confusion matrix represents correct predictions in the form of true negative and positive predictions. A distinct, visible diagonal is a favorable condition for the confusion matrix as this will likely indicate a favorable accuracy metric. Horizontal pixels represent false positives and vertical pixels represent false negatives. These pixels are barely visible; this is a favorable condition because it indicates a failed prediction, which minimizes the negative impact on the accuracy metric.
The confusion matrix is also partitioned by the primary art styles of the artists represented. This provides a method to determine which styles are confused. The primary art style for an artist is determined by the largest count of paintings of a given style for the artist. The first alphabetical style is used for ties. For example, “artist 1004” has the following painting styles by count: Impressionistic (13), Expressive and gestural (3), Urban and Pop (3), Abstract (2), Geometric (1), Organic (1), and Photorealistic (1). Therefore, ‘Impressionistic’ is attributed to “artist 1004”. The name “artist 1004” is used because the agreement with Artfinder, the provider of the data, is to keep the artist’s name anonymous.

3.2. Accuracy

The typical average accuracy calculation from the confusion matrix shown in Equation (1) cannot be used because it applies to binary classification [21]. Several techniques are combined from state-of-the-art multi-classification performance measure research to arrive at Equation (3). This equation represents macro-balanced accuracy, which provides reasonable accuracy because it prevents unbalanced majority and minority classes from influencing the overall accuracy [20,21]. The corresponding micro-balanced accuracy shown in Equation (2) is also available. This equation reduces to the average multi-classification recall calculation over all artists; thus, it is not used, even though it provides a better number for reporting. If the data were balanced on the front end of experiments, this metric would be legitimate and would converge with macro-balanced accuracy [20]. Micro-balanced accuracy is reported to demonstrate that it coincides with validation accuracy, which is unbalanced. This demonstrates that the model is not overfitting.
AvgAcc = i = 1 l t p i + t n i t p i + t n i + f n i + f p i l
BalAcc μ = i = 1 l t p i i = 1 l t p i + f n i = i = 1 l t p i Total Predictions
BalAcc M = i = 1 l t p i t p i + f n i l

4. Discussion

4.1. Multiclass Classifier as Binary Classifier

In Table 3, the artist count is inversely proportional to validation and test accuracy. Given the state-of-the-art research using multi-classification for image-only art authentication, this behavior is expected [3,4,23]. With these experiments, the goal is to reproduce the counterintuitive phenomenon wherein many classes can improve multi-classification metrics as the number of classes grows [24]. While this phenomenon was not observed, multi-classification for a binary classification art authentication problem is important because training multiple binary classifiers for the artists of interest is not required [25]. Moreover, training a model on more than one artist produces a model of an artist’s paintings in addition to what is not considered a painting by the artist in question. The overall model accuracy is reduced in these situations, but a true positive provides more information about the artists to whom the painting in question does not belong. To show this concept in these experiments, consider artist1051, which exists in all experiments. In 16 of the experiments, including the experiment with the most and least artists, the model predicts the artist with 100% accuracy with the test painting data. In three of the experiments, the model predicts the artist with 87.50% accuracy with the test painting data. In five of the experiments, the model predicts the artist with 85.71% accuracy with the test painting data. Given that the accuracy is high in all experiments, the test with the most artists is more meaningful because there are many other potential artists in the model that could confound the test.

4.2. True Negatives

Adding purposeful true negatives to experiments would be an interesting addition. This could be accomplished by adding a true negative in the form of a contemporary art forgery. Producing art forgeries is difficult because forgery paintings are difficult to acquire due to the obfuscation of the forgery and the rareness of the forgery event. It is also cost-prohibitive to commission forgeries due to constraints on time and money. True negatives could also be accomplished by keeping a random sample of paintings from the artists that were removed after each experiment’s annealing process. While changing the process to include true negatives from the previous experiments is a straightforward task, it is a time-prohibitive task to retrain the models.

4.3. Contemporary Art Performance

Performance with the Artfinder contemporary art dataset outperforms previous experiments with historical art datasets from WikiArt and Rijksmuseum. In Table 3, accuracy results are compared using the same macro-balanced accuracy metric. In all cases, the Artfinder experiments outperform WikiArt and Rijksmuseum experiments 10+% [3,4]. This increase in accuracy for contemporary art may be because experiments start with over twice as many artists, and the annealing process can select the best artists for classification once reaching a comparable artist count in previous experiments. It could also be because contemporary art has progressed from historical art in a way that provides more opportunities to learn artistic style.

4.4. Artist Style

For all experiments, Figure 3 shows the percentage of artist styles represented. This is important to show because it indicates a variety of art styles are represented (from the first experiment with 2368 artists to the last experiment with 100 artists). The evolution of painting style representations demonstrates that models do not favor a specific art style.

4.5. Uniqueness

This research aims to maximize the number of artists and the accuracy in these experiments. Maximizing both numbers yields the best accuracy with respect to as many artists as possible. It is helpful to have one metric that considers both numbers. Therefore, a uniqueness score is used as a metric to analyze experiments in terms of a ratio between the accuracy, artist count, and the number of paintings. The number of artists per painting, referred to as artist density, is taken into consideration. The artist’s painting density is chosen as the numerator and accuracy as the denominator because the focus is on how the artist painting density is distributed in relation to accuracy. In this manner, artist painting density approaches accuracy, in which case, the relative paintings uniquely define the artist. Uniqueness is calculated as the ratio of the artist painting density to accuracy, as seen in Equation (4). As hoped, the uniqueness score is not always proportional to the artist count in the experiments, as seen in Table 4; the ideal number of artists with respect to accuracy is 300. Therefore, an ideal artist count of 300 is recommended for model generation.
Uniqueness = Artist Count Painting Count Accuracy

5. Conclusions

In this paper, we contribute toward art authentication research using contemporary art from the Artfinder dataset by applying a performance annealing residual neural network to produce 24 baselines for the dataset. We demonstrate that it is possible to use ML-generated models to authenticate contemporary art from 2368 to 100 artists with an accuracy of 48.97% to 91.23%, respectively. We also provide a method to address very large confusion matrices that consider style and a uniqueness calculation, which provides a method to choose an optimal artist count for model generation.

Future Work

In future work, we would like to continue to explore new art datasets as they become available. We would like to compare these results with the results of the dataset in this paper and collaborate with artists to provide more detailed analyses for the art community. We also believe that exploring adversarial attacks on art authentication and investigating saliency maps in art authentication will be of interest.

Author Contributions

Conceptualization: T.D. and Z.R.; Data curation: T.D. and Z.R.; Formal analysis: T.D.; Funding acquisition: T.D. and Z.R.; Investigation: T.D.; Methodology: T.D.; Project administration: T.D.; Resources: Z.R.; Software: T.D.; Supervision: T.D. and Z.R.; Validation: T.D.; Visualization: I.C., T.D. and A.-A.-R.N.; writing—original draft preparation: T.D.; Writing—review & editing: I.C., T.D., A.-A.-R.N. and Z.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially supported by the National Science Foundation under grant IIP 1749105.

Data Availability Statement

The data and code for the training, validation, and test classes are available on GitHub via https://github.com/btdobbs/Contemporary-Art-Authentication-with-Large-Scale-Classification (accessed on 6 October 2023).

Acknowledgments

We would like to acknowledge the Technology Solutions Office (TSO) group at the University of North Carolina’s Charlotte College of Computing and Informatics for their assistance with this research.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLmachine learning
SVMsupport vector machine
k-NNk-nearest neighbor
PCAprincipal component analysis
OCToptical coherence tomography

References

  1. Lyu, S.; Rockmore, D.; Farid, H. A digital technique for art authentication. Proc. Natl. Acad. Sci. USA 2004, 101, 17006–17010. [Google Scholar] [CrossRef] [PubMed]
  2. Polak, A.; Kelman, T.; Murray, P.; Marshall, S.; Stothard, D.J.; Eastaugh, N.; Eastaugh, F. Hyperspectral imaging combined with data classification techniques as an aid for artwork authentication. J. Cult. Herit. 2017, 26, 1–11. [Google Scholar] [CrossRef]
  3. Dobbs, T.; Benedict, A.; Ras, Z. Jumping into the artistic deep end: Building the catalogue raisonné. AI Soc. 2022, 37, 873–889. [Google Scholar] [CrossRef]
  4. Dobbs, T.; Ras, Z. On art authentication and the Rijksmuseum challenge: A residual neural network approach. Expert Syst. Appl. 2022, 200, 116933. [Google Scholar] [CrossRef]
  5. Leonarduzzi, R.; Liu, H.; Wang, Y. Scattering transform and sparse linear classifiers for art authentication. Signal Process. 2018, 150, 11–19. [Google Scholar] [CrossRef]
  6. ydżba-Kopczyńska, B.I.; Szwabiński, J. Attribution Markers and Data Mining in Art Authentication. Molecules 2021, 27, 70. [Google Scholar] [CrossRef] [PubMed]
  7. Kim, S.; Park, S.M.; Bak, S.; Kim, G.H.; Kim, C.S.; Kim, C.E.; Kim, K. Advanced Art Authentication in Oil Paintings Using Precise 3D Morphological Analysis of Craquelure Patterns. Res. Sq. 2021. preprint. [Google Scholar] [CrossRef]
  8. Mar, B. Experts’ Role in Art Authentication; Seton Hall University eRepository: South Orange, NJ, USA, 2021. [Google Scholar]
  9. The MathWorks Inc. Optimization Toolbox Version: 9.4 (R2022b); The MathWorks Inc.: Natick, MA, USA, 2022. [Google Scholar]
  10. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Kai, L.; Li, F.-F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  11. Viswanathan, N. Artist Identification with Convolutional Neural Networks. In Standford193CS231N Report; Art Computer Science: Stanford, CA, USA, 2017. [Google Scholar]
  12. Chen, J. Comparison of Machine Learning Techniques for Artist Identification. In Standford193CS231N Report; Art Computer Science: Stanford, CA, USA, 2018. [Google Scholar]
  13. Cetinic, E.; Lipic, T.; Grgic, S. Fine-tuning Convolutional Neural Networks for fine art classification. Expert Syst. Appl. 2018, 114, 107–118. [Google Scholar] [CrossRef]
  14. Mensink, T.; van Gemert, J. The Rijksmuseum Challenge: Museum-Centered Visual Recognition. In Proceedings of the International Conference on Multimedia Retrieval. Association for Computing Machinery, ICMR ’14, Glasgow, UK, 1–4 April 2014. [Google Scholar] [CrossRef]
  15. van Noord, N.; Hendriks, E.; Postma, E. Toward Discovery of the Artist’s Style: Learning to recognize artists by their artworks. IEEE Signal Process. Mag. 2015, 32, 46–54. [Google Scholar] [CrossRef]
  16. Strezoski, G.; Worring, M. OmniArt: Multi-task Deep Learning for Artistic Data Analysis. arXiv 2017, arXiv:1708.00684. [Google Scholar]
  17. Kondo, K.; Hasegawa, T. CNN-based Criteria for Classifying Artists by Illustration Style. In Proceedings of the 2020 2nd International Conference on Image, Video and Signal Processing, Singapore, 20–22 March 2020; pp. 93–98. [Google Scholar] [CrossRef]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  19. Philipp, G.; Song, D.; Carbonell, J.G. The exploding gradient problem demystified-definition, prevalence, impact, origin, tradeoffs, 363 and solutions. arXiv 2017, arXiv:1712.05577. [Google Scholar] [CrossRef]
  20. Grandini, M.; Bagli, E.; Visani, G. Metrics for Multi-Class Classification: An Overview. arXiv 2008, arXiv:2008.05756. [Google Scholar]
  21. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  22. Powell, L.; Gelich, A.; Ras, Z.W. How to raise artwork prices using action rules, personalization and artwork visual features. J. Intell. Inf. Syst. 2021, 57, 583–599. [Google Scholar] [CrossRef]
  23. Dobbs, B.T.; Ras, Z.W. Machine Learning Approach to Art Authentication. In Encyclopedia of Data Science and Machine Learning; IGI Global: Hershey, PA, USA, 2023; pp. 1500–1513. [Google Scholar] [CrossRef]
  24. Abramovich, F.; Pensky, M. Classification with many classes: Challenges and pluses. J. Multivar. Anal. 2019, 174, 104536. [Google Scholar] [CrossRef]
  25. Honeine, P.; Noumir, Z.; Richard, C. Multiclass classification machines with the complexity of a single binary classifier. Signal Process. 2013, 93, 1013–1026. [Google Scholar] [CrossRef]
Figure 1. Process for image-only art authentication model attribute marker. p r i denotes paintings in the training set. p v i denotes paintings in the validation set. p t i denotes paintings in the test set. Paintings in the training, validation, and test sets are mutually exclusive.
Figure 1. Process for image-only art authentication model attribute marker. p r i denotes paintings in the training set. p v i denotes paintings in the validation set. p t i denotes paintings in the test set. Paintings in the training, validation, and test sets are mutually exclusive.
Bdcc 07 00162 g001
Figure 2. Pixel-based confusion matrix of the largest experiment.
Figure 2. Pixel-based confusion matrix of the largest experiment.
Bdcc 07 00162 g002
Figure 3. Artist Style Evolution.
Figure 3. Artist Style Evolution.
Bdcc 07 00162 g003
Table 1. ResNet Hyper Parameters Used for All Experiment Results.
Table 1. ResNet Hyper Parameters Used for All Experiment Results.
ParameterValuePurpose
Image Size224 × 224 × 3Resize to match network input
Training70%Baseline value
Validation10%Baseline value
Test20%Baseline value; performance measure source
Image Rotationrandomprevent overfitting
Image Scalingrandomprevent overfitting
Image Reflectionrandomprevent overfitting
Image Batch128Based on total image count and available resources
Maximum Epochs30Training Governor
Validation50 iterationsTraining Governor
Image shuffleeach epochHandles indivisible image partition
Initial Input WeightImageNet TLInitial weights for neural network
SolverSGDMAlgorithm that updates weights and biases to minimize the loss function
Learning Rate0.01Tuned to ensure training does not take too long or results do not diverge
Momentum0.9Parameter contribution of the previous iteration to the current iteration
Weight Decay Regularization0.0001Reduces overfitting
Table 2. Experiments Results.
Table 2. Experiments Results.
ArtistsVal AccTest Acc
( μ )
Test Acc
(M)
236867.62%65.33%48.97%
230068.09%66.02%50.93%
220068.67%67.20%52.88%
210069.15%67.63%54.84%
200069.71%68.37%57.35%
190070.49%68.95%59.35%
180071.42%70.23%61.05%
170072.49%71.47%63.66%
160073.29%72.76%65.34%
150074.29%73.41%66.80%
140075.76%74.41%68.34%
130076.66%75.93%70.51%
120077.81%77.43%71.77%
110078.83%78.46%74.01%
100079.59%79.57%75.40%
90081.34%81.57%77.20%
80082.49%82.35%78.36%
70083.75%83.35%80.33%
60085.59%85.71%82.66%
50086.46%86.85%83.60%
40088.15%88.51%85.47%
30091.11%91.30%88.88%
20093.17%93.36%91.15%
10096.20%96.29%91.23%
Table 3. Dataset accuracy comparison.
Table 3. Dataset accuracy comparison.
ArtistsArtfinder AccWikiArt AccRijks Acc
120071.77%n/a32.40%
100075.40%n/a40.51%
40085.47%n/a58.60%
30088.88%n/a46.70%
20091.15%n/a81.66%
10091.23%72.96%72.69%
Table 4. Uniqueness.
Table 4. Uniqueness.
Artists CountUniqueness Score
236811.37%
230010.76%
220010.04%
21009.32%
20008.57%
19008.00%
18007.53%
17006.99%
16006.58%
15006.19%
14005.82%
13005.56%
12005.29%
11004.85%
10004.55%
9004.24%
8003.99%
7003.75%
6003.41%
5003.12%
4002.95%
3002.59%
2002.68%
1003.99%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dobbs, T.; Nayeem, A.-A.-R.; Cho, I.; Ras, Z. Contemporary Art Authentication with Large-Scale Classification. Big Data Cogn. Comput. 2023, 7, 162. https://doi.org/10.3390/bdcc7040162

AMA Style

Dobbs T, Nayeem A-A-R, Cho I, Ras Z. Contemporary Art Authentication with Large-Scale Classification. Big Data and Cognitive Computing. 2023; 7(4):162. https://doi.org/10.3390/bdcc7040162

Chicago/Turabian Style

Dobbs, Todd, Abdullah-Al-Raihan Nayeem, Isaac Cho, and Zbigniew Ras. 2023. "Contemporary Art Authentication with Large-Scale Classification" Big Data and Cognitive Computing 7, no. 4: 162. https://doi.org/10.3390/bdcc7040162

APA Style

Dobbs, T., Nayeem, A. -A. -R., Cho, I., & Ras, Z. (2023). Contemporary Art Authentication with Large-Scale Classification. Big Data and Cognitive Computing, 7(4), 162. https://doi.org/10.3390/bdcc7040162

Article Metrics

Back to TopTop