Next Article in Journal
Using Macro- and Microscale Preservation in Vertebrate Fossils as Predictors for Molecular Preservation in Fluvial Environments
Previous Article in Journal
pH-Dependent Compaction of the Intrinsically Disordered Poly-E Motif in Titin
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Classification of Amazon Parrots by Deep Learning: A Potentially Useful Tool for Wildlife Conservation

Department of Biotechnology, Sangmyung University, Seoul 03016, Korea
*
Author to whom correspondence should be addressed.
Biology 2022, 11(9), 1303; https://doi.org/10.3390/biology11091303
Submission received: 14 July 2022 / Revised: 8 August 2022 / Accepted: 29 August 2022 / Published: 1 September 2022

Abstract

:

Simple Summary

Most parrot species are threatened with extinction because of habitat loss and commercial trade. Parrot conservation is vital because parrots play an important role in the ecosystem. The Amazon parrots are one of the most endangered parrot species. Monitoring their wild population and global trade is essential for their conservation. However, this is becoming more challenging because it requires manual analysis of large-scale image data. Furthermore, the morphological identification of the Amazon parrots can be difficult because they have similar morphological features. Deep learning-based object detection models are useful tools for monitoring wild populations and global trade. In this study, 26 Amazon parrot species were classified using eight object detection models. The object detection model, which showed the highest accuracy, classified the 26 Amazon parrot species at 90.7% on average. The continuous development of deep learning models for classifying Amazon parrots might help to improve the ability to monitor their wild populations and global trade.

Abstract

Parrots play a crucial role in the ecosystem by performing various roles, such as consuming the reproductive structures of plants and dispersing plant seeds. However, most are threatened because of habitat loss and commercial trade. Amazon parrots are one of the most traded and illegally traded parrots. Therefore, monitoring their wild populations and global trade is crucial for their conservation. However, monitoring wild populations is becoming more challenging because the manual analysis of large-scale datasets of images obtained from camera trap methods is labor-intensive and time consuming. Monitoring the wildlife trade is difficult because of the large quantities of wildlife trade. Amazon parrots can be difficult to identify because of their morphological similarity. Object detection models have been widely used for automatic and accurate species classification. In this study, to classify 26 Amazon parrot species, 8 Single Shot MultiBox Detector models were assessed. Among the eight models, the DenseNet121 model showed the highest mean average precision at 88.9%. This model classified the 26 Amazon parrot species at 90.7% on average. Continuous improvement of deep learning models classifying Amazon parrots may support monitoring wild populations and the global trade of these species.

1. Introduction

Parrots (order: Psittaciformes) play an important role in the ecosystem as consumers of the reproductive structures of plants [1]. They disperse seeds through external transport using their beaks and feet and via internal transport through feeding and excretion [1]. Parrots pollinate plants and protect them by feeding on plant-based parasites [2]. Most parrot species are threatened because by habitat loss [3] and the pet trade [4]. Particularly, parrots belonging to the family Psittacidae are reported to be one of the most traded birds [5]. Parrot conservation is important to preserve the ecosystem of their habitats. Therefore, various international conventions and conservation bodies, such as the Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES) and the International Union for the Conservation of Nature and Natural Resources (IUCN), aim to protect parrots from extinction and illegal trade. Among Psittacidae, Amazon parrots (genus Amazona), which are neotropical with a distribution from northern Mexico to much of South America through Mesoamerica and the Caribbean, are the most diverse parrot group, including 35 species [6,7,8]. According to the IUCN Red List of Threatened Species, three species are listed as “Critically Endangered,” six as “Endangered,” and nine as “Vulnerable” [9]. The population size of 27 wild Amazon parrot species has been decreasing [9]. According to the CITES checklist, 16 Amazon parrots are included in Appendix I, which means that their trade is prohibited [10]. Additionally, the Amazon parrots are one of the most traded parrots [11]. This high demand has made Amazon parrots one of the most illegally traded parrots [12]. Indeed, an average of 12,000 parrots in the Amazon region were exported annually to various countries [13], with the orange-winged Amazon parrot (Amazona amazonica) being the most exported species [13]. Following a European Union ban on the import of wild birds because of health and welfare risks in 2007 [14], the import of parrots to Asian countries, including Korea, rapidly increased [11,13]. According to the National Institute of Biological Resources, parrots, including Amazon parrots, are the most imported animals in Korea [15].
Monitoring wild populations is crucial for wildlife conservation. The camera trap method, used widely to monitor wildlife populations in the recent past [16], involves manual analysis to morphologically identify species using a large image dataset [16]. However, this is becoming more challenging because manually handling large-scale data is labor intensive and time consuming [16,17]. Furthermore, monitoring and controlling the wildlife trade are essential to conserving wildlife [18]. Therefore, the identification of species being traded should be first conducted [5,19]. Species identification of wildlife based on morphological features is a standard and effective method [19,20]. However, because of the large-scale trade in wildlife, the rapid and accurate identification of wildlife by morphological features is a challenge [21]. Additionally, the decline in the number of qualified morphological experts makes monitoring wild populations and global trade more difficult [19,22]. Particularly, parrots can be extremely difficult to identify and are sometimes misidentified during trading [7]. Amazon parrots are characterized by green bodies, with variable colors, dominantly red, yellow, white, and blue, on the head, breast, shoulders, and flight feathers [6,7,8]. Similar color combinations on their body can cause misidentification of the species [6,7]. Identifying some Amazon parrots can be challenging because of their similar morphological features [23,24]. These make it difficult to identify traded Amazon parrots based on morphological features. To overcome the limitations of morphological identification, DNA analysis methods, such as DNA barcoding, have been used to identify wildlife [25,26]. Although species identification via DNA barcoding is accurate, this tool is expensive; requires sample preparation from the feathers, hair follicles, feces, etc., and it is difficult to perform in situ [27,28]. Therefore, a method is needed for the rapid and accurate identification of Amazon parrots based on morphological features.
Image classification based on deep learning is potentially useful for enhancing the ability to monitor wildlife populations [28,29,30] and the wildlife trade [21,31]. Convolution neural networks (CNNs) are deep learning methods that were developed for image classification [32]. Object detection models, such as Faster R-CNN [33], You Only Look Once (YOLO) [34], and Single Shot MultiBox Detector (SSD) [35], were developed based on CNNs; they consider not only classification but also regression, which predicts objects in images. Among object detection models, two-stage detectors, such as Faster R-CNN, learn regression and classification separately and continuously, whereas one-stage detectors, such as YOLO and SSD learn regression and classification simultaneously. Hence, the one-stage detector processes data faster than the two-stage detector. Moreover, SSD shows faster data processing speed and accuracy than YOLO because SSD performs regression and classification using multiple feature maps from a CNN network, whereas YOLO performs those using the last feature map from the network [34,35]. Because of these advantages, the SSD model has been widely applied to species classification [36,37,38]. However, these studies used the SSD model to classify species belonging to different taxonomic groups, such as genus, family, and order. By contrast, in this study, we have applied the SSD models to classify species belonging to the same genus. This can be more challenging because species belonging to the same genus can normally be difficult to morphologically identify because of their similar morphological features.
The conservation of the Amazon parrots is essential for preserving the ecosystem of geographical regions ranging from northern Mexico to South America, including the Amazon region, which is considered the Earth’s lungs. However, very few studies have been performed on automatic and accurate species classification. In this study, an object detection model, SSD, using eight CNNs as backbone networks, was assessed to classify 26 Amazon parrot species. The application of deep learning to monitor the wild populations and global trade of Amazon parrots can assist in the conservation of this species.

2. Materials and Methods

2.1. Collection of Images

The images of 35 adult Amazon parrots were collected from the Internet (www.google.com accessed on 10 March 2022) because there was no standard dataset for these species. The image collection from the Internet has been used to establish a dataset for deep learning to obtain images of various individuals with diverse backgrounds when a standard dataset is not available [37,39]. For comprehensive image collection, the species and common names were used as keywords. The images were collected at the species level due to a lack of images at the subspecies level. Images collected from the Internet were identified using the morphological features of each species extracted from three books classifying parrot species written by experts. [6,7,8]. Images that could not accurately identify species were removed. Species with more than 100 images collected from the Internet were included in this study. Nine species were excluded because the number of images required to train the deep learning models was insufficient. Among the 26 Amazon parrot species, Amazona albifrons and Amazona viridigenalis show gender dimorphism [6,7,8]; therefore, images of both males and females of these species were included. Initially, 5968 images of the 26 species were included (Table S1), unified as 300 × 300 pixels images, which was the size required by the object detection model used in the study. Because the morphological features that classify the Amazon parrots are located all over their bodies, such as the colors of the crown, scapulars, and tail feathers, the whole body was labeled as a ground-truth bounding box using DarkLabel [40]. Then, the dataset was separated randomly into 70% of the training set, 15% of the validation set, and 15% of the test set. Data augmentation methods of horizontal flip, rotation, zoom-in, zoom-out, and transformation were applied to the training set to prevent overfitting (Figure 1). The horizontal flip method was applied once per image, and the other four methods were applied with different ranges to make the training set of each species > 10,000 images (Table S1). The rotation method was applied randomly between −10° and 10°; zoom-in and zoom-out methods were applied randomly from 100% to 200% and 50% to 100% of image sizes, respectively; and horizontal and vertical transformation method was applied randomly between −30 and +30 pixels. Additionally, the images generated through augmentation were removed if the object region in the images was out of the image range of 300 × 300 pixels. A final dataset was established containing 268,684 images for the training set, 867 images for the validation set, and 905 images for the testing set (Table 1).

2.2. Training of Deep Learning Models

The SSD model [36] was used to classify the 26 Amazona species (Figure 2). A predicted bounding box was identified using a feature map extracted from the backbone network and a feature map extracted using a bottleneck structure in the SSD model. Regression and classification were applied using a convolutional layer on a multiscale feature map. Additionally, the SSD used the Faster R-CNN anchor box concept [33], which creates a default box with various scales and aspect ratios and maps it to a multiscale feature map to apply the regression and classification functions. VGGNet with 16-layer [41]; three ResNet with 18-, 34-, and 50-layer [42]; and four DenseNet with 18-, 30-, 50-, and 121-layer [43] architectures were used as backbone networks in the SSD model. Tables S2–S4 present the structures of the CNN models. The experimental platform of these models is based on the Ubuntu 20.04 operating system, which uses two Intel Xeon Silver 4110 CPUs (Intel Inc., Santa Clara, CA, USA), RTX 2080 Ti Graphics with 11G video memory, and four 16 GB of REG.ECC DDR4 SDRAMs. The experimental program is based on Python 3.9.7 and runs on the PyCharm2021.1 software with the Keras–TensorFlow environments. The Keras Early Stop function was used to prevent overfitting.

2.3. Evaluation of Model Performances

The average precision (AP) values for each class were calculated from a precision–recall curve obtained by the measures of precision (true positive/true positive + false positive) and recall (true positive/true positive + false negative). Intersection over Union (IoU) was used to define true positives and the ratio intersection and union of the ground-truth bounding boxes labeled by hand and predicted bounding boxes suggested by the model. The model’s prediction was considered a true positive when the IoU was more than the threshold determined by the researcher. In this study, the threshold of the IoU was determined at 0.5 [44,45]. Finally, the mean AP (mAP) value was used to evaluate the performance of the model using Formula (1), where Q is the number of queries of the dataset, and AP(q) is the AP for the given query q.
mean   Average   Precision   mAP = q = 1 Q AP q Q
Additionally, the model inference time was calculated as the time to process a single image. The classification results of the models are shown using the confusion matrix. The classification result with the highest confidence value was chosen when models predicted multiple classification results.

3. Results

Four prediction results were obtained for the eight models (Figure 3). The prediction result with the highest confidence score was used for the classification result with multiple prediction bounding boxes. In Figure 3A, one prediction bounding box was predicted and classified correctly. In one case, multiple bounding boxes were predicted, and the classification result of the prediction bounding box with the highest confidence score was found to be correct (Figure 3B). Figure 3C shows that one prediction bounding box was predicted, but the classification result was incorrect. In one case, multiple bounding boxes were predicted and the classification result of the prediction bounding box with the highest confidence score was incorrect (Figure 3D). Figures S1–S8 show the precision-recall curves of eight models. Table 2 shows the performances of the eight SSD models incorporating different CNN backbone networks. The mAP of the models varied from 85.9% for the VGGNet16 model to 88.9% for the DenseNet121 model. In the VGGNet16 model, which showed the lowest mAP, the AP of each species varied from 74.4% for Amazona guatemalae to 96.4% for Amazona amazonica (Table S5). The AP of each species varied from 76.5% for Amazona vittata to 98.1% for Amazona amazonica in the DenseNet121 model, which showed the highest mAP (Table S5). Among the 26 Amazon parrot species, the A. amazonica showed the highest AP for the eight models, ranging from 96.4% for the VGGNet16 and DenseNet50 models to 98.3% for the ResNet18 model. By contrast, A. guatemalae showed the lowest AP for the eight models, ranging from 66.7% for the DenseNet18 model to 82.4% for the DenseNet121 model. The inference time of the eight models ranging from 22 to 48 ms. The ResNet18 model was the fastest to classify the 26 Amazona species, whereas the DenseNet121 model was the slowest.
The classification results of 26 Amazona species using 8 models are presented as a confusion matrix (Table 3 and Tables S6–S12). Prediction results with the highest confidence values were chosen when multiple prediction bounding boxes were present. The average correct classification rate of 26 Amazon parrot species in 8 models ranged from 84.4% for the VGGNet16 model to 91.3% for the DenseNet18 model. From the eight models, the lowest and highest correct classification rates on average were seen for A. vittata (71.4%) and A. amazonica (97.2%), respectively. In the confusion matrix of the DenseNet121 models, which showed the highest mAP, correct classification rates ranged from 75.0% for A. vittata to 100.0% for four species (Amazona dufresniana, Amazona festiva, A. guatemalae, and Amazona pretrei) (Table 3). A. vittata, which showed the lowest correct classification rate in the DenseNet121 model, was incorrectly classified as Amazona tucumana (16.7%) and Amazona ventralis (8.3%). Notably, the misclassification of A. vittata as A. tucumana was the most incorrectly classified result in the DenseNet121 model. The misclassification of Amazona barbadensis as Amazona oratrix was the second most incorrectly classified result (13.9%). Amazona mercenarius was incorrectly classified as Amazona auropalliata (11.1%). The misclassifications of A. auropalliata as Amazona ochrocephala and Amazona finschi as Amazona viridigenalis occurred at 10.6% and 10.0%, respectively. Figure 4 shows the representative images for the top five results of incorrect classification.

4. Discussion

The performance of the object detection model can differ depending on the CNN architecture used as the backbone network [46]. Indeed, values for mAP and inference time of the eight models assessed in this study were different (Table 2). The performance tended to be in proportion to the complexity of the CNN architecture. The models using DenseNet as the backbone network generally showed higher mAP and slower inference time than models using VGGNet and ResNet. Similarly, models using ResNet as the backbone network showed higher mAP than those using VGGNet. This might be because of the improved architectures of ResNet and DenseNet than VGGNet. The network performance of ResNet was improved by solving the degradation problem inherent to VGGNet. This has been achieved by using a skip connection that jumps over layers and adds features used in previous layers [42]. Moreover, DenseNet maximizes information delivery by directly connecting all layers and reusing all features of the previous layer [43]. However, the inference times of the ResNet18 and ResNet34 models were faster than that of the VGGNet16 model. This is because ResNet increases the computing speed by skip connection [42]. These proportion relationships between the performance of the object detection model and the complexity of CNN architectures used as backbone networks have been reported in previous studies [37,39].
The performance of deep learning-based image classification can be related to the number and quality of images used to train models [47,48]. The relationship that species trained with more images showed a lower misclassification rate was reported in a study [49]. However, a relationship between the number of images used for training models and misclassification rates was not found in this study. Nonetheless, the number of images used in the study was relatively small, which might have affected the performance of the models. Moreover, the images used in this study were collected from the Internet; therefore, the quality of the images could not be verified. This might lead to the misclassification of the models. Additionally, the images used in this study were integrated at a 1:1 aspect ratio which is the optimal aspect ratio for the image data used in CNN architectures [41]. The images might have been distorted during the integration of the aspect ratio [49]. Therefore, a 1:1 aspect ratio should be considered when collecting images for datasets where the aspect ratio and high resolution should be standardized [41].
Figure 4 shows the representative images of the top five misclassification results of the DeseNet121 model. Among these, four results might have been misclassified due to the morphological similarity between true and predicted species. A. vittata and A. tucumana have red foreheads and lores, green mantles and backs with dark margins, and green breasts with dark margins (Figure 4A) [6,7,8]. However, these two species can be distinguished by the feather color of the primary coverts. The primary coverts of A. vittata are blue, whereas those of A. tucumana are red [6,7,8]. The images of A. vittata, which clearly show the primary coverts, should be included more to improve the classification accuracy of this species in future studies. A. barbadensis and A. oratrix share similar morphological features on their head and wing speculum (Figure 4B) [6,7,8]. A. barbadensis has a yellow crown, lores, and cheeks. A. oratrix has an entirely yellow head, including the crown, lores, and cheeks. Both species have a red wing speculum. Nevertheless, they can be distinguished by features of the forehead and lesser wing coverts [6,7,8]. A. barbadensis has a white forehead and a yellow band on the lesser wing-coverts, whereas A. ortrix has a yellow forehead and orange-red intermixed yellow band on the lesser wing coverts [6,7,8]. Therefore, the images showing the forehead and lesser wing coverts of A. barbadensis should be included more to train the models in further studies. A. auropalliata and A. ochrocephala can be difficult to distinguish because of the presence of similar features on their foreheads and forecrowns (Figure 4D) [6,7,8,23]. A. auropalliata usually has a pale bluish-green forehead and forecrown, although sometimes it has a narrow yellow frontal band extending from the forehead to the forecrown. A. ochrocephala has a bright yellow forehead and forecrown. However, the color of the nape distinguishes the two species. A. auropalliata has a nape with a broad golden-yellow band, whereas A. ochrocephala has a green nape [6,7,8]. To increase the classification accuracy of A. auropalliata, the images showing the nape of this species should be included more during model training in future studies. Furthermore, A. finschi and A. viridigenalis can be difficult to distinguish (Figure 4E) [7]. These two species have red foreheads and lores, green cheeks and ear coverts, green mantles and backs with black tips, and green rumps and uppertail coverts [6,7,8]. However, A. viridigenalis can be distinguished from A. finschi by a predominantly green crown with blue largely confined to stripes over the eyes and fewer black-tipped feathers on the underparts, including the throat and breast [6,7,8]. Images of A. finschi with the feather color of the crowns, eyes, and underparts should be incorporated in future studies. Although A. mercenarius and A. auropalliata can be easily distinguished by the yellow nape of A. auropalliata, they do not share morphological features (Figure 4C) [6,7,8]. This might be due to the relatively low number of images of A. mercenarius used during model training. Therefore, more images of A. mercenarius from various angles, indicating morphological features of this species, should be included during model training to increase the classification accuracy of this species in further studies. The confusion between morphologically similar species has been widely discussed in the computer vision community as a fine-grained recognition field [50]. To increase classification accuracy between morphologically similar species, models developed for fine-grained recognition, such as bilinear CNN models, should be applied in further studies [50]. In addition, there are multiple standard datasets specifically for birds [51,52]. Pre-training models with these datasets could increase the classification accuracy of the models.
Although this study is the first to apply the object detection model for classifying Amazon parrots, it has limitations. The object detection model was used in this study. This model is needed in wildlife conservation because most of the images taken in the wild are with multiple objects. However, the dataset was mostly with only one object in the images. Therefore, the images with multiple objects on various scales should be collected more, and the models should be tested with these images, which are more realistic datasets in further studies. Although the images in this study were collected from the Internet and hand-picked because there was no dataset for Amazon parrots, these included diverse backgrounds, such as in the wild, in cages, and captive bred. Therefore, this dataset might be possible to extend to real-world applications for monitoring the wild populations and trade of Amazon parrots. However, for more extension to real-world applications of the models for the conservation of Amazon parrots, the images with multiple objects should be collected by taking photos of Amazon parrots in places where they are traded and captive bred, such as customs and zoos, using an unmanned camera. Data augmentation was applied beforehand to obtain more than ten thousand training images per class to overcome the limitation of the small number of images in the dataset. However, this method can limit the randomness of the data than online data augmentation during training deep learning models. Therefore, online data augmentation during model training should be applied to increase the randomness of augmentation in future studies. The subspecies of the Amazon parrots were not considered for the study because of a lack of images. However, subspecies have functioned as conservation units [53]. Therefore, the classification of the Amazon parrots at the subspecies level should be undertaken for more detailed monitoring of Amazon parrots. Additionally, only adult Amazon parrots were classified in this study. However, the bird trade includes juveniles [54]. Therefore, images of juveniles of these species should be included in future studies. Object detection models are rapidly evolving, and recently developed models have shown improved performance. They should be used to identify the best-fit model for classifying Amazon parrots in the future. In this study, the nine Amazon parrots, which lacked the number of initial images, were excluded because they might have decreased the performance of the model. However, deep learning models have been developed recently to overcome the limitation of imbalanced datasets, such as an iteratively updating recognition system [55]. The real-world data are usually imbalanced. Therefore, these models should be applied to the imbalanced dataset of Amazon parrots, including nine species excluded in this study, to expand to real-world applications for the conservation of these species in future studies.

5. Conclusions

In conclusion, 8 SSD models with different CNN backbone networks were assessed for the classification of 26 Amazon parrot species. Among them, the DenseNet121 model showed the highest mAP of 88.9%. The correct classification of the 26 Amazon parrot species by the DenseNet121 model varied from 75% for A. vittata to 100% for A. dufresniana, A. festiva, A. guatemalae, and A. pretrei. The relatively low classification accuracy for some species might be caused by the morphological similarity between true and predicted species, and the relatively low number of training set images clearly showing the morphological features. Among the top five incorrect classification results for the DenseNet121 model, four might be caused due to the morphological similarity between true and predicted species. The other result might be due to a lack of images showing the morphological features of the true species. In future studies, more images clearly showing the morphological features of these species should be included during model training to enhance classification accuracy. Additionally, high resolution images with standardized aspect ratios should be collected to improve the performance of the model. Moreover, recently developed object detection models should be applied to the classification of Amazon parrots. The continuous development of deep learning models classifying Amazon parrots may enhance our ability to monitor their wild populations and global trade to conserve these species.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biology11091303/s1, Figure S1: The precision–recall curves of the 26 Amazon parrot species for the VGGNet16 model; Figure S2: The precision–recall curves of the 26 Amazon parrot species for the ResNet18 model; Figure S3: The precision–recall curves of the 26 Amazon parrot species for the ResNet34 model; Figure S4: The precision–recall curves of the 26 Amazon parrot species for the ResNet50 model; Figure S5: The precision–recall curves of the 26 Amazon parrot species for the DenseNet18 model; Figure S6: The precision–recall curves of the 26 Amazon parrot species for the DenseNet30 model; Figure S7: The precision–recall curves of the 26 Amazon parrot species for the DenseNet50 model; Figure S8: The precision–recall curves of the 26 Amazon parrot species for the DenseNet121 model; Table S1: The augmentation rates and the number of training set after data augmentation of the 26 Amazon parrot species; Table S2; Structure of VGGNet based on SSD architecture. Each “Conv” layer in the table corresponds to the composite function sequence Conv-ReLU; Table S3: Structure of ResNet based on SSD architecture. Each “Conv” layer in the table corresponds to the composite function sequence BN-ReLU-Conv; Table S4: Structure of DenseNet based on SSD architecture. Growth rate K = 32 was used for each dense block. Each “Conv” layer in the table corresponds to the composite function sequence BN-ReLU-Conv; Table S5: The values of average precision (AP) of the assessed models for the 26 Amazon parrot species; Table S6: Confusion matrix of the VGGNet16 model for the classification of the 26 Amazon parrot species; Table S7: Confusion matrix of the ResNet18 model for the classification of the 26 Amazon parrot species; Table S8: Confusion matrix of the ResNet34 model for the classification of the 26 Amazon parrot species; Table S9: Confusion matrix of the ResNet50 model for the classification of the 26 Amazon parrot species; Table S10: Confusion matrix of the DenseNet18 model for the classification of the 26 Amazon parrot species; Table S11: Confusion matrix of the DenseNet30 model for the classification of the 26 Amazon parrot species; Table S12: Confusion matrix of the DenseNet50 model for the classification of the 26 Amazon parrot species.

Author Contributions

Conceptualization, J.-I.K. and C.-B.K.; investigation, J.-I.K. and J.-W.B.; data curation, J.-I.K. and J.-W.B.; writing—original draft preparation, J.-I.K. and C.-B.K. writing—review and editing, C.-B.K., supervision, C.-B.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a grant from the National Institute of Biological Resources (NIBR), funded by the Ministry of Environment (MOE) of the Republic of Korea (NIBR202231204).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

The authors thank Woohyuk Jang and Eui Chul Lee for their technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Blanco, G.; Hiraldo, F.; Tella, J.L. Ecological functions of parrots: An integrative perspective from plant life cycle to ecosystem functioning. Emu 2018, 118, 36–49. [Google Scholar] [CrossRef]
  2. Symes, C.T.; Perrin, M.R. Feeding biology of the greyheaded parrot, Poicephalus fuscicollis suahelicus (Reichenow), in Northern Province, South Africa. Emu 2003, 103, 49–58. [Google Scholar] [CrossRef]
  3. Olah, G.; Butchart, S.H.; Symes, A.; Guzmán, I.M.; Cunningham, R.; Brightsmith, D.J.; Heinsohn, R. Ecological and socio-economic factors affecting extinction risk in parrots. Biodivers. Conserv. 2016, 25, 205–223. [Google Scholar] [CrossRef]
  4. Bush, E.R.; Baker, S.E.; Macdonald, D.W. Global trade in exotic pets 2006–2012. Conserv. Biol. 2014, 28, 663–676. [Google Scholar] [CrossRef] [PubMed]
  5. Scheffers, B.R.; Oliveira, B.F.; Lamb, I.; Edwards, D.P. Global wildlife trade across the tree of life. Science 2019, 366, 71–76. Available online: https://www.science.or (accessed on 27 June 2022). [CrossRef]
  6. Forshaw, J.M. Parrots of the World; Princeton University Press: Princeton, NJ, USA, 2010; pp. 278–308. [Google Scholar]
  7. Parr, M.; Juniper, T. Parrots: A Guide to Parrots of the World; Bloomsbury Publishing: London, UK, 2010; pp. 523–554. [Google Scholar]
  8. Del Hoyo, J. All the Birds of the World; Lynx Edicions: Barcelona, Spain, 2020; pp. 364–366. [Google Scholar]
  9. BirdLife International Data Zone. Available online: http://datazone.birdlife.org/species/search (accessed on 27 June 2022).
  10. Checklist of CITES Species. Available online: https://checklist.cites.org/#/en (accessed on 20 June 2022).
  11. Chan, D.T.C.; Poon, E.S.K.; Wong, A.T.C.; Sin, S.Y.W. Global trade in parrots—Influential factors of trade and implications for conservation. Glob. Ecol. Conserv. 2021, 30, e01784. [Google Scholar] [CrossRef]
  12. Sánchez-Mercado, A.; Ferrer-Paris, J.R.; Rodríguez, J.P.; Tella, J.L. A literature synthesis of actions to tackle illegal parrot trade. Diversity 2021, 13, 191. [Google Scholar] [CrossRef]
  13. Sinovas, P.; Price, B.; King, E.; Hinsley, A.; Pavitt, A. Wildlife Trade in the Amazon Countries: An Analysis of Trade in CITES Listed Species; UN Environment—World Conservation Monitoring Centre: Cambridge, UK, 2017; pp. 31–34. [Google Scholar]
  14. Cardador, L.; Lattuada, M.; Strubbe, D.; Tella, J.L.; Reino, L.; Figueira, R.; Carrete, M. Regional bans on wild-bird trade modify invasion risks at a global scale. Conserv. Lett. 2017, 10, 717–725. [Google Scholar] [CrossRef]
  15. Seo, M.H.; Kim, H.J. The Guideline for Import and Export Review of CITES Species; National Institute of Biological Resources: Incheon, Korea, 2016; pp. 76–132. [Google Scholar]
  16. Schneider, S.; Taylor, G.W.; Kremer, S. Deep learning object detection methods for ecological camera trap data. In Proceedings of the 2018 15th Conference on Computer and Robot Vision (CRV), Toronto, ON, Canada, 9–11 May 2018; pp. 321–328. [Google Scholar]
  17. Rich, L.N.; Davis, C.L.; Farris, Z.J.; Miller, D.A.; Tucker, J.M.; Hamel, S.; Farhadinia, M.S.; Steenweg, R.; Di Bitetti, M.S.; Thapa, K. Assessing global patterns in mammalian carnivore occupancy and richness by integrating local camera trap surveys. Global Ecol. Biogeogr. 2017, 26, 918–929. [Google Scholar] [CrossRef]
  18. Tempier, L. Building worldwide expertise to detect and seize illegally traded wildlife. In UN Chronicle: Illegal Wildlife Trad; United Nations: New York, NY, USA, 2014; Volume LI, pp. 25–28. [Google Scholar]
  19. Trail, P.W. Morphological analysis: A powerful tool in wildlife forensic biology. Forensic Sci. Int. 2021, 1, 100025. [Google Scholar] [CrossRef]
  20. Mahendiran, M.; Parthiban, M.; Azeez, P.A.; Nagarajan, R. In situ measurements of animal morphological features: A non-invasive method. Methods Ecol. Evol. 2018, 9, 613–623. [Google Scholar] [CrossRef]
  21. Esipova, O.; Love, E.; Noakes, A.; Schatz, A.; Swartz, K.; Vallianos, C. Wildlife Trafficking Detection Tools: Best Practices and Application to the Illegal Rhino Horn Trade; Conservation International: Arlington, VA, USA, 2021; pp. 30–33. [Google Scholar]
  22. Hopkins, G.; Freckleton, R.P. Declines in the numbers of amateur and professional taxonomists: Implications for conservation. Anim. Conserv. 2002, 5, 245–249. [Google Scholar] [CrossRef]
  23. Eberhard, J.R.; Bermingham, E. Phylogeny and biogeography of the Amazona ochrocephala (Aves: Psittacidae) complex. Auk 2004, 121, 318–332. [Google Scholar] [CrossRef]
  24. Wenner, T.J.; Russello, M.A.; Wright, T.F. Cryptic species in a Neotropical parrot: Genetic variation within the Amazona farinosa species complex and its conservation implications. Conserv. Genet. 2012, 13, 1427–1432. [Google Scholar] [CrossRef]
  25. Abe, H.; Hayano, A.; Inoue-Murayama, M. Forensic species identification of large macaws using DNA barcodes and microsatellite profiles. Mol. Biol. Rep. 2012, 39, 693–699. [Google Scholar] [CrossRef]
  26. Kim, J.I.; Do, T.D.; Lee, D.; Yeo, Y.; Kim, C.B. Application of Cytochrome b Gene Sequences for Identification of Parrots from Korean Zoos. Anim. Syst. Evol. Divers. 2020, 36, 216–221. [Google Scholar] [CrossRef]
  27. Brandis, K.J.; Meagher, P.J.; Tong, L.J.; Shaw, M.; Mazumder, D.; Gadd, P.; Ramp, D. Novel detection of provenance in the illegal wildlife trade using elemental data. Sci. Rep. 2018, 8, 15380. [Google Scholar] [CrossRef]
  28. Wäldchen, J.; Mäder, P. Machine learning for image based species identification. Methods Ecol. Evol. 2018, 9, 2216–2225. [Google Scholar] [CrossRef]
  29. Schneider, S.; Taylor, G.W.; Linquist, S.; Kremer, S.C. Past, present and future approaches using computer vision for animal re-identification from camera trap data. Methods Ecol. Evol. 2019, 10, 461–470. [Google Scholar] [CrossRef]
  30. Christin, S.; Hervet, É.; Lecomte, N. Applications for deep learning in ecology. Methods Ecol. Evol. 2019, 10, 1632–1644. [Google Scholar] [CrossRef]
  31. Surkovic, E.; Smith, L.; Edbury, M.; Pilkington, J.; Taylor, L.; Kahunzire, A.; Kett, M. Science: Tackling the Illegal Wildlife Trade; The Royal Society: London, UK, 2018; pp. 23–25. [Google Scholar]
  32. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  33. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed]
  34. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  35. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  36. Hung, P.D.; Kien, N.N. SSD-Mobilenet Implementation for classifying fish species. In Intelligent Computing and Optimization; Vasant, P., Zelinka, I., Weber, G.W., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 399–408. [Google Scholar]
  37. Jang, W.; Lee, E.C. Multi-class parrot image classification including subspecies with similar appearance. Biology 2021, 10, 1140. [Google Scholar] [CrossRef] [PubMed]
  38. Akdemir, K.Ü.; Alaybeyoğlu, E. Classification of Red Mullet, Bluefish and Haddock caught in the Black sea by “Single Shot Multibox Detection”. In Proceedings of the IEEE International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; pp. 1–4. [Google Scholar]
  39. Patel, A.; Cheung, L.; Khatod, N.; Matijosaitiene, I.; Arteaga, A.; Gilkey, J.W., Jr. Revealing the unknown: Real-time recognition of Galápagos snake species using deep learning. Animals 2020, 10, 806. [Google Scholar] [CrossRef] [PubMed]
  40. DarkLabel. Available online: https://darkpgmr.tistory.com/16 (accessed on 5 March 2022).
  41. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  44. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  45. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  46. Hong, S.; Han, Y.; Kim, S.; Lee, A.; Kim, G. Application of deep-learning methods to bird detection using unmanned aerial vehicle imagery. Sensors 2019, 19, 1651. [Google Scholar] [CrossRef]
  47. Dodge, S.; Karam, L. Understanding how image quality affects deep neural networks. In Proceedings of the Eighth International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016; pp. 1–6. [Google Scholar]
  48. Spiesman, B.J.; Gratton, C.; Hatfield, R.G.; Hsu, W.H.; Jepsen, S.; McCornack, B.; Patel, K.; Wang, G. Assessing the potential for deep learning and computer vision to identify bumble bee species from images. Sci. Rep. 2021, 11, 7580. [Google Scholar] [CrossRef]
  49. Hou, J.; He, Y.; Yang, H.; Connor, T.; Gao, J.; Wang, Y.; Zeng, Y.; Zhang, J.; Huang, J.; Zheng, B. Identification of animal individuals using deep learning: A case study of giant panda. Biol. Conserv. 2020, 242, 108414. [Google Scholar] [CrossRef]
  50. Lin, T.Y.; Roy Chowdhury, A.; Maji, S. Bilinear CNN models for fine-grained visual recognition. In Proceeding of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1449–1457. [Google Scholar]
  51. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-UCSD Birds-200–2011 Dataset; Computation & Neural Systems Technical Report, CNS-TR; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  52. Van Horn, G.; Branson, S.; Farrell, R.; Haber, S.; Barry, J.; Ipeirotis, P.; Perona, P.; Belongie, S. Building a bird recognition app and large scale dataset with citizen scientists: The fine print in fine-grained dataset collection. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 595–604. [Google Scholar]
  53. Zink, R.M. The role of subspecies in obscuring avian biological diversity and misleading conservation policy. Proc. R. Soc. B Biol. Sci. 2004, 271, 561–564. [Google Scholar] [CrossRef]
  54. Johnson, R.N. Conservation Genetics and Wildlife Forensics of Birds. In Wildlife Forensics: Methods and Applications; Jane, E.H., John, R.W., Eds.; Wiley-Blackwell: Hoboken, NJ, USA, 2011; pp. 293–325. [Google Scholar]
  55. Miao, Z.; Liu, Z.; Gaynor, K.M.; Palmer, M.S.; Yu, S.X.; Getz, W.M. Iterative human and automated identification of wildlife images. Nat. Mach. Intell. 2021, 3, 885–895. [Google Scholar] [CrossRef]
Figure 1. Representative images of Amazona aestiva after data augmentation. The green boxes in the images represent the ground-truth bounding boxes. (A) Initial image, (B) Horizontal flipped image, (C) Rotated image, (D) Zoomed-in image, (E) Zoomed-out image, and (F) Translated image. Photo credit: Mauro Halpern.
Figure 1. Representative images of Amazona aestiva after data augmentation. The green boxes in the images represent the ground-truth bounding boxes. (A) Initial image, (B) Horizontal flipped image, (C) Rotated image, (D) Zoomed-in image, (E) Zoomed-out image, and (F) Translated image. Photo credit: Mauro Halpern.
Biology 11 01303 g001
Figure 2. Single Shot MultiBox Detector model architecture with different convolution neural network (CNN) backbone networks for the classification of the 26 Amazon parrot species.
Figure 2. Single Shot MultiBox Detector model architecture with different convolution neural network (CNN) backbone networks for the classification of the 26 Amazon parrot species.
Biology 11 01303 g002
Figure 3. Representative images of four cases of model prediction results. The green and yellow boxes on the images represent ground-truth and prediction bounding boxes, respectively. The values in the yellow boxes are confidence scores provided by the models, indicating the probability of the prediction being correct. (A) Image of Amazona aestiva, one prediction bounding box was predicted and classified correctly; (B) Image of A. aestiva, multiple prediction bounding boxes were predicted and classified correctly; (C) Image of Amazona vittata, one prediction bounding box was predicted and classified incorrectly; (D) Image of Amazona albifrons, multiple prediction bounding boxes were predicted and classified incorrectly. Photo credit: (A) Charles J. Sharp, (B) Bernard Dupont, (C) Tom MacKenzie, and (D) Charlottesville.
Figure 3. Representative images of four cases of model prediction results. The green and yellow boxes on the images represent ground-truth and prediction bounding boxes, respectively. The values in the yellow boxes are confidence scores provided by the models, indicating the probability of the prediction being correct. (A) Image of Amazona aestiva, one prediction bounding box was predicted and classified correctly; (B) Image of A. aestiva, multiple prediction bounding boxes were predicted and classified correctly; (C) Image of Amazona vittata, one prediction bounding box was predicted and classified incorrectly; (D) Image of Amazona albifrons, multiple prediction bounding boxes were predicted and classified incorrectly. Photo credit: (A) Charles J. Sharp, (B) Bernard Dupont, (C) Tom MacKenzie, and (D) Charlottesville.
Biology 11 01303 g003
Figure 4. Representative images of the top five results for incorrect classification using the DenseNet121 model. Images on the left and right represent the true and predicted results, respectively. (A) Amazona vittata (left) was predicted to be Amazona tucumana (right). (B) Amazona barbadensis (left) was predicted to be Amazona oratrix (right). (C) Amazona mercenarius (left) was predicted to be Amazona auropalliata (right). (D) Amazona auropalliata (left) was predicted to be Amazona ochrocephala (right). (E) Amazona finschi (left) was predicted to be Amazona viridigenalis (right). Photo credits: (A) Tom MacKenzie (left), Carlos Urdiales (right); (B) Emőke Dénes (left), David J. Stang (right); (C) Félix Uribe (left), Andrew Gwozdziewycz (right); (D) Andrew Gwozdziewycz (left), MAClarke21 (right); and (E) Cédric Allier (left), Roger Moore (right).
Figure 4. Representative images of the top five results for incorrect classification using the DenseNet121 model. Images on the left and right represent the true and predicted results, respectively. (A) Amazona vittata (left) was predicted to be Amazona tucumana (right). (B) Amazona barbadensis (left) was predicted to be Amazona oratrix (right). (C) Amazona mercenarius (left) was predicted to be Amazona auropalliata (right). (D) Amazona auropalliata (left) was predicted to be Amazona ochrocephala (right). (E) Amazona finschi (left) was predicted to be Amazona viridigenalis (right). Photo credits: (A) Tom MacKenzie (left), Carlos Urdiales (right); (B) Emőke Dénes (left), David J. Stang (right); (C) Félix Uribe (left), Andrew Gwozdziewycz (right); (D) Andrew Gwozdziewycz (left), MAClarke21 (right); and (E) Cédric Allier (left), Roger Moore (right).
Biology 11 01303 g004
Table 1. Dataset of 26 Amazon parrot species examined in this study.
Table 1. Dataset of 26 Amazon parrot species examined in this study.
No.SpeciesTraining SetValidation SetTest Set
1Amazona aestiva2194648
2Amazona albifrons2174647
3Amazona amazonica2896263
4Amazona auropalliata2154647
5Amazona autumnalis2024344
6Amazona barbadensis1643536
7Amazona brasiliensis1653536
8Amazona collaria781618
9Amazona dufresniana831719
10Amazona festiva952021
11Amazona finschi2284850
12Amazona guatemalae841819
13Amazona guildingii952426
14Amazona leucocephala2806061
15Amazona lilacina781618
16Amazona mercenarius791618
17Amazona ochrocephala1984244
18Amazona oratrix2555456
19Amazona pretrei1311921
20Amazona rhodocorytha1262728
21Amazona tucumana1052223
22Amazona ventralis1453132
23Amazona versicolor1072224
24Amazona vinacea1914142
25Amazona viridigenalis1803840
26Amazona vittata1082324
Total4096867905
Table 2. Values of mean average precision (mAP) and inference time of the eight models.
Table 2. Values of mean average precision (mAP) and inference time of the eight models.
ModelmAP (%)Inference Time (ms)
VGGNet1685.927
ResNet1887.822
ResNet3487.525
ResNet5087.231
DenseNet1887.631
DenseNet3086.834
DenseNet5088.645
DenseNet12188.948
Table 3. Confusion matrix of the DenseNet121 model for the classification of the 26 Amazon parrot species. Numbers from 1–26 indicate the 26 Amazon parrot species (shown in Table 1). The rows contain the actual species, and the columns contain the species predicted by the models. The prediction results for the models are shown as percentage values. The diagonal values indicate the correct predictions, with the other values being the incorrect predictions. The correct predictions are shaded in blue, and the incorrect predictions are in red. The deeper the blue, the higher the value of the correct prediction; the deeper the red, the higher the value of the incorrect prediction.
Table 3. Confusion matrix of the DenseNet121 model for the classification of the 26 Amazon parrot species. Numbers from 1–26 indicate the 26 Amazon parrot species (shown in Table 1). The rows contain the actual species, and the columns contain the species predicted by the models. The prediction results for the models are shown as percentage values. The diagonal values indicate the correct predictions, with the other values being the incorrect predictions. The correct predictions are shaded in blue, and the incorrect predictions are in red. The deeper the blue, the higher the value of the correct prediction; the deeper the red, the higher the value of the incorrect prediction.
Predicted Results
1234567891011121314151617181920212223242526
True Results191.70.02.10.00.00.00.00.00.00.00.00.00.02.10.00.02.10.00.00.00.00.00.00.00.02.1
20.091.50.00.00.00.00.00.00.00.00.00.00.08.50.00.00.00.00.00.00.00.00.00.00.00.0
30.01.696.81.60.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
42.10.00.083.00.00.00.00.00.00.00.02.10.00.00.00.010.60.00.00.00.00.00.02.10.00.0
50.00.00.00.084.10.02.30.00.00.02.30.00.00.00.00.00.00.04.54.50.00.00.00.02.30.0
62.80.00.00.00.080.60.00.00.00.00.00.00.00.00.00.02.813.90.00.00.00.00.00.00.00.0
70.00.00.00.00.00.094.42.80.02.80.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
80.00.00.00.00.00.00.094.45.60.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
90.00.00.00.00.00.00.00.0100.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
100.00.00.00.00.00.00.00.00.0100.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
110.00.00.00.00.00.02.00.00.02.082.00.00.00.04.00.00.00.00.00.00.00.00.00.010.00.0
120.00.00.00.00.00.00.00.00.00.00.0100.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0
130.00.00.00.00.00.00.00.00.00.00.00.096.20.00.00.00.00.00.00.00.00.03.80.00.00.0
140.00.00.00.00.00.01.60.00.00.00.00.00.093.40.00.00.00.00.00.03.30.00.01.60.00.0
150.00.00.00.05.60.00.00.00.00.00.00.00.00.088.90.00.00.00.05.60.00.00.00.00.00.0
160.00.00.011.10.00.00.00.00.00.00.05.60.00.00.083.30.00.00.00.00.00.00.00.00.00.0
170.00.00.04.50.00.00.00.00.00.00.04.50.00.00.00.088.60.00.00.00.00.02.30.00.00.0
180.01.80.00.00.03.60.00.00.00.00.00.00.00.00.00.00.092.90.01.80.00.00.00.00.00.0
190.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0100.00.00.00.00.00.00.00.0
200.00.00.00.03.60.03.60.03.60.00.00.00.00.00.00.00.00.00.089.30.00.00.00.00.00.0
210.00.00.00.00.00.00.00.00.00.00.04.30.00.00.00.00.00.00.00.091.30.00.00.00.04.3
220.00.03.10.00.00.00.03.13.10.00.00.00.00.00.00.00.00.00.00.00.087.53.10.00.00.0
230.00.00.00.00.00.00.00.04.20.00.00.00.04.20.00.00.00.00.00.00.00.087.50.04.20.0
240.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.02.40.00.095.22.40.0
250.00.00.00.00.00.05.00.00.00.00.00.00.00.00.00.00.00.00.02.52.50.00.00.090.00.0
260.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.016.78.30.00.00.075.0
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, J.-I.; Baek, J.-W.; Kim, C.-B. Image Classification of Amazon Parrots by Deep Learning: A Potentially Useful Tool for Wildlife Conservation. Biology 2022, 11, 1303. https://doi.org/10.3390/biology11091303

AMA Style

Kim J-I, Baek J-W, Kim C-B. Image Classification of Amazon Parrots by Deep Learning: A Potentially Useful Tool for Wildlife Conservation. Biology. 2022; 11(9):1303. https://doi.org/10.3390/biology11091303

Chicago/Turabian Style

Kim, Jung-Il, Jong-Won Baek, and Chang-Bae Kim. 2022. "Image Classification of Amazon Parrots by Deep Learning: A Potentially Useful Tool for Wildlife Conservation" Biology 11, no. 9: 1303. https://doi.org/10.3390/biology11091303

APA Style

Kim, J. -I., Baek, J. -W., & Kim, C. -B. (2022). Image Classification of Amazon Parrots by Deep Learning: A Potentially Useful Tool for Wildlife Conservation. Biology, 11(9), 1303. https://doi.org/10.3390/biology11091303

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop