Next Article in Journal
Therapeutic Approach of Whole-Body Vibration Exercise on Wound Healing in Animal Models: A Systematic Review
Previous Article in Journal
Polycyclic Aromatic Hydrocarbons (PAHs) in Roasted Pork Meat and the Effect of Dried Fruits on PAH Content
Previous Article in Special Issue
Preparation and Performance of Carbon-Based Ce-Mn Catalysts for Efficient Degradation of Acetone at Low Temperatures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fine-Grained Recognition Neural Network with High-Order Feature Maps via Graph-Based Embedding for Natural Bird Diversity Conservation

1
School of Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
2
State Environmental Protection Key Laboratory of Food Chain Pollution Control, Beijing Technology and Business University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2023, 20(6), 4924; https://doi.org/10.3390/ijerph20064924
Submission received: 20 January 2023 / Revised: 28 February 2023 / Accepted: 8 March 2023 / Published: 10 March 2023

Abstract

:
The conservation of avian diversity plays a critical role in maintaining ecological balance and ecosystem function, as well as having a profound impact on human survival and livelihood. With species’ continuous and rapid decline, information and intelligent technology have provided innovative knowledge about how functional biological diversity interacts with environmental changes. Especially in complex natural scenes, identifying bird species with a real-time and accurate pattern is vital to protect the ecological environment and maintain biodiversity changes. Aiming at the fine-grained problem in bird image recognition, this paper proposes a fine-grained detection neural network based on optimizing the YOLOV5 structure via a graph pyramid attention convolution operation. Firstly, the Cross Stage Partial (CSP) structure is introduced to a brand-new backbone classification network (GPA-Net) for significantly reducing the whole model’s parameters. Then, the graph pyramid structure is applied to learn the bird image features of different scales, which enhances the fine-grained learning ability and embeds high-order features to reduce parameters. Thirdly, YOLOV5 with the soft non-maximum suppression (NMS) strategy is adopted to design the detector composition, improving the detection capability for small targets. Detailed experiments demonstrated that the proposed model achieves better or equivalent accuracy results, over-performing current advanced models in bird species identification, and is more stable and suitable for practical applications in biodiversity conservation.

1. Introduction

Avian diversity refers to the variety of bird species that exist in natural environments. Birds are one of the most diverse groups of vertebrate animals on earth and play various roles in ecosystems, such as pollination, occupying the top or bottom of the food chain, and controlling populations [1]. The conservation of avian diversity plays a critical role in maintaining ecological balance and ecosystem function, in addition to having a profound impact on human survival and livelihood. With factors such as environmental change, habitat loss, habitat destruction, and human activities, many bird species are facing the threat of endangerment and extinction on a global scale. Therefore, protecting and maintaining avian diversity has become a global priority, especially in complex outdoor environments, where the accurate identification of bird species is essential to protect ecological environment security and maintain changes in avian biodiversity [2,3].
From the defined scope, biodiversity refers to all plants, animals, and microorganisms on the earth and their genes. The ecosystem and its ecological process are formed by interacting with various species and their habitats, including intra-species, inter-species, and ecosystem diversity [4]. Monitoring biodiversity is not only an effective way to monitor the changes between organisms but also an effective way to evaluate their conservation effectiveness. It can provide an essential basis for formulating action plans and management measures related to biodiversity conservation [5,6]. Bird species diversity is an ecological combination formed by interacting with birds, other organisms, and the environment, e.g., the relationship between birds and air quality [7], and the impact between birds and wetland environments [8]. It can characterize the composition and structure of bird communities and the quality of bird habitats, which is of great significance for understanding the stability of ecosystems [9]. Meanwhile, birds have unique ecological and system engineering effects, such as species conservation and environmental restoration, and are an integral part of wetland ecosystems, which are relevant to the healthy development of human civilization [10]. Because birds have a broad niche and are sensitive to changes in ecological factors, they are biological indicator species for detecting environmental quality [11]. The species composition of birds reflects the comprehensive utilization of the bird habitat environment to a certain extent [12]. In addition, birds provide humans with important cultural connotations that inspire people to watch them, which is fascinating in the life history of birds. One example is bird watching, a sport that is very beneficial to human physical and mental health and that actually focuses on the diversity of birds [13].
In forestry and wetland ecosystems, bird identification technology can effectively help related staff to realize automatic bird monitoring and count the number and species distribution of birds in the natural ecological environment, which plays a decisive role in bird protection and has essential research value [14]. Traditional bird identification mainly relies on the experience of experts, some of whom specialize in bird work and research and some of whom use bird watching as a recreational activity. But in any case, these people still constitute a minority group [15]. Although this method can ensure high accuracy, it needs to rely on experts to identify birds manually, which is costly. Moreover, there are many kinds of wild birds, and their habitats are complex. It is inefficient and difficult to guarantee the accuracy rate only by manually interpreting and counting their types and distributions. In addition, humans cannot wait in the wild all the time, so how to effectively measure the existence and distribution of the bird population has always been a complex problem.
With the progress of computer and artificial intelligence technology, artificial intelligence technology has been gradually applied to various intelligent identification fields, e.g., intelligent health monitoring [16], agricultural hazard detection [17], remote sensing satellite image processing [18], pest identification [19], and food traceability [20]. Similarly, bird image recognition technology has made significant progress. In the early days, researchers mainly relied on pattern recognition by manually extracting classification features to recognize bird images [21]. Researchers must select key features of different birds, e.g., texture [22] and the gradient histogram [23], and then design corresponding digital feature vectors. By comparing these feature vectors, the classification and prediction of different bird targets can be realized. However, to minimize the error, it is often necessary to manually extract the features of crucial points and subtle differences, dramatically increasing the workload of manual feature extraction. Moreover, due to the different requirements of each task, researchers focus on selecting varying bird features, which weakens the generalization ability of this kind of bird classification method. It is difficult for the algorithm to be further popularized and applied [24], and it is also tricky to significantly improve its accuracy, bringing significant challenges to bird feature recognition [25].
With deep learning and convolutional neural network development, target classification and recognition technology have also made significant breakthroughs. Deep learning is a method to learn big data samples’ inherent laws and representation levels. In the learning process, much information can be obtained for data interpretation. Deep learning has been studied in many fields, e.g., natural language [26], images [27], and sounds [28]. Deep learning can replace much repetitive work performed by human beings with mighty computer computing power and dig deeper into the subtle feature differences among different birds with its deep feature extraction capability, which significantly improves classification accuracy. Deep learning image classification distinguishes different images according to their semantic information. The difference lies in that their features are not extracted or designed manually. Still, hierarchical feature learning is carried out supervised or unsupervised through deep neural networks. The pixel information of the images is directly used for input, which keeps all the knowledge of the input images to the greatest extent. Through convolution operation, feature extraction and high-level feature abstraction are carried out on the images, and its image classification effect has surpassed that of human beings under certain conditions. It has significant advantages in the field of bird recognition.
However, the bird data obtained in the virtual environment are often affected by the actual environmental factors, e.g., occlusion, light, and shadow, which significantly limits the performance of the bird recognition algorithm. Meanwhile, in the actual environment, there are fine-grained phenomena with minor interspecific differences and large species differences among birds, which are also essential factors restricting bird identification. Moreover, because the equipment used for bird detection is often fixed in the virtual environment, this leads to a situation in which the target in the obtained bird image is tiny, which leads to a failure to detect the bird in the target detection algorithm correctly. Therefore, aiming at fine-grained bird recognition and detection, this paper proposes a bird recognition method based on improved YOLOV5 [29] via graph pyramid attention. This method is divided into three parts. Firstly, using a backbone network based on the CSP structure and cross-stage attention method, the CSP structure can streamline the model while attention can extract fine-grained features. Then, the graph pyramid structure is used to further extract features of different scales and for fine-grained recognition; finally, the optimized YOLOV5 architecture with Soft-NMS strategy is used for fine-grained image detection of various birds in complex natural environments.

2. Related Works

2.1. Traditional Bird Identification Dataset and Methods

Bird image recognition methods are mainly divided into machine learning and deep learning methods. The traditional recognition algorithm is based on manual extraction, feature formation, and a cascade feature classifier, and it is realized by shallow trainable architecture. A dataset is the basis of research. Many datasets for bird identification research have been published on the Internet. The BIRDS 400 bird dataset contains datasets of 400 birds, 58,388 training images, 2000 test images, and 2000 verification images. There is only one bird in each picture. All images are 224 × 224 × 3 color images in jpg format. The California Institute of Technology proposed the CUB200-2011 dataset in 2010 [30], the benchmark image dataset for fine-grained classification and recognition research. There are 11,788 bird images in this dataset, including 200 bird subcategories. The dataset is divided into training and testing, and the number of images is 5994 and 5794, respectively. Each image provides image label information, bird bounding box information, essential part information of birds, and attribute information. The Birdsnap dataset [31] contains pictures of 500 common birds in North America, with the number of pictures of each bird being about 100, for a total of 49,829 photos.
The NABirds V1 dataset is a collection of 48,000 annotated images of 400 species of birds common in North America [32]. There are more than 100 photos of each species, including individual annotations of males, females, and larvae. In the early days, people needed artificial feature designs for feature extraction, such as local binary patterns [33] and directional gradient histograms [34]. This method must first determine the features and parts to be extracted, for example, the bird’s outline, color, and texture, and 15 significant details of the bird’s mouth, feet, wings, and neck [35].
Huang et al. [36] use a graphical model considering significance to classify and annotate fine-grained bird species. The method first divides the image into several regions. Then, based on the regional and patch-level characteristics, GMS is used for classification and labeling. Finally, SVM is used as a classifier, and GMS is used for image classification. Berg et al. [37] proposed an algorithm for feature coding of local areas, automatically searching for information that plays a crucial role in image classification. However, this method requires high positioning accuracy of critical points, so only a 73.3% recognition accuracy can be obtained on CUB200-2011. These methods are suitable for simple images and a small amount of image data, but they perform poorly for images with complex backgrounds, such as the presence of occlusion or bad weather [38], and often fail to achieve the expected classification accuracy, and the generalization ability of the models are poor.

2.2. Bird Identification via Deep Learning Migration Technology

Thanks to the rapid development of deep learning, target recognition and classification technology have made remarkable progress. In recent years, networks, e.g., AlexNet [39], VGG [40], ResNet [41], GoogleNet [42], and DenseNet [43], have been launched one after another, which shows that convolutional neural networks have a compelling feature learning ability in image recognition. Therefore, many researchers also apply deep learning technology to classify and detect birds. In 2013, Zhang et al. [44] proposed a fine-grained classification model, R-CNN, based on parts. In this method, the depth convolution feature based on the bottom-up area suggestion calculation is used to overcome the labeling problem of the assumed bounding box in the attitude normalization process. The bounding box is modified and fine-tuned by learning the whole object and the component detector, and the geometric constraints between them are forcibly learned. Finally, a classification accuracy of 76.37% is achieved by implementing feature cascade fusion classification through AlexNet. Donahue et al. [45] analyzed the use of in-depth features in the semi-supervised multi-task framework, found that such features are superior to the most advanced feature methods based on complex multi-core learning technology and traditional manual engineering, and named them De CAF, which proves that convolutional neural network (CNN) features have more robust semantic information and better performance than artificial features. Since then, almost all bird image recognition algorithms have turned to CNN features.
Because of the apparent fine-grained problem in birds, it is also essential for research content to use the fine-grained method to identify birds. Huang et al. [46] proposed a PS-CNN method based on a partially laminated CNN structure, which, based on component-based R-CNN, modeled the subtle differences between components and objects to achieve accurate positioning of multiple target discrimination parts. By adopting a set of sharing strategies in calculating multiple object parts, the classification accuracy was 76%. Song et al. [47] introduced the feature enhancement and suppression module, which enabled the network to mine other potential features when learning the most significant features. The network could understand features of different significant levels through multilevel operations, which improved the recognition performance. Ji et al. [48] used a convolutional binary tree network based on the attention method to recognize fine-grained images. To calculate the calculation path from root to leaf in each tree, a convolution operation was added to the edge of the tree, so the final decision calculation was the synthesis of leaf node prediction.

2.3. Image Recognition Based on Graph Method

The graph neural network (GNN) is a new deep learning framework that has appeared in recent years. This framework can directly learn the graph structure data representing different local areas. Therefore, it is of tremendous research significance to use the advantages of graph neural networks in analyzing interpretable features of non-European data to understand the internal relations among various fine-grained features. With the development of graph neural network theory, graph neural networks have also derived many different network structures, such as the graph convolution neural network (GCN) [49] based on the spectral domain and spatial domain, graph neural network [50] based on gating, graph attention network [51] based on the attention method, graph spatiotemporal network [52] based on time and space, etc.
Zhao et al. [53] used a graph-based high-order relationship discovery method to learn the implicit connection between different areas. This method established a feature database through semantic and location perception. Then, using the semantic grouping strategy of a graph, the features of high-dimensional data are mapped to low-dimensional space, and the learning parameters are reduced. Lin et al. [54] constructed a graph convolution network method for weakly supervised fine-grained image classification based on correlation learning. This method learns the implicit relationship between different regions from the network transmission to fully mine and utilize the context relationship between other discrimination regions, thus improving the network’s recognition ability. Wang et al. [55] proposed a cross graph propagation sub-network to learn regional correlation and then weighted and aggregated other regions in a cross way to enhance each part. Chen et al. [56] used a fine-grained graph network method to represent the learning framework based on knowledge graph embedding. Firstly, this method builds a rich visual concept library as a knowledge map. Then, the gate graph neural network generates knowledge representation to realize fine-grained image recognition. Wang et al. [57] studied the confidence of graph neural networks. By applying the confidence correction structure to the graph convolution network, the faith of the graph convolution network was improved, and the classification ability of the graph was also enhanced.

3. Methods and Materials

The overall framework of this paper is mainly composed of three parts: data pre-processing, a fine-grained classifier, and an optimized detector. The framework structure is shown in Figure 1. Among them, the data pre-processing component introduces various strategies used to process the data. Part of the classifier (defined as GPA-Net) uses the graph-based pyramid attention method, which is mainly composed of a CSP-based backbone network and pyramid attention structure [58]. The Yolov5 detector is used as the detector in the detection part, and the Soft-NMS mechanism is used in the detection part to enhance the detection results.
The proposed fine-grained classifier, GPA-Net, mainly consists of the backbone structure based on the CSP structure and the fine-grained feature learning structure of graph pyramid attention, as shown in Figure 1. The specific functions of each part are as follows.

3.1. Lightweight Backbone Network

The improved CSP backbone network is used for processing as shown in Figure 2. CSP is a new variant of the ResNet network series. This structure prevents too-much-repeated gradient information by cutting off the gradient stream, strengthens CNN’s learning ability, eliminates the computational bottleneck, and effectively reduces the memory cost. The staging module of the basic CSPNet includes a primary branch and a spanning branch, and the characteristics of the two branches are spliced at the end of each stage.
The phase module of the basic CSPNet includes the basic branch and the spanning branch, and the features of the two branches are spliced at the end of each phase, as shown in Figure 2. The input of each stage first goes through two 1 × 1 convolutions. The calculation process is as follows:
X B a s e = F b a s e ( f 1 × 1 ( X ) )
where f 1 × 1 represents 1 × 1 convolution and F b a s e is a combination of n basic modules. Then, two groups of features of the basic branch and the generated branch are spliced together, and the information exchange between the two groups of features is increased by channel shuffling [48]. Finally, the output of each stage is obtained by the down-sampling module. The calculation process is as follows.
X ˜ = F d o w n ( S ( X B a s e , X C r o s s ) )
where X B a s e represents the output of the basic branch, X C r o s s represents the output across the branch, S represents the channel shuffling, and F d o w n represents the down-sampling.

3.2. Cross-Stage Three-Linear Attention Fine-Grained Feature Learning Module (CTA)

To mine plentiful features suitable for fine-grained image classification, inspired by [46], we propose a fine-grained feature learning module based on cross-stage trilinear attention as shown in Figure 3.
Through the CSP backbone network, we can obtain a feature map X ˜ W × H × C , of which W , H represent the feature map dimensions and the channel number of the feature map, respectively. Generally, the global average pool (GAP) or global maximum pool (GMP) is used to learn the final features X ˜ . One of the common problems with average or maximum pooling is that the interactive information between different semantic channels cannot be fully utilized.
Therefore, a common method is to use bilinear spatial attention. The second-order matrix F a of each position ( i , j ) is classified by merging itself X A W H × C A with another CNN stream X B W H × C B , and then the final feature is used in the final category through a fully connected (fc) layer.
F a = 1 W H i = 1 W j = 1 H v ( ( X A ) i , j ( X B ) i , j )
F b = W F a + b
where v : C A × C B C A C B × 1 represents the vectorization of the second-order matrix and W C A C B × N c l s is the learnable weight of the FC layer. Although abundant features are obtained through bilinear pooling, these high-dimensional features are easily optimized with poor usability. Therefore, in the method of trilinear attention, the cross-channel relationship is regarded as the attention diagram generated by the same feature map X T X C × C by non-local operation, and then the channel-perceived attention map gives different annotations to the original features to produce the third-order result S ( X T X ) X T W H × C , where S represents SoftMax normalization.
However, a disadvantage of the cross-channel relationship is that it ignores the learning between multi-scale features. Because different network layers have different scales of receptive fields, the later network layers have larger receptive fields. Inspired by [47], we use features φ ( X ) from the next layer of the same stage of the network to conduct cross-layer semantic learning of cross-channels (see Figure 3), which can be expressed as:
F c = N ( M ( 1 W H i = 1 W H ( X φ ( X ) ) ) ) C × H W
where φ ( X ) is the output of one layer after the same stage of CSP, N denotes SoftMax normalization, and M ( x ) = s i g n ( x ) x 1 / 2 denotes moment function normalization. Similar to trilinear attention, in order to make the feature map more consistent and enhance robustness, the spatial relationship is further integrated into the feature map by F c dot multiplication of F c and φ ( X ) , so a cross-stage trilinear attention map is obtained. The attention map can be expressed as:
F d = N ( M ( F c φ ( X ) ) ) C × H W
Therefore, this paper constructs an attention map F d , and each channel of the map F d represents an attention map F i d W × H . The pyramid method is used to learn objects with different scales in images:
F k d = CTA ( F k )
where CTA denotes cross-stage trilinear attention and F k represents the characteristics of the k level of the pyramid. The values of k are 2, 3, and 4, representing the second, third, and fourth stages of the network, respectively.

3.3. Graph-Based High-Order Feature Embedding (GFE)

Multi-scale fine-grained feature libraries can be generated through the cross-stage attention modules in Section 3.3. Inspired by [47], each element of these feature libraries can be regarded as a graph node, and then the adjacency matrix score of the node graph neural network is used to embed these features.
Firstly, every element in attention F d = { f 1 , , f C N } is regarded as a graph node, and each graph node shares a large amount of information, so we can aggregate these nodes in the following ways.
A i , j = τ ( f i ) T τ ( f j ) | | τ ( f i ) | | | | τ ( f j ) | |
where τ is the convolution used for dimension transformation and represents A i , j , the adjacency matrix score of nodes i and j . The adjacency matrix is A ˜ = A + I , where I C 1 × C 1 is the identity matrix.
Through similarity aggregation in this way, each node is updated as follows:
G 0 = Re L u ( D ˜ 1 2 A ˜ D ˜ 1 2 K d W d )
where W d is the graph node weight of d h learnable dimension and D ˜ d = j A ˜ i , j d is the diagonal matrix to be normalized in nodes. K d represents the matrix form of feature bank κ d .
Similarly, the features F k d in the pyramid structure also have embedded features G k d through similar operations, as shown below:
G k = Re L u ( D ˜ k 1 2 A ˜ k D ˜ 1 2 k   K k W k )
where W k is the graph node weight of learnable d h dimension, D ˜ k = j A ˜ i , j k is the diagonal matrix to be normalized, and K k represents the matrix form of feature library κ k .
Through the above graph propagation structure, we obtain the embedded features G = { G 0 , G 1 , , G k } of a multilevel structure. Considering that the features of different levels play different roles, we use an adaptive attention mechanism I = { I 0 , I 1 , , I k } to learn the importance between them, as shown below:
I = a t t ( G )
where a t t denotes attention, and its specific generation mode is as follows. There is a node g m i , where 0 m k and i denotes the i -th point. Secondly, we use nonlinear transformation to transform the embedding and then use an attention vector η to obtain the attention value, as shown below:
η m i = tanh ( W m ( g m i ) T + b m )
where tanh is a nonlinear function, W m is a learnable parameter weight, and b m is a bias. Through a similar operation, we can obtain all the embedded G = { G 0 , G 1 , , G k } attention values and normalize them with the SoftMax function:
I m i = s o f t m a x ( η m i ) = exp ( η m i ) m = 0 k η m i
The larger the I m i , the more important the embedding. For all n nodes. The attention vector I = [ I 0 , , I k ] can be obtained at last, which is selected to distinguish the importance of different levels of features through attention adaptation. The final result score is as follows:
C = m = 0 k I m s o f t max ( G m )

3.4. YOLOV5 Detector via Soft-NMS Optimization

The target detection task classifies and locates the objects in the image, generates a detection frame for the objects in the image, and generates corresponding confidence. YOLOV5 is used as the detector in this paper (Figure 4). When the model detects the detection frame of the object in the predicted image, multiple detection frames are generated for the same object, and each detection frame has a corresponding score. Having too many detection frames reduces the detection accuracy. Therefore, the NMS method is used to process the detection frame and obtain the final detection result. So far, this method is still a popular target detection algorithm and can effectively improve the recall rate of detection. Non-maximum suppression (NMS) is an integral part of the target detection model. The detection box is screened to reduce the repetition rate and improve the recall rate of detection.
The basic flow of this method is to sort the generated detection frames according to the scores, take out the detection frame M with the highest score, compare the rest with the selected detection frames, and suppress the detection frames whose overlap is higher than the threshold. This process is repeatedly applied to the rest of the detection frames until they are compared with the detection frames below it. The traditional non-maximum suppression method directly zeroes the score whose overlap ratio with the adjacent detection frame is greater than the threshold so that when two detected objects are close, the traditional non-maximum suppression method deletes the detection frame with the lower score, resulting in missed detection, which leads to the decrease in the accuracy of the detection results.
S i = S i ,   I O U ( M , b i ) < N t 0 ,   I O U ( M , b i ) N t
The traditional non-maximum suppression algorithm has the disadvantage of missing overlapping objects, so we adopt a new soft non-maximum suppression algorithm, which is improved on a conventional basis. In this algorithm, when the overlap of two detection frames is more significant than the threshold, it is deleted and improved, but an attenuation function is set. When the overlap between the rest of the detection frames and M is greater than the threshold, a very low weight is given. If only a tiny part of the detection frames overlap, it does not affect the detection results. Soft-NMS does not need additional training; it directly trains the model from end to end without increasing the model parameters.
When there is a high overlap with the detection box M, it is set to fractional attenuation instead of directly setting it to zero. The higher the overlap, the more serious the corresponding fractional attenuation. Therefore, the traditional non-maximum suppression method is improved, as shown in Formula (15).
S i = S i ,     I O U ( M , b i ) < N t S i ( 1 I O U ( M , b i ) ) ,   I O U ( M , b i ) N t
By improving the traditional non-maximum suppression method, a fractional attenuation method is proposed to solve the problem of missing detection when the detection frames overlap. It is verified on two existing public detection datasets. Through the experimental analysis, it is found that the improved non-maximum suppression method can identify overlapping objects and improve the detection accuracy of the model. The Soft-NMS method only improves the traditional non-maximum suppression algorithm, which does not increase the parameters of model operation and affects the detection speed of the model. It is easy to implement and can be embedded in any detection model.

3.5. Loss Function

In the training process, the cross-entropy loss is used as the loss function, and in order to reduce the risk of over-fitting, it is also applied to label smoothing technology, and the smoothed new label is used to replace the original label:
y = ( 1 ε ) y ˜ + ε μ
where y ˜ represents the sample label, ε is the smoothing factor, and u is a fraction of the category. Using label smoothing can make the classification probability result of the SoftMax activation function close to the correct classification, restrain the output difference between positive and negative samples, and make the network have better generalization ability.
The CIOU loss function of the detector used in YOLOV5 is to measure the loss of rectangular frames. The specific formula is as follows:
S 1 = ( min ( x p 2 , x l 2 ) max ( x p 1 , x l 1 ) ) × ( min ( y p 2 , y l 2 ) max ( y p 1 , y l 1 ) )
S 2 = ( x p 2 x p 1 ) × ( y p 2 y p 1 ) + ( x l 2 x l 1 ) × ( y l 2 y l 1 ) S 1
I O U = S 1 S 2
C L O U = I O U ρ 2 c 2 α v
v = 4 π 2 ( arctan w l h l arctan w p h p ) 2 = 4 π 2 ( arctan x l 2 x l 1 y l 2 y l 1 arctan x p 2 x p 1 y p 2 y p 1 ) 2
α = v 1 I O U + v
where ρ is the distance between the center points of box A and box B, c is the diagonal length of the minimum bounding rectangle of box A and box B, v is the width–height similarity ratio of frame A and frame B, and α is the influence factor of v . The calculation formula of CIOU loss can be obtained from the above:
l o s s C I O U = 1 C I O U
CIOU considers the distance between the target and the anchor, the overlap rate, the scale, and the penalty term so that the target box regression becomes more stable, and there is no divergence or other problems in the training process such as IOU and GIOU. The penalty factor considers the aspect ratio of the prediction frame and the aspect ratio of the target frame.

4. Experiment and Result

4.1. Implementation Details

In this experiment, we pre-trained our model on the ImageNet dataset. Pre-processing adopts random cutting and random turning. The figure size was resized to 448 × 448, the batch size was set to 128, training times were 150, and SGD was used as an optimizer. Meanwhile, the cosine annealing strategy was used to attenuate the learning rate, and the attenuation period was 20. The final experiment was conducted on 8 Nvidia P40 graphics cards.
In the classification experiment, we adopted two indexes: the accuracy rate (ACC) and the parameter size of the model parameters. The accuracy calculation formula is as follows:
A C C = T P + T N T P + T N + F P + F N
Here, TP represents true positive, TN represents true negative, FP represents false positive, and FP represents false negative. In the testing experiment, the precision (P), recall (P), and average accuracy are used as evaluation indexes, and the calculation formula is as follows:
P = T P T P + F P
R = T P T P + F N
AP = 0 1 P ( R ) d R
where P denotes precision and R denotes recall. Different evaluation indexes can not only evaluate the model in all aspects but also further prove the superiority and practicability of the proposed method.

4.2. Bird Classification Results

To verify the performance of the fine-grained methods, we used two fine-grained datasets for experiments and selected some open-source algorithms for performance comparison.
Dataset (1) CUB-200-2011 [35] contains 5994 images in the training process and 5794 images in the test process; (2) Bird-400 contains 58,388 images in training datasets and 2000 images in test sets. The comparison index selected the commonly used accuracy index of image classification. Meanwhile, the possibility of the fine-grained method is discussed according to a comparison of the parameters of different models as shown in Table 1.
From the classification results shown in Table 1, we can see that the results of GPA-Net are better than those of coarse-grained methods, e.g., VGG, ResNet, GoogleNet, and DenseNet, which shows that the GPA-Net can learn more new information and has greater vital feature extraction ability. Meanwhile, compared with some fine-grained methods, our method has more advantages. On the CUB-200-2011 dataset, GPA-Net has achieved the best results, with an accuracy rate of 89.6%. Compared with other fine-grained methods, the results on the Bird-400 dataset are not much different. This is because the Bird-400 dataset is more direct and has less interference than other datasets. Comparatively speaking, our method is better at feature extraction in complex backgrounds. The results show that FBSD and AP-CNN benefit from the simple backgrounds on the Bird-400 dataset. However, our method has more advantages in datasets with a more complex background, e.g., CUB2011 and AF-bird50. In terms of the parameters of the model, compared with the coarse-grained method, our model is slightly larger than the standard coarse-grained method, but its accuracy is better. Compared with other fine-grained methods, our model is lighter, which shows that our model has advantages in practical application. We drew a comparison chart of the loss change during model training to further explore and evaluate the model’s performance, as shown in Figure 5.
The downward trend of the loss is shown in Figure 5. Obviously, we know that GPA-Net can drop rapidly in the training process and reach a stable state later, indicating that the model has strong stability in the training process. Meanwhile, it can be seen that GPA-Net converges faster than other models in the early stage, and the loss value is lower after reaching the steady state, which shows that GPA-Net has better convergence and can learn more. We use a contrast histogram to analyze the parameter amounts of different models on three datasets, as shown in Figure 6.
The comparison results of model parameters are shown in Figure 6. The parameters of the GPA-Net model are slightly larger than those of ResNet50, Inception, and DenseNet because the attention module added to the GPA-Net model occupies a part of the parameters. At the same time, the parameters of GPA-Net are much smaller than those of other fine-grained algorithms. This is because the backbone network of our model uses a CSP structure, and the graph embedding method is used to reduce the parameters further, showing that our model is more applicable in practical applications.

4.3. Bird Detection Results

Although there are many bird species in existing public bird databases, they mainly rely on the Internet to crawl data. On the one hand, the scene of each image is an ideal shooting environment (many pictures are provided for photography enthusiasts), and different pictures have significant picture quality. On the other hand, these datasets are mainly collected for the purpose of bird image classification. A picture often has only a small sample of the same bird, which cannot solve the picture detection tasks that need to identify multiple samples of different birds appearing in the same photo. When a dataset is used to conduct deep learning network training, it cannot solve the real-scene application of actual environmental protection. In the remote monitoring system of environmental protection, the background of an effective picture is often very complicated. Many samples of different types of birds appear randomly together, and the proportion of each bird occupying the entire picture is usually small. There are also challenges such as mutual obstruction between goals and backgrounds or the appearance similarity of different birds. Therefore, deep learning models only relying on public datasets cannot effectively solve the image recognition task in practical applications.
Therefore, in order to solve the challenge of bird detection in actual environmental protection applications, we took the main bird species living in river and lake environments in Beijing, China, as research objects and built a new dataset, AF-bird50, for application demonstration experiments. AF-bird50 contains over 10,000 images, covering 50 different species. Our dataset comes from photos taken in three stations—Miyun Reservoir Station 1, Guanting Station 1, and Liudu Station in Fangshan District of Beijing—and these stations were set up starting in March 2022. All images were mainly collected through cameras, and sensors were also applied to obtain bird pictures in real scenes by ourselves. We tested the AF-bird50 dataset and selected SSD-300, Faster-R-CNN, and YOLOV3 as the comparison models. We divided this dataset into a ratio of 3:7; thus, there are 7000 training sets and 3000 test sets. The experimental results can be seen in Table 2.
From Table 2, we can see that compared with other target detection models, e.g., Faster-R-CNN, YOLOV3, etc., the optimized YOLOV5 achieves the highest results for different indicators, e.g., precision rate and recall rate. Thereby, the optimized YOLOV5 based on the proposed GPA-Net is suitable for accurate bird detection. Some bird detection results are shown in Figure 7. As shown in the picture, the species of these wild birds include sandpiper (scientific name “Tringa ochropus”, Chinese name “白腰草鹬”), heron (scientific name “Ardea cinerea”, Chinese name “苍鹭”), egret (scientific name “Egretta garzetta”, Chinese name “白鹭”), cormorant (scientific name “Phalacrocorax carbo”, Chinese name “鸬鹚”), etc. All pictures were obtained through remote monitoring cameras in natural environments such as lakes and rivers.
It can be seen from Figure 7 that the network can detect the target very accurately for the sample with only a single target and simple background in the image. On the other hand, in the presence of interference factors such as multiple targets, different poses, and background occlusion in the sample images, the network still achieves a sound effect, which indicates that the Soft-NMS mechanism can improve the performance of the network well and further demonstrates that the optimized YOLOV5 network can be used in actual bird recognition tasks. Furthermore, we draw a confusion matrix of the proposed detector’s accuracy to visualize the model’s performance, as shown in Figure 8.
The result of this confusion matrix is shown in Figure 8. It can be seen that the recognition rate of all birds is almost above 90%, which indicates that the model has high accuracy. At the same time, it can be seen that among the samples with model recognition errors, none identify a large number of samples as the existence of another bird, which indicates that the model has high fault tolerance. The results of recognition errors are mainly caused by vague pictures and severe occlusion, which lead to the model learning less than the required features.

5. Discussion

To verify the actual function of each module in the model, we set up an ablation experiment to analyze the model on the CUB-200-2011 dataset. The ablation experiment was executed for further analysis of module functions. When only relying on the CSP backbone network, the base model achieves an accuracy of 85.1%. After using the CTA module, the accuracy is improved by 0.8%, achieving a result of 86.3%. In contrast, after only using the GPE module, the accuracy of the model is up to 87.2%, which is 2.1% higher than that of the backbone network. When the CTA and GPE modules are applied to improve performance at the same time, the entire model can achieve an 89.6% accuracy, which is the highest result. This shows that the pyramid structure can help the model learn more fine-grained features and improve its fine-grained identification ability.
In particular, the CSP is used as the backbone of our method to learn the shallow feature information of images. Based on these shallow features in different stages, we can further learn high-order feature information with discriminant information. Meanwhile, the CTA attention module is used to mine the learned features further. The pyramid structure is used to learn the features of different scales for smaller targets. The graph structure is used to learn parts of different scales, and attention adaptive learning is used to understand the importance of different levels. The ablation experiment shows that the CTA module can strengthen feature learning and improve accuracy by 0.8%.
Meanwhile, the GPE module can adaptively distinguish the importance of different levels, reduce the influence of redundant features, and improve recognition accuracy. The analysis of the ablation experiment proves that our network can learn different regional features with discrimination information and improve the accuracy of fine-grained recognition. For visual analysis, we use a heat map of attention distribution to visualize the GPE module and explore its essential role. The visualization results of Stage2 to Stage4 for three natural samples are shown in Figure 9.
As shown in above Figure 9, if only the last layer of the backbone network is used as the classification feature, the network misses a large amount of other useful information, and other supplementary information is learned for classification by using the pyramid structure. Meanwhile, to prevent the features at the bottom stage from bringing too much noise, the attention mechanism is added to adaptively judge the importance of these features at different levels.

6. Conclusions

The avian species are essential to protecting the ecological environment and maintaining biodiversity change. This paper proposes a fine-grained bird recognition method based on a graph attention pyramid to solve the fine-grained problem in bird image recognition. Firstly, based on the CSP baseline network, the fine-grained feature of cross-stage attention module learning is designed. At the same time, the graphic pyramid structure is introduced to learn the multi-scale features for understanding the context information between different local parts and then building a novel fine-grained classifier GPA-Net. Finally, an optimized YOLOV5 with Soft-NMS strategy is designed as the detector composition, improving the detection capability for small targets. Massive experiments demonstrated that the performance of the proposed method for fine-grained bird identification is better than other comparative models. The accuracy rates of GPA-Net on CUB-200-2011, Bird-400, and AF-bird50 datasets are 89.6%, 99.3%, and 95.4%, respectively, reaching the optimal results. Additionally, the optimized YOLOV5 detector based on GPA-Net has the advantages of smaller identification error and lighter model parameters. This shows that the proposed method is more suitable for natural bird image recognition in complex scenes.
Optimizing the model structure and improving model performance will be our work in the future. At the same time, improving the model’s generalization ability is also an essential part of future work. In the future, we will explore the application potential of these methods in other fields, such as motion estimation [62], modeling optimization [63], and temporal prediction [64], etc.

Author Contributions

Conceptualization, J.-L.K.; methodology, X.X.; writing—original draft preparation, X.X. and Y.X.; writing—review and editing, J.-L.K.; data analysis and software, Y.X. and C.-C.Y.; funding acquisition and supervision, J.-L.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the National Natural Science Foundation of China (No. 62006008, 62173007), Beijing Natural Science Foundation (No. 6214034), and MOE (Ministry of Education in China) Project of Humanities and Social Sciences (No. 19YJC790028, 22YJCZH006).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets contained within the article are available at the following links. CUB-200-2011: https://www.vision.caltech.edu/datasets/cub_200_2011 (accessed on 19 January 2023). Bird-400: https://www.kaggle.com/datasets/gpiosenka/100-bird-species (accessed on 19 January 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Das, S.; Pradhan, B.; Shit, P.K.; Alamri, A.M. Assessment of wetland ecosystem health using the pressure–state–response (PSR) model: A case study of mursidabad district of West Bengal (India). Sustainability 2020, 12, 5932. [Google Scholar] [CrossRef]
  2. Li, G.; Hao, Y.; Yang, T.; Xiao, W.; Pan, M.; Huo, S.; Lyu, T. Enhancing bioenergy production from the raw and defatted microalgal biomass using wastewater as the cultivation medium. Bioengineering 2022, 9, 637. [Google Scholar] [CrossRef] [PubMed]
  3. Jin, X.-B.; Wang, Z.-Y.; Kong, J.-L.; Bai, Y.-T.; Su, T.-L.; Ma, H.-J.; Chakrabarti, P. Deep Spatio-Temporal Graph Network with Self-Optimization for Air Quality Prediction. Entropy 2023, 25, 247. [Google Scholar] [CrossRef] [PubMed]
  4. Little, C.J.; Rizzuto, M.; Luhring, T.M.; Monk, J.D.; Nowicki, R.J.; Paseka, R.E.; Stegen, J.C.; Symons, C.C.; Taub, F.B.; Yen, J.D. Movement with meaning: Integrating information into meta-ecology. Oikos 2022, 2022, e8892. [Google Scholar] [CrossRef]
  5. Muvengwi, J.; Fritz, H.; Mbiba, M.; Ndagurwa, H.G. Land use effects on phylogenetic and functional diversity of birds: Significance of urban green spaces. Landscape Urban Plan. 2022, 225, 104462. [Google Scholar] [CrossRef]
  6. Bełcik, M.; Lenda, M.; Amano, T.; Skórka, P. Different response of the taxonomic, phylogenetic and functional diversity of birds to forest fragmentation. Sci. Rep. 2020, 10, 20320. [Google Scholar] [CrossRef]
  7. Li, G.; Hu, R.; Wang, N.; Yang, T.; Xu, F.; Li, J.; Wu, J.; Huang, Z.; Pan, M.; Lv, T. Cultivation of microalgae in adjusted wastewater to enhance biofuel production and reduce environmental impact: Pyrolysis performances and life cycle assessment. J. Clean. Prod. 2022, 355, 131768. [Google Scholar] [CrossRef]
  8. Jin, X.-B.; Wang, Z.-Y.; Gong, W.-T.; Kong, J.-L.; Bai, Y.-T.; Su, T.-L.; Ma, H.-J.; Chakrabarti, P. Variational Bayesian Network with Information Interpretability Filtering for Air Quality Forecasting. Mathematics 2023, 11, 837. [Google Scholar] [CrossRef]
  9. Nugent, D.T.; Baker Gabb, D.J.; Leonard, S.W.; Morgan, J.W. Livestock grazing to maintain habitat of a critically endangered grassland bird: Is grazer species important? Ecol. Appl. 2022, 32, e2587. [Google Scholar] [CrossRef]
  10. Klaus, V.H.; Kiehl, K. A conceptual framework for urban ecological restoration and rehabilitation. Basic Appl. Ecol. 2021, 52, 82–94. [Google Scholar] [CrossRef]
  11. Morelli, F.; Reif, J.; Diaz, M.; Tryjanowski, P.; Ibáñez-álamo, J.D.; Suhonen, J.; Jokimäki, J.; Kaisanlahti-Jokimäki, M.; Møller, A.P.; Bussière, R. Top ten birds indicators of high environmental quality in European cities. Ecol. Indic. 2021, 133, 108397. [Google Scholar] [CrossRef]
  12. Randler, C.; Diaz-Morales, J.F.; Jokimäki, J.; Ortiz-Pulido, R.; Staller, N.; De Salvo, M.; Kaisanlahti-Jokimäki, M.L. Birding recreation specialization–A test of the factorial invariance in eight languages. J. Leis. Res. 2022, 1–7. [Google Scholar] [CrossRef]
  13. Pal, M.; Pop, P.; Mahapatra, A.; Bhagat, R.; Hore, U. Diversity and structure of bird assemblages along urban-rural gradient in Kolkata, India. Urban For. Urban Gree. 2019, 38, 84–96. [Google Scholar] [CrossRef]
  14. Wägele, J.W.; Bodesheim, P.; Bourlat, S.J.; Denzler, J.; Diepenbroek, M.; Fonseca, V.; Frommolt, K.; Geiger, M.F.; Gemeinholzer, B.; Glöckner, F.O. Towards a multisensor station for automated biodiversity monitoring. Basic Appl. Ecol. 2022, 59, 105–138. [Google Scholar] [CrossRef]
  15. Randler, C.; Tryjanowski, P.; Jokimäki, J.; Kaisanlahti-Jokimäki, M.L.; Staller, N. SARS-CoV2 (COVID-19) Pandemic lockdown influences nature-based recreational activity: The case of birders. Int. J. Environ. Res. Public Health 2022, 19, 7310. [Google Scholar] [CrossRef] [PubMed]
  16. Cai, W.W.; Gao, M.; Jiang, Y.Z.; Gu, X.Q.; Ning, X.; Qian, P.J.; Ni, T.G. Hierarchical Domain Adaptation Projective Dictionary Pair Learning Model for EEG Classification in IoMT Systems. IEEE Trans. Intell. Transp. 2022, 1–9. [Google Scholar] [CrossRef]
  17. Cai, W.W.; Wei, Z.G. Remote sensing image classification based on a cross-attention mechanism and graph convolution. IEEE Geosci. Remote Sens. Lett. 2022, 19, 21506569. [Google Scholar] [CrossRef]
  18. Kong, J.L.; Wang, H.X.; Wang, X.Y.; Jin, X.B.; Fang, X.; Lin, S. Multi-stream hybrid architecture based on cross-level fusion strategy for fine-grained crop species recognition in precision agriculture. Comput. Electron. Agr. 2021, 185, 106134. [Google Scholar] [CrossRef]
  19. Zheng, Y.Y.; Kong, J.L.; Jin, X.B.; Wang, X.Y.; Su, T.L.; Zuo, M. CropDeep: The crop vision dataset for deep-learning-based classification and detection in precision agriculture. Sensors 2019, 19, 1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Kong, J.L.; Yang, C.C.; Xiao, Y.; Lin, S.; Ma, K.; Zhu, Q.Z. A Graph-Related High-Order Neural Network Architecture via Feature Aggregation Enhancement for Identification Application of Diseases and Pests. Comput. Intell. Neurosci. 2022, 2022, 4391491. [Google Scholar] [CrossRef] [PubMed]
  21. Kong, J.L.; Yang, C.C.; Wang, J.L.; Wang, X.Y.; Zuo, M.; Jin, X.B.; Lin, S. Deep-stacking network approach by multisource data mining for hazardous risk identification in IoT-based intelligent food management systems. Comput. Intell. Neurosci. 2021, 2021, 16. [Google Scholar] [CrossRef]
  22. Jacob, I.J.; Darney, P.E. Design of deep learning algorithm for IoT application by image based recognition. J. ISMAC 2021, 3, 276–290. [Google Scholar] [CrossRef]
  23. Zhang, S.H.; Zhao, Z.; Xu, Z.Y.; Bellisario, K.; Pijanowski, B.C. Automatic bird vocalization identification based on fusion of spectral pattern and texture features. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018. [Google Scholar]
  24. Kumar, R.; Kumar, A.; Bhavsar, A. Bird Region Detection in Images with Multi-scale HOG Features and SVM Scoring. In Proceedings of the 2nd International Conference on Computer Vision & Image Processing, Roorkee, India, 9–12 September 2017; Chaudhuri, B., Kankanhalli, M., Raman, B., Eds.; Springer: Singapore, 2018; pp. 353–364. [Google Scholar]
  25. Kong, J.L.; Wang, H.X.; Yang, C.C.; Jin, X.B.; Zuo, M.; Zhang, X. A spatial feature-enhanced attention neural network with high-order pooling representation for application in pest and disease recognition. Agriculture 2022, 12, 500. [Google Scholar] [CrossRef]
  26. Jin, X.B.; Zheng, W.Z.; Kong, J.L.; Wang, X.Y.; Bai, Y.T.; Su, T.L.; Lin, S. Deep-learning forecasting method for electric power load via attention-based encoder-decoder with bayesian optimization. Energies 2021, 14, 1596. [Google Scholar] [CrossRef]
  27. Wang, L.; Cao, X.Y.; Li, M.H.; Zhao, J.N. English Letter Recognition Based on TensorFlow Deep Learning. J. Phys. Conf. Ser. 2020, 1627, 12012. [Google Scholar]
  28. Jiang, H.X. The analysis of plants image recognition based on deep learning and artificial neural network. IEEE Access 2020, 8, 68828–68841. [Google Scholar]
  29. Chen, C.; Liu, B.; Wan, S.H.; Qiao, P.; Pei, Q.Q. An edge traffic flow detection scheme based on deep learning in an intelligent transportation system. IEEE Trans. Intell. Transp. Syst. 2020, 22, 1840–1852. [Google Scholar] [CrossRef]
  30. Tumas, P.; Serackis, A. Automated image annotation based on YOLOv3. In Proceedings of the 2018 IEEE 6th Workshop on Advances in Information, Electronic and Electrical Engineering (AIEEE), Vilnius, Lithuania, 8–10 November 2018. [Google Scholar]
  31. The Caltech-Ucsd Birds-200–2011 Dataset. Available online: https://authors.library.caltech.edu/27452/1/CUB_200_2011.pdf (accessed on 7 January 2023).
  32. Berg, T.; Liu, J.X.; Woo Lee, S.; Alexander, M.L.; Jacobs, D.W.; Belhumeur, P.N. Birdsnap: Large-scale fine-grained visual categorization of birds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  33. Naumov, S.; Yaroslavtsev, G.; Avdiukhin, D. Objective-Based Hierarchical Clustering of Deep Embedding Vectors. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021. [Google Scholar]
  34. Karthikeyan, C.; Jabber, B.; Deepak, V.; Vamsidhar, E. Image Processing based Improved Face Recognition for Mobile Devices by using Scale-Invariant Feature Transform. In Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020. [Google Scholar]
  35. Essa, A.; Asari, V.K. Face recognition based on modular histogram of oriented directional features. In Proceedings of the 2016 IEEE National Aerospace and Electronics Conference (NAECON) and Ohio Innovation Summit (OIS), Dayton, OH, USA, 25–29 July 2016. [Google Scholar]
  36. Huang, C.; Luo, B.; Tang, L.Z.; Liu, Y.N.; Ma, J.X. Topic model based bird breed classification and annotation. In Proceedings of the 2013 International Conference on Communications, Circuits and Systems (ICCCAS), Chengdu, China, 15–17 November 2013. [Google Scholar]
  37. Berg, T.; Belhumeur, P.N. Poof: Part-based one-vs.-one features for fine-grained categorization, face verification, and attribute estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013. [Google Scholar]
  38. Tryjanowski, P.; Murawiec, S.; Randler, C. No such Thing as Bad Birding Weather, but Depends on Personal Experience. Leis. Sci. 2023, 1–13. [Google Scholar] [CrossRef]
  39. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2012, 25, 1097–1105. [Google Scholar] [CrossRef] [Green Version]
  40. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  41. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  42. Szegedyc, L. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  43. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  44. Zhang, N.; Donahue, J.; Girshick, R.; Darrell, T. Part-based R-CNNs for fine-grained category detection. In European Conference on Computer Vision, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Cham, Switzerland, 2014; pp. 834–849. [Google Scholar]
  45. Donahue, J.; Jia, Y.Q.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. Decaf: A deep convolutional activation feature for generic visual recognition. In Proceedings of the International Conference on Machine Learning, Beijing, China, 21–26 June 2014. [Google Scholar]
  46. Huang, S.L.; Xu, Z.; Tao, D.C.; Zhang, Y. Part-stacked CNN for fine-grained visual categorization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  47. Feature Boosting, Suppression, and Diversification for Fine-Grained Visual Classification. Available online: https://arxiv.org/abs/2103.02782 (accessed on 7 January 2023).
  48. Ji, R.Y.; Wen, L.Y.; Zhang, L.B.; Du, D.W.; Wu, Y.J.; Zhao, C.; Liu, X.L.; Huang, F.Y. Attention convolutional binary neural tree for fine-grained visual categorization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  49. Yang, B.; Pan, H.W.; Yu, J.Y.; Han, K.; Wang, Y.A. Classification of medical images with synergic graph convolutional networks. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering Workshops (ICDEW), Macao, China, 8–12 April 2019. [Google Scholar]
  50. Tai, K.S.; Socher, R.; Manning, C.D. Improved Semantic Representations from Tree-Structured Long Short-Term Memory Networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; Volume 1. Long Papers. [Google Scholar]
  51. Zhang, T.; Liu, B.; Niu, D.; Lai, K.F.; Xu, Y. Multiresolution graph attention networks for relevance matching. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino, Italy, 22–26 October 2018. [Google Scholar]
  52. Yu, B.; Yin, H.T.; Zhu, Z.X. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  53. Zhao, Y.F.; Yan, K.; Huang, F.Y.; Li, J. Graph-based high-order relation discovery for fine-grained recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  54. Lin, S.; Xiu, Y.; Kong, J.; Yang, C.; Zhao, C. An Effective Pyramid Neural Network Based on Graph-Related Attentions Structure for Fine-Grained Disease and Pest Identification in Intelligent Agriculture. Agriculture 2023, 13, 567. [Google Scholar] [CrossRef]
  55. Wang, Z.H.; Wang, S.J.; Li, H.J.; Dou, Z.; Li, J.J. Graph-propagation based correlation learning for weakly supervised fine-grained image classification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  56. Chen, T.S.; Lin, L.; Chen, R.Q.; Wu, Y.; Luo, X.N. Knowledge-embedded representation learning for fine-grained image recognition. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  57. Wang, X.; Liu, H.R.; Shi, C.; Yang, C. Be confident! towards trustworthy graph neural networks via confidence calibration. Adv. Neural Inf. Process. Syst. 2021, 34, 23768–23779. [Google Scholar]
  58. Wang, C.Y.; Liao, H.; Wu, Y. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Nashville, TN, USA, 19-25 June 2020; pp. 390–391. [Google Scholar]
  59. Ding, Y.F.; Ma, Z.Y.; Wen, S.G.; Xie, J.Y.; Chang, D.L.; Si, Z.W.; Wu, M.; Ling, H.B. AP-CNN: Weakly supervised attention pyramid convolutional neural network for fine-grained visual classification. IEEE Tran. Image Process. 2021, 30, 2826–2836. [Google Scholar] [CrossRef]
  60. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  61. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  62. Ding, F. Combined state and least squares parameter estimation algorithms for dynamic systems. Appl. Math. Model. 2014, 38, 403–412. [Google Scholar] [CrossRef]
  63. Ding, F. Hierarchical multi-innovation stochastic gradient algorithm for Hammerstein nonlinear system modeling. Appl. Math. Model. 2013, 37, 1694–1704. [Google Scholar] [CrossRef]
  64. Kong, J.-L.; Fan, X.-M.; Jin, X.-B.; Su, T.-L.; Bai, Y.-T.; Ma, H.-J.; Zuo, M. BMAE-Net: A Data-Driven Weather Prediction Network for Smart Agriculture. Agronomy 2023, 13, 625. [Google Scholar] [CrossRef]
Figure 1. The GPA-Net framework based on a graph attention pyramid. We generate a library through the cross-stage attention module (CTA) and use feature maps of different levels to form an attention pyramid. Then, the pyramid layer is embedded into features by the graph method to form a graph attention pyramid network (GPE). Finally, multiple features are connected to obtain attention representation.
Figure 1. The GPA-Net framework based on a graph attention pyramid. We generate a library through the cross-stage attention module (CTA) and use feature maps of different levels to form an attention pyramid. Then, the pyramid layer is embedded into features by the graph method to form a graph attention pyramid network (GPE). Finally, multiple features are connected to obtain attention representation.
Ijerph 20 04924 g001
Figure 2. CSP network structure.
Figure 2. CSP network structure.
Ijerph 20 04924 g002
Figure 3. Cross-stage three-linear attention fine-grained feature learning module (CTA).
Figure 3. Cross-stage three-linear attention fine-grained feature learning module (CTA).
Ijerph 20 04924 g003
Figure 4. Optimized YOLOV5 detector structure.
Figure 4. Optimized YOLOV5 detector structure.
Ijerph 20 04924 g004
Figure 5. (a) The loss curves of each model on CUB-200-2011. (b) The loss curves of each model on Bird-400. (c) The loss curves of each model on AF-bird50.
Figure 5. (a) The loss curves of each model on CUB-200-2011. (b) The loss curves of each model on Bird-400. (c) The loss curves of each model on AF-bird50.
Ijerph 20 04924 g005aIjerph 20 04924 g005b
Figure 6. Parameter comparison chart of different models.
Figure 6. Parameter comparison chart of different models.
Ijerph 20 04924 g006
Figure 7. The detaching results obtained by optimized YOLOV5 model.
Figure 7. The detaching results obtained by optimized YOLOV5 model.
Ijerph 20 04924 g007
Figure 8. Confusion matrix on AF-bird50.
Figure 8. Confusion matrix on AF-bird50.
Ijerph 20 04924 g008
Figure 9. Heat map visualization of attention distribution for GPE module.
Figure 9. Heat map visualization of attention distribution for GPE module.
Ijerph 20 04924 g009
Table 1. The classification results of three datasets.
Table 1. The classification results of three datasets.
CUB-200-2011Bird-400AF-Bird50
MethodsACC (%)Parameter (M)ACC (%)Parameter (M)ACC (%)Parameter (M)
Coarse-grainedVGG19 [42]76.4575.595.454980.1564.8
ResNet50 [43]85.297.196.893.187.295.4
Inception [44]85.948.997.634.188.344.2
DenseNet [45]86.149.297.447.490.146.3
Fine-grainedFBSD [49]89.216499.295.394.3159.7
AP-CNN [59]88.117899.3199.694.6172.1
GPA-Net89.610099.312895.498.2
Table 2. Bird detection results on AF-bird50.
Table 2. Bird detection results on AF-bird50.
MethodP(%)R(%)AP(%)
SSD300 [60]77.3879.4778.43
Faster-R-CNN [61]80.0181.6980.52
YOLOV3 [31]85.8987.9286.91
Optimized YOLOV588.4390.6489.37
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, X.; Yang, C.-C.; Xiao, Y.; Kong, J.-L. A Fine-Grained Recognition Neural Network with High-Order Feature Maps via Graph-Based Embedding for Natural Bird Diversity Conservation. Int. J. Environ. Res. Public Health 2023, 20, 4924. https://doi.org/10.3390/ijerph20064924

AMA Style

Xu X, Yang C-C, Xiao Y, Kong J-L. A Fine-Grained Recognition Neural Network with High-Order Feature Maps via Graph-Based Embedding for Natural Bird Diversity Conservation. International Journal of Environmental Research and Public Health. 2023; 20(6):4924. https://doi.org/10.3390/ijerph20064924

Chicago/Turabian Style

Xu, Xin, Cheng-Cai Yang, Yang Xiao, and Jian-Lei Kong. 2023. "A Fine-Grained Recognition Neural Network with High-Order Feature Maps via Graph-Based Embedding for Natural Bird Diversity Conservation" International Journal of Environmental Research and Public Health 20, no. 6: 4924. https://doi.org/10.3390/ijerph20064924

APA Style

Xu, X., Yang, C. -C., Xiao, Y., & Kong, J. -L. (2023). A Fine-Grained Recognition Neural Network with High-Order Feature Maps via Graph-Based Embedding for Natural Bird Diversity Conservation. International Journal of Environmental Research and Public Health, 20(6), 4924. https://doi.org/10.3390/ijerph20064924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop