Next Article in Journal
Beta-Carotene Supplementation for Sows: Reproductive Parameters and Productive Performance of Piglets
Next Article in Special Issue
Assessing the Impact of Spraying an Enterococcus faecium-Based Probiotic on Day-Old Broiler Chicks at Hatch on the Incidence of Bacterial Chondronecrosis with Osteomyelitis Lameness Using a Staphylococcus Challenge Model
Previous Article in Journal
Identification of Pre-Emptive Biosecurity Zone Areas for Highly Pathogenic Avian Influenza Based on Machine Learning-Driven Risk Analysis
Previous Article in Special Issue
In Ovo Feeding Techniques of Green Nanoparticles of Silver and Probiotics: Evaluation of Performance, Physiological, and Microbiological Responses of Hatched One-Day-Old Broiler Chicks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DFCCNet: A Dense Flock of Chickens Counting Network Based on Density Map Regression

1
College of Mathematics and Informatics, South China Agricultural University, Guangzhou 510642, China
2
Key Laboratory of Smart Agriculture of Guangzhou, South China Agricultural University, Guangzhou 510642, China
*
Author to whom correspondence should be addressed.
Animals 2023, 13(23), 3729; https://doi.org/10.3390/ani13233729
Submission received: 23 September 2023 / Revised: 23 November 2023 / Accepted: 28 November 2023 / Published: 1 December 2023

Abstract

:

Simple Summary

In large-scale chicken farming, monitoring flock density can help optimize feeding management and improve animal welfare. Traditional manual counting methods are not only prone to errors and omissions but also can cause harm to chickens. In this paper, a counting model for dense chicken flocks is proposed. The method can effectively estimate the number of chickens in real dense scenes.

Abstract

With the development of artificial intelligence, automatically and accurately counting chickens has become a reality. However, insufficient lighting, irregular sizes, and dense flocks make this a challenging task. The existing methods cannot perform accurate and stable counting. In this article, a dense flock of chickens counting network (DFCCNet) is proposed based on density map regression, where features from different levels are merged using feature fusion to obtain more information for distinguishing chickens from the background, resulting in more stable counting results. Multi-scaling is used to detect and count chickens at various scales, which can improve the counting accuracy and ensure stable performance for chickens of different sizes. Feature convolution kernels are adopted to convolve feature maps, which can extract more accurate target information, reduce the impact of occlusion, and achieve more reliable and precise results. A dataset of dense flocks of chickens (namely Dense-Chicken) has been collected and constructed, which contains 600 images of 99,916 chickens, with labeled points and boxes. It can be accessed by researchers as benchmark data. The proposed method was compared with some state-of-the-art algorithms, to validate its effectiveness. With its robustness being verified by counting in three kinds of density situations, with the mean absolute error being 4.26, 9.85, and 19.17, respectively, and a speed of 16.15 FPS. DFCCNet provides an automatic and fast approach to counting chickens in a dense farming environment. It can be easily embedded into handheld devices for application in agricultural engineering.

1. Introduction

Chicken is one of the most popular meats in the world. Therefore, the chicken breeding industry has great economic value. In the process of chicken breeding, cageless rearing is more conducive to the growth of chickens than cage rearing [1,2]. Proper breeding density can improve the growth performance of chickens, as well as their immunity and carcass yield [3,4,5], while too intensive feeding can negatively affect the health of chickens [6,7]. Moreover, many steps such as welfare breeding [8], feed feeding, stocking, and slaughtering are needed to obtain accurate quantities. Therefore, rapid and accurate estimation of flock density is a very important research field.
Traditional poultry farming requires artificial observation of behavior and health status, manual feeding, and counting. This requires a lot of labor, which increases the cost of breeding. An irregular and inaccurate manual operation may cause harm to chickens. Moreover, due to the small space and high density of chickens, if the staff do not carry out complete disinfection before entering the chicken house, they may pass germs to the chickens, which spread quickly, causing serious losses [9]. It is obviously difficult for chicken farm managers to count manually, which can lead to wrong amounts, low efficiency, and subjective influences. In particular, overlapping chickens and different perspectives can cause large errors with complex backgrounds. These problems are common in large-scale breeding enterprises. Therefore, it is necessary to develop and implement efficient, automated, and highly accurate counting methods, so as to fully improve efficiency and truly realize intelligent and automated management.
Digital technology with intelligent monitoring methods is widely used in poultry health and welfare management [10], which can realize rapid, accurate, automatic, non-invasive monitoring in the process of poultry breeding, and help replace some labor-intensive tasks in poultry breeding. A novel fully automated and non-invasive monitoring system was proposed to detect lame behavior according to the important correlation between characteristic variables and broiler gait scores [11]. A real-time automatic feed weight monitoring system was developed to automatically detect the intake and weight of a single turkey, to study the feed conversion rate and feeding behavior of a single turkey in a population environment [12]. A segmentation algorithm could effectively separate a broiler from the background, extract the pose information of the broiler, and accurately and quickly identify the health status of the broiler with an accuracy rate of 99.47% [13]. However, these methods belong to shallow learning, which limits performance when dealing with complex problems, due to a lack of deep topology and big data support.
Deep learning is a method of data representation-based learning that consists of multiple processing layers, to learn data representations with multiple levels of abstraction [14]. An attention encoder and convolutional neural network (CNN) were used to detect and classify chicks in different living stages, with an accuracy of 95.2% [15]. CNN was used to monitor the opening beak behavior, spatial dispersion, and movement of chickens [16]. A deep neural network (DNN) and cross-entropy in information theory were used to train rooster and hen classifiers with an average accuracy of 96.85%, which provided a feasible method for the estimation of sex ratio [17]. A camera-based system was developed to monitor chickens and detect injuries caused by pecking at each other using neural networks [18]. A DNN-based pose estimation method was first applied to classify and identify the poses of broilers [19]. Experiments showed that chickens in standing, walking, running, eating, and resting states could be identified.
Counting refers to estimating the number of objects in a target area, to obtain information and perform timely control operations, and is currently widely used in various fields [20,21,22]. In the field of chicken counting, some researchers have conducted related studies. One study applied a localized fully convolutional network (LCFCN) to count and locate chickens from images of the pins with an average absolute error (MAE) of 0.56 [23]. A fully convolutional network termed DenseFCN was designed, and a point-supervised method was used to count chickens in the image, with an accuracy of 93.84% and speed and 9.27 frames per second (FPS) [24]. A novel framework called YOLOX-birth growth death (Y-BGD) was proposed for automatic and accurate cage-free broiler counting, with an average accuracy of 98.13% [25]. A automatic system based on YOLOv5 was applied to count chickens in the image with an average accuracy rate of 95.87% [26]. We analyzed the datasets used in these studies, in which [23,24,25] counted chickens in sparse scenes, with an average number of chickens per image of 27.89, 24.36, and 2.38, respectively. Ref. [26] counted chickens in both sparse and dense scenes, with an average number of chickens per picture of 155.37. However, the counting objects in all of these studies were large chickens. In real farming environments, large and small chickens are not the same in farming scenes; the same number of small chickens are farmed in smaller areas, and small chickens are more likely to congregate together, thus creating more serious occlusion and shadowing. More research is needed to address this situation.
At present, the mainstream counting technologies can be divided into two types, namely object detection and density map regression. Object detection can identify objects in images and determine their size and position using prediction boxes. The number of prediction boxes can be used to count. Object detection is mainly applied to two counting scenes: one is to directly detect and count chickens in sparse scenes [26], and the other is to count chickens passing through the counting line in the corridor by combining tracking algorithms [25]. But the accuracy is greatly affected in dense environments with heavy occlusion. Density map regression is mainly used in high-density scenes, which can obtain accurate counting by learning the mapping relationship between the picture and the real density map. This method has been applied in aquaculture. A lightweight fish counting model based on density plot regression was used for counting high-density fish [27]. Multi-scale modules and attention mechanisms were integrated into a network based on density map regression in fish counting, with an accuracy of 97.1% [28]. Although these experiments were conducted in a dense environment, their experimental environment was also ideal. Serious occlusion and complex backgrounds could still lead to poor stability and accuracy of the model, and it is necessary to further improve the performance.
Lightweight models are crucial in practical applications. Deep learning models usually require a large amount of computational resources, such as CPU, GPU, and so on. If the computational burden is too heavy, this will cause the model to run slowly, or even not be able to run on resource-constrained devices. This will make the model unable to meet the requirements of real-time calculation and portability, hindering its application in practice. Therefore, it is important to take into account the weight of the model, while ensuring the correctness of the model.
Aiming at avoiding heavy computational effort, and further improving the chicken counting accuracy under serious occlusion and complex environments, a lightweight model based on density map regression is proposed. The main contributions can be summarized as follows:
(1)
A lightweight framework, DFCCNet, was designed by improving a feature convolution kernels module and proposing a density map module, satisfying the requirements for fast computing power and high detection accuracy in dense-flock chicken-counting tasks;
(2)
Feature fusion is adopted to obtain more feature information with insufficient lighting; a multi-scaling mechanism can be used for solving irregular sizes; and feature convolution kernels are employed to address serious occlusion;
(3)
A self-built dataset called Dense-Chicken was collected and constructed in a complex feeding environment, which contains 600 images of 99,916 chickens, in dot annotation files and box annotation files. It can be shared for researchers.
The rest of this article is organized as follows: Section 2 provides an overall description of the proposed models and methods. Section 3 describes the details of dataset preparation, which was used to train and evaluate the proposed network. The experimental results and a performance analysis of the proposed network are presented. Finally, Section 4 concludes the work by providing a summary and prospects of the proposed technique.

2. Materials and Methods

In this section, the overall framework of DFCCNet is first introduced, and then the feature convolution kernel generation module and counting module are explained. Finally, the loss function is demonstrated. The code is available at https://github.com/2226450890/DFCCNet (accessed on 22 November 2023).

2.1. Overall Framework

The overall framework of DFCCNet is shown in Figure 1. First, the original image and the cropped image are fed into the backbone module for feature extraction, to obtain image features and cropped image features, respectively. Then, in the correlation map generation module, the feature map of the cropped image is used as a feature convolution kernel to convolve the image features, to generate the correlation map. Finally, in the density map generation module, the resolution of the correlation map is restored to the original image size and the density map is generated. In particular, the parameters of the feature convolution kernel remain constant throughout training and testing, and it only needs to be generated once before the counting model is trained. The model makes direct calls to the feature convolution kernel during training and testing.Therefore, depending on whether the model parameters are changed and the order of execution, DFCCNet is divided into two parts: the feature convolution kernel generation module, and the counting module.

2.2. Feature Convolution Kernel Generation Module

In this study, the images contain a lot of useless background information that can adversely affect the counting of chickens. In order to count objects from different categories, Ref. [29] proposed an adaptive method. The input information for this method consists of the image and the box information of the counting target. Before generating the density map, image features located within these boxes are used as convolution kernels to perform convolution operations with the overall image features, to generate a correlation map focused on the counting objects. Inspired by this approach, feature convolution kernels are generated for enhancing the correlation between counting objects and feature maps, while filtering useless background information. Unlike the original method, in this study, since the counting objects are chickens of the same breed and age, and the morphological differences between different chickens in the images are small, generating different convolution kernels to improve the correlation of the feature maps during training was unnecessary. In this paper, the feature convolution kernel generation module is independent as the pre-module for counting, and the structure of this module is shown in Figure 2, which first adopts multiple images of a single chicken for feature extraction, and then the feature maps of different images are resized to the same size and fused to form a feature convolution kernel. During the whole model training and testing period, the feature convolution kernel only needs to be generated once. Although this can reduce the computational consumption, the quality of the feature convolution kernel needs to be considered. ResNet [30] has a wide range of applications in agriculture due to its excellent model performance [31,32,33]. ImageNet is a large-scale dataset, and by pretraining on this dataset, a model can be given good feature extraction capability [34]. Therefore, the pretrained ResNet in the feature convolutional kernel generation module is used for feature extraction, to ensure the quality of the convolutional kernel. Various versions of ResNet were used for comparative experiments, and finally ResNet50 was selected.
Although the experimental subjects were all chickens of the same breed and age, they presented different sizes in the images. Single-scale feature extraction can lead to the loss of multi-scale information. In order to solve this problem, multi-scale scaling methods are widely adopted [35,36]. In this paper, multi-scale scaling is applied to the feature convolution kernel, which ensures that the subsequently generated correlation maps can focus on the multi-scale chicken information, and thus accurately count chickens of different sizes in the images.

2.3. Counting Module

The framework of the counting module is presented in Figure 3. First, the image is input to ResNet50 to extract features, and feature convolution kernels are used to generate the correlation maps. Then, the correlation maps are adjusted to the same size and concatenated together for input into the density map prediction module. Finally, the predicted density map is generated. During the training phase, the loss between the predicted density map and the ground truth is calculated to update the model parameters. During the testing phase, the sum of the pixel values of the predicted density map is used as the counting result. In the counting module, two strategies are used to improve the counting performance.
(1) Transfer learning. In this research, data annotation became a time-consuming and cumbersome task, due to the high density and complex environment. Although a significant amount of time and effort was spent on data annotation, limited data still affected the performance of the model. Transfer learning provided us with new possibilities and can effectively apply knowledge learned in one task or field to other related tasks or fields, thereby greatly improving the learning ability and generalization of the model [37]. Through transfer learning, the trained models could be used as a base to quickly retrain new tasks, without the need to train the entire network from scratch. This method not only saved a lot of time and computational resources, but also improved the performance of the model under limited data conditions [38,39]. As with the feature convolution kernel generation module, the pretrained ResNet50 is used for feature extraction.
(2) Feature fusion. The feature maps output from different depths of the convolutional layers have different information: shallow feature maps are rich in spatial information, and deep feature maps are rich in semantic information. Fusion of feature maps from different layers can improve network representation and enhance information acquisition, effectively improving model performance [40,41]. In this paper, the feature maps and feature convolution kernels output from the third residual block and the fourth residual block of ResNet50 are utilized to generate the correlation maps of different layers, which are fused to obtain the final correlation map. This correlation map can be used to generate a high-quality density map for counting.
The details the of counting module are shown in Figure 4, which is divided into three modules; i.e., feature extraction, correlation map generation, and density map prediction. The feature convolution kernels and feature map obtained by the third block are used to obtain a series of correlation maps. The same operation is performed on the fourth block. In order to finally concatenate the correlation maps, the correlation maps output by the fourth block need to be upsampled. In the density map prediction module, five convolution kernels and three upsampling operations are used, where the first four convolution kernels are designed to restore the output size to the input size and the last convolution kernel is designed to predict the density map. Finally, the sum of the pixel values of the density map is used as the counting result.

2.4. Loss Function

In this paper, mean squared error (MSE) is used as the main body of the loss function, which means the average of the squared distance between the predicted value and the true value. MSE is defined in Equation (1).
M S E = 1 N i = 1 N ( Z i Z ^ i ) 2
where N represents the number of images, Z i represents the real density map of the i-th image, and Z ^ i represents the predicted density map. However, MSE is sensitive to outliers and susceptible to noise (including different sizes, occlusion, etc.) in the dense-counting field, which leads to inaccurate prediction in high-density areas. Thus, we added a constraint to the MSE that uses the random occlusion target for additional MSE to reduce noise interference. The random occlusion target is labeled on each image in the training set. The loss function is defined in Equation (2).
L o s s = M S E + λ M S E ( o c c l u s i o n _ t a r g e t )
The former MSE calculates the error between the predicted density map and the true density map. The latter part is the error between the predicted density map corresponding to the random occlusion target and its real density map. λ is a hyperparameter set as 1 × 10 9 .

3. Experiments and Discussion

In this section, a series of experiments are reported. (1) Each strategy was incrementally added to the counting network to determine its effectiveness, and the most appropriate parameter settings were identified. (2) The validation and testing sets were divided into different densities to verify the robustness of the model. (3) Comparison experiments between DFCCNet and some state-of-the-art methods were executed, to show the counting performance of DFCCNet.

3.1. Dataset

The experimental data were collected from a chicken farm in Guangdong Province in China in October 2021. The collected subjects were week-old chickens. The equipment for data acquisition is shown in Figure 5. The height of the chicken farm fence was 50 cm. During the data collecting process, our devices mainly included a mobile phone with the function of recording video and a bracket. The bracket was used to fix the mobile phone during the video process, and the length of the bracket could be adjusted in the range of 50–100 cm. Then, diverse data could be obtained from different angles in different breeding areas, which enhanced the generalization ability of the model. Finally, we obtained 20 videos with a resolution of 1280 × 720 pixels and a duration of 3 min, which had a frame rate of 30 frames per second.

3.1.1. Dataset Construction

A total of 1800 images were intercepted in videos by cutting 1 frame every 60 frames. All images were then examined and images with motion blur and a similar chicken distribution were removed, retaining a total of 600 images. The data were divided into a training set, validation set, and testing set, in a ratio of 7:1:2. And the images in the different sets were from different videos. The distribution of the dataset is shown in Table 1. In the data labeling stage, two image annotation tools, i.e., Labelme and LabelImg, were adopted to obtain JSON files containing point coordinate information and XML files containing box coordinate information. The representative dataset is available at https://stuscaueducn-my.sharepoint.com/:u:/g/personal/3170062_stu_scau_edu_cn/ETT-vDigmvZBu6EgSRtSn0sBnNHLojY_tDmiVaoZteVP3g?e=rGa2yO (accessed on 15 December 2022).
Some scenes in the dataset are shown in Figure 6. The challenges of counting a dense flock of chickens are presented as follows:
(1)
Complex background in the pictures: Chickens in different areas are exposed to different intensities of light, in which it is difficult to distinguish chickens from the background;
(2)
Occlusion situation: Since some feeders are placed in the breeding area, these devices can block the chickens to some extent. On the other hand, the high density of chickens and their habit of gathering also cause occlusion;
(3)
Pixel blur: Chickens like to move and sometimes move quickly, which can cause pixel blur;
(4)
Different scales: Since the object has the characteristics of being large in proximity and small in the distance, there will be many chickens of different sizes in the image. This requires a model to have strong feature extraction capabilities;
(5)
Different numbers: The density and number of chickens in different regions are different, which will challenge the generalization of a model.

3.1.2. Data Preprocessing

The spatial distribution information of the chickens needed to be generated according to the ground truth before training. Gaussian smoothing with an adaptive size window was used to generate a density map. First, a point annotation map was given, in which each point was located in the approximate center of the object. Second, the distances between each point and its nearest neighbor were calculated and averaged. Finally, this average distance was used as the size of the Gaussian window to generate the target density map.
In the model, feature convolution kernels fusing the features of single chickens were adopted, to generate correlation maps with better correlation with chickens.
Therefore, single chickens with different sizes and orientations were obtained by cropping images. The process of cropping image is shown in Figure 7.

3.2. Experimental Environmental and Metrics

The experimental environment of DFCCNet is listed in Table 2. The PyTorch framework for deep learning was adopted.
In this paper, mean absolute error (MAE), root mean squared error (RMSE), and mean normalized absolute error (NAE) were used as the main evaluation metrics in the counting experiments [42,43,44]. MAE is the average error between the predicted values and ground truth in counting, which was used to evaluate the accuracy of the model. RMSE is the dispersion of the error between the predicted values and ground truth, reflecting the stability of the model. NAE is the normalized MAE. They are defined in Equations (3), (4) and (5), respectively.
M A E = 1 N i = 1 N y i y ^ i
R M S E = 1 N i = 1 N ( y i y ^ i ) 2
N A E = 1 N i = 1 N y i y ^ i y i
where N is the number of images, y i represents the real number of chickens in the i-th image, and y ^ i represents the predicted number of chickens in the i-th image. The smaller MAE and RMSE, the higher the accuracy and stability of the model.
In addition, the parameter amount and frames per second (FPS) were used for evaluating the performance of the model. Parameter amount determines the size of the model. If a model is for use on a mobile platform, fewer parameters should be set to meet lightweight requirements. FPS was used to test the efficiency and speed of the model.

3.3. Model Validation with Different Strategies

In this section, transfer learning, feature convolution kernel, feature fusion, and improved loss function were sequentially added to the network for counting.

3.3.1. Transfer Learning

In this paper, a pretrained ResNet50 network was used to extract image features.ResNet is a deep learning network with many different versions, such as ResNet18, ResNet34, ResNet50, and ResNet101, which have different depths and sizes. The performance of a model improves as the depth of the network increases, but the model becomes more complex and has a larger size. In order to select the most appropriate version of ResNet and to verify the effectiveness of transfer learning, different models were used to extract image features. The experimental results are shown in Table 3, where the MAE of the models were all >20 without using transfer learning, while the MAE decreased dramatically when transfer learning was used. In addition, the MAE of ResNet50 and ResNet101 were significantly smaller than those of ResNet18 and ResNet34 when transfer learning was used. Although the error of ResNet101 was slightly smaller than ResNet50, its size was much larger. Therefore, considering the performance and size of the models, the pretrained ResNet50 was used to extract image features.

3.3.2. Feature Convolution Kernel

In order to improve the relevance of the image features to the chickens, and to reduce the interference of the complex background, the ordinary convolution kernel was replaced with a feature convolution kernel. During the feature convolution generation process, features from multiple single chicken images were fused and multi-scale scaling was performed. The following experiments explored these factors:
(1) Feature Fusion. Although the counting objects in this paper were chickens of the same breed and age, the feature map of one image did not provide an accurate representation of all chickens, due to their different postures. A feature convolution kernel that fuses enough features is beneficial for improving the relevance of the image features to chickens during the counting process. To find the most suitable number of features, a series of experiments were executed. The experimental results are shown in Table 4. The results show that the performance of the model improved as the number of features increased. However, the model performance no longer improved after a certain number was reached. This is because extracting a large number of image features is conducive to enriching the feature information of the chickens, but excessive information will lead to redundant feature information. Finally, the features of twenty single chicken images were extracted for the subsequent experiments.
(2) Features Multi-scaling. A multi-scaling strategy was adopted to obtain rich target features by adding convolution kernels with different scales. The length and width of most chickens in the image were between 50 and 150 pixels. In the stage of feature convolution kernel generation, the feature maps could represent the features of single chickens with 100 × 100 pixels. Therefore, the scales were set from 0.5 to 1.5, to allow the model to accurately count chickens with different scales. A series of experiments were implemented to find the most suitable strategy. The results are shown in Table 5. The magnification and reduction factors were set consistently. New scales were added to the combination of scales with the best result. This is conducive to counting chickens of different scales, through adding feature convolution kernels with multiple scales. But too much accumulation can produce redundant information, such that the performance cannot be improved. Finally, the best result was obtained by setting multiple scales of 0.8, 0.9, 1.0, 1.1, and 1.2.

3.3.3. Feature Fusion

In addition to the feature fusion used due to the introduction of feature convolution kernels, feature fusion was used to concatenate features from the different layers during the correlation map generation stage. Extracting features using the fifth residual block of ResNet50 resulted in a significant increase in the number of model parameters; therefore, in order to reduce the computational consumption, the outputs of the third and fourth residual blocks were used for the experiments. The experimental results are shown in Table 6. It is obvious that concatenating the correlation maps of the third block and the fourth block obtained the best result. The richer the feature information obtained, the higher the quality of the density map generated using the fusion strategy.

3.3.4. Improved Loss Function

The loss function in this article used additional MSE. Different values of the hyperparameter ( λ ) were set for the experiments. The experimental results are shown in Table 7, the best results were obtained when the value of λ was set to 1 × 10 9 . Compared to the original MSE, the addition of the extra MSE reduced the MAE by 0.29 for the validation set and 0.32 for the test set.
In the series of experiments above, various strategies were sequentially added to the network to perform the counting task, and the experiments verified their effectiveness. DFCCNet adopted all of these strategies and achieved accurate counting results, as shown in Figure 8.

3.4. Robustness Testing

In order to verify the robustness of DFCCNet, the validation set and testing set were divided into different groups with different density levels, according to the number of targets. The detailed distribution of the data is shown in Figure 9. The data with three density levels were used as validation sets, to verify the robustness of the model. CSRNet [45] was compared as a baseline. The results are presented in Table 8. DFCCNet achieved smaller counting errors than CSRNet at the same level, and the NAE was less than 0.1 for all levels. Therefore, DFCCNet could maintain stable performance at different densities.

3.5. Comparison with Other Methods

To further analyze the performance of DFCCNet, six state-of-the-art methods, including YOLOv5x [46], YOLOv7x [47], multi-column convolutional neural network (MCNN) [48], CSRNet, context aware network (CAN) [49], and segmentation guided attention networks (SGANet) [50] were used to compare with DFCCNet on the same training set, validation set, and testing set. The experimental results are shown in Table 9.
First, DFCCNet was compared with two advanced object detection algorithms, YOLOv5x and YOLOv7x. However, it was difficult for these two methods to detect accurately individual objects in dense groups, due to occlusion phenomena. It is obvious that they had large MAE values in the experiment. Compared to YOLOv5x, the MAE of DFCCNet was reduced by 2.66 for the validation set and 3.41 for the test set. Although DFCCNet’s FPS decreased by 20.55 compared to YOLOv5x, its number of parameters decreased by 69.31 MB. For the other density map regression methods, it can be seen that the parameters of MCNN were only 0.13 MB, and FPS reached 56.68. However, the three-column convolution kernels used by MCNN were too large in size. It could not extract features well for some small targets, resulting in a low accuracy. CSRNet, CANNet, and SGANet had a higher accuracy. But they are more complex, making them slow, and had a FPS of less than 10. Compared to CSRNet, DFCCNet showed a significant improvement in performance. The MAE was reduced by 1.12 on the validation set and 2.13 on the test set, and the FPS was improved by 7.72, while the parameters were increased by only 1.63 MB. Counting tasks in real scenery need to be implemented on mobile devices, which requires important considerations such as accuracy, size, and speed. In comparison, DFCCNet had the best accuracy, while maintaining acceptable parameters with 17.89 MB and FPS with 16.15, which can better meet the requirements of mobile devices. As shown in Figure 10, some renderings of DFCCNet, YOLOv5x, and CSRNet are presented. The first row shows some images were counted, and the last three rows list the counting results of YOLOv5x, CSRNet, and DFCCNet, respectively. It can be seen that YOLOv5x could obtain good counting result when the occlusion was not serious in scene1, but accurate counting was hindered in the case of serious occlusion in scene3 and scene4. In sparse scenes, density map regression, such as in DFCCNet and CSRNet, was not superior to object detection method, i.e., YOLOv5x. But they could achieve a higher counting accuracy than YOLOv5x in dense scenes. It is obvious that the counting results of DFCCNet were better than those of the classical model, i.e., CSRNet.
In addition, the values of MAE with the parameters and FPS respectively for the different models are shown in Figure 11. The methods with a density map have fewer parameters and smaller errors, while the object detection methods are faster. It can be seen that DCCNet is more suitable for the counting task for dense flocks of chickens.

3.6. Mobile Application

To enable DFCCNet to run on mobile devices and make the system easily accessible to end users, a mobile application (named Chickens-Counting) was developed. During the deployment of the model, the format of the model was transformed into Core ML (IOS), a format that can be used by mobile frameworks. In order to improve the performance of the model and reduce memory usage, quantization methods were used to compress the trained model. Sample screens of the app are shown in Figure 12. The app can ask the camera to take a picture or select a picture from the album, then the picture is passed to the API of the counting module. The API returns the time consumed by the counting process and the result of counting, and a visual image of the counting result is displayed in the screen. The app was tested on an iPhone 12 and the time required to count a single picture was 0.3 ± 0.05 s, which meets the counting requirements for mobile devices.

4. Conclusions

In modern farms, fast and accurate automatic counting is essential for chicken farming. In this paper, a dense-flock chicken-counting network termed DFCCNet was proposed based on density map regression, which implements accurate counting through generating high-quality density maps. In addition, a dataset of dense flocks of chickens in complex environments, called Dense-Chicken, was collected and constructed for training and evaluation. The experimental results showed that the density map regression method was superior to the object detection method in dense situations. Compared with some advanced density map regression methods, DFCCNet achieved good results in dense and complex scenes, with a small number of parameters and fast counting speed. As a result, DFCCNet can be embedded and deployed on mobile devices used for chicken farming. It can also provide scientific and theoretical guidance for fast and accurate counting of dense flocks of chickens.

Author Contributions

Conceptualization, J.L. and J.W.; methodology, J.L. and J.W.; software, J.L.; validation, J.L., J.W. and C.P.; formal analysis, Q.H.; investigation, J.L.; resources, J.W.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, J.W. and Q.H.; supervision, J.W.; project administration, J.W.; funding acquisition, Q.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Scientific Research Platforms and Projects of Guangdong Provincial Education Department (2021ZDZX1078), the Guangzhou Key Laboratory of Intelligent Agriculture (201902010081).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Acknowledgments

The authors thank all editors and reviewers for the suggestions and comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, Q.; Saatkamp, H.W.; Cortenbach, J.; Jin, W. Comparison of Chinese broiler production systems in economic performance and animal welfare. Animals 2020, 10, 491. [Google Scholar] [CrossRef] [PubMed]
  2. Damaziak, K.; Musielak, M.; Musielak, C.; Riedel, J.; Gozdowski, D.; Grzybek, W. Effect of different rearing system on eggs production, hatchability, and offspring quality in layer breeders. Poult. Sci. 2021, 100, 101101. [Google Scholar] [CrossRef] [PubMed]
  3. Li, X.M.; Zhang, M.H.; Liu, S.M.; Feng, J.H.; Ma, D.D.; Liu, Q.X.; Zhou, Y.; Wang, X.J.; Xing, S. Effects of stocking density on growth performance, growth regulatory factors, and endocrine hormones in broilers under appropriate environments. Poult. Sci. 2019, 98, 6611–6617. [Google Scholar] [CrossRef] [PubMed]
  4. Nasr, M.A.; Alkhedaide, A.Q.; Ramadan, A.A.; Abd-El Salam, E.H.; Hussein, M.A. Potential impact of stocking density on growth, carcass traits, indicators of biochemical and oxidative stress and meat quality of different broiler breeds. Poult. Sci. 2021, 100, 101442. [Google Scholar] [CrossRef] [PubMed]
  5. Thema, K.K.; Mnisi, C.M.; Mlambo, V. Stocking density-induced changes in growth performance, blood parameters, meat quality traits, and welfare of broiler chickens reared under semi-arid subtropical conditions. PLoS ONE 2022, 17, e0275811. [Google Scholar] [CrossRef] [PubMed]
  6. Estevez, I. Density allowances for broilers: Where to set the limits? Poult. Sci. 2007, 86, 1265–1272. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, B.; Min, Z.; Yuan, J.; Zhang, B.; Guo, Y. Effects of dietary tryptophan and stocking density on the performance, meat quality, and metabolic status of broilers. J. Anim. Sci. Biotechnol. 2014, 5, 44. [Google Scholar] [CrossRef]
  8. Aitor, A.; Elyse, M.; Tina, M.W.; Michele, T.G.; Elijah, G.K.; Stephanie, T. The effect of alternative feeding strategies for broiler breeder pullets: 1. Welfare and performance during rearing. Poult. Sci. 2019, 98, 3377–3390. [Google Scholar] [CrossRef]
  9. Scott, A.B.; Singh, M.; Groves, P.; Hernandez-Jover, M.; Barnes, B.; Glass, K.; Moloney, B.; Black, A.; Toribio, J.A. Biosecurity practices on Australian commercial layer and meat chicken farms: Performance and perceptions of farmers. PLoS ONE 2018, 13, e0195582. [Google Scholar] [CrossRef]
  10. Ojo, R.O.; Ajayi, A.O.; Owolabi, H.A.; Oyedele, L.O.; Akanbi, L.A. Internet of Things and Machine Learning techniques in poultry health and welfare management: A systematic literature review. Comput. Electron. Agric. 2022, 200, 107266. [Google Scholar] [CrossRef]
  11. Aydin, A. Development of an early detection system for lameness of broilers using computer vision. Comput. Electron. Agric. 2017, 136, 140–146. [Google Scholar] [CrossRef]
  12. Tu, X.; Du, S.; Tang, L.; Xin, H.; Wood, B. A real-time automated system for monitoring individual feed intake and body weight of group housed turkeys. Comput. Electron. Agric. 2011, 75, 313–320. [Google Scholar] [CrossRef]
  13. Zhuang, X.; Bi, M.; Guo, J.; Wu, S.; Zhang, T. Development of an early warning algorithm to detect sick broilers. Comput. Electron. Agric. 2018, 144, 102–113. [Google Scholar] [CrossRef]
  14. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  15. Ren, Y.; Huang, Y.; Wang, Y.; Zhang, S.; Qu, H.; Ma, J.; Wang, L.; Li, L. A High-Performance Day-Age Classification and Detection Model for Chick Based on Attention Encoder and Convolutional Neural Network. Animals 2022, 12, 2425. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, B.L.; Kuo, Y.F. Early Warning System for Open-beaked Ratio, Spatial dispersion, and Movement of Chicken Using CNNs. In Proceedings of the 2023 ASABE Annual International Meeting, Omaha, NE, USA, 9–12 July 2023; p. 1. [Google Scholar] [CrossRef]
  17. Yao, Y.; Yu, H.; Mu, J.; Li, J.; Pu, H. Estimation of the gender ratio of chickens based on computer vision: Dataset and exploration. Entropy 2020, 22, 719. [Google Scholar] [CrossRef]
  18. Volkmann, N.; Zelenka, C.; Devaraju, A.M.; Brünger, J.; Stracke, J.; Spindler, B.; Kemper, N.; Koch, R. Keypoint detection for injury identification during turkey husbandry using neural networks. Sensors 2022, 22, 5188. [Google Scholar] [CrossRef]
  19. Fang, C.; Zhang, T.; Zheng, H.; Huang, J.; Cuan, K. Pose estimation and behavior classification of broiler chickens based on deep neural networks. Comput. Electron. Agric. 2021, 180, 105863. [Google Scholar] [CrossRef]
  20. Rong, J.; Zhou, H.; Zhang, F.; Yuan, T.; Wang, P. Tomato cluster detection and counting using improved YOLOv5 based on RGB-D fusion. Comput. Electron. Agric. 2023, 207, 107741. [Google Scholar] [CrossRef]
  21. Tian, M.; Guo, H.; Chen, H.; Wang, Q.; Long, c.; Ma, Y. Automated pig counting using deep learning. Comput. Electron. Agric. 2019, 163, 104840. [Google Scholar] [CrossRef]
  22. Zhang, L.; Li, W.; Liu, C.; Zhou, X.; Duan, Q. Automatic fish counting method using image density grading and local regression. Comput. Electron. Agric. 2020, 179, 105844. [Google Scholar] [CrossRef]
  23. Abuaiadah, D.; Switzer, A.; Bosu, M.; Liu, Y. Automatic counting of chickens in confined area using the LCFCN algorithm. In Proceedings of the IEEE International Conference on Intelligent Systems and Computer Vision (ISCV), Fez, Morocco, 18–20 May 2022; pp. 1–7. [Google Scholar] [CrossRef]
  24. Cao, L.; Xiao, Z.; Liao, X.; Yao, Y.; Wu, K.; Mu, J.; Li, J.; Pu, H. Automated chicken counting in surveillance camera environments based on the point supervision algorithm: LC-DenseFCN. Agriculture 2021, 11, 493. [Google Scholar] [CrossRef]
  25. Li, X.; Zhao, Z.; Wu, J.; Huang, Y.; Wen, J.; Sun, S.; Xie, H.; Sun, J.; Gao, Y. Y-BGD: Broiler counting based on multi-object tracking. Comput. Electron. Agric. 2022, 202, 107347. [Google Scholar] [CrossRef]
  26. Zhu, X.; Wu, C.; Yang, Y.; Yao, Y.; Wu, Y. Automated Chicken Counting Using YOLO-v5x Algorithm. In Proceedings of the 2022 8th International Conference on Systems and Informatics (ICSAI), Kunming, China, 10–12 December 2022; pp. 1–6. [Google Scholar] [CrossRef]
  27. Zhao, Y.; Li, W.; Li, Y.; Qi, Y.; Li, Z.; Yue, J. LFCNet: A lightweight fish counting model based on density map regression. Comput. Electron. Agric. 2022, 203, 107496. [Google Scholar] [CrossRef]
  28. Yu, X.; Wang, Y.; An, D.; Wei, Y. Counting method for cultured fishes based on multi-modules and attention mechanism. Aquac. Eng. 2022, 96, 102215. [Google Scholar] [CrossRef]
  29. Ranjan, V.; Sharma, U.; Nguyen, T.; Hoai, M. Learning to count everything. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3394–3403. [Google Scholar] [CrossRef]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  31. Degu, M.Z.; Simegn, G.L. Smartphone based detection and classification of poultry diseases from chicken fecal images using deep learning techniques. Smart Agric. Technol. 2023, 4, 100221. [Google Scholar] [CrossRef]
  32. Sreenivasulu, K.; Khan, H.A.S.; Damini, K.; Akhila, M.; Bharathi, G. Detection of Chicken Disease Based on Day-Age Using Pre Trained Model of CNN. In Proceedings of the International Conference on Mining Intelligence and Knowledge Exploration, Kristiansand, Norway, 28–30 June 2023; pp. 13–24. [Google Scholar] [CrossRef]
  33. Wu, D.; Ying, Y.; Zhou, M.; Pan, J.; Cui, D. Improved ResNet-50 deep learning algorithm for identifying chicken gender. Comput. Electron. Agric. 2023, 205, 107622. [Google Scholar] [CrossRef]
  34. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  35. Lu, L.; Li, H.; Ding, Z.; Guo, Q. An improved target detection method based on multiscale features fusion. Microw. Opt. Technol. Lett. 2020, 62, 3051–3059. [Google Scholar] [CrossRef]
  36. Yang, J.; Wu, C.; Du, B.; Zhang, L. Enhanced multiscale feature fusion network for HSI classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10328–10347. [Google Scholar] [CrossRef]
  37. Ray, A.; Kolekar, M.H.; Balasubramanian, R.; Hafiane, A. Transfer learning enhanced vision-based human activity recognition: A decade-long analysis. Int. J. Inf. Manag. Data Insights 2023, 3, 100142. [Google Scholar] [CrossRef]
  38. Gulzar, Y. Fruit image classification model based on MobileNetV2 with deep transfer learning technique. Sustainability 2023, 15, 1906. [Google Scholar] [CrossRef]
  39. Mamat, N.; Othman, M.F.; Abdulghafor, R.; Alwan, A.A.; Gulzar, Y. Enhancing image annotation technique of fruit classification using a deep learning approach. Sustainability 2023, 15, 901. [Google Scholar] [CrossRef]
  40. Fan, X.; Luo, P.; Mu, Y.; Zhou, R.; Tjahjadi, T.; Ren, Y. Leaf image based plant disease identification using transfer learning and feature fusion. Comput. Electron. Agric. 2022, 196, 106892. [Google Scholar] [CrossRef]
  41. Ouyang, H.; Zeng, J.; Leng, L. Inception Convolution and Feature Fusion for Person Search. Sensors 2023, 23, 1984. [Google Scholar] [CrossRef]
  42. Ma, Z.; Wei, X.; Hong, X.; Gong, Y. Bayesian loss for crowd count estimation with point supervision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6142–6151. [Google Scholar] [CrossRef]
  43. Ranjan, V.; Le, H.; Hoai, M. Iterative crowd counting. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 270–285. [Google Scholar] [CrossRef]
  44. Wang, Q.; Gao, J.; Lin, W.; Li, X. NWPU-crowd: A large-scale benchmark for crowd counting and localization. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2141–2149. [Google Scholar] [CrossRef]
  45. Li, Y.; Zhang, X.; Chen, D. Csrnet: Dilated convolutional neural networks for understanding the highly congested scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1091–1100. [Google Scholar] [CrossRef]
  46. Glenn, J. YOLOv5. Git Code. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 1 October 2022).
  47. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, QC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  48. Zhang, Y.Y.; Zhou, D.S.; Chen, S.Q.; Gao, S.H.; Ma, Y. Single-Image Crowd Counting via Multi-Column Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Cvpr), Las Vegas, NV, USA, 27–30 June 2016; pp. 589–597. [Google Scholar] [CrossRef]
  49. Liu, W.; Salzmann, M.; Fua, P. Context-aware crowd counting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5099–5108. [Google Scholar]
  50. Wang, Q.; Breckon, T.P. Crowd counting via segmentation guided attention networks and curriculum loss. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15233–15243. [Google Scholar] [CrossRef]
Figure 1. The overall framework of DFCCNet.
Figure 1. The overall framework of DFCCNet.
Animals 13 03729 g001
Figure 2. The structure of the feature convolution kernel generation module. The whole process is divided into three stages: feature extraction, concatenating, and scaling. Where the green and orange blocks represent the output of the third and fourth residual blocks, respectively.
Figure 2. The structure of the feature convolution kernel generation module. The whole process is divided into three stages: feature extraction, concatenating, and scaling. Where the green and orange blocks represent the output of the third and fourth residual blocks, respectively.
Animals 13 03729 g002
Figure 3. The structure of counting module.
Figure 3. The structure of counting module.
Animals 13 03729 g003
Figure 4. Details of the counting module. In this figure, h and w refer to the height and width of the input image, and a and b refer to the characteristic convolution kernel sizes of the output of the third and fourth blocks. Since the feature convolution kernel generated earlier uses 20 cropped images, the dimension of the convolution kernel is 20 in the correlation map generation module. The scale_factor for all upsampling operations was set to 2.
Figure 4. Details of the counting module. In this figure, h and w refer to the height and width of the input image, and a and b refer to the characteristic convolution kernel sizes of the output of the third and fourth blocks. Since the feature convolution kernel generated earlier uses 20 cropped images, the dimension of the convolution kernel is 20 in the correlation map generation module. The scale_factor for all upsampling operations was set to 2.
Animals 13 03729 g004
Figure 5. Experimental data acquisition device.
Figure 5. Experimental data acquisition device.
Animals 13 03729 g005
Figure 6. Images and annotation files in different scenes.
Figure 6. Images and annotation files in different scenes.
Animals 13 03729 g006
Figure 7. Image cropping. The chickens cut out are intact, with different sizes and directions.
Figure 7. Image cropping. The chickens cut out are intact, with different sizes and directions.
Animals 13 03729 g007
Figure 8. Visualization of counting results. The first line shows the original image, the second line shows the ground truth, and the last line shows the visualization output of DFCCNet. For better observation, the gray scale map was combined with the density map.
Figure 8. Visualization of counting results. The first line shows the original image, the second line shows the ground truth, and the last line shows the visualization output of DFCCNet. For better observation, the gray scale map was combined with the density map.
Animals 13 03729 g008
Figure 9. Distribution of data for different density levels.
Figure 9. Distribution of data for different density levels.
Animals 13 03729 g009
Figure 10. Comparison of yolov5x, CSRNet, and DFCCNet. For the same image, the result of DFCCNet was closer to the ground truth.
Figure 10. Comparison of yolov5x, CSRNet, and DFCCNet. For the same image, the result of DFCCNet was closer to the ground truth.
Animals 13 03729 g010
Figure 11. Comparing the test MAE, parameters, and FPS of the different methods. (a) The relationship between MAE and parameters. (b) The relationship between MAE and FPS.
Figure 11. Comparing the test MAE, parameters, and FPS of the different methods. (a) The relationship between MAE and parameters. (b) The relationship between MAE and FPS.
Animals 13 03729 g011aAnimals 13 03729 g011b
Figure 12. Chickens counting mobile app user interface. (a) Picture acquisition screen, (b) counting result, (c) visualization result.
Figure 12. Chickens counting mobile app user interface. (a) Picture acquisition screen, (b) counting result, (c) visualization result.
Animals 13 03729 g012
Table 1. The amounts and distribution of the proposed dataset.
Table 1. The amounts and distribution of the proposed dataset.
DatasetsImagesChickensAverage
Complete dataset60099,916166
Training Set41670,777170
Validation Set619657158
Testing Set12319,482158
Table 2. Experimental environment.
Table 2. Experimental environment.
ConfigurationParameter
CPUIntel Core i7-12700
GPUNVIDIA RTX A5000
Operating systemWindows10
Deep learning frameworkPytorch 1.11.0
Programming languagePython 3.8
Table 3. Model performance using different ResNets. A ✔ indicates that transfer learning is used.
Table 3. Model performance using different ResNets. A ✔ indicates that transfer learning is used.
Transfer LearningMethodVal MAEVal RMSETest MAETest RMSESize (MB)
ResNet1825.0832.6832.4042.3044.7
ResNet3427.7734.1437.5646.0183.3
ResNet5024.0130.9530.2841.0197.8
ResNet10124.4334.5627.6139.41171
ResNet1817.2125.1522.6734.5944.7
ResNet3415.6719.8020.7926.5683.3
ResNet5012.1519.1817.7225.1597.8
ResNet10112.0517.7516.6325.87171
Table 4. Model performance that fuses different numbers of features.
Table 4. Model performance that fuses different numbers of features.
NumberVal MAEVal RMSETest MAETest RMSE
1012.5817.9014.6221.04
1510.3116.2014.8723.72
2010.7115.4514.4122.55
2510.5015.8114.6523.40
3011.0615.2515.3623.27
Table 5. Model performance with different scale feature convolution kernels.
Table 5. Model performance with different scale feature convolution kernels.
ScaleVal MAEVal RMSETest MAETest RMSE
1.010.7115.4514.4122.55
0.5, 1.0, 1.510.4714.9814.4220.70
0.6, 1.0, 1.410.0914.6114.0820.08
0.7, 1.0, 1.39.9414.9114.5422.91
0.8, 1.0, 1.29.9315.0113.7422.02
0.9, 1.0, 1.110.5215.5713.9920.70
0.5, 0.8, 1.0, 1.2, 1.510.1914.8213.3420.57
0.6, 0.8, 1.0, 1.2, 1.49.4215.4716.4524.78
0.7, 0.8, 1.0, 1.2, 1.39.9514.9816.0824.81
0.8, 0.9, 1.0, 1.1, 1.210.1114.6112.7419.40
0.5, 0.8, 0.9, 1.0, 1.1, 1.2, 1.59.9614.8514.2122.11
0.6, 0.8, 0.9, 1.0, 1.1, 1.2, 1.410.9215.9913.7220.82
0.7, 0.8, 0.9, 1.0, 1.1, 1.2, 1.310.2914.5613.9520.39
Table 6. Model performance when extracting features from different blocks.
Table 6. Model performance when extracting features from different blocks.
BlockVal MAEVal RMSETest MAETest RMSE
315.5822.4521.1330.71
410.1114.6112.7419.40
3, 49.3212.4312.3918.81
Table 7. Model performance for different values of λ .
Table 7. Model performance for different values of λ .
λ Val MAEVal RMSETest MAETest RMSE
09.3212.4312.3918.81
1 × 10 7 9.4312.8112.6418.46
1 × 10 8 9.1912.6112.2418.33
1 × 10 9 9.0312.0912.0718.20
1 × 10 10 9.2912.4612.4118.77
Table 8. The performance of DFCCNet and CSRNet at different density levels.
Table 8. The performance of DFCCNet and CSRNet at different density levels.
MethodDensity LevelMAENAE
CSRNet (baseline)level17.030.121
level210.100.073
level321.460.077
DFCCNetlevel14.260.074
level29.850.071
level319.170.068
Table 9. Experimental results of chicken counting using different methods.
Table 9. Experimental results of chicken counting using different methods.
MethodVal MAEVal RMSETest MAETest RMSEParameters (MB)FPS
yolov5x11.6916.8815.4824.9287.2036.70
yolov7x15.6124.1418.5429.0670.7841.39
MCNN12.8317.7117.0422.920.1356.68
CSRNet10.1513.7814.2021.1016.268.43
CANNet10.7814.3114.1718.7818.107.90
SGANet9.7713.0714.7722.6221.795.90
Ours9.0312.0912.0718.2017.8916.15
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, J.; Wang, J.; Peng, C.; Huang, Q. DFCCNet: A Dense Flock of Chickens Counting Network Based on Density Map Regression. Animals 2023, 13, 3729. https://doi.org/10.3390/ani13233729

AMA Style

Lv J, Wang J, Peng C, Huang Q. DFCCNet: A Dense Flock of Chickens Counting Network Based on Density Map Regression. Animals. 2023; 13(23):3729. https://doi.org/10.3390/ani13233729

Chicago/Turabian Style

Lv, Jinze, Jinfeng Wang, Chaoda Peng, and Qiong Huang. 2023. "DFCCNet: A Dense Flock of Chickens Counting Network Based on Density Map Regression" Animals 13, no. 23: 3729. https://doi.org/10.3390/ani13233729

APA Style

Lv, J., Wang, J., Peng, C., & Huang, Q. (2023). DFCCNet: A Dense Flock of Chickens Counting Network Based on Density Map Regression. Animals, 13(23), 3729. https://doi.org/10.3390/ani13233729

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop