Next Article in Journal
Eight-Year Study of Haemogregarina stepanowi Infection in Poached European Pond Turtles (Emys orbicularis) Held in Belgrade Zoo Quarantine
Next Article in Special Issue
Improving Known–Unknown Cattle’s Face Recognition for Smart Livestock Farm Management
Previous Article in Journal
Preductal Segmental Tubular Aortic Hypoplasia in Perinatally Died Stabyhoun Puppies
Previous Article in Special Issue
Multiview Monitoring of Individual Cattle Behavior Based on Action Recognition in Closed Barns Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessment of Preference Behavior of Layer Hens under Different Light Colors and Temperature Environments in Long-Time Footage Using a Computer Vision System

by
Vanessa Kodaira
1,
Allan Lincoln Rodrigues Siriani
2,
Henry Ponti Medeiros
3,
Daniella Jorge De Moura
1,* and
Danilo Florentino Pereira
1,4
1
Graduate Program in Agricultural Engineering, Faculty of Agricultural Engineering, Campinas State University, Campinas 13083-875, SP, Brazil
2
Graduate Program in Agribusiness and Development, School of Science and Engineering, São Paulo State University, Tupã 17602-496, SP, Brazil
3
Department of Agricultural and Biological Engineering, University of Florida, Gainesville, FL 32611, USA
4
Department of Management, Development and Technology, School of Science and Engineering, São Paulo State University, Tupã 17602-496, SP, Brazil
*
Author to whom correspondence should be addressed.
Animals 2023, 13(15), 2426; https://doi.org/10.3390/ani13152426
Submission received: 2 July 2023 / Revised: 22 July 2023 / Accepted: 26 July 2023 / Published: 27 July 2023
(This article belongs to the Special Issue Artificial Intelligence Tools to Optimize Livestock Production)

Abstract

:

Simple Summary

Preference behavior can be an important indicator of animal welfare. The effects of different wavelengths of light on laying hens are not completely known. We performed a preference test with few laying hens with three different light sources (green, red, and white) in three thermal environmental conditions (cold, comfort and heat stress). We recorded video footage during the entire period of the experiment and automatically estimated the permanence of the birds in each room of the Environmental Preference Chamber using YOLO v4. The results showed that birds prefer white and red light under thermal comfort conditions and do not show preference under heat stress conditions.

Abstract

As for all birds, the behavior of chickens is largely determined by environmental conditions. In many production systems, light intensity is low and red feather strains have low contrast with the background, making it impossible to use conventional image segmentation techniques. On the other hand, studies of chicken behavior, even when using video camera resources, depend on human vision to extract the information of interest; and in this case, reduced samples are observed, due to the high cost of time and energy. Our work combined the use of advanced object detection techniques using YOLO v4 architecture to locate chickens in low-quality videos, and we automatically extracted information on the location of birds in more than 648 h of footage. We develop an automated system that allows the chickens to transition among three environments with different illuminations equipped with video cameras to monitor the presence of birds in each compartment, and we automatically count the number of birds in each compartment and determine their preference. Our chicken detection algorithm shows a mean average precision of 99.9%, and a manual inspection of the results showed an accuracy of 98.8%. Behavioral analysis results based on bird unrest index and permanence time indicate that chickens tend to prefer white light and disfavor green light, except in the presence of heat stress when no clear preference can be observed. This study demonstrates the potential of using computer vision techniques with low-resolution, low-cost cameras to monitor chickens in low-light conditions.

1. Introduction

Illumination is a key environmental factor impacting egg production, as it alters the bird’s physiology, behavior, and well-being [1,2]. Light intensity and color are two important parameters that affect bird behavior [3]. The wavelength defines the colors perceived by the animals, impacting their behavior and consequently their response to the rearing environment, whether in stress relief or in improving the immune response [4]. The avian eye has photoreceptors in the retinal cone that are sensitive to a broader portion of the light spectrum than humans, allowing birds to perceive wavelengths between 350 nm and 700 nm [5]. This wavelength range encompasses ultraviolet rays [6], which increase productivity and encourage various desirable behaviors [7,8].
White light is commonly used for rearing laying hens [9,10] because it has a balanced distribution of the light spectrum in all wavelengths. Several studies indicate that short-wavelength light (green and blue) stimulates bird growth, improves animal welfare [11,12,13,14], and leads to higher quality eggs with increased weight, thickness, and shell strength [9,15]. Long-wavelength light (red and orange) increases reproductive hormone levels, favoring the development of sexual organs, influencing the age of sexual maturation of chickens, and improving productive performance [11,16,17,18]. On the other hand, these wavelengths also promote unwanted behaviors, such as aggressive pecking and cannibalism [9,12].
It is possible to monitor the behavior of chickens in different environmental conditions using video cameras. These behaviors can be analyzed either manually or automatically, using computer vision techniques and deep learning models for object detection [19]. Traditionally, the application of automated object detection techniques required substantial expertise in the design of visual feature extraction methods and feature similarity evaluation strategies. Deep learning methods changed that paradigm. It is now possible to apply computational models composed of multiple layers of abstraction that can directly represent the relationships between the inputs and outputs of the problem to be solved based on observed data. Such models are especially effective for problems for which there are no solutions based on traditional artificial intelligence and machine learning methods [20].
New tools based on machine vision are being developed with the aim of extracting information on the behavior of birds automatically in recorded videos. Sudebi et al. [21] developed and evaluated a model based on YOLO v5 for detecting pecking behavior in cage-free chickens and used 1924 images, 1300 for training, 324 for validation, and 300 for testing and obtained an accuracy of approximately 88%. Yang et al. [22] used YOLO v5 to classify six broiler behaviors in 9600 images for training and 2400 images for validation, obtaining global accuracy above 95%. Guo et al. [23] found accuracy above 96% in a brown chicken detection model in 720 images with different numbers of chickens previously selected. Pu et al. [24] used deep learning models to classify the behavior of birds in feeders, using color and depth images (i.e., a Kinect camera), obtaining a classification accuracy of 99%. Lin et al. [25] implemented a system to detect heat stress in birds, monitoring bird activity, temperature, and relative humidity. Wang et al. [26] designed a bird behavior classifier, achieving an accuracy of 95% in the detection of six classes of behaviors. An image-based system for detecting sick birds was developed by Zhuang and Zhang [27] and showed an accuracy of 99.7%. The aforementioned studies, as well as the vast majority of research using machine vision to detect chickens and analyze the behavior of these animals, consist of proofs of concept that test the effectiveness of the models in small image samples.
Commercial laying aviaries normally operate with low illumination levels. Red hues, more common in cage-free production systems, provide low contrast between the bird and the background of the image in these low-light conditions, making it difficult to recognize the birds using computer vision techniques. The bird detection model developed by Siriani et al. [28], based on the YOLO v4 architecture, effectively solved this problem, detecting birds in low-contrast images with an accuracy of 99.9%.
Our hypothesis is that laying hens have preferences for specific wavelength of illumination and that the ambient temperature can affect this preference behavior. Thus, we use the bird detection system proposed by Siriani et al. [28] to analyze footage of an experiment in a climatic chamber and evaluate the preference behavior and activity level of chickens for environments with different illumination sources (white, green, and red), under three environmental conditions (thermoneutral, cold, and hot).

2. Materials and Methods

2.1. Experimental Design

Six H&N Brown Nick layer hens, 35-weeks old at the beginning of the pilot experiment, were housed in an Environmental Preference Chamber (EPC) at the Faculty of Agricultural Engineering of the Campinas State University. The EPC comprises three mutually insulated compartments, measuring 1.6 m × 1.4 m × 3.0 m (L × W × H), enriched with a wood shavings bed (5 cm deep), a perch (60 cm length and 30 cm height), and a nest. The compartments are interconnected through automatic doors that allow the birds to move freely and choose their preferred environment, as shown in Figure 1. Before we initiated the experiments, the hens went through an adaptation period of seven days to gain familiarity with the EPC and learn to access its compartments.
After the training period, the birds were subjected to treatments with different light colors and ambient temperature, according to the experimental design shown in Table 1. Each compartment was illuminated using an LED light source of different color (white 3000 K, red, and green) and the same luminance of ~5 lux. The photoperiod was 16 h of light and 8 h of complete darkness. Bird preference behavior was monitored for three temperatures: 35 °C (Hot), 24 °C (Thermoneutral), and 17 °C (Cold). In each experiment, the temperature setting in the EPC was the same for all the compartments, so that the birds’ preference for the location was determined solely by the lighting conditions in that environment. The combined light and temperature treatments were rotated in the EPC compartments, with each treatment lasting two days, according to the protocol proposed by Ma et al. [29]. Each experiment listed in Table 1 was repeated three times, for a total of 54 days of experimentation.

2.2. Preference Behavior Monitoring

The behavior of the birds was recorded using low-cost video cameras with a resolution of 702 × 480 pixels and a capture rate of 30 fps (CCD sensor 1/4″, 2.6 mm lens, Intelbras® iM5, São José, SC, Brazil). We installed one ceiling-mounted camera at the center of each EPC compartment with the camera optical axis perpendicular to the floor of the compartment so that the field of view of the camera covers the entire floor area of its respective compartment.
Although videos of the three compartments were recorded uninterruptedly for the duration of the experiments, we select two periods of two hours per day to monitor the behavior of the birds. As first observed by Bizeray et al. [30] and more recently by Grebey et al. [31], monitoring bird behavior at different time periods helps to compensate for natural behavior variations throughout the day. Hence, to minimize the impact of observation time on the behavior of the birds, we monitored one interval in the morning (9:30–11:30) and one in the afternoon (15:00–17:00). This resulted in a total of 648 h of videos across the three EPC compartments. Given the high correlation among subsequent video frames, we down-sample the observed videos to 15 fps. Thus, our evaluation dataset comprises approximately 35 million video frames.

2.3. Bird Detection Workflow

The workflow for our automatic chicken detection system is shown in Figure 2. A hard disk drive was used to store the videos in MP4 format. We developed a Python script to scan these video files and arrange them in a queue for further processing. A separate script corresponding to the bird detection model identifies the boundaries of birds as bounding boxes and calculates their centroids. After detection, the number of centroids present in each frame is stored in a structured CSV file, which also associates the results with the treatment information and respective EPC compartment.

2.3.1. Computer Vision Model Design

Our bird detection model is based on the YOLO (You Only Look Once) single-stage framework for object classification and detection [32,33,34]. Unlike two-stage methods, in which image pixels are first coarsely segmented into regions of interest that are processed individually, the YOLO framework analyzes all the image pixels in a single step, which makes the model substantially faster than its two-stage counterparts. YOLO v4 [35] is a recent variant of the YOLO framework. It detects objects of different sizes in low-resolution images with an accuracy comparable to that of most state-of-the-art object detectors in a fraction of the time required by these methods. In this work, we apply the methodology for bird detection using YOLO v4 described by Siriani et al. [28], which follows the steps shown in Figure 3.
The number of frames used to train our model was defined using the sample size estimation strategy proposed by Agranonik and Hirakata [36], which is calculated as:
n = p 1 p Z 2 N ε 2 ( N 1 ) + Z 2   p ( 1 p )
where n is the sample size, p is the expected proportion of frames containing birds, Z is the normal distribution value for a given confidence level, N is the population size, and ε is the desired margin of error. In this work, we set the confidence level to 99%, which corresponds to Z = 2.576 and a margin of error ε = 4 % . Since no prior information is available regarding the distribution of the birds, we conservatively set p = 0.5 . Thus, for a population size of N = 35   M , which represents the total number of video frames acquired in our experiments, we obtain a sample size of n = 1041 frames.
We selected 1041 random frames containing birds to compose our training dataset. Our validation dataset consisted of 200 additional frames selected at random from the original videos. Figure 4a shows one of the original frames from our training set. For both datasets, we use the MakeSense tool (www.makesense.ai, accessed on 10 March 2021) to delimit a rectangular region representing the space occupied by birds in the video frames and to manually annotate the bounding boxes corresponding to each bird, as illustrated in Figure 4b.
Our object detection system is based on the Python implementation of the YOLO v4 model provided in the Darknet framework [37]. We modify the model architecture to predict only one object category and train it using the dataset described above. We configure our model to generate 16 subdivision per image, and train it using a batch size of 16, weight decay of 5 × 10−4, and a multi-step learning rate scheduling policy, with an initial learning rate of 1.3 × 10−3, which is divided by 10 after 8000 and 9000 iterations. The performance of this model applied to chicken detection within the EPC compartments is discussed in Section 3.1.

2.3.2. Validation of the Bird Detection Model

We assess the effectiveness of the YOLO v4 model based on its mean average precision (mAP) on the 200 frames of our validation set. The mAP is a widely used metric to assess the accuracy of detection models based on the YOLO network [38]. It is calculated as:
m A P = 1 n i = 1 n A P i
where A P i is the model precision for the i -th object category and n is the number of categories. Since our application contains only one category, n = 1 and m A P = A P 1 , which is simply the average percentage of detected birds on the validation set.
We further validate the model by manually inspecting an additional 299 randomly selected scenes. Each scene comprises one video frame from each EPC compartment, for a total of 897 synchronized frames. Since the video frames are acquired simultaneously, the same bird may occasionally be partially observed in two cameras, as illustrated in Figure 5. In these scenarios, we consider the bird to be present in the compartment where its head is located and disregard the corresponding detection from the frame where the remainder of its body is observed. The manually annotated bounding boxes are considered the gold standard for comparison with the results obtained by the YOLO v4 network for the same video frames. We compute the number of correct and incorrect detections and summarize them using a confusion matrix.

2.4. Preference Behavior Analysis

Upon validation, we used the YOLO v4 model to detect chickens in the 648 h of videos acquired in the EPC compartments. These detections were used to estimate the mobility of the birds according to two criteria: unrest index and permanence time.

2.4.1. Unrest Index Computation

Del Valle et al. [39] developed an index for the analysis of the agitation of chickens observed in subsequent video frames, which is calculated as:
U n r e s t   I n d e x ( i , i 1 ) = k . max d F i , F i 1 ,   d F i 1 , F i  
where k is a proportionality factor, F ( i ) and F ( i 1 ) are the sets of bounding boxes for chickens detected at frames observed at times i and i 1 , respectively, and d ( · , · ) is the Euclidean distance between the centroids of the bounding boxes in the two frames. The proportionality factor is calculated according to the camera’s field of view. In our system, the cameras have a focal length of 2.8 mm, resulting in a field of view of 90°, and k = 1 . One important characteristic of the unrest index is that it does not depend on the individual identities of the chickens, which our detection algorithm is unable to determine. Thus, we use the centroids of the bounding boxes generated by our detector at each video frame to calculate the unrest index and compare its value among different treatment groups using Tukey’s test at a 5% significance level.

2.4.2. Permanence Time Estimation

To further assess the mobility of the chickens, we also calculate the cumulative time spent by the birds in each compartment, according to:
T L = j = 1 m t j
where T L is the total time of permanence of the birds in light treatment L , tj is the time of permanence of the j -th bird, and m is the number of birds observed in the compartment during all the recordings of one experimental scenario described in Table 1.
Since our algorithm is unable to differentiate individual birds, dwell time was approximated by summing the total number of chickens identified in each frame, divided by the capture rate of 30 frames per second. The result is the total time in seconds that the chickens spend in the video. Given the high frame rate of the videos collected in our experiments and the high precision of the bird detection model, we expect this approach to accurately estimate permanence time. Differences between the total length of stay of the birds were tested among the different light and temperature treatments using one-way ANOVA and later compared using Tukey’s mean test at a significance level of 5%.

3. Results

3.1. Detection Model Validation

Figure 6a summarizes the performance of our detection model as a function of the training iterations on the 200 images composing our validation set. After approximately 5000 iterations, the loss value stabilizes at 0.5, and the mAP plateaus at ~99.9% after 3500 iterations. Although we cannot rule out model overfitting considering the similarity of the images in the training and validations sets, in this experimental environment this behavior is acceptable because all the frames used for the analysis were obtained under virtually identical conditions.
Figure 6b shows the confusion matrix for the visually inspected video frames used to manually validate the model. As expected, between zero and six birds were present in each scene, and the model predictions reflect this with an accuracy of 98.8%. As explained in Section 3.1, we accounted for chickens observed in two frames in the generation of our ground truth annotations. However, that does not preclude the automatic algorithm from detecting both parts of the chicken in both chambers. This would result in the system occasionally overestimating the number of birds in one of the chambers by one. Figure 6b indeed seems to support that hypothesis, as all of the mistakes made by the algorithm indeed represent the prediction of one additional bird in one of the frames. However, this problem only affected 1.23% of the frames under consideration (11 out of 897) and can thus be safely disregarded. Overall, these results indicate that it is possible to rely on the automatic detection of chickens for the behavioral analysis carried out in the next step.

3.2. Unrest Index Analysis

Figure 7a shows the unrest index values for each treatment calculated according to Equation (3). The figure also shows the result of Tukey’s statistical analysis. It is possible to observe that under heat stress conditions the birds do not express significant differences in movement between the light treatments, while in the thermoneutral and cold conditions, the birds move less under green lighting. There were no differences in the movement of the birds between the temperature treatments when the light treatment was fixed. ANOVA did not indicate differences in bird movement between different temperatures, but it did indicate differences for lighting, with birds moving less under green light.

3.3. Bird Permanence Time

Figure 7b shows the cumulative permanence time of the birds for each treatment computed using Equation (4). It can be observed that the preference of birds for light color is affected by the temperature of the environment. In the cold environment, the birds crowded more in the EPC compartment illuminated with white light, while in the hot environment they did not show a clear preference for any light color. The ANOVA analysis showed that there is an interaction between light and temperature treatments, which can be confirmed in Table 2, which presents the differences in the average length of stay of hens in the combined treatments.

4. Discussion

Illumination exerts an important influence on the behavior and welfare of birds. Our work analyzed the behavior of laying hens in more than 35 million frames or 648 h of recorded videos, positioning itself as one of the first studies to use computer vision to extract knowledge about the preference of hens to different light treatments. Our measured bird detection accuracy was 99%, and we used low-quality images obtained by low-cost cameras.
Several studies have evaluated the behavior of chickens using video cameras, but analysis always depends on human observation. Mendoza et al. [40] verified the effects of UV light on the behavior of laying hens at 1 min intervals for 6 min before and 6 min after application of light treatments. Geng et al. [41] evaluated the effects of light on circadian rhythmic behavior by positioning two observers on a walkway out of the hens’ field of vision so as not to affect the birds’ normal activities. Sun et al. [42] evaluated the nesting behavior of hens in enriched cages of different sizes, in videos recorded between 6:00 and 11:00 a.m. two days a week, for 20 weeks. These works are examples of the limitation of sample sizes, and yet, they are extremely labor intensive for the observers who recorded the behavior of the animals.
The use of low-resolution images can be a viable option for environments where capturing high-quality images is not possible, either due to equipment or budget limitations. This approach can be applied in different areas of animal monitoring. One of the main advantages of using low-resolution images is the reduction in image file size, which makes storing and sharing collected data easier. This is especially useful in environments with limited bandwidth for data transfer, such as remote or hard-to-reach areas. One more benefit is the low-cost of capture equipment and the low-cost of computers for processing these data.
However, it is important to note that using low-resolution images can affect the accuracy of object detection, especially in situations where the objects to be detected are small or have complex features. Therefore, it is necessary to carefully evaluate the limitations and advantages of using low-resolution images in each specific context. In summary, the bird detection model in low-resolution images is an effective and economical solution for animal monitoring in real environments. The application of advanced image processing and machine learning techniques can maximize detection accuracy, allowing the implementation of more animal monitoring projects in different parts of the world.
The experimental results indicate that chicken have a preference for white light, followed by red, and lastly green. According to Govardovskii and Zueva [43] and Hart et al. [44], birds have four cones in their vision system and are sensitive to the light spectrum range between 350 nm and 700 nm. White light provides a strong stimulus for the perception of the environment, as it involves a broad range of the birds’ visual perception system [5,45]. This is in agreement with the studies by Lewis and Morris [6] and Gunnarsson et al. [46], in which chickens preferred white light. However, Prayitno et al. [47] reports different results. In their study, the chickens preferred green light instead of white or red.
In all the scenarios under consideration, green light was the least preferred by the birds. These results are similar to those presented by Rierson [48], who also observed the preference for white light, followed by red, while green was the least preferred by broilers. Prayitno et al. [47] reports no difference in preference between red and white, but under red light the birds showed a more aggressive behavior in comparison to white, green, and blue light. Analyzing the unrest index (Figure 7a), we observed that the birds were indeed more agitated in environments with red light, except in the presence of heat stress, which is in agreement with the results obtained by Sultana et al. [49], Hesham et al. [50], and Khaliq et al. [51].
Thermal discomfort seems to prevail over lighting preference for chickens. In the cold environment, the birds express a clear preference for white light, concentrating in groups of three or four birds in the EPC compartment illuminated with white light. On the other hand, under heat stress, the birds were equally distributed among the light treatments in groups of two birds per compartment, on average. Red lighting has the potential to reduce stress in laying hens and reduce the occurrence of feather pecking and cannibalism [7,14,52], which may have contributed to the choice and permanence of birds. Under green light, birds tend to explore the environment less [7], which naturally occurs when birds are under heat stress. These are hypotheses that we raised about the reasons why we did not find a clear preference for a type of lighting under heat stress.
Ours study is limited to an EPC experiment with few animals, allowing assessment of the preference behavior of few chickens. However, we were able to accurately assess flock preference in long footage, using a computer vision system based on deep learning, allowing us to confirm our initial hypothesis. The results presented correspond only to these chickens and to the conditions described in the experiment and can subsidize other studies of preferred behavior both in the method of analysis approached and for the comparison of the results shown.

5. Conclusions

We present a high-performance system to automatically monitor the behavior of layer hens using computer vision techniques. We detect the birds using the YOLO v4 model with nearly perfect accuracy and use these automatic detections to determine the distribution of birds among compartments over 648 h of footage, surpassing any previous behavioral assessment experiment ever published. Our behavior preference analysis revealed that birds prefer white light, followed by red, under the experimental conditions described in this article. This study also indicates that chickens generally avoided green light and were more agitated under red illumination. However, temperature influenced the behavior of the birds more than the color of the light source. Under heat stress, the birds preferred to move away from one another, evenly splitting between the three light treatments. Under cold stress, on the other hand, the birds concentrated more on the chamber with white illumination.
This study indicates that light color may be an important environmental factor influencing the well-being and consequently the productivity of layer hens. Most importantly, this work highlights the importance of techniques to mitigate thermal stress, whose impact is significantly more prevalent than light color differences. Overall, our findings shed light on the interplay between temperature and illumination conditions as potential stressors for commercial birds. Hence, integrated management strategies for commercial layer chicken operations should be applied. We demonstrated the technical viability of using low-resolution standard security video cameras (which are consequently low-cost and for commercial use) in poor lighting conditions, in a high-performance computer vision system that allowed us to accurately study the preference behavior of chickens.

Author Contributions

The Conceptualization, Investigation and Data Curation was conducted by V.K. and D.J.D.M.; the Methodology was written by V.K., A.L.R.S., D.J.D.M. and D.F.P.; the Formal and Statistical Analysis was carried out by V.K., H.P.M., D.J.D.M. and D.F.P.; the Validation was made by A.L.R.S., H.P.M. and D.F.P.; the Original Draft, Proofreading and Editing of the manuscript was carried out by A.L.R.S., H.P.M., D.J.D.M. and D.F.P.; Visualization and Supervision was performed by D.J.D.M. and D.F.P. This study is part of the PhD thesis of V.K., under the supervision of D.J.D.M. All authors have read and agreed to the published version of the manuscript.

Funding

Funding was provided by National Council for Scientific and Technological Development—CNPq (Grant # 304085/2021-9 and # 308177/2021-7).

Institutional Review Board Statement

This study was conducted according to the guidelines of the Brazilian National Council for the Control of Animal Experimentation (CONCEA) and approved by the State University of Campinas Animal Ethics Committee (protocol number 5012-1/2018, CEUA-UNICAMP).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be available upon request to the corresponding author.

Acknowledgments

We thank the support of the technicians of the Laboratory of Environmental Comfort of the School of Agricultural Engineering of UNICAMP.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mohammed, H.; Grashorn, M.; Bessei, W. The effects of lighting conditions on the behaviour of laying hens. Arch. Geflügelkunde 2010, 74, 197–202. [Google Scholar]
  2. Jácome, I.; Rossi, L.; Borille, R. Influence of artificial lighting on the performance and egg quality of commercial layers: A review. Braz. J. Poult. Sci. 2014, 16, 337–344. [Google Scholar] [CrossRef]
  3. Elkomy, H.E.; Taha, A.E.; Basha, H.A.; Abo-Samaha, M.I.; Sharaf, M.M. Growth and reproduction performance of japanese quails (Coturnix coturnix japonica) under various environments of light colors. Slov. Vet. Res. 2019, 56, 119–127. [Google Scholar]
  4. Xie, D.; Wang, Z.; Dong, Y.; Cao, J.; Wang, J.; Chen, J.; Chen, Y. Effects of monochromatic light on immune response of broilers. Poult. Sci. 2008, 87, 1535–1539. [Google Scholar] [CrossRef]
  5. Prescott, N.; Wathes, C.M.; Jarvis, J. Light, vision and the welfare of poultry. Anim. Welf. 2003, 12, 269–288. [Google Scholar] [CrossRef]
  6. Lewis, P.; Morris, T. Poultry and coloured light. World’s Poult. Sci. J. 2000, 56, 189–207. [Google Scholar] [CrossRef]
  7. Huber-Eicher, B.; Suter, A.; Spring-Stähli, P. Effects of colored light-emitting diode illumination on behavior and performance of laying hens. Poult. Sci. 2013, 92, 869–873. [Google Scholar] [CrossRef] [PubMed]
  8. Baxter, M.; O’Connell, N.E. Testing ultra-wideband technology as a method of tracking fast-growing broilers under commercial conditions. Appl. Anim. Behav. Sci. 2020, 233, 105150. [Google Scholar] [CrossRef]
  9. Hassan, M.R.; Sultana, S.; Choe, H.S.; Ryu, K.S. Effect of combinations of monochromatic led light color on the performance and behavior of laying hens. J. Poult. Sci. 2014, 51, 321–326. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, K.; Xin, H.; Sekhon, J.; Wang, T. Effect of fluorescent vs. poultry-specific light-emitting diode lights on production performance and egg quality of w-36 laying hens. Poult. Sci. 2018, 97, 834–844. [Google Scholar] [CrossRef] [PubMed]
  11. Hassan, M.R.; Sultana, S.; Choe, H.S.; Ryu, K.S. Effect of monochromatic and combined light colour on performance, blood parameters, ovarian morphology and reproductive hormones in laying hens. Ital. J. Anim. Sci. 2013, 12, e56. [Google Scholar] [CrossRef] [Green Version]
  12. Rozenboim, I.; Piestun, Y.; Mobarkey, N.; Barak, M.; Hoyzman, A.; Halevy, O. Monochromatic light stimuli during embryogenesis enhance embryo development and posthatch growth. Poult. Sci. 2004, 83, 1413–1419. [Google Scholar] [CrossRef]
  13. Zhang, Z.; Cao, J.; Wang, Z.; Dong, Y.; Chen, Y. Effect of a combination of green and blue monochromatic light on broiler immune response. J. Photochem. Photobiol. B Biol. 2014, 138, 118–123. [Google Scholar] [CrossRef]
  14. Shi, H.; Li, B.; Tong, Q.; Zheng, W.; Zeng, D.; Feng, G. Effects of LED light color and intensity on feather pecking and fear responses of layer breeders in natural mating colony cages. Animals 2019, 9, 814. [Google Scholar] [CrossRef] [Green Version]
  15. Er, D.; Wang, Z.; Cao, J.; Chen, Y. Effect of monochromatic light on the egg quality of laying hens. J. Appl. Poult. Res. 2007, 16, 605–612. [Google Scholar] [CrossRef]
  16. Pyrzak, R.; Snapir, N.; Goodman, G.; Perek, M. The effect of light wavelength on the production and quality of eggs of the domestic hen. Theriogenology 1987, 28, 947–960. [Google Scholar] [CrossRef]
  17. Gongruttananun, N. Influence of red light on reproductive performance, eggshell ultrastructure, and eye morphology in thai-native hens. Poult. Sci. 2011, 90, 2855–2863. [Google Scholar] [CrossRef]
  18. Min, J.K.; Hossan, M.S.; Nazma, A.; Jae, C.N.; Han, T.B.; Hwan, K.K.; Dong, W.K.; Hyun, S.C.; Hee, C.C.; Ok, S.S. Effect of monochromatic light on sexual maturity, production performance and egg quality of laying hens. Avian Biol. Res. 2012, 5, 69–74. [Google Scholar] [CrossRef]
  19. Tulbure, A.A.; Tulbure, A.A.; Dulf, E.H. A review on modern defect detection models using dcnns–deep convolutional neural networks. J. Adv. Res. 2022, 35, 33–48. [Google Scholar] [CrossRef] [PubMed]
  20. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef] [Green Version]
  21. Sudebi, S.; Bist, R.; Yang, X.; Chai, L. Tracking pecking behaviors and damages of cage-free laying hens with machine vision technologies. Comput. Electron. Agric. 2023, 204, 107545. [Google Scholar] [CrossRef]
  22. Yang, X.; Bist, R.; Sudebi, S.; Wu, Z.; Liu, T.; Chai, L. An automatic classifier for monitoring applied behaviors of cage-free laying hens with deep learning. Eng. Appl. Artif. Intell. 2023, 123, 106377. [Google Scholar] [CrossRef]
  23. Guo, Y.; Regmi, P.; Ding, Y.; Bist, R.B.; Chai, L. Automatic Detection of Brown Hens in Cage Free Houses with Deep Learning Methods. Poult. Sci. 2023, 102, 102784. [Google Scholar] [CrossRef] [PubMed]
  24. Pu, H.; Lian, J.; Fan, M. Automatic recognition of flock behavior of chickens with convolutional neural network and kinect sensor. Int. J. Pattern Recognit. Artif. Intell. 2018, 32, 1850023. [Google Scholar] [CrossRef]
  25. Lin, C.Y.; Hsieh, K.W.; Tsai, Y.C.; Kuo, Y.F. Monitoring chicken heat stress using deep convolutional neural networks. In Proceedings of the 2018 ASABE Annual International Meeting, American Society of Agricultural and Biological Engineers, Detroit, MI, USA, 29 July–1 August 2018; p. 1. [Google Scholar]
  26. Wang, J.; Shen, M.; Liu, L.; Xu, Y.; Okinda, C. Recognition and classification of broiler droppings based on deep convolutional neural network. J. Sens. 2019, 2019, 3823515. [Google Scholar] [CrossRef] [Green Version]
  27. Zhuang, X.; Zhang, T. Detection of sick broilers by digital image processing and deep learning. Biosyst. Eng. 2019, 179, 106–116. [Google Scholar] [CrossRef]
  28. Siriani, A.L.R.; Kodaira, V.; Mehdizadeh, S.A.; de Alencar Nääs, I.; de Moura, D.J.; Pereira, D.F. Detection and tracking of chickens in low-light images using YOLO network and kalman filter. Neural Comput. Appl. 2022, 34, 21987–21997. [Google Scholar] [CrossRef]
  29. Ma, H.; Xin, H.; Zhao, Y.; Li, B.; Shepherd, T.; Alvarez, I. Assessment of lighting needs by w-36 laying hens via preference test. Animal 2016, 10, 671–680. [Google Scholar] [CrossRef]
  30. Bizeray, D.; Estevez, I.; Leterrier, C.; Faure, J. Effects of increasing environmental complexity on the physical activity of broiler chickens. Appl. Anim. Behav. Sci. 2002, 79, 27–41. [Google Scholar] [CrossRef]
  31. Grebey, T.C.; Ali, A.B.; Swanson, J.C.; Widowski, T.M.; Siegford, J.M. Dust bathing in laying hens: Strain, proximity to, and number of conspecifics matter. Poult. Sci. 2020, 99, 4103–4112. [Google Scholar] [CrossRef] [PubMed]
  32. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  33. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  34. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  35. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  36. Agranonik, M.; Hirakata, V.N. Cálculo de tamanho de amostra: Proporções. Clin. Biomed. Res. 2011, 31, 382–388. [Google Scholar]
  37. Redmon, J. Darknet: Open-Source Neural Networks. 2013–2016. Available online: http://pjreddie.com/darknet/ (accessed on 10 March 2021).
  38. Kumar, A.; Kalia, A.; Kalia, A. ETL-YOLO v4: A face mask detection algorithm in era of COVID-19 pandemic. Optik 2022, 259, 169051. [Google Scholar] [CrossRef]
  39. Del Valle, J.E.; Pereira, D.F.; Neto, M.M.; Gabriel Filho, L.R.A.; Salgado, D.D. Unrest index for estimating thermal comfort of poultry birds (Gallus gallus domesticus) using computer vision techniques. Biosyst. Eng. 2021, 206, 123–134. [Google Scholar] [CrossRef]
  40. Mendoza, A.V.; Weimer, S.; Williams, Z. Can UV light induce movement in cage-free laying hens? J. Appl. Poult. Res. 2023, 32, 100350. [Google Scholar] [CrossRef]
  41. Geng, A.L.; Zhang, Y.; Zhang, J.; Wang, H.H.; Chu, Q.; Yan, Z.X.; Liu, H.G. Effects of light regime on circadian rhythmic behavior and reproductive parameters in native laying hens. Poult. Sci. 2022, 101, 101808. [Google Scholar] [CrossRef]
  42. Sun, H.; Zhao, P.; Su, Y.; Zhao, Q.; Li, X.; Bao, J. Effects of different cage sizes and nest settings on laying hen nesting behavior and the proportion of dirty eggs in furnished cages. Appl. Anim. Behav. Sci. 2023, 262, 105917. [Google Scholar] [CrossRef]
  43. Govardovskii, V.; Zueva, L. Visual pigments of chicken and pigeon. Vis. Res. 1977, 17, 537–543. [Google Scholar] [CrossRef] [PubMed]
  44. Hart, N.; Partridge, J.; Cuthill, I. Visual pigments, cone oil droplets, ocular media and predicted spectral sensitivity in the domestic turkey (Meleagris gallopavo). Vis. Res. 1999, 39, 3321–3328. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Prescott, N.; Wathes, C. Spectral sensitivity of the domestic fowl (gallus g. domesticus). Br. Poult. Sci. 1999, 40, 332–339. [Google Scholar] [CrossRef] [PubMed]
  46. Gunnarsson, S.; Heikkilä, M.; Hultgren, J.; Valros, A. A note on light preference in layer pullets reared in incandescent or natural light. Appl. Anim. Behav. Sci. 2008, 112, 395–399. [Google Scholar] [CrossRef]
  47. Prayitno, D.; Phillips, C.; Stokes, D. The effects of color and intensity of light on behavior and leg disorders in broiler chickens. Poult. Sci. 1997, 76, 1674–1681. [Google Scholar] [CrossRef] [PubMed]
  48. Rierson, R.D. Broiler Preference for Light Color and Feed Form, and the Effect of Light on Growth and Performance of Broiler Chicks. Master’s Thesis, College of Agriculture, Kansas State University, Manhattan, KS, USA, 2011; p. 62. [Google Scholar]
  49. Sultana, S.; Hassan, M.R.; Choe, H.S.; Ryu, K.S. The effect of monochromatic and mixed led light colour on the behaviour and fear responses of broiler chicken. Avian Biol. Res. 2013, 6, 207–214. [Google Scholar] [CrossRef]
  50. Hesham, M.; El Shereen, A.; Enas, S. Impact of different light colors in behavior, welfare parameters and growth performance of fayoumi broiler chickens strain. J. Hell. Vet. Med. Soc. 2018, 69, 951–958. [Google Scholar] [CrossRef] [Green Version]
  51. Khaliq, T.; Khan, A.; Dar, P.A.; Nazir, T.; Afzal, I.; Bilal, M.; Tarique, P. Behavioral study of broilers reared under different colours of light in the evening hours. J. Entomol. Zool. Stud. 2018, 6, 1624–1627. [Google Scholar]
  52. Archer, G.S. How does red light affect layer production, fear, and stress? Poult. Sci. 2019, 98, 3–8. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Environmental Preference Chamber (EPC) layout. The EPC comprises three identical compartments: (AC).
Figure 1. Environmental Preference Chamber (EPC) layout. The EPC comprises three identical compartments: (AC).
Animals 13 02426 g001
Figure 2. Model execution diagram and corresponding output files.
Figure 2. Model execution diagram and corresponding output files.
Animals 13 02426 g002
Figure 3. Diagram of the bird detection methodology used to build the bird behavioral analysis model.
Figure 3. Diagram of the bird detection methodology used to build the bird behavioral analysis model.
Animals 13 02426 g003
Figure 4. (a) Sample image illustrating the training set. (b) Manual annotation of chickens presents in the images that compose the training dataset.
Figure 4. (a) Sample image illustrating the training set. (b) Manual annotation of chickens presents in the images that compose the training dataset.
Animals 13 02426 g004
Figure 5. Sample frames simultaneously acquired from the three EPC compartments to manually validate the YOLO v4 detection model. The left and center frames illustrate the scenario where one bird can be partially observed in two compartments simultaneously.
Figure 5. Sample frames simultaneously acquired from the three EPC compartments to manually validate the YOLO v4 detection model. The left and center frames illustrate the scenario where one bird can be partially observed in two compartments simultaneously.
Animals 13 02426 g005
Figure 6. Manual performance validation of the YOLO v4 detection model. (a) Validation loss and mAP as a function of training iterations. (b) Confusion matrix for the 299 manually validated simultaneous scenes.
Figure 6. Manual performance validation of the YOLO v4 detection model. (a) Validation loss and mAP as a function of training iterations. (b) Confusion matrix for the 299 manually validated simultaneous scenes.
Animals 13 02426 g006
Figure 7. (a) Unrest index values for each combined temperature and illumination condition and Tukey statistical test result—the letters indicate significant difference (p < 0.05) in Tukey’s test between illumination treatments for each temperature condition. (b) Cumulative permanence time in hours for each combined light and temperature treatment.
Figure 7. (a) Unrest index values for each combined temperature and illumination condition and Tukey statistical test result—the letters indicate significant difference (p < 0.05) in Tukey’s test between illumination treatments for each temperature condition. (b) Cumulative permanence time in hours for each combined light and temperature treatment.
Animals 13 02426 g007
Table 1. Combinations of illumination color and thermal conditions for the three EPC compartments used in our experiments. W, G, and R stand for white, green, and red illumination. Each experiment lasted two days and was repeated three times (n = 3).
Table 1. Combinations of illumination color and thermal conditions for the three EPC compartments used in our experiments. W, G, and R stand for white, green, and red illumination. Each experiment lasted two days and was repeated three times (n = 3).
Exp.ThermalCompartment
#Env.ABC
1 WGR
2Hot (35 °C)RWG
3 GRW
4 WGR
5Neutral (24 °C)RWG
6 GRW
7 WGR
8Cold (17 °C)RWG
9 GRW
Table 2. Average time (minutes) spent per chicken in each combined temperature and light treatment.
Table 2. Average time (minutes) spent per chicken in each combined temperature and light treatment.
LightTemperature
ColdComfortHeat
Green291.3 ± 65.7A,B,c453.9 ± 93.0A,b824.1 ± 85.7A
Red670.0 ± 207.0b936.0 ± 140.0a728.0 ± 108.0
White1354.0 ± 165.0a1131.0 ± 127.0a974.0 ± 134.0
Lowercase letters indicate differences between light treatments and uppercase letters indicate differences between temperature treatments in the Tukey test (p < 0.05).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kodaira, V.; Siriani, A.L.R.; Medeiros, H.P.; De Moura, D.J.; Pereira, D.F. Assessment of Preference Behavior of Layer Hens under Different Light Colors and Temperature Environments in Long-Time Footage Using a Computer Vision System. Animals 2023, 13, 2426. https://doi.org/10.3390/ani13152426

AMA Style

Kodaira V, Siriani ALR, Medeiros HP, De Moura DJ, Pereira DF. Assessment of Preference Behavior of Layer Hens under Different Light Colors and Temperature Environments in Long-Time Footage Using a Computer Vision System. Animals. 2023; 13(15):2426. https://doi.org/10.3390/ani13152426

Chicago/Turabian Style

Kodaira, Vanessa, Allan Lincoln Rodrigues Siriani, Henry Ponti Medeiros, Daniella Jorge De Moura, and Danilo Florentino Pereira. 2023. "Assessment of Preference Behavior of Layer Hens under Different Light Colors and Temperature Environments in Long-Time Footage Using a Computer Vision System" Animals 13, no. 15: 2426. https://doi.org/10.3390/ani13152426

APA Style

Kodaira, V., Siriani, A. L. R., Medeiros, H. P., De Moura, D. J., & Pereira, D. F. (2023). Assessment of Preference Behavior of Layer Hens under Different Light Colors and Temperature Environments in Long-Time Footage Using a Computer Vision System. Animals, 13(15), 2426. https://doi.org/10.3390/ani13152426

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop