Next Article in Journal
A Physical-Layer Watermarking Scheme Based on 5G NR
Next Article in Special Issue
Usability Analysis of a Virtual Reality Exposure Therapy Serious Game for Blood Phobia Treatment: Phobos
Previous Article in Journal
Applications of Multi-Agent Systems in Unmanned Surface Vessels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Eye Aspect Ratio for Real-Time Drowsiness Detection to Improve Driver Safety

1
Department of Information Technology, Satya Wacana Christian University, 52-60 Diponegoro Rd, Salatiga City 50711, Indonesia
2
Department of Information Management, Chaoyang University of Technology, Taichung 41349, Taiwan
3
Department of Computer Science and Information Engineering, Chaoyang University of Technology, Taichung 41349, Taiwan
4
Department of Mathematics and Computer Science, University of Münster, D-48149 Münster, Germany
5
School of Creative Technologies, University of Portsmouth, Portsmouth PO1 2UP, UK
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(19), 3183; https://doi.org/10.3390/electronics11193183
Submission received: 8 September 2022 / Revised: 26 September 2022 / Accepted: 30 September 2022 / Published: 4 October 2022
(This article belongs to the Special Issue Recent Advances in Metaverse and Computer Vision)

Abstract

:
Drowsiness is a major risk factor for road safety, contributing to serious injury, death, and economic loss on the road. Driving performance decreases because of increased drowsiness. In several different applications, such as facial movement analysis and driver safety, blink detection is an essential requirement that is used. The extremely rapid blink rate, on the other hand, makes automatic blink detection an extremely challenging task. This research paper presents a technique for identifying eye blinks in a video series recorded by a car dashboard camera in real time. The suggested technique determines the facial landmark positions for each video frame and then extracts the vertical distance between the eyelids from the facial landmark positions. The algorithm that has been proposed estimates the facial landmark positions, extracts a single scalar quantity by making use of Eye Aspect Ratio (EAR), and identifies the eye closeness in each frame. In the end, blinks are recognized by employing the modified EAR threshold value in conjunction with a pattern of EAR values in a relatively short period of time. Experimental evidence indicates that the greater the EAR threshold, the worse the AUC’s accuracy and performance. Further, 0.18 was determined to be the optimum EAR threshold in our research.

1. Introduction

The technology for detecting eye blinks is important and has been used in a variety of areas, including drowsiness detection [1,2]. driver safety [3,4,5], computer vision [6,7], and anti-spoofing protection in face recognition systems [8,9]. Drowsiness is one of the most significant variables that jeopardize road safety and contributes to serious injuries, deaths, and economic losses on the road. Due to increased drowsiness, driving performance decreases. Accidents involving serious injury or death occur because of inattention caused by an involuntary shift from waking to sleep. Individuals with normal vision exhibit spontaneous eye blinking at a specific frequency. Improvements in information and signal processing technologies have a positive impact on autonomous driving (AD), making driving safer while reducing the challenges faced by human drivers as a result of newly developed artificial intelligence (AI) techniques [10,11]. Over the decades, the development of autonomous vehicles has resulted in life-changing breakthroughs. In reality, there will be noticeable societal effects from its adoption in the areas of accessibility, safety, security, and ecology [12].
Eye blinking is influenced by various factors, including eyelid conditions, eye conditions, the presence of disease, the presence of contact lenses, psychological conditions, the surrounding environment, drugs, and other stimuli. The number of blinks per minute ranges between 6 and 30 [13]. According to [14], the human blink rate varies depending on the circumstances. During normal activity, a person’s average blink rate is 17 blinks per minute. There is variation in blink rate, with the highest being 26 blinks per minute and the lowest being 4–5 blinks per minute. In this way, it becomes clear that a person’s blink rate varies based on the environment he is in and his concentration on the task at hand.
While driving, one must maintain the maximum level of concentration on the road, which results in a reduction in the blink rate. When driving, the average blinking speed is about 8–10 blinks per minute. A person’s blink rate is also affected by their age group, gender, and the amount of time they spend blinking. There are real-time facial landmark detectors that can capture most of the distinctive aspects of human facial photographs. These features include the corner-of-the-eye angles and the eyelids [15,16].
The size of an individual’s eye does not correspond to the size of his or her body. Take, for instance, two people who are physically identical save for the sizes of their eyes: one can have enormous eyes, and the other, small eyes. The eye height when closed is the same for all people, no matter whether the size of the eye is big or small. This problem will inevitably influence the experimental findings. In response, we present a simple but very successful method for identifying the blink of an eye using a facial landmark detector with Eye Aspect Ratio (EAR). One easy way is to use the Eye Aspect Ratio (EAR) algorithm. Further, the EAR requires only basic calculations based on the ratio of the distances between the eye’s facial landmarks. This technique for detecting the blink of an eye is fast, efficient, and easy to practice. Dewi et al. [17] built their own eye dataset, which had several challenges, including small eyes, wearing glasses, and driving a car. This dataset is adapted to the characteristics of small eyes. We used this dataset in our experiment.
The model proposed in [18] is one in which the eye is modeled in conjunction with its surrounding context. The first step is a visual context pattern-based eye model, and the second step is semi-supervised boosting for high-precision detection. The approach consists of these two steps. Lee et al. [19] tried to provide an estimation of the condition of an eye, which may include whether or not an eye is open or closed.
When analyzing patterns in a visual setting, it is important to maintain as much consistency as possible in what is visible. Another approach is presented in [20], where an eye filter is utilized for finding all the eye candidate’s points. The non-negative matrix factorization (NMF) is reduced to its smallest possible value because of this reconstruction of the mistake.
Pioneering work was performed by Fan Li et al. [21], who investigated the effects of data quality on eye-tracking-based fatigue indicators and proposed a hierarchical-based interpolation approach to extract eye-tracking-based fatigue indicators from low-quality eye-tracking data. This work was considered groundbreaking because it investigated the effects of data quality on eye-tracking-based fatigue indicators. Gracia et al. [22] conducted an experiment using eye closure for separate frames, which was then subsequently used in a sequence for the detection of blinks. We built our algorithm upon the successful methods of Eye Aspect Ratio [23,24] and facial landmarks [25,26].
Learning and normalized cross-correlation are used to create templates with open and/or closed eyes. Eye blinks can also be identified by measuring ocular parameters, such as by fitting ellipses to eye pupils using a variation of the algebraic distance algorithm for conic approximation. Eye blinks can be detected by monitoring ocular parameters [27,28].
The following is a list of the most significant contributions that this article has made: (1) We propose a method to automatically classify blink types by determining the different EAR thresholds (0.18, 0.2, 0.225, 0.25). (2) Adjustments were made to the Eye Aspect Ratio to improve the detection of eye blinking based on facial landmarks. (3) We conducted an in-depth analysis of the experiment’s findings using TalkingFace, Eyeblink8, and Eye Blink datasets. (4) Our experimental results show that using 0.18 as the EAR threshold provides the best performance.
The following is the structure of this research work. The Materials and Methods section covers related work and the methodology we applied in this research. Results and Discussion describes our experimental setting and results. In the final section, conclusions are drawn and suggestions for future research are made.

2. Materials and Methods

Drowsiness is characterized by yawning, heavy eyelids, daydreaming, eye rubbing, an inability to concentrate, and lack of attention. The percentage of eyelid closure over the pupil over time (PERCLOS) [29] is one of the most widely used parameters in computer-vision-based drowsiness detection in driving scenarios [30]. In reference [31], a convolutional neural network (CNN) was used to develop a tiredness detection system. The program trained the first network to distinguish between human and non-human eyes, then used the second network to locate the eye feature points and calculate the eye-opening degree. Efficient algorithms for detecting drowsiness are presented in this article. In addition, facial landmarks can be retrieved using the Dlib toolkit. By identifying the different EAR thresholds, we present a method for automatically classifying blink types.

2.1. Facial Landmarks for Eye Blink Detection

Deep-learning-based facial landmark detection systems have made impressive strides in recent years [32,33]. A cascaded convolutional network model, as proposed by Sun et al. [34], consists of a total of 23 CNN models. This model has very high computational complexity during training and testing.
To detect and track important facial features, identification of facial markers must be performed on the subject. As a result of head movements and facial expressions, facial tracking is stronger for rigid facial deformations. Facial landmark identification is a computer vision job in which we try to identify and track key points on the human face using computer vision algorithms [35]. Multi-Block Color-Binarized Statistical Image Features (MB-C-BSIF) is discussed in [36]; this is a novel approach to single-sample facial recognition (SSFR) that makes use of local, regional, global, and textured-color properties [37,38].
Drowsiness can be measured on a computational eyeglass that can continually sense fine-grained measures such as blink duration and percentage of eye closure (PERCLOS) at high frame rates of about 100 fps. This work can be used for a variety of problems. Facial landmarks are used to localize and represent salient regions of the face, including eyes, eyebrows, nose, mouth, and jawline.
Blinking occurs repeatedly and involuntarily throughout the day to maintain a certain thickness of the tear film on the cornea [39]. The act of blinking is a reflex that involves the fast closure and opening of the eyelids in rapid succession. Blinking is also known as blepharospasm. The act of blinking is performed subconsciously. The synchronization of several different muscles is required for the act of blinking one’s eyes.
While keeping the cornea healthy is an important function of blinking, there are other benefits as well [40], and this is supported by the fact that adults and infants blink their eyes at different rates. A person’s blink frequency changes in response to their level of activity. The number of blinks increases when a person reads a certain phrase aloud or performs a visually given information exercise, whereas the number of blinks decreases when a person focuses on visual information or reads words quietly [41].
In our investigation, we made use of the 68 facial landmarks from Dlib [42]. Estimating the 68 (x,y)-coordinates corresponding to the facial structure on the face was carried out with the help of a pre-trained facial landmark detector found in the Dlib library. Figure 1 displays that the jaw points range from 0 to 16, the right brow points range from 17 to 21, and the left brow points range from 22 to 26. The nose points range from 27 to 35, the right eye points range from 36 to 41, and the left eye points range from 42 to 47. The mouth points range from 48 to 60, and the lip points range from 61 to 67. Dlib is a library that helps implement computer vision and machine learning techniques. The C++ programming language serves as the foundation for this library.
The process of locating facial landmark points with the use of Dlib’s 68-landmark model consists of two stages, all of which are described in the following order: (1) The first way to locate a human face, face detection, is done by returning a value in the form of x, y, w, and h coordinates, which together form a rectangle. (2) A facial landmark: Once we have determined the location of a face within an image, we must then place points within the rectangle. This annotation is included in the 68-point iBUG 300-W dataset, which serves as the basis for the training of the Dlib facial landmark predictor. The Dlib framework can be utilized to train form predictors on input training data, regardless of the dataset that is selected to be trained on.

2.2. Eye Aspect Ratio (EAR)

The Eye Aspect Ratio, or EAR, is a scalar value that responds, particularly for opening and closing the eyes [43]. During the flashing process, we can see that the EAR value grows rapidly or decreases significantly. Interesting findings in terms of robustness were obtained when EAR was used to detect blinks in [44]. Studies in the past have employed a predetermined EAR threshold to establish when subjects blink (EAR threshold at 0.2). This approach is impractical when dealing with a wide range of individuals, due to inter-subject variation in appearance and features such as natural eye openness, as in this study. Our works used an EAR threshold value to detect a rapid increase or decrease in the EAR value caused by blinking, based on the findings of previous studies.
We used the varying EAR threshold to automatically categorize the various sorts of blinks (0.18, 0.2, 0.225, 0.25). After that, we analyzed the experimental result and determined the best EAR threshold for our dataset. Each frame of the video stream is used to estimate the EAR. Furthermore, when the user shuts their eyes, the EAR drops and then returns to a regular level when the eyes are opened again. This technique is used to determine both blinks and eye opening. As the EAR formula is insensitive to both the direction of the face and the distance between it and the observer, it can be used to detect faces from a considerable distance. The EAR value can be calculated by entering six coordinates surrounding the eyes, as shown in Figure 2, and Equations (1) and (2) [30,45].
E A R = P 2 P 6 + P 3 P 5 2 P 1 P 4
A V G   E A R = 1 2 ( E A R L e f t + E A R R i g h t )
The EAR equations are described by Equation (1), where P1 through P6 stand in for the locations of the 2D landmarks on the retina. P2, P3, P5, and P6 were utilized to measure the height of the eye, whereas P1 and P4 were utilized to measure the breadth of the eye. This is depicted in Figure 2. When the eyes are closed, the EAR value quickly drops to virtually zero, in contrast to when the eyes are open, in which case the EAR value remains constant. This behavior is seen in Figure 2b.

2.3. Research Workflow

Our system architecture is divided into two steps, namely, data preprocessing and eye blink detection, as described in Figure 3. In the data preprocessing step, the video labeling procedure using Eyeblink Annotator 3.0 by Andrej Fogelton [46] is shown in Figure 4. OpenCV version 2.4.6 is utilized by the annotation tool. Both video 1 and video 3 were recorded at a frame rate of 27.97 frames per second. Video 2 was captured with 24 fps. Video 1 has a length of 1829 frames, totaling 29.6 MB. Video 2 has a length of 1302 frames and a file size of 12.4 MB. Next, video 3 has 2195 frames and a file size of 38.6 MB. The Talking Face and Eyeblink8 datasets contain 5000 frames and 10,712 frames, respectively. Video information is explained in Table 1.
People who wear glasses and have relatively tiny eyes are represented in our dataset in a unique way. The people who operate automobiles make up the environment. This dataset may be utilized for additional research endeavors. Based on what we know, it is difficult to locate a dataset of persons who have tiny eyes, wear spectacles, and drive cars. We have the footage from the dashboard camera installed in a vehicle in the Wufeng District of Taichung, Taiwan. We have verified that informed consent was received from each individual who participated in the video dataset collection. Our data collection includes five films and one individual performing a driving scenario.
The annotations start with line “#start,” and rows consist of the following information: frame ID: blink ID: NF: LE_FC: LE_NV: RE_FC: RE_NV: F_X: F_Y: F_W: F_H: LE_LX: LE_LY: LE_RX: LE_RY: RE_LX: RE_LY: RE_RX: RE_RY. An example of a frame with a blink is: 118: −1: X: X: X: X: X: 394: 287: 220: 217: 442: 369: 468: 367: 516: 364: 546: 363. The eyes may or may not be completely closed during a blink. According to the website blinkmatters.com, the range of fully closed eyes during a blink is between ninety percent to one hundred percent [46]. The row will be like: 415: 5: X: C: X: C: X: 451: 294: 182: 191: 491: 362: 513: 363: 554: 365: 577: 367. In this particular study, our experiments were only interested in the blink ID and eye completely closed (FC) columns; as a result, we ignored any additional information that may be provided. Table 2 provides an explanation of the features included in the dataset.
The Eyeblink8 dataset is more complex because it includes facial expressions, head gestures, and staring down at a keyboard. According to [46], this dataset has a total of 408 blinks across 70,992 video frames at 640 × 480-pixel resolution. This clip has an average length of between 5000 and 11,000 frames and was shot at 30 frames per second. There is only one video of a single participant chatting to the camera, so to speak, in the Talking Face dataset. In the video, the person can be seen smiling, laughing, and doing a “funny face” in a variety of situations. In addition, the frame rate is 30 frames per second, the resolution is 720 by 576, and 61 blinks that have been labeled.

2.4. Eye Blink Detection Flowchart

Figure 5 illustrates the blink detection method, and the frame-by-frame breakdown of the video is the initial stage. The facial landmarks feature [47] was implemented with the help of Dlib to detect the face. The detector used here is made up of classic histogram of oriented gradients (HOG) [48] feature along with a linear classifier. In order to identify face characteristics, including the ears, eyes, and nose, a facial landmarks detector was built inside Dlib [49,50]. Moreover, with the help of two lines, our research was able to identify blinks. The lines dividing the eye are drawn in two directions: horizontal and vertical. Blinking is the act of briefly closing the eyes and shifting the eyelids from one side to the other. Blinking is a natural thing to happen.
The eyes are closed or blinking when the eyeballs are not visible, the eyelids are closed, the upper and lower eyelids are fused, and the upper and lower eyelids are not connected. Further, when the eyes are opened, the vertical and horizontal lines are almost the same, but the vertical lines narrow or almost disappear when the eyes are closed. We may consider eye blinking if the EAR is less than the modified EAR threshold for three seconds. To perform our experiment, we used four alternative threshold values: 0.18, 0.2, 0.225, and 0.25. Additionally, we experimented with different EAR cutoffs and both video datasets.

3. Results

Table 3 summarizes the statistics for the predictions and test sets of video 1, video 2, and video 3. For the EAR threshold of 0.18, the total frame count for the prediction set of videos 1 is 1829, the number of closed frames analyzed was 23, and the number of blinks found was 2. On the other hand, the statistics for the test set state that there are 58 closed frames, and there are 14 blinks. This experiment exhibited an accuracy of 95.5% and an area under the curve (AUC) of 0.613. Furthermore, video 3 has 1302 frames with 182 closed frames. The maximum accuracy was obtained while implementing the 0.18 EAR threshold: 86.1%. Moreover, video 3 contains 2192 frames, totaling 73.24 seconds. Using an EAR threshold of 0.18 resulted in 89% accuracy for this dataset. In the third video, we see the minimal results of 47.5% accuracy and 0.594 AUC for an EAR threshold of 0.25 being used.
Moreover, Table 4 describes the statistics for the prediction and test sets of Talking Face and Eyeblink8 datasets. Talking Face has 5000 frames with 227 closed frames. The optimum accuracy was achieved while employing the 0.18 EAR threshold: 97.1% accuracy and 0.974 AUC. For Eyeblink8, video 8, the highest accuracy was obtained when using the 0.18 EAR threshold: 86.1% accuracy and 0.732 AUC. This dataset has 1302 frames with 182 closed frames and 18 blinks.
The best EAR threshold in our experiment was 0.18. This value provided the best accuracy and AUC values in all experiments. Hence, 0.25 is the worst EAR threshold value because it obtained the minimum accuracy and AUC values. Based on the experimental results, it can be concluded that the higher the EAR threshold, the lower the accuracy and AUC performance. In previous studies, it was said that the EAR threshold of 0.2 is the best value, but it was not for our experiment. In fact, 0.18 was the best EAR threshold in our work. Our dataset is unique because of the small eyes. The size of the eyes will certainly affect the EAR and EAR threshold values. Therefore, our dataset also has some challenges, namely, people driving cars and people wearing glasses.
Figure 6 shows the confusion matrix for the video 1, video 2, and video 3. Figure 6a describes the false positive (FP) value of 58 out of 58 positive labels (1.0000%) and false negative (FN) rate of 23 out of 1771 negative labels (0.0130%) for video 1 and EAR threshold 0.18. Figure 6c explains the false positive (FP) rate of 35 out of 61 positive labels (0.5738%) and false negative (FN) rate of 206 out of 2131 negative labels (0.0967%).
In our experiment, we analyzed videos frame by frame and identified eye blinks every three frames, as shown in Figure 7. The results of the experiment only show the blinks at the beginning, middle, and end frames. For instance, Figure 7a illustrates the 1st blink started in the 3rd frame, the middle of the action was in the 5th frame, and it ended in the 7th frame. Next, Figure 7b describes that the 2nd blink started in the 1555th frame, the middle of action was in the 1556th frame, and it ended in the 1557th frame. Moreover, Figure 7c explains the 3rd blink started in the 1563rd frame, the middle of the action was in the 1564th frame, and it ended in the 1565th frame.
Figure 8 exhibits the Video 3 eye blink prediction result frame by frame with the EAR threshold of 0.18. The 1st blink started in the 69th frame, the middle of the action was in the 71st frame, and it ended in the 72nd frame as shown in Figure 8a. The 2nd blink started in the 227th frame, the middle of the action was in the 230th frame, and it ended in the 232nd frame, as described in Figure 8b. Next, Figure 8c illustrates that the 3rd blink started in the 272nd frame, the middle of the action was in the 273rd frame, and it ended in the 274th frame.
Table 5 and Table 6 present the specific results that were obtained for each video dataset. These tables include the precision, recall, and F1-score measurements. In our testing, the best EAR threshold was 0.18, which was excellent. In all tests, this number yielded the highest accuracy and AUC values, respectively. As mentioned, 0.25 was the worst EAR threshold value, since it only achieved the bare minimum in terms of accuracy and AUC.
Researchers normally choose 0.2 or 0.3 as the EAR threshold, despite the fact that not everyone’s eye size is the same. As a result, it is better to recalculate the EAR threshold to detect whether the eye is closed or open in order to identify the blinks more accurately. For video 1 we achieved 96% accuracy, followed by video 3, with 89% accuracy, and video 2, with 86% accuracy, as shown in Table 5. Using Talking Face and Eyeblink8 datasets, we obtained the same accuracy, 97%, by employing the 0.18 EAR threshold, as shown in Table 6.

4. Discussion

The EAR and error analysis of the video 1 dataset is presented in Figure 9. The EAR threshold for this dataset was set to 0.18. Assume that the linear regression’s optimum slope is m ≥ 0. All the data from video 1 were plotted in our experiment, and the result is m = 0. Infrequent blinking has a small impact on the overall trend in EAR measurements depicted in Figure 10a. However, the cumulative error is irrelevant for blinks, owing to its delayed effect. Nevertheless, mistakes behave more similarly to correctly dispersed data than the EAR values in Figure 10b.
The effectiveness of the proposed method for detecting eye blinks was evaluated in this work by contrasting the detected blinks with the ground-truth blinks using the three video datasets. Overall, the output samples may be divided into three categories. The samples that were correctly identified are referred to as true positives (TP), the samples that had the wrong identification are referred to as false positives (FP), and the samples that were correctly not recognized are referred to as true negatives (TN) [51,52]. Precision (P) and recall (R) are represented by [53,54] in Equations (3) and (4).
P = TP TP + FP
R = TP TP + FN
Another evaluation index, F1 [55], is shown in Equation (5).
F 1 = 2   ×   Precision   ×   Recall Precision   +   Recall
There is a possibility that if the driver does not blink for a long time and his EAR value decreases without any blinking in the initial period, the algorithm will not return an error. Further, our work calculates errors as errors = calibration – linear and cumsum (). This function will return the cumulative sum of the elements along a given axis. A new array holding the result, in which case a reference to out is returned, is returned unless out is specified. The result has the same size and shape as if the axis were none or a 1d array. Cumulative errors are not very important for blinking, as their effects are delayed. However, typical errors can be exploited to detect anomalies. Figure 10 describes the EAR and error analysis of the video 3 dataset with an EAR threshold of 0.18. The average EAR value for video 3 was 0.25, as shown in Figure 9b. This average value is slightly different from the average value in Figure 10a, which is close to 0.30. During our tests, we found that the optimal EAR threshold is 0.18. This figure yielded the greatest accuracy and AUC values in all tests, both excellent results. The statistics are listed in Table 5 and Table 6.
Furthermore, Table 7 describes the evaluation of the proposed method in comparison to existing research. Our proposed method achieved peak average accuracies of 97.10% with the TalkingFace dataset, 97.00% with the Eyeblink8 dataset, and 96% with the Eye Video 1 dataset. We improved on the performances of previous methods.

5. Conclusions

In this article, we provide a method for automatically classifying blink types which includes establishing a threshold based on the Eye Aspect Ratio value. We call it Real-Time Blink Detection for Driver Safety using Eye Aspect Ratio. Using our eye blink dataset, we conducted a thorough analysis of the method. According to the experimental findings, the higher the EAR threshold, the worse the accuracy and AUC. Previously published studies showed that an EAR threshold of 0.2 was the optimal value; however, this was not the case in our experiment. In our study, 0.18 was the optimal EAR threshold. The experimental findings suggest that the EAR threshold for identifying whether the eyes are open or closed should be recalculated. Machine-learning techniques could be a viable alternative. In our future research, we will use explainable artificial intelligence (XAI) to explain our model when making certain predictions, as this is as important as prediction accuracy. In addition, our future work will explore the use of generative adversarial networks (GANs) to generate new synthetic data samples and improve image representation or quality [56,57]. We may consider coloring an image in grayscale, enhancing coloring, denoising, segmenting, or removing occlusion by objects [58].

Author Contributions

Conceptualization, C.D., R.-C.C. and X.J.; data curation, C.D. and S.-H.W.; formal analysis, C.D., S.-H.W. and H.Y.; investigation, C.D. and C.-W.C.; methodology, C.D.; project administration, R.-C.C. and X.J.; resources, C.D. and C.-W.C.; software, C.D., C.-W.C. and X.J.; supervision, R.-C.C., S.-H.W. and H.Y.; visualization, H.Y.; writing—original draft, C.D.; writing—review and editing, C.D. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the Ministry of Science and Technology, Taiwan. The numbers are MOST-111-2221-E-324 -020, and MOST-110 -2927-I-324 -501- Taiwan. Additionally, this study was partially funded by the EU Horizon 2020 program RISE Project ULTRACEPT under grant 778062.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to all subjects gave their informed consent for inclusion before they participated in the study.

Informed Consent Statement

All subjects gave their informed consent for inclusion before they participated in the study.

Data Availability Statement

Taiwan Eye Blink Dataset (https://drive.google.com/drive/u/1/folders/1U2qlw-ViqdW1pny77aJGLUIEf0B-HKzZ (accessed on 13 January 2021)).

Acknowledgments

The authors would like to thank the support and help from Chaoyang University of Technology, Satya Wacana Christian University, and others that took part in this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. LaRocco, J.; Le, M.D.; Paeng, D.G. A Systemic Review of Available Low-Cost EEG Headsets Used for Drowsiness Detection. Front. Neuroinform. 2020, 14, 42. [Google Scholar] [CrossRef]
  2. Rahman, A.; Sirshar, M.; Khan, A. Real Time Drowsiness Detection Using Eye Blink Monitoring. In Proceedings of the 2015 National Software Engineering Conference, NSEC 2015, Rawalpindi, Pakistan, 25–26 November 2015; pp. 1–7. [Google Scholar]
  3. Lemke, M.K.; Apostolopoulos, Y.; Sönmez, S. Syndemic Frameworks to Understand the Effects of COVID-19 on Commercial Driver Stress, Health, and Safety. J. Transp. Health 2020, 18, 100877. [Google Scholar] [CrossRef] [PubMed]
  4. Gagnon, S.; Stinchcombe, A.; Curtis, M.; Kateb, M.; Polgar, J.; Porter, M.M.; Bédard, M. Driving Safety Improves after Individualized Training: An RCT Involving Older Drivers in an Urban Area. Traffic Inj. Prev. 2019, 20, 595–600. [Google Scholar] [CrossRef] [PubMed]
  5. Koesdwiady, A.; Soua, R.; Karray, F.; Kamel, M.S. Recent Trends in Driver Safety Monitoring Systems: State of the Art and Challenges. IEEE Trans. Veh. Technol. 2017, 66, 4550–4563. [Google Scholar] [CrossRef]
  6. Al Tawil, L.; Aldokhayel, S.; Zeitouni, L.; Qadoumi, T.; Hussein, S.; Ahamed, S.S. Prevalence of Self-Reported Computer Vision Syndrome Symptoms and Its Associated Factors among University Students. Eur. J. Ophthalmol. 2020, 30, 189–195. [Google Scholar] [CrossRef]
  7. Drutarovsky, T.; Fogelton, A. Eye Blink Detection Using Variance of Motion Vectors. In Proceedings of the Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Braga, Portugal, 28 September–1 October 2015; Volume 8927, pp. 436–448. [Google Scholar]
  8. Pan, G.; Sun, L.; Wu, Z.; Lao, S. Eyeblink-Based Anti-Spoofing in Face Recognition from a Generic Webcamera. In Proceedings of the IEEE International Conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  9. Dewi, C.; Chen, R.C.; Yu, H. Weight Analysis for Various Prohibitory Sign Detection and Recognition Using Deep Learning. Multimed. Tools Appl. 2020, 79, 32897–32915. [Google Scholar] [CrossRef]
  10. Muhammad, K.; Ullah, A.; Lloret, J.; Ser, J.D.; De Albuquerque, V.H.C. Deep Learning for Safe Autonomous Driving: Current Challenges and Future Directions. IEEE Trans. Intell. Transp. Syst. 2021, 22, 4316–4336. [Google Scholar] [CrossRef]
  11. Dewi, C.; Chen, R.C.; Liu, Y.T. Wasserstein Generative Adversarial Networks for Realistic Traffic Sign Image Generation. In Proceedings of the Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Phuket, Thailand, 7–10 April 2021; pp. 479–493. [Google Scholar]
  12. Mimouna, A.; Alouani, I.; Ben, K.A.; El Hillali, Y.; Taleb-Ahmed, A.; Menhaj, A.; Ouahabi, A.; Amara, N.E. Ben OLIMP: A Heterogeneous Multimodal Dataset for Advanced Environment Perception. Electronics 2020, 9, 560. [Google Scholar] [CrossRef] [Green Version]
  13. Rosenfield, M. Computer Vision Syndrome: A Review of Ocular Causes and Potential Treatments. Ophthalmic Physiol. Opt. 2011, 31, 502–515. [Google Scholar] [CrossRef]
  14. Bentivoglio, A.R.; Bressman, S.B.; Cassetta, E.; Carretta, D.; Tonali, P.; Albanese, A. Analysis of Blink Rate Patterns in Normal Subjects. Mov. Disord. 1997, 12, 1028–1034. [Google Scholar] [CrossRef]
  15. Čech, J.; Franc, V.; Uřičář, M.; Matas, J. Multi-View Facial Landmark Detection by Using a 3D Shape Model. Image Vis. Comput. 2016, 47, 60–70. [Google Scholar] [CrossRef]
  16. Dong, X.; Yu, S.I.; Weng, X.; Wei, S.E.; Yang, Y.; Sheikh, Y. Supervision-by-Registration: An Unsupervised Approach to Improve the Precision of Facial Landmark Detectors. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 360–368. [Google Scholar]
  17. Dewi, C.; Chen, R.-C.; Jiang, X.; Yu, H. Adjusting Eye Aspect Ratio for Strong Eye Blink Detection Based on Facial Landmarks. PeerJ Comput. Sci. 2022, 8, e943. [Google Scholar] [CrossRef] [PubMed]
  18. Song, M.; Tao, D.; Sun, Z.; Li, X. Visual-Context Boosting for Eye Detection. IEEE Trans. Syst. Man. Cybern. B Cybern. 2010, 40, 1460–1467. [Google Scholar] [CrossRef] [PubMed]
  19. Lee, W.O.; Lee, E.C.; Park, K.R. Blink Detection Robust to Various Facial Poses. J. Neurosci. Methods 2010, 193, 356–372. [Google Scholar] [CrossRef]
  20. Park, C.W.; Park, K.T.; Moon, Y.S. Eye Detection Using Eye Filter and Minimisation of NMF-Based Reconstruction Error in Facial Image. Electron. Lett. 2010, 46, 130–132. [Google Scholar] [CrossRef] [Green Version]
  21. Li, F.; Chen, C.H.; Xu, G.; Khoo, L.P. Hierarchical Eye-Tracking Data Analytics for Human Fatigue Detection at a Traffic Control Center. IEEE Trans. Human-Mach. Syst. 2020, 50, 465–474. [Google Scholar] [CrossRef]
  22. García, I.; Bronte, S.; Bergasa, L.M.; Almazán, J.; Yebes, J. Vision-Based Drowsiness Detector for Real Driving Conditions. In Proceedings of the IEEE Intelligent Vehicles Symposium, Madrid, Spain, 3–7 June 2012; pp. 618–623. [Google Scholar]
  23. Maior, C.B.S.; das Chagas Moura, M.J.; Santana, J.M.M.; Lins, I.D. Real-Time Classification for Autonomous Drowsiness Detection Using Eye Aspect Ratio. Exp. Syst. Appl. 2020, 158, 113505. [Google Scholar] [CrossRef]
  24. Mehta, S.; Dadhich, S.; Gumber, S.; Jadhav Bhatt, A. Real-Time Driver Drowsiness Detection System Using Eye Aspect Ratio and Eye Closure Ratio. In Proceedings of the International Conference on Sustainable Computing in Science, Technology and Management (SUSCOM), Amity University Rajasthan, Jaipur, India, 26–28 February 2019; pp. 1333–1339. [Google Scholar]
  25. Wu, Y.; Ji, Q. Facial Landmark Detection: A Literature Survey. Int. J. Comput. Vis. 2019, 127, 115–142. [Google Scholar] [CrossRef] [Green Version]
  26. Dewi, C.; Chen, R.; Liu, Y.; Yu, H. Various Generative Adversarial Networks Model for Synthetic Prohibitory Sign Image Generation. Appl. Sci. 2021, 11, 2913. [Google Scholar] [CrossRef]
  27. Bergasa, L.M.; Nuevo, J.; Sotelo, M.A.; Barea, R.; Lopez, M.E. Real-Time System for Monitoring Driver Vigilance. IEEE Trans. Intell. Transp. Syst. 2006, 7, 63–77. [Google Scholar] [CrossRef]
  28. Dewi, C.; Chen, R.-C.; Jiang, X.; Yu, H. Deep Convolutional Neural Network for Enhancing Traffic Sign Recognition Developed on Yolo V4. Multimed. Tools Appl. 2022, 81, 37821–37845. [Google Scholar] [CrossRef]
  29. Fu, R.; Wang, H.; Zhao, W. Dynamic Driver Fatigue Detection Using Hidden Markov Model in Real Driving Condition. Exp. Syst. Appl. 2016, 63, 397–411. [Google Scholar] [CrossRef]
  30. You, F.; Li, X.; Gong, Y.; Wang, H.; Li, H. A Real-Time Driving Drowsiness Detection Algorithm with Individual Differences Consideration. IEEE Access 2019, 7, 179396–179408. [Google Scholar] [CrossRef]
  31. Zhao, X.; Meng, C.; Feng, M.; Chang, S.; Zeng, Q. Eye Feature Point Detection Based on Single Convolutional Neural Network. IET Comput. Vis. 2018, 12, 453–457. [Google Scholar] [CrossRef]
  32. Zhang, Z.; Luo, P.; Loy, C.C.; Tang, X. Learning Deep Representation for Face Alignment with Auxiliary Attributes. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 918–930. [Google Scholar] [CrossRef] [Green Version]
  33. Yue, X.; Li, J.; Wu, J.; Chang, J.; Wan, J.; Ma, J. Multi-Task Adversarial Autoencoder Network for Face Alignment in the Wild. Neurocomputing 2021, 437, 261–273. [Google Scholar] [CrossRef]
  34. Sun, Y.; Wang, X.; Tang, X. Deep Convolutional Network Cascade for Facial Point Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2013; pp. 3476–3483. [Google Scholar]
  35. Adjabi, I.; Ouahabi, A.; Benzaoui, A.; Taleb-Ahmed, A. Past, Present, and Future of Face Recognition: A Review. Electronics 2020, 9, 1188. [Google Scholar] [CrossRef]
  36. Adjabi, I.; Ouahabi, A.; Benzaoui, A.; Jacques, S. Multi-block Color-binarized Statistical Images for Single-sam-Ple Face Recognition. Sensors 2021, 21, 728. [Google Scholar] [CrossRef]
  37. El Morabit, S.; Rivenq, A.; Zighem, M.E.N.; Hadid, A.; Ouahabi, A.; Taleb-Ahmed, A. Automatic Pain Estimation from Facial Expressions: A Comparative Analysis Using off-the-Shelf Cnn Architectures. Electronics 2021, 10, 1926. [Google Scholar] [CrossRef]
  38. Jimenez-Pinto, J.; Torres-Torriti, M. Face Salient Points and Eyes Tracking for Robust Drowsiness Detection. Robotica 2012, 30, 731–741. [Google Scholar] [CrossRef]
  39. Lawrenson, J.G.; Birhah, R.; Murphy, P.J. Tear-Film Lipid Layer Morphology and Corneal Sensation in the Development of Blinking in Neonates and Infants. J. Anat. 2005, 206, 265–270. [Google Scholar] [CrossRef] [PubMed]
  40. Perelman, B.S. Detecting Deception via Eyeblink Frequency Modulation. PeerJ 2014, 2, e260. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Lin, X.; Wan, J.; Xie, Y.; Zhang, S.; Lin, C.; Liang, Y.; Guo, G.; Li, S.Z. Task-Oriented Feature-Fused Network with Multivariate Dataset for Joint Face Analysis. IEEE Trans. Cybern. 2020, 50, 1292–1305. [Google Scholar] [CrossRef]
  42. Kazemi, V.; Sullivan, J. One Millisecond Face Alignment with an Ensemble of Regression Trees. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1867–1874. [Google Scholar]
  43. Sugawara, E.; Nikaido, H. Properties of AdeABC and AdeIJK Efflux Systems of Acinetobacter Baumannii Compared with Those of the AcrAB-TolC System of Escherichia Coli. Antimicrob. Agents Chemother. 2014, 58, 7250–7257. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Rakshita, R. Communication Through Real-Time Video Oculography Using Face Landmark Detection. In Proceedings of the International Conference on Inventive Communication and Computational Technologies, ICICCT 2018, Coimbatore, India, 20–21 April 2018; pp. 1094–1098. [Google Scholar]
  45. Noor, A.Z.M.; Jafar, F.A.; Ibrahim, M.R.; Soid, S.N.M. Fatigue Detection among Operators in Industry Based on Euclidean Distance Computation Using Python Software. Int. J. Emerg. Trends Eng. Res. 2020, 8, 6375–6379. [Google Scholar] [CrossRef]
  46. Fogelton, A.; Benesova, W. Eye Blink Detection Based on Motion Vectors Analysis. Comput. Vis. Image Underst. 2016, 148, 23–33. [Google Scholar] [CrossRef]
  47. Tang, X.; Guo, F.; Shen, J.; Du, T. Facial Landmark Detection by Semi-Supervised Deep Learning. Neurocomputing 2018, 297, 22–32. [Google Scholar] [CrossRef]
  48. Dhiraj; Jain, D.K. An Evaluation of Deep Learning Based Object Detection Strategies for Threat Object Detection in Baggage Security Imagery. Pattern Recognit. Lett. 2019, 120, 112–119. [Google Scholar] [CrossRef]
  49. King, D.E. Dlib-Ml: A Machine Learning Toolkit. J. Mach. Learn. Res. 2009, 10, 1755–1758. [Google Scholar]
  50. Eriksson, M.; Papanikolopoulos, N.P. Eye-Tracking for Detection of Driver Fatigue. In Proceedings of the IEEE Conference on Intelligent Transportation Systems Proceedings, ITSC, Boston, MA, USA, 9–12 November 1997. [Google Scholar]
  51. Dewi, C.; Chen, R.-C.; Liu, Y.-T.; Tai, S.-K. Synthetic Data Generation Using DCGAN for Improved Traffic Sign Recognition. Neural Comput. Appl. 2021, 33, 1–15. [Google Scholar] [CrossRef]
  52. Chen, R.C.; Saravanarajan, V.S.; Hung, H. Te Monitoring the Behaviours of Pet Cat Based on YOLO Model and Raspberry Pi. Int. J. Appl. Sci. Eng. 2021, 18, 1–12. [Google Scholar] [CrossRef]
  53. Yang, H.; Chen, L.; Chen, M.; Ma, Z.; Deng, F.; Li, M.; Li, X. Tender Tea Shoots Recognition and Positioning for Picking Robot Using Improved YOLO-V3 Model. IEEE Access 2019, 7, 180998–181011. [Google Scholar] [CrossRef]
  54. Yuan, Y.; Xiong, Z.; Wang, Q. An Incremental Framework for Video-Based Traffic Sign Detection, Tracking, and Recognition. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1918–1929. [Google Scholar] [CrossRef]
  55. Tian, Y.; Yang, G.; Wang, Z.; Wang, H.; Li, E.; Liang, Z. Apple Detection during Different Growth Stages in Orchards Using the Improved YOLO-V3 Model. Comput. Electron. Agric. 2019, 157, 417–426. [Google Scholar] [CrossRef]
  56. Khan, A.; Jin, W.; Haider, A.; Rahman, M.; Wang, D. Adversarial Gaussian Denoiser for Multiple-Level Image Denoising. Sensors 2021, 21, 2998. [Google Scholar] [CrossRef]
  57. Khaldi, Y.; Benzaoui, A.; Ouahabi, A.; Jacques, S.; Taleb-Ahmed, A. Ear Recognition Based on Deep Unsupervised Active Learning. IEEE Sens. J. 2021, 21, 20704–20713. [Google Scholar] [CrossRef]
  58. Khaldi, Y.; Benzaoui, A. A New Framework for Grayscale Ear Images Recognition Using Generative Adversarial Networks under Unconstrained Conditions. Evol. Syst. 2021, 12, 923–934. [Google Scholar] [CrossRef]
Figure 1. Eye identification made possible by employing facial landmarks (right eye points = 36–41, left eye points = 42–47).
Figure 1. Eye identification made possible by employing facial landmarks (right eye points = 36–41, left eye points = 42–47).
Electronics 11 03183 g001
Figure 2. Open and closed eyes with facial landmarks (P1, P2, P3, P4, P5, P6). (a) Open eye. (b) Close eye.
Figure 2. Open and closed eyes with facial landmarks (P1, P2, P3, P4, P5, P6). (a) Open eye. (b) Close eye.
Electronics 11 03183 g002
Figure 3. The system architecture.
Figure 3. The system architecture.
Electronics 11 03183 g003
Figure 4. Video labeling process with Eyeblink Annotator 3.0.
Figure 4. Video labeling process with Eyeblink Annotator 3.0.
Electronics 11 03183 g004
Figure 5. Eye blink detection flowchart.
Figure 5. Eye blink detection flowchart.
Electronics 11 03183 g005
Figure 6. Confusion matrix (Eye Blink dataset). (a) Video 1 with 0.18 EAR threshold, (b) Video 2 with 0.18 EAR threshold, (c) Video 3 with 0.18 EAR threshold, (d) Video 1 with 0.2 EAR threshold, (e) Video 2 with 0.2 EAR threshold, (f) Video 3 with 0.2 EAR threshold, (g) Video 1 with 0.225 EAR threshold, (h) Video 2 with 0.225 EAR threshold, (i) Video 3 with 0.225 EAR threshold, (j) Video 1 with 0.25 EAR threshold, (k) Video 2 with 0.25 EAR threshold, (l) Video 3 with 0.25 EAR threshold.
Figure 6. Confusion matrix (Eye Blink dataset). (a) Video 1 with 0.18 EAR threshold, (b) Video 2 with 0.18 EAR threshold, (c) Video 3 with 0.18 EAR threshold, (d) Video 1 with 0.2 EAR threshold, (e) Video 2 with 0.2 EAR threshold, (f) Video 3 with 0.2 EAR threshold, (g) Video 1 with 0.225 EAR threshold, (h) Video 2 with 0.225 EAR threshold, (i) Video 3 with 0.225 EAR threshold, (j) Video 1 with 0.25 EAR threshold, (k) Video 2 with 0.25 EAR threshold, (l) Video 3 with 0.25 EAR threshold.
Electronics 11 03183 g006
Figure 7. Video 1 eye blink prediction results frame by frame (threshold = 0.18). (a) The 1st blink started in the 3rd frame, the middle of the action in the 5th frame, and it ended in the 7th frame. (b) The 2nd blink started in the 1555th frame, the middle of the action was in the 1556th frame, and it ended in the 1557th frame. (c) The 3rd blink started in the 1563rd frame, the middle of the action was in the 1564th frame, and it ended in the 1565th frame.
Figure 7. Video 1 eye blink prediction results frame by frame (threshold = 0.18). (a) The 1st blink started in the 3rd frame, the middle of the action in the 5th frame, and it ended in the 7th frame. (b) The 2nd blink started in the 1555th frame, the middle of the action was in the 1556th frame, and it ended in the 1557th frame. (c) The 3rd blink started in the 1563rd frame, the middle of the action was in the 1564th frame, and it ended in the 1565th frame.
Electronics 11 03183 g007
Figure 8. Video 2 Eye Blink Prediction result frame by frame (Threshold = 0.18). (a) The 1st blink started at the 69th frame, the middle of the action was in the 71st frame, and it ended in the 72nd frame. (b) The 2nd blink started at the 227th frame, the middle of the action was in the 230th frame, and it ended in the frame. (c) The 3rd blink started at the 272tnd frame, the middle of the action was in the 273rd frame, and it ended in the 274th frame.
Figure 8. Video 2 Eye Blink Prediction result frame by frame (Threshold = 0.18). (a) The 1st blink started at the 69th frame, the middle of the action was in the 71st frame, and it ended in the 72nd frame. (b) The 2nd blink started at the 227th frame, the middle of the action was in the 230th frame, and it ended in the frame. (c) The 3rd blink started at the 272tnd frame, the middle of the action was in the 273rd frame, and it ended in the 274th frame.
Electronics 11 03183 g008
Figure 9. EAR and error analysis (Video 1, EAR threshold = 0.18). (a) Average EAR. (b) Error and EAR.
Figure 9. EAR and error analysis (Video 1, EAR threshold = 0.18). (a) Average EAR. (b) Error and EAR.
Electronics 11 03183 g009
Figure 10. EAR and error analysis (Video 3, EAR threshold = 0.18). (a) Average EAR. (b) Error and EAR.
Figure 10. EAR and error analysis (Video 3, EAR threshold = 0.18). (a) Average EAR. (b) Error and EAR.
Electronics 11 03183 g010
Table 1. Video dataset information.
Table 1. Video dataset information.
Video InfoVideo 1Video 2Video 3Talking FaceEyeblink8 Video 8
FPS29.972429.973030
Frame Count182913022195500010,712
Durations (s)61.0354.2573.24166.67357.07
Size (MB)29.612.438.62218.6
Table 2. Dataset features.
Table 2. Dataset features.
NoDescriptionFeatures
1Alternatively, a frame counter may be used to get a timestamp in a different file.frame ID
2A unique blink ID is defined as a sequence of blink ID frames that are all the identical. The time between two consecutive blinks is measured in terms of a sequence of identical blink ID frames.blink ID
3A change from X to N occurs in the provided variable while the person is looking sideways and blinking.non frontal face (NF)
4Left Eye.left eye (LE),
5Right Eye.right eye (RE),
6Face.face (F)
7The given flag will transition from X to C if the subject’s eye closure percentage is between 90% and 100%.eye fully closed (FC)
8This variable changes from X to N when the subject’s eye is covered (by the subject’s hand, by low lighting, or by the subject’s excessive head movement).eye not visible (NV)
9x and y coordinates, width, height.face bounding box (F_X, F_Y, F_W, F_H)
10RX (right corner x coordinate), LY (left corner y coordinate)left and right eye corners positions
Table 3. Statistics for prediction and test sets (Eye Blink Dataset).
Table 3. Statistics for prediction and test sets (Eye Blink Dataset).
DatasetVideo 1Video 2Video 3
EAR Threshold (t)0.180.20.2250.250.180.20.2250.250.180.20.2250.25
Statistics on the prediction set are
Total Number of Frames Processed182918291829182913021302130213022192219221922192
Number of Closed Frames23561312811823426148842324407911177
Number of Blinks269161839736525497989
Statistics on the test set are
Total Number of Frames Processed182918291829182913021302130213022192219221922192
Number of Closed Frames585858583535353561616161
Number of Blinks14141414999910101010
Eye Closeness Frame by Frame Test Scores
Accuracy0.9550.9380.8970.8200.8610.740.5430.3400.8900.7970.6450.475
AUC0.6130.5810.5280.5010.7320.6920.6540.5910.6640.6410.6260.594
Table 4. Statistics on prediction and test (Talking Face and Eyeblink8 datasets).
Table 4. Statistics on prediction and test (Talking Face and Eyeblink8 datasets).
DatasetTalking FaceEyeblink8 Video 8
EAR Threshold (t)0.180.20.2250.250.180.20.2250.25
Statistics on the prediction set are
Total Number of Frames Processed500050005000500010,66310,66310,66310,663
Number of Closed Frames22729235248440452910552002
Number of Blinks31424959374385126
Statistics on the test set are
Total Number of Frames Processed500050005000500010,66310,66310,66310,663
Number of Closed Frames153153153153107107107107
Number of Blinks6161616130303030
Eye Closeness Frame by Frame Test Scores
Accuracy0.9710.9680.9590.9330.9700.9590.9110.911
AUC0.9740.9680.9530.9460.9630.9610.9550.955
Table 5. Precision, recall, and F1-score (Eye Blink dataset).
Table 5. Precision, recall, and F1-score (Eye Blink dataset).
EvaluationVideo 1Video 2Video 3
PrecisionRecallF1-Score SupportPrecisionRecall F1-ScoreSupportPrecisionRecallF1-ScoreSupport
EAR Threshold (t) = 0.18EAR Threshold (t) = 0.18EAR Threshold (t) = 0.18
00.970.990.9817710.980.870.9212670.980.900.942131
10.000.000.00580.100.510.17350.110.430.1861
Macro avg0.480.490.4918290.540.690.5513020.550.660.562192
Weight avg0.940.960.9518290.960.860.9013020.960.890.922192
Accuracy 0.961829 0.861302 0.892192
EAR Threshold (t) = 0.2EAR Threshold (t) = 0.2EAR Threshold (t) = 0.2
00.970.970.9717710.990.750.8512670.980.810.892131
10.020.020.02580.070.710.13350.070.480.1261
Macro avg0.490.490.4918290.530.730.4913020.520.640.502192
Weight avg0.940.940.9418290.960.750.8313020.960.800.862192
Accuracy 0.941829 0.751302 0.802192
EAR Threshold (t) = 0.225EAR Threshold (t) = 0.225EAR Threshold (t) = 0.225
00.970.930.9517710.990.540.7012670.980.650.782131
10.010.020.01580.040.770.08350.050.610.0961
Macro avg0.490.470.4818290.520.650.3913020.510.630.432192
Weight avg0.940.900.9218290.960.540.6813020.960.650.762192
Accuracy 0.901829 0.541302 0.652192
EAR Threshold (t) = 0.25EAR Threshold (t) = 0.25EAR Threshold (t) = 0.25
00.970.840.9017710.990.330.4912670.980.470.632131
10.020.090.03580.030.850.07350.040.720.0761
Macro avg0.490.470.4718290.510.590.2813020.510.590.352192
Weight avg0.940.820.8718290.960.340.4813020.960.480.622192
Accuracy 0.821829 0.341302 0.482192
Table 6. Precision, recall, and F1-score (Talking Face and Eyeblink8 datasets).
Table 6. Precision, recall, and F1-score (Talking Face and Eyeblink8 datasets).
EvaluationTalking FaceEyeblink8 Video 8
PrecisionRecallF1-Score SupportPrecisionRecallF1-ScoreSupport
EAR Threshold (t) = 0.18EAR Threshold (t) = 0.18
00.990.980.9948471.000.970.9810,556
10.520.770.621530.240.920.38107
Macro avg0.760.870.8050000.620.940.6810,663
Weight avg0.980.970.9750000.990.970.9810,663
Accuracy 0.975000 0.9710,663
EAR Threshold (t) = 0.2EAR Threshold (t) = 0.2
01.000.970.9848471.000.960.9810,556
10.490.930.641530.190.960.32107
Macro avg0.740.950.8150000.600.960.6510,663
Weight avg0.980.970.9750000.990.960.9710,663
Accuracy 0.975000 0.9610,663
EAR Threshold (t) = 0.225EAR Threshold (t) = 0.225
01.000.960.9848471.000.910.9510,556
10.430.990.601530.101.000.18107
Macro avg0.710.970.7950000.550.960.5710,663
Weight avg0.980.960.9750000.990.910.9510,663
Accuracy 0.965000 0.9110,663
EAR Threshold (t) = 0.25EAR Threshold (t) = 0.25
01.000.930.9648471.000.910.9510,556
10.321.000.481530.101.000.18107
Macro avg0.660.970.7250000.550.960.5710,663
Weight avg0.980.930.9550000.990.910.9510,663
Accuracy 0.935000 0.9110,663
Table 7. Evaluation of the proposed method in comparison to existing research.
Table 7. Evaluation of the proposed method in comparison to existing research.
ReferenceDataset Accuracy (%)
Talking FaceEyeblink8Eye Video 1
Drutarovskys et al. [7]92.2079.00-
Fogelton et al. [46]95.0094.69-
Proposed Method97.1097.0096.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dewi, C.; Chen, R.-C.; Chang, C.-W.; Wu, S.-H.; Jiang, X.; Yu, H. Eye Aspect Ratio for Real-Time Drowsiness Detection to Improve Driver Safety. Electronics 2022, 11, 3183. https://doi.org/10.3390/electronics11193183

AMA Style

Dewi C, Chen R-C, Chang C-W, Wu S-H, Jiang X, Yu H. Eye Aspect Ratio for Real-Time Drowsiness Detection to Improve Driver Safety. Electronics. 2022; 11(19):3183. https://doi.org/10.3390/electronics11193183

Chicago/Turabian Style

Dewi, Christine, Rung-Ching Chen, Chun-Wei Chang, Shih-Hung Wu, Xiaoyi Jiang, and Hui Yu. 2022. "Eye Aspect Ratio for Real-Time Drowsiness Detection to Improve Driver Safety" Electronics 11, no. 19: 3183. https://doi.org/10.3390/electronics11193183

APA Style

Dewi, C., Chen, R. -C., Chang, C. -W., Wu, S. -H., Jiang, X., & Yu, H. (2022). Eye Aspect Ratio for Real-Time Drowsiness Detection to Improve Driver Safety. Electronics, 11(19), 3183. https://doi.org/10.3390/electronics11193183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop