Next Article in Journal
A Multitask Learning Approach for Named Entity Recognition by Exploiting Sentence-Level Semantics Globally
Previous Article in Journal
Adaptive Control Method of Sensorless Permanent Magnet Synchronous Motor Based on Super-Twisting Sliding Mode Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Facial Action Unit Recognition by Prior and Adaptive Attention

1
School of Computer Science and Technology, China University of Mining and Technology, Xuzhou 221116, China
2
Engineering Research Center of Mine Digitization, Ministry of Education of the People’s Republic of China, Xuzhou 221116, China
3
Xuzhou Guanglian Technology Co., Ltd., Xuzhou 221116, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(19), 3047; https://doi.org/10.3390/electronics11193047
Submission received: 31 August 2022 / Revised: 19 September 2022 / Accepted: 21 September 2022 / Published: 24 September 2022
(This article belongs to the Topic Computer Vision and Image Processing)

Abstract

:
Facial action unit (AU) recognition remains a challenging task, due to the subtlety and non-rigidity of AUs. A typical solution is to localize the correlated regions of each AU. Current works often predefine the region of interest (ROI) of each AU via prior knowledge, or try to capture the ROI only by the supervision of AU recognition during training. However, the predefinition often neglects important regions, while the supervision is insufficient to precisely localize ROIs. In this paper, we propose a novel AU recognition method by prior and adaptive attention. Specifically, we predefine a mask for each AU, in which the locations farther away from the AU centers specified by prior knowledge have lower weights. A learnable parameter is adopted to control the importance of different locations. Then, we element-wise multiply the mask by a learnable attention map, and use the new attention map to extract the AU-related feature, in which AU recognition can supervise the adaptive learning of a new attention map. Experimental results show that our method (i) outperforms the state-of-the-art AU recognition approaches on challenging benchmark datasets, and (ii) can accurately reason the regional attention distribution of each AU by combining the advantages of both the predefinition and the supervision.

1. Introduction

Facial action unit (AU) recognition involves the prediction for occurrence or non-occurrence of each AU, which is an important task in the communities of computer vision and affective computing [1,2,3,4]. As defined in the facial action coding system (FACS) [5,6], each AU is a local facial action with one or more atomic muscle actions. Due to the subtlety and non-rigidity, the appearance of AUs are diversely changed across persons and expressions. For instance, as shown in Figure 1, AU 1 (inner brow raiser), AU 2 (outer brow raiser), and AU 4 (brow lowerer) occur in brow regions with overlaps, in which it is difficult to distinguish each AU from the fused appearance. In the literature, facial AU recognition is still a challenging task.
Considering AUs appear in local facial regions, one intuitive solution is to localize the correlated regions so as to extract features for AU recognition. Since the locations of AU centers can be specified by correlated facial landmarks via prior knowledge, Li et al. [2,7] predefined a regional attention map for each AU, in which a position with a farther distance to the AU centers is given a lower attention value. However, different AUs share the same attention distribution, which ignores the divergences across AUs. Furthermore, correlated landmarks only can determine the central locations of AUs, while a few potentially correlated regions very far away from the centers are neglected.
As deep neural networks have a self-attention mechanism [8] during training, Shao et al. [9] only resorted to the supervision of AU recognition to capture correlated regions of AUs. In this approach, some irrelevant regions are included since AUs are subtle and do not have distinct contours. Shao et al. [3,10] proposed adaptively modifying the predefined attention map of each AU, which is a pioneering work of combining predefined attention and supervised attention. However, directly convoluting on a predefined attention map only works as smoothing, in which positions outside of the predefined region of interest (ROI) obtain similar attention weights and thus are regarded as similar importance. In this way, correlated regions very distant to the AU centers are still not emphasized.
To tackle the above issues, we develop a novel facial AU recognition method named PAA by prior and adaptive attention. In particular, we first predefine a mask for each AU by assigning lower weights to the positions farther away from the predefined AU centers. Since the regions outside the predefined ROI should not be neglected, we use a learnable parameter to adaptively control the importance of different positions. Then, we element-wise multiply the mask by a learnable attention map, and employ the new attention map to extract AU-related features. In this process, the new attention map is adaptively learned under the guidance of AU recognition. By integrating the advantages of the predefinition and the supervision, our method can precisely capture correlated regions of each AU, in which correlated locations are included and uncorrelated locations are discarded.
This paper has three main contributions:
  • We propose to combine the constraint of prior knowledge and the supervision of AU recognition to adaptively learn the regional attention distribution of each AU.
  • We propose a learnable parameter to adaptively control the importance of different positions in the predefined mask, which is beneficial for choosing an appropriate constraint of prior knowledge.
  • We conduct extensive experiments on challenging benchmark datasets, which demonstrate that our method outperforms the state-of-the-art AU recognition approaches, and can precisely learn the attention map of each AU.

2. Related Works

In this section, we review other approaches that are strongly relevant to our work, including facial landmark-aided AU recognition approaches and attention learning-based AU recognition approaches.

2.1. Facial Landmark-Aided AU Recognition

Since facial landmarks have prior location relationships with AUs, landmarks can help to learn AU-related features. Benitez-Quiroz et al. [11] integrated the geometry and local texture feature for AU recognition, in which the geometry feature contains the normalized distances among landmarks as well as the angles of Delaunay mask constructed by landmarks. Zhao et al. [12] extracted scale-invariant feature transform (SIFT) [13] features from local facial regions centered at relevant landmarks as AU-related features.
Facial landmarks can also facilitate AU recognition in other ways. Niu et al. [14] relied on landmarks to construct a shape regularization to AU recognition. Ma et al. [15] introduced typical object detection tasks into AU recognition by employing landmarks to define bounding boxes for AUs, in which each AU is predicted to occur in which bounding box. If one AU is absent, it should be predicted as non-occurrence for all bounding boxes.
These approaches all demonstrate the contribution of landmarks to AU recognition. In this paper, we use landmarks to predefine a regional mask with a learnable control parameter for each AU.

2.2. Attention Learning-Based AU Recognition

Considering AUs are subtle and have no distinct contour and texture, it is impracticable to manually annotate their regional attention distribution. An intuitive solution is to use the prior knowledge for attention predefinition. Li et al. [2,7] generated an attention map for each AU by predefining two Gaussian distributions centered around the two AU centers due to the symmetry, in which a position farther away from the centers has a smaller attention weight. Different AUs have identical attention distributions, which ignore the differences among AUs. Furthermore, attention predefinition cannot highlight potentially correlated regions far away from the predefined ROI.
As important region beyond the predefined ROI may be neglected, Shao et al. [9] directly learned the attention map of each AU without the prior knowledge, while Shao et al. [3,10] modified the predefined attention map under the supervision of AU recognition. However, the attention distribution learned in [9] contains quite a few uncorrelated locations, and the attention modification in [3,10] seems to be the smoothing of predefintion and still cannot emphasize correlated locations very distant to the predefined ROI. On the contrary, our approach can include both strongly correlated positions near the AU centers and weakly correlated positions scattered on the global face.

3. PAA for Facial AU Recognition

3.1. Overview

Our main goal is to predict the occurrence probability of total m AUs for an input image: p ^ = ( p ^ 1 , , p ^ m ) . The structure of our network is illustrated in Figure 2. Specifically, two hierarchical and multi-scale region layers [3,10] with each, followed by a max-pooling layer, are firstly used to extract a multi-scale feature, which is beneficial for adapting to AUs with diverse sizes in different local facial regions. Then, each AU has one branch to predict the occurrence probability, in which the predefined mask M i , as the prior knowledge, constrains the learning of new attention map M ^ i ( 1 ) during training. In our framework, both the prior knowledge and the AU recognition guidance are exploited to learn the regional attention distribution of each AU.

3.2. Constraint of Attention Predefinition

In the branch of the i-th AU, three convolutional layers are firstly adopted, where i = 1 , , m . Then, a convolutional layer with one channel is used to learn the attention map M ^ i ( 0 ) . According to the prior knowledge, the central locations of AUs can be determined by correlated facial landmarks [2,3], as illustrated in Figure 3. To exploit the prior knowledge, we predefine a mask M i for the i-th AU.
Since the i-th AU has two centers ( a ¯ i ( 1 ) , b ¯ i ( 1 ) ) and ( a ¯ i ( 2 ) , b ¯ i ( 2 ) ) , we take the predefined mask M ˜ i ( 1 ) of the first center as an example. In particular, we use a Gaussian distribution with a standard deviation δ centered around the location ( a ¯ i ( 1 ) , b ¯ i ( 1 ) ) to compute the value at each location ( a , b ) :
M ˜ i a b ( 1 ) = exp ( ( a a ¯ i ( 1 ) ) 2 + ( b b ¯ i ( 1 ) ) 2 2 δ 2 ) .
We next incorporate M ˜ i ( 1 ) and M ˜ i ( 2 ) by choosing the larger value at each location:
M ˜ i a b = max ( M ˜ i a b ( 1 ) , M ˜ i a b ( 2 ) ) ( 0 , 1 ] .
In M ˜ i , the positions with values significantly larger than zero constitute the ROI of the i-th AU, while other approximately zero-valued positions are ignored. However, the positions beyond the predefined ROI that we do not want are completely discarded during the constraint for attention learning. We introduce a learnable—instead of fixed—control parameter ϵ i to give appropriate importance to the positions outside of the ROI:
M i a b = M ˜ i a b + ϵ i 1 + ϵ i ( 0 , 1 ] ,
where ϵ i 0 and a larger ϵ i give larger importance to the positions beyond the predefined ROI. Note that the relative size between different positions in M ˜ i is unchanged, and different AUs have independent control parameters. As illustrated in Figure 2, M 1 and M m are adaptively learned with different attention distributions.

Supervision of AU Recognition

After obtaining the predefined mask M i , we generate the new attention map M ^ i ( 1 ) by element-wise multiplying M i by M ^ i ( 0 ) :
M ^ i ( 1 ) = M i M ^ i ( 0 ) .
Considering that deep neural networks have a self-attention mechanism [8], we exploit AU recognition to guide the learning of M ^ i ( 1 ) .
Specifically, as shown in Figure 2, we element-wise multiply M ^ i ( 1 ) with the fourth convolutional channel map to emphasize AU-related features. Then, another convolutional layer, as well as a global average pooling layer [16], are adopted to extract the AU feature with the size of 12 c . Finally, we predict the occurrence probability p ^ i of the i-th AU by using a fully-connected layer with one dimension followed by a Sigmoid function, and define the AU recognition loss as:
L u = i = 1 m w i [ v i p i log p ^ i + ( 1 p i ) log ( 1 p ^ i ) ] ,
where a weighting strategy is employed, and p i , w i , and v i denote the ground-truth occurrence probability, the weight, and the weight for occurrence of the i-th AU, respectively. There are two types of data imbalance issues [17] in most existing AU datasets [18,19,20]: inter-AU data imbalance that different AUs have different occurrence rates, and intra-AU data imbalance that AUs often have smaller occurrence rates than non-occurrence rates. To alleviate these data imbalance issues, w i and v i are defined as:
w i = n n i o c c / k = 1 m n n k o c c , v i = n n i o c c n i o c c ,
where n i o c c / n is the occurrence rate of the i-th AU, and n i o c c and n denote the number of images appearing in the i-th AU and the number of all images in the training set, respectively.
By the constraint of attention predefinition and the guidance of AU recognition, the adaptively learned AU attention map M ^ i ( 1 ) can capture both strongly relevant regions predefined by prior knowledge as well as scattered relevant regions on the global face. In this case, our AU recognition method can work well under the subtlety and non-rigidity of AUs due to the accuracy of AU-related features.

4. Experiments

4.1. Datasets and Settings

4.1.1. Datasets

In this paper, we evaluate our PAA on three popular benchmark datasets. Besides AU annotations, each dataset is also annotated with facial landmarks.
  • Binghamton-Pittsburgh 4D (BP4D) [18] includes 41 subjects, including 23 women and 18 men, in which 328 videos with about 140,000 frames are captured in total by placing each subject into 8 sessions. Each frame is labeled with the AUs of occurrence or non-occurrence. Similar to the previous approaches [1,2,3], we conduct subject-exclusive three-fold cross-validation with two folds for training and the remaining one for testing on 12 AUs. Our method uses the same partitions of subjects as the previous works [1,2,3].
  • Denver Intensity of Spontaneous Facial Action (DISFA) [19] contains 12 women and 15 men, in which each subject is recorded by a video with 4845 frames. Each frame is labeled with AU intensities ranging from 0 to 5. Following the previous methods [1,2,3], we treat an AU as occurrence if its intensity is equal or larger than two, and treat it as non-occurrence otherwise. We conduct subject-exclusive three-fold cross-validation on eight AUs. Our method uses the same partitions of subjects as the previous works [1,2,3].
  • Sayette Group Formation Task (GFT) [20] includes 96 subjects with each subject captured by one video, whose images are more challenging than BP4D and DISFA due to unscripted interactions in 32 three-subject teams. Each frame is labeled with AU occurrences. We adopt the official training and testing partitions of subjects [20], in which about 108,000 frames of 78 subjects are used for training, and about 24,600 frames of 18 subjects are used for testing.

4.1.2. Implementation Details

Our PAA is implemented via PyTorch [21], in which each convolutional layer adopts 3 × 3 convolutional filters with a stride of 1, a padding of 1, and each max-pooling layer processes 2 × 2 spatial fields with a stride of 2. We normalize each image to 3 × 200 × 200 by similarity transformation, and randomly crop the normalized image to 3 × l × l with a random horizontal flip. The image size l, the network parameter c, as well as the standard deviation δ in Equation (1) are set to 176, 8, and 3, respectively.
Similar to JÂA-Net [3], we employ the stochastic gradient descent (SGD) solver with a Nesterov momentum [22] of 0.9 , a weight decay of 0.0005 , and a mini-batch size of 8 to train PAA with 12 epochs. The learning rate is initialized to be 0.006 and is multiplied by 0.3 at every 2 epochs during training. Following the settings in [1,2,3], we use the parameters of the well-trained model on BP4D for initialization when training on DISFA.

4.1.3. Evaluation Metrics

We evaluate methods via a popular metric frame-based F1-score (F1-frame):
F 1 = 2 P R P + R ,
where P denotes the precision, and R denotes the recall. We also report the average results of the F1-frame over all AUs (shortly written as Avg). We omit “%” in all the F1-frame results for simplicity in the experimental results.

4.2. Comparison with State-of-the-Art Methods

In this section, we compare our PAA approach with state-of-the-art AU recognition methods, including LSVM [23], APL [24], JPML [12], AlexNet [25], DRML [1], EAC-Net [2], DSIN [26], CMS [27], LP-Net [14], ARL [9], SRERL [28], AU R-CNN [15], TCAE [29], AU-GCN [30], Ertugrul et al. [31], JÂA-Net [3], UGN-B [32], HMP-PS [33], and GeoCNN [34]. Notice that these works often adopt external training data, while our PAA uses the benchmark dataset only. Specifically, EAC-Net, SRERL, AU R-CNN, UGN-B, HMP-PS, and GeoCNN use pre-trained ImageNet models [35,36], CMS employs external thermal images, LP-Net pre-trains on a face recognition dataset [37], and GeoCNN utilizes a pre-trained 3D morphable model (3DMM) [38,39]. Several related works such as R-T1 [7] are not compared since they process a sequence of frames instead of a single frame.

4.2.1. Evaluation on BP4D

Table 1 shows the results of different methods on the BP4D benchmark. We can observe that our PAA performs better than most of the previous works, especially for the average F1-frame. Compared to other methods using external training data, such as AU R-CNN and UGN-B, PAA uses benchmark training images only, while achieving better performance. Although GeoCNN is slightly better than our method, it relies on a pre-trained 3DMM to obtain additional 3D manifold information to facilitate AU recognition.

4.2.2. Evaluation on DISFA

Table 2 shows the results on DISFA, from which we can see that our PAA achieves competitive performance. It can also be found that many methods such as AU-GCN exhibit more fluctuated results across AUs on DISFA than on BP4D, and work well on BP4D, but show poor results on DISFA. This is because DISFA is more challenging with a more severe data imbalance problem than BP4D. In this case, our PAA achieves a more stable performance among different AUs than most of the previous works, and performs consistently well on BP4D and DISFA with 63.4 and 62.9 average F1-frame results, respectively.

4.2.3. Evaluation of GFT

The comparison results on GFT are presented in Table 3. It can be observed that our PAA outperforms all other approaches. Notice that GFT images are often in large poses, which are more challenging than BP4D and DISFA images with near-frontal poses. In this case, PAA still achieves good performance with the highest average F1-frame of 55.8 .

4.3. Ablation Study

In this section, we investigate the effectiveness of each component in our PAA. Table 4 summarizes the structures of different variants of PAA, in which Baseline does not have the structure of learning the attention map M ^ i ( 1 ) and does not utilize the weighting strategy in Equation (5) with w i = 1 / m and v i = 1 . The results of different variants of PAA on the BP4D benchmark are presented in Table 5, which use the same hyperparameters, as detailed in Section 4.1.2.

4.3.1. Weighting Strategy for Suppressing Data Imbalance

We can observe that Baseline+ W ( a u ) performs better with the average F1-frame of 59.9 than Baseline. This demonstrates the effectiveness of the introduced weighting strategy by alleviating both inter-AU data imbalance and intra-AU data imbalance.

4.3.2. Supervision of AU Recognition for Attention Learning

Besides the structure of Baseline+ W ( a u ) , AA further adaptively learns the attention map M ^ i ( 0 ) under the supervision of AU recognition and directly element-wise multiplies M ^ i ( 0 ) with each channel of the fourth convolutional feature map. We can see that AA significantly improves the average F1-frame to 61.5 , which shows that the attention map only under the guidance of AU recognition can already capture much useful AU information.

4.3.3. Attention Predefinition

Another variant over Baseline+ W ( a u ) is PA, which only uses the prior attention by directly element-wise multiplying the predefined mask M i with each channel of the fourth convolutional feature map. Since the guidance of AU recognition is not available, ϵ i = 0 is fixed for M i . We can see that PA achieves good performance with the average F1-frame of 61.9 . This indicates that the prior knowledge is beneficial for AU recognition by specifying the ROI of each AU. We next further explore the effectiveness of combining prior attention and adaptive attention.
After employing the predefined mask M i with the adaptively learned ϵ i over AA, our PAA achieves the highest average F1-frame 63.4 . To investigate the usefulness of learnable ϵ i , we implement PAA ( f i x ) by fixing ϵ i = 0 in Equation (3) for each AU. In this case, the potentially relevant regions beyond the predefined ROI are ignored, in which the average F1-frame is degraded to 62.0 . Therefore, the design of adaptive learning for the control parameter is effective since our PAA can adaptively learn which AU has correlated positions beyond the predefined ROI to be emphasized.

4.4. Visual Results

In Figure 4, we visualize the attention maps learned by recent attention learning-based methods, EAC-Net [2], JÂA-Net [3] and ARL [9], as well as M ^ i ( 1 ) , M ^ i ( 0 ) , and M i for each AU learned by our PAA. Owing to the subtlety and non-rigidity of AUs, different AUs have different appearances, including shapes and sizes, and AU appearances are also varied across persons and expressions. In this case, AUs should have diverse regional attention distributions. For example, the two images in Figure 4 are both in happy expression, while AU 14 co-occurs with AUs 6, 7, 10, 12, and 17 in the first image, and only occurs alone in the second image. It is expected that the same AU of the two images should have different attention maps.
We can observe that different AUs of different images for EAC-Net have the same attention distribution except for different distribution centers, in which the divergences across AUs are ignored and the locations beyond the predefined ROIs are also ignored with zero attention weights. Although JÂA-Net tries to modify the predefined attention maps, it seems that the modification works as the smoothing of the predefinition, in which the positions very distant to the AU centers have smoothed attention weights. Correlated and uncorrelated positions beyond the predefined ROIs are regarded as similar importance, in which the learning of AU features are still often inaccurate. Another solution with a different perspective is ARL, in which the learned attention maps are dense with almost all correlated positions included. However, many potentially uncorrelated positions are mistakenly emphasized.
In contrast with these methods, by combining the advantages of M i as the predefinition and M ^ i ( 0 ) as the supervision for each AU, our PAA can capture both strongly relevant positions specified by landmarks and weakly relevant positions distributed globally in the face. Moreover, we can see that different AUs often have different attention distributions in M i since we employ a learnable instead of fixed control parameter for each AU. In this way, we can more adaptively learn the attention weights at different locations of different AUs, especially for those far away from the predefined ROIs. Furthermore, we find that there are quite a few overlaps among the relevant locations in the attention maps of occurred AUs for our PAA. In this case, our combined attention distribution is clean and lies in the highlighted attention range of the predefined combined attention map, which demonstrates that our method can precisely capture the correlated positions of each AU.

5. Conclusions

In this paper, we have proposed a novel AU recognition method by prior and adaptive attention, which is beneficial for integrating the advantages of the constraint of prior knowledge and the supervision of AU recognition. We have also proposed a learnable parameter to adaptively control the importance of different positions in the predefined mask of each AU. In this case, we can adaptively learn an appropriate constraint of the prior knowledge.
We have compared our approach against state-of-the-art methods on popular challenging benchmarks, which shows that our approach outperforms most of the previous methods. Besides, we have conducted an ablation study, in which each component in our framework is demonstrated to be contributed to AU recognition. Moreover, the visual results indicate that our approach can accurately reason the regional attention distribution of each AU.

Author Contributions

Conceptualization, Z.S. and Y.Z.; methodology, Z.S. and H.Z.; software, Z.S.; validation, W.-L.D. and R.Y.; formal analysis, W.-L.D.; investigation, Z.S.; resources, Z.S.; data curation, H.C.; writing—original draft preparation, Z.S.; writing—review and editing, Z.S., Y.Z. and H.Z.; visualization, Z.S.; supervision, Y.Z.; project administration, Y.Z.; funding acquisition, Y.Z. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62106268), the High-Level Talent Program for Innovation and Entrepreneurship (ShuangChuang Doctor) of Jiangsu Province (No. JSSCBS20211220), and the Talent Program for Deputy General Manager of Science and Technology of Jiangsu Province (No. FZ20220440). It was also partially supported by the National Natural Science Foundation of China (No. 62101555, No. 62002360, and No. 62172417), the Natural Science Foundation of Jiangsu Province (No. BK20201346 and No. BK20210488), the Fundamental Research Funds for the Central Universities (No. 2021QN1072), and the joint project of Guanglian Technology and China University of Mining and Technology.

Data Availability Statement

The experiment uses three public AU datasets, including BP4D, DISFA, and GFT. BP4D: http://www.cs.binghamton.edu/~lijun/Research/3DFE/3DFE_Analysis.html (accessed on 30 August 2022); DISFA: http://mohammadmahoor.com/disfa (accessed on 30 August 2022); GFT: https://osf.io/7wcyz (accessed on 30 August 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, K.; Chu, W.S.; Zhang, H. Deep region and multi-label learning for facial action unit detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3391–3399. [Google Scholar]
  2. Li, W.; Abtahi, F.; Zhu, Z.; Yin, L. EAC-Net: Deep Nets with Enhancing and Cropping for Facial Action Unit Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2583–2596. [Google Scholar] [CrossRef] [PubMed]
  3. Shao, Z.; Liu, Z.; Cai, J.; Ma, L. JÂA-Net: Joint Facial Action Unit Detection and Face Alignment via Adaptive Attention. Int. J. Comput. Vis. 2021, 129, 321–340. [Google Scholar] [CrossRef]
  4. Shao, Z.; Cai, J.; Cham, T.J.; Lu, X.; Ma, L. Unconstrained facial action unit detection via latent feature domain. IEEE Trans. Affect. Comput. 2022, 13, 1111–1126. [Google Scholar] [CrossRef]
  5. Ekman, P.; Friesen, W.V. Facial Action Coding System: A Technique for the Measurement of Facial Movement; Consulting Psychologists Press: Palo Alto, CA, USA, 1978. [Google Scholar]
  6. Ekman, P.; Friesen, W.V.; Hager, J.C. Facial Action Coding System; Manual and Investigator’s Guide; Research Nexus: Salt Lake City, UT, USA, 2002. [Google Scholar]
  7. Li, W.; Abtahi, F.; Zhu, Z. Action unit detection with region adaptation, multi-labeling learning and optimal temporal fusing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6766–6775. [Google Scholar]
  8. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  9. Shao, Z.; Liu, Z.; Cai, J.; Wu, Y.; Ma, L. Facial action unit detection using attention and relation learning. IEEE Trans. Affect. Comput. 2022, 13, 1274–1289. [Google Scholar] [CrossRef]
  10. Shao, Z.; Liu, Z.; Cai, J.; Ma, L. Deep Adaptive Attention for Joint Facial Action Unit Detection and Face Alignment. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 725–740. [Google Scholar]
  11. Benitez-Quiroz, C.F.; Srinivasan, R.; Martinez, A.M. EmotioNet: An Accurate, Real-Time Algorithm for the Automatic Annotation of a Million Facial Expressions in the Wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5562–5570. [Google Scholar]
  12. Zhao, K.; Chu, W.S.; De la Torre, F.; Cohn, J.F.; Zhang, H. Joint patch and multi-label learning for facial action unit and holistic expression recognition. IEEE Trans. Image Process. 2016, 25, 3931–3946. [Google Scholar] [CrossRef] [PubMed]
  13. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–25 September 1999; pp. 1150–1157. [Google Scholar]
  14. Niu, X.; Han, H.; Yang, S.; Huang, Y.; Shan, S. Local Relationship Learning With Person-Specific Shape Regularization for Facial Action Unit Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11917–11926. [Google Scholar]
  15. Ma, C.; Chen, L.; Yong, J. AU R-CNN: Encoding expert prior knowledge into R-CNN for action unit detection. Neurocomputing 2019, 355, 35–47. [Google Scholar] [CrossRef]
  16. Lin, M.; Chen, Q.; Yan, S. Network in network. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  17. Shao, Z.; Zhou, Y.; Tan, X.; Ma, L.; Liu, B.; Yao, R. Survey of Expression Action Unit Recognition Based on Deep Learning. Acta Electron. Sin. 2022, 50, 2003–2017. [Google Scholar]
  18. Zhang, X.; Yin, L.; Cohn, J.F.; Canavan, S.; Reale, M.; Horowitz, A.; Liu, P.; Girard, J.M. Bp4d-spontaneous: A high-resolution spontaneous 3d dynamic facial expression database. Image Vis. Comput. 2014, 32, 692–706. [Google Scholar] [CrossRef]
  19. Mavadati, S.M.; Mahoor, M.H.; Bartlett, K.; Trinh, P.; Cohn, J.F. Disfa: A spontaneous facial action intensity database. IEEE Trans. Affect. Comput. 2013, 4, 151–160. [Google Scholar] [CrossRef]
  20. Girard, J.M.; Chu, W.S.; Jeni, L.A.; Cohn, J.F. Sayette group formation task (gft) spontaneous facial expression database. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Washington, DC, USA, 30 May–3 June 2017; pp. 581–588. [Google Scholar]
  21. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 8024–8035. [Google Scholar]
  22. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1139–1147. [Google Scholar]
  23. Fan, R.E.; Chang, K.W.; Hsieh, C.J.; Wang, X.R.; Lin, C.J. LIBLINEAR: A library for large linear classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
  24. Zhong, L.; Liu, Q.; Yang, P.; Huang, J.; Metaxas, D.N. Learning multiscale active facial patches for expression analysis. IEEE Trans. Cybern. 2015, 45, 1499–1510. [Google Scholar] [CrossRef] [PubMed]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  26. Corneanu, C.A.; Madadi, M.; Escalera, S. Deep Structure Inference Network for Facial Action Unit Recognition. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 309–324. [Google Scholar]
  27. Sankaran, N.; Mohan, D.D.; Setlur, S.; Govindaraju, V.; Fedorishin, D. Representation Learning Through Cross-Modality Supervision. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Lille, France, 14–18 May 2019; pp. 1–8. [Google Scholar]
  28. Li, G.; Zhu, X.; Zeng, Y.; Wang, Q.; Lin, L. Semantic relationships guided representation learning for facial action unit recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 8594–8601. [Google Scholar]
  29. Li, Y.; Zeng, J.; Shan, S.; Chen, X. Self-Supervised Representation Learning From Videos for Facial Action Unit Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10924–10933. [Google Scholar]
  30. Liu, Z.; Dong, J.; Zhang, C.; Wang, L.; Dang, J. Relation modeling with graph convolutional networks for facial action unit detection. In Proceedings of the International Conference on Multimedia Modeling, Daejeon, Korea, 5–8 January 2020; pp. 489–501. [Google Scholar]
  31. Ertugrul, I.O.; Cohn, J.F.; Jeni, L.A.; Zhang, Z.; Yin, L.; Ji, Q. Crossing Domains for AU Coding: Perspectives, Approaches, and Measures. IEEE Trans. Biom. Behav. Identity Sci. 2020, 2, 158–171. [Google Scholar] [CrossRef] [PubMed]
  32. Song, T.; Chen, L.; Zheng, W.; Ji, Q. Uncertain Graph Neural Networks for Facial Action Unit Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 5993–6001. [Google Scholar]
  33. Song, T.; Cui, Z.; Zheng, W.; Ji, Q. Hybrid Message Passing With Performance-Driven Structures for Facial Action Unit Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 6267–6276. [Google Scholar]
  34. Chen, Y.; Song, G.; Shao, Z.; Cai, J.; Cham, T.J.; Zheng, J. GeoConv: Geodesic guided convolution for facial action unit recognition. Pattern Recognit. 2022, 122, 108355. [Google Scholar] [CrossRef]
  35. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. Vggface2: A dataset for recognising faces across pose and age. In Proceedings of the IEEE International Conference on Automatic Face & Gesture Recognition, Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar]
  38. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 187–194. [Google Scholar]
  39. Guo, Y.; Zhang, J.; Cai, J.; Jiang, B.; Zheng, J. Cnn-based real-time dense face reconstruction with inverse-rendered photo-realistic face images. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1294–1307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Example images for different action units (AUs) with overlapped regions. If the regions of multiple AUs have overlaps, the appearance of each AU fuses for a new combined appearance.
Figure 1. Example images for different action units (AUs) with overlapped regions. If the regions of multiple AUs have overlaps, the appearance of each AU fuses for a new combined appearance.
Electronics 11 03047 g001
Figure 2. The overview of our PAA framework. An input image firstly goes through two hierarchical and multi-scale region layers [3,10], each of which is followed by a max-pooling layer. Then, m branches are used to predict AU occurrence probabilities, in which the learned attention map M ^ i ( 0 ) is element-wise multiplied by the predefined mask M i to obtain the new attention map M ^ i ( 1 ) for the i-th AU. We overlay the attention maps as well as the masks on the input image for a better view. “★” refers to element-wise multiplication. The expression c × l × l refers to the layer dimensions are c, l, and l, respectively.
Figure 2. The overview of our PAA framework. An input image firstly goes through two hierarchical and multi-scale region layers [3,10], each of which is followed by a max-pooling layer. Then, m branches are used to predict AU occurrence probabilities, in which the learned attention map M ^ i ( 0 ) is element-wise multiplied by the predefined mask M i to obtain the new attention map M ^ i ( 1 ) for the i-th AU. We overlay the attention maps as well as the masks on the input image for a better view. “★” refers to element-wise multiplication. The expression c × l × l refers to the layer dimensions are c, l, and l, respectively.
Electronics 11 03047 g002
Figure 3. Definition of the central locations of 15 popular AUs for a normalized face with eye centers on the same horizontal level [2,3], in which “scale” denotes the distance between two inner eye corners. Each AU has two symmetric centers specified by two correlated facial landmarks, in which landmarks are in white and AU centers are in other colors.
Figure 3. Definition of the central locations of 15 popular AUs for a normalized face with eye centers on the same horizontal level [2,3], in which “scale” denotes the distance between two inner eye corners. Each AU has two symmetric centers specified by two correlated facial landmarks, in which landmarks are in white and AU centers are in other colors.
Electronics 11 03047 g003
Figure 4. Visualization of learned attention maps of different methods, in which AUs 6, 7, 10, 12, 14, and 17 appear in the first image, and AU 14 appears in the second image. Each row lists the attention maps of 12 AUs, as well as the combined attention map of occurred AUs, for the corresponding method. Attention weights from zero to one are visualized using colors from blue to red, and are overlaid on the input images for a better view.
Figure 4. Visualization of learned attention maps of different methods, in which AUs 6, 7, 10, 12, 14, and 17 appear in the first image, and AU 14 appears in the second image. Each row lists the attention maps of 12 AUs, as well as the combined attention map of occurred AUs, for the corresponding method. Attention weights from zero to one are visualized using colors from blue to red, and are overlaid on the input images for a better view.
Electronics 11 03047 g004
Table 1. F1-frame results on Binghamton-Pittsburgh 4D (BP4D) [18]. The results of LSVM [23] and JPML [12] are from [1], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method performs better than most of the previous works.
Table 1. F1-frame results on Binghamton-Pittsburgh 4D (BP4D) [18]. The results of LSVM [23] and JPML [12] are from [1], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method performs better than most of the previous works.
AU1246710121415172324Avg
LSVM [23]23.222.823.127.247.177.263.764.318.433.019.420.735.3
JPML [12]32.625.637.442.350.572.274.165.738.140.030.442.345.9
DRML [1]36.441.843.055.067.066.365.854.133.248.031.730.048.3
EAC-Net [2]39.035.248.676.172.981.986.258.837.559.135.935.855.9
DSIN [26]51.740.456.076.173.579.985.462.737.362.938.841.658.9
CMS [27]49.144.150.379.274.780.988.363.944.460.341.451.260.6
LP-Net [14]43.438.054.277.176.783.887.263.345.360.548.154.261.0
ARL [9]45.839.855.175.777.282.386.658.847.662.147.455.461.1
SRERL [28]46.945.355.677.178.483.587.660.652.263.947.153.362.9
AU R-CNN [15]50.243.757.078.578.582.687.067.749.162.450.449.363.0
AU-GCN [30]46.838.560.180.179.584.888.067.352.063.240.952.862.8
JÂA-Net [3]53.847.858.278.575.882.788.263.743.361.845.649.962.4
UGN-B [32]54.246.456.876.276.782.486.164.751.263.148.553.663.3
HMP-PS [33]53.146.156.076.576.982.186.464.851.563.049.954.563.4
GeoCNN [34]48.444.259.978.475.683.686.765.053.064.749.554.163.6
PAA50.147.755.074.078.982.287.263.851.462.452.155.863.4
Table 2. F1-frame results on Denver Intensity of Spontaneous Facial Action (DISFA) [19]. The results of LSVM [23] and APL [24] are from [1], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method achieves competitive performance, and achieves a more stable performance among different AUs than most of the previous works.
Table 2. F1-frame results on Denver Intensity of Spontaneous Facial Action (DISFA) [19]. The results of LSVM [23] and APL [24] are from [1], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method achieves competitive performance, and achieves a more stable performance among different AUs than most of the previous works.
AU12469122526Avg
LSVM [23]10.810.021.815.711.570.412.022.121.8
APL [24]11.412.030.112.410.165.921.426.923.8
DRML [1]17.317.737.429.010.737.738.520.126.7
EAC-Net [2]41.526.466.450.78.589.388.915.648.5
DSIN [26]42.439.068.428.646.870.890.442.253.6
CMS [27]40.244.353.257.150.373.581.159.757.4
LP-Net [14]29.924.772.746.849.672.993.865.056.9
ARL [9]43.942.163.641.840.076.295.266.858.7
SRERL [28]45.747.859.647.145.673.584.343.655.9
AU R-CNN [15]32.125.959.855.339.867.777.452.651.3
AU-GCN [30]32.319.555.757.961.462.790.960.055.0
JÂA-Net [3]62.460.767.141.145.173.590.967.463.5
UGN-B [32]43.348.163.449.548.272.990.859.060.0
HMP-PS [33]38.045.965.250.950.876.093.367.661.0
GeoCNN [34]65.565.867.248.651.472.680.944.962.1
PAA56.157.059.039.749.474.695.671.962.9
Table 3. F1-frame results of Sayette Group Formation Task (GFT) [20]. The results of LSVM [23] and AlexNet [25] are from [20], those of EAC-Net [2] and ARL [9] are from [3], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method outperforms all the other approaches.
Table 3. F1-frame results of Sayette Group Formation Task (GFT) [20]. The results of LSVM [23] and AlexNet [25] are from [20], those of EAC-Net [2] and ARL [9] are from [3], and those of other previous methods are reported in their original papers. The best results of each AU, as well as the average across methods, are shown in bold. Our PAA method outperforms all the other approaches.
AU1246101214152324Avg
LSVM [23]3832136764781529494442.9
AlexNet [25]44462737282519434242.8
EAC-Net [2]15.556.60.181.076.184.00.138.557.851.246.1
TCAE [29]43.949.56.371.076.279.510.728.534.541.744.2
ARL [9]51.945.913.779.275.582.80.144.959.247.550.1
Ertugrul et al. [31]43.744.919.874.676.579.850.033.916.812.945.3
JÂA-Net [3]46.549.319.279.075.084.844.133.554.950.753.7
PAA64.645.49.877.974.882.845.453.358.945.055.8
Table 4. The architectures of different variants of our PAA. HM: two hierarchical and multi-scale region layers, each of which is followed by a max-pooling layer. C: five successive convolutional layers in each AU branch. W ( a u ) : weighting strategy in Equation (5). M ^ ( 0 ) : attention map M ^ i ( 0 ) for the i-th AU. M ( f i x ) : predefined mask M i with fixed ϵ i = 0 for the i-th AU. M ( a d a ) : predefined mask M i with adaptively learned ϵ i for the i-th AU. Baseline does not have the structure of obtaining M ^ i ( 0 ) , M i , and M ^ i ( 1 ) , and does not utilize the weighting strategy in Equation (5) with w i = 1 / m and v i = 1 .
Table 4. The architectures of different variants of our PAA. HM: two hierarchical and multi-scale region layers, each of which is followed by a max-pooling layer. C: five successive convolutional layers in each AU branch. W ( a u ) : weighting strategy in Equation (5). M ^ ( 0 ) : attention map M ^ i ( 0 ) for the i-th AU. M ( f i x ) : predefined mask M i with fixed ϵ i = 0 for the i-th AU. M ( a d a ) : predefined mask M i with adaptively learned ϵ i for the i-th AU. Baseline does not have the structure of obtaining M ^ i ( 0 ) , M i , and M ^ i ( 1 ) , and does not utilize the weighting strategy in Equation (5) with w i = 1 / m and v i = 1 .
MethodHMC W ( au ) M ^ ( 0 ) M ( f i x ) M ( a d a ) L u
Baseline
Baseline+ W ( a u )
AA
PA
PAA ( f i x )
PAA
Table 5. F1-frame results for 12 AUs of different variants of PAA on BP4D [18]. The best results of each AU, as well as the average across methods, are shown in bold. The performance is gradually improved after adding the proposed components.
Table 5. F1-frame results for 12 AUs of different variants of PAA on BP4D [18]. The best results of each AU, as well as the average across methods, are shown in bold. The performance is gradually improved after adding the proposed components.
AU1246710121415172324Avg
Baseline47.842.151.472.673.479.585.658.345.359.940.848.858.8
Baseline+ W ( a u ) 49.044.252.973.974.979.384.559.148.061.141.850.459.9
AA50.541.855.075.175.680.486.060.650.061.150.551.361.5
PA48.850.148.975.377.681.685.762.552.561.944.853.361.9
PAA ( f i x ) 46.745.653.975.578.982.086.960.252.061.648.252.562.0
PAA50.147.755.074.078.982.287.263.851.462.452.155.863.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shao, Z.; Zhou, Y.; Zhu, H.; Du, W.-L.; Yao, R.; Chen, H. Facial Action Unit Recognition by Prior and Adaptive Attention. Electronics 2022, 11, 3047. https://doi.org/10.3390/electronics11193047

AMA Style

Shao Z, Zhou Y, Zhu H, Du W-L, Yao R, Chen H. Facial Action Unit Recognition by Prior and Adaptive Attention. Electronics. 2022; 11(19):3047. https://doi.org/10.3390/electronics11193047

Chicago/Turabian Style

Shao, Zhiwen, Yong Zhou, Hancheng Zhu, Wen-Liang Du, Rui Yao, and Hao Chen. 2022. "Facial Action Unit Recognition by Prior and Adaptive Attention" Electronics 11, no. 19: 3047. https://doi.org/10.3390/electronics11193047

APA Style

Shao, Z., Zhou, Y., Zhu, H., Du, W. -L., Yao, R., & Chen, H. (2022). Facial Action Unit Recognition by Prior and Adaptive Attention. Electronics, 11(19), 3047. https://doi.org/10.3390/electronics11193047

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop