Comparing Approaches for Explaining DNN-Based Facial Expression Classifications
Abstract
:1. Introduction
- 1.
- Developing a new method of visually and textually explaining DNN predictions based on geometric features.
- 2.
- Making a direct comparison between the interpretability of a CNN and a DNN trained for an emotion classification task.
- 3.
- Performing a user study to evaluate and compare the quality of the explanations.
2. Background and Related Work
2.1. Background on Explainable AI
2.1.1. SHAP
2.1.2. Grad-CAM
2.2. XAI in Affective Computing and FER
3. Proposed Method
3.1. Geometric Features-Based DNN Modeling
3.1.1. Geometric Feature Extraction
3.1.2. SHAP-Based Explanation Generation
3.2. End-to-End CNN Modeling
3.3. User Survey
4. Experimental Setting
4.1. Dataset
4.2. DNN-Based System Development
4.2.1. Hyperparameter Tuning
4.2.2. Splitting on Pose
4.2.3. Feature Selection
4.3. CNN-Based System Development
4.3.1. Preprocessing
4.3.2. Fine-Tuning
4.4. User Survey Construction
4.4.1. Environment
4.4.2. Questions
- 1.
- The output representations help me understand how the model works. (adapted from [8])
- 2.
- The output representations of how the model works are satisfying. (adapted from [8])
- 3.
- The output representations are sufficiently detailed. (adapted from [8])
- 4.
- The output representations let me know how confident the model is for individual predictions.
- 5.
- The output representations let me know how trustworthy the model is. (adapted from [8])
- 6.
- I found the output representations unnecessarily complex. (adapted from [48])
- 7.
- I think I would need an expert to give me additional explanations. (adapted from [48])
- 8.
- The outputs of the model are very predictable. (adapted from [8])
- 9.
- The model can perform the task better than a novice human. (adapted from [8])
- 10.
- I am confident in the model. I believe it works well. (adapted from [8])
- 1.
- The explanations for model 1 are more understandable than those for model 2.
- 2.
- I trust model 1 more than model 2.
- 3.
- I would prefer the explanations of model 1 over those for model 2.
- 4.
- The explanations for model 1 are more detailed than those for model 2.
- 5.
- The explanations for model 1 are clearer on the model’s accuracy than those for model 2.
- 6.
- The explanations for model 1 reflect the model’s confidence on each prediction better than those of model 2.
- 7.
- Model 1’s explanations are more unnecessarily complex than those of model 2.
- 8.
- The explanations for model 1 were more precise than those for model 2.
- 9.
- I would follow model 1’s advice over that of model 2.
- 10.
- The outputs of model 1 were more predictable than those of model 2.
4.4.3. Hypotheses
4.4.4. Statistical Tests
5. Results
5.1. Experimental Results for Geometric Features-Based DNN Models
5.1.1. Comparative Results Using Pose-Based Models
5.1.2. Feature Selection Results
5.2. Geometric Feature Explanation Results
5.3. Experimental Results for CNN Models
- Rotation range: 50
- Shear range: 0.5
- Zoom range: 0.5
- Horizontal flip
5.4. Comparing Explanations for the DNN and CNN
- 1.
- Left eye aspect ratio (ratio between eye width and eye height).
- 2.
- Angle from bottom mouth to left upper mouth.
- 3.
- Angle from left mouth corner to top of the mouth.
- 4.
- Distance between the centre of the left eye and the left inner eyebrow.
- 5.
- Left lower eye outer angle.
5.5. User Study Results
5.6. Discussion
6. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
CNN | Convolutional Neural Network |
DNN | Deep Neural Network |
FER | Facial Emotion Recognition |
FR | Frontal |
FS | Feature Selection |
FSFS | Forward Sequential Feature Selection |
Grad-CAM | Gradient-weighted Class Activation Mapping |
HL | Half Left |
HR | Half Right |
KDEF | Karolinska Directed Emotional Faces |
LIME | Local Interpretable Model-Agnostic Explanations |
LRP | Layer-wise Relevance Propagation |
ReLU | Rectified Linear Unit |
RFE | Recursive Feature Elimination |
SCS | System Causability Scale |
SGD | Stochastic Gradient Descent |
SHAP | SHapley Additive exPlanation |
SUS | System Usability Scale |
XAI | eXplainable Artificial Intelligence |
Appendix A. Final DNN Configurations
Hyperparameter | Frontal | HL | HR |
---|---|---|---|
Number of hidden layers | 2 | 1 | 1 |
Learning rate | 0.001 | 0.001 | 0.001 |
No. neurons hidden layer 1 | 352 | 64 | 352 |
Regularisation rate hidden layer 1 | 0.01 | 0.01 | 0.001 |
Dropout hidden layer 1 | 0.3 | 0 | 0.6 |
No. neurons hidden layer 2 | 256 | - | - |
Regularisation rate hidden layer 2 | 0.01 | - | - |
Dropout hidden layer 2 | 0.8 | - | - |
Appendix B. User Study
Appendix B.1. Research Description
Appendix B.2. Consent Form
- I voluntarily agree to participate in the research project.
- I agree that I will not be paid for my participation.
- I have been informed of the nature of the research project.
- I understand that statistical data gathered from this survey can be used in a scientific publication.
- I understand that my participation will remain anonymous.
- I agree that my data can be shared with other researchers to answer possible other research questions.
Appendix B.3. General Questions
- No degree
- Elementary school
- High school
- MBO
- HBO
- Bachelor’s degree
- Master’s degree
- Doctorate degree
- 1—Strongly disagree
- 2—Disagree
- 3—Neutral
- 4—Agree
- 5—Strongly agree
Appendix C. All Geometric Features
Feature # | Description | Landmarks | Feature Type |
---|---|---|---|
1 | Eye aspect ratio (L) | [19, 24] | Distance |
2 | Eye aspect ratio (R) | [25, 30] | Distance |
3 | Mouth aspect ratio | [31, 34, 37, 40] | Distance |
4 | Upper lip angle (L) | [31, 34] | Angle |
5 | Upper lip angle (R) | [34, 37] | Angle |
6 | Nose tip—mouth corner angle (L) | [16, 31] | Angle |
7 | Nose tip—mouth corner angle (R) | [16, 37] | Angle |
8 | Lower lip angle (L) | [31, 41] | Angle |
9 | Lower lip angle (R) | [37, 39] | Angle |
10 | Eyebrow slope (L) | [0, 4] | Angle |
11 | Eyebrow slope (R) | [5, 9] | Angle |
12 | Lower eye outer angles (L) | [19, 24] | Angle |
13 | Lower eye inner angles (L) | [22, 23] | Angle |
14 | Lower eye outer angles (R) | [28, 29] | Angle |
15 | Lower eye inner angles (R) | [25, 30] | Angle |
16 | Mouthe corner—mouth bottom angle (L) | [31, 40] | Angle |
17 | Mouth corner—mouth bottom angle (R) | [37, 40] | Angle |
18 | Upper mouth angles (L) | [33, 40] | Angle |
19 | Upper mouth angles (R) | [35, 40] | Angle |
20 | Curvature of lower-outer lips (L) | [31, 41, 42] | Curvature |
21 | Curvature of lower-outer lips (R) | [37, 38, 39] | Curvature |
22 | Curvature of lower-inner lips (L) | [31, 40, 41] | Curvature |
23 | Curvature of lower-inner lips (R) | [37, 39, 40] | Curvature |
24 | Bottom lip curvature | [31, 37, 40] | Curvature |
25 | Mouth opening/mouth width | [43–48] | Distance |
26 | Mouth up/down | [34, 40, 44] | Distance |
27 | Eye—middle eyebrow distance (L) | [0, 4, 19, 22] | Distance |
28 | Eye—middle eyebrow distance (R) | [5, 9, 25, 28] | Distance |
29 | Eye—inner eyebrow distance (L) | [4, 19, 22] | Distance |
30 | Eye—inner eyebrow distance (R) | [5, 25, 28] | Distance |
31 | Inner eye—eyebrow centre (L) | [2, 22] | Distance |
32 | Inner eye—eyebrow centre (R) | [7, 25] | Distance |
33 | Inner eye—mouth top distance (L) | [22, 34] | Distance |
34 | Inner eye—mouth top distance (R) | [25, 34] | Distance |
35 | Mouth width | [31, 37] | Distance |
36 | Mouth height | [34, 40] | Distance |
37 | Upper mouth height | [34, 44, 47] | Distance |
38 | Lower mouth height | [40, 44, 47] | Distance |
39 | Outer mid eyebrow slope (L) * | [0, 2] | Slope |
40 | Outer mid eyebrow slope (R) * | [7, 9] | Slope |
References
- Tao, J.; Tan, T. Affective computing: A review. In Proceedings of the International Conference on Affective computing and Intelligent Interaction, Beijing, China, 22–24 October 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 981–995. [Google Scholar]
- Ko, B.C. A brief review of facial emotion recognition based on visual information. Sensors 2018, 18, 401. [Google Scholar] [CrossRef] [PubMed]
- Wang, W.; Yang, Y.; Wang, X.; Wang, W.; Li, J. Development of convolutional neural network and its application in image classification: A survey. Opt. Eng. 2019, 58, 040901. [Google Scholar] [CrossRef] [Green Version]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef] [Green Version]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Letham, B.; Rudin, C.; McCormick, T.H.; Madigan, D. Interpretable classifiers using rules and bayesian analysis: Building a better stroke prediction model. Ann. Appl. Stat. 2015, 9, 1350–1371. [Google Scholar] [CrossRef]
- Weitz, K.; Schiller, D.; Schlagowski, R.; Huber, T.; André, E. ‘Do you trust me?’ Increasing user-trust by integrating virtual agents in explainable AI interaction design. In Proceedings of the 19th ACM International Conference on Intelligent Virtual Agents, Paris, France, 2–5 July 2019; pp. 7–9. [Google Scholar] [CrossRef] [Green Version]
- Hoffman, R.R.; Mueller, S.T.; Klein, G.; Litman, J. Metrics for explainable AI: Challenges and prospects. arXiv 2018, arXiv:1812.04608. [Google Scholar]
- Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
- Weitz, K.; Hassan, T.; Schmid, U.; Garbas, J. Towards explaining deep learning networks to distinguish facial expressions of pain and emotions. In Forum Bildverarbeitung; Institut für Industrielle Informationstechnik (IIIT): Karlsruhe, Germany, 2018; pp. 197–208. [Google Scholar]
- Gund, M.; Bharadwaj, A.R.; Nwogu, I. Interpretable Emotion Classification Using Temporal Convolutional Models. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 6367–6374. [Google Scholar]
- Escalante, H.J.; Guyon, I.; Escalera, S.; Jacques, J.; Madadi, M.; Baró, X.; Ayache, S.; Viegas, E.; Güçlütürk, Y.; Güçlü, U.; et al. Design of an explainable machine learning challenge for video interviews. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3688–3695. [Google Scholar]
- Escalante, H.J.; Kaya, H.; Salah, A.A.; Escalera, S.; Güçlütürk, Y.; Güçlü, U.; Baró, X.; Guyon, I.; Jacques, J.C.; Madadi, M.; et al. Modeling, Recognizing, and Explaining Apparent Personality from Videos. IEEE Trans. Affect. Comput. 2022, 13, 894–911. [Google Scholar] [CrossRef]
- Kaya, H.; Gürpinar, F.; Afshar, S.; Salah, A.A. Contrasting and combining least squares based learners for emotion recognition in the wild. In Proceedings of the 2015 ACM International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 459–466. [Google Scholar]
- Dresvyanskiy, D.; Ryumina, E.; Kaya, H.; Markitantov, M.; Karpov, A.; Minker, W. End-to-End Modeling and Transfer Learning for Audiovisual Emotion Recognition in-the-Wild. Multimodal Technol. Interact. 2022, 6, 11. [Google Scholar] [CrossRef]
- Selbst, A.D.; Powles, J. Meaningful information and the right to explanation. Int. Data Priv. Law 2017, 7, 233–242. Available online: http://xxx.lanl.gov/abs/https://academic.oup.com/idpl/articlepdf/7/4/233/22923065/ipx022.pdf (accessed on 1 May 2022). [CrossRef]
- Doshi-Velez, F.; Kim, B. Towards a rigorous science of interpretable machine learning. arXiv 2017, arXiv:1702.08608. [Google Scholar]
- Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef] [Green Version]
- Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 2018, 51, 1–42. [Google Scholar] [CrossRef] [Green Version]
- Jacovi, A.; Goldberg, Y. Towards Faithfully Interpretable NLP Systems: How Should We Define and Evaluate Faithfulness? In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL); Association for Computational Linguistics: Online, 2020; pp. 4198–4205. [Google Scholar] [CrossRef]
- Nori, H.; Jenkins, S.; Koch, P.; Caruana, R. Interpretml: A unified framework for machine learning interpretability. arXiv 2019, arXiv:1909.09223. [Google Scholar]
- Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA; Volume 30, pp. 4768–4777. [Google Scholar]
- Shapley, L.S. A value for n-person games. In Class. Game Theory; Kuhn, H.W., Ed.; Princeton University Press: Princeton, NJ, USA, 1997; pp. 69–79. [Google Scholar]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
- Bach, S.; Binder, A.; Montavon, G.; Klauschen, F.; Müller, K.R.; Samek, W. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE 2015, 10, 1–46. [Google Scholar] [CrossRef] [Green Version]
- Ventura, C.; Masip, D.; Lapedriza, A. Interpreting CNN models for apparent personality trait regression. In Proceedings of the CVPR Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 55–63. [Google Scholar]
- Bobek, S.; Tragarz, M.M.; Szelażek, M.; Nalepa, G.J. Explaining Machine Learning Models of Emotion Using the BIRAFFE Dataset. In Artificial Intelligence and Soft Computing; Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 290–300. [Google Scholar]
- Liew, W.S.; Loo, C.K.; Wermter, S. Emotion Recognition Using Explainable Genetically Optimized Fuzzy ART Ensembles. IEEE Access 2021, 9, 61513–61531. [Google Scholar] [CrossRef]
- Prajod, P.; Schiller, D.; Huber, T.; André, E. Do Deep Neural Networks Forget Facial Action Units?–Exploring the Effects of Transfer Learning in Health Related Facial Expression Recognition. arXiv 2021, arXiv:2104.07389. [Google Scholar]
- Xiong, X.; De la Torre, F. Supervised Descent Method and Its Application to Face Alignment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 532–539. [Google Scholar]
- Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. VGGFace2: A Dataset for Recognising Faces across Pose and Age. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar] [CrossRef] [Green Version]
- Kollias, D.; Tzirakis, P.; Nicolaou, M.A.; Papaioannou, A.; Zhao, G.; Schuller, B.; Kotsia, I.; Zafeiriou, S. Deep affect prediction in-the-wild: Aff-wild database and challenge, deep architectures, and beyond. Int. J. Comput. Vis. 2019, 127, 907–929. [Google Scholar] [CrossRef] [Green Version]
- Korobov, M.; Lopuhin, K. ELI5. 2016. Available online: https://eli5.readthedocs.io/en/latest/ (accessed on 1 May 2022).
- Likert, R. A technique for the measurement of attitudes. Arch. Psychol. 1932, 140, 55. [Google Scholar]
- Lundqvist, D.; Flykt, A.; Öhman, A. The Karolinska Directed Emotional Faces-KDEF; CD ROM from Department of Clinical Neuroscience, Psychology section; Karolinska Institutet: Solna, Sweden, 1998; ISBN 91-630-7164-9. [Google Scholar]
- Ekman, P. Basic emotions. Handb. Cogn. Emot. 1999, 98, 16. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference for Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 1 May 2022).
- Li, L.; Jamieson, K.; DeSalvo, G.; Rostamizadeh, A.; Talwalkar, A. Hyperband: A novel bandit-based approach to hyperparameter optimization. J. Mach. Learn. Res. 2017, 18, 6765–6816. [Google Scholar]
- Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
- Ng, A.Y. Feature selection, L 1 vs. L 2 regularization, and rotational invariance. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 78. [Google Scholar]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene selection for cancer classification using support vector machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
- Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
- Deng, J.; Guo, J.; Ververas, E.; Kotsia, I.; Zafeiriou, S. RetinaFace: Single-Shot Multi-Level Face Localisation in the Wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5203–5212. [Google Scholar]
- Zheng, E. Batch Face. 2020. Available online: https://github.com/elliottzheng/batch-face (accessed on 1 May 2022).
- Brooke, J. SUS: A quick and dirty usability scale. Usability Eval. Ind. 1995, 189, 6. [Google Scholar]
- Holzinger, A.; Carrington, A.; Müller, H. Measuring the quality of explanations: The system causability scale (SCS). In KI-Künstliche Intelligenz; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1–6. [Google Scholar]
- Jamieson, S. Likert scales: How to (ab)use them? Med Educ. 2004, 38, 1217–1218. [Google Scholar] [CrossRef]
- Norman, G. Likert scales, levels of measurement and the “laws” of statistics. Adv. Health Sci. Educ. 2010, 15, 625–632. [Google Scholar] [CrossRef]
- Wilcoxon, F. Individual Comparisons by Ranking Methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
- Mann, H.B.; Whitney, D.R. On a Test of Whether one of Two Random Variables is Stochastically Larger than the Other. Ann. Math. Stat. 1947, 18, 50–60. [Google Scholar] [CrossRef]
- Burg, K.t. Explaining DNN Based Facial Expression Classifications. BSc Thesis, Utrecht University, Utrecht, The Netherlands, 2021. [Google Scholar]
- Davis, B.; Glenski, M.; Sealy, W.; Arendt, D. Measure Utility, Gain Trust: Practical Advice for XAI Researchers. In Proceedings of the 2020 IEEE Workshop on TRust and EXpertise in Visual Analytics (TREX), Salt Lake City, Utah, USA, 25 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
- Toisoul, A.; Kossaifi, J.; Bulat, A.; Tzimiropoulos, G.; Pantic, M. Estimation of continuous valence and arousal levels from faces in naturalistic conditions. Nat. Mach. Intell. 2021, 3, 42–50. [Google Scholar] [CrossRef]
- Mahmud, F.; Islam, B.; Hossain, A.; Goala, P.B. Facial region segmentation based emotion recognition using K-nearest neighbors. In Proceedings of the International Conference on Innovation in Engineering and Technology (ICIET), Dhaka, Bangladesh, 27–28 December 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–5. [Google Scholar]
- Kandeel, A.A.; Abbas, H.M.; Hassanein, H.S. Explainable Model Selection of a Convolutional Neural Network for Driver’s Facial Emotion Identification. In ICPR Workshops and Challenges; Del Bimbo, A., Cucchiara, R., Sclaroff, S., Farinella, G.M., Mei, T., Bertini, M., Escalante, H.J., Vezzani, R., Eds.; Springer: Cham, Switzerland, 2021; pp. 699–713. [Google Scholar]
- Puthanidam, R.V.; Moh, T.S. A Hybrid approach for facial expression recognition. In Proceedings of the 12th International Conference on Ubiquitous Information Management and Communication, Langkawi, Malaysia, 5–7 January 2018; pp. 1–8. [Google Scholar]
- Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 1, p. I. [Google Scholar]
- Yu, Z.; Zhang, C. Image based static facial expression recognition with multiple deep network learning. In Proceedings of the 2015 ACM International Conference on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 435–442. [Google Scholar]
Model | Accuracy |
---|---|
Frontal | 0.780 |
Half Left | 0.790 |
Half Right | 0.828 |
Overall pose-based | 0.794 |
No pose split | 0.756 |
Frontal | Half Left | Half Right | ||||
---|---|---|---|---|---|---|
Algorithm | # Feats. | Accuracy | # Feats. | Accuracy | # Feats. | Accuracy |
FSFS | 25 | 0.7857 | 25 | 0.8070 | 30 | 0.8448 |
RFE | 30 | 0.7976 | 25 | 0.8070 | 20 | 0.8276 |
SHAP | 35 | 0.7857 | 35 | 0.7895 | 35 | 0.8362 |
Handpicked | - | - | 24 | 0.8070 | 24 | 0.8017 |
No FS | 40 | 0.7798 | 40 | 0.7895 | 40 | 0.8276 |
Question # | W-Value | p-Value | |
---|---|---|---|
1 | greater | 0.0 | 0.001 * |
2 | greater | 6.5 | 0.100 |
3 | greater | 4.5 | 0.015 * |
4 | less | 39.5 | 0.021 * |
5 | greater | 16.5 | 0.415 |
6 | greater | 0.0 | 0.118 |
7 | less | 23.0 | 0.061 |
8 | greater | 14.5 | 0.198 |
9 | greater | 2.0 | 0.718 |
10 | greater | 9.0 | 0.327 |
Question # | U-Value | p-Value | |
---|---|---|---|
1 | less | 42.0 | 0.031 * |
2 | less | 66.0 | 0.364 |
3 | less | 42.0 | 0.031 * |
4 | less | 18.0 | 0.0003 * |
5 | unequal | 48.0 | 0.105 |
6 | unequal | 54.0 | 0.154 |
7 | less | 36.0 | 0.011 * |
8 | less | 24.0 | 0.001 * |
9 | less | 66.0 | 0.364 |
10 | less | 9.0 | 0.327 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
ter Burg, K.; Kaya, H. Comparing Approaches for Explaining DNN-Based Facial Expression Classifications. Algorithms 2022, 15, 367. https://doi.org/10.3390/a15100367
ter Burg K, Kaya H. Comparing Approaches for Explaining DNN-Based Facial Expression Classifications. Algorithms. 2022; 15(10):367. https://doi.org/10.3390/a15100367
Chicago/Turabian Styleter Burg, Kaya, and Heysem Kaya. 2022. "Comparing Approaches for Explaining DNN-Based Facial Expression Classifications" Algorithms 15, no. 10: 367. https://doi.org/10.3390/a15100367
APA Styleter Burg, K., & Kaya, H. (2022). Comparing Approaches for Explaining DNN-Based Facial Expression Classifications. Algorithms, 15(10), 367. https://doi.org/10.3390/a15100367