Computer Vision and Human Behaviour, Emotion and Cognition Detection: A Use Case on Student Engagement
Abstract
:1. Introduction
1.1. Background and Rationale behind the Study
1.2. Possibilities of Computer Vision Techniques to Detect Human Behaviour, Cognition or Emotion: State of the Art
1.3. Engagement and Its Relevance for Learning
1.4. Measuring Engagement
2. Research Gaps and Aim of This Study
- (1)
- At the individual (student) level:
- (1a)
- Describing a methodology to apply computer vision to detect student-level indicators for engagement (hand-raising, note-taking, etc.) and evaluating how well computer vision can detect these, in terms of precision and recall;
- (1b)
- evaluating how well these indicators can measure self-reported engagement.
- (2)
- At the collective (classroom) level:
- (2a)
- Describing a methodology to apply computer vision to detect classroom-level indicators for engagement (a measure for synchronicity, students’ reaction times and shared eye-gaze intersections);
- (2b)
- evaluating how well these indicators can measure self-reported engagement.
3. Methodology
3.1. Participants, Setting, Procedure, Self-Reporting and Annotations
3.2. High-Level Overview of Data-Processing Pipeline
3.3. Estimating Body Poses of Students in the Classroom
3.4. Recognising Individual Behaviour from Students’ Body Poses
3.5. Quantifying Collective Behaviour from Students’ Body Poses
3.5.1. Estimating Intermediate Individual State Representations
3.5.2. Measuring Collective States
3.5.3. Collective State Transition: A Measure to Detect Synchrony and Educational Events
- k is the total number of clusters, obtained by applying unsupervised clustering to students’ joints and body parts;
- Δt is the time interval between two consecutive calculations (e.g., 0.5 s for a stride of 2 frames per second);
- is the proportion of students that are assigned to cluster si at time t;
- is the proportion of students that are assigned to cluster si at time t + Δt.
3.5.4. From Collective Behaviour to Individual Behaviour: The Variance in Students’ Reaction Time to Classroom Events
- M is the total number of students.
- is the time for student m to change cluster upon educational event e.
- is the average time across students to change cluster upon educational event e.
3.6. Assessing the Number of Eye-Gaze Intersections That Students Share
4. Results
4.1. Recognising Individual Behaviour
4.2. Measuring Engagement through Students’ Individual Behaviour
4.3. Recognising Collective Behaviour
4.3.1. Unsupervised Clustering
4.3.2. Collective State Transition: A Measure to Detect Synchrony and Educational Events
4.4. Measuring Engagement through Students’ Collective Behaviour
4.4.1. CST
4.4.2. The Variance in Students’ Reaction Time to Educational Events
4.4.3. Shared Eye-Gaze Intersections
5. Discussion
5.1. Discussion of This Study’s Main Findings on Individual Behaviour
5.2. Discussion of This Study’s Main Findings on Collective Behaviour
5.3. Limitations of the Study
5.4. Assets of the Study
5.5. Implications for Practice and Future Research
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Zhang, H.; Zhang, Y.-X.; Zhong, B.; Lei, Q.; Yang, L.; Du, J.-X.; Chen, D.-S. A Comprehensive Survey of Vision-Based Human Action Recognition Methods. Sensors 2019, 19, 1005. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Muthalagu, R.; Bolimera, A.; Kalaichelvi, V. Lane detection technique based on perspective transformation and histogram analysis for self-driving cars. Comput. Electr. Eng. 2020, 85, 106653. [Google Scholar] [CrossRef]
- Peter, C.; Beale, R. Affect and Emotion in Human-Computer Interaction: From Theory to Applications, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
- Fredricks, A.J.; Blumenfeld, P.C.; Paris, A.H. School Engagement: Potential of the Concept, State of the Evidence. Rev. Educ. Res. 2004, 74, 59–109. [Google Scholar] [CrossRef] [Green Version]
- Raes, A.; Vanneste, P.; Pieters, M.; Windey, I.; Noortgate, W.V.D.; Depaepe, F. Learning and instruction in the hybrid virtual classroom: An investigation of students’ engagement and the effect of quizzes. Comput. Educ. 2020, 143, 103682. [Google Scholar] [CrossRef]
- Cao, Z.; Hidalgo Martinez, G.; Simon, T.; Wei, S.-E.; Sheikh, Y.A. OpenPose: Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 172–186. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ekman, R. What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS), 2nd ed.; Series in affective science; Oxford University Press: New York, NY, USA, 1997. [Google Scholar]
- Tarnowski, P.; Kołodziej, M.; Majkowski, A.; Rak, R.J. Emotion recognition using facial expressions. Procedia Comput. Sci. 2017, 108, 1175–1184. [Google Scholar] [CrossRef]
- El Kaliouby, R.; Robinson, P. Real-Time Inference of Complex Mental States from Facial Expressions and Head Gestures. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; Volume 2. [Google Scholar] [CrossRef]
- Van Acker, B.B.; Bombeke, K.; Durnez, W.; Parmentier, D.D.; Mateus, J.C.; Biondi, A.; Saldien, J.; Vlerick, P. Mobile pupillometry in manual assembly: A pilot study exploring the wearability and external validity of a renowned mental workload lab measure. Int. J. Ind. Ergon. 2020, 75, 102891. [Google Scholar] [CrossRef]
- Dobbins, C.; Denton, P. MyWallMate: An Investigation into the use of Mobile Technology in Enhancing Student Engagement. TechTrends 2017, 11, 142–549. [Google Scholar] [CrossRef] [Green Version]
- Gobert, J.D.; Baker, R.S.; Wixon, M.B. Operationalizing and Detecting Disengagement within Online Science Microworlds. Educ. Psychol. 2015, 50, 43–57. [Google Scholar] [CrossRef]
- Connell, J.P.; Spencer, M.B.; Aber, J.L. Educational Risk and Resilience in African-American Youth: Context, Self, Action, and Outcomes in School. Child Dev. 1994, 65, 493. [Google Scholar] [CrossRef]
- Marks, H.M. Student Engagement in Instructional Activity: Patterns in the Elementary, Middle, and High School Years. Am. Educ. Res. J. 2000, 37, 153–184. [Google Scholar] [CrossRef]
- Furlong, M.J.; Christenson, S.L. Engaging students at school and with learning: A relevant construct forall students. Psychol. Sch. 2008, 45, 365–368. [Google Scholar] [CrossRef]
- Boyle, F.; Kwon, J.; Ross, C.; Simpson, O. Student–student mentoring for retention and engagement in distance education. Open Learn. J. Open Distance e-Learn. 2010, 25, 115–130. [Google Scholar] [CrossRef]
- Mazer, J.P. Associations among Teacher Communication Behaviors, Student Interest, and Engagement: A Validity Test. Commun. Educ. 2013, 62, 86–96. [Google Scholar] [CrossRef]
- Christophel, D.M. The relationships among teacher immediacy behaviors, student motivation, and learning. Commun. Educ. 1990, 39, 323–340. [Google Scholar] [CrossRef]
- Raes, A.; Detienne, L.; Windey, I.; Depaepe, F. A systematic literature review on synchronous hybrid learning: Gaps identified. Learn. Environ. Res. 2019, 23, 269–290. [Google Scholar] [CrossRef]
- Lee, O.; Anderson, C.W. Task engagement and conceptual change in middle school science classrooms. Am. Educ. Res. J. 1993, 30, 585–610. [Google Scholar] [CrossRef]
- Whitehill, J.; Serpell, Z.; Lin, Y.-C.; Foster, A.; Movellan, J.R. The Faces of Engagement: Automatic Recognition of Student Engagement from Facial Expressions. IEEE Trans. Affect. Comput. 2014, 5, 86–98. [Google Scholar] [CrossRef]
- Spanjers, D.M.; Burns, M.K.; Wagner, A.R. Systematic Direct Observation of Time on Task as a Measure of Student Engagement. Assess. Eff. Interv. 2008, 33, 120–126. [Google Scholar] [CrossRef]
- Revere, L.; Kovach, J.V. Online technologies for engaged learning, a meaningful synthesis for educators. Q. Rev. Distance Educ. 2011, 12, 113–124. [Google Scholar]
- Pirsoul, T.; Parmentier, M.; Nils, F. The rocky road to emotion measurement in learning and career development: On the use of self-reports. In Proceedings of the 18th Biennial EARLI Conference for Research on Learning and Instruction, Aachen, Germany, 12–16 August 2019; Available online: https://hdl.handle.net/2078.1/218801 (accessed on 23 November 2020).
- Brown, G.T.; Andrade, H.L.; Chen, F. Accuracy in student self-assessment: Directions and cautions for research. Assess. Educ. Princ. Policy Pr. 2015, 22, 444–457. [Google Scholar] [CrossRef]
- Skinner, E.A.; Kindermann, T.A.; Furrer, C.J. A Motivational Perspective on Engagement and Disaffection. Conceptualization and assessment of children’s behavioral and emotional participation in academic activities in the classroom. Educ. Psychol. Meas. 2008, 69, 493–525. [Google Scholar] [CrossRef] [Green Version]
- Liao, W.; Xu, W.; Kong, S.; Ahmad, F.; Liu, W. A Two-stage Method for Hand-Raising Gesture Recognition in Classroom. In Proceedings of the 2019 8th International Conference on Educational and Information Technology (ICEIT 2019), Cambridge, UK, 2–4 March 2019; pp. 38–44. [Google Scholar] [CrossRef]
- Lin, J.; Jiang, F.; Shen, R. Hand-Raising Gesture Detection in Real Classroom. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6453–6457. [Google Scholar]
- Böheim, R.; Knogler, M.; Kosel, C.; Seidel, T. Exploring student hand-raising across two school subjects using mixed methods: An investigation of an everyday classroom behavior from a motivational perspective. Learn. Instr. 2020, 65, 101250. [Google Scholar] [CrossRef]
- Chen, Y.-L.E.; Kraklow, D. Taiwanese College Students’ Motivation and Engagement for English Learning in the Context of Internationalization at Home. J. Stud. Int. Educ. 2014, 19, 46–64. [Google Scholar] [CrossRef]
- Barbadekar, A.; Gaikwad, V.; Patil, S.; Chaudhari, T.; Deshpande, S.; Burad, S.; Godbole, R. Engagement Index for Classroom Lecture using Computer Vision. In Proceedings of the 2019 Global Conference for Advancement in Technology (GCAT), Bangalore, India, 18–20 October 2019; pp. 1–5. [Google Scholar]
- Canedo, D.; Trifan, A.; Neves, A.J.R. Monitoring Students’ Attention in a Classroom through Computer Vision. In Highlights of Practical Applications of Agents, Multi-Agent Systems, and Complexity: The PAAMS Collection. PAAMS 2018. Communications in Computer and Information Science; Springer: Cham, Switzerland, 2018; Volume 887, pp. 371–378. [Google Scholar]
- Li, W.; Jiang, F.; Shen, R. Sleep Gesture Detection in Classroom Monitor System. In Proceedings of the (ICASSP 2019) 2019 IEEE International Conference on Acoustics, Speech and Signal Processing, Brighton, UK, 12–17 May 2019; pp. 7640–7644. [Google Scholar]
- Nezami, O.M.; Dras, M.; Hamey, L.; Richards, D.; Wan, S.; Paris, C. Automatic Recognition of Student Engagement Using Deep Learning and Facial Expression. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Dublin, Ireland, 10–14 September 2018; pp. 273–289. [Google Scholar]
- MacHardy, Z.; Syharath, K.; Dewan, P. Engagement Analysis through Computer Vision. In Proceedings of the 7th International Conference on Collaborative Computing: Networking, Applications and Worksharing (CollaborateCom 2011), Orlando, FL, USA, 15–18 October 2012; pp. 535–539. [Google Scholar]
- Bosch, N.; D’Mello, S.; Ocumpaugh, J.; Baker, R.S.; Shute, V. Using Video to Automatically Detect Learner Affect in Computer-Enabled Classrooms. ACM Trans. Interact. Intell. Syst. 2016, 6, 1–26. [Google Scholar] [CrossRef]
- Raca, M.; Tormey, R.; Dillenbourg, P. Sleepers’ lag—study on motion and attention. In Proceedings of the 4th International Conference on Learning Analytics and Knowledge (LAK ’14), Indianapolis, IN, USA, 24–28 March 2014; pp. 36–43. [Google Scholar]
- Matthews, G.; De Winter, J.; Hancock, P.A. What do subjective workload scales really measure? Operational and representational solutions to divergence of workload measures. Theor. Issues Ergon. Sci. 2020, 21, 369–396. [Google Scholar] [CrossRef] [Green Version]
- Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6299–6308. [Google Scholar]
- Wang, Z.; Jiang, F.; Shen, R. An Effective Yawn Behavior Detection Method in Classroom. In Mining Data for Financial Applications; Springer: Cham, Switzerland, 2019; Volume 11953, pp. 430–441. [Google Scholar]
- Macey, W.H.; Schneider, B. The Meaning of Employee Engagement. Ind. Organ. Psychol. 2008, 1, 3–30. [Google Scholar] [CrossRef]
- Organisation for Economic Co-Operation and Development. Student Engagement at School: A Sense of Belonging and Participation; Organisation for Economic Co-Operation and Development (OECD): Paris, France, 2003. [Google Scholar]
- Finn, J.D.; Rock, D.A. Academic success among students at risk for school failure. J. Appl. Psychol. 1997, 82, 221–234. [Google Scholar] [CrossRef] [Green Version]
- Gendron, M.; Roberson, D.; Van Der Vyver, J.M.; Barrett, L.F. Perceptions of emotion from facial expressions are not culturally universal: Evidence from a remote culture. Emotion 2014, 14, 251–262. [Google Scholar] [CrossRef]
- Golann, J.W.; Darling-Aduana, J. Toward a multifaceted understanding of Lareau’s “sense of entitlement”: Bridging sociological and psychological constructs. Sociol. Compass 2020, 14, e12798. [Google Scholar] [CrossRef]
- Derry, S.J.; Pea, R.D.; Barron, B.; Engle, R.A.; Erickson, F.; Goldman, R.; Hall, R.; Koschmann, T.; Lemke, J.L.; Sherin, M.G.; et al. Conducting Video Research in the Learning Sciences: Guidance on Selection, Analysis, Technology, and Ethics. J. Learn. Sci. 2010, 19, 3–53. [Google Scholar] [CrossRef]
- Luckin, R. AI is coming: Use it or lose to it. Times Educ. Suppl. 2018, 5306. [Google Scholar]
Annotated Samples | Precision | Recall | |
---|---|---|---|
Raising hand | 85 | 0.67 | 0.59 |
Taking notes | 315 | 0.69 | 0.63 |
Hand on face | 185 | 0.35 | 0.32 |
Working with laptop | 76 | 0.26 | 0.31 |
Looking back | 84 | 0.50 | 0.53 |
Fiddling with hair | 45 | 0.17 | 0.11 |
Playing with cellphone | 37 | 0.22 | 0.38 |
Crossing arms | 48 | 0.27 | 0.50 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Vanneste, P.; Oramas, J.; Verelst, T.; Tuytelaars, T.; Raes, A.; Depaepe, F.; Van den Noortgate, W. Computer Vision and Human Behaviour, Emotion and Cognition Detection: A Use Case on Student Engagement. Mathematics 2021, 9, 287. https://doi.org/10.3390/math9030287
Vanneste P, Oramas J, Verelst T, Tuytelaars T, Raes A, Depaepe F, Van den Noortgate W. Computer Vision and Human Behaviour, Emotion and Cognition Detection: A Use Case on Student Engagement. Mathematics. 2021; 9(3):287. https://doi.org/10.3390/math9030287
Chicago/Turabian StyleVanneste, Pieter, José Oramas, Thomas Verelst, Tinne Tuytelaars, Annelies Raes, Fien Depaepe, and Wim Van den Noortgate. 2021. "Computer Vision and Human Behaviour, Emotion and Cognition Detection: A Use Case on Student Engagement" Mathematics 9, no. 3: 287. https://doi.org/10.3390/math9030287
APA StyleVanneste, P., Oramas, J., Verelst, T., Tuytelaars, T., Raes, A., Depaepe, F., & Van den Noortgate, W. (2021). Computer Vision and Human Behaviour, Emotion and Cognition Detection: A Use Case on Student Engagement. Mathematics, 9(3), 287. https://doi.org/10.3390/math9030287