A Real-time Engagement Assessment for Learner in Asynchronous Distance Learning

Authors

  • Shofiyati Nur Karimah Japan Advanced Institute of Science and Technology (JAIST)
  • Shinobu Hasegawa

DOI:

https://doi.org/10.52731/liir.v003.064

Abstract

Early notice of a learner’s disengagement during learning is an essential signal for an educator to either change the pedagogy or give personal support. Hence, assessing learner’s engagement is important to avoid dropout. However, assessing learner’s engagement in distance learning is a challenge due to the learner-educator interaction limit. To address the challenge, in this paper we proposed a real-time automatic engagement estimation system to assess learner’s engagement from facial landmarks and body pose during his/her learning activity in asynchronous distance learning, where there is no direct interaction between learner and educator. A web-based application has been developed as an early stage implementation in a real education setting. The prototype has been successfully recognizing learner’s engagement in three-level: Very Engaged, Normal Engaged, and Not Engaged.

References

J. Reeve and C.-M. Tseng, ”Agency as a fourth aspect of students’ engagement

during learning activities,” Contemporary Educational Psychology, 36(4):257–267,

2011. ISSN 0361476X. doi: 10.1016/j.cedpsych.2011.05.002. URL https: //linkinghub.

elsevier.com/retrieve/pii/S0361476X11000191.

J. A. Fredricks, P. C. Blumenfeld, and A. H. Paris, ”School Engagement: Potential of

the Concept, State of the Evidence,” Review of Educational Research, 74(1):59–109, 3

ISSN 0034-6543. doi: 10.3102/00346543074001059.

K. L. Alexander, D. R. Entwisle, and C. S. Horsey, ”From First Grade

Forward: Early Foundations of High School Dropout,” Sociology of Education,

(2):87, 4 1997. ISSN 00380407. doi: 10.2307/2673158. URL

https://www.jstor.org/stable/2673158?origin=crossref.

H. Monkaresi, N. Bosch, R. A. Calvo, and S. K. D’Mello, ”Automated Detection of

Engagement Using Video-Based Estimation of Facial Expressions and Heart Rate,”

IEEE Transactions on Affective Computing, 8(1):15–28, 1 2017. ISSN 1949-3045. doi:

1109/TAFFC.2016.2515084. URL http://ieeexplore.ieee.org/document/7373578/.

J. Whitehill, Z. Serpell, Y. C. Lin, A. Foster, and J. R. Movellan, ”The faces of engagement:

Automatic recognition of student engagement from facial expressions,”

IEEE Transactions on Affective Computing, 5(1):86–98, 2014. ISSN 19493045. doi:

1109/TAFFC.2014.2316163.

J. Shen, H. Yang, J. Li, and Z. Cheng, ”Assessing learning engagement based on facial

expression recognition in MOOC’s scenario,” Multimedia Systems, 28(2):469–478, 4

ISSN 0942-4962. doi: 10.1007/s00530-021-00854-x.

O. Sumer, P. Goldberg, S. D’Mello, P. Gerjets, U. Trautwein, and E. Kasneci, ”Multimodal

Engagement Analysis from Facial Videos in the Classroom,” IEEE Transactions

on Affective Computing, 2021. ISSN 19493045. doi: 10.1109/TAFFC.2021.3127692.

URL https://ieeexplore.ieee.org/document/9613750.

A. Ben-Youssef, C. Clavel, and S. Essid, ”Early Detection of User Engagement

Breakdown in Spontaneous Human-Humanoid Interaction,” IEEE Transactions

on Affective Computing, 12(3):776–787, 7 2021. ISSN 1949-3045. doi:

1109/TAFFC.2019.2898399. URL https://ieeexplore.ieee.org/document/8637822.

P. Goldberg, Sumer, K. Sturmer, W. Wagner, R. Gollner, P. Gerjets, E. Kasneci, and

U. Trautwein, ”Attentive or Not? Toward a Machine Learning Approach to Assessing

Students’ Visible Engagement in Classroom Instruction,” Educational Psychology Review,

(1):27–49, 3 2021. ISSN 1040-726X. doi: 10.1007/s10648-019-09514-z. URL

https://link.springer.com/article/10.1007/s10648-019-09514-z#citeas.

J. Liao, Y. Liang, and J. Pan, ”Deep facial spatiotemporal network for engagement

prediction in online learning,” Applied Intelligence, 51(10):6609–6621,

2021. ISSN 0924-669X. doi: 10.1007/s10489-020-02139-8. URL

https://link.springer.com/10.1007/s10489-020-02139-8.

W. H. Yun, D. Lee, C. Park, J. Kim, and J. Kim, ”Automatic Recognition of Children

Engagement from Facial Video Using Convolutional Neural Networks,” IEEE

Transactions on Affective Computing, 11(4):696–707, 10 2020. ISSN 19493045. doi:

1109/TAFFC.2018.2834350.

A. Gupta, A. D’Cunha, K. Awasthi, and V. Balasubramanian, ”DAiSEE: Towards

User Engagement Recognition in theWild,” 14(8):1–12, 9 2016. doi: https://doi.org/10.

/arXiv.1609.01885. URL https://arxiv.org/abs/1609.01885.

A. Dhall, A. Kaur, R. Goecke, and T. Gedeon, ”EmotiW 2018: Audio-Video, Student

Engagement and Group-Level Affect Prediction,” In Proceedings of the 2018 on International

Conference on Multimodal Interaction - ICMI ’18, number October, pages

–656, New York, New York, USA, 2018. ACM Press. ISBN 9781450356923. doi:

1145/3242969.3264993.

B. Woolf, W. Burleson, I. Arroyo, T. Dragon, D. Cooper, and R. Picard. Affect-aware

tutors: recognising and responding to student affect. International Journal of Learning

Technology, 4(3/4):129, 2009

Downloads

Published

2023-03-08