Scholarly open access journals, Peer-reviewed, and Refereed Journals, Impact factor 8.14 (Calculate by google scholar and Semantic Scholar | AI-Powered Research Tool) , Multidisciplinary, Monthly, Indexing in all major database & Metadata, Citation Generator, Digital Object Identifier(DOI)
Recognizing and enhancing student engagement is crucial for improving learning outcomes, particularly in the context of online classes where monitoring can be challenging. Traditional methods of attendance tracking, such as calling out names, are impractical and susceptible to manipulation in the virtual environment. This project is to develop a novel design using the AI based FFCNN (Face Fiducial Convolution Neural Network) model to capture face biometric randomly from students’ video stream and record their attendance automatically. In the proposed system, the learner’s face is monitored by a video camera while attending a video lecture using light Gradient Boosting Machine (LightGBM) to predict the attention and engagement level of the student by facial feature behavior analysis and random QA system.
The present study aimed to develop a method for estimating students’ attentional state from facial expressions during online lectures. We estimated the level of attention while students watched a video lecture by measuring reaction time (RT) to detect a target sound that was irrelevant to the lecture. We assumed that RT to such a stimulus would be longer when participants were focusing on the lecture compared with when they were not. We sought to estimate how much learners focus on a lecture using RT measurement. In the experiment, the learner’s face was recorded by a video camera while watching a video lecture. Facial features were analyzed to predict RT to a task-irrelevant stimulus, which was assumed to be an index of the level of attention. We applied a machine learning method, light Gradient Boosting Machine (LightGBM), to estimate RTs from facial features extracted as action units (AUs) corresponding to facial muscle movements by an open-source software (OpenFace). The model obtained using LightGBM indicated that RTs to the irrelevant stimuli can be estimated from AUs, suggesting that facial expressions are useful for predicting attentional states while watching lectures. We re-analyzed the data while excluding RT data with sleepy faces of the students to test whether decreased general arousal caused by sleepiness was a significant factor in the RT lengthening observed in the experiment. The results were similar regardless of the inclusion of RTs with sleepy faces, indicating that facial expression can be used to predict learners’ level of attention to video lectures.
Keywords:
Cite Article:
"Analysis of Facial Expressions to Estimate the Level of Engagement in Online Lectures", International Journal for Research Trends and Innovation (www.ijrti.org), ISSN:2455-2631, Vol.10, Issue 5, page no.a18-a23, May-2025, Available :http://www.ijrti.org/papers/IJRTI2505003.pdf
Downloads:
000467
ISSN:
2456-3315 | IMPACT FACTOR: 8.14 Calculated By Google Scholar| ESTD YEAR: 2016
An International Scholarly Open Access Journal, Peer-Reviewed, Refereed Journal Impact Factor 8.14 Calculate by Google Scholar and Semantic Scholar | AI-Powered Research Tool, Multidisciplinary, Monthly, Multilanguage Journal Indexing in All Major Database & Metadata, Citation Generator