Learner's Engagement Analysis For E-Learning Platform: ISSN: 2455-2631 2020 IJSDR - Volume 5, Issue
Learner's Engagement Analysis For E-Learning Platform: ISSN: 2455-2631 2020 IJSDR - Volume 5, Issue
Abstract: Adaptive learning is an educational method that utilizes computers as an interactive teaching device. In existing
most educational agents do not monitor engagement explicitly, but rather assume engagement and feedback by interaction
based on the student’s responses to questionnaires, quiz and tasks. Thus, dynamic learner engagement analysis is a first step
towards an automated course feedback tool helpful to the instructor for measuring student engagement. In our system, we
propose a hybrid architecture system invoking student facial emotion recognition, eye gaze monitoring, head movements
identifications based analyzing dynamic student engagement in classroom and towards a specific course at e-learning
platforms. Our proposed architecture uses feature extraction algorithms for facial emotion recognition, Haar Cascade for
pupil detection and Local Binary Patterns for recognizing head movements. For a machine learning approach and to
provide accurate results we propose a Convolutional Neural Network. Experimental results have been implemented using
PyCharm IDE.
Keywords: Video processing, e-learning, emotion recognition, eye tracking, head pose identification
I. INTRODUCTION
Recognizing faces and distinguishing between faces is an inborn ability of humans. By introducing that ability to computers is
feasible by means machine learning and computer vision techniques. This can open mountain of applications like distracted driver
detection, access and security to personalized devices, human robot interaction for youngsters with autism, entertainment industries,
E-Learning feedback, allow payments to be processed without physical cards, enable criminal identification and permit personalized
healthcare like patient monitoring and other services. Emotions on an individual's face by using most advanced image processing
tools is read using emotion recognition phenomenon. The proposed model tries to recognize emotions based on facial expressions
such as happy, anger, disgust, fear, joy, surprise as shown in Fig. 1.
Facial expression detection in images and videos is the most impetuous and ordinary way to identify emotions of an individual.
As the cost of education like tuition, fees and living expenses has skyrocketed over the past few decades, prolonged graduation time
has become a crucial contributing factor to the ever-growing student graduation. Recent studies show that only 50 of more than 580
public four-year institutions in the United States have on-time graduation rates at or above 50 percent for their full-time students.
To make college cheaper, it’s thus important to ensure that several students graduate on time through prior involvement on
students whose performance are unlikely to satisfy the graduation criteria of the degree program on time. A critical step towards
effective intervention is to develop a system which will continuously keep track of student’s attention levels and accurately predict
their mood of listening and supply teachers with information on progress and achievements of their students. Thus, teachers can
assess the impact of their instructional strategies they use.
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 156
ISSN: 2455-2631 © March 2020 IJSDR | Volume 5, Issue 3
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 157
ISSN: 2455-2631 © March 2020 IJSDR | Volume 5, Issue 3
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 158
ISSN: 2455-2631 © March 2020 IJSDR | Volume 5, Issue 3
Convolution neural network is one of the representative network structures in deep learning and has become a hotspot in the
field of speech analysis and image recognition. CNN can take real-time raw image frame input, thus eliminating feature extraction
and data reconstruction process in the standard deep learning methodologies. Weight network layer’s structure of CNN architecture
matches the biological neural network of humans, which reduces the difficulty of the network model and lowers the number of
weights.
D. Algorithm
● Step 1: The learner logins into the e-learning portal and the web-camera begins frame acquisition.
● Step 2: The face of the learner is detected and then processed.
● Step 3: The eyes region and other regions of interest are detected and cropped.
● Step 4: Student’s head position is identified
● Step 5: Students eye position is tracked whether left, right or up.
● Step 6: Student emotion is predicted.
● Step 7: The student’s attention state is classified into concentrated or distracted based on the valid combinations of head,
eye and student’s emotion.
● Step 8: Finally, if the student’s engagement level is engaged course URL is activated for video lecture otherwise an URL
is activated for course documentation.
E. List of Hyperparameters
The below table TABLE I. shows the list of hyper-parameters and their corresponding variants involved in the proposed
convolutional neural network of the emotion classifier module.
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 159
ISSN: 2455-2631 © March 2020 IJSDR | Volume 5, Issue 3
V. RESULTS DISCUSSION
The results obtained from the implementation of the proposed model has about 63.8% of test accuracy. The output from the three
basic modules of the proposed system are shown below in Fig. 4.
The output combinations from the above three modules are used to detect the status of the learner as concentrated or distracted
as represented in TABLE II. If the student is classified as concentrated then the course URL for the video lecture can be activated.
Otherwise the course documentation URL will be activated for the distracted learner.
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 160
ISSN: 2455-2631 © March 2020 IJSDR | Volume 5, Issue 3
VI. CONCLUSION
In this paper, we have proposed a system that uses a progressive and repetitive method to monitor and record the learning
individual’s engagement level thereby analyzing the learner’s concentration to the instructional content and initiates a response to
redirect the learner’s attention when they become distracted. The quality of the course content can be enhanced forward based on
the level of engagement recognized using eye tracking and head movement detection. The proposed system is highly effective in
identifying facial expression, emotions, eye-gaze and head-pose in real-time e-learning environments and detects the level of focus
or distraction of the student. Moreover, the common methodology can be used for other aspects like eye-gaze and head-pose
movement identifications. Future improvements can be made by adding facial area recognition to uniquely identify the individual
student and provide personal feedback and evaluation of learner focus during the e-learning period of time. Also, the same proposed
system can be modified to evaluate individual engagement in a classroom environment and provide the teacher with this information
in real-time or as a summarized report at the end of a lecture. Further, this method can also be applied to any kind of real-time
learning situation.
REFERENCES
[1] Fabri, M., Moore, D.J., Hobbs, D.J “Mediating the Expression of Emotion in Educational Collaborative Virtual Environments:
An Experimental Study” (2014)
[2] M. Feidakis, T. Daradoumis, S. Caballé and J. Conesa “Measuring the Impact of Emotion Awareness on e-learning Situations”
(2013)
[3] J. Yu “An Infrastructure for Real-Time Interactive Distance E-Learning Environment” (2010)
[4] S. Al. Hend, G. K. Remya “Eye Tracking and e-Learning: Seeing Through Your Students” (2010)
[5] I.E. Haddioui, and M. Khaldi “Learner Behavior Analysis through Eye Tracking” (2011)
[6] Pushpaja V. Saudagar, D.S. Chaudhari “Facial Expression Recognition using Neural Network –An Overview “(2012)
[7] Mohamed Sathik M, Sofia G “Identification of student comprehension using forehead wrinkles” (2011)
[8] Bailenson J, Beall A, Blascovich J, Raimundo M, Weishbush M “Intelligent agents who wear your face: User’s reactions to
the virtual self (2010)
[9] Goodfellow, I.J., Erhan, D., Carrier, P.L., Courville, A., Mirza, M., Hamner, B., Cukierski, W., Tang, Y., Thaler, D., Lee, D.H.,
et al.: Challenges in representation learning: A report on three machine learning contests. In: ICONIP. pp. 117–124. Springer
(2013)
IJSDR2003030 International Journal of Scientific Development and Research (IJSDR) www.ijsdr.org 161