| Literature DB >> 35590984 |
Xavier Solé-Beteta1, Joan Navarro1, Brigita Gajšek2, Alessandro Guadagni3, Agustín Zaballos1.
Abstract
In face-to-face learning environments, instructors (sub)consciously measure student engagement to obtain immediate feedback regarding the training they are leading. This constant monitoring process enables instructors to dynamically adapt the training activities according to the perceived student reactions, which aims to keep them engaged in the learning process. However, when shifting from face-to-face to synchronous virtual learning environments (VLEs), assessing to what extent students are engaged to the training process during the lecture has become a challenging and arduous task. Typical indicators such as students' faces, gestural poses, or even hearing their voice can be easily masked by the intrinsic nature of the virtual domain (e.g., cameras and microphones can be turned off). The purpose of this paper is to propose a methodology and its associated model to measure student engagement in VLEs that can be obtained from the systematic analysis of more than 30 types of digital interactions and events during a synchronous lesson. To validate the feasibility of this approach, a software prototype has been implemented to measure student engagement in two different learning activities in a synchronous learning session: a masterclass and a hands-on session. The obtained results aim to help those instructors who feel that the connection with their students has weakened due to the virtuality of the learning environment.Entities:
Keywords: digital interactions; student engagement; virtual learning environments
Mesh:
Year: 2022 PMID: 35590984 PMCID: PMC9103305 DOI: 10.3390/s22093294
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Methodological approach to model engagement in virtual learning environments.
Answers from the behavioral engagement category.
| ID | Answer |
|---|---|
| B1 | Interactions between teachers and students |
| B2 | Participation to the activities and exercises proposed by the teacher (during lesson) |
| B3 | Participation to project works proposed by the teacher (after lesson) |
| B4 | Total duration of the microphone on for each participant |
| B5 | Number of raised hands [ |
| B6 | Number of posts in meeting chat |
| B7 | Number of screen sharing |
| B8 | Number of interactions with student’s environment |
| B9 | Percentage of students that have participated in tasks |
| B10 | Lip movement time |
| B11 | Number of interactions among the students and the teacher |
| B12 | Answer to online form or pools |
| B13 | Number of interactions among the students |
| B14 | Number of student questions via chat |
| B15 | Whether or not the camera is on |
| B16 | Number of chat interactions |
| B17 | Number of attendees |
| B18 | Number of voice interventions |
Answers from the emotional category.
| ID | Answer |
|---|---|
| E1 | Type of stickers and emotions posted in chat |
| E2 | Yawning |
| E3 | Tone of voice in conversations |
| E4 | Playing with hands, playing with hair, pens⋯ |
| E5 | Students not watching on the screen of the device |
| E6 | Students talking with others |
| E7 | Students switching off the camera |
| E8 | Students watching around |
| E9 | Non-verbal signs of appreciation, rejection and tension |
| E10 | Student emotion |
| E11 | Background noise |
| E12 | Mean Opinion Score (MOS metric) |
| E13 | Time in meeting |
| E14 | Audio Discontinuity |
Answers from the cognitive engagement category.
| ID | Answer |
|---|---|
| C1 | Time to persevere with the task |
| C2 | Number of answers which are: true/right/correct |
| C3 | Number of correct answers to forms and pools |
| C4 | Time to answer to online forms or pools |
| C5 | Correct answers in debates/polls/surveys |
Answers identified as either implicit or out of scope.
| ID | Answer | Reason |
|---|---|---|
| B1 | Interactions between teachers and students | Implicit |
| B2 | Participation to the activities and exercises proposed by the teacher (during lesson) | Implicit |
| B9 | Percentage of students that have participated in tasks | Implicit |
| B3 | Participation in project works proposed by the teacher (after lesson) | Out of scope |
| B12 | Answer to online form or pools | Out of scope |
| E4 | Playing with hands, playing with hair, pens… | Out of scope |
| E9 | Non-verbal signs of appreciation, rejection and tensions | Out of scope |
| C1 | Time to persevere with the task | Out of scope |
| C2 | Number of answers which are: true/right/correct | Out of scope |
| C3 | Number of correct answers to forms and pools | Out of scope |
| C4 | Time to answer to online forms or pools | Out of scope |
| C5 | Correct answers in debates/polls/surveys | Out of scope |
Resulting set of features obtained at the Feature Identification stage.
| ID | Feature | Engagement |
|---|---|---|
| B4 | Total duration of the microphone on for each participant | Behavioral |
| B5 | Number of raised hands | Behavioral |
| B6 | Number of posts in meeting chat | Behavioral |
| B7 | Amount of screen sharing | Behavioral |
| B8 | Number of interactions with student’s environment | Behavioral |
| B10 | Lip movement time | Behavioral |
| B11 | Number of interactions among the students and the teacher | Behavioral |
| B13 | Number of interactions among the students | Behavioral |
| B14 | Number of student’s questions via chat | Behavioral |
| B15 | Whether or not the camera is on | Behavioral |
| B16 | Number of chat interactions | Behavioral |
| B17 | Number of attendees | Behavioral |
| B18 | Number of voice interventions | Behavioral |
| E1 | Type of stickers and emotions posted in chat | Emotional |
| E2 | Yawning | Emotional |
| E3 | Tone of voice in conversations | Emotional |
| E5 | Students not watching on the screen of the device | Emotional |
| E6 | Students talking with others | Emotional |
| E7 | Students switching off the camera | Emotional |
| E8 | Students watching around | Emotional |
| E10 | Student emotion | Emotional |
| E11 | Background noise | Emotional |
| E12 | Mean Opinion Score (MOS metric) | Emotional |
| E13 | Time in meeting | Emotional |
| E14 | Audio Discontinuity | Emotional |
Result of grouping the 25 features obtained at the Feature Identification stage into 10 digital categories.
| Digital Category | Engagement | No./% Total | Features |
|---|---|---|---|
| Attendance | Behavioral/Emotional | 2/8% | (B17) Number of attendees, (E13) Time in meeting |
| Camera usage | Behavioral/Emotional | 2/8% | (B15) Whether or not the camera is on, (E7) Students switching off the camera |
| Voice interactions | Behavioral | 4/16% | (B4) Total duration of the microphone on for each participant, (B11) Number of interactions among the students and the teacher, (B13) number of interactions among the students, (B18) Number of voice interventions |
| Hand rising | Behavioral | 1/4% | (B5) Number of raised hands |
| Screen sharing | Behavioral | 1/4% | (B7) Number of screen sharing |
| Chat interactions | Behavioral/Emotional | 4/16% | (B6) Number of posts in meeting chat, (B14) Number of student’s questions via chat, (B16) Number of chat interactions, (E1) Type of stickers and emotions posted in chat |
| Sound analysis | Behavioral | 4/16% | (E3) Tone of voice in conversations, (E11) Background noise, (E12) Mean Opinion Score (MOS metric), (E14) Audio Discontinuity |
| Facial Emotion | Emotional | 1/4% | (E10) Student emotion |
| Mouth-movement analysis | Behavioral/Emotional | 3/12% | (B10) Lip movement time, (E2) Yawning, (E6) Students talking with others |
| Eye gaze | Emotional | 3/12% | (E6) Students talking with others, (E5) Students not watching on the screen of the device, (E8) Students watching around |
| TOTAL | 25/100% |
Set of digital features defined for “Attendance” category, corresponding to features B17 and E13.
| Level | Measurable Digital Feature |
|---|---|
| Group | (ATG1) Current number of attendees/number of enrolled students |
| Group | (ATG2) Maximum number of attendees/number of enrolled students |
| Group | (ATG3) Minimum number of attendees/number of enrolled students |
| Individual | (ATI1) Time in meeting |
Set of digital features defined for “camera usage” category, corresponding to features B15 and E7.
| Level | Measurable Digital Feature |
|---|---|
| Group | (CUG1) Number of students/% of the time with camera on |
| Group | (CUG2) Number of students/% of the time with camera off |
| Group | (CUG3) Number of students/% of students that have switched the camera off once or more times |
| Individual | (CUI1) Current state (on/off) of the camera |
| Individual | (CUI2) Time camera on/% of the session elapsed time |
| Individual | (CUI3) Time camera off/% of the session elapsed time |
| Individual | (CUI4) Times that the camera has been switched off |
Set of digital features defined for the “voice interactions” category, corresponding to features B4, B11, B13 and B18.
| Level | Measurable Digital Feature |
|---|---|
| Group | (VIG1) Overall spoken minutes by the lecturer/% of the session elapsed time |
| Group | (VIG2) Number of students that have spoken/% of the current attendees |
| Group | (VIG3) Overall spoken minutes by learners/% of the session elapsed time |
| Group | (VIG4) Overall silence minutes/% of the session elapsed time |
| Individual | (VII1) Whether or not has spoken |
| Individual | (VII2) Minutes spoken/% of the session elapsed time |
| Individual | (VII3) Times that has spoken |
Set of digital features defined for “Hand rising” category, corresponding to the feature B5.
| Level | Measurable Digital Feature |
|---|---|
| Group | (HRG1) Number of students that have raised hand/% of the current attendees |
| Individual | (HRI1) Whether or not has raised hand |
| Individual | (HRI2) Number of times he has raised hand |
Set of digital features defined for “Screen sharing” category, corresponding to the feature B7.
| Level | Measurable Digital Feature |
|---|---|
| Group | (SSG1) Number of students that have shared screen/% of the current attendees |
| Individual | (SSI1) Whether or not has shared the screen |
| Individual | (SSI2) Number of times he has shared the screen |
Set of digital features defined for “chat interactions” category, corresponding to features B6, B14, B16 and E1.
| Level | Measurable Digital Feature |
|---|---|
| Group | (CIG1) Number of students that have written posts/% of the current attendees |
| Group | (CIG2) Number of posts written by students |
| Group | (CIG3) Type of stickers used |
| Individual | (CII1) Whether or not has written posts |
| Individual | (CII2) How many posts has written |
| Individual | (CII3) How many of those posts have been questions |
| Individual | (CII4) Type of stickers used |
Set of digital features defined for “facial emotion” category, corresponding to the features E10.
| Level | Measurable Digital Feature |
|---|---|
| Group | (FEG1) Number of students for each of the emotions and % of the current attendees |
| Individual | (FEI1) Real-time main emotion |
Set of digital features defined for “sound analysis” category, corresponding to features E11, E12, E13 and E14.
| Level | Measurable Digital Feature |
|---|---|
| Group | (SAG1) MOS |
| Group | (SAG2) Background noise |
| Group | (SAG3) Loudness |
| Group | (SAG4) Discontinuity |
Set of digital features defined for “mouth-movement analysis” category, corresponding to features B10, E2 and E6.
| Level | Measurable Digital Feature |
|---|---|
| Group | (MMG1) Number of students that have yawned/% of the current attendees |
| Group | (MMG2) Number of students that have had interacted with other people outside the virtual session/% of the current attendees |
| Individual | (MMI1) How many times the student has yawned |
| Individual | (MMI2) Whether or not the student has had interaction with people outside the virtual session |
Set of digital features defined for “eye gaze” category, corresponding to features E6, E5 and E8.
| Level | Measurable Digital Feature |
|---|---|
| Group | (EGG1) Number of students that have leaved the area captured by the camera/% of the current attendees |
| Group | (EGG2) Number of students that have been tagged as not watching at the screen at some moment/% of the current attendees |
| Individual | (EGI1) Times that the student leaves the area captured by the camera |
| Individual | (EGI2) Whether or not the student has not been watching at the screen⋯ |
| Individual | (EGI3) Percentage of attention looking at the screen for each student |
Categorization of the identified features, 25 features classified in 10 digital categories.
| Digital Category | Engagement | No./% Initial Features | No./% Digital Features |
|---|---|---|---|
| Attendance | Behavioral/Emotional | 2/8% | 4/9% |
| Camera usage | Behavioral/Emotional | 2/8% | 7/15% |
| Voice interactions | Behavioral | 4/16% | 7/15% |
| Hand rising | Behavioral | 1/4% | 3/7% |
| Screen sharing | Behavioral | 1/4% | 3/7% |
| Chat interactions | Behavioral/Emotional | 4/16% | 7/15% |
| Sound analysis | Behavioral | 4/16% | 4/9% |
| Facial Emotion | Emotional | 1/4% | 2/4% |
| Mouth-movement analysis | Behavioral/Emotional | 3/12% | 4/9% |
| Eye gaze | Emotional | 3/12% | 5/10% |
| TOTAL | 25/100% | 46/100% |
Relevance of each category according to its weight in the teaching activity.
| Lecture | Tutorial | Laboratory Class | Seminar | Doubts Session | PI | |
|---|---|---|---|---|---|---|
|
| VR | VR | VR | VR | VR | VR |
|
| I | I | I | R | R | VR |
|
| R | R | R | R | R | R |
|
| R | R | R | R | VR | I |
|
| I | I | R | I | R | I |
|
| I | I | I | R | VR | I |
|
| VR | SR | SR | VR | VR | VR |
|
| VR | VR | VR | VR | VR | VR |
|
| VR | R | R | VR | R | VR |
|
| VR | SR | SR | VR | SR | I |
Figure 2System prototype conceptualization, components and interactions.
Figure 3Software prototype, masterclass results.
Figure 4Software prototype, hands-on results.