Literature DB >> 33968373

Affective rating of audio and video clips using the EmojiGrid.

Alexander Toet1,2, Jan B F van Erp1,3.   

Abstract

Background: In this study we measured the affective appraisal of sounds and video clips using a newly developed graphical self-report tool: the EmojiGrid. The EmojiGrid is a square grid, labeled with emoji that express different degrees of valence and arousal. Users rate the valence and arousal of a given stimulus by simply clicking on the grid.
Methods: In Experiment I, observers (N=150, 74 males, mean age=25.2±3.5) used the EmojiGrid to rate their affective appraisal of 77 validated sound clips from nine different semantic categories, covering a large area of the affective space. In Experiment II, observers (N=60, 32 males, mean age=24.5±3.3) used the EmojiGrid to rate their affective appraisal of 50 validated film fragments varying in positive and negative affect (20 positive, 20 negative, 10 neutral).
Results: The results of this study show that for both sound and video, the agreement between the mean ratings obtained with the EmojiGrid and those obtained with an alternative and validated affective rating tool in previous studies in the literature, is excellent for valence and good for arousal. Our results also show the typical universal U-shaped relation between mean valence and arousal that is commonly observed for affective sensory stimuli, both for sound and video. Conclusions: We conclude that the EmojiGrid can be used as an affective self-report tool for the assessment of sound and video-evoked emotions. Copyright:
© 2021 Toet A and van Erp JBF.

Entities:  

Keywords:  EmojiGrid; affective response; arousal; audio clips; valence; video clips

Mesh:

Year:  2020        PMID: 33968373      PMCID: PMC8080979          DOI: 10.12688/f1000research.25088.2

Source DB:  PubMed          Journal:  F1000Res        ISSN: 2046-1402


Introduction

In daily human life, visual and auditory input from our environment significantly determines our feelings, behavior and evaluations ( Fazio, 2001; Jaquet ; Turley & Milliman, 2000, for a review see: Schreuder ). The assessment of the affective response of users to the auditory and visual characteristics of for instance (built and natural) environments ( Anderson ; Huang ; Kuijsters ; Ma & Thompson, 2015; Medvedev ; Toet ; Watts & Pheasant, 2015) and their virtual representations ( Houtkamp & Junger, 2010; Houtkamp ; Rohrmann & Bishop, 2002; Toet ; Westerdahl ), multimedia content ( Baveye ; Soleymani ), human-computer interaction systems ( Fagerberg ; Hudlicka, 2003; Jaimes & Sebe, 2010; Peter & Herbon, 2006; Pfister ) and (serious) games ( Anolli ; Ekman & Lankoski, 2009; Garner ; Geslin ; Tsukamoto ; Wolfson & Case, 2000) is an essential part of their design and evaluation and requires efficient methods to assess whether the desired experiences are indeed achieved. A wide range of physiological, behavioral and cognitive measures is currently available to measure the affective response to sensorial stimuli, each with their own advantages and disadvantages (for a review see: Kaneko ). The most practical and widely used instruments to measure affective responses are questionnaires and rating scales. However, their application is typically time-consuming and requires a significant amount of mental effort (people typically find it difficult to name their emotions, especially mixed or complex ones), which affects the experience itself ( Constantinou ; Lieberman, 2019; Lieberman ; Taylor ; Thomassin ; for a review see: Torre & Lieberman, 2018) and restricts repeated application. While verbal rating scales are typically more efficient than questionnaires, they also require mental effort since users are required to relate their affective state to verbal descriptions (labels). Graphical rating tools however allow users to intuitively project their feelings to figural elements that correspond to their current affective state. Arousal and pleasantness (valence) are principal dimensions of affective responses to environmental stimuli ( Mehrabian & Russell, 1974). A popular graphical affective self-report tool is the Self-Assessment Mannikin (SAM) ( Bradley & Lang, 1994): a set of iconic humanoid figures representing different degrees of valence, arousal, and dominance. Users respond by selecting from each of the three scales the figure that best expresses their own feeling. The SAM has previously been used for the affective rating of video fragments (e.g., Bos ; Deng ; Detenber ; Detenber ; Ellard ; Ellis & Simons, 2005; Fernández ; Soleymani ) and auditory stimuli (e.g., Bergman ; Bradley & Lang, 2000; Lemaitre ; Morris & Boone, 1998; Redondo ; Vastfjall ). Although the SAM is validated and widely used, users often misunderstand the depicted emotions ( Hayashi ; Yusoff ): especially the arousal dimension (shown as an ‘explosion’ in the belly area) is often interpreted incorrectly ( Betella & Verschure, 2016; Broekens & Brinkman, 2013; Chen ; Toet ). The SAM also requires a successive assessment of the stimulus on each of its individual dimensions. To overcome these problems we developed an alternative intuitive graphical self-report tool to measure valence and arousal: the EmojiGrid ( Toet ). The EmojiGrid is a square grid (resembling the Affect Grid: Russell ), labeled with emoji that express various degrees of valence and arousal. Emoji are facial icons that can elicit the same range of neural ( Gantiva ) and emotional ( Moore ) responses as real human faces. In contrast to photographs, emoji are not associated with overgeneralization (the misattribution of emotions and traits to neutral human faces that merely bear a subtle structural resemblance to emotional expressions: Said ), or racial, cultural and sexual biases. Although some facial emoji can be poly-interpretable ( Miller ; Tigwell & Flatla, 2016) it has been found that emoji with similar facial expressions are typically attributed similar meanings ( Jaeger & Ares, 2017; Moore ) that are also to a large extent language independent ( Novak ). Emoji have a wide range of different applications, amongst others in psychological research ( Bai ). Emoji based rating tools are increasingly becoming popular tools as self-report instruments ( Kaye ) to measure for instance user and consumer experience (e.g. www.emojiscore.com). Since facial expressions can communicate a wide variety of both basic and complex emotions emoji-based self-report tools may also afford the measurement and expression of mixed (complex) emotions that are otherwise hard to verbalize ( Elder, 2018). However, while facial images and emoji are processed in a largely equivalent manner, suggesting that some non-verbal aspects of emoji are processed automatically, further research is required to establish whether they are also emotionally appraised on an implicit level ( Kaye ). The EmojiGrid enables users to rate the valence and arousal of a given stimulus by simply clicking on the grid. It has been found that the use of emoji as scale anchors facilitates affective over cognitive responses ( Phan ). Previous studies on the assessment of affective responses to food images ( Toet ) and odorants ( Toet ) showed that the EmojiGrid is self-explaining: valence and arousal ratings did not depend on framing and verbal instructions ( Kaneko ; Toet ). The current study was performed to investigate the EmojiGrid for the affective appraisal of auditory and visual stimuli. Sounds can induce a wide range of affective and physiological responses ( Bradley & Lang, 2000; Gomez & Danuser, 2004; Redondo ). Ecological sounds have a clear association with objects or events. However, music can also elicit emotional responses that are as vivid and intense as emotions that are elicited by real-world events ( Altenmüller ; Gabrielsson & Lindström, 2003; Krumhansl, 1997) and can activate brain regions associated with reward, motivation, pleasure and the mediation of dopaminergic levels ( Blood & Zatorre, 2001; Brown ; Menon & Levitin, 2005; Small ). Even abstract or highly simplified sounds can convey different emotions ( Mion ; Vastfjall ) and can elicit vivid affective mental images when they have some salient acoustic properties in common with the actual sounds. As a result, auditory perception is emotionally biased ( Tajadura-Jiménez ; Tajadura-Jiménez & Västfjäll, 2008). Video clips can also effectively evoke various affective and physiological responses ( Aguado ; Carvalho ; Rottenberg ; Schaefer ). While sounds and imagery individually elicit various affective responses that recruit similar brain structures ( Gerdes ), a wide range of non-linear interactions at multiple processing levels in the brain make that their combined effects are not a priori evident (e.g., Spreckelmeyer ; for a review see: Schreuder ). Several standardized and validated affective databases have been presented to enable a systematic investigation of sound ( Bradley & Lang, 1999; Yang ) and video ( Aguado ; Carvalho ; Hewig ; Schaefer ) elicited affective responses. This study evaluates the EmojiGrid as a self-report tool for the affective appraisal of auditory and visual events. In two experiments, participants were presented with different sound and video clips, covering both a large part of the valence scale and a wide range of semantic categories. The video clips were stripped of their sound channel (silent) to avoid interaction effects. After perceiving each stimulus, participants reported their affective appraisal (valence and arousal) using the EmojiGrid. The sound samples ( Yang ) and video clips ( Aguado ) had been validated in previous studies in the literature using 9-point SAM affective rating scales. This enables an evaluation of the EmojiGrid by directly comparing the mean affective ratings obtained with it to those that were obtained with the SAM. In this study we also investigate how the mean valence and arousal ratings for the different stimuli are related. Although the relation between valence and arousal for affective stimuli varies between individuals and cultures ( Kuppens ), it typically shows a quadratic (U-shaped) form across participants (i.e., at the group level): stimuli that are on average rated either high or low on valence are typically also rated as more arousing than stimuli that are on average rated near neutral on valence ( Kuppens ; Mattek ). For the valence and arousal ratings obtained with the EmojiGrid, we therefore also investigate to what extent a quadratic form describes their relation at the group level.

Methods

Participants

English speaking participants from the UK were recruited via the Prolific database ( https://www.prolific.co/). Exclusion criteria were age (outside the range of 18–35 years old) and hearing or (color) vision deficiencies. No further attempts were made to eliminate any sampling bias. We estimated the sample size required for this study with the “ ICC.Sample.Size” R-package, assuming an ICC of 0.70 (generally considered as ‘moderate’: Landis & Koch, 1977), and determined that sample sizes of 57 (Experiment 1) and 23 (Experiment 2) would yield a 95% confidence interval of sufficient precision (±0.07; Landis & Koch, 1977). Because the current experiment was run online and not in a well-controlled laboratory environment, we aimed to recruit about 2–3 times the minimum required number of participants. This study was approved by the by TNO Ethics Committee (Application nr: 2019-012), and was conducted in accordance with the Helsinki Declaration of 1975, as revised in 2013 ( World Medical Association, 2013). Participants electronically signed an informed consent by clicking “ I agree to participate in this study”, affirming that they were at least 18 years old and voluntarily participated in the study. The participants received a small financial compensation for their participation.

Measures

The participants in this study reported their nationality, gender and age. The EmojiGrid is a square grid (similar to the Affect Grid: Russell ), labeled with emoji that express various degrees of valence and arousal ( Figure 1). Users rate their affective appraisal (i.e., the valence and arousal) of a given stimulus by pointing and clicking at the location on the grid that that best represents their impression. The EmojiGrid was originally developed and validated for the affective appraisal of food stimuli, since the SAM appeared to be frequently misunderstood in that context ( Toet ). It has since also been used and validated for the affective appraisal of odors ( Toet ).
Figure 1.

The EmojiGrid.

The iconic facial expressions range from disliking (unpleasant) via neutral to liking (pleasant) along the horizontal (valence) axis, while their intensity increases along the vertical (arousal) axis. This figure has been reproduced with permission from Toet .

The EmojiGrid.

The iconic facial expressions range from disliking (unpleasant) via neutral to liking (pleasant) along the horizontal (valence) axis, while their intensity increases along the vertical (arousal) axis. This figure has been reproduced with permission from Toet .

Procedure

Participants took part in two anonymous online surveys, created with the Gorilla experiment builder ( Anwyl-Irvine ). After thanking the participants for their interest, the surveys first gave a general introduction to the experiment. The instructions asked the participants to perform the survey on a computer or tablet (but not on a device with a small screen such as a smartphone) and to activate the full-screen mode of their browser. This served to maximize the resolution of the questionnaire and to prevent distractions by other programs running in the background. In Experiment I (sounds) the participants were asked to turn off any potentially disturbing sound sources in their room. Then the participants were informed that they would be presented with a given number of different stimuli (sounds in Experiment I and video clips in Experiment II) during the experiment and they were asked to rate their affective appraisal of each stimulus. The instructions also mentioned that it was important to respond seriously, while there would be no correct or incorrect answers. Participants could electronically sign an informed consent. By clicking “ I agree to participate in this study ”, they confirmed that they were at least 18 years old and that their participation was voluntary. The survey then continued with an assessment of the demographic variables (nationality, gender, age). Next, the participants were familiarized with the EmojiGrid. First, it was explained how the tool could be used to rate valence and arousal for each stimulus. The instructions were: “ To respond, first place the cursor inside the grid on a position that best represents how you feel about the stimulus, and then click the mouse button.” Note that the dimensions of valence and arousal were not mentioned here. Then the participants performed two practice trials. In Experiment I, these practice trials also allowed the repeated playing of the sound stimulus. This was done to allow the participants to adjust the sound level of their computer system. The actual experiment started immediately after the practice trials. The stimuli were presented in random order. The participants rated each stimulus by clicking at the appropriate location on the EmojiGrid. The next stimulus appeared immediately after clicking. There were no time restrictions. On average, each experiment lasted about 15 minutes. Experiment I: Sounds This experiment served to validate the EmojiGrid as a rating tool for the affective appraisal of sound-evoked emotions. Thereto, participants rated valence and arousal for a selection of sounds from a validated sound database using the EmojiGrid. The results are compared with the corresponding SAM ratings provided for each sound in the database. The sound stimuli used in this experiment are 77 sound clips from the expanded version of the validated International Affective Digitized Sounds database (IADS-E, available upon request; Yang ). The sound clips were selected from 9 different semantic categories: scenarios (2), breaking sounds (8), daily routine sounds (8), electric sounds (8), people (8), sound effects (8), transport (8), animals (9), and music (10). For all sounds, Yang provided normative ratings for valence and arousal, obtained with 9-point SAM scales and collected by at least 22 participants from a total pool of 207 young Japanese adults (103 males, 104 females, mean age 21.3 years, SD=2.4). The selection used in the current study was such that the mean affective (valence and arousal) ratings provided for stimuli in the same semantic category were maximally distributed over the two-dimensional affective space (ranging from very negative like a car horn, hurricane sounds or sounds of vomiting, via neutral like people walking up a stairs, to very positive music). As a result, the entire stimulus set is a representative cross-section of the IADS-E covering a large area of the affective space. All sound clips had a fixed duration of 6s. The exact composition of the stimulus set is provided in the Supplementary Material. Each participant rated all sound clips. A total of 150 participants (74 males, 76 females) participated in this experiment. All participants were UK nationals. Their mean age was 25.2 (SD= 3.5) years. Experiment II: Video clips This experiment served to validate the EmojiGrid as a self-report tool for the assessment of emotions evoke by (silent) video clips. Participants rated valence and arousal for a selection of video clips from a validated set of video fragments using the EmojiGrid. The results are compared with the corresponding SAM ratings for the video clips ( Aguado ). The stimuli comprised of a set of 50 film fragments with different affective content (20 positive ones like a coral reef with swimming fishes and jumping dolphins, 10 neutral ones like a man walking in the street or an elevator going down, and 20 negative ones like someone being attacked or a car accident scene). All video clips had a fixed duration of 10 s and were stripped of their soundtracks (for detailed information about the video clips and their availability see Aguado ). Aguado obtained normative ratings for valence and arousal, collected by 38 young adults (19 males, 19 females, mean age 22.3 years, SD=2.2) using 9-point SAM scales. In the present study, each participant rated all video clips using the EmojiGrid. A total of 60 participants (32 males, 28 females) participated in this experiment. All participants were UK nationals. Their mean age was 24.5 (SD= 3.3) years.

Data analysis

The response data (i.e., the horizontal or valence and vertical or arousal coordinates of the check marks on the EmojiGrid) were quantified as integers between 0 and 550 (the size of the square EmojiGrid in pixels), and then scaled between 1 and 9 for comparison with the results of Yang obtained with a 9-point SAM scale (Experiment I), or between 0 and 8 for comparison with the results of Aguado , also obtained with a 9-point SAM scale (Experiment II). All statistical analyses were performed with IBM SPSS Statistics 26 ( www.ibm.com) for Windows. The computation of the intraclass correlation coefficient (ICC) estimates with their associated 95% confidence intervals was based on a mean-rating (k = 3), consistency, 2-way mixed-effects model ( Koo & Li, 2016; Shrout & Fleiss, 1979). ICC values less than 0.5 indicate poor reliability, values between 0.5 and 0.75 suggest moderate reliability, values between 0.75 and 0.9 represent good reliability, while values greater than 0.9 indicate excellent reliability ( Koo & Li, 2016; Landis & Koch, 1977). For all other analyses a probability level of p < 0.05 was considered to be statistically significant. MATLAB 2020a was used to further investigate the data. The mean valence and arousal responses were computed across all participants and for each of the stimuli. MATLAB’s Curve Fitting Toolbox (version 3.5.7) was used to compute least-squares fits to the data points. Adjusted R-squared values were calculated to quantify the agreement between the data and the curve fits.

Results

Experiment I

Figure 2 shows the correlation plots between the mean valence and arousal ratings for the 77 affective IADS-E sounds used in the current study, obtained with the EmojiGrid (this study) and with a 9-point SAM scale ( Yang . This figure illustrates the overall agreement between the affective ratings obtained with both self-assessment tools for affective sound stimuli.
Figure 2.

Relation between mean valence (left) and arousal (right) ratings obtained with the SAM and EmojiGrid for selected sounds from the IADS-E database.

Labels correspond to the original identifiers of the stimuli ( Yang ). The line segments represent linear fits to the data points.

Relation between mean valence (left) and arousal (right) ratings obtained with the SAM and EmojiGrid for selected sounds from the IADS-E database.

Labels correspond to the original identifiers of the stimuli ( Yang ). The line segments represent linear fits to the data points. The linear (two-tailed) Pearson correlation coefficients between the valence and arousal ratings obtained with the EmojiGrid (present study) and with the SAM ( Yang ) were, respectively, 0.881 and 0.760 (p<0.001). To further quantify the agreement between both rating tools we computed intraclass correlation coefficients (ICC) with their 95% confidence intervals for the mean valence and arousal ratings between both studies. The ICC value for valence is 0.936 [0.899–0.959] while the ICC for arousal is 0.793 [0.674–0.868], indicating both studies show an excellent agreement for valence and a good agreement for arousal (even though the current study was performed via the internet and therefore did not provide the amount of control over many experimental factors as one would have in a lab experiment). Figure 3 shows the relation between the mean valence and arousal ratings for the 77 IADS-E sounds used as stimuli in the current study, measured both with the EmojiGrid (this study) and with a 9-point SAM scale ( Yang . The curves in this figure represent least-squares quadratic fits to the data points. The adjusted R-squared values are 0.62 for results obtained with the EmojiGrid and 0.22 for the SAM results. Hence, both methods yield a relation between mean valence and arousal ratings that can indeed be described by a quadratic (U-shaped) relation at the nomothetic (group) level.
Figure 3.

Relation between mean valence and arousal ratings for selected sounds from the IADS-E database.

Labels correspond to the original identifiers of the stimuli ( Yang ). Blue labels represent data obtained with the SAM ( Yang ), while red labels represent data obtained with the EmojiGrid (this study). The curves represent quadratic fits to the corresponding data points.

Relation between mean valence and arousal ratings for selected sounds from the IADS-E database.

Labels correspond to the original identifiers of the stimuli ( Yang ). Blue labels represent data obtained with the SAM ( Yang ), while red labels represent data obtained with the EmojiGrid (this study). The curves represent quadratic fits to the corresponding data points.

Experiment II

Figure 4 shows the correlation plots between the mean valence and arousal ratings for the 50 affective video clips used in the current study, obtained with the EmojiGrid (this study) and with a 9-point SAM scale ( Aguado ). This figure illustrates the overall agreement between the affective ratings obtained with both self-assessment tools for affective sound stimuli.
Figure 4.

Relation between mean valence (left) and arousal (right) ratings obtained with the SAM and EmojiGrid for 50 affective video clips ( Aguado ).

Labels correspond to the original identifiers of the stimuli ( Yang ). The line segments represent linear fits to the data points.

Relation between mean valence (left) and arousal (right) ratings obtained with the SAM and EmojiGrid for 50 affective video clips ( Aguado ).

Labels correspond to the original identifiers of the stimuli ( Yang ). The line segments represent linear fits to the data points. The linear (two-tailed) Pearson correlation coefficients between the valence and arousal ratings obtained with the EmojiGrid (present study) and with the SAM ( Aguado ) were respectively 0.963 and 0.624 (p<0.001). To further quantify the agreement between both rating tools we computed intraclass correlation coefficients (ICC) with their 95% confidence intervals for the mean valence and arousal ratings between both studies. The ICC value for valence is 0.981 [0.967 – 0.989] while the ICC for arousal is 0.721 [0.509 – 0.842], indicating both studies show an excellent agreement for valence and a good agreement for arousal. Figure 5 shows the relation between the mean valence and arousal ratings for the 50 video clips tested. The curves in this figure represent quadratic fits to the data points. The adjusted R-squared values are respectively 0.68 and 0.78. Hence, both methods yield a relation between mean valence and arousal ratings that can be described by a quadratic (U-shaped) relation at the nomothetic (group) level.
Figure 5.

Mean valence and arousal ratings for affective film clips.

Labels correspond to the original identifiers of the stimuli ( Aguado ). Blue labels represent data obtained with the SAM ( Aguado ) while red labels represent data obtained with the EmojiGrid (this study). The curves show quadratic fits to the corresponding data points.

Mean valence and arousal ratings for affective film clips.

Labels correspond to the original identifiers of the stimuli ( Aguado ). Blue labels represent data obtained with the SAM ( Aguado ) while red labels represent data obtained with the EmojiGrid (this study). The curves show quadratic fits to the corresponding data points. Raw data from each experiment are available as Underlying data ( Toet, 2020).

Conclusion

In this study we evaluated the recently developed EmojiGrid self-report tool for the affective rating of sounds and video. In two experiments, observers rated their affective appraisal of sound and video clips using the EmojiGrid. The results show a close correspondence between the mean ratings obtained with the EmojiGrid and those obtained with the validated SAM tool in previous validation studies in the literature: the agreement is excellent for valence and good for arousal, both for sound and video. Also, for both sound and video, the EmojiGrid yields the universal U-shaped (quadratic) relation between mean valence and arousal that is typically observed for affective sensory stimuli. We conclude that the EmojiGrid is an efficient affective self-report tool for the assessment of sound and video-evoked emotions. A limitation of the EmojiGrid is the fact that it is based on the circumplex model of affect which posits that positive and negative feelings are mutually exclusive ( Russell, 1980). Hence, in its present form, and similar to other affective self-report tools like the SAM or VAS scales, the EmojiGrid only allows the measurement of a single emotion at a time. However, emotions are not strictly bipolar and two or more same or opposite valenced emotions can co-occur together ( Larsen & McGraw, 2014; Larsen ). Mixed emotions consisting of opposite feelings can in principle be registered with the EmojiGrid by allowing participants to enter multiple responses. Another limitation of this study is the fact that the comparison of the SAM and EmojiGrid ratings were based on ratings from different populations (akin to a comparison of two independent samples). Hence, our current regression estimates are optimized based on the particular samples that were used. Future studies should investigate a design in which the same participants use both self-report tools to rate the same set of stimuli. Future applications of the EmojiGrid may involve the real-time evaluation of affective events or the provision of affective feedback. For instance, in studies on affective communication in human-computer interaction (e.g., Tajadura-Jiménez & Västfjäll, 2008), the EmojiGrid can be deployed as a continuous response tool by moving a mouse-controlled cursor over the grid while logging the cursor coordinates. Such an implementation may also afford the affective annotation of multimedia ( Chen ; Runge ), and could be useful for personalized affective video retrieval or recommender systems ( Hanjalic & Xu, 2005; Koelstra ; Lopatovska & Arapakis, 2011; Xu ), for real-time affective appraisal of entertainment ( Fleureau ) or to provide affective input to serious gaming applications ( Anolli ) and affective music generation ( Kim & André, 2004). Sensiks ( www.sensiks.com) has adopted a simplified version of the EmojiGrid in its Sensory Reality Pod to enable the user to select and tune multisensory (visual, auditory, tactile and olfactory) affective experiences.

Data availability

Underlying data

Open Science Framework: Affective rating of audio and video clips using the EmojiGrid. https://doi.org/10.17605/OSF.IO/GTZH4 ( Toet, 2020). File ‘Results_sound_video’ (XLSX) contains the EmojiGrid co-ordinates selected by each participant following each stimulus. Open Science Framework: Additional data on affective rating of audio and video clips using the EmojiGrid. https://doi.org/10.17605/OSF.IO/6HQTR File ‘sound_results.xlsx’ contains the mean valence and arousal ratings, obtained with the SAM ( Yang ) and the EmojiGrid (this study), together with graphs in which each of the stimuli are labelled for easy identification. File ‘video_results.xlsx’ contains the mean valence and arousal ratings, obtained with the SAM ( Aguado ) and the EmojiGrid (this study), together with graphs in which each of the stimuli are labelled for easy identification. Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0). Thank you very much the updated paper. The authors has addressed my concerns adequately. Is the work clearly and accurately presented and does it cite the current literature? Yes If applicable, is the statistical analysis and its interpretation appropriate? Partly Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Partly Are the conclusions drawn adequately supported by the results? Partly Are sufficient details of methods and analysis provided to allow replication by others? Partly Reviewer Expertise: Survey methodology, rating formats, Emotions, and Emojis. I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. Thank you to the authors for making relevant changes as noted from the first round of reviews. I am satisfied that these have been addressed sufficiently. Is the work clearly and accurately presented and does it cite the current literature? Partly If applicable, is the statistical analysis and its interpretation appropriate? Yes Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Yes Are the conclusions drawn adequately supported by the results? Yes Are sufficient details of methods and analysis provided to allow replication by others? Partly Reviewer Expertise: Psychology of emoji; cyberpsychology; online behaviour I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. This is an interesting study that seeks to validate the EmojiGrid for use with auditory and video stimuli. Thank you to the authors for providing the research resources on OSF as this is helpful when reviewing the research. Overall, the research has merits but would benefit from being more detailed especially in the introductory and discussion sections. I also have a methodological query but this may be rectified from additional clarity in the writing of this section. The introduction could do with additional literature about the emotional affordances of emoji. That is, the research is presented as assuming that emoji are emotional stimuli but does not provide a review of the literature which can support this. Interestingly, recent evidence (Kaye et al., 2021) suggests that emoji may not be processed emotionally on an implicit level, so the authors should be careful about their assumptions in this regard. Relevant sources that may be useful: Bai, Q., Dan, Q., Mu, Z., & Yang, M. (2019). A systematic review of emoji: Current research and future perspectives.  Frontiers in  Psychology, 10, e2221.   doi:10.3389/fpsyg.2019.02221 [1] Derks, D., Fischer, A. H., & Bos, A. E. R. (2008). The role of emotion in computer-mediated communication: A review. Computers in Human Behavior, 24 (3), 766-785 [2] Kaye, L. K., Rodriguez Cuadrado, S., Malone, S. A., Wall, H. J., Gaunt, E., Mulvey, A. L., & Graham, C. (2021). How emotional are emoji?: Exploring the effect of emotional valence on the processing of emoji stimuli. Computers in Human Behavior, 116, 106648 [3] Novak, P. K., Smailović, J., Sluban, B., & Mozetič, I. (2015). Sentiment of emojis. PLoS ONE, 10 (12), e0144296 [4] With regards to the data presented (e.g., Fig 2), it is not made explicitly clear how numerical values were determined based on the responses from the EmojiGrid. E.g., how are each of the emoji symbols based on their position on the axis determined numerically? From Fig 1, it looks like this ranges from 1 to 5 based on the number of emoji on each axis. However, looking in the methodology, the SAM scale is outlined as being a 9-item response scale so it isn’t clear how Fig 2 & 3 can present the data from these two scales on the same axis if the response scales are different. The discussion could benefit from further elaboration. E.g., To what extent do the findings contribute theoretically to the literature? What are the limitations of the work? Minor In the methodology, it is more typical to use the term “participants” rather than “persons Is the work clearly and accurately presented and does it cite the current literature? Partly If applicable, is the statistical analysis and its interpretation appropriate? Yes Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Yes Are the conclusions drawn adequately supported by the results? Yes Are sufficient details of methods and analysis provided to allow replication by others? Partly Reviewer Expertise: Psychology of emoji; cyberpsychology; online behaviour I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above. Dear Dr Kaye, Thank you for your critical remarks and valuable suggestions which definitely helped us to improve our initial draft paper. Also, we appreciate the fact that you spent your valuable time on this review. 1.  Literature about the emotional affordances of emoji Thank you for this suggestion. We agree that reviewing literature about the emotional affordances of emoji will be a valuable addition to the Introduction, helping the reader to better place the current findings in their context. We therefore added the following text to the Introduction: “Emoji are facial icons that can elicit the same range of neural (Gantiva, Sotaquirá, Araujo, & Cuervo, 2020) and emotional (Moore, Steiner, & Conlan, 2013) responses as real human faces. In contrast to photographs, emoji are not associated with overgeneralization (the misattribution of emotions and traits to neutral human faces that merely bear a subtle structural resemblance to emotional expressions: Said, Sebe, & Todorov, 2009), or racial, cultural and sexual biases. Although some facial emoji can be poly-interpretable (Miller et al., 2016; Tigwell & Flatla, 2016) it has been found that emoji with similar facial expressions are typically attributed similar meanings (Jaeger & Ares, 2017; Moore et al., 2013) that are also to a large extent language independent (Kralj Novak, Smailović, Sluban, & Mozetič, 2015). Emoji have a wide range of different applications, amongst others in psychological research (Bai, Dan, Mu, & Yang, 2019). Emoji based rating tools are increasingly becoming popular tools as self-report instruments (Kaye, Malone, & Wall, 2017) to measure for instance user and consumer experience (e.g. www.emojiscore.com). Since facial expressions can communicate a wide variety of both basic and complex emotions emoji-based self-report tools may also afford the measurement and expression of mixed (complex) emotions that are otherwise hard to verbalize (Elder, 2018). However, while facial images and emoji are processed in a largely equivalent manner, suggesting that some non-verbal aspects of emoji are processed automatically, further research is required to establish whether they are also emotionally appraised on an implicit level (Kaye et al., 2021).” 2.  How numerical values were determined Thank you for pointing out this omission. We now include the following explanation of the scaling in the section on data analysis: “The response data (i.e., the horizontal or valence and vertical or arousal coordinates of the check marks on the EmojiGrid) were quantified as integers between 0 and 550 (the size of the square EmojiGrid in pixels), and then scaled between 1 and 9 for comparison with the results of  Yang et al. (2018) obtained with a 9-point SAM scale (Experiment I), or between 0 and 8 for comparison with the results of Aguado et al. (2018), also obtained with a  9-point SAM scale (Experiment II).” 3.  Contribution and limitations We now address some limitations of the present study (e.g. related to the measurement of mixed emotions and the study design itself) in the Discussion section (see also our reply to the comments of dr Pham). 4. Minor points We replaced “persons” by “participants” throughout the text. Thank you for the opportunity to review the manuscript: “Affective rating of audio and video clips using the EmojiGrid.” This paper is primarily focused on validating the extension of a scale format (EmojiGrid) to a broader range of stimuli (audio and video). Overall, the paper makes some useful methodological contributions such as (1) the potentially greater ease for respondents for rating their emotions; (2) capturing both arousal and valence simultaneously; and (3) the use of more familiar contemporary symbols (emojis) compared to the SAM (Bradley & Lang, 1994). [1] I do have a few suggestions and concerns regarding the paper. 1. Limitation of the EmojiGrid in measuring single discrete emotions. The EmojiGrid is useful for respondents when selecting which area of the grid corresponds to their current felt emotion. However, emotions are not bipolar in nature and can often co-occur together, e.g., feeling bitter-sweet (Larsen et al., 2001; Larsen & McGraw, 2014) [2] , [3]. Thus, the current form of the EmojiGrid is limited to assessing stimuli that invoke single discrete emotions and may not be as suited for assessing real-time affective reactions (e.g., to entertainment or news). This limitation can potentially be highlighted in the discussion. Importantly, this limitation can be solved by future and different operationalizations of the grid structure when mixed emotions are the object of inquiry. 2. Details regarding the stimuli selected. Related to the first point, I note that the majority of the stimuli in both experiments (in particular experiment 1) seem to have a moderate amount of valence and arousal. Without knowing which stimuli were used, it is difficult to assess whether the emotion felt by the respondent was truly neutral or potential mix of emotions. To help the reader, please include two things potentially using tables in the supplementary material if needed. First, a greater description of which stimuli selected was expected to invoke which emotion in terms of both valence and arousal for both experiments. Second, please use a different numbering/labeling/coloring scheme that corresponds to the stimuli instead of dots for figures 1 and 2 when comparing the results from this study to previous work. Both are important because it allows the reader to visually assess the extent an expected emotion of stimuli (e.g., high arousal and positive valence) truly maps onto the mean scores and for the potential discrepancy between the two scale formats for the same stimuli to be obvious. This is important for replication but also because there is a greater dispersion when the SAM rating format is used. 3. Comparing current data and alternate (future) research design. When comparing data from the current experiments to previous experiments the regression estimates are locally optimized based on the sample used to generate them. Thus, a caveat and clarification to potentially include are that the comparisons made are akin to that of two independent samples. Relatedly, an alternate design to consider would be doing a 4-block repeated measures design. Where participants rate the same stimuli using the two rating formats twice as: 1. A then A 2. B then B 3. A then B 4. B then A Blocks 3 and 4 would allow more direct comparisons between two different rating formats, especially given the greater dispersion in ratings observed when the SAM format is used. 4. Free response clicks within the EmojiGrid I note that participants are free to click anywhere within the space of the EmojiGrid. I am curious as to variability/freedom that having no fixed anchor points generates. When participants respond do they more typically engage in: (1) subconsciously select a point close to one of the 25 potential points implied by the 5 X 5 grid of emojis, or (2) freely select a space with the grid, e.g., selecting point that corresponds to 2.30 arousal and 5.80 in valence? I ask this because the reliability of a scale is linked to the number of response points available (Preston & Colman, 2000; Schutz & Rucker, 1975) [4] , [5]. If respondents are truly giving their ratings as (2) then greater reliability would be a potential additional advantage of using the EmojiGrid. If it were (1) the design of the EmojiGrid could include finer lines (i.e., more grid lines) to help respondents more easily locate their emotions on the Grid. Minor points Nationality information was collected from participants how was this information used? What was the distribution of nationalities for the participants? I appreciate the way the authors determined their sample sizes. I enjoyed reading your paper and hope you will find my comments helpful! Is the work clearly and accurately presented and does it cite the current literature? Yes If applicable, is the statistical analysis and its interpretation appropriate? Partly Are all the source data underlying the results available to ensure full reproducibility? Yes Is the study design appropriate and is the work technically sound? Partly Are the conclusions drawn adequately supported by the results? Partly Are sufficient details of methods and analysis provided to allow replication by others? Partly Reviewer Expertise: Survey methodology, rating formats, Emotions, and Emojis. I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above. Dear Dr Phan, Thank you for your helpful suggestions and constructive remarks, which helped us to improve the quality of our initial draft paper. In addition, we appreciate the fact that you spent your valuable time on this review. 1.         Mixed emotions We thank the reviewer for raising this important issue. We now address this limitation in the Conclusion section as follows: “A limitation of the EmojiGrid is the fact that it is based on the circumplex model of affect which posits that positive and negative feelings are mutually exclusive (Russell, 1980).  Hence, in its present form, and similar to other affective self-report tools like the SAM or VAS scales, the EmojiGrid only allows the measurement of a single emotion at a time. However, emotions are not strictly bipolar and two or more same or opposite valenced emotions can co-occur together  (Larsen & McGraw, 2014; Larsen, McGraw, & Cacioppo, 2001). Mixed emotions consisting of opposite feelings can in principle be registered with the EmojiGrid by allowing participants to enter multiple responses. “ 2.  Stimulus details Thank you for bringing this limitation to our attention. We agree that a labelling scheme (e.g. using the original stimuli identifier) makes a visual comparison between the experiments much easier. We therefore replaced the original graph with labelled graphs in the paper to allows the readers to visually assess and verify the expected emotions induced by the stimuli. We also added correlation plots for the mean valence and arousal ratings obtained both with the SAM and EmojiGrid to enable a direct comparison within each dimension. In addition, we now also provide a more detailed description of the selected stimuli in a new set of Excel notebooks that we uploaded to the Open Science Framework  These notebooks include a brief description of the nature and content of all stimuli, their original affective classification, and their mean valence and arousal values (1) as provided by the authors of the (sound and video) databases and (2) as measured in this study. The notebooks also contain several graphs in which each of the stimuli is represented by the index number for easy identification. The graphs include plots showing (1) the relation between the mean valence measures obtained with the SAM and EmojiGrid, (2) the  relation between the mean arousal measures obtained with the SAM and EmojiGrid, (3) the relation between the mean valence and arousal measures obtained with the SAM, and (4) the relation between the mean valence and arousal measures obtained with the EmojiGrid. 3. Comparing data Thank you for pointing out this limitation.  We now address this issue in the Conclusion section as follows: “Another limitation of this study is the fact that the comparison of the SAM and EmojiGrid ratings were based on ratings from different populations (akin to a comparison of two independent samples). Hence, our current regression estimates are optimized based on the particular samples that were used. Future studies should investigate a design in which the same participants use both self-report tools to rate the same set of stimuli. “ 4. Free response clicks Thank you for raising this potentially important issue. We plotted the raw response data for visual inspection. The overall response pattern appears truly random and shows no regularities or evidence for attraction to any of the individual emojis lining the grid area or to any of the grid lines inside the grid area. Minor points Thank you for drawing our attention to this omission. All participants had the UK nationality. We now report this in the text. Nationality information was collected to check if the participants adhered to the recruitment restrictions as specified through Prolific. Thank you for your positive appraisal. Your comments were quite valuable, and definitely served to improve the quality of our paper.
  43 in total

1.  Optimal number of response categories in rating scales: reliability, validity, discriminating power, and respondent preferences.

Authors:  C C Preston; A M Colman
Journal:  Acta Psychol (Amst)       Date:  2000-03

2.  Affective and physiological responses to environmental noises and music.

Authors:  Patrick Gomez; Brigitta Danuser
Journal:  Int J Psychophysiol       Date:  2004-07       Impact factor: 2.997

3.  Combined perception of emotion in pictures and musical sounds.

Authors:  Katja N Spreckelmeyer; Marta Kutas; Thomas P Urbach; Eckart Altenmüller; Thomas F Münte
Journal:  Brain Res       Date:  2006-01-05       Impact factor: 3.252

4.  A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research.

Authors:  Terry K Koo; Mae Y Li
Journal:  J Chiropr Med       Date:  2016-03-31

5.  Structural resemblance to emotional expressions predicts evaluation of emotionally neutral faces.

Authors:  Christopher P Said; Nicu Sebe; Alexander Todorov
Journal:  Emotion       Date:  2009-04

Review 6.  Emotional pictures and sounds: a review of multimodal interactions of emotion cues in multiple domains.

Authors:  Antje B M Gerdes; Matthias J Wieser; Georg W Alpers
Journal:  Front Psychol       Date:  2014-12-01

7.  A Systematic Review of Emoji: Current Research and Future Perspectives.

Authors:  Qiyu Bai; Qi Dan; Zhe Mu; Maokun Yang
Journal:  Front Psychol       Date:  2019-10-15

8.  The Affective Slider: A Digital Self-Assessment Scale for the Measurement of Human Emotions.

Authors:  Alberto Betella; Paul F M J Verschure
Journal:  PLoS One       Date:  2016-02-05       Impact factor: 3.240

9.  Sentiment of Emojis.

Authors:  Petra Kralj Novak; Jasmina Smailović; Borut Sluban; Igor Mozetič
Journal:  PLoS One       Date:  2015-12-07       Impact factor: 3.240

Review 10.  Methods for Evaluating Emotions Evoked by Food Experiences: A Literature Review.

Authors:  Daisuke Kaneko; Alexander Toet; Anne-Marie Brouwer; Victor Kallen; Jan B F van Erp
Journal:  Front Psychol       Date:  2018-06-08
View more
  1 in total

1.  Linking Categorical and Dimensional Approaches to Assess Food-Related Emotions.

Authors:  Alexander Toet; Erik Van der Burg; Tim J Van den Broek; Daisuke Kaneko; Anne-Marie Brouwer; Jan B F Van Erp
Journal:  Foods       Date:  2022-03-27
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.