Literature DB >> 36168447

Variations in consumer acceptance, sensory engagement and method practicality across three remote consumer-testing modalities.

Marta Albiol Tapia1, Soo-Yeun Lee1.   

Abstract

In the context of the COVID-19 pandemic, it has become challenging for sensory scientists to conduct in-person sensory tests, particularly large central location tests. Sensory literature comparing central location and home use tests shows no clear consensus about how each methodology affects sample ratings and panelist engagement. Research on instructional delivery suggests that the most effective method of increasing engagement involves interactive video conferencing. The objective of this study was to compare three methods of remote consumer testing regarding sample acceptance, sensory engagement, and method practicality. Eighty-four participants rated five chocolate-chip cookie products on a 9-pt hedonic scale in each of three methods: 1) a live (synchronous) Zoom session, 2) an asynchronous video-guided session, and 3) a fully written protocol session. Results showed no significant differences in sample liking pattern across the methods used. Engagement scores approached the limit of significance for the Active Involvement dimension, indicating panelists were least likely to feel distracted, zoned out or lose interest in the written protocol method. There were no significant differences in the time spent on the test by the panelists across the three methods. Asynchronous methods showed to be most suitable in terms of the convenience of the time of day at which the tests were completed, but showed no significant differences in other aspects of method practicality. Overall, a written protocol method of remote consumer testing is recommended, as it is less time-consuming for researchers while providing similar acceptance and engagement as other methods.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  Cookies; Engagement; Live Zoom; Remote testing; Video-guided; Written protocol

Year:  2022        PMID: 36168447      PMCID: PMC9499737          DOI: 10.1016/j.foodqual.2022.104616

Source DB:  PubMed          Journal:  Food Qual Prefer        ISSN: 0950-3293            Impact factor:   6.345


Introduction

In the context of the COVID-19 pandemic, it has become challenging for sensory scientists to conduct in-person sensory tests, particularly large central location tests. The COVID-19 virus is thought to spread through respiratory droplets between people who are within 6 feet from each other (Centers for Disease Control and Prevention, 2020), even in people who are asymptomatic. The use of masks and social distancing have been enforced in public settings, and consequently, many businesses have transferred to remote/online platforms. Given the fact that sensory tests usually involve panelists smelling and tasting food samples, the food industry has had to quickly overcome many challenges to be able to conduct sensory tests. Home-use tests (HUT) offer a good alternative to in-person testing; however, they present numerous challenges including the inability to control experimental conditions as well as the limitations in terms of shelf life and packaging. Many research studies have compared sensory perception and consumer acceptance of different samples in central location test (CLT) and HUT. Some of these studies show no effect of testing location (Lee and Lee, 2021, Schouteten et al., 2021), whereas others have observed a significant effect (Schouteten et al., 2019, Sveinsdóttir et al., 2010, Zhang et al., 2020). Inconsistencies shown in previous literature may be due to the type of products tested and also to the level at which a given food or beverage product is associated to a specific context in consumer’s minds. A previous study comparing HUT, CLT, and laboratory test found no difference in attribute liking, but did find differences in the perceived intensity of certain attributes (Pound, Duizer, & McDowell, 2000). Another study that compared CLT and HUT found different results depending on the product type and its usual context of consumption (Boutrolle, Delarue, Arranz, Rogeaux, & Köster, 2007). Previous studies that compare these two methods show no consensus about how each methodology affects sample ratings. Some studies have even shown that consumers display a higher discrimination (McDaniel & Sawyer, 1981) and are more critical of certain attributes (Miller et al., 1995) in home panels compared to laboratory panels. Another study showed that CLT yield more robust results than HUT, and lower average liking scores (Boutrolle, Arranz, Rogeaux, & Delarue, 2005). Recent studies that have adopted the approach of online/remote testing have focused on viewing photographs instead of tasting samples (Oliveira e Silva, do Carmo Rouxinol, & da Silva Coutinho Patarata, 2020), so comparison to previous literature is limited. This scientific discipline has had to find a way to suddenly shift its standard practices, so it is valuable for the food industry to have a method comparison between supervised/synchronous and unsupervised/asynchronous online consumer tests. It is essential to collect data on sensory engagement and consumer behavior in each of the studied methods to evaluate which type of treatment assists panelists best in their evaluation of samples. This involves panelists not getting distracted or losing interest in the sample evaluation, as well as finding the task meaningful and enjoyable. Previous research on instructional delivery suggests that the most effective method of increasing engagement involves instructor visibility through interactive video conferencing (Carr, 2014). Sensory protocol also determines that testing conditions and execution can be a source of variability if there is a lack of control (Moskowitz, Munoz, & Gacula, Maximo, 2003). The objective of this study was to compare three methods of online consumer testing in terms of sample perception, sensory engagement, and method practicality. It was hypothesized that a live, facilitated method of remote consumer testing will result in: 1) less variability across panelists, due to testing conditions being monitored live by the researcher; 2) a higher degree of sensory engagement due to a lower likelihood of losing interest during the task, and 3) a lower degree of practicality overall, due to the inconvenience of being scheduled for a testing session, in comparison to asynchronous written and video-guided methods of remote testing.

Material and methods

Samples and subjects

Five commercially available chocolate-chip cookie products were selected as test samples. This product category was selected due to its popularity and wide product variation. All five cookie products contained semi-sweet chocolate chips. More sample information can be seen in Table 1 .
Table 1

Product information for the five cookie products tested.

SampleWeight per cookie (g)Calories per cookieType of brandPrice per 357–397 g of cookies ($)
C11470National9.58
C21153.3Generic2.29
C31470National6.58
C41580National3.49
C51153.3National2.89
Product information for the five cookie products tested. Eighty-four consumers were recruited through email, social media, and university campus newsletters. The participation requirements were: 1) be 18 years of age or older, 2) have no known food/beverage allergies, 3) have no dietary restrictions, 4) not be pregnant or nursing, 5) not have any pre-existing or new smell/taste disorders, 6) have access to the Internet, a computer with a working microphone and camera, and a Zoom account, 7) be a frequent consumer of cookies, which was defined as consuming cookies more than once a month on average, 8) be on or near the University of Illinois Urbana-Champaign (UIUC) campus to pick up sample kits, and 9) be available for four remote sessions. Requirement #5 (not have pre-existing or new smell/taste disorders), was included as a means to screen out any consumers that may have been infected with COVID-19 or may still be recovering from the infection. Recent studies have demonstrated that loss of taste and smell are strong predictors of COVID-19 infection (Dawson et al., 2021, Menni et al., 2020, Printza and Constantinidis, 2020). In addition to meeting the screening requirements, panelists had to show proof of a recent negative COVID test in order to enter the building where they were instructed to pick up their sample kit. Due to the number of strict participation requirements, the researchers were only able to recruit eighty-four panelists. However, there is a certain range for subject number in published literature. Depending on the experimental design and nature of each consumer test, the number of participants varies from around eighty (Giacalone et al., 2019, Hall et al., 2003) to several hundred (Cardello et al., 2022, Crisosto et al., 2003). The demographic characteristics of this study's participants can be seen in Table 2 .
Table 2

Demographic information of subjects.

Count (n = 84)Percentage
GenderFemale6476.2
Male1922.6
Non-binary / third gender11.2
Prefer not to say00.0
Age18–25 years old1619.0
26–35 years old2327.4
36–45 years old2529.8
46–55 years old1416.7
56–65 years old44.8
66–75 years old22.4
76 years old or older00.0
EthnicityHispanic or Latino33.6
Not Hispanic or Latino8196.4
I don't know / I prefer not to say00.0
Racial backgroundAmerican Indian or Alaskan Native00.0
Asian or Asian American1720.2
Black or African American11.2
Native Hawaiian or Pacific Islander00.0
White or Caucasian6273.8
Mixed or Other22.4
I don't know / I prefer not to say22.4
Frequency of consumption of cookies (any type)Daily1011.9
Weekly4756.0
Monthly2732.1
Frequency of consumption of chocolate-chip cookiesDaily33.6
Weekly2631.0
Monthly4553.6
Every four to six months1011.9
Once a year or less00.0
Never00.0
Demographic information of subjects. Panelists received a gift card payment for their participation in the study. This research was reviewed and approved by the Institutional Review Board at the University of Illinois Urbana-Champaign, and informed consent was obtained from each subject prior to their participation in the study.

Test variables

The study consisted of four remote sessions in a time span of two weeks. The first session did not involve tasting; it was an introductory session aimed at providing an overview of the testing sessions and ensuring the panelists’ cameras were in working condition. Additionally, the researcher commented on general instructions and expectations of the study and provided details on how to pick up the sample kits. Lastly, each panelist signed a consent form and completed a demographic questionnaire. This was the first session for all panelists, and it was held on the video-conferencing platform Zoom (Zoom Video Communications, Inc., San Jose, CA, USA). All electronic questionnaires were completed on Qualtrics (Qualtrics, Provo, UT, USA), an online survey software. The other three sessions involved the sensory evaluation of the same five cookie samples, with different instructional delivery methods. The sessions were randomized across panelists, and samples were labeled with different three-digit codes in each session. The first type of testing session was a synchronous, facilitated session through Zoom. In this session, the researcher provided live, step-by-step instructions to all panelists on completing the sensory evaluation of the samples. This included a brief welcome, an explanation of how to follow the rinse protocol, a description of the hedonic scale, and a short introduction to each evaluation question. The researcher tried adapting the pace of the test to all panelists. Given the fact that it was required that all subjects had their cameras on during the entire session, it was possible to know when most people had finished tasting each sample. Participants were free to ask questions at any time during the session. Panelists were scheduled for one of nine live testing sessions according to their availability, which means there were between seven and eleven panelists in each live Zoom session. The researcher followed the same script in all sessions. The second session type was an asynchronous video-guided session. Panelists received their questionnaire link by email and were given 24 h to complete the sensory test at their preferred time of day. This questionnaire had short, embedded videos that were meant to walk the panelists through the sensory evaluation process. The researcher was the same as for the live sessions, and the videos were recorded in the same background as the one used for the live Zoom sessions. The researcher was also wearing a lab coat and gloves in all videos and all live sessions, to maintain consistency and to ensure that the delivery method was the only variable. The third and last type of delivery method was an asynchronous, fully written protocol. The questionnaire was the same as the other two sessions, but this time the instructions were typed in detail, with no visual aids such as videos or interactive conference calls. The typed instructions were the same as the ones provided verbally by the researcher in the live session, and taped in the video-guided session.

Test protocol

Cookie products were purchased from nearby grocery stores and opened the same day the sample kits were prepared. One cookie from each one of the five brands used was placed in 163 mL clear plastic cups with lids (Dart Container Corporation, Mason, MI, USA), labeled with random three-digit codes. Five unsalted crackers were placed in a separate cup. The six cups were placed in a gallon resealable plastic bag (Ziploc, S.C. Johnson & Sons Inc., Racine, WI, USA) together with a 207 mL wax-coated paper cup (Solo Cup Company, Lake Forest, IL, USA) and a 9-cm × 12-cm white dispenser napkin. One bag was prepared for each panelist two days before each one of their testing sessions. The day after the sample bags were prepared, panelists picked them up from a designated room in a building on the university campus. Panelists were given a 10-hour window to pick up their sample bags on each one of the three designated pick-up days. The next day, they would have one of their three testing sessions (live, video-guided, or written protocol). The process was repeated until all panelists had completed all three session types, which were randomized across the panelists.

Sensory evaluation procedure

Panelists were instructed to complete each test in one sitting, so they were asked not to begin the test until they were ready to complete it. They were asked to have all the necessary materials in front of them before beginning the test, and, if possible, to choose a quiet, distraction-free environment to complete the session. They were also reminded of the importance of matching the code in the screen to the code in the cup for each question. Panelists were also instructed to follow the rinse protocol before tasting each cookie sample. The rinse protocol consisted of biting half a cracker, chewing and swallowing, and then taking a sip of room temperature water. They were also instructed to not consume the whole sample, but to leave approximately half for re-tasting. After typing in their start time, panelists would rate each of the five samples on overall liking on a modified 9-point hedonic scale, with 1 being ‘dislike extremely’, 9 being ‘like extremely’, and 5 being ‘neither like nor dislike’. The order of appearance of the overall liking questions was randomized among panelists. After completing this set of questions, panelists moved on to evaluating each sample on liking of aroma, taste, and texture, also on a 9-point hedonic scale. To do this, panelists re-tasted each sample. After typing in their end time, panelists completed a validated engagement questionnaire (Hannum & Simons, 2020), where they rated their level of agreement or disagreement with a series of ten statements measuring three different dimensions of sensory engagement on a 7-point scale. Panelists also answered questions about their impressions of the practicality of each one of the methods used. Lastly, panelists were given the opportunity to type in any comments about the test in a text box. The full research questionnaire is available as supplementary material.

Data analysis

XLSTAT Sensory version 2019.1.1 (Addinsoft, NY, USA) was the data analysis software used to identify significant differences across samples evaluated using different remote testing methods. Analysis of Variance (ANOVA) was used to analyze consumer ratings with ‘panelist’ as a random factor and ‘testing method’, and ‘sample’ as fixed factors. Interaction among test types and samples was explored for significant impact on liking scores. Post hoc tests were conducted by the Fisher Least Significant Difference (LSD) method, with a confidence interval of 95%. To group subjects into statistically similar clusters according to their overall liking ratings of each product, Agglomerative Hierarchical Clustering (AHC) was used with the Euclidean distance for the dissimilarity scale by Ward’s method. The cluster truncation was automatic based on entropy. Data from the engagement questionnaire were averaged to generate three types of dimensional scores for each type of testing method: 1) Active Involvement, 2) Purposeful Intent, and 3) Affective Value. The questions that were part of the Active Involvement dimension were reverse coded so that high values would indicate an equal response for each item. A two-way ANOVA with ‘panelist’ and ‘testing method’ as main effects was conducted for each dimension to determine any significant effects of the testing method on sensory engagement. Agglomerative Hierarchical Clustering was also conducted on engagement scores to categorize panelists according to their level of overall engagement in each one of the three testing methods. The scores for each engagement dimension were averaged to calculate an average factor score. Analysis of Variance was run on the data extracted from each cluster, and post-hoc Fisher LSD tests were conducted to identify significant differences within (not across) each cluster. Clusters formed by a single panelist were combined in a logical way following the corresponding dendrogram. Ratings from the method practicality questionnaire were analyzed using ANOVA, with ‘panelist’ as a random factor, and ‘method’ and ‘time of day’ as fixed factors. ‘Time of day’ was included as a factor in order to determine any differences in liking according to the time chosen by panelists to complete the test. Fisher LSD was conducted to resolve the direction of any significant effects and was set to a significance level of 0.05.

Results

Product acceptance

Based on the Analysis of Variance shown in Table 3 , the variable ‘testing method’ did not bring significant information to explain the variability of the dependent variables of overall liking, or liking of any of the specific attributes tested. There was also no significant interaction of ‘testing method’ with ‘sample’, as can be seen in Fig. 1 , which displays acceptance scores of each sample in the different modalities used. The average acceptance scores for each sample can be seen in Fig. 2 , where on average, sample C3 was liked the most, and samples C2 and C4 were liked the least.
Table 3

Analysis of Variance (ANOVA) F-values of overall liking and specific attribute scores.

Attribute LikingPanelist (P)df = 83Testing Method (M)df = 2Sample (S)df = 4M*Sdf = 8
Overall4.26***0.8722.87***0.39
Aroma5.92***2.2722.69***0.43
Taste4.86***0.4629.68***0.64
Texture6.73***1.3616.59***0.26

*, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom.

Fig. 1

Average overall liking ratings on 9-pt hedonic scale for each sample tested in each of the three testing modalities. Sample*Method interaction not significant (p = 0.925).

Fig. 2

Average overall liking ratings on 9-pt hedonic scale for each sample tested. Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference).

Analysis of Variance (ANOVA) F-values of overall liking and specific attribute scores. *, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom. Average overall liking ratings on 9-pt hedonic scale for each sample tested in each of the three testing modalities. Sample*Method interaction not significant (p = 0.925). Average overall liking ratings on 9-pt hedonic scale for each sample tested. Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference). Panelists were clustered according to their overall liking ratings across testing methods. The results of the AHC analysis (Fig. 3 ) identified three groups of panelists from their taste preferences, which varied slightly across testing methods. The major cluster in every testing method (Cluster 3) had a similar level of acceptance for all samples, where the average overall liking rating was approximately between 6 and 7.5. The major cluster represented 39% of panelists in the live Zoom method, 54% in the video-guided method and 56% in the written protocol method. Cluster 2 was also similar in all methods, where approximately a third of the panelists rated samples C1 and C3 higher than the rest of the samples. The testing methods varied primarily in their smallest cluster (Cluster 1). In the written protocol session, panelists in Cluster 1 displayed a high degree of hedonic discrimination across samples, with C1 being the least liked and C5 being the most liked. The smallest cluster in the live Zoom testing session exhibited a low degree of hedonic discrimination among samples C2, C3 and C4, with C1 remaining the least liked and C5 the most liked. Lastly, the smallest cluster in the video-guided session was characterized by panelists who liked sample C3 the most and sample C1 the least. Therefore, Cluster 1 in each of the analyses represented a percentage of consumers who did change their preferences based on the testing modality, which ranged from 11% in the written protocol method to 27% in the live Zoom method.
Fig. 3

Agglomerative Hierarchical Clustering of overall liking scores. C1-C5 = Cookie samples 1 through 5 for a) Live Zoom, b) Video-Guided, and c) Written Protocol methods. Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference).

Agglomerative Hierarchical Clustering of overall liking scores. C1-C5 = Cookie samples 1 through 5 for a) Live Zoom, b) Video-Guided, and c) Written Protocol methods. Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference).

Sensory engagement

Results of the Engagement Questionnaire (EQ) indicated that sensory engagement was not significantly affected by the testing method, as shown in Table 4 . The ‘Active Involvement’ dimension closely approached the limit of significance (p = 0.069), where the written protocol method received higher engagement scores (mean of 6.59) than the other methods (mean of 6.51 for live Zoom and 6.31 for video-guided).
Table 4

Analysis of Variance (ANOVA) F-values Engagement scores.

Engagement DimensionQuestions included in each dimensionPanelist (P)df = 83Testing Method (M)df = 2
Active InvolvementI lost interest in the task2.38***2.72
I was distracted
I felt myself zoning out during the task
Purposeful IntentI found the task meaningful5.57***1.23
I felt dedicated to finish the task
I wanted to devote my full attention to the task
My contribution was significant to the outcome of the task
Affective ValueI found the task captivating6.79***0.88
During the task, I was enjoying myself
I was motivated to expend extra effort during the task

*, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom; ‡ indicates a question was reverse coded for analysis.

Analysis of Variance (ANOVA) F-values Engagement scores. *, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom; ‡ indicates a question was reverse coded for analysis. When panelists were clustered according to their level of overall engagement in each testing method, three clusters appeared (Fig. 4 ). The largest cluster (Cluster 3) was characterized by panelists who displayed a high level of engagement in all methods, with a mean of 6.7 out of 7. In the second to largest cluster (Cluster 2), panelists also displayed a similar degree of engagement in all three testing methods, with the average score being lower than the largest cluster (5.9 out of 7). The last cluster (Cluster 1), comprised of only seven panelists, displayed a higher engagement in the live Zoom session compared to the other two sessions, though not significantly higher upon Fisher LSD analysis. Cluster 1 displays large error bars given the small number of panelists included in this cluster. It also demonstrates that only a small percentage of the subjects were not as engaged during their participation in this study. The majority of panelists displayed high sensory engagement scores overall.
Fig. 4

Agglomerative Hierarchical Clustering of engagement scores (1 = strongly disagree to 7 = strongly agree). Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference).

Agglomerative Hierarchical Clustering of engagement scores (1 = strongly disagree to 7 = strongly agree). Means labeled with the same letter indicate no significant differences in overall liking (Fisher Least Significant Difference).

Method practicality

The three testing methods proved to be similar regarding overall practicality (Table 5 ). The only variable that showed a significant difference across methods was the ‘convenience of the time of day’. Both asynchronous methods (video-guided and written protocol) were rated significantly higher in convenience than the synchronous method through Zoom. The time of day that showed to be most convenient for panelists to complete the remote consumer test was the evening, with an average convenience rating of 6.8 out of 7.
Table 5

Analysis of Variance (ANOVA) F-values of practicality questionnaire.

Practicality QuestionPanelist (P)df = 83Testing Method (M)df = 2Time of Day (T)df = 3M*Tdf = 4
Easiness of following instructions1.100.070.740.29
Clarity of instructions1.360.110.571.47
Adequacy of session length1.071.140.721.37
Convenience of time of day2.12***15.09***3.79*0.90
Rating of experience1.90***0.790.930.49

*, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom.

Analysis of Variance (ANOVA) F-values of practicality questionnaire. *, **, and *** indicate significance at p < 0.05, p < 0.01 and p < 0.001, respectively; df = degrees of freedom. Fisher LSD analysis showed that the time spent on each session was not significantly different across the three methods, with an average of 13.8 min spent on each questionnaire.

Discussion

Acceptance of the cookie products did not differ across the three remote testing methods. This outcome was expected, as previous research from this laboratory has shown acceptance to be rather robust and absolute, and not greatly be affected by context effect (Albiol Tapia, Baik, & Lee, 2021), which in this study was created by the different delivery methods. There were also no differences in overall liking of samples across session orders. Comparison to previous literature is limited, as not many research studies have explored different delivery methods of remote consumer testing with the same samples. Most studies focus on comparing test locations, such as comparing CLT with HUT. Home Use Test, in comparison to CLT, is generally characterized by natural (though uncontrolled) eating conditions, pleasant ambiance, consumption of a larger amount of food, repeated exposures, and a multi-session monadic sequential presentation of products (Boutrolle et al., 2007). Additionally, HUT allow consumers to choose the moment of consumption as well as the way the product is consumed, the possibility of a social experience and the chance of allowing post-ingestional effects before providing an opinion of the food (Boutrolle et al., 2007). In this study, the live Zoom method had aspects of a CLT in that the researcher could ensure that the samples were being tasted and the rinse protocol was being followed, since panelists were performing the sensory evaluation live on camera. In addition, panelists were not completely free to choose the time of day to complete the test, as they were scheduled for a session ten days prior, based on their reported availability in the screening questionnaire. The two other methods (video-guided and written protocol sessions) can be described as closely following the protocols of a HUT, in the sense that the consumers are able to choose an appropriate time of day to consume the food, and the conditions in which they decide to complete the sensory evaluation are completely uncontrolled. Previous studies comparing CLT to HUT have resulted in different outcomes. For products like fermented milk beverages (Boutrolle et al., 2005), salted crackers and sparkling water (Boutrolle et al., 2007), HUT results showed higher acceptance ratings than CLT. Studies testing other snack products showed product-specific results, where chips were liked more in HUT, juice was liked more in CLT, and yogurt showed no contextual differences (Schouteten et al., 2019). In another study, consumers that tested cod also averaged higher liking results in HUT setting, although the scores were comparable when the cooking method was similar to the method used in CLT (Sveinsdóttir et al., 2010). No significant differences in sensory engagement were found across testing methods. A potential reason for this outcome is that the testing methods were very standardized, with the delivery of instructions being the only difference. This was designed purposefully to test the delivery method as the only variable. The ‘Active Involvement’ dimension, which closely approached the limit of significance, showed that overall, panelists felt less ‘zoned out’ in the written protocol session. This may be because to receive the instructions, panelists had to actively read text on their screens, as opposed to passively listen in the video-guided and live Zoom formats. Once the panelist was familiar with the sensory evaluation dynamics, it may have been easier to lose focus in the video-guided and live sessions. The ability to skim or even skip instructions is also higher in the written protocol session compared to the video-guided session. Panelists may have opted for reading at a faster pace after being familiarized with the instructions. When looking solely at engagement scores for panelists’ first testing session, we observe lower scores in the ‘purposeful intent’ dimension, which assesses panelists’ perceived personal relevance of the sensory evaluation. This effect was driven by the video-guided session, which received higher engagement scores when it was the second or third session for panelists. Previous research comparing different types of instructional delivery is mostly focused on student learning, especially in children. Some studies found that, during teacher-directed instruction, children were less likely to attend to instructions and to ask questions. During seat work, however, they were more likely to get disorganized (Stright & Supplee, 2010). In research on adults, our results conflict with two of Richard Mayer’s principles of multimedia design, which involve replacing visual text with spoken text (modality principle), and adding visual cues relating elements of a picture to the text (multimedia principle) (Mayer, 1999). These principles have been shown to increase the effectiveness of multimedia instructions regarding less mental effort spent on understanding explanations. A study testing these principles on college students, however, found lower retention and transfer scores when visual text was replaced by spoken text (Tabbers, Martens, & van Merriënboer, 2004). This study argued that visual text may be more appropriate for presenting procedural information (which in the case of the present study, can apply to the instructions on how to proceed with the sensory test) than spoken test, because the subject has more time to reflect on the information presented. It is likely that a written protocol method accompanied with visual images of the instructions would have resulted in higher engagement than the video-guided method. Another reason for the lower engagement scores in the live Zoom method could be that panelists disliked being on camera, or were distracted by watching other panelists on camera. Several panelists reported this at the end of their questionnaires, as there was a text box for them to type any comments they had about each session. The videoconferencing software used does not offer an option where the host can view all participants’ video, but the participants can only view the host’s video. This feature would be ideal to use in remote consumer tests, as it would allow the researcher to monitor test performance while preventing the bias that comes from panelists viewing each other on camera. It was also reported that the videos were somewhat repetitive, and that a written protocol encouraged the panelists to focus more. Additionally, ‘Zoom fatigue’ could also explain why engagement was not higher in the live session. Videoconferencing technologies have allowed students to continue their education remotely during the pandemic and have also enhanced our ability to connect when in-person gatherings were discouraged. However, the prolonged use of this technology has also reported drawbacks. Examples include a greater need to concentrate, difficulties relaxing into natural conversations, expectations of instant responses, and even increased self-evaluation from continuously staring at video of oneself, which could lead to loss of self-esteem (Bailenson, 2021, Wiederhold, 2020, Williams, 2021). Given that the only significant difference found in the analysis of method practicality questions was the convenience of the time of day, the authors recommend an asynchronous method for remote consumer tests of products like cookies, which do not need any form of cooking and can be easily portioned into sample cups. The evening, defined by the researchers as 6 pm to midnight, resulted in the highest convenience ratings, with a mean score of 6.76 out of 7. Table 6 shows panelists’ choices for the time of day to complete each asynchronous testing session. It is notable that, even though more panelists chose the morning to complete the test, the evening received higher convenience ratings.
Table 6

Choice of time of day for the completion of asynchronous sessions.

Asynchronous Testing MethodChosen time of dayNumber of panelistsPercentage
Video-GuidedMorning3744
Afternoon3036
Evening1619
Night11
Written ProtocolMorning3238
Afternoon2631
Evening2530
Night11

Morning: 6 am–12 pm, Afternoon: 12 pm–6 pm, Evening: 6 pm–12 am, Night: 12 am–6 am.

Choice of time of day for the completion of asynchronous sessions. Morning: 6 am–12 pm, Afternoon: 12 pm–6 pm, Evening: 6 pm–12 am, Night: 12 am–6 am. Even though the focus of this study is to maximize the method practicality for the subjects, it is worth noting that the practicality for the sensory scientist conducting the research is also much higher in a fully written protocol method. To prepare for this session, the researcher is not required to spend time pre-recording, repeating, editing, uploading and embedding videos in the research questionnaires, nor does the researcher have to schedule panelists according to their availability. In addition, it seems that panelists did not rush the evaluations in the asynchronous sessions, as the average time spent on each session did not vary significantly. Overall, a written protocol session that includes some images supporting the instructions would be an engaging, practical, and effective method of remote consumer testing.

Conclusions

The method of remote consumer testing that proved to be the most practical in this study was an asynchronous, fully written protocol method. With similar product acceptance ratings as well as sensory engagement scores, this method scored highest in convenience of time of day. A written protocol method also has the most benefits for the sensory scientist, who will not need to schedule panelists for live videoconferencing or pre-record instructional videos to embed them in the research questionnaires. These findings allow sensory scientists in academia and industry to justify the use of asynchronous methods of remote consumer testing during the pandemic. This study also proves that safety can be prioritized over controlled conditions of testing, and that remote testing using a detailed, written protocol can be considered an alternative until in-person testing is deemed safe again. The main limitation of this study is that results were not compared to different methods of in-person testing, so it is unclear whether acceptance, engagement and practicality would be comparable when tasting chocolate-chip cookies in an in-person setting. In addition, the practicality questions that were used in this study are not part of a validated questionnaire, so it is unclear whether the questions chosen by the researchers were the most appropriate to address subjects’ practicality experiences. Lastly, the size of the population used in this research study is also limited (n = 84). The objective of this research study is methodological in nature, and as such not meant to be representative of the general population. Additionally, published literature ranges widely in participant number for consumer tests from around 80 to several hundred, so our sample size fits within the published range. Future research can include an additional method consisting of a written protocol with visual images (still or animated) that aid in the description of the instructions. Ultimately, a comparison of scores of in-person and remote testing will be key in determining whether these methods can be comparable for different food products.

CRediT authorship contribution statement

Marta Albiol Tapia: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Resources, Visualization, Writing – original draft. Soo-Yeun Lee: Conceptualization, Methodology, Resources, Supervision, Writing – review & editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  8 in total

1.  Multimedia instructions and cognitive load theory: effects of modality and cueing.

Authors:  Huib K Tabbers; Rob L Martens; Jeroen J G van Merriënboer
Journal:  Br J Educ Psychol       Date:  2004-03

2.  Influence of organic labels on consumer's flavor perception and emotional profiling: Comparison between a central location test and home-use-test.

Authors:  Joachim J Schouteten; X Gellynck; H Slabbinck
Journal:  Food Res Int       Date:  2018-09-18       Impact factor: 6.475

3.  Connecting Through Technology During the Coronavirus Disease 2019 Pandemic: Avoiding "Zoom Fatigue".

Authors:  Brenda K Wiederhold
Journal:  Cyberpsychol Behav Soc Netw       Date:  2020-06-18

4.  Comparison of a central location test versus a home usage test for consumer perception of ready-to-mix protein beverages.

Authors:  M T Zhang; Y Jo; K Lopetcharat; M A Drake
Journal:  J Dairy Sci       Date:  2020-02-20       Impact factor: 4.034

5.  Loss of Taste and Smell as Distinguishing Symptoms of Coronavirus Disease 2019.

Authors:  Patrick Dawson; Elizabeth M Rabold; Rebecca L Laws; Erin E Conners; Radhika Gharpure; Sherry Yin; Sean A Buono; Trivikram Dasu; Sanjib Bhattacharyya; Ryan P Westergaard; Ian W Pray; Dongni Ye; Scott A Nabity; Jacqueline E Tate; Hannah L Kirking
Journal:  Clin Infect Dis       Date:  2021-02-16       Impact factor: 9.079

6.  The role of self-reported smell and taste disorders in suspected COVID‑19.

Authors:  Athanasia Printza; Jannis Constantinidis
Journal:  Eur Arch Otorhinolaryngol       Date:  2020-05-23       Impact factor: 2.503

7.  Real-time tracking of self-reported symptoms to predict potential COVID-19.

Authors:  Cristina Menni; Ana M Valdes; Claire J Steves; Tim D Spector; Maxim B Freidin; Carole H Sudre; Long H Nguyen; David A Drew; Sajaysurya Ganesh; Thomas Varsavsky; M Jorge Cardoso; Julia S El-Sayed Moustafa; Alessia Visconti; Pirro Hysi; Ruth C E Bowyer; Massimo Mangino; Mario Falchi; Jonathan Wolf; Sebastien Ourselin; Andrew T Chan
Journal:  Nat Med       Date:  2020-05-11       Impact factor: 53.440

8.  Comparison of Home Use Tests with Differing Time and Order Controls.

Authors:  Nahyung Lee; Jeehyun Lee
Journal:  Foods       Date:  2021-06-03
  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.