| Literature DB >> 25876027 |
Christoph Bartneck1, Andreas Duenser2, Elena Moltchanova1, Karolina Zawieska3.
Abstract
Computer and internet based questionnaires have become a standard tool in Human-Computer Interaction research and other related fields, such as psychology and sociology. Amazon's Mechanical Turk (AMT) service is a new method of recruiting participants and conducting certain types of experiments. This study compares whether participants recruited through AMT give different responses than participants recruited through an online forum or recruited directly on a university campus. Moreover, we compare whether a study conducted within AMT results in different responses compared to a study for which participants are recruited through AMT but which is conducted using an external online questionnaire service. The results of this study show that there is a statistical difference between results obtained from participants recruited through AMT compared to the results from the participant recruited on campus or through online forums. We do, however, argue that this difference is so small that it has no practical consequence. There was no significant difference between running the study within AMT compared to running it with an online questionnaire service. There was no significant difference between results obtained directly from within AMT compared to results obtained in the campus and online forum condition. This may suggest that AMT is a viable and economical option for recruiting participants and for conducting studies as setting up and running a study with AMT generally requires less effort and time compared to other frequently used methods. We discuss our findings as well as limitations of using AMT for empirical studies.Entities:
Mesh:
Year: 2015 PMID: 25876027 PMCID: PMC4397064 DOI: 10.1371/journal.pone.0121595
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1A LEGO Minifigure.
Fig 2Questionnaire layout a) MT b) Qualtrics.
Comparison of the conditions on various aspects as listed in the leftmost column.
| MT | MT Batch | Campus | Online | |
|---|---|---|---|---|
| Country | USA | USA | New Zealand | Unknown |
| Location | Online | Online | Univ. Library | Online |
| Duration | ∼ 30 minutes | variable | ∼ 30 minutes | ∼ 30 minutes |
| System | Qualtrics | AMT | Qualtrics | Qualtrics |
| Task Type | Questionnaire | Questionnaire | Questionnaire | Questionnaire |
| Recruitment | AMT | AMT | in person | forums |
| Participants | variable (see [ | variable (see [ | Univ. students | Unknown |
| Sample size | 29 and 31 | 209 and 83 | 32 | 185 |
| Reward | Money | Money | None | None |
Number of responses by condition.
| min | max | median | mean | |
|---|---|---|---|---|
| MT | 94 | 94 | 94 | 94 |
| MT Batch | 1 | 94 | 6 | 14.5 |
| campus | 92 | 94 | 94 | 93.8 |
| online | 1 | 94 | 20 | 36.5 |
Fig 3Plate Diagram for the simplified Bayesian multinomial logistic random effects model.
Fig 4Estimated average distribution of emotional assignment by method (campus, online, MT, MT Batch).
The numbers of Minifigures, for which the three methods produce strongly contradictory results (ΔDIC ≥ 10), mildly contradictory results (5 < ΔDIC ≤ 10) and agree (ΔDIC ≤ 5).
The rows where two methods are listed test for the respective pairwise difference. The rows where three methods are listed test for at least one being different from the rest.
| Strongly Contradictory | Mildly Contradictory | Agree | |
|---|---|---|---|
| Δ | 5 < Δ | Δ | |
| MT Batch vs. Campus | 14 | 18 | 62 |
| MT Batch vs. Online | 19 | 19 | 56 |
| Campus vs Online | 4 | 18 | 72 |
| MT Batch vs. Online vs. Campus | 39 | 16 | 39 |
| MT vs. Campus | 33 | 19 | 42 |
| MT vs. Online | 28 | 25 | 41 |
| MT vs. Campus vs. Online | 39 | 13 | 42 |
| MT vs. MT Batch | 11 | 5 | 78 |
Fig 5Frequencies of emotional categories for two Minifigures
a) Minifigure 03753 b) Minifigure 00948