| Literature DB >> 25337457 |
Aaron Fisher1, G Brooke Anderson2, Roger Peng1, Jeff Leek1.
Abstract
Scatterplots are the most common way for statisticians, scientists, and the public to visually detect relationships between measured variables. At the same time, and despite widely publicized controversy, P-values remain the most commonly used measure to statistically justify relationships identified between variables. Here we measure the ability to detect statistically significant relationships from scatterplots in a randomized trial of 2,039 students in a statistics massive open online course (MOOC). Each subject was shown a random set of scatterplots and asked to visually determine if the underlying relationships were statistically significant at the P < 0.05 level. Subjects correctly classified only 47.4% (95% CI [45.1%-49.7%]) of statistically significant relationships, and 74.6% (95% CI [72.5%-76.6%]) of non-significant relationships. Adding visual aids such as a best fit line or scatterplot smooth increased the probability a relationship was called significant, regardless of whether the relationship was actually significant. Classification of statistically significant relationships improved on repeat attempts of the survey, although classification of non-significant relationships did not. Our results suggest: (1) that evidence-based data analysis can be used to identify weaknesses in theoretical procedures in the hands of average users, (2) data analysts can be trained to improve detection of statistically significant results with practice, but (3) data analysts have incorrect intuition about what statistically significant relationships look like, particularly for small effects. We have built a web tool for people to compare scatterplots with their corresponding p-values which is available here: http://glimmer.rstudio.com/afisher/EDA/.Entities:
Keywords: Data visualization; Education; Evidenced based data analysis; MOOC; Randomized trial; Statistical significance; Statistics; p-values
Year: 2014 PMID: 25337457 PMCID: PMC4203023 DOI: 10.7717/peerj.589
Source DB: PubMed Journal: PeerJ ISSN: 2167-8359 Impact factor: 2.984
Figure 1Examples of plots shown to users.
Plot categories shown to users.
| Reference | 100 data points (e.g., |
| Smaller | 35 data points |
| Larger | 200 data points |
| Best-fit line | 100 data points, with best fit line added |
| Lowess | 100 data points, with smooth lowess curve added (using R “lowess” function) |
| Axis Scale | 100 data points, with the axis range increased to 1.5 standard deviations outside |
| Axis Label | 100 data points, with fictional |
Figure 2Accuracy of significance classifications under different conditions.
Point estimates and confidence intervals for classification accuracy for each presentation style (Table 1). Accuracy rates for plots with truly significant underlying relationships (sensitivity) are shown in blue, and accuracy rates for plots with non-significant underlying relationships (specificity) are shown in red.
Figure 3Classification accuracy on repeat attempts of the survey.
Each plot shows point estimates and confidence intervals for accuracy rates of human visual classifications of statistical significance on the first and second attempt of the survey. For the truly significant underlying P-values, users showed a significant increase in accuracy (sensitivity) on the second attempt of the survey for the “Reference,” “Smaller n,” and “Best Fit” presentation styles. For non-significant underlying P-values, accuracy (specificity) decreased significantly for the “Smaller n” category. Because these accuracy rates were estimated only based on the data from students who submitted more than one response to the survey, the confidence intervals here are wider than those in Fig. 2.