| Literature DB >> 28251868 |
Sophie Engle1, Sean Whalen2, Alark Joshi3, Katherine S Pollard2,4.
Abstract
BACKGROUND: Cluster heatmaps are commonly used in biology and related fields to reveal hierarchical clusters in data matrices. This visualization technique has high data density and reveal clusters better than unordered heatmaps alone. However, cluster heatmaps have known issues making them both time consuming to use and prone to error. We hypothesize that visualization techniques without the rigid grid constraint of cluster heatmaps will perform better at clustering-related tasks.Entities:
Keywords: Bioinformatics visualization; Data clustering; Hierarchy data; Qualitative evaluation; Quantitative evaluation; Systems biology/omics data
Mesh:
Year: 2017 PMID: 28251868 PMCID: PMC5333167 DOI: 10.1186/s12859-016-1442-6
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1Alternatives to cluster heatmaps. We used these 5 alternatives (in addition to cluster heatmaps) in our final study. All five alternatives depicted here are for the same dataset. From left to right: (a) gapmap [11, 17] (b) circle packing [24], (c) sunburst [21], (d) radial dendrogram, and (e) force-directed tree [20]. Leaf nodes are filled to indicate the original value from the data matrix using the PRGn ColorBrewer scheme [66]. Positive values are green in color, and negative values are purple in color. The root node, if depicted, is indicated by a black outline. Inner nodes have no fill color and a gray outline. Edge tapering is used to indicate parent-child relationships in the node-link diagrams [37]
Fig. 2Unboxing Approach. Illustrates how the data matrix is “unboxed” and embedded into the hierarchical clustering of a symmetric matrix. The process is similar for asymmetric matrices, except there are no redundant cells to remove. Top Left: A standard cluster heatmap of a correlation matrix. Top Middle: The cluster heatmap with non-redundant information highlighted. Top Right: The cluster heatmap without the redundant information. Bottom Left: Hierarchical clustering of the variables (rows and columns) from the cluster heatmap, shown as a dendrogram. Bottom Middle: Hierarchical clustering of values (individual cells) and variables, shown as a dendrogram. Bottom Right: Hierarchical clustering of values and variables, shown as a treemap
Fig. 3Mechanical Turk Example. Shows example images presented to Amazon Mechanical Turk participants for question 15 (see Table 1 for details). The question asked, “Which two of the highlighted elements are more closely clustered?” The correct answer is the pair K and S, since these clusters connect at a lower depth in the tree than cluster C
Mechanical Turk user study analysis
| Type | Task | Number | Question text | Nodes | Type | Mean |
| df |
| ||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Training | 1 | 1 | Is the highlighted cluster mostly positive or mostly negative? | Clusters | N/A | ||||||
| Training | 1 | 2 | What is the height of the tree? | N/A | N/A | ||||||
| Timed | 1 | 3 | Which of the highlighted elements has the highest value? | Leaves | Score | 0.380 | 4.578 | 5 | 4.695 | E-01 | |
| Timed | 1 | 4 | Is the highlighted cluster mostly positive or mostly negative? | Clusters | Score | 0.661 | 14.650 | 5 | 1.197 | E-02 | ∗ |
| Timed | 1 | 5 | What is the height of the tree? | N/A | Error | 3.083 | 23.278 | 5 | 2.987 | E-04 | ∗∗∗ |
| Training | 2 | 6 | Which of the highlighted elements is furthest away from the root? | Leaves | N/A | ||||||
| Training | 2 | 7 | Which of the highlighted elements are siblings? | Leaves | N/A | ||||||
| Timed | 2 | 8 | Which of the highlighted elements is furthest away from the root? | Leaves | Score | 0.605 | 63.540 | 5 | 2.250 | E-12 | ∗∗∗ |
| Timed | 2 | 9 | Which of the highlighted elements is furthest away from the root? | Clusters | Score | 0.732 | 14.705 | 5 | 1.170 | E-02 | ∗ |
| Timed | 2 | 10 | Which of the highlighted elements are siblings? | Clusters | Score | 0.864 | 21.662 | 5 | 6.070 | E-04 | ∗∗∗ |
| Training | 3 | 11 | Which two of the highlighted elements are more closely clustered? | Clusters | N/A | ||||||
| Training | 3 | 12 | How many visually distinct clusters do you see in this visualization? | N/A | N/A | ||||||
| Timed | 3 | 13 | Which two of the highlighted elements are more closely clustered? | Siblings | Score | 0.738 | 29.499 | 5 | 1.850 | E-05 | ∗∗∗ |
| Timed | 3 | 14 | Which two of the highlighted elements are more closely clustered? | Leaves | Score | 0.275 | 12.775 | 5 | 2.558 | E-02 | ∗ |
| Timed | 3 | 15 | Which two of the highlighted elements are more closely clustered? | Clusters | Score | 0.352 | 9.539 | 5 | 8.941 | E-02 | · |
| Timed | 3 | 16 | Which of the highlighted elements is least similar to its neighbors? | Clusters | Score | 0.283 | 13.726 | 5 | 1.745 | E-02 | ∗ |
| Timed | 3 | 17 | How many visually distinct clusters do you see in this visualization? | N/A | Value | 8.794 | 31.138 | 5 | 8.796 | E-06 | ∗∗∗ |
Shows the type of question (training or timed), the task set, question number and text, the node type of the choices (leaf nodes, sibling nodes, or cluster nodes), the value type (score, absolute error, or raw value), overall average, and results (χ 2-test statistic, degrees of freedom, and p-value) from the per-question Kruskal-Wallis tests by technique. See Figs. 4 and 5 for the distribution of values for these questions broken down by technique
Legend: *** p≤ 0.001, ** p≤ 0.01, * p≤ 0.05, . p≤ 0.1
Fig. 4Mechanical Turk Average Scores. Shows the distribution of scores for clustering-related questions. The circle and black bar indicate the average score and standard error for each technique. The dotted line indicates random performance (not including the “Unsure” option). Our analysis shows that cluster heatmaps perform worse on average on question 13 than sunbursts, radial dendrograms, and force directed trees. Cluster heatmaps perform better on average than radial dendrograms in question 14. Question 15 did not have statistically significant differences in our post-hoc analysis. Gapmap performed better than radial dendrograms in question 16. Performance is worse than random for certain techniques in questions 14, 15, and 16—indicating their difficulty. However, there is at least one technique for each of these questions where the performance is better than random. See Tables 1 and 2 for more details
Fig. 5Mechanical Turk Average Clusters. Shows the distribution of responses to question 17, “How many visually distinct clusters do you see in this visualization?” The circle and black bar indicate the mean estimate and standard error for each technique. Our analysis shows that cluster heatmaps and sunbursts produce lower estimates on average compared to the other techniques. See Tables 1 and 2 for more details
Mechanical Turk post hoc analysis
| Q13 | Q14 | Q15 | Q16 | Q17 | |||||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Technique Pairs | est | err |
|
| est | err |
|
| est | err |
|
| est | err |
|
| est | err |
|
| |||||
| Cluster Heatmap, Gapmap | 1.056 | 0.542 | 1.949 | 0.361 | –0.274 | 0.503 | –0.546 | 0.994 | –0.186 | 0.497 | –0.374 | 0.999 | 0.916 | 0.531 | 1.726 | 0.511 | 3.121 | 0.945 | 3.303 | 0.014 | ∗ | ||||
| Cluster Heatmap, Circle Packing | –0.065 | 0.504 | –0.128 | 1.000 | 0.657 | 0.525 | 1.252 | 0.808 | 0.658 | 0.512 | 1.286 | 0.792 | 0.405 | 0.601 | 0.675 | 0.984 | –3.745 | 0.960 | –3.901 | 0.002 | ∗∗ | ||||
| Cluster Heatmap, Sunburst | 2.708 | 0.811 | 3.338 | 0.010 | ∗ | –1.191 | 0.571 | –2.086 | 0.290 | –1.159 | 0.537 | –2.160 | 0.256 | –0.292 | 0.579 | –0.504 | 0.996 | 0.205 | 0.952 | 0.215 | 1.000 | ||||
| Cluster Heatmap, Radial Dendrogram | 2.269 | 0.702 | 3.232 | 0.015 | ∗ | –2.032 | 0.698 | –2.912 | 0.041 | ∗ | –1.159 | 0.537 | –2.160 | 0.256 | –0.965 | 0.662 | –1.457 | 0.688 | 3.003 | 0.960 | 3.128 | 0.025 | ∗ | ||
| Cluster Heatmap, Force Directed Tree | 1.504 | 0.573 | 2.623 | 0.087 | · | –0.903 | 0.533 | –1.694 | 0.530 | –1.041 | 0.524 | –1.989 | 0.348 | 0.550 | 0.529 | 1.040 | 0.903 | 3.455 | 0.938 | 3.683 | 0.004 | ∗∗ | |||
| Gapmap, Circle Packing | 0.992 | 0.546 | 1.817 | 0.443 | 0.383 | 0.538 | 0.711 | 0.980 | 0.473 | 0.516 | 0.916 | 0.942 | 1.322 | 0.581 | 2.275 | 0.202 | –0.624 | 0.960 | –0.650 | 0.987 | |||||
| Gapmap, Sunburst | 1.652 | 0.838 | 1.972 | 0.347 | –0.916 | 0.583 | –1.571 | 0.612 | –0.973 | 0.541 | –1.801 | 0.464 | –1.208 | 0.559 | –2.163 | 0.252 | –2.917 | 0.952 | –3.063 | 0.030 | ∗ | ||||
| Gapmap, Radial Dendrogram | 1.213 | 0.732 | 1.656 | 0.550 | –1.758 | 0.708 | –2.483 | 0.126 | –0.973 | 0.541 | –1.801 | 0.464 | –1.881 | 0.644 | –2.921 | 0.040 | ∗ | –0.118 | 0.960 | –0.123 | 1.000 | ||||
| Gapmap, Force Directed Tree | –0.448 | 0.610 | –0.734 | 0.977 | 0.629 | 0.546 | 1.151 | 0.857 | 0.856 | 0.528 | 1.622 | 0.583 | 0.366 | 0.506 | 0.724 | 0.979 | –0.333 | 0.938 | –0.355 | 0.999 | |||||
| Circle Packing, Sunburst | 2.644 | 0.814 | 3.248 | 0.014 | ∗ | –0.533 | 0.603 | –0.885 | 0.949 | –0.501 | 0.555 | –0.903 | 0.946 | 0.113 | 0.625 | 0.181 | 1.000 | –3.540 | 0.967 | –3.660 | 0.004 | ∗∗ | |||
| Circle Packing, Radial Dendrogram | 2.204 | 0.705 | 3.127 | 0.021 | ∗ | –1.375 | 0.724 | –1.899 | 0.397 | –0.501 | 0.555 | –0.903 | 0.946 | –0.560 | 0.703 | –0.796 | 0.968 | –0.742 | 0.975 | –0.761 | 0.974 | ||||
| Circle Packing, Force Directed Tree | 1.440 | 0.577 | 2.495 | 0.119 | –0.246 | 0.567 | –0.433 | 0.998 | –0.383 | 0.542 | –0.707 | 0.981 | 0.956 | 0.579 | 1.650 | 0.561 | –0.290 | 0.953 | –0.305 | 1.000 | |||||
| Sunburst, Radial Dendrogram | 0.439 | 0.949 | 0.463 | 0.997 | 0.842 | 0.758 | 1.110 | 0.875 | 0.000 | 0.577 | 0.000 | 1.000 | 0.673 | 0.685 | 0.983 | 0.922 | –2.798 | 0.967 | –2.893 | 0.048 | ∗ | ||||
| Sunburst, Force Directed Tree | 1.204 | 0.859 | 1.402 | 0.716 | –0.288 | 0.610 | –0.472 | 0.997 | –0.118 | 0.565 | –0.208 | 1.000 | –0.842 | 0.557 | –1.513 | 0.653 | –3.250 | 0.945 | –3.438 | 0.009 | ∗∗ | ||||
| Force Directed Tree, Radial Dendrogram | 0.765 | 0.756 | 1.011 | 0.910 | –1.129 | 0.730 | –1.547 | 0.628 | –0.118 | 0.565 | –0.208 | 1.000 | –1.515 | 0.642 | –2.359 | 0.169 | –0.452 | 0.953 | –0.474 | 0.997 | |||||
Shows the significance results (estimate, standard error, t-value, and p-value) from running a post-hoc analysis using Tukey’s HSD test on clustering-related questions. There are several statistically significant differences in means for questions 13 and 17. The differences between the best and worst performers in questions 14 and 16 are also significant
Legend: *** p≤ 0.001, ** p≤ 0.01, * p≤ 0.05, . p≤ 0.1