Literature DB >> 24051849

A systematic review on the practice of evaluating visualization.

Tobias Isenberg1, Petra Isenberg, Jian Chen, Michael Sedlmair, Torsten Möller.   

Abstract

We present an assessment of the state and historic development of evaluation practices as reported in papers published at the IEEE Visualization conference. Our goal is to reflect on a meta-level about evaluation in our community through a systematic understanding of the characteristics and goals of presented evaluations. For this purpose we conducted a systematic review of ten years of evaluations in the published papers using and extending a coding scheme previously established by Lam et al. [2012]. The results of our review include an overview of the most common evaluation goals in the community, how they evolved over time, and how they contrast or align to those of the IEEE Information Visualization conference. In particular, we found that evaluations specific to assessing resulting images and algorithm performance are the most prevalent (with consistently 80-90% of all papers since 1997). However, especially over the last six years there is a steady increase in evaluation methods that include participants, either by evaluating their performances and subjective feedback or by evaluating their work practices and their improved analysis and reasoning capabilities using visual tools. Up to 2010, this trend in the IEEE Visualization conference was much more pronounced than in the IEEE Information Visualization conference which only showed an increasing percentage of evaluation through user performance and experience testing. Since 2011, however, also papers in IEEE Information Visualization show such an increase of evaluations of work practices and analysis as well as reasoning using visual tools. Further, we found that generally the studies reporting requirements analyses and domain-specific work practices are too informally reported which hinders cross-comparison and lowers external validity.

Entities:  

Mesh:

Year:  2013        PMID: 24051849     DOI: 10.1109/TVCG.2013.126

Source DB:  PubMed          Journal:  IEEE Trans Vis Comput Graph        ISSN: 1077-2626            Impact factor:   4.579


  5 in total

1.  A Systematic Review of Patient-Facing Visualizations of Personal Health Data.

Authors:  Meghan Reading Turchioe; Annie Myers; Samuel Isaac; Dawon Baik; Lisa V Grossman; Jessica S Ancker; Ruth Masterson Creber
Journal:  Appl Clin Inform       Date:  2019-10-09       Impact factor: 2.342

2.  Validation of SplitVectors Encoding for Quantitative Visualization of Large-Magnitude-Range Vector Fields.

Authors:  Garnett W Bryant; Wesley Griffin; Judith E Terrill
Journal:  IEEE Trans Vis Comput Graph       Date:  2016-03-09       Impact factor: 4.579

3.  Information architecture for a patient-specific dashboard in head and neck tumor boards.

Authors:  Alexander Oeser; Jan Gaebel; Andreas Dietz; Susanne Wiegand; Steffen Oeltze-Jafra
Journal:  Int J Comput Assist Radiol Surg       Date:  2018-03-28       Impact factor: 2.924

4.  Open up: a survey on open and non-anonymized peer reviewing.

Authors:  Lonni Besançon; Niklas Rönnberg; Jonas Löwgren; Jonathan P Tennant; Matthew Cooper
Journal:  Res Integr Peer Rev       Date:  2020-06-26

5.  Effects of ensemble and summary displays on interpretations of geospatial uncertainty data.

Authors:  Lace M Padilla; Ian T Ruginski; Sarah H Creem-Regehr
Journal:  Cogn Res Princ Implic       Date:  2017-10-04
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.