Narratives in Crowdsourced Evaluation of Visualizations: A Double-Edged Sword? - Archive ouverte HAL Access content directly
Conference Papers Year : 2017

Narratives in Crowdsourced Evaluation of Visualizations: A Double-Edged Sword?

Evanthia Dimara
Anastasia Bezerianos
Pierre Dragicevic

Abstract

We explore the effects of providing task context when evaluating visualization tools using crowdsourcing. We gave crowd-workers i) abstract information visualization tasks without any context, ii) tasks where we added semantics to the dataset, and iii) tasks with two types of backstory narratives: an analytic narrative and a decision-making narrative. Contrary to our expectations , we did not find evidence that adding data semantics increases accuracy, and further found that our backstory narratives can even decrease accuracy. Adding dataset semantics can however increase attention and provide subjective benefits in terms of confidence, perceived easiness, task enjoyability and perceived usefulness of the visualization. Nevertheless, our backstory narratives did not appear to provide additional subjective benefits. These preliminary findings suggest that narratives may have complex and unanticipated effects, calling for more studies in this area.
Fichier principal
Vignette du fichier
framing-paper-chi-2016.pdf (1.73 Mo) Télécharger le fichier
Origin : Files produced by the author(s)
Loading...

Dates and versions

hal-01448095 , version 1 (31-01-2017)

Licence

Attribution - ShareAlike

Identifiers

Cite

Evanthia Dimara, Anastasia Bezerianos, Pierre Dragicevic. Narratives in Crowdsourced Evaluation of Visualizations: A Double-Edged Sword?. ACM Conference on Human Factors in Computing Systems (CHI) , May 2017, Denver, United States. ⟨10.1145/3025453.3025870⟩. ⟨hal-01448095⟩
737 View
1167 Download

Altmetric

Share

Gmail Facebook X LinkedIn More