We designed a controlled study in which we manipulated a single variable: visualisation type (the conventional visualisation condition, or CV, versus the data storytelling condition, or DS). We opted for a within-subjects study design, wherein the same participant would be presented with both conventional visualisations and data stories, with the order of presentation counterbalanced through randomisation. That is, each participant served as their own control, enhancing the statistical power and sensitivity in detecting effects as noted in [
19]. This approach is particularly important, considering we also incorporated individual’s visualisation literacy into the analysis of RQ5.
The study was structured to be completed within 40 minutes, and participants received AUD $14 as compensation for their time. Additionally, although participants were compensated for 40 minutes of their time, they were encouraged to complete the study "as fast as possible" for efficient earning. The study consisted of four sections: i) an introduction, including demographic and background questions; ii) a visualisation literacy test (addressing RQ5); iii) the main comparative study (RQ1, RQ2, RQ3, and RQ5); and iv) questions concerning perceptions of data storytelling elements (RQ4). Ethics approval was obtained from Anonymous University (Project ID: Anonymised). Participants were asked to consent before participating in the study. Figure
2 illustrates the study procedure and maps each section to our research questions (RQs). Details are presented below (further details are presented in the supplementary material of this paper for replication purposes).
3.3.3 Section 3 – The comparative study.
Conventional visualisations versus data stories. This consisted of the core questions of this survey (see Figure
2-Section 3). We designed 6 pairs of visualisations, each showcasing different data but conveying the same insight within a pair. Specifically, we created six conventional visualisations (cv1-6) alongside six alternate versions that applied data storytelling principles (ds1-6). These data stories were created following the transformation process to convert conventional visualisations to those with data storytelling elements, as proposed by [
99] and detailed in Section
3.1. Two researchers participated in several discussion sessions to collaboratively identify the insights from each visualisation based on the data and metadata provided in the
Our World in Data repository (phase i) to then apply the transformation process (phases ii and iii), ensuring that aspects such as the type of chart, decluttering actions, emphasis techniques, annotations, and explanatory titles all effectively contributed to conveying the insights. Later, a third researcher validated the insights and the data stories, and further participated in discussion sessions with the other researchers to agree on the final set of visualisations. All visualisations included the complete range of data storytelling elements discussed above. Two example conventional visualisation–data story pairs (cv1–ds1 and cv2–ds2) are shown in Figure
3. To facilitate a balanced exposure to both conventional and data storytelling visualisations, each participant was randomly presented with three items from each condition, and in a random sequence. For instance, one participant might be shown visualisations cv1, cv2, cv3 paired with ds4, ds5, ds6, while another could encounter cv4, cv5, cv6 and ds1, ds2, ds3. This method not only ensured an even distribution of both conventional visualisations and data stories but also helped mitigate any learning or practice effects. No explicit indication was provided to participants regarding which visualisation was conventional and which was a data story.
The data visualisations were designed to be diverse in terms of both visualisation complexity and frequency of use across varied audiences. To achieve this, and as described in the previous section, we included a range of chart types: line charts, bar charts, pie charts, stacked bar charts, stacked area charts, and bubble charts. According to Segel and Heer [
85], line charts, bar charts, and pie charts are the most commonly used visualisation techniques and are relatively easy to understand [
56]. Stacked bar charts and stacked area charts have been found to be slightly more challenging for participants, as indicated by the difficulty scores calculated by Lee et al. [
56], yet they are also commonly used. Bubble charts have been considered more difficult to interpret, according to their difficulty scores, and are less commonly used [
56]. This assortment of chart types provided a comprehensive range of visualisations (see details in Table
1). In our study, the distribution mirrored the commonality of these charts: we included three chart types that are most common, two that are relatively common, and one that is less commonly used. The choice of chart for the conventional visualisations was also influenced by the specific data being displayed. We intentionally designed the conventional mirroring of the original data source made public through
Our World in Data.
We included text annotations, data point annotations, colour emphasis, and explanatory titles in all visualisations with data storytelling elements. Moreover, since the change of chart type can significantly affect visualisation and different chart types require distinct visualisation skills from participants [
56], we limited the change of chart types to half of the visualisation sets (3 sets) (see details in Table
1, columns 1 and 2).
Types of questions. Each time a participant was presented with a visualisation, they were asked to respond to four multiple-choice questions. Each question was categorical, and the correct response was only one out of four predefined options. For each pair of visualisations, the same set of questions was asked consistently for both versions –the conventional visualisation and the data story – to different participants.
We defined these questions based on Bloom’s taxonomy [
6]. Bloom’s taxonomy has been extensively used to categorise questions into varying levels of complexity (i.e., knowledge, comprehension, application, analysis, synthesis, and evaluation). It has also been proposed for assessing the kinds of tasks that data visualisation can enable [
2,
15], as well as for supporting the development of visualisation skills [
16,
63]. Following the guidelines proposed by Burns et al. [
15], we designed two types of questions based on the first two levels (knowledge and comprehension) of Bloom’s taxonomy. We selected these two levels because they align with the goal of this research (RQ3). Further levels require other types of interventions and study designs that go beyond the purpose of the current study. In the context of information visualisation, knowledge (Level 1) questions in Bloom’s taxonomy serve to assess whether users can identify and select relevant data points, categories, or trends from a data set. This aligns with our goal to assess whether data storytelling supports
information retrieval tasks. Comprehension questions (Level 2) relate to questions that assess the ability to understand, interpret, and derive meaningful insights from the data. Data comprehension, in this sense, goes beyond mere retrieval and involves interpretation, assessment, and decision-making. It involves identifying various data points, making comparisons, and understanding the underlying meaning, which is aligned with the
comprehension task. Questions of this type were subdivided into two sub-types. The first sub-type asked participants about a
single insight that the data visualisation aimed to provide and allowed the participants to choose the correct answer. The second sub-type requires participants to gain a comprehensive understanding of
multiple insights. The question presents four statements, and the participants were required to identify the invalid statement.
For each pair of visualisations, we added two information retrieval (Level 1) questions and two comprehension (Level 2) questions (see Table
1, column 3). Table
2 (row 1) shows an example of an information retrieval question that requires the participants to retrieve a single data point from the visualisation and choose the correct value among the response options [
1]. Table
2 (row 2) shows an example of a comprehension question that asks the participants to identify a single insight by comparing various data points. The comprehension question sub-types were equally distributed across all pairs of visualisations and depended on the number of insights that could be extracted from the data shown in each visualisation pair (see Table
1, column 3). Table
2 (row 3) presents an example in which a participant needs to identify the incorrect insight among the other three correct insights communicated in the conventional visualisation or the data story. Two researchers collaboratively designed these questions independently of the type of visualisation (conventional or data story). A third researcher validated these questions. Discussion sessions were conducted among the three researchers to reach a consensus on the final set of questions. They ensured there was an alignment between the questions asked and the information highlighted in the visualisations with data storytelling elements.
The time taken to complete each question was recorded to assess the efficiency of data storytelling (RQ1) [
36,
105], while the correctness rate of responses was recorded to shed light on its effectiveness (RQ2) [
36,
105].