Thank you to everyone who is attending this morning both in person and online. I will be sharing my Ph.D. work on the human perception of exponentially increasing data displayed on a log scale evaluated through experimental graphics tasks.
I will first start with the motivation and background of this work and I will introduce the research objectives. Next, I will give a high level overview of the two studies related to perception and prediction as I have previously shared work related to these two chapters. Then I will discuss the development and results from the last study as it relates to estimation and graph comprehension. Lastly, I will provide an overall conclusion and discuss future directions for this work.
Thank you to everyone who is attending this morning both in person and online. I will be sharing my Ph.D. work on the human perception of exponentially increasing data displayed on a log scale evaluated through experimental graphics tasks.
Motivation and Background
Research Objectives
Perception through Lineups
Prediction with 'You Draw It'
Numerical Translation and Estimation
Overall conclusion
Future Work
I will first start with the motivation and background of this work and I will introduce the research objectives. Next, I will give a high level overview of the two studies related to perception and prediction as I have previously shared work related to these two chapters. Then I will discuss the development and results from the last study as it relates to estimation and graph comprehension. Lastly, I will provide an overall conclusion and discuss future directions for this work.
Data visualizations played an important role in during the COVID-19 pandemic (Rost, 2020; Romano, Sotis, Dominioni, and Guidi, 2020; Bavel, Baicker, Boggio, Capraro, Cichocka, Cikara, Crockett, Crum, Douglas, Druckman, and others, 2020).
Dashboards displayed: case counts, transmission rates, outbreak regions.
(Fagen-Ulmschneider, 2020)
The motivation for this work came from the large impact graphics and charts have had during the COVID-19 pandemic. Many of these graphics helped guide decision makers to implement policies such as shut-downs or mandated mask wearing, as well as facilitated communication with the public to increase compliance.
As graphics began to play an important role in sharing information with the public, it was crucial to start asking what made some of these charts better than others? Creators of the graphics were faced with many design choices in order to ensure their charts were effective at accurately communicating the current status of the pandemic.
In order to make educated decisions when designing a chart, we need to establish guidelines through experimentation in order to ensure the graphic is effective at communicating the intended results.
Misleading Graphics
Graphical Guidelines (Wickham, 2013; Andrews, 2022)
In order to ensure your chart is effective, it is important to first understand the many ways in which plots may inaccurately display the data and be ineffective or misleading in sharing information and results. A few ways a chart may be ineffective are by selecting bad form (such as a 3D pie chart), including too much clutter or chartjunk, or by selecting bad axes. This might be by the baseline selected, as shown here, which changes the perceptual interpretation of the chart or by the scale selected to be used on the axes.
Recommendations and guidelines emerged in the early 1900s to help improve the overall quality of graphics.
What is interesting about some of these guidelines, is that even though they were stated in the early 20th century, it took, in some cases, over 100 years to establish and ground these guiding principles through experimentation. A few recommendations made include to keep the symbols to a minimum (avoiding chartjunk), using lengths over areas over volumes (we all may have previously heard that bar charts are better than pie charts), and use proper scales. Here, the ASME advocated for careful consideration of the limiting lines for curves drawn on logarithmic coordinates.
I also want to draw your attention to the minor grid line breaks here drawn at 2 million, 3 million, 4 million, etc. but since they are displayed on the log scale, they visually appear unequal in spacial distance although they are equally spaced numerically.
Evaluate design choices and understand cognitive biases through the use of visual tests (Cleveland and McGill, 1984; Lewandowsky and Spence, 1989; Spence, 1990; Vanderplas, Cook, and Hofmann, 2020).
Could ask participants to:
All of these types of tests require different levels of use and manipulation of the information presented in the chart.
One way in which we establish guidelines is through the use of graphical testing methods. This allows researchers to conduct studies geared at understanding human ability to conduct tasks related to the perception of statistical charts. These tests may take many forms such as identifying differences in graphs, accurately reading information off a chart, using data to make correct real-world decisions, or predicting the next few observations. All of these types of tests require different levels of use and manipulation of the information presented in the chart.
A major issue we encountered in the creation of COVID-19 plots was how to display data from a wide range of values.
Benefits were seen in spring 2020, during the early stages of the COVID-19 pandemic.
Pitfalls were exposed as the pandemic evolved, and the case counts were no longer spreading exponentially.
Benefits were seen in spring 2020, during the early stages of the COVID-19 pandemic.
Pitfalls were exposed as the pandemic evolved, and the case counts were no longer spreading exponentially.
While this is one such example, there is a long history of using log scales to display results in ecology, psychophysics, engineering, and physics. Given the widespread use of logarithmic scales, it is important to understand the implications of their use in order to provide guidelines for best use.
Our perception is logarithmic at first, but transitions to a linear scale later in development (Dehaene, Izard, Spelke, and Pica, 2008; Siegler and Braithwaite, 2017; Varshney and Sun, 2013).
Assumption: If we perceive logarithmically by default, it is a natural way to display information and should be easy to read and understand/use.
When we first learn to count, we begin counting by ones, then by tens, and advancing to hundreds, following the base10 order of magnitude system.
Research shows our perception and mapping of numbers to a number line is logarithmic at first, but transitions to a linear scale later in development, with formal mathematics education.
Regardless of training, we also know our visual system is vulnerable to biases related to different stimuli such as weight, light, or sound. Weber's law established that we do not notice absolute changes in stimuli, but instead, we notice the relative change. For example, if the volume of a speaker is set quiet and you turn it up slightly, you are going to think the change was large compared to if the volume of the speaker was already blasting and you turned it up the same amount as when it was quiet. The Weber-Fechner law extended this discovery and stated that the relationship between the perceived intensity (as sensed by the person) is logarithmic to the stimulus intensity (as outputted by the object source) when observed above a minimal threshold of perception.
Assuming there is a direct relationship between perceptual and cognitive processes, it is reasonable to assume numerical representations should also be displayed on a nonlinear, compressed number scale. Therefore, if we perceive logarithmically by default, it is a natural (and presumably low effort) way to display information and should be easy to read and understand/use.
(Von Bergmann, 2021)
Estimation and prediction of exponential growth is underestimated when presented both numerically and graphically (Jones, 1979; MacKinnon and Wearing, 1991; Wagenaar and Sagaria, 1975).
In addition to biases which result from the use of log scales, there is a general misinterpretation of exponential growth. This visual illustrates how individuals in public health interpret exponential growth distinctly different from scientists during early, middle, and late stages of growth.
Early studies explored the estimation and prediction of exponential growth and found that growth is underestimated* when presented both numerically and graphically (Jones, 1979; MacKinnon and Wearing, 1991; Wagenaar and Sagaria, 1975).
Can log transforming the data help?
In this section, we will take a look at how different graphical tasks are related to different levels of cognition and how that relates to my research objectives.
(Carpenter and Shah, 1998) identifies pattern recognition, interpretative processes, and integrative processes as strategies and processes required to complete tasks of varying degrees of complexity.
Pattern recognition requires the viewer to encode graphic patterns.
Interpretive processes operate on those patterns to construct meaning.
Integrative processes then relate the meanings to the contextual scenario as inferred from labels and titles.
In order to understand how our visual system perceives statistical charts, we must first consider the complexity of the graphic and how viewers are interacting with the data and information being displayed. Pattern recognition requires the viewer to encode graphic patterns while interpretive processes operate on those patterns to construct meaning. Integrative processes then relate the meanings to the contextual scenario as inferred from labels and titles. We will see how these three levels of task complexity relate to the research objectives of our study.
Big Idea: Are there benefits to displaying exponentially increasing data on a log scale rather than a linear scale?
Perception through Lineups
Prediction with You Draw It
Estimation by Numerical Translation
In this research, we conducted a series of three graphical studies to evaluate the impact displaying data on the log scale has on human perception of exponentially increasing trends compared to displaying data on the linear scale. Each study was related to a different graphical task, each requiring a different level of interaction and cognitive use of the data being presented.
Combined, the three studies provide a comprehensive evaluation of the impact of displaying exponentially increasing data on a log scale as it relates to perception, prediction, and estimation. The results of these studies help us make recommendations and provide guidelines for the use of log scales.
Pilot studies have been conducted throughout this research to test out the studies. Many of you were participants in some of these pilot studies.
The final participant recruitment and study deployment were conducted via Prolific, a crowd sourcing website, in March 2022. Participants were provided a link to an R Shiny application that asked them demographic information and then guided them through all three studies. For the final data collection, we collected data for about 300 participants for each study; this was the final data set used for analysis in my research.
I previously presented development processes and pilot study results for the first two chapters so I am going to briefly share a high level overview of the study and final data set results for these.
Study Participant Prompt: Which plot is most different?
We used statistical lineups to test participants perception of exponentially increasing data. 'Lineups' are named after the 'police lineup' of criminal investigations where witnesses are asked to identify the criminal from a set of individuals. Similarly, a statistical lineup is a plot consisting of smaller panels where the viewer is asked to identify the panel containing the real data from among a set of decoy null plots.
Here we see an example of statistical lineups as shown in the first study. Notice how it is much easier to pick out panel 13 as being most different when displayed on the log scale (right) than on the linear scale (left).
Curvature:
\(3!\cdot 2!= 6\) curvature combinations
\(\times 2\) lineup data sets per combination \(=\) 12 test data sets
\(\times 2\) scales (log & linear) \(=\) 24 different lineup plots
Experimental Design: 12 lineup plots per participant
\(6\) test parameter combinations per participant \(\times 2\) scales \(= 12\) test lineups
We used a 'goldilocks' type of approach to select coefficients related to three levels of curvature (low, medium, and high) for our data simulation.
The lineups were then created by looking at combinations of each of these curvature levels where the data in the target panel was simulated with equation A and the data in the null panels were simulated with equation B for a total of 6 curvature combinations.
We simulated 2 replicate data sets of each curvature combinations and each data set was plotted on botht he log and linear scales for a total of 24 different lineup plots to select from. Each participant ended up evaluating 12 lineup plots and ideally saw one of each of the 6 curvature combinations on both scales.
Define \(Y_{ijkl}\) to be the event that participant \(l\) correctly identifies the target plot for data set \(k\) with curvature \(j\) plotted on scale \(i\).
$$\text{logit }P(Y_{ijk}) = \eta + \delta_i + \gamma_j + \delta \gamma_{ij} + s_l + d_k$$ where
We assume that random effects for data set and participant are independent.
Each lineup plot evaluated was assigned a value based on the participant response (correct = 1, not correct = 0). The binary response was analyzed using generalized linear mixed model following a binomial distribution with a logit link function. Here we included both main effects and interactions for the curvature combination and scale while account for the participant and data specific characteristics with random effects.
Estimated (log) odds ratio of successfully identifying the target panel on the log scale compared to the linear scale. The y-axis indicates the the model parameters used to simulate the null plots with the target plot model parameter selection designated by shape and shade of green. The thumbnail figures on the right display the curvature combination as shown previously on both scales (linear - left, log - right).
The results indicated the choice of scale changes the contextual appearance of the data leading to slight perceptual advantages for both scales depending on the curvatures of the trend lines being compared.
The second study evaluated participants ability to make forecast predictions of exponentially increasing data.
Study Participant Prompt: Use your mouse to fill in the trend in the yellow box region.
In 2015, the New York Times introduced an interactive feature, called ‘You Draw It’, where readers input their own assumptions about various metrics and compare how these assumptions relate to reality. The Times team utilizes Data Driven Documents (D3) that allows readers to predict these metrics through the use of drawing a line on their computer screen with their mouse. We adapted this feature as a way to measure the patterns we see in data and compare intuitive visual model fitting and statistical model fitting. In the study, participants were shown a series of 'You Draw It' interactive task plots and asked to "Use your mouse to fill in the trend in the yellow box region." The yellow box region moves along as the participant draws their trend-line until the yellow region disappears, providing a visual cue for the task.
Study Participant Prompt: Use your mouse to fill in the trend in the yellow box region.
A sub-study validated a new method, 'You Draw It', as a tool for graphical testing statistical graphics and introduced an appropriate statistical analysis method for comparing visually fitted trend lines to statistical regression results (Eye Fitting Straight Lines in the Modern Era).
In 2015, the New York Times introduced an interactive feature, called ‘You Draw It’, where readers input their own assumptions about various metrics and compare how these assumptions relate to reality. The Times team utilizes Data Driven Documents (D3) that allows readers to predict these metrics through the use of drawing a line on their computer screen with their mouse. We adapted this feature as a way to measure the patterns we see in data and compare intuitive visual model fitting and statistical model fitting. In the study, participants were shown a series of 'You Draw It' interactive task plots and asked to "Use your mouse to fill in the trend in the yellow box region." The yellow box region moves along as the participant draws their trend-line until the yellow region disappears, providing a visual cue for the task.
A sub-study validated a new method, 'You Draw It', as a tool for graphical testing statistical graphics and introduced an appropriate statistical analysis method for comparing visually fitted trend lines to statistical regression results. We are working on revisions for the manuscript related to this sub-study.
Data were simulated based on a one-parameter exponential model: \(N = 30\) points \((x_i,y_i), i = 1,...,N\) where $$y_i = e^{\beta x_i + e_i}$$ with \(e_i \sim N(0,\sigma^2)\).
2 x 2 x 2 factorial:
For this study, data were simulated based on a one-parameter exponential model. We designed a 2x2x2 factorial treatment design where we are looking at the effect of growth rate, having the aid of points, and of course the effect of scale. This creates a total of 8 treatment combinations for 'You Draw It' plots in which participants were shown in random order.
For each participant, the final data set used for analysis contains:
for:
Vertical residuals between the drawn and fitted values were calculated as:
Out of the study, we save the participant feedback data to a data base for analysis. Here we see a spaghetti plots of results. Participants drawn lines on the linear scale are shown in blue and the log scale are shown in orange. Variability in the statistically fitted regression lines occurred due to a unique data set being simulated for each individual; the gray band shows the range fitted values from the statistically fitted regression lines. Here we can see a strong underestimation on the linear scale (blue lines).
Fit separate for each growth rate (low = 1, high = 2), the GAMM equations for residuals are given by: \begin{equation} e_{1jklm,NLS} = \tau_{1jk} + s_{1jk}(x_{1jklm}) + p_{l} + s_{l}(x_{1jklm}) \end{equation} \begin{equation} e_{2jklm,NLS} = \tau_{2jk} + s_{2jk}(x_{2jklm}) + p_{l} + s_{l}(x_{2jklm}) \end{equation}
where
Allowing for flexibility, the we fit a Generalized Additive Mixed Model to estimate smoothing splines for the vertical residuals from the participant drawn line in relation to the NLS fitted values. We included an intercept related to each treatment combination along with a smoothing spline adjustment for each treatment combination. We accounted for participant variability with random intercept and random spline effects.
The results from the analysis showed a clear underestimation of forecasting trends with high exponential growth rates when participants were asked to make predictions on the linear scale.
The previous two studies explored the use of log scales through differentiation and visual prediction of trends. These did not depend on context and instead focused on how our visual system percieves and identifies patterns in exponential growth.
In order to understand the cognitive implications of displaying exponentially increasing data on a log scale, we evaluated graph comprehension as it relates to the contextual scenario of the data shown.
Graph Comprehension
The three behaviors related to graph comprehension involve (Curcio, 1987; Friel, Curcio, and Bright, 2001; Glazer, 2011; Jolliffe, 1991; Wood, 1968).
Estimation Biases
This study involved integrative processes which required viewers to translate the visual features into conceptual relations by interpreting titles, labels, and scales.
An important consideration in understanding graph comprehension is the questions being asked of the viewer. Three levels of graph comprehension have emerged from mathematics education research: literal reading of the data (elementary level), reading between the data (intermediate level), and reading beyond the data (advanced level).
While not exclusive to extracting numerical values from charts, mathematics education also places an emphasis on quantitative estimation skills such as providing an estimated length of a string or weight of a box.
With open ended estimation tasks, certain biases including anchoring and rounding to multiples of five or ten arise.
Question type | Tribble scenario | Ewok scenario |
---|---|---|
Open Ended | Between stardates 4530 and 4540, how does the population of Tribbles change? | Between 30 and 40 ABY, how does the population of Ewoks change? |
Elementary Q1 | What is the population of Tribbles in stardate 4510? | What is the population of Ewoks in 10 ABY? |
Elementary Q2 | In what stardate does the population of Tribbles reach 4,000? | In what ABY does the population of Ewoks reach 4,000? |
Intermediate Q1 | From 4520 to 4540, the population increases by ____ Tribbles. | From 20 ABY to 40 ABY, the population increases by ____ Ewoks. |
Intermediate Q2 | How many times more Tribbles are there in 4540 than in 4520? | How many times more Ewoks are there in 40 ABY than in 20 ABY? |
Intermediate Q3 | How long does it take for the population of Tribbles in stardate 4510 to double? | How long does it take for the population of Ewoks in 10 ABY to double? |
We selected two scenarios and provided text to participants describing a context in which populations of fuzzy fictional characters from star trek and star wars are growing exponentially over periods of time. We wanted to go with a fun context to avoid any personal opinions or biases affecting the results.
We then selected 6 questions related to each scenario. The open ended question was meant to prompt the participants to take time to inspect the graph while the two intermediate questions are related to the literal reading and estimation of population and year respectively. There were three intermediate questions selected to address the additive increase in population between two years, the multiplicative change in population between two years, and the amount of time it took the population to double from a certain year.
We did not focus on advanced level questioning since extrapolation and interpolation were addressed in the second study.
Each of these questions will again be presented in the following slides of results.
Two unique data sets were simulated based on a three parameter exponential model with multiplicative errors:
\(y_i=\alpha e^{\beta x_i + \epsilon_i} + \theta\) with \(\epsilon_i \sim N(0, \sigma^2)\)
for
The time unit labels on the x-axis reflected 0 to 50 ABY (After Battle of Yavin) for the Ewok scenario and were adjusted to 4500 to 4550 stardates for the Tribble Scenario to align with the associated popular media depiction of each figure.
Two unique data sets were generated with the same underlying parameter coefficients, but different errors randomly generated from the same error distribution. Notice in particular that some of the simulated points fall farther from the "true value" or underlying curve. For example, year 40 in data set 2.
This figure displays scatter plots of the two unique data sets on both the linear and log base two scales. A log base two was selected in order to aid in participants estimation of time until the population doubled in the third intermediate question.
I will refer to the time unit as year, but during the study, the time unit labels on the x-axis aligned with the associated popular media depiction of each figure. This was also meant to disguise the use of the same underlying data simulation model and estimation questions across both scenarios.
Participants were shown two examples with associated sketches (available throughout study).
The scale of the graphic and data set displayed was randomly assigned to scenarios for each individual.
The order in which the participant saw the scenario, scale, and data set combination was randomly selected.
Questioning:
During the study, participants were first shown two examples with associated sketches that were available to them throughout the study.
The scale of the graphic and data set displayed was randomly assigned to scenarios for each individual.
The order of the two scenarios and their assigned data set and scale was randomly assigned to each individual.
In each scenario, participants were first asked an open ended question, which required them to spend time exploring the data displayed in the graphic, followed by a random order of two elementary level questions and three intermediate level questions.
We recorded the inputted and evaluated calculations and scratch work of each participant in order to better understand participant strategies for estimation.
For quantitative estimation questions, we provided basic calculator and scratch pad resources for participants to use. We recorded the inputted and evaluated calculations and scratch work of each participant in order to better understand participant strategies for estimation.
The data set used for analysis contained the unique participant identification and indicated the scenario, scale, data set, and estimation question along with the participant text response or quantitative estimate, calculation input and evaluation, and associated scratch work.
An array of graphical displays allow for visual inspection of participant responses and provide suggestions about the cognitive implications of displaying exponentially increasing data on the log scale.
I am now going to present results from the study as they relate to each of the six questions asked.
"Between year 30 and 40, how does the population of [creatures] change?"
Before participants were asked to estimate numeric quantities, they were asked to provide an open ended response and describe how the population changed over between year 30 and 40. We used a comparison word cloud to illustrates the general terminology participants used when describing the scatter-plots shown on each scale.
Participants more frequently referred to terms such as ‘exponential’ and ‘rapid’ when shown the scatter plot on the linear scale while ‘double’ and ‘quadruple’ were often used to describe the graphic when shown on the log scale; indicating participants read the y-axis labels and noticed the doubling grid lines.
"What is the population in year 10?"
In order to examine the effect of scale on literal reading of the data, participants were asked, “What is the population in year 10?”
Here we see a sketch of the estimation procedure. Participants first locate 10 along the x-axis and move upward until they believe they have found the correct location on the curve; then participants look to the y-axis for their estimated population.
We used density plots were to illustrate the distribution of the quantitative estimates for the population in year 10 provided by participants. The colors are associated to the scale - linear (blue) and log (orange) - and vertical lines indicate the true value based on the underlying model equation (black solid), closest point value based on the simulated data set (black dashed), and grid lines shown on the graphs (blue/orange dotted). A jittered rug plot along the x-axis shows where participant estimates were made. The two unique data sets are shown separately.
The figure reveals a larger variance in quantitative population estimates made on the linear scale compared to the log scale. As expected, it is clear that participants were anchoring to grid lines and base ten values as highlighted by the high density of estimates at 512 and 500 on the log scale as well as local maximums near multiples of ten such as 500 and 1,000.
One such explanation for this difference might be the location of year 10 along the x-axis. This year corresponds to a low magnitude where the population is condensed in a small region on the linear scale as opposed to later in time when larger magnitudes in population can be seen. Thus, it is important to evaluate the accuracy of estiamtes along the full domain.
In two estimation questions related to intermediate level reading between the data, participants are asked to provide an increase and change in population between years 20 and 40, thus requiring participants to make first-level estimates at these locations. To understand the effect of the location along the domain and in turn the magnitude of the population being estimated, we extracted first-level estimates for years 20 and 40 from participant calculations and scratch work.
These figures show evaluations of participants estimates of the population at years 10, 20, and 40 for each data set separately. Spaghetti plots are displayed on the linear scale (top) and log scale (bottom) with both scale estimates shown on each - linear (blue), log (orange). The year was calculated from the underlying model equation based on the population estimate provided by the participant. Gray arrows indicate the true value and closest point value as demonstrated in the sketch.
As the year increases, we observe an increased accuracy in estimates made on the linear scale while estimates made on the log scale suffer in accuracy due to strong anchoring to grid lines and the larger quantitative difference between grid lines as population magnitudes increase.
This is important to recognize becuase inaccurate first-level estimations can lead to consequences in estimations which require participants to make comparisons between two points (e.g. Intermediate Q1 and Q2).
Participant calculations and scratch work provides support that participants equated halfway spatially as halfway numerically.
2048 − 1024 = 1024
1024/2 = 512
512 + 1024 = 1536
32768 − 16384 = 16384
32768 − 16384 = 16384
16384 ∗ 2 = 32768
16384/2 = 8192
8192 + 16384 = 24576.
In extracting participant first-level estimates from their calculation and scratch work, we observed participants were resistant to estimating between grid lines and had a greater tendency to anchor their estimates to the grid line estimates on the log scale. When estimates between grid lines were made, participant calculations and scratch work revealed a lack of understanding of logarithmic mapping due to considering spatial distance as indicative of numerical distance.
"In what year does the population reach 4,000?"
In addition to estimating the population from a given year, participants were asked, “In what year does the population reach 4,000?” Participants first located 4000 along the y-axis and move to the right until they believe they have found the correct location on the curve; then participants look down to the x-axis for their estimated year.
The density plot shows anchoring occurred to the closest point as shown by an increase in density around the dashed lines as opposed to the solid black line. We also saw density peaks occur at whole values indicating rounding errors.
From 20 to 40, the population increases by __ [creatures].
Intermediate level questions required participants to read between the data and make comparisons between points. Participants were asked, “From 20 to 40, the population increases by _ [creatures].” (Fig. 4.10). The questioning was selected carefully to prompt participants to make an additive comparison of populations between two years.
We saw during the sketch on the previous slide for first level estimates that this procedure required participants to make first-level population estimates at years 20 and 40, then calculate the difference between the two values by subtraction.
These two figures display the density for estimated increases in population as made by participants for data set one and two respectively.
Here we also see the subjectivity of results due to the unique data points as the point associated with year 40 appears much further from the underlying data trend for data set 2 than it does for data set 1.
How many times more [creatures] are there in 40 than in 20?
Participants were asked, “How many times more [creatures] are there in 40 than in 20?” The questioning was selected carefully to prompt participants to make a multiplicative comparison between two years.
Here we see a similar sketch as before where participants make first-level population estimates at years 20 and 40, but then calculate the ratio between the two values rather than the difference.
The density plots shown for each data set separately, show participants tended to be make more accurate and less variable estimates on the log scale than on the linear scale. Evaluating some of the scratch-work for this question, it was interesting to see how many participants used a "guess-and-check" sort of method of multiplying their population estimate from year 20 by different multipliers until they felt the result matched the population at year 40.
Additive Increase Incorrect Logic
2048 − 1024 = 1024
1024 + 512 = 1536
16384 − 1536 = 14848
14848/1536 = 9.67.
Multiplicative Change Incorrect Logic
16384 − 1536 = 14848
Overall, there seemed to be misunderstanding of log logic as demonstrated by misinterpretations and miscalculations of the previous two questions.
This figure displays the observed estimated change in population for Intermediate Q2.The plot shows a substantial number of participants provided estimates that more closely reflected that of the additive increase in population rather than the multiplicative change. We can also see scratch work providing support that miscalculations were made.
However, further investigation is necessary to determine if the misunderstanding occurs due to the ambiguity of language or scale.
How long does it take for the population in year 10 to double?
An alternative multiplicative comparison between two points was to determine the amount of time it took for a value to double. Participants were asked, “How long does it take for the population in year 10 to double?”
This required participants to make a first-level population estimate at year 10, then double this population and estimate the year in which that occurs. Lastly, participants subtracted year 10 to determine how long it took for the population to double. Another estimation strategy could involve participants spatially judging the double of population on the log base2 scale; however, examining the scratch-work revealed most participants took the first approach to estimation.
The estimated number of years until the population doubled from a reference year of 10 resulted in lower variability on the log scale as opposed to a larger variability on the linear scale. One explanation might be the low magnitude of population in year 10 and further exploration would be needed to justify for other reference years. Common responses strongly suggest participants were anchoring to multiples of 5 years.
Understanding log logic is difficult.
Accuracy greatly depends on the location of the value being estimated in relation to the magnitude.
Strong anchoring and rounding effects.
Participants were resistance to estimate between grid lines on the log scale.
Inaccurate representation of equating spatial distance to quantitative difference.
Inaccurate first level estimations can lead to consequences in estimations which require participants to make comparisons between two points.
Estimates were subjective to the simulated data set.
This study was intended to aid in understanding the cognitive implications of displaying exponentially increasing data on a log scale.
Some of the main take aways from this study are shown here. Overall, we provide argument that understanding log logic is difficult and that the accuracy greatly depends on the location of the value being estimated in relation to the magnitude.
We found biases related to strong anchoring and rounding effects as participants were resistant to estimate between grid lines on the log scale and had inaccurate representations of equating that spatial distance to the quantitative difference.
This inaccuracy in first-level estimations showed a clear consequence in estimations which required participants to make comparisons between two points.
The use of two unique data sets allowed us to look at how estimates are very subjective to the simulated data set as some participants are actually reading the data off of the plot and not basing estimates off the underlying trend.
Perception
Prediction
Estimation
This research evaluated the use of log scales to display exponentially increasing data from three different angles and levels of complexity: perception, prediction, and estimation. Each study provided us insight into the advantages and disadvantages of displaying exponentially increasing data on a log scale and in what context each choice of scale might be more appropriate.
The first study laid the foundation for the future studies. Results from this study indicated the choice of scale changes the contextual appearance of the data leading to slight perceptual advantages for both scales depending on the curvatures of the trend lines being compared.
The second study tested participant’s abilities to make forecast predictions for exponentially increasing trends on both scales. The results from the analysis showed a clear underestimation of forecasting trends with high exponential growth rates when participants were asked to make predictions on the linear scale and improvements in forecasts were made when participants were asked to make predictions on the log scale.
The third study evaluated graph comprehension as it relates to the contextual scenario of the data shown. Overall, our results suggested that log logic is difficult and that anchoring and rounding biases result in a sacrifice in accuracy in estimates made on the log scale for large magnitudes.
Perceptual advantages of the use of log scales due to the change in contextual appearance.
Our understanding of log logic is flawed when translating the information into context.
We recommend consideration of both user needs and graph specific tasks when presenting data on the log scale.
Caution should be taken when interpretation of large magnitudes is required, but advantages may appear when it is necessary to visually identify and interpret small magnitudes on the chart.
This research evaluated the use of log scales to display exponentially increasing data from three different angles and levels of complexity: perception, prediction, and estimation. Each study provided us insight into the advantages and disadvantages of displaying exponentially increasing data on a log scale and in what context each choice of scale might be more appropriate.
The results are instrumental in establishing guidelines for making design choices about scale which result in data visualizations effective at communicating the intended results.
We recommend understanding the contextual appearance log scales have on the data being displayed and the perceptual advantages and disadvantages that come with each. It is also important to recognize that our understanding of log logic is flawed when translating this information into context.
In general, it is important to consideration of both user needs and graph specific tasks when presenting data on the log scale. Caution should be taken when interpretation of large magnitudes is required, but advantages may appear when it is necessary to visually identify and interpret small magnitudes on the chart.
Potential follow-up studies:
Expand the use of log scales to data which does not follow an exponential trend, but where log scales are otherwise appropriate.
Evaluate the implications of transforming or displaying the \(x\)-axis on a log scale.
More widespread use of comprehensive graphical studies conducted on the same type of data and plot.
Follow-up studies for this research could provide further insight into the strategies of estimation by providing users the ability to interact with the graph with visual aids such as the arrows shown in sketches from the estimation study and how we make decisions based on data displayed on a log scale to evaluate the effect of scale on risk adversity.
This research stands as a model for conducting an extensive series of graphical tasks on the same type of data and plot in order to gain a comprehensive understanding of both the perceptual and cognitive implications of the design choices.
Andrews, R. (2022a). Standards for graphic presentation. URL: https://infowetrust.com/project/1915-standards.
— (2022b). Standards for graphic presentation. URL: https://infowetrust.com/project/1915-standards.
Curcio, F. (1987). “Comprehension of mathematical relationships expressed in graphs”. In: Journal for research in mathematics education 18.5, pp. 382–393.
Dehaene, S., V. Izard, E. Spelke, et al. (2008). “Log or linear? Distinct intuitions of the number scale in Western and Amazonian indigene cultures”. In: science 320.5880, pp. 1217–1220.
Dongarra, J. J., H. W. Meuer, E. Strohmaier, et al. (1997). “TOP500 supercomputer sites”. In: Supercomputer 13, pp. 89–111.
Friel, S., F. Curcio, and G. Bright (2001). “Making sense of graphs: Critical factors influencing comprehension and instructional implications”. In: Journal for Research in mathematics Education 32.2, pp. 124–158.
Glazer, N. (2011). “Challenges with graph interpretation: A review of the literature”. In: Studies in science education 47.2, pp. 183–210.
Jolliffe, F. (1991). “Assessment of the understanding of statistical concepts”. In: Proceedings of the third international conference on teaching statistics. Vol. 1. , pp. 461–466.
Jones, G. (1979). “A generalized polynomial model for perception of exponential series.” In: Perception & Psychophysics.
MacKinnon, A. and A. Wearing (1991). “Feedback and the forecasting of exponential change”. In: Acta Psychologica 76.2, pp. 177–191.
Siegler, R. and D. Braithwaite (2017). “Numerical development”. In: Annual review of psychology 68, pp. 187–213.
Tufte, E. R. (1985). “The visual display of quantitative information”. In: The Journal for Healthcare Quality (JHQ) 7.3, p. 15.
Varshney, L. and J. Sun (2013). “Why do we perceive logarithmically?” In: Significance 10.1, pp. 28–31.
Villa, M. (2022). The rouble collapse against US dollar. URL: https://twitter.com/emmevilla/status/1498365717720223751.
Von Bergmann, J. (2021). xkcd_exponential: Public Health vs Scientists. URL: https://github.com/mountainMath/xkcd_exponential.
Wagenaar, W. and S. Sagaria (1975). “Misperception of exponential growth”. In: Perception & Psychophysics 18.6, pp. 416–422.
Wickham, H. (2013). “Graphical criticism: some historical notes”. In: Journal of Computational and Graphical Statistics 22.1, pp. 38–44.
Wood, R. (1968). “Objectives in the teaching of mathematics”. In: Educational Research 10.2, pp. 83–98.
Advisors
Committee Members
Other
Tribble scenario
Hi, we’re Tribbles! We were taken from our native planet, Iota Germinorum IV, and brought abroad Starfleet in stardate 4500. A Starfleet scientist, Edward Larkin, genetically engineered us to increase our reproductive rate in an attempt to solve a planetary food shortage. The Tribble population on Starfleet over the next 50 stardates (equivalent to 1 week universe time) is illustrated in the graph. We need your help answering a few questions regarding the population of Tribbles.
Ewok scenario
Hi, we’re Ewoks! We are native to the forest moon of Endor. After the Galactic Civil War, some Ewoks traveled offworld to help Rebel veterans as `therapy Ewoks’ and began to repopulate. The Ewok population After the Battle of Yavin (ABY) is illustrated in the graph. We need your help answering a few questions regarding the population of Ewoks offworld.
Fictional illustrations of the figures were modified from artwork by Allison Horst (2021) and included on the main page for each scenario.
Previous results have found few differences between expert and non-expert participants in this context (VanderPlas and Hofmann, 2015).
Motivation and Background
Research Objectives
Perception through Lineups
Prediction with 'You Draw It'
Numerical Translation and Estimation
Overall conclusion
Future Work
I will first start with the motivation and background of this work and I will introduce the research objectives. Next, I will give a high level overview of the two studies related to perception and prediction as I have previously shared work related to these two chapters. Then I will discuss the development and results from the last study as it relates to estimation and graph comprehension. Lastly, I will provide an overall conclusion and discuss future directions for this work.
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
Esc | Back to slideshow |