Award Abstract # 1749266
CAREER: Enhancing Critical Reflection on Data by Integrating Users' Expectations in Visualization Interaction

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: UNIVERSITY OF WASHINGTON
Initial Amendment Date: April 9, 2018
Latest Amendment Date: April 9, 2018
Award Number: 1749266
Award Instrument: Continuing Grant
Program Manager: Dan Cosley
dcosley@nsf.gov
 (703)292-8832
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: June 1, 2018
End Date: July 31, 2019 (Estimated)
Total Intended Award Amount: $523,516.00
Total Awarded Amount to Date: $104,038.00
Funds Obligated to Date: FY 2018 = $38,133.00
History of Investigator:
  • Jessica Hullman (Principal Investigator)
    jhullman@northwestern.edu
Recipient Sponsored Research Office: University of Washington
4333 BROOKLYN AVE NE
SEATTLE
WA  US  98195-1016
(206)543-4043
Sponsor Congressional District: 07
Primary Place of Performance: University of Washington
Seattle
WA  US  98195-2350
Primary Place of Performance
Congressional District:
07
Unique Entity Identifier (UEI): HD1WMN6945W6
Parent UEI:
NSF Program(s): HCC-Human-Centered Computing
Primary Program Source: 01001819DB NSF RESEARCH & RELATED ACTIVIT
01001920DB NSF RESEARCH & RELATED ACTIVIT

01002021DB NSF RESEARCH & RELATED ACTIVIT

01002122DB NSF RESEARCH & RELATED ACTIVIT

01002223DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7367
Program Element Code(s): 736700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Interactive graphs, charts, and other visual representations of data are increasingly common in public life. People bring their own expectations and assumptions about the data and situations these visualizations represent, though most visualizations do not take these expectations into account. Comparing these expectations to actual data is a powerful tool for checking those assumptions, developing better understanding of situations, and making better decisions. To support such expectation visualizations, the project team will use a combination of experiments, software development, and design activities to develop toolkits and best practices for developing visualizations that allow viewers to represent, interact with, and see feedback on their own predictions about the data. The work will focus on helping people better understand scientific research and expert analysis around topics such as health decisions that might impact their own lives. The work will also support a broader educational goal of data literacy education, through course modules that can be inserted into introductory informatics and data science courses and a new course on thinking with data, and outreach goals through developing a research and development platform where designers, researchers, and developers can work together to improve expectation visualization techniques.

To do this, the project has three main research goals. The first is to develop a suite of empirical findings on the effects of expectation visualization, through a series of experiments on how predicting data, receiving personalized feedback on those predictions, and reflecting on gaps between predictions and data affect people's later memory of the data and future expectations. The second thrust builds on the first, using these empirical results along with design studies and comprehensive reviews of existing tools and literature to build a design space with software examples characterizing key decisions in designing expectation visualizations. These decisions will include a range of techniques for graphically eliciting people's expectations, contextualization techniques that help people learn to use those techniques and constrain their choices appropriately, and feedback or reflection techniques that help call attention to places where expectations did and did not match the underlying data. The third thrust is to put these principles into practice by developing applications to support the communication of uncertainty in experimental results, the reduction of spurious pattern discoveries in data analysis, and the integration of problem context and expert analysis with the visualization itself.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Kim, Yea-Seul and Walls, Logan A. and Krafft, Peter and Hullman, Jessica "A Bayesian Cognition Approach to Improve Data Visualization" ACM Conference on Computer Human Interaction , 2019 10.1145/3290605.3300912 Citation Details
Kim, Y.S. "A Bayesian Cognition Approach to Improve Data Visualization" Proceedings of 2019 ACM Computer-Human Interaction (CHI) , 2019 Citation Details
Nguyen, F. "Belief-Driven Data Journalism" Computation+Journalism , 2019 Citation Details
Phelan, Chanda and Hullman, Jessica and Kay, Matthew and Resnick, Paul "Some Prior(s) Experience Necessary: Templates for Getting Started With Bayesian Analysis" ACM Conference on Computer Human Interaction , 2019 10.1145/3290605.3300709 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page