Award Abstract # 1631428
NCS-FO: Collaborative Research: Operationalizing Students' Textbooks Annotations to Improve Comprehension and Long-Term Retention

NSF Org: DRL
Division of Research on Learning in Formal and Informal Settings (DRL)
Recipient: THE REGENTS OF THE UNIVERSITY OF COLORADO
Initial Amendment Date: August 17, 2016
Latest Amendment Date: May 10, 2017
Award Number: 1631428
Award Instrument: Standard Grant
Program Manager: Gregg Solomon
gesolomo@nsf.gov
 (703)292-8333
DRL
 Division of Research on Learning in Formal and Informal Settings (DRL)
EDU
 Directorate for STEM Education
Start Date: September 1, 2016
End Date: August 31, 2021 (Estimated)
Total Intended Award Amount: $299,976.00
Total Awarded Amount to Date: $307,176.00
Funds Obligated to Date: FY 2016 = $299,976.00
FY 2017 = $7,200.00
History of Investigator:
  • Michael Mozer (Principal Investigator)
    mozer@cs.colorado.edu
Recipient Sponsored Research Office: University of Colorado at Boulder
3100 MARINE ST
Boulder
CO  US  80309-0001
(303)492-6221
Sponsor Congressional District: 02
Primary Place of Performance: University of Colorado Boulder
3100 Marine Street, Room 479
Boulder
CO  US  80303-1058
Primary Place of Performance
Congressional District:
02
Unique Entity Identifier (UEI): SPVKK1RC2MZ3
Parent UEI:
NSF Program(s): ECR-EDU Core Research,
IntgStrat Undst Neurl&Cogn Sys
Primary Program Source: 01001617DB NSF RESEARCH & RELATED ACTIVIT
04001617DB NSF Education & Human Resource

04001718DB NSF Education & Human Resource
Program Reference Code(s): 8089, 8091, 8551, 8816, 8817, 9251
Program Element Code(s): 798000, 862400
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.076

ABSTRACT

While traditional textbooks are designed to transmit information from the printed page to the learner, contemporary digital textbooks offer the opportunity to study learners as they interpret and process information being read. With a better understanding of a learner's state of mind, textbooks can make personalized recommendations for further study and review. How can the learner's state of mind be determined? Open a used printed textbook and the answer is clear: students feel compelled to engage with their texts by annotating key passages with highlights, tags, questions, and notes. Despite students' spontaneous desire to annotate as they read, this form of interaction has reaped few educational benefits in the past. At best, highlighted passages are re-read to study for exams, a strategy not nearly as effective as other strategies such as self-quizzing. This project will develop a new methodology that: assesses student knowledge level automatically based on annotations, transforms highlighted passages into appropriate study questions, and provides each student with well-timed, personalized review. Because the project is based on free, peer-reviewed, openly licensed materials from OpenStax that have been widely adopted at a range of institutions, particularly community colleges, the technology will reach beyond elite institutions to provide a broad spectrum of underserved students with access to a potentially powerful learning tool.

This project adopts a big-data approach that involves collecting annotations from a population of learners to draw inferences about individual learners. The project will determine how to exploit these data to model cognitive state, enabling the team to infer students' depth of understanding of facts and concepts, predict subsequent test performance, and perform interventions that improve learning outcomes. A tool will be developed that administers appropriately timed quizzes on material related to a student's highlights. A collaborative-filtering methodology will be employed that leverages population data to suggest specific passages for an individual to review. The proposed tool will reformulate selected passages into review questions that encourage the active reconstruction and elaboration of knowledge. The design and implementation of the tool will be informed by both randomized controlled studies within the innovative OpenStax textbook platform and coordinated laboratory studies. These studies will address basic scientific questions pertaining to why students annotate, how to improve their annotation skills, and techniques to optimize the use of annotations for guiding active review.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 15)
Attarian, M. "Transforming neural network representations to predict human judgments of similarity" Workshop on Shared Visual Representations in Human and Machine Intelligence (SVRHM 2020) , 2020 Citation Details
Beckage, Nicole M. and Mozer, Michael C. and Colunga, Eliana "Quantifying the role of vocabulary knowledge in predicting future word learning" IEEE transactions on cognitive and developmental systems , 2019 10.1109/TCDS.2019.2928023 Citation Details
Davidson, G. and Mozer, M. C. "Sequential mastery of multiple visual tasks: Networks naturally learn to learn and forget to forget" IEEE Computer Society Conference on Computer Vision and Pattern Recognition , 2020 Citation Details
Khajah, M. M. and Mozer, M. C. and Kelly, S. and Milne, B. "Boosting engagement with educational software using near wins" Nineteenth International Conference on Artificial Intelligence in Education , v.19 , 2018 Citation Details
Kim, D. Y. and Scott, T. R. and Mallick, D. and Mozer, M. C. "Using semantics of textbook highlights to predict student comprehension and knowledge retention" Proceedings of the Third International Workshop on Intelligent Textbooks (iTextbooks) , 2021 Citation Details
Kim, D. Y. and Winchell, A. and Water, A. E. and Grimaldi, P. J. and Baraniuk, R. and Mozer, M. C. "Inferring student comprehension from highlighting patterns in digital textbooks: An exploration in an authentic learning platform" Intelligent Textbooks 2020 , 2020 Citation Details
Kneusel, Ronald T. and Mozer, Michael C. "Improving Human-Machine Cooperative Visual Search With Soft Highlighting" ACM Transactions on Applied Perception , v.15 , 2017 10.1145/3129669 Citation Details
Montero, S. and Arora, A. and Kelly, S. and Milne, B. and Mozer, M. C. "Does deep knowledge tracing model interactions among skills?" Proceedings of the Eleventh International Conference on Educational Data Mining , v.11 , 2018 Citation Details
Mozer, Michael C. and Wiseheart, Melody and Novikoff, Timothy P. "Artificial intelligence to support human instruction" Proceedings of the National Academy of Sciences , v.116 , 2019 10.1073/pnas.1900370116 Citation Details
Ridgeway, Karl and Mozer, Michael C. and Bowles, Anita R. "Forgetting of Foreign-Language Skills: A Corpus-Based Analysis of Online Tutoring Software" Cognitive Science , v.41 , 2017 10.1111/cogs.12385 Citation Details
Roads, Brett D. and Mozer, Michael C. "Obtaining psychological embeddings through joint kernel and metric learning" Behavior Research Methods , v.51 , 2019 10.3758/s13428-019-01285-3 Citation Details
(Showing: 1 - 10 of 15)

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

While traditional textbooks are designed to transmit information from the printed page to the learner, contemporary digital textbooks offer the opportunity to study learners as they interpret and process information being read, make personalized recommendations for further study and review, and to thereby promote long-term retention and conceptual understanding. Our project explored this objective by instrumenting digital textbooks and e-readers to collect data from learners as they read and interact with the material. Two data sources were considered: (1) highlighting annotations that learners make as they read, indicating material the learner considers to be key content, and (2) and finger-scrolling actions using an electronic e-reader.

Although students believe that highlighting and subsequent review of the highlights will further their educational goals, the psychological literature provides little evidence of benefits. Nonetheless, students' choice of text for highlighting may serve as a window into their mental state—their level of comprehension, grasp of the key ideas, reading goals, and so on. We explored this hypothesis via an experiment in which 400 participants read three sections from a college-level biology text, briefly reviewed the text, and then took a quiz on the material. During initial reading, participants were able to highlight words, phrases, and sentences, and these highlights were displayed along with the complete text during the subsequent review. Consistent with past research, the amount of highlighted material is unrelated to quiz performance. However, we found that the specific content highlighted could help us to predict quiz performance. We explored a range of models and a range of highlighting representations to determine which combination yielded the best predictive accuracy. We also explored matrix factorization models that allowed us to show that individuals' highlighting pattern is informative of what they highlight elsewhere, which could be useful for content-recommendation engines.

We moved from laboratory experiments to an analysis of a large-scale corpus of highlighting data collected in a genuine educational context by a digital open-access platform, OpenStax. The corpus consisted of data from over 11,000 students and nearly 900 sections from College Biology and Physics textbooks. In addition to highlighting data, the corpus provided scores on delayed quizzes associated with the sections. We considered two approaches to encoding highlighting patterns. A positional representation indicated where highlights were made in the stream of text for a particular section. A semantic representation was based on a state-of-the-art deep-learning sentence embedding technique (SBERT) that captures the content-based similarity between quiz questions and highlighted (as well as non-highlighted) sentences in the text. We construct regression models that include latent variables for student skill level and question difficulty and augment the models with highlighting features. We find that both positional and semantic highlighting features reliably boost model performance, with semantic features being the more predictive. We conduct experiments that validate models on held-out questions, students, and student-questions and find strong generalization for the latter two but not for held-out questions. Surprisingly, highlighting features improve models for questions at all levels of the Bloom taxonomy, from straightforward recall questions to inferential synthesis/evaluation/creation questions.

Highlights explain only a portion of observed variability in performance. As is typical in big-data applications, one seeks multiple weak predictor variables which in combination can make strong predictions. Toward this goal, we examined readers' scrolling patterns on a personal e-reader (table, phone). We collected a data set consisting of 20 readers reading 20 newspaper articles. The data indicated the exact timing of finger motions and flicks, allowing us to determine how long text appeared on the screen, and thus allowed us to infer content-specific reading rates. We found that we could predict difficulty of content from reading rates and we could identify individual differences in scrolling patterns. In follow up work, we plan to explore the relationship between scrolling patterns and comprehension and retention.

Our long-term goal is to design digital textbooks that serve not only as conduits of information into the reader’s mind but also allow us to draw inferences about the reader at a point where interventions may increase the effectiveness of the material. Interventions might include alterting the instructor, posing questions that allow students to evaluate their understanding, or guiding students to improve by suggesting content to review.


Last Modified: 10/29/2021
Modified by: Michael C Mozer

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page