Award Abstract # 2223839
EFRI BRAID: Principles of sleep-dependent memory consolidation for adaptive and continual learning in artificial intelligence

NSF Org: EFMA
Office of Emerging Frontiers in Research and Innovation (EFRI)
Recipient: UNIVERSITY OF CALIFORNIA, SAN DIEGO
Initial Amendment Date: September 16, 2022
Latest Amendment Date: September 16, 2022
Award Number: 2223839
Award Instrument: Standard Grant
Program Manager: Edda Thiels
ethiels@nsf.gov
 (703)292-8167
EFMA
 Office of Emerging Frontiers in Research and Innovation (EFRI)
ENG
 Directorate for Engineering
Start Date: October 1, 2022
End Date: September 30, 2026 (Estimated)
Total Intended Award Amount: $2,000,000.00
Total Awarded Amount to Date: $2,000,000.00
Funds Obligated to Date: FY 2022 = $2,000,000.00
History of Investigator:
  • Maksim Bazhenov (Principal Investigator)
    mbazhenov@ucsd.edu
  • Brian Smith (Co-Principal Investigator)
  • Hong Lei (Co-Principal Investigator)
  • Theodore Pavlic (Co-Principal Investigator)
  • Giri Prashanth Krishnan (Co-Principal Investigator)
Recipient Sponsored Research Office: University of California-San Diego
9500 GILMAN DR
LA JOLLA
CA  US  92093-0021
(858)534-4896
Sponsor Congressional District: 50
Primary Place of Performance: University of California, San Diego
9500 GILMAN DR
SAN DIEGO
CA  US  92093-0021
Primary Place of Performance
Congressional District:
50
Unique Entity Identifier (UEI): UYTTZT6G9DT1
Parent UEI:
NSF Program(s): EFRI Research Projects,
Modulation
Primary Program Source: 01002223DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 8091
Program Element Code(s): 763300, 771400
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.041, 47.074

ABSTRACT

Artificial Neural Networks (ANNs) are a form of Artificial Intelligence (AI) used in applications from self-driving cars to medicine to robotic systems. Although they can match and even exceed human performance on some learning tasks, they fail to reproduce important characteristics of the human mind, such as quick and continual learning, transfer of knowledge to the new tasks, and energy efficiency. Indeed, ANNs commonly forget what they knew when new information is learned, and so they need to be taught from scratch to re-learn. In real-life applications in changing and unpredictable environments, ANNs can only reach near human-level performance if they are trained on all possible scenarios that could happen in life. This level of training is inefficient and unrealistic. In natural brains, sleep is thought to be important for intelligence. During sleep, the brain repeats and replays what was learned during the day, and this helps to prevent forgetting, to generalize to new situations, and to create new emerging knowledge. In this project, principles learned from the biology of sleep will be used to develop powerful new algorithms for AI systems that can learn continuously and from few examples, transfer knowledge learned from old tasks to new tasks, and be robust and efficient. Because AI and ANNs are so fundamental to the modern world, from healthcare to electronics to national defense, this project has the potential to make a significant societal impact. The project takes a multi-disciplinary approach and supports broader participation of underrepresented groups in STEM research through a range of educational activities focused on high school, undergraduate, including community college, and graduate students.

This project aims to translate insights from the study of sleep to improvements in deep-learning systems necessary for continual learning, generalization, and transfer of knowledge in artificial intelligence (AI). Taking advantage of the architectural similarities between information processing in ANNs and the honeybee brain, the main goals of this project are: (a) to characterize multi-phasic sleep in the honeybee brain in vivo and in biophysical in silico models in fine spatio-temporal detail to reveal the critical principles of the role of sleep in memory consolidation, and (b) to apply these results to support the development of novel machine-learning algorithms for adaptive and continual learning in complex and dynamic environments. The study will develop an empirically grounded theory of multi-phasic sleep that will be then applied to artificial neural networks, and the process of developing ?sleep for AI? will help to strengthen connections between engineering, computational neuroscience, and neuroethology for researchers at a range of career stages. To accomplish this goal, the project team also plans a four-tiered educational approach targeting students in high schools, community colleges, bachelor?s degree programs, and graduate-level programs to introduce a wider range of students to the topics in AI, sleep biology, and computational neuroscience.

This project is funded jointly by the Emerging Frontiers in Research and Innovation Brain-Inspired Dynamics for Engineering Energy-Efficient Circuits and Artificial Intelligence Program of the Engineering Directorate and the Neural Systems/Modulation Program of the Biological Sciences Directorate.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Bazhenov, Anthony and Dewasurendra, Pahan and Krishnan, Giri and Delanois, Jean Erik "Sleep-Like Unsupervised Replay Improves Performance When Data Are Limited or Unbalanced (Student Abstract)" Proceedings of the AAAI Conference on Artificial Intelligence , v.38 , 2024 https://doi.org/10.1609/aaai.v38i21.30420 Citation Details
Bazhenov, Anthony and Dewasurendra, Pahan and Krishnan, Giri P and Delanois, Jean Erik "Unsupervised Replay Strategies for Continual Learning with Limited Data" , 2024 https://doi.org/10.1109/IJCNN60899.2024.10650116 Citation Details
Delanois, Jean Erik and Ahuja, Aditya and Krishnan, Giri P and Tadros, Timothy and McAuley, Julian and Bazhenov, Maxim "Improving Robustness of Convolutional Networks Through Sleep-Like Replay" , 2023 https://doi.org/10.1109/ICMLA58977.2023.00043 Citation Details
Golden, Ryan and Delanois, Jean Erik and Sanda, Pavel and Bazhenov, Maxim "Sleep prevents catastrophic forgetting in spiking neural networks by forming a joint synaptic weight representation" PLOS Computational Biology , v.18 , 2022 https://doi.org/10.1371/journal.pcbi.1010628 Citation Details
Hong, Jinyung and Park, Keun Hee and Pavlic, Theodore P "Concept-Centric Transformers: Enhancing Model Interpretability through Object-Centric Concept Learning within a Shared Global Workspace" , 2024 https://doi.org/10.1109/WACV57701.2024.00481 Citation Details
Hong, Jinyung and Pavlic, Theodore P "Randomly Weighted Neuromodulation in Neural Networks Facilitates Learning of Manifolds Common Across Tasks" , 2023 Citation Details
Tadros, Timothy and Krishnan, Giri P. and Ramyaa, Ramyaa and Bazhenov, Maxim "Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks" Nature Communications , v.13 , 2022 https://doi.org/10.1038/s41467-022-34938-7 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page