Award Abstract # 1639995
E2CDA: Type II: Self-Adaptive Reservoir Computing with Spiking Neurons: Learning Algorithms and Processor Architectures

NSF Org: CCF
Division of Computing and Communication Foundations
Recipient: TEXAS A&M ENGINEERING EXPERIMENT STATION
Initial Amendment Date: August 26, 2016
Latest Amendment Date: July 5, 2018
Award Number: 1639995
Award Instrument: Continuing Grant
Program Manager: Sankar Basu
sabasu@nsf.gov
 (703)292-7843
CCF
 Division of Computing and Communication Foundations
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: September 1, 2016
End Date: September 30, 2019 (Estimated)
Total Intended Award Amount: $332,742.00
Total Awarded Amount to Date: $332,742.00
Funds Obligated to Date: FY 2016 = $110,914.00
FY 2017 = $6,156.00

FY 2018 = $0.00
History of Investigator:
  • Peng Li (Principal Investigator)
    lip@ece.ucsb.edu
Recipient Sponsored Research Office: Texas A&M Engineering Experiment Station
3124 TAMU
COLLEGE STATION
TX  US  77843-3124
(979)862-6777
Sponsor Congressional District: 10
Primary Place of Performance: Texas A&M Engineering Experiment Station
College Station
TX  US  77843-3128
Primary Place of Performance
Congressional District:
10
Unique Entity Identifier (UEI): QD1MX6N5YTN4
Parent UEI: QD1MX6N5YTN4
NSF Program(s): Energy Efficient Computing: fr
Primary Program Source: 01001617DB NSF RESEARCH & RELATED ACTIVIT
01001718DB NSF RESEARCH & RELATED ACTIVIT

01001819DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7945
Program Element Code(s): 015Y00
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

While computing has become increasingly data centric across many disciplines, conventional computer architectures have limited potential in meeting the escalating performance and energy efficiency needs in this era of data-driven science and engineering. This project aims to develop brain-inspired neural models of computation and adaptive processor architectures to enable intelligent data processing and learning in a wide range of applications. While being strongly interdisciplinary, this work will bridge neuroscience, artificial neural networks, computer architecture, and hardware engineering. The planned research will provide rich training and educational opportunities to students, and produce new curriculum. Research participation from undergraduate and underrepresented students will be promoted. The outcomes of this project will be broadly disseminated. Research collaboration with the US industry will be actively pursued via interaction with the Semiconductor Research Corporation.

This work is aimed at attaining brain-like learning performance by imitating how the brain represents, processes, and learns from information, and more specifically, by developing models of computation based on the third-generation spiking neural networks, and efficient adaptive processor architectures. Within the framework of so called reservoir computing, the proposed neural models mimic key characteristics of the brain such as information processing based on spike timing. Furthermore, this project will develop brain-inspired learning mechanisms to allow training of complex recurrent spiking neural networks. Self-adaptive processor architectures with integrated on-chip learning, light-weight runtime learning performance prediction, and energy management will be developed to maximize system energy efficiency while providing a guarantee of performance.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Jin, Y and Li, P "Calcium-modulated supervised spike-timing-dependent plasticity for readout training and sparsification of the liquid state machine" Proceedings of ... International Joint Conference on Neural Networks , 2017 Citation Details
Liu, Y and Jin, Y and Li, P "Exploring sparsity of firing activities and clock gating for energy-efficient recurrent spiking neural processors" Proceedings - International Symposium on Low Power Electronics and Design , 2017 Citation Details
Liu, Yu and Jin, Yingyezhe and Li, Peng "Online Adaptation and Energy Minimization for Hardware Recurrent Spiking Neural Networks" ACM Journal on Emerging Technologies in Computing Systems , v.14 , 2018 10.1145/3145479 Citation Details
Mahadevuni, A and Li, P "Navigating mobile robots to target in near shortest time using reinforcement learning with spiking neural networks" Proceedings of ... International Joint Conference on Neural Networks , 2017 Citation Details
Shim, Myung S and Li, Peng "Biologically inspired reinforcement learning for mobile robot collision avoidance" Proceedings of ... International Joint Conference on Neural Networks , 2017 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page