
NSF Org: |
CNS Division Of Computer and Network Systems |
Recipient: |
|
Initial Amendment Date: | September 16, 2016 |
Latest Amendment Date: | September 16, 2016 |
Award Number: | 1626008 |
Award Instrument: | Standard Grant |
Program Manager: |
Rita Rodriguez
CNS Division Of Computer and Network Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | October 1, 2016 |
End Date: | September 30, 2019 (Estimated) |
Total Intended Award Amount: | $339,174.00 |
Total Awarded Amount to Date: | $339,174.00 |
Funds Obligated to Date: |
|
History of Investigator: |
|
Recipient Sponsored Research Office: |
3451 WALNUT ST STE 440A PHILADELPHIA PA US 19104-6205 (215)898-7293 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
3330 Walnut STreet Philadelphia PA US 19104-3409 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | Major Research Instrumentation |
Primary Program Source: |
|
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
This project, developing a new instrument to enable an accurate quantitative analysis of the movement of animals and vocal expressions in real world scenes, aims to facilitate innovative research in the study of animal behavior and neuroscience in complex realistic environments. While much progress has been made investigating brain mechanisms of behavior, these have been limited primarily to studying individual subjects in relatively simple settings. For many social species, including humans, understanding neurobiological processes within the confines of these more complex environments is critical because their brains have evolved to perceive and evaluate signals within a social context. Indeed, today's advances in video capture hardware and storage and in algorithms in computer vision and network science make this facilitation with animals possible. Past work has relied on subjective and time-consuming observations from video streams, which suffer from imprecision, low dimensionality, and the limitations of the expert analyst's sensory discriminability. This instrument will not only automate the process of detecting behaviors but also provide an exact numeric characterization in time and space for each individual in the social group. While not explicitly part of the instrument, the quantitative description provided by our system will allow the ability to correlate social context with neural measurements, a task that may only be accomplished when sufficient spatiotemporal precision has been achieved.
The instrument enables research in the behavioral and neural sciences and development of novel algorithms in computer vision and network theory. In the behavioral sciences, the instrumentation allows the generation of network models of social behavior in small groups of animals or humans that can be used to ask questions that can range from how the dynamics of the networks influence sexual selection, reproductive success, and even health messaging to how vocal decision making in individuals gives rise to social dominance hierarchies. In the neural sciences, the precise spatio-temporal information the system would provide can be used to evaluate the neural bases of sensory processing and behavioral decision under precisely defined social contexts. Sensory responses to a given vocal stimulus, for example, can be evaluated by the context in which the animal heard the stimulus and both his and the sender's prior behavioral history in the group. In computer vision, we propose novel approaches for the calibration of multiple cameras "in the wild", the combination of appearance and geometry for the extraction of exact 3D pose and body parts from video, the learning of attentional focus among animals in a group, and the estimation of sound source and the classification of vocalizations. New approaches will be used on hierarchical discovery of behaviors in graphs, the incorporation of interactions beyond the pairwise level with simplicial complices, and a novel theory of graph dynamics for the temporal evolution of social behavior. The instrumentation benefits behavioral and neural scientists. Therefore, the code and algorithms developed will be open-source so that the scientific community can extend them based on the application. The proposed work also impacts computer vision and network science because the fundamental algorithms designed should advance the state of the art. For performance evaluation of other computer vision algorithms, established datasets will be employed.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
Neural and behavioral sciences are entering a phase where new advances in artificial intelligence now make it a possibility to quantify behaviors within complex naturalistic environments and relate them to underlying neural mechanisms. Because social animals, including humans, have evolved to perceive and evaluate signals within a social context, the ability to link neural function with the precise social context that surrounds the individual promises to be transformative.
This MRI project produced the world?s first computational observatory for birds, a smart aviary. Birds have widespread impacts on our everyday lives, ranging from agriculture to science and the arts. Although studying birds' motion is a critical step in many of these areas, current methods were not able to robustly estimate pose and shape of birds maneuvering in three dimensions and anytime during the day and any season.
To facilitate continuous monitoring of behavior the aviary was equipped with eight cameras and 24 microphones that can record life in the aviary 24/7. Inspired by the unparalleled performance of articulated mesh-based models for pose and shape estimation in humans, a multi-view dataset was collected that includes over 1000 images of 15 cowbirds housed together in an outdoor aviary. A novel articulated 3D bird model was introduced as well as a deep learning pipeline that estimates the parameters of the 3D bird model from segmented bird silhouettes and characteristic keypoints. The dataset can be used by biologists and neuroscientists in studying bird posture in social settings as well as the social graph of birds based on their position, orientation, and singing.
Last Modified: 02/13/2020
Modified by: Kostas Daniilidis
Please report errors in award information by writing to: awardsearch@nsf.gov.