Award Abstract # 1626008
MRI: Development of an Observatory for Quantitative Analysis of Collective Behavior in Animals

NSF Org: CNS
Division Of Computer and Network Systems
Recipient: TRUSTEES OF THE UNIVERSITY OF PENNSYLVANIA, THE
Initial Amendment Date: September 16, 2016
Latest Amendment Date: September 16, 2016
Award Number: 1626008
Award Instrument: Standard Grant
Program Manager: Rita Rodriguez
CNS
 Division Of Computer and Network Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: October 1, 2016
End Date: September 30, 2019 (Estimated)
Total Intended Award Amount: $339,174.00
Total Awarded Amount to Date: $339,174.00
Funds Obligated to Date: FY 2016 = $339,174.00
History of Investigator:
  • Kostas Daniilidis (Principal Investigator)
    kostas@cis.upenn.edu
  • Jianbo Shi (Co-Principal Investigator)
  • Marc Schmidt (Co-Principal Investigator)
  • Daniel Lee (Co-Principal Investigator)
  • Danielle Bassett (Co-Principal Investigator)
Recipient Sponsored Research Office: University of Pennsylvania
3451 WALNUT ST STE 440A
PHILADELPHIA
PA  US  19104-6205
(215)898-7293
Sponsor Congressional District: 03
Primary Place of Performance: University of Pennsylvania
3330 Walnut STreet
Philadelphia
PA  US  19104-3409
Primary Place of Performance
Congressional District:
03
Unique Entity Identifier (UEI): GM1XX56LEP58
Parent UEI: GM1XX56LEP58
NSF Program(s): Major Research Instrumentation
Primary Program Source: 01001617DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1189
Program Element Code(s): 118900
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

This project, developing a new instrument to enable an accurate quantitative analysis of the movement of animals and vocal expressions in real world scenes, aims to facilitate innovative research in the study of animal behavior and neuroscience in complex realistic environments. While much progress has been made investigating brain mechanisms of behavior, these have been limited primarily to studying individual subjects in relatively simple settings. For many social species, including humans, understanding neurobiological processes within the confines of these more complex environments is critical because their brains have evolved to perceive and evaluate signals within a social context. Indeed, today's advances in video capture hardware and storage and in algorithms in computer vision and network science make this facilitation with animals possible. Past work has relied on subjective and time-consuming observations from video streams, which suffer from imprecision, low dimensionality, and the limitations of the expert analyst's sensory discriminability. This instrument will not only automate the process of detecting behaviors but also provide an exact numeric characterization in time and space for each individual in the social group. While not explicitly part of the instrument, the quantitative description provided by our system will allow the ability to correlate social context with neural measurements, a task that may only be accomplished when sufficient spatiotemporal precision has been achieved.

The instrument enables research in the behavioral and neural sciences and development of novel algorithms in computer vision and network theory. In the behavioral sciences, the instrumentation allows the generation of network models of social behavior in small groups of animals or humans that can be used to ask questions that can range from how the dynamics of the networks influence sexual selection, reproductive success, and even health messaging to how vocal decision making in individuals gives rise to social dominance hierarchies. In the neural sciences, the precise spatio-temporal information the system would provide can be used to evaluate the neural bases of sensory processing and behavioral decision under precisely defined social contexts. Sensory responses to a given vocal stimulus, for example, can be evaluated by the context in which the animal heard the stimulus and both his and the sender's prior behavioral history in the group. In computer vision, we propose novel approaches for the calibration of multiple cameras "in the wild", the combination of appearance and geometry for the extraction of exact 3D pose and body parts from video, the learning of attentional focus among animals in a group, and the estimation of sound source and the classification of vocalizations. New approaches will be used on hierarchical discovery of behaviors in graphs, the incorporation of interactions beyond the pairwise level with simplicial complices, and a novel theory of graph dynamics for the temporal evolution of social behavior. The instrumentation benefits behavioral and neural scientists. Therefore, the code and algorithms developed will be open-source so that the scientific community can extend them based on the application. The proposed work also impacts computer vision and network science because the fundamental algorithms designed should advance the state of the art. For performance evaluation of other computer vision algorithms, established datasets will be employed.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Esteves, Carlos and Daniilidis, Kostas and Makadia, Ameesh and Allec-Blanchette, Christine "Learning SO(3) Equivariant Representations with Spherical CNNs" Proceedings of the European Conference on Computer Vision (ECCV) , 2018
G Pavlakos, X Zhou, KG Derpanis, K Daniilidis "Coarse-to-Fine Volumetric Prediction for Single-Image 3D Human Pose" IEEE Computer Vision Pattern Recognition , 2017
Nikos Kolotouros, Georgios Pavlakos, Kostas Daniilidis "Convolutional Mesh Regression for Single-Image Human Shape Reconstruction" IEEE Computer Vision Pattern Recognition , 2020
Pavlakos, Georgios and Zhu, Luyang and Zhou, Xiaowei and Daniilidis, Kostas "Learning to Estimate 3D Human Pose and Shape from a Single Color Image" IEEE Comp. Vis. Pattern Recog. (CVPR) , 2018
Pavlakos, G., Zhou, X., Derpanis, K. G., & Daniilidis, K. "Harvesting Multiple Views for Marker-less 3D Human Pose Annotations" IEEE Computer Vision Pattern Recognition , 2017

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Neural and behavioral sciences are entering a phase where new advances in artificial intelligence now make it a possibility to quantify behaviors within complex naturalistic environments and relate them to underlying neural mechanisms. Because social animals, including humans, have evolved to perceive and evaluate signals within a social context, the ability to link neural function with the precise social context that surrounds the individual promises to be transformative. 

 

This MRI project produced the world?s first computational observatory for birds, a smart aviary. Birds have widespread impacts on our everyday lives, ranging from agriculture to science and the arts. Although studying birds' motion is a critical step in many of these areas, current methods were not able to robustly estimate pose and shape of birds maneuvering in three dimensions and anytime during the day and any season.  

 

To facilitate continuous monitoring of behavior  the aviary was equipped with eight cameras and 24 microphones that can record life in the aviary 24/7. Inspired by the unparalleled performance of articulated mesh-based models for pose and shape estimation in humans, a multi-view dataset was collected that includes over 1000 images of 15 cowbirds housed together in an outdoor aviary. A novel articulated 3D bird model was introduced as well as a deep learning pipeline that estimates the parameters of the 3D bird model from segmented bird silhouettes and characteristic keypoints. The dataset can be used by biologists and neuroscientists in studying bird posture in social settings as well as the social graph of birds based on their position, orientation, and singing.


 

 


Last Modified: 02/13/2020
Modified by: Kostas Daniilidis

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page