
NSF Org: |
IIS Division of Information & Intelligent Systems |
Recipient: |
|
Initial Amendment Date: | September 11, 2012 |
Latest Amendment Date: | August 7, 2013 |
Award Number: | 1218749 |
Award Instrument: | Continuing Grant |
Program Manager: |
Sylvia Spengler
sspengle@nsf.gov (703)292-7347 IIS Division of Information & Intelligent Systems CSE Directorate for Computer and Information Science and Engineering |
Start Date: | October 1, 2012 |
End Date: | September 30, 2017 (Estimated) |
Total Intended Award Amount: | $299,979.00 |
Total Awarded Amount to Date: | $299,979.00 |
Funds Obligated to Date: |
FY 2013 = $203,207.00 |
History of Investigator: |
|
Recipient Sponsored Research Office: |
926 DALNEY ST NW ATLANTA GA US 30318-6395 (404)894-4819 |
Sponsor Congressional District: |
|
Primary Place of Performance: |
GA US 30332-0002 |
Primary Place of
Performance Congressional District: |
|
Unique Entity Identifier (UEI): |
|
Parent UEI: |
|
NSF Program(s): | Info Integration & Informatics |
Primary Program Source: |
01001314DB NSF RESEARCH & RELATED ACTIVIT |
Program Reference Code(s): |
|
Program Element Code(s): |
|
Award Agency Code: | 4900 |
Fund Agency Code: | 4900 |
Assistance Listing Number(s): | 47.070 |
ABSTRACT
Many modern applications ranging from computer vision to biology require modeling and inferring high-dimensional continuous variables based on distributions with multimodality, skewness, and rich latent structures. Most existing models in this regime rely heavily on parametric assumptions where the components of the model are typically assumed to be discrete or multivariate Gaussian, or the relations between variables are linear, which may be very different from the actual data generating processes. Furthermore, existing algorithms for discovering the latent dependency structures and learning the latent parameters largely are restricted to local search heuristics such as expectation maximization. Conclusions inferred under these restricted assumptions and suboptimal solutions can be misleading, if the underlying assumptions are violated or if the suboptimal solutions differ greatly from the globally optimal ones. This project aims to develop a novel framework which can (i) discover and take advantage of latent structures in the data, while (ii) allowing parts to handle near-arbitrary distributions, and (iii) allowing the models to scale to modern massive datasets in a local-minimum-free fashion.
The key innovation in the project is a novel nonparametric latent variable modeling framework based on kernel embedding of distributions. The basic idea is to map distributions into infinite dimensional feature spaces using kernels, such that subsequent comparisons and manipulations of distributions can be achieved via feature space operations, such as inner products, distances, projections, linear transformations and spectral analysis. Conceptually, the framework represents components from latent variable models, such as marginal distributions over a single variable, joint distributions over variable pairs, triplets and more variables, as infinite dimensional vectors, matrices, tensors and high-order tensors respectively. Probabilistic relations between these components, i.e., conditional distributions, Sum Rule, Product Rule etc. become linear transformations and relations between these feature space components.
The framework supports modeling data with diverse statistical features without the need for making restrictive assumptions about the type of distributions and relations. It supports the application of a large pool of linear and multi-linear algebraic (tensor) tools for addressing challenging graphical model problems in the presence of latent variables, including structure discovery, inference, parameter learning and latent feature extraction. The framework applies not only to general continuous variables, but also to variables that take values on strings, graphs, groups, compact manifolds, and other domains on which kernels may be defined.
Besides advancing the state of the art in machine learning,the new non-parametric methods resulting from the project find applications in image data and understanding and gene expression data analysis. It also contributes to research-based training of graduate and undergraduate students at Georgia Tech and CMU.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH
Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external
site maintained by the publisher. Some full text articles may not yet be available without a
charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from
this site.
PROJECT OUTCOMES REPORT
Disclaimer
This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.
This project developed a novel nonparametric latent vari-able modeling framework based on kernel embedding of distributions. The key idea is to map distributions into infinite dimensional feature spaces using kernels, such that subsequent comparisons and manipulations of distributions can be achieved via feature space operations, such as inner products, distances,projections, linear transformations and spectral analysis. The developed method can be used to address a range of latent variablemodel problems, including latent structure discovery, latent parameter learning, inference with latent variablesand latent feature extraction.
The developed framework allows us to model data with diverse statistical features without the need formaking restrictive assumptions about the type of distributions and relations. Furthermore, it allows us to apply a large pool of linear and multi-linear algebraic (tensor) tools for addressing challenging graphical model prob-lems in the presence of latent variables, including structure discovery, inference, parameter learning, and latentfeature extraction. For instance, by making novel use of the spectral properties of the embedded distributions,we can design fast and local-minimum-free algorithms for discovering latent structures and learning latent pa-rameters. Another advantage of our framework is that it applies not only to general continuous variables, but also generalizes to variables which may take values on strings, graphs, groups, compact manifolds, and otherdomains on which kernels may be defined.
This project has advanced the principles and technologies of latent structure analysisfor high-dimensional data with general distributions. It is a necessary step towards the ultimate, long term goalof understanding and exploiting latent structures and rich diverse statistical features prevalent in many modernapplications, such as computer vision, computational biology, social sciences, music research, and material sciences.
As an interdisciplinary research effort, this project has not only led to development of new methodology, but also provide rich opportunities formulti-disciplinary educational and research training, at both undergraduate and graduate levels.
Last Modified: 01/01/2018
Modified by: Le Song
Please report errors in award information by writing to: awardsearch@nsf.gov.