Award Abstract # 1317788
NRI: Small: Collaborative Research: Active Sensing for Robotic Cameramen

NSF Org: IIS
Division of Information & Intelligent Systems
Recipient: REGENTS OF THE UNIVERSITY OF MINNESOTA
Initial Amendment Date: September 12, 2013
Latest Amendment Date: April 29, 2015
Award Number: 1317788
Award Instrument: Standard Grant
Program Manager: Jie Yang
jyang@nsf.gov
 (703)292-4768
IIS
 Division of Information & Intelligent Systems
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: September 15, 2013
End Date: August 31, 2017 (Estimated)
Total Intended Award Amount: $292,000.00
Total Awarded Amount to Date: $300,000.00
Funds Obligated to Date: FY 2013 = $292,000.00
FY 2015 = $8,000.00
History of Investigator:
  • Ibrahim Isler (Principal Investigator)
    isler@cs.utexas.edu
Recipient Sponsored Research Office: University of Minnesota-Twin Cities
2221 UNIVERSITY AVE SE STE 100
MINNEAPOLIS
MN  US  55414-3074
(612)624-5599
Sponsor Congressional District: 05
Primary Place of Performance: University of Minnesota-Twin Cities
200 Union Street SE
Minneapolis
MN  US  55455-0159
Primary Place of Performance
Congressional District:
05
Unique Entity Identifier (UEI): KABJZBBJ4B54
Parent UEI:
NSF Program(s): Robust Intelligence,
NRI-National Robotics Initiati
Primary Program Source: 01001314DB NSF RESEARCH & RELATED ACTIVIT
01001516DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7923, 8086, 9251
Program Element Code(s): 749500, 801300
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

With advances in camera technologies, and as cloud storage, network bandwidth and protocols become available, visual media are becoming ubiquitous. Video recording became de facto universal means of instruction for a wide range of applications such as physical exercise, technology, assembly, or cooking. This project addresses the scientific and technological challenges of video shooting in terms of coverage and optimal views planning while leaving high level aspects including creativity to the video editing and post-production stages.

Camera placement and novel view selection challenges are modeled as optimization problems that minimize the uncertainty in the location of actors and objects, maximize coverage and effective appearance resolution, and optimize object detection for the sake of semantic annotation of the scene. New probabilistic models capture long range correlations when the trajectories of actors are only partially observable. Quality of potential novel views is modeled in terms of resolution that is optimized by maximizing the coverage of a 3D orientation histogram while an active view selection process for object detection minimizes a dynamic programming objective function capturing the loss due to classification error as well as the resources spent for each view.

The project advances active sensing and perception and provides the technology for further automation on video capturing. Such technology has broader impact on the production of education videos for online courses as well as in telepresence applications. Research results are integrated into robotics and digital media programs addressing K-12 students.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Nicolai Hani and Volkan Isler "Visual Servoing in Orchard Settings" 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems , 2016
Patrick Plonski and Volkan Isler "Approximation Algorithms for Tours of Height-varying View Cones" In Proceedings of the Workshop on Algorithmic Foundations on Robotics (WAFR 2016) , 2016
Pravakar Roy and Volkan Isler "Active View Planning for Counting Apples in Orchards" In Proc. International Conference on Intelligent Robots and Systems (IROS) , 2017
Pravakar Roy and Volkan Isler "Surveying Apple Orchards with a Monocular Vision System" 12th Conference on Automation Science and Engineering , 2016

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Many computer vision tasks such as scene coverage and object detection rely on the acquisition of high quality images related to the task. In this project, we explored mechanisms to actively control cameras to ensure the capture of such images.  The algorithms can be used at a variety of scales ranging from planning the trajectories of a camera mounted on an aerial vehicle inspecting a field to the case of a camera mounted on a robot arm for visual inspection.

In terms of intellectual merit, the primary contributions of the project  included an algorithm with theoretical performance guaranteess to visit a given set of viewing cones in the shortest amount of time. This problem arises in a number of settings including visual inspection of farms where the cones correspond to the viewing regions of areas of interest (e.g. potential infestations). We also developed a novel visual servoing platform where a camera is mounted on a robot arm and is controlled by interpreting the images acquired in an online fashion. We demonstrated the algorithm in an orchard inspection task which is challenging because the scene is not static (due to wind) and light conditions can be varying drastically: there can be bright, direct sunlight or many shadows.

In terms of broader impact, such view planning algorithms can be useful in a variety of applications of societal importance. These include agriculture as mentioned above, industrial inspection or search and rescue. Graduate and undergraduate students (some of whom were underrepresented minorities) were trained as part of this project.


Last Modified: 11/30/2017
Modified by: Ibrahim V Isler

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page