Award Abstract # 2104319
CDSE: Collaborative: Cyber Infrastructure to Enable Computer Vision Applications at the Edge Using Automated Contextual Analysis

NSF Org: OAC
Office of Advanced Cyberinfrastructure (OAC)
Recipient: LOYOLA UNIVERSITY OF CHICAGO
Initial Amendment Date: April 22, 2021
Latest Amendment Date: August 30, 2021
Award Number: 2104319
Award Instrument: Standard Grant
Program Manager: Sheikh Ghafoor
sghafoor@nsf.gov
 (703)292-7116
OAC
 Office of Advanced Cyberinfrastructure (OAC)
CSE
 Directorate for Computer and Information Science and Engineering
Start Date: September 1, 2021
End Date: August 31, 2025 (Estimated)
Total Intended Award Amount: $174,749.00
Total Awarded Amount to Date: $209,624.00
Funds Obligated to Date: FY 2021 = $209,624.00
History of Investigator:
  • George Thiruvathukal (Principal Investigator)
    gkt@cs.luc.edu
Recipient Sponsored Research Office: Loyola University of Chicago
820 N MICHIGAN AVE
CHICAGO
IL  US  60611-2147
(773)508-2471
Sponsor Congressional District: 05
Primary Place of Performance: Loyola University of Chicago
1032 W. Sheridan Road
Chicago
IL  US  60660-1537
Primary Place of Performance
Congressional District:
09
Unique Entity Identifier (UEI): CVNBL4GDUKF3
Parent UEI:
NSF Program(s): CDS&E
Primary Program Source: 01002122DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1504, 8084, 026Z
Program Element Code(s): 808400
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Digital cameras are deployed as network edge devices, gathering visual data for such tasks as autonomous driving, traffic analysis, and wildlife observation. Analyzing the vast amount of visual data is a challenge. Existing computer vision methods require fast computers that are beyond the computational capabilities of many edge devices. This project aims to improve the efficiency of computer vision methods so that they can run on battery-powered edge devices. Based on the visual data and complementary metadata (e.g., geographical location, local time), the project first extracts contextual information (such as a city street is expected to be busy at rush hour). The contextual information can help assist determine whether analysis results are correct. For example, a wild animal is not expected on a city street. Moreover, contextual information can improve efficiency. Only certain pixels need to be analyzed (pixels on the road are useful for detecting cars, while pixels in the sky are not) and this can significantly reduce the amount of computation, thus enabling analysis on edge devices. This project constructs a cyberinfrastructure for three services: (1) understand contextual information to reduce the search space of analysis methods, (2) reduce computation by considering only necessary pixels, and (3) automate evaluation of analysis results based on the contextual information without human effort.

Understanding contextual information is achieved by using background segmentation, GPS-location-dependent logic, and image depth maps. Background analysis leverages semantic segmentation and analysis over time to identify the background pixels and then generate inference rules via a background-implies-foreground relationship. If a pixel is consistently marked by the same semantic label across a long period of time, this pixel is classified as a background pixel. The background information can infer certain types of foreground objects. For example, if the background is city streets, the foreground objects can be vehicles or pedestrians; if a bison is detected, this is likely a mistake. This project processes only the foreground pixels by adding masks to the neural network layers. Masking convolution can substantially reduce the amount of computation with no loss of accuracy and no additional training is needed. Meanwhile, hierarchical neural networks can skip sections of a model based on context. For example, pixels in the sky only need to be processed by the hierarchy nodes that classify airplanes. The project provides an online service that can accept input data and analysis programs for automatic evaluation of the programs, without human created labels. The evaluation is based on the correlations of background and foreground objects.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Goel, Abhinav and Tung, Caleb and Eliopoulos, Nick and Thiruvathukal, George K. and Wang, Amy and Lu, Yung-Hsiang and Davis, James C. "Tree-Based Unidirectional Neural Networks for Low-Power Computer Vision" IEEE Design & Test , v.40 , 2023 https://doi.org/10.1109/MDAT.2022.3217016 Citation Details
Allcroft, Shane and Metwaly, Mohammed and Berg, Zachery and Ghodgaonkar, Isha and Bordwell, Fischer and Zhao, Xinxin and Liu, Xinglei and Xu, Jiahao and Chakraborty, Subhankar and Banna, Vishnu and Chinnakotla, Akhil and Goel, Abhinav and Tung, Caleb and "Observing Human Mobility Internationally During COVID-19" Computer , v.56 , 2023 https://doi.org/10.1109/MC.2022.3175751 Citation Details
Tung, Caleb and Goel, Abhinav and Bordwell, Fischer and Eliopoulos, Nick John and Hu, Xiao and Thiruvathukal, George K and Lu, Yung-Hsiang "Why Accuracy Is Not Enough: The Need for Consistency in Object Detection" IEEE MultiMedia , 2022 https://doi.org/10.1109/MMUL.2022.3175239 Citation Details
Tung, Caleb and Eliopoulos, Nicholas and Jajal, Purvish and Ramshankar, Gowri and Yang, Cheng-Yun and Synovic, Nicholas and Zhang, Xuecen and Chaudhary, Vipin and Thiruvathukal, George K and Lu, Yung-Hsiang "An automated approach for improving the inference latency and energy efficiency of pretrained CNNs by removing irrelevant pixels with focused convolutions" Asia and South Pacific Design Automation Conference (ASP-DAC) , 2024 https://doi.org/10.1109/ASP-DAC58780.2024.10473884 Citation Details
Thiruvathukal, George K. and Lu, Yung-Hsiang "Efficient Computer Vision for Embedded Systems" Computer , v.55 , 2022 https://doi.org/10.1109/MC.2022.3145677 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page