Award Abstract # 2144194
CAREER: Vision Systems for an Evolving World

NSF Org: IIS
Div Of Information & Intelligent Systems
Recipient: GEORGIA TECH RESEARCH CORP
Initial Amendment Date: August 24, 2022
Latest Amendment Date: July 17, 2023
Award Number: 2144194
Award Instrument: Continuing Grant
Program Manager: Jie Yang
jyang@nsf.gov
 (703)292-4768
IIS
 Div Of Information & Intelligent Systems
CSE
 Direct For Computer & Info Scie & Enginr
Start Date: September 1, 2022
End Date: August 31, 2027 (Estimated)
Total Intended Award Amount: $589,854.00
Total Awarded Amount to Date: $589,854.00
Funds Obligated to Date: FY 2022 = $114,634.00
FY 2023 = $475,220.00
History of Investigator:
  • Judy Hoffman (Principal Investigator)
    judy@gatech.edu
Recipient Sponsored Research Office: Georgia Tech Research Corporation
926 DALNEY ST NW
ATLANTA
GA  US  30318-6395
(404)894-4819
Sponsor Congressional District: 05
Primary Place of Performance: Georgia Institute of Technology
225 North Avenue
Atlanta
GA  US  30332-0365
Primary Place of Performance
Congressional District:
05
Unique Entity Identifier (UEI): EMW9FC8J3HN4
Parent UEI: EMW9FC8J3HN4
NSF Program(s): Robust Intelligence
Primary Program Source: 01002223DB NSF RESEARCH & RELATED ACTIVIT
01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 1045, 7495
Program Element Code(s): 749500
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

When transitioning into a dark room most people initially struggle to see, but are quickly able to adjust and see in the new setting. In fact, people continue to see and understand the world around them even as the appearance of the world changes in many different ways. In contrast, our computer vision systems have limited ability to understand the world if it changes. Imagine if the first time one drove at twilight or in the snow one could no longer recognize the road. No doubt, driver?s training would require many hours of sitting in the passenger seat at all times of day, within different weather conditions, across different cities, and so on before allowing a new driver behind the wheel. This project aims to study and build new models, algorithms, and measures of success enabling the next generation of visual recognition systems to be resilient to an evolving visual world. The project will integrate research with education and outreach to K-12 students.

This project advocates for resilient vision systems through a new integrated approach which iterates between generalizing across available visual domains and rapidly adapting given new domain data. Prior approaches optimize for independent criteria, either generalization across multiple domains, or adaptation to a new target domain, which limits advancement towards the larger goal of creating vision systems that can operate in more domains over time. Further, existing solutions are slow to adapt, relying on substantial new observations before updates can be made. The project will work on: 1) Model design and learning approaches for multi-domain generalization that facilitates future adaptation. 2) Transformative visual domain adaptation algorithms that are capable of rapidly adapting to a target domain using limited target observations and without accessing a large auxiliary source of data, reducing compute demands. 3) Algorithms that enable vision systems to expand the set of domains they can successfully operate in over time. Finally, this project will introduce a benchmark and new evaluation protocols to measure the resilience of visual recognition models to changing domains.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Chattopadhyay, P and Goyal, B and Ecsedi, B and Prabhu, V and Hoffman, J "AUGCAL: Improving Sim2Real Adaptation by Uncertainty Calibration on Augmented Synthetic Images" International Conference on Learning Representations , 2024 Citation Details
Kareer, S and Vijaykumar, V and Maheshwari, H and Chattopadhyay, P and Hoffman, J and Prabhu, V "We're Not Using Videos Effectively: An Updated Domain Adaptive Video Segmentation Baseline" Transactions on machine learning research , 2024 Citation Details
Prabhu, V and Yenamandra, S and Chattopadhyay, P and Hoffman, J "LANCE: stress-testing visual models by generating language-guided counterfactual images" International Conference on Neural Information Processing Systems , 2023 Citation Details
Stoica, G and Bolya, D and Bjorner, J and Ramesh, P and Hearn, T and Hoffman, J "ZipIt! Merging Models from Different Tasks without Training" International Conference on Learning Representations , 2024 Citation Details

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page