Award Abstract # 1943251
CAREER: Understanding the Inductive Biases in Modern Machine Learning
| NSF Org: |
IIS
Div Of Information & Intelligent Systems
|
| Awardee: |
JOHNS HOPKINS UNIVERSITY, THE
|
| Initial Amendment Date: |
February 11, 2020 |
| Latest Amendment Date: |
February 18, 2021 |
| Award Number: |
1943251 |
| Award Instrument: |
Continuing Grant |
| Program Manager: |
Rebecca Hwa
rhwa@nsf.gov
(703)292-7148
IIS
Div Of Information & Intelligent Systems
CSE
Direct For Computer & Info Scie & Enginr
|
| Start Date: |
February 15, 2020 |
| End Date: |
January 31, 2025 (Estimated) |
| Total Intended Award Amount: |
$500,000.00 |
| Total Awarded Amount to Date: |
$192,804.00 |
| Funds Obligated to Date: |
FY 2020 = $104,712.00
FY 2021 = $88,092.00
|
| History of Investigator: |
-
Raman
Arora
(Principal Investigator)
arora@cs.jhu.edu
|
| Awardee Sponsored Research Office: |
Johns Hopkins University
1101 E 33rd St
Baltimore
MD
US
21218-2686
(443)997-1898
|
| Sponsor Congressional District: |
07
|
| Primary Place of Performance: |
Johns Hopkins University
3400 North Charles Street
Baltimore
MD
US
21218-2608
|
Primary Place of Performance Congressional District: |
07
|
| DUNS ID: |
001910777
|
| Parent DUNS ID: |
001910777
|
| NSF Program(s): |
Robust Intelligence
|
| Primary Program Source: |
040100 NSF RESEARCH & RELATED ACTIVIT
040100 NSF RESEARCH & RELATED ACTIVIT
|
| Program Reference Code(s): |
1045,
7495
|
| Program Element Code(s): |
7495
|
| Award Agency Code: |
4900
|
| Fund Agency Code: |
4900
|
| Assistance Listing Number(s): |
47.070
|
ABSTRACT

Recent advances in modern machine learning (deep learning in particular) are ushering in the era of artificial intelligence, which has the potential to revolutionize every aspect of our daily lives. However, much like the early days of the steam engine, a satisfactory understanding of deep learning has so far been elusive. We currently lack a formal theory of deep learning, one that could explain why we can train overly complex models with seemingly not enough training data and still find solutions that generalize to previously unseen data, or why models trained for one task also perform well on another related task, or why trained models are so vulnerable to slight, nearly imperceptible, corruptions of data. This project aims to address this need by developing an explanatory and prescriptive theory of deep learning that is tightly integrated with and motivated by the practice. Rather than view learning as simply a black-box optimization problem, the approach investigates the inner workings by shedding light on algorithmic heuristics that potentially play an equally important role in endowing the trained models with excellent generalization properties. Given the broad applicability of deep learning and the complementary nature of theoretical analyses and empirical studies in the proposed research, the project is particularly suited for integrating research into education and outreach. The proposed educational activities include curriculum development, summer internships, hackathons, and instructor's outreach through local Baltimore programs.
The project investigates the role of explicit algorithmic regularization in the form of early stopping, batch normalization, and dropout, as well as the choice of optimization algorithms and network architecture in providing an adequate inductive bias that helps with generalization. A second overarching goal of the project is to understand, more broadly, the generalization phenomenon in deep learning. It seeks to understand why systems that memorize the training data can still generalize well, how the neural network architecture enables transfer learning, and how to design robust algorithms that will guarantee that deep learning solutions generalize despite adversarial corruption to data.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:
When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
Sulam, Jeremias and Muthukumar, Ramchandran and Arora, Raman
"Adversarial Robustness of Supervised Sparse Coding"
Advances in neural information processing systems
, 2020
https://doi.org/
Citation Details
Mianjy, Poorya and Arora, Raman
"On Convergence and Generalization of Dropout Training"
Advances in neural information processing systems
, 2020
https://doi.org/
Citation Details
Rothchild, Daniel and Panda, Ashwinee and Ullah, Enayat and Ivkin, Nikita and Stoica, Ion and Braverman, Vladimir and Gonzalez, Joseph and Arora, Raman
"FetchSGD: Communication-Efficient Federated Learning with Sketching"
Proceedings of Machine Learning Research
, 2020
https://doi.org/
Citation Details
Rothchild, Daniel and Panda, Ashwinee and Ullah, Enayat and Ivkin, Nikita and Stoica, Ion and Braverman, Vladimir and Gonzalez, Joseph and Arora, Raman
"FetchSGD: Communication-Efficient Federated Learning with Sketching."
Proceedings of Machine Learning Research
, 2020
https://doi.org/
Citation Details
Wang, Yunjuan and Mianjy, Poorya and Arora, Raman
"Robust Learning for Data Poisoning Attacks"
Proceedings of Machine Learning Research
, v.139
, 2021
Citation Details
Arora, Raman and Bartlett, Peter and Mianjy, Poorya and Srebro, Nathan
"Dropout: Explicit Forms and Capacity Control"
Proceedings of Machine Learning Research
, v.139
, 2021
Citation Details
Rothchild, Daniel and Panda, Ashwinee and Ullah, Enayat and Ivkin, Nikita and Stoica, Ion and Braverman, Vladimir and Gonzalez, Joseph and Arora, Raman
"FetchSGD: Communication-Efficient Federated Learning with Sketching"
International Conference on Machine Learning
, 2020
https://doi.org/
Citation Details
Please report errors in award information by writing to: awardsearch@nsf.gov.