Award Abstract # 1546482
BIGDATA: Collaborative Research: F: Stochastic Approximation for Subspace and Multiview Representation Learning

NSF Org: IIS
Div Of Information & Intelligent Systems
Awardee: JOHNS HOPKINS UNIVERSITY, THE
Initial Amendment Date: September 3, 2015
Latest Amendment Date: September 3, 2015
Award Number: 1546482
Award Instrument: Standard Grant
Program Manager: Sylvia Spengler
sspengle@nsf.gov
 (703)292-7347
IIS
 Div Of Information & Intelligent Systems
CSE
 Direct For Computer & Info Scie & Enginr
Start Date: September 1, 2015
End Date: August 31, 2021 (Estimated)
Total Intended Award Amount: $704,672.00
Total Awarded Amount to Date: $704,672.00
Funds Obligated to Date: FY 2015 = $704,672.00
History of Investigator:
  • Raman  Arora (Principal Investigator)
    arora@cs.jhu.edu
Awardee Sponsored Research Office: Johns Hopkins University
1101 E 33rd St
Baltimore
MD  US  21218-2686
(443)997-1898
Sponsor Congressional District: 07
Primary Place of Performance: Johns Hopkins University
3400 N CHARLES ST
Baltimore
MD  US  21218-2608
Primary Place of Performance
Congressional District:
07
DUNS ID: 001910777
Parent DUNS ID: 001910777
NSF Program(s): Big Data Science &Engineering
Primary Program Source: 040100 NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7433, 8083
Program Element Code(s): 8083
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.070

ABSTRACT

Unsupervised learning of useful features, or representations, is one of the most basic challenges of machine learning. Unsupervised representation learning techniques capitalize on unlabeled data which is often cheap and abundant and sometimes virtually unlimited. The goal of these ubiquitous techniques is to learn a representation that reveals intrinsic low-dimensional structure in data, disentangles underlying factors of variation by incorporating universal AI priors such as smoothness and sparsity, and is useful across multiple tasks and domains.

This project aims to develop new theory and methods for representation learning that can easily scale to large datasets. In particular, this project is concerned with methods for large-scale unsupervised feature learning, including Principal Component Analysis (PCA) and Partial Least Squares (PLS). To capitalize on massive amounts of unlabeled data, this project will develop appropriate computational approaches and study them in the "data-laden" regime. Therefore, instead of viewing representation learning as dimensionality reduction techniques and focusing on an empirical objective on finite data, these methods are studied with the goal of optimizing a population objective based on sample. This view suggests using Stochastic Approximation approaches, such as Stochastic Gradient Descent (SGD) and Stochastic Mirror Descent, that are incremental in nature and process each new sample with a computationally cheap update. Furthermore, this view enables a rigorous analysis of benefits of stochastic approximation algorithms over traditional finite-data methods. The project aims to develop stochastic approximation approaches to PCA and PLS and related problems and extensions, including deep, and sparse variants, and analyze these problems in the data-laden regime.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 38)
Adrian Benton, Raman Arora, and Mark Dredze "Learning Multiview Embeddings of Twitter Users" Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL) , 2016
Mo Yu, Mark Dredze, Raman Arora, and Matthew Gormley "Embedding Lexical Features via Low-rank Tensors" Proceedings of the 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT) , 2016
Raman Arora, Poorya Mianjy, and Teodor Marinov "Stochastic Optimization for Multiview Representation Learning using Partial Least Squares" Proceedings of the 33rd International Conference on Machine Learning (ICML) , 2016
Weiran Wang, Raman Arora, Karen Livescu, and Nathan Srebro "Stochastic Optimization for Deep CCA via Nonlinear Orthogonal Iterations" Proceedings of the 53rd Annual Allerton Conference on Communication, Control, and Computing (ALLERTON) , 2015
Xingguo Li, Tuo Zhao, Raman Arora, Han Liu, and Jarvis Haupt "Stochastic Variance Reduced Optimization for Nonconvex Sparse Learnin" Proceedings of The 33rd International Conference on Machine Learning (ICML) , 2016
Xingguo Li, Tuo Zhao, Raman Arora, Han Liu, and Mingyi Hong "An Improved Convergence Analysis of Cyclic Block Coordinate Descent-type Methods for Strongly Convex Minimization" Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS) , 2016
Peter Schulam and Raman Arora "Disease Trajectory Maps" Proceedings of the The Thirtieth Annual Conference on Neural Information Processing Systems (NIPS) , 2016
Leonardo Badino, Luca Franceschi, Raman Arora, Michele Donini, Massimiliano Pontil "A Speaker Adaptive DNN Training Approach for Speaker-independent Acoustic Inversion" Conference of the International Speech Communication Association (INTERSPEECH) , 2017
Blake Woodworth, Nathan Srebro "Tight Complexity Bounds for Optimizing Composite Objectives" Neural Information Processing Systems (NIPS) , 2017
Dan Garber, Ohad Shamir, Nathan Srebro "Communication-efficient Algorithms for Distributed Stochastic Principal Component Analysis" Proceedings of the 34th International Conference on Machine Learning (ICML) , 2017
Jialei Wang, Jason D Lee, Mehrdad Mahdavi, Mladen Kolar, Nathan Srebro "Sketching Meets Random Projection in the Dual: A Provable Recovery Algorithm for Big and High -dimensional Data" Artificial Inteligence and Statistics (AISTATS) , 2017
(Showing: 1 - 10 of 38)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page