Loading…
Loading grant details…
| Funder | National Science Foundation (US) |
|---|---|
| Recipient Organization | Florida State University |
| Country | United States |
| Start Date | Jun 01, 2021 |
| End Date | May 31, 2025 |
| Duration | 1,460 days |
| Number of Grantees | 1 |
| Roles | Principal Investigator |
| Data Source | National Science Foundation (US) |
| Grant ID | 2053454 |
The research in this project will focus on a particular class of algorithms for machine learning and data science. In particular, the investigators consider the large class of Markov Chain Monte Carlo (MCMC) methods which arise in several contexts in machine learning and data science. The project will develop new algorithms within the subclass called Langevin MCMC methods.
These new algorithms will be scalable to high dimensions and large datasets and will be faster than traditional ones. The features of scalability and fast convergence are important for use in Bayesian statistical inference as well as in non-convex stochastic optimization methods for machine learning. The algorithms will allow efficient training and calibration of predictive machine learning models from large-scale data and have a direct impact on a broad range of data-driven application areas from information technology to computer vision. Graduate students will be trained and involved in research.
In this project, the PIs investigate a new class of algorithms within the class of Langevin MCMC methods. These algorithms can be applied in three contexts of machine learning and data science. First, they can be used for Bayesian (learning) inference problems with high-dimensional models, where the objective is to sample from a posterior distribution given a prior distribution on the parameter space and the likelihood of the observed data.
Second, they can be used for solving stochastic non-convex optimization problems including the challenging problems arising in deep learning. Third, they arise in modeling and approximating workhorse algorithms in data science such as stochastic gradient descent methods. By leveraging out the connections between stochastic gradient algorithms and MCMC algorithms, the proposed approach results in a new class of stochastic gradient algorithms called Hamiltonian Accelerated Stochastic Gradient that can outperform existing methods in deep learning practice.
A first goal of the project is to study theoretical convergence properties of the proposed algorithms further to fill out the current gap between theory and practice, as well as to develop new scalable algorithms that can extend the existing framework. A second goal is to investigate existing Langevin algorithms further to provide non-asymptotic rigorous performance guarantees relevant to machine learning and data science practice.
This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
Florida State University
Complete our application form to express your interest and we'll guide you through the process.
Apply for This Grant