IMA Data Science Seminar: Consistent Sparse Deep Learning: Theory and Computation

Data science seminars hosted by the The Institute for Mathematics and Its Applications (IMA) take place on Tuesdays from 1:25 p.m. - 2:25 p.m.

This week, Faming Liang (Purdue University), will be giving a lecture titled "Consistent Sparse Deep Learning: Theory and Computation".

Registration is required to access the Zoom webinar.

Abstract

Deep learning has been the engine powering many successes of data science. However, the deep neural network (DNN), as the basic model of deep learning, is often excessively over-parameterized, causing many difficulties in training, prediction and interpretation. We propose a frequentist-like method for learning sparse DNNs and justify its consistency under the Bayesian framework: the proposed method could learn a sparse DNN with at most O(n/log(n))O(n/log⁡(n)) connections and nice theoretical guarantees such as posterior consistency, variable selection consistency and asymptotically optimal generalization bounds. In particular, we establish posterior consistency for the sparse DNN with a mixture Gaussian prior, show that the structure of the sparse DNN can be consistently determined using a Laplace approximation-based marginal posterior inclusion probability approach, and use Bayesian evidence to elicit sparse DNNs learned by an optimization method such as stochastic gradient descent in multiple runs with different initializations. The proposed method  is computationally more efficient than standard Bayesian methods for large-scale sparse DNNs.  The numerical results indicate that the proposed method can perform very well for large-scale network compression and high-dimensional nonlinear variable selection, both advancing interpretable machine learning.  The talk is based on a joint work with Yan Sun and Qifan Song.

Biography

Faming Liang is Professor of Statistics at Purdue University. Before joining Purdue, he held a faculty position at University of Florida and Texas A&M University. Faming has wide research interests, including machine learning, Monte Carlo methods, bioinformatics, high-dimensional statistics, and big data. He is ASA fellow and IMS fellow, and has published over 120 journal papers.

View the full list of IMA data science seminars.

Start date
Tuesday, March 16, 2021, 1:25 p.m.
End date
Tuesday, March 16, 2021, 2:25 p.m.
Location

Online

Share