Stochastics and Statistics Seminar

Views Navigation

Event Views Navigation

Two Approaches Towards Adaptive Optimization

Ashia Wilson, MIT
E18-304

Abstract: This talk will address to recent projects I am excited about. The first describes efficient methodologies for hyper-parameter estimation in optimization algorithms. I will describe two approaches for how to adaptively estimate these parameters that often lead to significant improvement in convergence. The second describes a new method, called Metropolis-Adjusted Preconditioned Langevin Algorithm for sampling from a convex body. Taking an optimization perspective, I focus on the mixing time guarantees of these algorithms — an essential theoretical property for…

Find out more »

Finite-Particle Convergence Rates for Stein Variational Gradient Descent

Krishna Balasubramanian, University of California - Davis
E18-304

Abstract: Stein Variational Gradient Descent (SVGD) is a deterministic, interacting particle-based algorithm for nonparametric variational inference, yet its theoretical properties remain challenging to fully understand. This talk presents two complementary perspectives on SVGD. First, we introduce Gaussian-SVGD, a framework that projects SVGD onto the family of Gaussian distributions using a bilinear kernel. We establish rigorous convergence results for both mean-field dynamics and finite-particle systems, proving linear convergence to equilibrium in strongly log-concave settings. This framework also unifies recent algorithms for…

Find out more »

How should we do linear regression?

Richard Samworth, University of Cambridge
E18-304

Abstract: In the context of linear regression, we construct a data-driven convex loss function with respect to which empirical risk minimisation yields optimal asymptotic variance in the downstream estimation of the regression coefficients. Our semiparametric approach targets the best decreasing approximation of the derivative of the log-density of the noise distribution. At the population level, this fitting process is a nonparametric extension of score matching, corresponding to a log-concave projection of the noise distribution with respect to the Fisher divergence.…

Find out more »


MIT Statistics + Data Science Center
Massachusetts Institute of Technology
77 Massachusetts Avenue
Cambridge, MA 02139-4307
617-253-1764