Views Navigation

Event Views Navigation

Calendar of Events

S Sun

M Mon

T Tue

W Wed

T Thu

F Fri

S Sat

0 events,

1 event,

2017 Charles River Lectures on Probability and Related Topics

0 events,

0 events,

0 events,

1 event,

Statistics and Data Science Seminar Youssef Marzouk (MIT)

0 events,

0 events,

0 events,

0 events,

0 events,

0 events,

1 event,

Statistics and Data Science Seminar Galen Reeves (Duke University)

0 events,

0 events,

0 events,

0 events,

0 events,

1 event,

Statistics and Data Science Seminar John Cunningham (Columbia)

1 event,

Statistics and Data Science Seminar Sayan Mukherjee (Duke)

0 events,

0 events,

0 events,

0 events,

0 events,

0 events,

1 event,

0 events,

0 events,

0 events,

0 events,

1 event,

Statistics and Data Science Seminar Pierre Jacob (Harvard)

0 events,

1 event,

Statistics and Data Science Seminar Joan Bruna Estrach (NYU)

0 events,

2017 Charles River Lectures on Probability and Related Topics

The Charles River Lectures on Probability and Related Topics will be hosted by Harvard University on Monday, October 2, 2017 in Cambridge, MA. The lectures are jointly organized by Harvard University, Massachusetts Institute of Technology and Microsoft Research New England for the benefit of the greater Boston area mathematics community. The event features five lectures…

Find out more »

Transport maps for Bayesian computation

Youssef Marzouk (MIT)
E18-304

Abstract: Integration against an intractable probability measure is among the fundamental challenges of Bayesian inference. A useful approach to this problem seeks a deterministic coupling of the measure of interest with a tractable "reference" measure (e.g., a standard Gaussian). This coupling is induced by a transport map, and enables direct simulation from the desired measure simply…

Find out more »

Additivity of Information in Deep Generative Networks: The I-MMSE Transform Method

Galen Reeves (Duke University)
E18-304

Abstract:  Deep generative networks are powerful probabilistic models that consist of multiple stages of linear transformations (described by matrices) and non-linear, possibly random, functions (described generally by information channels). These models have gained great popularity due to their ability to characterize complex probabilistic relationships arising in a wide variety of inference problems. In this talk, we…

Find out more »

Structure in multi-index tensor data: a trivial byproduct of simpler phenomena?

John Cunningham (Columbia)
E18-304

Abstract:  As large tensor-variate data become increasingly common across applied machine learning and statistics, complex analysis methods for these data similarly increase in prevalence.  Such a trend offers the opportunity to understand subtler and more meaningful features of the data that, ostensibly, could not be studied with simpler datasets or simpler methodologies.  While promising, these advances…

Find out more »

Inference in dynamical systems and the geometry of learning group actions

Sayan Mukherjee (Duke)
E18-304

Abstract: We examine consistency of the Gibbs posterior for dynamical systems using a classical idea in dynamical systems called the thermodynamic formalism in tracking dynamical systems. We state a variation formulation under which there is a unique posterior distribution of parameters as well as hidden states using using classic ideas from dynamical systems such as pressure and joinings.…

Find out more »

On Learning Theory and Neural Networks

Amit Daniely (Google)
E18-304

Abstract:  Can learning theory, as we know it today, form a theoretical basis for neural networks. I will try to discuss this question in light of two new results -- one positive and one negative. Based on joint work with Roy Frostig, Vineet Gupta and Yoram Singer, and with Vitaly Feldman Biography: Amit Daniely is an Assistant…

Find out more »

Unbiased Markov chain Monte Carlo with couplings

Pierre Jacob (Harvard)
E18-304

Abstract: Markov chain Monte Carlo methods provide consistent approximations of integrals as the number of iterations goes to infinity. However, these estimators are generally biased after any fixed number of iterations, which complicates both parallel computation. In this talk I will explain how to remove this burn-in  bias by using couplings of Markov chains and…

Find out more »

Statistics, Computation and Learning with Graph Neural Networks

Joan Bruna Estrach (NYU)
E18-304

Abstract:  Deep Learning, thanks mostly to Convolutional architectures, has recently transformed computer vision and speech recognition. Their ability to encode geometric stability priors, while offering enough expressive power, is at the core of their success. In such settings, geometric stability is expressed in terms of local deformations, and it is enforced thanks to localized convolutional operators…

Find out more »


MIT Statistics + Data Science Center
Massachusetts Institute of Technology
77 Massachusetts Avenue
Cambridge, MA 02139-4307
617-253-1764