BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//MIT Statistics and Data Science Center - ECPv5.12.2//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:MIT Statistics and Data Science Center
X-ORIGINAL-URL:https://stat.mit.edu
X-WR-CALDESC:Events for MIT Statistics and Data Science Center
BEGIN:VTIMEZONE
TZID:America/New_York
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20190310T070000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20191103T060000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/New_York:20190911T160000
DTEND;TZID=America/New_York:20190911T170000
DTSTAMP:20220128T103112
CREATED:20190910T180928Z
LAST-MODIFIED:20190910T190537Z
UID:3453-1568217600-1568221200@stat.mit.edu
SUMMARY:Automated Data Summarization for Scalability in Bayesian Inference
DESCRIPTION:IDS.190 – Topics in Bayesian Modeling and Computation \nAbstract: \nMany algorithms take prohibitively long to run on modern\, large datasets. But even in complex data sets\, many data points may be at least partially redundant for some task of interest. So one might instead construct and use a weighted subset of the data (called a “coreset”) that is much smaller than the original dataset. Typically running algorithms on a much smaller data set will take much less computing time\, but it remains to understand whether the output can be widely useful. (1) In particular\, can running an analysis on a smaller coreset yield answers close to those from running on the full data set? (2) And can useful coresets be constructed automatically for new analyses\, with minimal extra work from the user? We answer in the affirmative for a wide variety of problems in Bayesian inference. We demonstrate how to construct “Bayesian coresets” as an automatic\, practical pre-processing step. We prove that our method provides geometric decay in relevant approximation error as a function of coreset size. Empirical analysis shows that our method reduces approximation error by orders of magnitude relative to uniform random subsampling of data. Though we focus on Bayesian methods here\, we also show that our construction can be applied in other domains. \nBiography: \nTamara Broderick is an Associate Professor in EECS at MIT. \n**Meetings are open to any interested researcher. \n**Taking IDS.190 satisfies the seminar requirement for students in MIT’s Interdisciplinary Doctoral Program in Statistics (IDPS)\, but formal registration is open to any graduate student who can register for MIT classes. For more information and an up-to-date schedule\, please see https://stellar.mit.edu/S/course/IDS/fa19/IDS.190/\n \n \n \n
URL:https://stat.mit.edu/calendar/automated-data-summarization-for-scalability-in-bayesian-inference/
LOCATION:E18-304\, United States
CATEGORIES:IDS.190 Topics in Bayesian Modeling and Computation
GEO:42.3620185;-71.0878444
END:VEVENT
END:VCALENDAR