BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//UM//UM*Events//EN
CALSCALE:GREGORIAN
BEGIN:VTIMEZONE
TZID:America/Detroit
TZURL:http://tzurl.org/zoneinfo/America/Detroit
X-LIC-LOCATION:America/Detroit
BEGIN:DAYLIGHT
TZOFFSETFROM:-0500
TZOFFSETTO:-0400
TZNAME:EDT
DTSTART:20070311T020000
RRULE:FREQ=YEARLY;BYMONTH=3;BYDAY=2SU
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0400
TZOFFSETTO:-0500
TZNAME:EST
DTSTART:20071104T020000
RRULE:FREQ=YEARLY;BYMONTH=11;BYDAY=1SU
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTAMP:20250416T130417
DTSTART;TZID=America/Detroit:20250418T103000
DTEND;TZID=America/Detroit:20250418T233000
SUMMARY:Workshop / Seminar:Bridging the theory-practice gap in machine learning: new results in sampling and optimization
DESCRIPTION:Meteoric progress in machine learning over the last decade has outpaced our foundational understanding\, limiting our ability to harness the technology effectively in applications that require performance guarantees\, and inviting the development of theory to enable such applications. At the heart of this progress is a highly productive connection to gradient-based optimization\, the efficacy of which we are so far unable to fully explain. Motivated by this issue\, in the first part of the talk\, I will briefly describe an interpretable and computationally efficient adaptive step-size method for gradient-based optimization that relies on ideas from the numerical analysis of ordinary differential equations. I will show how this method connects studies of popular optimizers for machine learning in continuous time—where they are often more amenable to analysis—with their practical discrete-time implementations.\n\nSurprising recent developments in machine learning include the ability to generate---or sample---perceptual data such as natural images and language. Markov Chain Monte Carlo (MCMC) algorithms have long provided a generic recipe for sampling from probability distributions of interest. The Gibbs sampler is a specific limit of an MCMC algorithm and is the natural choice for sampling from a simple model of image patches. In the bulk of this talk\, I will focus on a new mixing time bound for Gibbs sampling from well-conditioned log-concave distributions. I will outline the proof of the bound and place it within the context of ongoing efforts in the broader community to understand the efficacy of diffusion-based image generation methods. Time permitting\, I will discuss potential applications of these efforts to problems in cosmology and biophysics.
UID:135076-21876044@events.umich.edu
URL:https://events.umich.edu/event/135076
CLASS:PUBLIC
STATUS:CONFIRMED
CATEGORIES:Mathematics
LOCATION:East Hall - 3866
CONTACT:
END:VEVENT
END:VCALENDAR