BEGIN:VCALENDAR
VERSION:2.0
PRODID:-// - ECPv6.15.18//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-ORIGINAL-URL:https://tilos.ai
X-WR-CALDESC:Events for 
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/Los_Angeles
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20210314T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20211107T090000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20220313T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20221106T090000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20230312T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20231105T090000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20220921T100000
DTEND;TZID=America/Los_Angeles:20220921T110000
DTSTAMP:20260403T080132
CREATED:20250904T172224Z
LAST-MODIFIED:20250904T172224Z
UID:7356-1663754400-1663758000@tilos.ai
SUMMARY:TILOS Seminar: Non-convex Optimization for Linear Quadratic Gaussian (LQG) Control
DESCRIPTION:Yang Zheng\, Assistant Professor\, UC San Diego \nAbstract: Recent studies have started to apply machine learning techniques to the control of unknown dynamical systems. They have achieved impressive empirical results. However\, the convergence behavior\, statistical properties\, and robustness performance of these approaches are often poorly understood due to the non-convex nature of the underlying control problems. In this talk\, we revisit the Linear Quadratic Gaussian (LQG) control and present recent progress towards its landscape analysis from a non-convex optimization perspective. We view the LQG cost as a function of the controller parameters and study its analytical and geometrical properties. Due to the inherent symmetry induced by similarity transformations\, the LQG landscape is very rich yet complicated. We show that 1) the set of stabilizing controllers has at most two path-connected components\, and 2) despite the nonconvexity\, all minimal stationary points (controllable and observable controllers) are globally optimal. Based on the special non-convex optimization landscape\, we further introduce a novel perturbed policy gradient (PGD) method to escape a large class of suboptimal stationary points (including high-order saddles). These results shed some light on the performance analysis of direct policy gradient methods for solving the LQG problem. The talk is based on our recent papers: https://arxiv.org/abs/2102.04393 and https://arxiv.org/abs/2204.00912.
URL:https://tilos.ai/event/tilos-seminar-non-convex-optimization-for-linear-quadratic-gaussian-lqg-control/
LOCATION:Virtual
CATEGORIES:TILOS Seminar Series
ATTACH;FMTTYPE=image/jpeg:https://tilos.ai/wp-content/uploads/2023/09/zheng-yang-e1757006534476.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20220928T100000
DTEND;TZID=America/Los_Angeles:20220928T110000
DTSTAMP:20260403T080132
CREATED:20250904T172904Z
LAST-MODIFIED:20250904T172904Z
UID:7355-1664359200-1664362800@tilos.ai
SUMMARY:TILOS Seminar: On Policy Optimization Methods for Control
DESCRIPTION:Maryam Fazel\, Professor\, University of Washington \nAbstract: Policy Optimization methods enjoy wide practical use in reinforcement learning (RL) for applications ranging from robotic manipulation to game-playing\, partly because they are easy to implement and allow for richly parameterized policies. Yet their theoretical properties\, from optimality to statistical complexity\, are still not fully understood. To help develop a theoretical basis for these methods\, and to bridge the gap between RL and control theoretic approaches\, recent work has studied whether gradient-based policy optimization can succeed in designing feedback control policies. In this talk\, we start by showing the convergence and optimality of these methods for linear dynamical systems with quadratic costs\, where despite nonconvexity\, convergence to the optimal policy occurs under mild assumptions. Next\, we make a connection between convex parameterizations in control theory on one hand\, and the Polyak-Lojasiewicz property of the nonconvex cost function\, on the other. Such a connection between the nonconvex and convex landscapes provides a unified view towards extending the results to more complex control problems.
URL:https://tilos.ai/event/tilos-seminar-on-policy-optimization-methods-for-control/
LOCATION:Virtual
CATEGORIES:TILOS Seminar Series
ATTACH;FMTTYPE=image/jpeg:https://tilos.ai/wp-content/uploads/2023/10/fazel-maryam.jpg
END:VEVENT
END:VCALENDAR