BEGIN:VCALENDAR
VERSION:2.0
PRODID:-// - ECPv6.15.18//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-ORIGINAL-URL:https://tilos.ai
X-WR-CALDESC:Events for 
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/Los_Angeles
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20210314T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20211107T090000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20220313T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20221106T090000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20230312T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20231105T090000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20220928T100000
DTEND;TZID=America/Los_Angeles:20220928T110000
DTSTAMP:20260406T101018
CREATED:20250904T172904Z
LAST-MODIFIED:20250904T172904Z
UID:7355-1664359200-1664362800@tilos.ai
SUMMARY:TILOS Seminar: On Policy Optimization Methods for Control
DESCRIPTION:Maryam Fazel\, Professor\, University of Washington \nAbstract: Policy Optimization methods enjoy wide practical use in reinforcement learning (RL) for applications ranging from robotic manipulation to game-playing\, partly because they are easy to implement and allow for richly parameterized policies. Yet their theoretical properties\, from optimality to statistical complexity\, are still not fully understood. To help develop a theoretical basis for these methods\, and to bridge the gap between RL and control theoretic approaches\, recent work has studied whether gradient-based policy optimization can succeed in designing feedback control policies. In this talk\, we start by showing the convergence and optimality of these methods for linear dynamical systems with quadratic costs\, where despite nonconvexity\, convergence to the optimal policy occurs under mild assumptions. Next\, we make a connection between convex parameterizations in control theory on one hand\, and the Polyak-Lojasiewicz property of the nonconvex cost function\, on the other. Such a connection between the nonconvex and convex landscapes provides a unified view towards extending the results to more complex control problems.
URL:https://tilos.ai/event/tilos-seminar-on-policy-optimization-methods-for-control/
LOCATION:Virtual
CATEGORIES:TILOS Seminar Series
ATTACH;FMTTYPE=image/jpeg:https://tilos.ai/wp-content/uploads/2023/10/fazel-maryam.jpg
END:VEVENT
END:VCALENDAR