Control meets learning: scalable and reliable AI for autonomy

9:30am - 10:30am
Zoom Seminar:

Artificial Intelligence (AI), particularly Reinforcement Learning (RL), has achieved great success in domains such as gameplay. However, RL has major scalability and reliability issues which makes it challenging for RL to make an impact in safety-critical and large-scale systems such as power grids, transportation, smart city. In this talk, we show that incorporating model-structure and model-based control can help address the scalability and reliability issues of RL. In the first part of the talk, we consider a networked multi-agent setting and we propose a Scalable Actor Critic framework that provably addresses the scalability issue of multi-agent RL. The key is to exploit a form of local interaction structure widely present in networked systems. In the second part, we consider a nonlinear control setting where the dynamics admit an approximate linear model, which is true for many systems such as the power grid. We show that exploiting this approximate linear model can greatly improve the reliability of an important class of RL algorithms.

讲者/ 表演者:
Dr. Guannan Qu
CMI and Resnick postdoctoral scholar in the Department of Computing and Mathematical Sciences, California Institute of Technology

Guannan Qu received his B.S. degree in Electrical Engineering from Tsinghua University in Beijing, China in 2014, and his Ph.D. in Applied Mathematics from Harvard University in Cambridge, MA in 2019. Since 2019 he has been a CMI and Resnick postdoctoral scholar in the Department of Computing and Mathematical Sciences at California Institute of Technology. He is the recipient of Simoudis Discovery Award, PIMCO Fellowship, and AI4Science Fellowship at Caltech, and Best Student Paper Reward from IEEE SmartGridComm. His research interest lies in control, optimization, and machine/reinforcement learning with applications to power systems, multi-agent systems, smart city, etc.

语言
英文
适合对象
教职员
研究生
主办单位
电子及计算器工程学系
新增活动
请各校内团体将活动发布至大学活动日历。