Withdraw
Loading…
Reinforcement learning for multi-agent and robust control systems
Zhang, Kaiqing
Loading…
Permalink
https://hdl.handle.net/2142/113153
Description
- Title
- Reinforcement learning for multi-agent and robust control systems
- Author(s)
- Zhang, Kaiqing
- Issue Date
- 2021-07-08
- Director of Research (if dissertation) or Advisor (if thesis)
- Basar, Tamer
- Doctoral Committee Chair(s)
- Basar, Tamer
- Committee Member(s)
- Srikant, Rayadurgam
- Dullerud, Geir
- Raginsky, Maxim
- Hu, Bin
- Department of Study
- Electrical & Computer Eng
- Discipline
- Electrical & Computer Engr
- Degree Granting Institution
- University of Illinois at Urbana-Champaign
- Degree Name
- Ph.D.
- Degree Level
- Dissertation
- Keyword(s)
- Reinforcement Learning
- Multi-agent Systems
- Robust Control
- Abstract
- Recent years have witnessed phenomenal accomplishments of reinforcement learning (RL) in many prominent sequential decision-making problems, such as playing the game of Go, playing real-time strategy games, robotic control, and autonomous driving. Motivated by these empirical successes, research toward theoretical understandings of RL algorithms has also re-gained great attention in recent years. In this dissertation, our goal is to contribute to these efforts through new approaches and tools, by developing RL algorithms for multi-agent and robust control systems, which find broad applications in the aforementioned examples and other diverse areas. In this dissertation, we consider specifically two different and fundamental settings that in general fall into the realm of RL for multi-agent and robust control systems: (i) decentralized multi-agent RL (MARL) with networked agents; (ii) H2/H∞-robust control synthesis. We develop new RL algorithms for these settings, supported by theoretical convergence guarantees. In setting i, a team of collaborative MARL agents is connected via a communication network, without the coordination of any central controller. With only neighbor-to-neighbor communications, we introduce decentralized actor-critic algorithms for each agent, and establish their convergence guarantees when linear function approximation is used. Setting ii corresponds to a classical robust control problem, with linear dynamics and robustness concerns in the H∞-norm sense. In contrast to existing solvers, we introduce policy-gradient methods to solve the robust control problem, with global convergence guarantees, despite its nonconvexity. More interestingly, we show that two of these methods enjoy the implicit regularization property: the iterates of the controller automatically preserve a certain level of robustness stability, by following such policy search directions. This robustness-on-the-fly property is crucial for learning in safety-critical robust control systems. We then study the model-free regime, where we develop derivative-free policy gradient methods to solve the finite-horizon version of the problem, with sampled trajectories from the system and sample complexity guarantees. Interestingly, this robust control problem also unifies several other fundamental settings in control theory and game theory, including risk-sensitive linear control, i.e., linear exponential quadratic Gaussian (LEQG) control, and linear quadratic zero-sum dynamic games. The latter can be viewed as a benchmark setting for competitive multi-agent RL. Hence, our results provide policy-search methods for solving these problems unifiedly. Finally, we provide numerical results to demonstrate the computational efficiency of our policy search algorithms, compared to several existing robust control solvers.
- Graduation Semester
- 2021-08
- Type of Resource
- Thesis
- Permalink
- http://hdl.handle.net/2142/113153
- Copyright and License Information
- Copyright 2021 Kaiqing Zhang
Owning Collections
Graduate Dissertations and Theses at Illinois PRIMARY
Graduate Theses and Dissertations at IllinoisDissertations and Theses - Electrical and Computer Engineering
Dissertations and Theses in Electrical and Computer EngineeringManage Files
Loading…
Edit Collection Membership
Loading…
Edit Metadata
Loading…
Edit Properties
Loading…
Embargoes
Loading…