Withdraw
Loading…
Reinforcement learning under general function approximation and novel interaction settings
Chen, Jinglin
Loading…
Permalink
https://hdl.handle.net/2142/121992
Description
- Title
- Reinforcement learning under general function approximation and novel interaction settings
- Author(s)
- Chen, Jinglin
- Issue Date
- 2023-11-17
- Director of Research (if dissertation) or Advisor (if thesis)
- Jiang, Nan
- Doctoral Committee Chair(s)
- Jiang, Nan
- Committee Member(s)
- Banerjee, Arindam
- Raginsky, Maxim
- Krishnamurthy, Akshay
- Department of Study
- Computer Science
- Discipline
- Computer Science
- Degree Granting Institution
- University of Illinois at Urbana-Champaign
- Degree Name
- Ph.D.
- Degree Level
- Dissertation
- Keyword(s)
- reinforcement learning
- sample complexity analysis
- machine learning
- artificial intelligence
- Abstract
- Reinforcement Learning (RL) is an area of machine learning where an intelligent agent solves sequential decision-making problems based on experience. Recent advances in the applications of RL have been witnessed not only in the field of games, control systems, and healthcare, but also in language models and addressing open problems in mathematics. Despite the recent progress, RL algorithms are deemed to be data-hungry and unstable in training. Therefore, a flurry of research has focused on building fundamental guarantees for provably efficient RL algorithms. This line of research aims to provide a theoretical backbone and deepen our understanding of RL methods, which is crucial for the success of applications of RL in real-world scenarios. The main theme of this thesis is to propose and analyze sample efficient algorithms under different structural complexity measures and various interaction settings. In the thesis, we examine numerous setups: from pure offline RL to online RL with exploration, from Markov Decision Processes (MDPs) with low-rank structures to MDPs under more general function approximators, and from reward-aware to reward-free learning. More concretely, this thesis shows theoretical results in the sequel. The first part of this thesis revisits data coverage and function class representability assumptions in offline RL and makes an important conjecture that initiates future research. The second part provides a positive result in offline RL under non-exploratory data and weak function approximation. The third part focuses on the reward-free learning framework under general non-linear function approximation, bridges the sizeable gap in our understanding of reward-aware and reward-free settings, and shows an exponential separation between low-rank and linear completeness settings. Lastly, in the fourth part, we consider representation learning and reward-free learning in low-rank MDPs from the angle of the density candidate feature class. When establishing statistical complexity guarantees for RL algorithms under structural assumptions, we leverage powerful tools from machine learning theory literature. Throughout the thesis, we tackle complex scenarios in RL, where the agent faces challenges such as the absence of online data collection or a lack of reward information as guidance while exploring the environment. Overall, this thesis proposes sample efficient algorithms and presents theoretical results under general function approximation and novel interaction protocols.
- Graduation Semester
- 2023-12
- Type of Resource
- Thesis
- Copyright and License Information
- Copyright 2023 Jinglin Chen
Owning Collections
Graduate Dissertations and Theses at Illinois PRIMARY
Graduate Theses and Dissertations at IllinoisManage Files
Loading…
Edit Collection Membership
Loading…
Edit Metadata
Loading…
Edit Properties
Loading…
Embargoes
Loading…