Papers
arxiv:2312.16730
Foundations of Reinforcement Learning and Interactive Decision Making
Published on Dec 27, 2023
Authors:
Abstract
These lecture notes give a statistical perspective on the foundations of reinforcement learning and interactive decision making. We present a unifying framework for addressing the exploration-exploitation dilemma using frequentist and Bayesian approaches, with connections and parallels between supervised learning/estimation and decision making as an overarching theme. Special attention is paid to function approximation and flexible model classes such as neural networks. Topics covered include multi-armed and contextual bandits, structured bandits, and reinforcement learning with high-dimensional feedback.
Models citing this paper 0
No model linking this paper
Cite arxiv.org/abs/2312.16730 in a model README.md to link it from this page.
Datasets citing this paper 0
No dataset linking this paper
Cite arxiv.org/abs/2312.16730 in a dataset README.md to link it from this page.
Spaces citing this paper 0
No Space linking this paper
Cite arxiv.org/abs/2312.16730 in a Space README.md to link it from this page.