Skip to yearly menu bar Skip to main content


Poster

State-free Reinforcement Learning

Mingyu Chen · Aldo Pacchiano · Xuezhou Zhang

[ ]
Wed 11 Dec 11 a.m. PST — 2 p.m. PST

Abstract: In this work, we study the \textit{state-free RL} problem, where the algorithm does not have the states information before interacting with the environment. Specifically, denote the reachable state set by $\mathcal{S}^\Pi := \{ s|\max_{\pi\in \Pi}q^{P, \pi}(s)>0 \}$, we design an algorithm which requires no information on the state space $S$ while having a regret that is completely independent of $\mathcal{S}$ and only depend on $\mathcal{S}^\Pi$. We view this as a concrete first step towards \textit{parameter-free RL}, with the goal of designing RL algorithms that require no hyper-parameter tuning.

Live content is unavailable. Log in and register to view live content