Towards a Theory for Sample-efficient Reinforcement Learning with Rich Observations

1 860
28.2
Следующее
18.09.18 – 3 14122:24
Fireside Chat with Cynthia Dwork
Популярные
Опубликовано 18 сентября 2018, 15:49
How can we tractably solve sequential decision making problems where the learning agent receives rich observations? We begin with a new model called Contextual Decision Processes (CDPs) for studying such problems, and show that it encompasses several prior setups to study RL such as MDPs and POMDPs. Several special cases of CDPs are, however, known to be provably intractable in their sample complexities. To overcome this challenge, we further propose a structural property of such processes, called the Bellman Rank. We find that the Bellman Rank of a CDP (and an associated class of functions) provides an intuitive measure of the hardness of a problem in terms of sample complexity and is small in several practical settings. In particular, we propose an algorithm, whose sample complexity scales with the Bellman Rank of the process, and is completely independent of the size of the observation space of the agent. We also show that our techniques are robust to our modeling assumptions, and make connections to several known results as well as highlight novel consequences of our results.

Finally, we also discuss the computational difficulties which arise in operationalizing the resulting methods and directions for further progress.

See more at microsoft.com/en-us/research/v...
автотехномузыкадетское