website/content/paper/2406.01.md

1.4 KiB
Raw Permalink Blame History

draft title authors date publish_date conference isbn doi language pdf_url abstract
false Guiding Hiearchical Reinforcement Learning in Partially Observable Environments with AI Planning
Brandon Rozek
Junkyu Lee
Harsha Kokel
Michael Katz
Shirin Sohrabi
2024-06-02 2024/06/02 International Workshop on Bridging the Gap Between AI Planning and Reinforcement Learning (PRL) English https://prl-theworkshop.github.io/prl2024-icaps/papers/12.pdf Partially observable Markov decision processes challenge reinforcement learning agents since observations provide an limited view of the environment. This often requires an agent to explore collecting observations to form the necessary state information to complete the task. Even assuming knowledge is monotonic, it is difficult to know when to stop exploration. We integrate AI planning within hierarchical reinforcement learning to aide in the exploration of partially observable environments. Given a set of unknown state variables, their potential valuations, along with which abstract operators may discover them, we create an abstract fully-observable non-deterministic planning problem which captures the agents abstract belief state. This decomposes the POMDP into a tree of semi-POMDPs based on sensing outcomes. We evaluate our agents performance on a MiniGrid domain and show how guided exploration may improve agent performance.