website/content/paper/2406.01.md

21 lines
No EOL
1.4 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
draft: false
title: "Guiding Hiearchical Reinforcement Learning in Partially Observable Environments with AI Planning"
authors: [
"Brandon Rozek",
"Junkyu Lee",
"Harsha Kokel",
"Michael Katz",
"Shirin Sohrabi"
]
date: 2024-06-02
publish_date: "2024/06/02"
conference: "International Workshop on Bridging the Gap Between AI Planning and Reinforcement Learning (PRL)"
isbn: ""
doi: ""
language: "English"
pdf_url: "https://prl-theworkshop.github.io/prl2024-icaps/papers/12.pdf"
abstract: "Partially observable Markov decision processes challenge reinforcement learning agents since observations provide an limited view of the environment. This often requires an agent to explore collecting observations to form the necessary state information to complete the task. Even assuming knowledge is monotonic, it is difficult to know when to stop exploration. We integrate AI planning within hierarchical reinforcement learning to aide in the exploration of partially observable environments. Given a set of unknown state variables, their potential valuations, along with which abstract operators may discover them, we create an abstract fully-observable non-deterministic planning problem which captures the agents abstract belief state. This decomposes the POMDP into a tree of semi-POMDPs based on sensing outcomes. We evaluate our agents performance on a MiniGrid domain and show how guided exploration may improve agent performance."
---