website/_posts/research/2020-06-01-ood-classification.md
2024-11-10 12:16:11 +01:00

1.3 KiB

layout, title, categories, excerpt, header
layout title categories excerpt header
single Policy Entropy for OOD Classification research PEOC for reliably detecting unencountered states in deep RL
teaser
assets/figures/6_ood_pipeline.jpg

PEOC Pipeline{:style="display:block; margin-left:auto; margin-right:auto"}

One critical prerequisite for the deployment of reinforcement learning systems in the real world is the ability to reliably detect situations on which the agent was not trained. Such situations could lead to potential safety risks when wrong predictions lead to the execution of harmful actions. In this work, we propose PEOC, a new policy entropy based out-of-distribution classifier that reliably detects unencountered states in deep reinforcement learning. It is based on using the entropy of an agent's policy as the classification score of a one-class classifier. We evaluate our approach using a procedural environment generator. Results show that PEOC is highly competitive against state-of-the-art one-class classification algorithms on the evaluated environments. Furthermore, we present a structured process for benchmarking out-of-distribution classification in reinforcement learning. {% cite sedlmeier2020peoc %}

PEOC Performance{:style="display:block; margin-left:auto; margin-right:auto"}