Quantifying Differences in Reward Functions

June 24, 2020 ยท Entered Twilight ยท ๐Ÿ› International Conference on Learning Representations

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"Last commit was 5.0 years ago (โ‰ฅ5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .circleci, .codespell.skip, .dockerignore, .gitignore, CONTRIBUTING.md, Dockerfile, LICENSE, README.md, codecov.yml, examples, pylintrc, pyproject.toml, requirements-dev.txt, requirements.txt, runners, scripts, setup.cfg, setup.py, src, tests

Authors Adam Gleave, Michael Dennis, Shane Legg, Stuart Russell, Jan Leike arXiv ID 2006.13900 Category cs.LG: Machine Learning Cross-listed cs.AI, stat.ML Citations 74 Venue International Conference on Learning Representations Repository https://github.com/HumanCompatibleAI/evaluating-rewards โญ 67 Last Checked 1 month ago
Abstract
For many tasks, the reward function is inaccessible to introspection or too complex to be specified procedurally, and must instead be learned from user data. Prior work has evaluated learned reward functions by evaluating policies optimized for the learned reward. However, this method cannot distinguish between the learned reward function failing to reflect user preferences and the policy optimization process failing to optimize the learned reward. Moreover, this method can only tell us about behavior in the evaluation environment, but the reward may incentivize very different behavior in even a slightly different deployment environment. To address these problems, we introduce the Equivalent-Policy Invariant Comparison (EPIC) distance to quantify the difference between two reward functions directly, without a policy optimization step. We prove EPIC is invariant on an equivalence class of reward functions that always induce the same optimal policy. Furthermore, we find EPIC can be efficiently approximated and is more robust than baselines to the choice of coverage distribution. Finally, we show that EPIC distance bounds the regret of optimal policies even under different transition dynamics, and we confirm empirically that it predicts policy training success. Our source code is available at https://github.com/HumanCompatibleAI/evaluating-rewards.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning