Meta-Learning for Multi-objective Reinforcement Learning

November 08, 2018 Β· Declared Dead Β· πŸ› IEEE/RJS International Conference on Intelligent RObots and Systems

πŸ‘» CAUSE OF DEATH: Ghosted
No code link whatsoever

"No code URL or promise found in abstract"

Evidence collected by the PWNC Scanner

Authors Xi Chen, Ali Ghadirzadeh, MΓ₯rten BjΓΆrkman, Patric Jensfelt arXiv ID 1811.03376 Category cs.AI: Artificial Intelligence Citations 64 Venue IEEE/RJS International Conference on Intelligent RObots and Systems Last Checked 1 month ago
Abstract
Multi-objective reinforcement learning (MORL) is the generalization of standard reinforcement learning (RL) approaches to solve sequential decision making problems that consist of several, possibly conflicting, objectives. Generally, in such formulations, there is no single optimal policy which optimizes all the objectives simultaneously, and instead, a number of policies has to be found each optimizing a preference of the objectives. In other words, the MORL is framed as a meta-learning problem, with the task distribution given by a distribution over the preferences. We demonstrate that such a formulation results in a better approximation of the Pareto optimal solutions in terms of both the optimality and the computational efficiency. We evaluated our method on obtaining Pareto optimal policies using a number of continuous control problems with high degrees of freedom.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Artificial Intelligence

Died the same way β€” πŸ‘» Ghosted