Lifelong Reinforcement Learning with Modulating Masks

December 21, 2022 ยท Entered Twilight ยท ๐Ÿ› Trans. Mach. Learn. Res.

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, Dockerfile, LICENSE, README.md, README_deeprl.md, deep_rl, env_configs, eval_continualworld.py, eval_ctgraph.py, eval_minigrid.py, paper_experiments.txt, requirements.txt, setup.py, train_continualworld.py, train_ctgraph.py, train_minigrid.py

Authors Eseoghene Ben-Iwhiwhu, Saptarshi Nath, Praveen K. Pilly, Soheil Kolouri, Andrea Soltoggio arXiv ID 2212.11110 Category cs.LG: Machine Learning Cross-listed cs.AI, stat.ML Citations 27 Venue Trans. Mach. Learn. Res. Repository https://github.com/dlpbc/mask-lrl โญ 3 Last Checked 1 month ago
Abstract
Lifelong learning aims to create AI systems that continuously and incrementally learn during a lifetime, similar to biological learning. Attempts so far have met problems, including catastrophic forgetting, interference among tasks, and the inability to exploit previous knowledge. While considerable research has focused on learning multiple supervised classification tasks that involve changes in the input distribution, lifelong reinforcement learning (LRL) must deal with variations in the state and transition distributions, and in the reward functions. Modulating masks with a fixed backbone network, recently developed for classification, are particularly suitable to deal with such a large spectrum of task variations. In this paper, we adapted modulating masks to work with deep LRL, specifically PPO and IMPALA agents. The comparison with LRL baselines in both discrete and continuous RL tasks shows superior performance. We further investigated the use of a linear combination of previously learned masks to exploit previous knowledge when learning new tasks: not only is learning faster, the algorithm solves tasks that we could not otherwise solve from scratch due to extremely sparse rewards. The results suggest that RL with modulating masks is a promising approach to lifelong learning, to the composition of knowledge to learn increasingly complex tasks, and to knowledge reuse for efficient and faster learning.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Machine Learning