Learning to Modulate pre-trained Models in RL

May 3, 2023

Speakers

About

Reinforcement Learning (RL) has experienced great success in complex games and simulations. However, RL agents are often highly specialized for a particular task, and it is difficult to adapt a trained agent to a new task.In supervised learning, an established paradigm is multi-task pre-training followed by fine-tuning.A similar trend is emerging in RL, where agents are pre-trained on data collections that comprise a multitude of tasks.Despite these developments, it remains an open challenge how to adapt such pre-trained agents to novel tasks while retaining performance on the pre-training tasks.In this regard, we pre-train an agent on a set of tasks from the Meta-World benchmark suite and adapt it to tasks from Continual-World. We conduct a comprehensive comparison of fine-tuning methods originating from supervised learning in our setup.Our findings show that fine-tuning is feasible, but for existing methods, performance on previously learned tasks often deteriorates.Therefore, we propose a novel approach that avoids forgetting by modulating the information flow of the pre-trained model. Our method outperforms existing fine-tuning approaches, and achieves state-of-the-art performance on the Continual-World benchmark.To facilitate future research in this direction, we collect datasets for all Meta-World tasks and make them publicly available.

Organizer

Like the format? Trust SlidesLive to capture your next event!

Professional recording and live streaming, delivered globally.

Sharing

Recommended Videos

Presentations on similar topic, category or speaker

Interested in talks like this? Follow ICLR 2023