Towards continual reinforcement learning: A review and perspectives

K Khetarpal, M Riemer, I Rish, D Precup - Journal of Artificial Intelligence …, 2022 - jair.org
In this article, we aim to provide a literature review of different formulations and approaches
to continual reinforcement learning (RL), also known as lifelong or non-stationary RL. We …

Deep reinforcement learning: a survey

H Wang, N Liu, Y Zhang, D Feng, F Huang, D Li… - Frontiers of Information …, 2020 - Springer
Deep reinforcement learning (RL) has become one of the most popular topics in artificial
intelligence research. It has been widely used in various fields, such as end-to-end control …

A survey of meta-reinforcement learning

J Beck, R Vuorio, EZ Liu, Z **ong, L Zintgraf… - arxiv preprint arxiv …, 2023 - arxiv.org
While deep reinforcement learning (RL) has fueled multiple high-profile successes in
machine learning, it is held back from more widespread adoption by its often poor data …

Meta-learning in neural networks: A survey

T Hospedales, A Antoniou, P Micaelli… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
The field of meta-learning, or learning-to-learn, has seen a dramatic rise in interest in recent
years. Contrary to conventional approaches to AI where tasks are solved from scratch using …

Efficient off-policy meta-reinforcement learning via probabilistic context variables

K Rakelly, A Zhou, C Finn, S Levine… - … on machine learning, 2019 - proceedings.mlr.press
Deep reinforcement learning algorithms require large amounts of experience to learn an
individual task. While meta-reinforcement learning (meta-RL) algorithms can enable agents …

Human-timescale adaptation in an open-ended task space

AA Team, J Bauer, K Baumli, S Baveja… - arxiv preprint arxiv …, 2023 - arxiv.org
Foundation models have shown impressive adaptation and scalability in supervised and self-
supervised learning problems, but so far these successes have not fully translated to …

Human-timescale adaptation in an open-ended task space

J Bauer, K Baumli, F Behbahani… - International …, 2023 - proceedings.mlr.press
Foundation models have shown impressive adaptation and scalability in supervised and self-
supervised learning problems, but so far these successes have not fully translated to …

Fast context adaptation via meta-learning

L Zintgraf, K Shiarli, V Kurin… - International …, 2019 - proceedings.mlr.press
We propose CAVIA for meta-learning, a simple extension to MAML that is less prone to meta-
overfitting, easier to parallelise, and more interpretable. CAVIA partitions the model …

Varibad: A very good method for bayes-adaptive deep rl via meta-learning

L Zintgraf, K Shiarlis, M Igl, S Schulze, Y Gal… - arxiv preprint arxiv …, 2019 - arxiv.org
Trading off exploration and exploitation in an unknown environment is key to maximising
expected return during learning. A Bayes-optimal policy, which does so optimally, conditions …

Offline Meta Reinforcement Learning--Identifiability Challenges and Effective Data Collection Strategies

R Dorfman, I Shenfeld, A Tamar - Advances in Neural …, 2021 - proceedings.neurips.cc
Consider the following instance of the Offline Meta Reinforcement Learning (OMRL)
problem: given the complete training logs of $ N $ conventional RL agents, trained on $ N …