← Back to Benchmarks
simmediumatarimetric · varies

Augmenting Replay in World Models for Continual Reinforcement Learning

Description

Continual RL requires an agent to learn new tasks without forgetting previous ones, while improving on both past and future tasks. The most common approaches use model-free algorithms and replay buffers can help to mitigate catastrophic forgetting, but often struggle with scalability due to large memory requirements. Biologically inspired replay suggests replay to a world model, aligning with model-based RL; as opposed to the common setting of replay in model-free algorithms. Model-based RL offe

Source

http://arxiv.org/abs/2401.16650v3