← Back to Benchmarks
simmediumlocomotionmetric · varies

Fast Adaptation with Behavioral Foundation Models

Description

Unsupervised zero-shot reinforcement learning (RL) has emerged as a powerful paradigm for pretraining behavioral foundation models (BFMs), enabling agents to solve a wide range of downstream tasks specified via reward functions in a zero-shot fashion, i.e., without additional test-time learning or planning. This is achieved by learning self-supervised task embeddings alongside corresponding near-optimal behaviors and incorporating an inference procedure to directly retrieve the latent task embed

Source

http://arxiv.org/abs/2504.07896v1