← Back to Benchmarks
simmediumoffline-rlmetric · varies

Sparse-Reg: Improving Sample Complexity in Offline Reinforcement Learning using Sparsity

Description

In this paper, we investigate the use of small datasets in the context of offline reinforcement learning (RL). While many common offline RL benchmarks employ datasets with over a million data points, many offline RL applications rely on considerably smaller datasets. We show that offline RL algorithms can overfit on small datasets, resulting in poor performance. To address this challenge, we introduce "Sparse-Reg": a regularization technique based on sparsity to mitigate overfitting in offline r

Source

http://arxiv.org/abs/2506.17155v2