← Back to Benchmarks
simmediumoffline-rlmetric · varies

ShiQ: Bringing back Bellman to LLMs

Description

The fine-tuning of pre-trained large language models (LLMs) using reinforcement learning (RL) is generally formulated as direct policy optimization. This approach was naturally favored as it efficiently improves a pretrained LLM, seen as an initial policy. Another RL paradigm, Q-learning methods, has received far less attention in the LLM community while demonstrating major success in various non-LLM RL tasks. In particular, Q-learning effectiveness comes from its sample efficiency and ability t

Source

http://arxiv.org/abs/2505.11081v1