We propose PTGM (Pre-Training Goal-based Models), a pre-training approach for goal-based models that achieves sample-efficient reinforcement learning. Our method significantly reduces the sample complexity of RL algorithms through effective pre-training strategies that leverage goal-conditioned learning.
Sample efficiency remains one of the most significant challenges in reinforcement learning, particularly when deploying algorithms in real-world scenarios where data collection is expensive and time-consuming. Traditional RL methods often require millions of interactions with the environment to learn effective policies.
We provide theoretical justification for our approach, showing that goal-conditioned pre-training leads to better representation learning and faster convergence guarantees.