In model-based reinforcement learning (MBRL), a reward model is a learned function, often parameterized by a neural network, that estimates the immediate or cumulative reward an agent will receive for taking a specific action in a given state. It serves as a critical component of an agent's internal world model, alongside a dynamics model (or transition model) that predicts state transitions. By learning this function from interaction data, the agent can simulate and evaluate the outcomes of potential action sequences without costly real-world trials, enabling more sample-efficient planning and policy optimization.
