A world model is an internal, learned representation within an artificial intelligence agent that predicts future environmental states and expected rewards based on current states and potential actions. This learned dynamics model allows the agent to simulate or 'imagine' the consequences of action sequences without costly, real-world interaction. By compressing sensory experience into a latent representation, it enables efficient planning, such as via Model Predictive Control (MPC) or training policies through imagined rollouts, directly addressing the challenge of sample efficiency in reinforcement learning.
