Model-Based Reinforcement Learning (MBRL) is a machine learning paradigm where an agent learns an internal, predictive model of its environment's dynamics and reward function, which it then uses for planning and policy optimization to improve sample efficiency. Unlike model-free methods that learn a policy or value function directly from experience, MBRL agents can simulate potential futures via their model to evaluate actions without costly real-world interaction, making them advantageous for domains like robotics and autonomous systems where data collection is expensive or risky.
