In model-based reinforcement learning (MBRL), uncertainty quantification involves estimating both epistemic uncertainty (model uncertainty due to limited data) and aleatoric uncertainty (inherent environmental stochasticity) in a learned dynamics model's predictions. This allows an agent to distinguish between what it knows and what it does not, enabling robust planning by avoiding states where predictions are unreliable and guiding exploration towards regions of high model error to improve sample efficiency.
