A Partially Observable Markov Decision Process (POMDP) is a mathematical framework for modeling sequential decision-making problems where an agent cannot directly observe the true state of the environment and must maintain a belief state, a probability distribution over possible states. It extends the Markov Decision Process (MDP) by incorporating observations that provide noisy, incomplete information about the underlying state, formalizing the core challenge of acting under uncertainty.
