Multi-Objective Reinforcement Learning (MORL) is a framework where an agent learns a policy by interacting with an environment to optimize a vector-valued reward signal, representing multiple, often conflicting, objectives. Unlike standard RL, which seeks to maximize a single cumulative reward, MORL aims to find policies that make optimal trade-offs across all objectives, typically defined by the Pareto front of non-dominated solutions.
