Differentiable planning formulates planning problems—such as generating sequences of actions to achieve a goal—within a differentiable computational graph. This allows gradients from a downstream loss function (e.g., task success) to flow backward through the planning steps, enabling the planner's parameters and the world model it uses to be optimized via gradient descent. This bridges the gap between symbolic, discrete search and continuous neural network optimization.
