AIXI is a theoretical, mathematical formulation of an optimal reinforcement learning agent that maximizes expected future rewards by combining Solomonoff induction for sequence prediction with sequential decision theory. It represents a Bayesian ideal: an agent that maintains a mixture of all computable environment models, updates its beliefs via Bayesian inference, and chooses actions that maximize the reward sum over its future horizon. However, AIXI is incomputable in practice, serving primarily as a gold standard for evaluating the optimality of real-world algorithms.
