Experience replay is a reinforcement learning technique where an agent stores its past experiences—each a tuple of (state, action, reward, next state)—in a fixed-size memory buffer and later randomly samples from this buffer to perform learning updates. This mechanism decouples the data-generation process (acting in the environment) from the learning process, breaking harmful temporal correlations in sequential observations and allowing the same valuable experience to be reused multiple times, dramatically improving data efficiency and training stability.
