Catastrophic forgetting, also known as catastrophic interference, is the tendency of an artificial neural network to overwrite previously learned weights and representations when trained sequentially on new tasks or data distributions. This phenomenon occurs because standard gradient-based optimization lacks mechanisms to protect consolidated knowledge, treating all parameters as equally plastic. It is a primary obstacle in continual learning and lifelong learning systems, which aim to accumulate knowledge over time without full retraining on all past data.
