Catastrophic forgetting is the phenomenon where an artificial neural network loses previously acquired knowledge when trained on new data or tasks, effectively overwriting old weights. This occurs because standard backpropagation and gradient descent algorithms optimize for the current training distribution without explicit mechanisms to preserve past learning. In sequential learning or continual learning scenarios, this leads to a drastic, non-gradual drop in accuracy on earlier tasks, undermining the model's ability to accumulate knowledge over time.
