Self-supervised learning (SSL) is a machine learning paradigm where a model generates its own supervisory signals from the inherent structure of unlabeled data, typically by solving a pretext task like predicting masked input segments or future data points. This approach enables the learning of rich, general-purpose data representations, forming a crucial foundation for world models and other advanced agentic cognitive architectures that require a compressed understanding of their environment.
