Out-of-distribution detection is the task of identifying whether an input query or data point falls outside the known statistical distribution that a machine learning model, such as an embedding model, was trained on. This is crucial for monitoring embedding quality and system robustness, as models often make unreliable predictions or generate poor-quality embeddings when faced with unfamiliar data. In agentic memory systems, detecting OOD inputs prevents the retrieval of irrelevant or low-confidence information.
