Data deduplication is a storage optimization technique that identifies and eliminates duplicate copies of repeating data, replacing them with references to a single stored instance. In agentic memory and context management, this process is critical for reducing the storage footprint of vector embeddings, knowledge graph triples, and episodic logs, thereby lowering costs and improving retrieval latency. It operates at the file, block, or byte level, often using cryptographic hashing to detect identical data segments.
