Memory chunking is the process of grouping individual units of information—such as words, tokens, or data points—into larger, meaningful wholes called chunks. In cognitive science, this explains how human short-term memory holds ~7±2 chunks. In AI systems, it is a preprocessing algorithm applied to documents, conversations, or data streams before storage in a vector database or knowledge graph. Effective chunking balances semantic integrity with practical constraints like context window limits and embedding model input sizes.
