Context summarization is a memory compression technique for long-context models that creates a condensed representation of past information to manage context window limits. It functions as a form of lossy compression, selectively preserving salient facts, decisions, and entity relationships from a prior interaction history. This condensed context is then re-injected into the model's limited input window, enabling extended multi-turn reasoning without exceeding token constraints.
