Embedding normalization is the preprocessing step of scaling a vector embedding to have a unit norm (a length of 1). This operation transforms any non-zero vector into a direction-only representation on the surface of a hypersphere, which is essential for computing similarity metrics like cosine similarity efficiently as a simple dot product. It is a standard practice in retrieval-augmented generation (RAG) and semantic search pipelines.
