Sentence-BERT (SBERT) is a modification of the BERT model that uses siamese and triplet network structures to derive semantically meaningful sentence embeddings. Unlike BERT, which processes sentence pairs together for tasks like NLI, SBERT creates fixed-size, standalone embeddings for individual sentences. These embeddings can then be compared using efficient metrics like cosine similarity, making large-scale semantic search and clustering computationally feasible. This architecture is foundational for semantic indexing and dense retrieval in systems like Retrieval-Augmented Generation (RAG).
