YaRN (Yet another RoPE extensioN) is a parameter-efficient fine-tuning technique designed to extend the functional context window of pre-trained large language models that utilize Rotary Positional Embedding (RoPE). It combines theoretical insights from NTK-aware scaling with a practical temperature-tuning strategy, allowing models to generalize to sequences significantly longer than their original training length with minimal additional training data and compute. This makes it a highly efficient alternative to full model retraining for long-context applications.
