Adapter layers are small, trainable neural network modules inserted into a pre-trained model to adapt it for a new task or modality, enabling parameter-efficient fine-tuning without modifying the bulk of the original model weights. Typically consisting of a down-projection, a non-linearity, and an up-projection, they create a bottleneck that captures task-specific knowledge. During fine-tuning, only the adapter parameters and a new task head are updated, freezing the foundational model to preserve its general knowledge and prevent catastrophic forgetting.
