LoRA (Low-Rank Adaptation) is a parameter-efficient fine-tuning technique that adds small trainable matrices to frozen pretrained weights. Instead of updating all parameters, LoRA decomposes weight updates into low-rank matrices, dramatically reducing trainable parameters and memory.
LoRA enables fine-tuning large models on consumer hardware while maintaining quality close to full fine-tuning.