Back to glossary

QLoRA (Quantized LoRA)

An extension of LoRA that combines 4-bit quantization of the base model with low-rank adaptation, enabling fine-tuning of large language models on a single consumer GPU.

QLoRA pushes the accessibility of fine-tuning even further by quantizing the base model to 4-bit precision before applying LoRA adapters. This means a 65B parameter model that normally requires 130GB of VRAM can be fine-tuned on a single 48GB GPU. The 4-bit NormalFloat (NF4) data type used by QLoRA is specifically designed to minimize information loss for normally distributed neural network weights.

The technique introduces two key innovations: 4-bit NormalFloat quantization that is information-theoretically optimal for normally distributed weights, and double quantization that further compresses the quantization constants themselves. Together, these reduce the memory footprint of the base model to roughly 0.5 bytes per parameter while maintaining near-full-precision quality.

For teams that want to fine-tune models without cloud GPU infrastructure, QLoRA is transformative. You can fine-tune a 7B model on a laptop GPU (8GB VRAM) or a 70B model on a single A100. This democratizes custom model development, making it feasible for startups and small teams to create specialized models that compete with much larger organizations. The quality trade-off versus full-precision LoRA is minimal, typically under 1% on benchmarks.

Related Terms