Techniques

What is LoRA (Low-Rank Adaptation)?

An efficient method to fine-tune AI models using much less compute and memory.

Definition

LoRA (Low-Rank Adaptation) is a technique for efficiently fine-tuning large AI models by only modifying a small number of additional parameters rather than updating the entire model. This dramatically reduces the computational cost, memory requirements, and training time needed for customization while achieving results comparable to full fine-tuning.

๐Ÿ’ก Example

Instead of fine-tuning all 70 billion parameters of Llama 3, LoRA might only train 10 million additional parameters (0.01%) to adapt the model for medical diagnosis. This can be done on a single GPU instead of requiring a cluster.

Related concepts

LLM (Large Language Model)

A type of AI trained on massive text datasets to understand and generate human language.

โ†’
Fine-Tuning

Training a pre-trained AI model on specialized data to improve performance on specific tasks.

โ†’

Explore AI tools

Find tools that use lora (low-rank adaptation) in practice.

Browse all tools โ†’ Back to glossary
What is LoRA (Low-Rank Adaptation)?

LoRA (Low-Rank Adaptation) is a technique for efficiently fine-tuning large AI models by only modifying a small number of additional parameters rather than updating the entire model. This dramatically reduces the computational cost, memory requirements, and training time needed for customization while achieving results comparable to full fine-tuning.

How does LoRA (Low-Rank Adaptation) work in practice?

Instead of fine-tuning all 70 billion parameters of Llama 3, LoRA might only train 10 million additional parameters (0.01%) to adapt the model for medical diagnosis. This can be done on a single GPU instead of requiring a cluster.