The process of taking a pre-trained model and continuing to train it on a smaller, specific dataset to adapt it for a particular task or domain.
The process of taking a pre-trained model and continuing to train it on a smaller, specific dataset to adapt it for a particular task or domain. Much cheaper and faster than training from scratch. LoRA and QLoRA are popular efficient fine-tuning methods that only update a small fraction of parameters.
Fine-tuning takes a pre-trained model and trains it further on a specific dataset for a particular task. The model already knows language from its initial training — fine-tuning specializes that knowledge. It's like taking someone who speaks English fluently and teaching them medical terminology for a specific job.
The process involves feeding the model task-specific examples and adjusting its weights. You might fine-tune a base LLM on thousands of customer service conversations to make it better at handling support tickets. Or fine-tune on legal documents to create a model that understands legal language. The model retains its general abilities while becoming much better at the target domain.
Fine-tuning used to require serious hardware and expertise, but techniques like LoRA have made it much more accessible. You can now fine-tune a 7-billion parameter model on a single GPU in a few hours. The key decisions are: what data to use (quality matters more than quantity), how long to train (too much leads to overfitting), and which base model to start from. For many use cases, prompt engineering or RAG might be good enough — fine-tuning is worth the effort when you need consistent behavior at scale.
"We fine-tuned LLaMA on 10,000 medical Q&A pairs, and it now outperforms the base model on clinical questions by a wide margin."
The initial, expensive phase of training where a model learns general patterns from a massive dataset.
Low-Rank Adaptation.
Using knowledge learned from one task to improve performance on a different but related task.
A mathematical function applied to a neuron's output that introduces non-linearity into the network.
An optimization algorithm that combines the best parts of two other methods — AdaGrad and RMSProp.
Artificial General Intelligence.
Browse our complete glossary or subscribe to our newsletter for the latest AI news and insights.