Category · AI
Fine-tuning
Continuing training of an LLM on a specific dataset.
LoRA, full fine-tuning, RLHF, DPO. Useful for style, output format, or highly specific domains (legal code, medical). Cost: GPUs + labeled dataset + evaluation. Often, a solid RAG + prompting gets the job done without fine-tuning.
// In action with our clients
Related articles
// See also
