Abbeal

Category · AI

Fine-tuning

Continuing training of an LLM on a specific dataset.

LoRA, full fine-tuning, RLHF, DPO. Useful for style, output format, or highly specific domains (legal code, medical). Cost: GPUs + labeled dataset + evaluation. Often, a solid RAG + prompting gets the job done without fine-tuning.

// In action with our clients

// See also

Want us to apply this for you?

Talk to an architect