Fine-tuning vs Prompting.
Two ways to make an AI model do your specific task. Prompting (giving the model good instructions) is cheap and fast — usually the right answer. Fine-tuning (retraining the model on your data) is expensive and slow — needed in a small number of cases.
In plain English
There are two main ways to adapt an LLM to your specific task: prompting (giving the model good instructions for each request) and fine-tuning (retraining the model on examples of your task). For most SMB applications, prompting is the right answer. Fine-tuning is needed in a small minority of cases.
Prompting works because modern LLMs are big enough and capable enough to follow detailed instructions reliably. You don't need to train the model to extract data from your specific invoices — you just need to tell it what fields you want, give it some examples, and define the output format. This is called few-shot prompting and it handles 90% of business use cases.
Fine-tuning makes sense when: the task is very specific to your domain (specialist medical or legal language), you need to consistently match a very particular output style, you're processing such high volumes that the cost of long prompts becomes meaningful, or you have specific data that the base model genuinely doesn't know.
For 95% of SMB use cases — document extraction, enquiry triage, quote generation, internal assistants — well-engineered prompts with good context (often via RAG) are faster, cheaper, and just as accurate as fine-tuning. Fine-tuning takes weeks, costs thousands, and locks you to a specific model version. Prompting takes hours, costs almost nothing, and lets you switch between Anthropic, OpenAI, and Google as the models improve.
If an AI consultancy proposes fine-tuning for your invoice matching, ask them why prompting wouldn't work first.
Real examples
What this looks like in practice.
- Prompting (95% of cases): structured prompt with examples → invoice extracted into your schema, no training required.
- Fine-tuning (rare): specialist medical coding where the base model genuinely doesn't know the domain language.
- Prompting + RAG: assistant queries your live business data and uses it in the response — no fine-tuning needed.
- Fine-tuning (rare): matching a very specific brand voice across millions of outputs where prompt context isn't practical.
See in action
Where we deliver this for clients.
Related terms
Adjacent concepts.
Apply it
Want this in your business?
Book a free 15-minute call. We'll talk through whether this pattern actually fits your workflow — and tell you honestly if it doesn't.
Let's see if we can help.
A 15-minute chat with Chris & Kay. No slides. No pitch deck. You tell us what's on your plate; we follow up by email with real thinking.