The Question Every Team Faces
You have a large language model that works well out of the box, but it does not know your proprietary data, your domain terminology, or your specific use cases. You need it to. The two dominant approaches are retrieval-augmented generation (RAG) and fine-tuning. Both work. Neither is universally better. The wrong choice costs months and six figures.
Large language models know Shakespeare, Python syntax, and general medical terminology. What they do not know is your proprietary data: internal documentation, customer interaction histories, product specifications, or regulatory filings.
Understanding the Technical Landscape
Before choosing between RAG and fine-tuning, it is essential to understand how each approach works at a technical level. Both leverage large language models, but they do so in fundamentally different ways, with different implications for system architecture, operational overhead, and long-term maintenance.
Large language models like GPT-4, Claude, and Llama are trained on vast public corpora. They know Shakespeare, Python syntax, and general medical terminology. What they do not know is your proprietary data: internal documentation, customer interaction histories, product specifications, or regulatory filings. Bridging this gap is the core challenge that RAG and fine-tuning address.
What RAG Actually Does
RAG keeps the base model frozen and feeds relevant context into the prompt at inference time. When a user asks a question, the system retrieves documents from a vector store, ranks them by relevance, and includes the top results in the prompt context. The model then generates an answer grounded in that retrieved information.
The architecture typically involves several components working together. First, an embedding model converts your documents into high-dimensional vectors. These vectors are stored in a vector database such as Chroma, Pinecone, Weaviate, or pgvector. When a query arrives, the same embedding model converts the query into a vector, and the database performs a similarity search to find the most relevant documents.
When RAG Wins
- Data changes frequently. If your knowledge base updates daily, RAG reflects those changes immediately without retraining.
- You need source attribution. RAG naturally points to the documents that informed the answer, which is critical for compliance and trust.
- Domain is narrow but deep. Legal, medical, and technical domains where specific document retrieval outperforms general knowledge.
- Cost control matters. No GPU training, no model hosting beyond inference. Vector storage is cheap.
What Fine-Tuning Actually Does
Fine-tuning adjusts the model's weights on a domain-specific dataset, teaching it patterns, terminology, and reasoning styles unique to your context. The model internalises the knowledge rather than retrieving it on demand.
The process involves collecting a high-quality dataset of input-output pairs that represent your domain. For customer support, this might be thousands of historical tickets and their resolutions. For legal analysis, this might be case summaries and their outcomes. The base model is then trained on this dataset using techniques like supervised fine-tuning (SFT) or, for more advanced cases, reinforcement learning from human feedback (RLHF).
This produces a model that generates text in your organisation's voice, understands internal acronyms without explanation, and can reason about proprietary concepts without explicit context in every prompt. The model becomes an extension of your institutional knowledge, encoded in its billions of parameters.
When Fine-Tuning Wins
- Style and tone matter. Customer-facing chatbots that must sound like your brand, legal writing with specific phrasing.
- Latency is critical. Shorter prompts mean faster inference. A fine-tuned model needs less context to produce correct output.
- Tasks are repetitive and well-defined. Form filling, code generation in a specific stack, structured data extraction.
- Data is stable and proprietary. Internal knowledge that never leaves your environment and changes slowly.
The Decision Framework
We use a five-dimension framework with clients to navigate this choice. Each dimension represents a trade-off that shapes the architecture, cost structure, and operational model of your AI system. No dimension is decisive alone; the interaction between them determines the optimal approach.
1. Data Volatility
High volatility strongly favours RAG. If your product catalogue, documentation, or regulations change weekly, fine-tuning becomes a treadmill. RAG updates happen at the vector store level in minutes.
2. Attribution Requirements
If users or auditors need to verify where an answer came from, RAG is the only practical choice. Fine-tuned models are black boxes; you cannot trace a specific output back to a specific training example.
3. Latency Budget
RAG adds retrieval time to inference time. For sub-second response requirements, fine-tuning eliminates the retrieval step. However, efficient vector stores and caching can bring RAG latency within acceptable bounds for most applications.
4. Cost Structure
Fine-tuning has high upfront costs (GPU time, data preparation) but lower per-query costs. RAG has minimal upfront costs but pays retrieval and longer-context inference on every query. For high-volume applications, the crossover point where fine-tuning becomes cheaper typically arrives at 100,000+ queries per month.
5. Expertise Available
RAG requires vector database expertise, embedding model selection, and retrieval tuning. Fine-tuning requires ML engineering, hyperparameter optimisation, and evaluation infrastructure. Most teams find RAG easier to get right the first time.
The Hybrid Reality
In production, the best systems often combine both approaches. We frequently fine-tune a base model for style and task understanding, then use RAG for factual grounding. The fine-tuned model generates better questions for retrieval and synthesises retrieved information more coherently.
This hybrid approach costs more to build but produces the best user experience: fast, accurate, on-brand, and attributable. The investment is justified when the stakes are high and the use case is complex.
Common Failure Modes
- RAG without relevance ranking. Dumping the top-5 vector matches into context without re-ranking produces noisy, contradictory prompts. We have seen models generate answers that blend unrelated documents because the retrieval stage was not sophisticated enough.
- Fine-tuning on too little data. Fewer than 1,000 high-quality examples rarely produces reliable results and often degrades base model performance. We have seen teams train on 200 examples and wonder why the model performs worse than the base.
- Ignoring prompt engineering. Neither approach replaces solid prompt design. The best RAG system fails with poorly structured prompts. The best fine-tuned model still benefits from clear instructions and few-shot examples.
- No evaluation framework. Teams ship without measuring accuracy, latency, or cost per query. All three drift over time. We recommend continuous evaluation against a held-out test set, with automated regression detection.
- Treating fine-tuning as a one-off. Models drift as the world changes. Without regular retraining schedules, fine-tuned models become outdated. We have seen models trained six months ago produce answers based on old pricing, deprecated products, or superseded policies.
Our Recommendation
Start with RAG unless you have a clear, measurable reason not to. It is faster to implement, easier to iterate, and provides the attribution that business stakeholders demand. Most teams can ship a working RAG system in 2-4 weeks, compared to 2-3 months for fine-tuning.
Build RAG first, then add fine-tuning as a performance optimisation. Not the other way around. The RAG system gives you a baseline for evaluation. Without it, you cannot measure whether fine-tuning actually improves anything.
Finally, invest in evaluation infrastructure before you ship. Define what good looks like, measure it continuously, and be prepared to revert if metrics degrade. The organisations that succeed with AI treat evaluation as a first-class concern, not an afterthought.