- 51% of enterprise AI deployments use RAG in production; only 9% use fine-tuning — the gap reflects missing decision criteria, not a verdict on which approach is better.
- RAG, fine-tuning, and long-context prompting solve different problems: knowledge access, behavioural consistency, and corpus injection each require a distinct architectural response.
- A four-variable decision framework — knowledge volatility, query volume, team capability, and output constraints — determines the correct architecture for each enterprise workload.
Key Claim: The most common cause of LLM architecture regret in enterprise deployments is conflating what a model can see (retrieval) with how it behaves (fine-tuning) — two structurally different problems that require different solutions.
Fifty-one percent of enterprise AI deployments now use retrieval-augmented generation (RAG) in production. Only 9% use fine-tuning. Those figures, from the Menlo Ventures 2024 State of Generative AI in the Enterprise report, are not a story about one technology winning — they are a story about missing decision criteria. Most enterprise teams default to RAG not because they evaluated the alternatives but because no framework told them when to choose differently. What they lack is a decision model that accounts for knowledge volatility, query volume, team capability, and cost.
The framing of “RAG vs. fine-tuning” has always been slightly wrong. The two approaches solve different problems. RAG changes what a model can see for a specific query. Fine-tuning changes how a model behaves across every query it will ever handle. Conflating them — choosing one primarily because a vendor recommended it or because a tutorial covered it first — is the single most common cause of LLM architecture regret in enterprise deployments.
A third option, underweighted in most coverage, has become genuinely viable: loading a knowledge base directly into a long-context window with prompt caching. For certain workloads, it eliminates the need for retrieval infrastructure entirely.
The State of Production Adoption
The Menlo Ventures data makes the current reality concrete. As of their 2024 survey, 51% of enterprise AI deployments use RAG in production — up from 31% the prior year. Fine-tuning is used in only 9% of production models. Despite the vendor noise around fine-tuning as a necessary step toward enterprise customisation, the overwhelming majority of production deployments are using the model out of the box with retrieval, not modifying its weights.
This gap between marketing and production is informative. RAG dominates partly because it is the path of least resistance — no ML infrastructure required, knowledge updates are immediate, and the retrieval pipeline can be built by data engineers rather than model trainers. But it also dominates because, for most enterprise use cases, it is the genuinely correct choice. The 9% fine-tuning figure is not evidence of underinvestment; it is a signal about which problems actually require fine-tuning.
Three Architectural Modes
Mode 1: Prompt Engineering and Long-Context Injection
For knowledge bases that fit within a model’s context window — roughly up to 200,000 tokens as a working threshold — full-document injection combined with prompt caching can be faster, cheaper, and simpler than building retrieval infrastructure.
Anthropic’s prompt caching delivers up to 90% cost reduction and up to 85% latency reduction for repeated long-context prompts, per documented benchmarks. The economics are compelling for internal copilots, document analysis tools, and knowledge bases queried by a small number of users. A legal team running contract review against a fixed set of standard forms is a better fit for this architecture than for a retrieval pipeline.
The ceiling is query volume. At enterprise scale — 50 million tokens of corpus, 10,000 queries per day, 500,000 tokens per query — full long-context injection generates approximately $4.5 million per year in inference costs for a single use case. Beyond that threshold, retrieval is not optional.
Mode 2: Retrieval-Augmented Generation
RAG remains the correct default for enterprise knowledge access when the knowledge base is large, changes frequently, or is queried at scale. It solves for knowledge currency — the model always has access to the most recent version of a document without retraining — and for auditability, since the retrieved sources are visible and citable.
The 2025 LaRA benchmark, published at ICML (Alibaba NLP), tested 11 large language models across 2,326 cases at 32K and 128K context lengths. Its finding: there is no definitive conclusion regarding the superiority of RAG versus long-context approaches. Optimal performance depends on context length, model capability, task type, and retrieval quality. This is relevant for enterprise teams building on the assumption that more capable models will solve retrieval problems — they will not.
RAG’s primary failure mode is not hallucination; it is retrieval error. When the retrieved context is wrong or incomplete, the model’s output will be confidently wrong. Teams that measure RAG performance by end-to-end accuracy and find it wanting often have a retrieval problem they are diagnosing as a model problem.
Mode 3: Fine-Tuning
Fine-tuning is warranted in four specific scenarios: when behavioural consistency across outputs cannot be enforced by prompting alone; when output format or style constraints are too rigid for few-shot examples to reliably enforce; when latency requirements rule out the overhead of a retrieval call; and when the model needs to learn a domain-specific reasoning pattern absent from the base model’s training data.
It is not warranted for knowledge injection. Loading proprietary facts into model weights through fine-tuning is a documented failure mode — the model will exhibit inconsistent recall, and the knowledge cannot be updated without retraining. This is the correct use case for RAG, not fine-tuning.
Parameter-efficient methods — LoRA and QLoRA in particular — have reduced the compute and data requirements for fine-tuning to the point where teams without dedicated GPU infrastructure can attempt it on smaller models. But “can run it” and “should run it” are different questions. The operational overhead of managing a fine-tuned model, its retraining cadence, and its evaluation pipeline is non-trivial. Most enterprise teams underestimate it before they begin.
The Hybrid Case
UC Berkeley, Microsoft, and Meta Research published RAFT (Retrieval Augmented Fine-Tuning) in 2024, demonstrating a specific hybrid approach: fine-tune a model to be a better RAG reader. Rather than using fine-tuning to inject knowledge, RAFT trains the model to distinguish relevant retrieved documents from distractor documents — and to cite evidence accurately from the context window. The training dataset uses an 80/20 split: 80% of examples include the oracle document, 20% do not, forcing the model to develop retrieval-independent reasoning as a fallback.
Tested on PubMed, HotpotQA, and the Gorilla benchmark, RAFT consistently outperformed RAG alone — though generalisation across all enterprise task types is not yet established. The key insight is the combination’s division of labour: RAG handles knowledge currency and scale; fine-tuning handles the reasoning quality applied to retrieved context. This is a materially different use of fine-tuning than knowledge injection.
In production, the hybrid architecture requires both ML and data engineering capability simultaneously. That is why, despite demonstrably better performance, hybrid remains a minority pattern. This connects directly to a broader pattern in enterprise AI: teams choosing architectures based on what their team can maintain, not what is theoretically optimal.
A Four-Variable Decision Framework
Knowledge volatility. If the knowledge base changes faster than a retraining cycle — daily, weekly, any cadence requiring near-real-time currency — the knowledge must live in retrieval, not weights. Fine-tuning is not viable. The choice is between RAG and long-context injection.
Query volume. As a rough heuristic based on current per-token pricing: if query volume is low (on the order of tens to low hundreds of queries per day) against a sub-200K-token knowledge base, evaluate long-context with prompt caching first — the total cost may undercut a RAG pipeline. As volume grows, the economics shift decisively toward retrieval. Calculate your own crossover point from your provider’s cache-hit pricing and expected query volume.
Team capability. RAG requires data engineering. Fine-tuning requires ML infrastructure and evaluation. Long-context requires neither — it requires a larger API budget. A practical litmus test: if your team cannot answer “who owns the evaluation pipeline and what does it measure,” you are not ready for fine-tuning. Teams being pushed toward fine-tuning by a vendor without a dedicated ML function should default to RAG for dynamic knowledge and prompt engineering for behavioural consistency. This capability gap also explains why enterprise AI production gaps persist even after meaningful investment.
Output constraints. If the use case requires rigid output formats, specific tone patterns, or reasoning structures that few-shot prompting cannot reliably enforce at scale — structured JSON extraction from messy documents, for example, or a consistent clinical note format — fine-tuning is worth the overhead. A practical threshold: if your current few-shot prompt achieves greater than 95% format compliance in testing, fine-tuning for format consistency is unlikely to justify the overhead.
What to Watch
The emerging pressure point is not the RAG-versus-fine-tuning question but the retrieval quality question. As enterprise deployments scale, the bottleneck shifts from model capability to the precision of the retrieval layer — chunk size, embedding model quality, re-ranking architecture, and query reformulation. Teams that built minimal viable RAG pipelines during the 2024 adoption surge are now hitting retrieval ceilings that require engineering rather than model upgrades to address.
The second trend worth tracking is the continued reduction in fine-tuning cost. As LoRA variants improve and smaller models approach larger model performance on narrow tasks, the break-even point for fine-tuning will shift. For teams with defined, stable output requirements, that calculus is likely to move in fine-tuning’s favour — an editorial projection based on the current trajectory of LoRA variant improvements, not a sourced forecast.
The teams that will navigate this best are those that have separated the knowledge question from the behaviour question in their architecture. Knowledge belongs in retrieval. Behaviour belongs in the model — whether through prompting, fine-tuning, or both. That framing also bears on how enterprises should think about reasoning model deployments, where the cost-vs-capability trade-off is sharpest.
This article was produced with AI assistance and reviewed by the editorial team.



