Research · Tech Industry & Practitioner
Back to sweepResearch sweep · deep · 2025 – 2026
Agentic RAG — Evolution, Challenges, and Decision Criteria
Agentic RAG between November 2025 and May 2026: how retrieval-augmented generation is shifting toward agent-driven architectures, the operational problems (token burn, context management, latency, reliability), information-organisation patterns such as context catalogues and semantic categorisation, parallels with traditional data warehousing (dimensions, measures, star schemas), the evolving RAG tooling landscape, and decision criteria for switching to pure agentic workflows.
- academic
- frontier
- tech
- blogs
- vc
Synthesised 2026-05-10
Narrative
Practitioner coverage from late 2025 through May 2026 reveals a clear inflection. The Thoughtworks Technology Radar's Volume 33 (November 2025) is the most authoritative single signal: after RAG dominated Volume 32 in April 2025, Volume 33 shifted its central theme to agents and MCP, with Thoughtworks CTO Rachel Laycock framing it as a 'step change' toward context engineering. That editorial move compressed what had been a gradual architectural transition into a widely recognised industry event. InfoQ's April 2026 field report from three financial-services deployments (Q4 2025, ~1,500 multi-hop queries) supplied the sharpest empirical grounding: ~30% silent failure rate under static RAG, and roughly 60% of hallucinations tracing to unhandled execution errors rather than model reasoning — a finding with direct implications for where engineering effort should be directed.
On tooling, the picture is one of consolidation and functional specialisation. The benchmark evidence (AIMultiple, January 2026) shows a 53% token-count difference between the most and least efficient frameworks — Haystack at 1.57k tokens versus LangChain at 2.40k — which compounds materially at scale. LangGraph reached 1.0 stability in October 2025 and now functions as the de facto stateful agent orchestration layer, while LlamaIndex handles retrieval depth; MarsDevs and others report the dominant 2026 production pattern is the two-framework combination of LlamaIndex plus LangGraph. Evaluation tooling has matured in parallel: Ragas (400k+ monthly downloads), LangSmith, Arize Phoenix, and Langfuse form a three-layer observability stack, and MarsDevs quantifies production targets as faithfulness ≥0.9, answer relevancy ≥0.85, and context precision ≥0.8.
The cost-and-latency evidence is the most practically consequential finding for practitioners deciding when to switch. MarsDevs reports that agentic loops with three to four iterations take 8–12 seconds against the 1–2 second baseline for standard RAG, a 3–10x token cost multiplier, and a worse p95 latency profile. Redis quantifies semantic caching as recovering up to 73% of cost in high-repetition workloads. CSO Online reports 72–80% of enterprise RAG implementations underperformed or failed within their first year, with 51% of all enterprise AI failures in 2025 being RAG-related — a failure rate that drove both the agentic migration and the growth of specialist evaluation infrastructure. On accuracy, a 2025 MDPI study across 250 clinical vignettes showed a 55-percentage-point multi-hop accuracy gap: 34% for static RAG versus 89% for agentic RAG, described as a 'categorical capability gap' rather than a marginal improvement.
The information organisation dimension is where the least practitioner consensus exists but the most structurally interesting parallels emerge. The Towards Data Science piece from October 2025 documents the M&A evidence — ServiceNow's acquisition of data.world, Salesforce's $8 billion Informatica purchase — as market confirmation that knowledge graphs and metadata management are becoming the semantic backbone for AI. The RAGFlow year-end review introduces 'Context Engineering' as the successor to RAG optimisation, shifting focus from retrieval algorithms to the systematic design of the full retrieval–context assembly–reasoning pipeline. Thoughtworks independently surfaces the semantic layer as critical infrastructure for agentic text-to-SQL, noting that Snowflake Semantic Views, Databricks Metric Views, and dbt MetricFlow are converging around Open Semantic Interchange (OSI) v1.0 — a standardisation signal that partially mirrors dimensional modelling's role in the data warehouse era.
Sources
| ID | Title | Outlet | Date | Significance |
|---|---|---|---|---|
| p1 | Building Hierarchical Agentic RAG Systems: Multi-Modal Reasoning with Autonomous Error Recovery | InfoQ | 2026-04 | Field report from three financial-services deployments (Q4 2025, n=~1,500 multi-hop queries) showing 30% silent-failure rate and finding that ~60% of hallucinations originated from unhandled execution errors rather than model reasoning, directly grounding operational failure-mode claims. |
| p2 | Agentic Retrieval-Augmented Generation: A Survey on Agentic RAG (arXiv 2501.09136 v4) | arXiv (Aditi Singh et al.) | 2026-04 | The most cited practitioner-adjacent survey of agentic RAG architectures; introduces a principled taxonomy by agent cardinality, control structure, and autonomy, with comparative trade-off analysis across healthcare, finance, and enterprise document processing use cases. |
| p3 | Thoughtworks Technology Radar Volume 33 — Themes: Rise of Agents Elevated by MCP, Context Engineering, AI Antipatterns | Thoughtworks Technology Radar | 2025-11 | Authoritative practitioner signal that RAG dominated Volume 32 conversation while Volume 33 shifted to agents and MCP, confirming the industry inflection from static retrieval to agentic workflows as observed by Thoughtworks CTO Rachel Laycock. |
| p4 | Thoughtworks Technology Radar Highlights The Rapid Evolution of AI Assistance in 2025 (Vol. 33 press release) | Thoughtworks / PR Newswire | 2025-11 | Official release statement confirming Volume 33's shift from RAG and prompt engineering (Vol. 32) to context engineering, MCP, and agentic systems, citing the growth of agentic workflows and enterprise AI antipatterns as the dominant themes. |
| p5 | Thoughtworks Technology Radar — Techniques: Semantic Layer for AI (LLM text-to-SQL, dbt MetricFlow, Snowflake Semantic Views) | Thoughtworks Technology Radar | 2025-11 | Practitioner evidence that semantic layers — the closest analogue to dimensional modelling in agentic systems — are now a first-class concern, with Thoughtworks warning that naive LLM text-to-SQL produces incorrect results when business rules live outside the schema. |
| p6 | Thoughtworks Technology Radar Volume 32 — Supervised Agents, RAG Techniques, Data Product Thinking | Thoughtworks | 2025-04 | Volume 32 spotlighted corrective RAG, Fusion-RAG, Self-RAG, and FastGraphRAG as Trial-level techniques, and introduced 'data product thinking' as the data management analogue of product management — directly relevant to context catalogue design. |
| p7 | Thoughtworks Technology Radar — Platforms: Graphiti, Databricks Agent Bricks, Rhesis testing | Thoughtworks Technology Radar | 2026-04 | Practitioner-assessed platform blips including Graphiti (temporal knowledge graph for LLM memory) and Databricks Agent Bricks; explicitly flags that flat vector stores in RAG pipelines fail to track how facts change over time. |
| p8 | Themes from Technology Radar Vol. 33 — Podcast: Infrastructure Automation, Rise of Agents, MCP, AI Antipatterns | Thoughtworks | 2025-11 | Neal Ford and Ken Mugrage explain the editorial reasoning behind Volume 33's shift from RAG to agents and MCP, including the concept of context engineering — 'how do you tell the agents what they're supposed to do and give them roles.' |
| p9 | RAG in 2026: The UK/EU Enterprise Guide to Grounded GenAI | Data Nucleus | 2026-01 | Practitioner guide situating EU AI Act and GDPR obligations alongside agentic RAG architecture guidance, covering framework selection (LangGraph, LlamaIndex, AutoGen, CrewAI), access control patterns, and ReAct/Tree-of-Thoughts retrieval reasoning. |
| p10 | Agentic RAG: The 2026 Production Guide | MarsDevs | 2026-05 | Production-focused guide with quantified latency benchmarks (standard RAG 1–2 s; agentic loop 8–12 s; 3–10x token cost multiplier) and three-layer evaluation architecture using Ragas, Arize Phoenix, and Langfuse — the most numerically grounded cost/latency source in the sweep. |
| p11 | Next-Generation Agentic RAG with LangGraph (2026 Edition) | Medium (Vinod Rane) | 2026-03 | Detailed implementation guide for stateful agentic RAG using LangGraph directed cyclic graphs, with per-node RAGAS observability instrumentation (critic_score, retrieval_round, iteration_count, token_budget_used) and production metric targets. |
| p12 | RAG Framework Benchmark: LangChain vs LangGraph vs LlamaIndex vs Haystack vs DSPy | AIMultiple | 2026-01 | Standardised 100-query benchmark across five frameworks with identical models (GPT-4.1-mini) and retriever (Qdrant), isolating framework overhead and token efficiency: DSPy 3.53 ms overhead; Haystack 1.57k tokens vs LangChain 2.40k tokens — a 53% token difference that compounds at scale. |
| p13 | LangChain vs LlamaIndex (2026): Complete Production RAG Comparison | PremAI Blog | 2026-03 | Documents the architectural split between LangChain/LangGraph (workflow-first, stateful graphs) and LlamaIndex (retrieval-first, data-centric agents), noting LangGraph reached 1.0 stability in October 2025 and effectively superseded original chain-based LangChain for production agentic work. |
| p14 | Why LLM Frameworks Like LangChain and LlamaIndex Are Being Replaced by Agent SDKs | MindStudio | 2026-03 | Analyses the structural disruption of heavyweight RAG frameworks by native tool-calling, expanded context windows, MCP standardisation, and agent SDKs; includes LlamaIndex co-founder Jerry Liu's public acknowledgement that the framework era is ending. |
| p15 | LLM Frameworks Compared (2026): LangChain, LlamaIndex, DSPy and More | Morph | 2026-03 | Documents framework consolidation into four categories (orchestration, agents, optimisation, code-specific), reports LangChain at 100K+ GitHub stars and 34.5 million monthly LangGraph downloads, and warns that stacking three or more frameworks signals overengineering. |
| p16 | The 5 Best RAG Evaluation Tools You Should Know in 2026 | Maxim AI | 2026-02 | Comparative review of the five dominant evaluation platforms (Maxim AI, LangSmith, Arize Phoenix, Ragas, DeepEval), noting RAGAS exceeds 400,000 monthly downloads and 20 million evaluations, and that LangSmith's tight LangChain coupling creates friction in mixed-framework environments. |
| p17 | Top RAG Evaluation Tools in 2026 | Goodeye Labs | 2026-03 | Independent ranking of seven evaluation platforms including Weights & Biases Weave and Braintrust, with evidence that a 2025 study found Microsoft Copilot gave medically incorrect advice 26% of the time — illustrating the real-world stakes of inadequate RAG evaluation. |
| p18 | 7 RAG Evaluation Tools You Must Know | Iguazio | 2025-12 | Practitioner-oriented tool guide covering Ragas, LangSmith, Arize Phoenix, TruLens, and Promptfoo for continuous RAG evaluation in CI/CD pipelines, relevant to the maturing DevOps practices around agentic AI quality assurance. |
| p19 | Is RAG Dead? The Rise of Context Engineering and Semantic Layers for Agentic AI | Towards Data Science | 2025-10 | Documents the M&A consolidation around knowledge graph and semantic layer infrastructure (ServiceNow's acquisition of data.world, Salesforce's $8bn Informatica purchase), with Gartner's May 2025 recommendation that data engineering teams adopt ontologies and knowledge graphs to support AI. |
| p20 | From RAG to Context: A 2025 Year-End Review of RAG | RAGFlow | 2025-12 | Engineering team's year-end synthesis introducing 'Context Engineering' as the successor discipline to RAG optimisation, describing the shift from tuning single retrieval algorithms to systematic design of the end-to-end retrieval–context assembly–model reasoning pipeline. |
| p21 | Why 2025's Agentic AI Boom Is a CISO's Worst Nightmare | CSO Online | 2026-02 | Reports that 72–80% of enterprise RAG implementations significantly underperform or fail within their first year, and that 51% of all enterprise AI failures in 2025 were RAG-related; also identifies the '20,000-document cliff' latency and accuracy degradation pattern. |
| p22 | LLM Token Optimization: Cut Costs and Latency in 2026 | Redis | 2026-02 | Vendor-authored technical guide quantifying that semantic caching achieves up to 73% cost reduction in high-repetition agentic workloads, with benchmarks contrasting cache-hit millisecond response against seconds-scale fresh LLM inference. |
| p23 | Agentic RAG: When Static Retrieval Is No Longer Enough | Medium | 2026-03 | Cites MDPI Electronics 2025 study across 12 RAG variants and 250 clinical vignettes showing Self-RAG at 5.8% hallucination rate and a 55-percentage-point multi-hop accuracy gap between static RAG (34%) and agentic RAG (89%), providing the strongest quantitative case for the capability differential. |
| p24 | The Next Frontier of RAG: How Enterprise Knowledge Systems Will Evolve (2026–2030) | NStarX Inc. | 2025-12 | Engineering team's forward-looking analysis reframing RAG as a 'knowledge runtime' analogous to Kubernetes for application workloads, with governance, retrieval quality gates, and audit trails as mandatory infrastructure — the most explicit articulation of RAG-as-operational-infrastructure. |
| p25 | 10 RAG Architectures in 2026: Enterprise Use Cases and Strategy | Techment | 2026-03 | Practitioner decision framework for CTO/CDO selection across ten RAG architectures, explicitly stating that Agentic RAG is 'only necessary for complex, multi-step workflows' and that most enterprise search performs well with Hybrid RAG — the clearest published decision threshold for switching. |