The scaffolding layer that developers once needed to ship LLM applications β indexing layers, query engines, retrieval pipelines, carefully orchestrated agent loops β is collapsing. And according to Jerry Liu, co-founder and CEO of LlamaIndex, that's not a problem. It's the point.
βAs a result, there's less of a need for frameworks to actually help users compose these deterministic workflows in a light and shallow manner,β Jerry Liu, co-founder and CEO of LlamaIndex, explains in a new VentureBeat Beyond the Pilot podcast.Β
Context is becoming the moat
Liuβs LlamaIndex is one of the foremost retrieval-augmented generation (RAG) frameworks connecting private, custom, and domain-specific data to LLMs. But even he acknowledges that these types of frameworks are becoming less relevant.Β
With every new release, models demonstrate incremental capabilities to reason over βmassive amountsβ of unstructured data, and theyβre getting better at it than humans, he notes. They can be trusted to reason extensively, self-correct, and perform multi-step planning; Modern Context Protocol (MCP) and Claude Agent Skills plug-ins allow models to discover and use tools without requiring integrations for every one independently.Β
Agent patterns have consolidated toward what Liu calls a "managed agent diagram" β a harness layer combined with tools, MCP connectors, and skills plug-ins, rather than custom-built orchestration for every workflow.
Further, coding agents excel at writing code, meaning devs donβt need to rely on extensive libraries. In fact, about 95% of LlamaIndex code is generated by AI. βEngineers are not actually writing real code,β Liu said. βThey're all typing in natural language.β This means the layers between programmers and non-programmers is collapsing, because βthe new programming language is essentially English.βΒ
Instead of manual coding or struggling to understand API and document integration, devs can just point Claude Code at it. βThis type of stuff was either extremely inefficient or just would break the agent three years ago,β said Liu. βIt's just way easier for people to build even relatively advanced retrieval with extremely simple primitives.β
So whatβs the core differentiator when the stack collapses?Β
Context, Liu says. Agents need to be able to decipher file formats to extract the right information. Providing higher accuracy and cheaper parsing becomes key, and LlamaIndex is well-positioned here, he contends, because of its developments with agentic document processing via optical character recognition (OCR).Β
βWe've really identified that there's a core set of data that has been locked up in all these file format containers,β he said. Ultimately, βwhether you use OpenAI Codex or Claude Code doesn't really matter. The thing that they all need is context.β
Keeping stacks modular
Thereβs growing concern about builders like Anthropic locking in session data; in light of this, Liu emphasizes the importance of modularity and agnosticism. Builders shouldnβt bet on any one frontier model, or overbuild in a way that overcomplicates components of the stack.Β
Retrieval has evolved into βagent-plus-sandbox,β as he describes it, and enterprises must ensure that their code bases are tech debt free and adaptable to changing patterns. They also have to acknowledge that some parts of the stack will eventually need to be thrown away as a matter of course.Β
βBecause with every new model release, there's always a different model that is kind of the winner,β Liu said. βYou want to make sure you actually have some flexibility to take advantage of it.β
Listen to the podcast to hear more about:Β
-
LlamaIndexβs beginnings as a βtoy projectβ with initially only about 40% accuracy;Β
-
How SaaS companies can tap into complicated workflows that must be standardized and repeatable for average knowledge workers;
-
Why vertical AI companies are taking off and why βbuild versus buyβ is still a very valid question in the agent age.Β
You can also listen and subscribe to Beyond the Pilot on Spotify, Apple or wherever you get your podcasts.



