Blog - K2view

It’s not the LLM, it’s the data architecture

Written by Iris Zarecki | January 11, 2026
Why operational GenAI changes data architecture requirements, and why architecture, not models, determines trust, latency, cost, and adoption.


Operational GenAI is breaking old assumptions

When GenAI apps have difficulty in production, the instinct is often to look at models, prompts, or fine-tuning. In practice, however, the root cause is almost always the data architecture.

This post reframes the problem. Operational GenAI introduces new requirements on enterprise data architectures, and that is why architecture – not models – ultimately determines trust, latency, cost, and adoption in production GenAI use cases.

The following diagram illustrates a fundamental shift introduced by operational GenAI.

On the left is the old predictable world. A user asks a known question. The system knows which backend holds the answer, which API to invoke, which query to run, and what shape the response will take. The data flow is linear and deterministic. Architecture can be designed ahead of time because the interaction pattern is fixed.

On the right is a world of unpredictable interactions.

Today, users expect to interact with systems through natural language. Questions are no longer constrained by forms, screens, or predefined workflows. They are open-ended, conversational, and impossible to enumerate in advance.

While an initial question may appear simple, small variations in wording often represent very different requests. Each follow-up introduces new conditions, comparisons, or what-if scenarios that require additional data, more systems to access, and more complex reasoning.

Although the unpredictable questions may sound similar, they aren’t from an architectural perspective. One may require a single system call. Another may require assembling context across loans, payments, income changes, policies, and eligibility rules in real time.

What changed is not just the interface, but the nature of data access.

Operational GenAI does not execute a predefined workflow. It reasons dynamically, deciding at runtime which data is relevant, which systems must be accessed, and how context should be assembled based on the conversation, the user, and the current state of the business. The architecture can no longer assume it knows the question, the data path, or the logic upfront.

What architectural requirements have changed?

Traditional enterprise data architectures were designed for a world where access patterns were predictable.

In that world, architecture could assume that questions were known in advance, data paths could be designed ahead of time, and performance could be optimized for fixed workloads. APIs exposed specific functions. Data platforms centralized information for analysis. Governance was applied at ingestion or through role-based access.

Those assumptions no longer hold for operational GenAI.

Operational GenAI introduces a new set of architectural requirements that existing patterns were not designed to encounter.

Data access must be dynamic rather than predefined. The architecture can no longer rely on fixed APIs or curated queries because questions evolve during the interaction. GenAI must discover and assemble the right context at runtime rather than follow a predefined integration path.

Latency becomes a hard constraint. In analytics, seconds or even minutes are acceptable. In operational GenAI, responses must support live conversations and automated decisions. Architectures optimized for batch processing or broad scans struggle to meet these expectations.

Freshness is no longer optional. Operational GenAI reasons about the current state of a customer, an account, or a transaction. Data that is minutes or hours old can lead to incorrect or misleading answers, even if it is technically accurate.

Governance must be enforced throughout every interaction. Access controls, privacy rules, and policy constraints cannot be applied only upstream. They must be evaluated dynamically for each request, based on who is asking, what they are asking, and why.

Why must the unit of access change?

Traditional architectures expose systems, tables, or endpoints. Data is accessed where it lives, shaped by how applications and databases were designed. Any higher-level business context is reconstructed by the consuming application through joins, orchestration logic, or custom code.

Operational GenAI cannot work this way. It does not ask for a table or an API response. It asks for business context.

That context is entity centric. A customer, an account, a loan, or a policy spans multiple systems, includes relationships and recent activity, and must reflect the current state of the business whenever a question is asked.

From an architectural perspective, this means GenAI requires a consistent, unified view of entities that is assembled dynamically at runtime. The same entity must look the same regardless of which system the data originated from, which question is being asked, or which reasoning path is taken.

Without entity-level access, every GenAI interaction becomes an exercise in manual reconstruction. Each new question requires stitching together partial views, reconciling inconsistencies, and reapplying rules. That increases latency, cost, and risk, and makes consistent, trustworthy answers difficult to achieve at scale.

This shift in the unit of access is subtle, but it’s one of the most fundamental reasons why traditional data architectures struggle with operational GenAI.

These requirements are not incremental improvements. They represent a fundamental shift in how data must be accessed, assembled, and governed. This is why architectures built for APIs, lakes, and warehouses struggle as soon as GenAI moves from pilots to production.

Why it’s not the LLM, but the data architecture

By the time GenAI reaches production, model choice is rarely the limiting factor. Modern LLMs are already capable of reasoning, summarizing, and generating high-quality responses.

What determines success or failure is whether the data architecture can support those capabilities in the real world.

Operational GenAI places new demands on data access. Questions are open-ended. Context must be assembled dynamically. Data must be fresh, consistent, and governed whenever it’s used. Latency must support live interactions. Answers must be trustworthy enough to influence real decisions.

Traditional architectures weren’t designed for this. They assume predictable access patterns, predefined integrations, and data prepared in advance. Those assumptions break as soon as GenAI becomes operational.

That’s why GenAI is often unsuccessful in production environments, even as models continue to improve. The gap is architectural.

Until data architectures evolve to serve real-time, entity-level, governed context on demand, GenAI will remain impressive in demos and fragile in production.

Discover how AI-ready data powers
operational GenAI data architectures.