Comprehension Lock-In — The Enterprise Trap Nobody Sees Coming

OpenAI leaked GPT-5.4 and everyone stared at the model. The real play is a context platform that makes Salesforce's lock-in look quaint.

35 min read

35 min read

Published 5 March 2026

Blog Image

The Leak That Mattered for the Wrong Reasons

Last week, OpenAI engineers committed internal code to a public GitHub repository — twice in five days — and accidentally confirmed that GPT-5.4 exists. The internet did what it always does: prediction markets spiked, speculation threads caught fire, and a thousand LinkedIn posts appeared with the words "generational leap" in them.

None of that matters.

What matters is the thing the model is a component of. Not the intelligence itself, but what OpenAI intends to do with it. Buried in the press release accompanying their $110 billion fundraise was a single phrase that should have set off alarm bells across every enterprise CTO's office: stateful runtime environment. That phrase, combined with the AWS partnership underpinning it, describes something far more consequential than a smarter chatbot. It describes the foundation for a new kind of enterprise platform — one that doesn't just store your data, but comprehends your organisation.

And if you run a commerce team, an agency, or any business that depends on institutional knowledge, you need to understand what's coming. Because the lock-in it creates makes Salesforce's data moat look like a magazine subscription.

The Filing Cabinet Problem Nobody Admits

Here's a question worth sitting with: where does your organisation's actual knowledge live?

I don't mean the documented kind. Documentation is a polite fiction — everyone knows the real knowledge lives somewhere else. It's in the senior developer who built the checkout flow three years ago and knows, because they were in the room, that the retry logic interacts with the rate limiter in a way that's never been written down. It's in the Slack thread from February where someone explained why the migration was paused. It's in the head of the account manager who left last quarter and took twenty client relationships' worth of context with them.

Right now, organisational knowledge is fragmented across a dozen filing cabinets: code in GitHub, architectural decisions in Confluence (if you're lucky), customer context in Salesforce, project status in Jira, and the reasoning behind all of it — the why — scattered across meeting transcripts nobody reads and chat threads that scrolled past months ago.

The information exists. It exists in abundance, actually. What's missing is the synthesis layer. And today, the synthesis layer is human brains. Human brains that are bandwidth-limited, context-switching-impaired, and increasingly likely to hand in their notice when a recruiter offers 15% more.

When a senior engineer leaves, every filing cabinet remains full. What vanishes is the person who knew which cabinets to open and how to connect the contents in a way that led to meaningful decisions. Anyone who's lived through that departure — and I've seen it happen dozens of times in 26 years of commerce — knows the damage isn't immediate. It's a slow bleed. Decisions get made with less context. Assumptions go unchallenged. The organisation starts confidently doing things based on how it used to work.

The Platform That Replaces the Room

Now imagine a system that does the synthesis for you. Not a search engine. Not a chatbot. A system that continuously ingests from every filing cabinet in your business, maintains a coherent model of your organisation's knowledge, and reasons about it at a depth no individual human can match.

That's the play. OpenAI's stateful runtime environment, built on a $100 billion expanded AWS partnership and backed by an $840 billion valuation, is designed to become exactly this. And when it works — if it works — the consequences reshape the entire enterprise software stack.

In this future, Jira isn't where project knowledge lives. It's a data source. The agent ingests its signal and integrates it with code changes, customer feedback, and strategic priorities to produce coherent understanding. Salesforce isn't the system of record for customer relationships — it's a filing cabinet the synthesis layer reads from. The intelligence, the value, the margin all move to the context platform.

Salesforce — fresh off record-breaking FY2026 earnings — is worth roughly $250 billion for owning customer data. ServiceNow is worth $200 billion for owning IT workflow data. The company that owns the synthesis layer across all enterprise data is worth considerably more than both combined. Not because data isn't valuable, but because data was never where the real value lived. The value was always in connecting the dots.

Four Bets That Have to Pay Off Simultaneously

This isn't a straightforward engineering project. It requires four separate technical bets to come good at the same time, and each one is individually a hard research problem.

Bet one: intelligence makes context multiplicative. Give a mediocre model a million tokens of organisational history and it'll pattern-match on surface-level similarity. It'll find a discussion that sounds related but was about a different service in a different context, and synthesise from it with perfect confidence. Long context paired with weak reasoning is actively harmful — enterprises will and should run away from it. A strong reasoning model changes the equation. Each increment of reasoning expands the scope of context the model can productively use, and the returns become non-linear. This is why every GPT-5.x point release is load-bearing for the broader strategy, even if benchmarks look incremental.

Bet two: memory that doesn't rot. Today's AI memory is a colleague who remembers your coffee order but forgets the substance of last week's conversation. What the context platform needs is institutional memory at a depth that's never existed in software. And organisational context isn't static — the decision that was correct six months ago may have been superseded. The architectural pattern recommended last quarter might have been abandoned after performance testing. Memory that preserves context without updating it is worse than no memory at all. It's institutional hallucination: the AI equivalent of the ten-year veteran who confidently explains how things work based on what they remember from last year.

Bet three: retrieval at a scale nobody benchmarks. This is the crux. When your agent has trillions of tokens of organisational history, retrieval-augmented generation (RAG) cannot solve the problem. RAG works for factual lookup. It breaks for enterprise-scale organisational context in specific, well-understood ways. It can't handle relational queries across time — finding the chain of decisions that led to a current vulnerability requires understanding temporal sequence and causation across months. It can't distinguish current context from context about systems that no longer exist. And all of this degrades as the corpus grows: more false positives, more near-miss retrievals, more confident synthesis from irrelevant context. The solution probably requires a hybrid architecture: structured indexing, hierarchical memory, temporal state tracking, and possibly state-space compression for long-horizon context. And here's the strategic kicker — retrieval quality at enterprise scale is invisible in current benchmarks. Nobody runs evaluations on "find 2,000 relevant tokens in 10 trillion when relevance is defined by causal chains across eight months." The company that solves this first has a lead competitors can't even assess from the outside.

Bet four: execution at the speed of trust. When an agent runs autonomously across hundreds of tasks for weeks, even a 5% per-task failure rate compounds into systemic risk. The target for sustained agentic workflows at this scale is closer to 99.5% accuracy across diverse tasks, including situations where organisational context is ambiguous, contradictory, or incomplete. Every capability reinforces the others: better retrieval means more relevant context, better intelligence means more careful reasoning, more coherent memory means context reflects reality. The compound either improves together or it all falls apart.

Why Commerce Teams Should Be Paying Attention, Not Later — Now

If you work in ecommerce — whether you're running a DTC brand, managing a Shopify Plus estate, or leading an agency — you might be thinking this is an enterprise AI problem that doesn't touch you yet. You'd be wrong.

Consider what organisational knowledge looks like inside a commerce operation. It's the merchandiser who knows that the spring campaign always underperforms when launched before the bank holiday because shipping cutoffs create a gap in the fulfilment window. It's the developer who remembers that the payments integration was built to handle a specific edge case with multi-currency orders that's documented nowhere except a Notion page three people have access to. It's the agency account director who understands that the client's real objection to a platform migration isn't technical — it's that their CFO had a disastrous ERP migration in 2019 and now vetoes anything that sounds similar.

All of that context currently lives in people's heads. When those people leave, the context evaporates. When the account director moves agencies, three years of client relationship nuance walks out the door. The filing cabinets — the CRM, the project management tool, the support tickets — remain full. But the synthesis is gone.

Now extrapolate this to the platform level. If OpenAI (or Anthropic, or Google) succeeds in building a context platform that captures and synthesises this kind of knowledge, the implications for commerce are staggering. An agency whose AI context layer understands not just each client's tech stack but the why behind every architectural decision, every campaign iteration, every seasonal pattern across a portfolio of accounts — that agency has a structural advantage that's nearly impossible to replicate.

But it also creates a dependency that should terrify founders. Because the moment your organisational understanding lives on someone else's platform, you've created a switching cost that dwarfs anything Shopify or Salesforce Commerce Cloud has ever imposed.

Comprehension Lock-In: Deeper Than Data, Impossible to Export

Let me be precise about what kind of lock-in we're discussing, because it's qualitatively different from anything enterprise software has produced before.

Salesforce's lock-in comes from data. Your customer records, your pipeline stages, your custom objects — it's painful to migrate, but it's technically portable. You can export it. Consultancies will charge you a fortune to move it, but the data is yours and it can be moved.

The context platform's lock-in comes from understanding. Not data points, but the synthesis of how data points relate to each other, how they've evolved, what they imply for current decisions. When an enterprise's organisational understanding lives on the context platform, switching to anything else means losing the layer that connects every other system in the stack. The agent that knows how Salesforce data relates to GitHub decisions relates to the board deck — that understanding can't be exported. It's not a CSV. It's not even a model's weights. It's the accumulated product of months or years of reasoning about your specific organisation.

Data is portable. Comprehension isn't.

Fast-forward to 2028. An enterprise has been running on the context platform for two years. New engineers onboard in days rather than weeks because the agent can contextualise their work against two years of institutional memory. Management decisions are informed by synthesis across every department. The switching cost isn't the subscription fee — it's the understanding. The months of accumulated synthesis, decision histories, cross-team connections, pattern recognition from hundreds of code reviews and incidents. All of that would disappear in a migration. The enterprise would go back to humans as the integration layer and reset from scratch.

There is no natural ceiling. The longer you stay, the deeper the understanding and the higher the switching cost. This is institutional capture at a depth enterprise software has never seen.

Anthropic's Accidental Head Start

Here's where it gets interesting — and where the conventional wisdom about OpenAI's dominance may be wrong.

While OpenAI is building its context platform architecturally from the top down, with AWS infrastructure and CIO-level enterprise contracts, Anthropic has stumbled into something potentially more valuable: bottom-up context accumulation through daily usage.

Claude Code has captured over half of the enterprise coding market. Every day, millions of developers are generating CLAUDE.md files, building workflow patterns, and creating organisational context as a byproduct of doing their actual jobs. That context isn't currently productised — it's not being harvested into an institutional memory layer — but it's immensely valuable. And Anthropic knows it.

The distinction matters. Context accumulated organically through daily usage may be more valuable than context captured architecturally from day one. The developer who's been using Claude Code for six months has built workflows deeply integrated into their actual process. A runtime capturing context from scratch captures context about workflows that haven't adapted to its existence yet.

If Anthropic can figure out how to productise this accumulated context in the next six to nine months — before OpenAI's stateful runtime ships — they have a genuine chance to establish the kind of comprehension lock-in I've described, but built bottom-up from real work rather than top-down from infrastructure.

OpenAI is betting on capital. Anthropic is betting on product-market fit. Capital buys infrastructure; it doesn't necessarily buy understanding.

What This Means for Commerce — and What to Do About It

I'm not going to pretend this is a 2026 problem with a 2026 solution. The four bets I've described — multiplicative intelligence, non-rotting memory, enterprise-scale retrieval, and sustained execution accuracy — likely need another 12 to 18 months of research progress before any vendor can deliver them together. But that doesn't mean you should wait.

Three things to think about from your chair, wherever you sit:

First: where is your organisational understanding actually accumulating? If your engineers are on Claude Code, your product team is on ChatGPT, and your analysts are on Gemini, you're building valuable context on each individual team but you're building zero common understanding. Think about this now. You don't need a trillion-token context layer to start — getting to a few million tokens of properly structured, well-tagged organisational knowledge with decent retrieval will accelerate your team's collective capability significantly. A structured Notion workspace with consistent taxonomy, an internal wiki with enforced templates, a deliberately curated repository of decision records — these are primitive context layers, but they're real ones.

Second: are you building a flywheel or just running experiments? There's a meaningful difference between letting people try AI tools and see what sticks versus intentionally building compound improvement. Is retrieval getting better? Is execution against AI-assisted projects getting more reliable? Are you evaluating what requires sustained context versus what's a point-use tool? If the answer to most of these is "I don't know," this is a conversation worth having — and not just in the C-suite. AI champions at every level should have a voice, because there is no leadership halo when it comes to assessing what's actually working on the ground.

Third: what are your understanding switching costs today? If you started building a primitive context layer internally and got it to the point where it captured even 20-30% of your organisation's working understanding, how painful would it be to switch? How portable is that context? If OpenAI or Anthropic offers you a beta in 12 months, at what point would you be willing to hand over your organisational understanding to a third party? And if you operate in a regulated industry where that's not an option, you need to invest more in your internal context layer now, because the gap between organisations that have one and those that don't is about to become the gap that determines competitive survival.

GPT-5.4 will drop when it drops. It'll be faster, smarter, and better at benchmarks. Commentary will call it a generational leap or a disappointment, depending on who's talking. None of that is the game.

GPT-5.4 will drop when it drops. It'll be faster, smarter, and better at benchmarks. Commentary will call it a generational leap or a disappointment, depending on who's talking. None of that is the game.

The game is the race to own enterprise comprehension. OpenAI is building from the top, with infrastructure muscle and an $840 billion valuation to justify. Anthropic is accumulating from the bottom, with product-market fit that OpenAI's Codex is chasing but hasn't matched. Google is somewhere in between, with data advantages that could matter if they figure out the product.

And the prize — the new system of record for organisational understanding — is worth more than Salesforce, ServiceNow, and SAP combined. Because those companies own data. The winner of this race owns the thing that makes data useful.

For commerce teams, the practical implication is simple: the organisations that start building their context advantage now — even primitive, even imperfect — will have a structural edge when these platforms arrive. The ones that wait for a vendor to solve it will discover that the vendor's solution comes with lock-in so deep it makes your current platform dependency feel like choosing between two coffee shops.

The pieces are on the board. The clock is running. Most of us are staring at the wrong chess piece. Don't be that person.

Explore Topics

Icon

0%

Explore Topics

Icon

0%