What is vendor lock-in (in AI)? Why it matters for your business

Two colleagues at an office table reviewing a printed contract, one pointing at a clause
TL;DR

Vendor lock-in is anything that makes leaving a provider expensive or impractical. With AI it accumulates across five vectors: prompts tuned to one model, fine-tunes that do not transfer, RAG knowledge bases tied to one embedding space, agent tool-calling formats, and contractual commitments. By month twelve a 30-staff services firm has typically built up 45,000 to 85,000 pounds of switching cost. The fix is to price it in before you sign.

Key takeaways

- Vendor lock-in is anything that makes switching providers expensive or impractical, and with AI it shows up in five distinct places rather than one. - The visible costs (multi-year commits, exit fees) are negotiable. The architectural costs (prompt tuning, fine-tunes, RAG embeddings, agent tool calls) are the ones that quietly compound. - A 12-month deployment for a 30-staff services firm typically carries 300 to 500 engineering hours of switching cost, or 45,000 to 85,000 pounds of direct labour. - Open-weight models (Llama, Mistral, DeepSeek, Qwen) reduce lock-in at the model layer but do not remove prompt or RAG lock-in, and they shift the operational burden onto your team. - Below 25,000 pounds a year of AI spend, accept some lock-in as the cost of moving fast. Above that, architect for partial portability from day one.

A 30-staff financial services firm I spoke with last quarter had spent fourteen months building an internal compliance assistant on Claude. The prompts were tuned for Claude’s behaviour. The retrieval layer was embedded with the recommended embedding model. The fine-tuned helper that summarised FCA-language documents was trained on Anthropic’s API. Then OpenAI shipped a feature the team genuinely needed, and the operations director asked the obvious question: what would it take to switch?

Her engineering lead came back with a number. Prompt rewrites, three to four weeks of senior time. Re-embedding the knowledge base into a different vector space, two weeks of compute plus re-evaluation. Rebuilding the fine-tune, twelve to eighteen thousand pounds. Rebuilding the evaluation harness, because the two models scored differently on the firm’s own benchmarks. Total: roughly 67,000 pounds and four months. The firm decided not to switch. The procurement decision they had made fourteen months earlier had committed them to an architecture, which is a bigger commitment than picking a vendor.

That is what vendor lock-in actually looks like in 2026.

What is vendor lock-in?

Vendor lock-in is anything that makes switching providers expensive or impractical. With AI it operates differently to traditional cloud lock-in. Cloud lock-in is mostly about visible infrastructure: compute, storage, networking, contracts. AI lock-in is mostly about decisions you have already made: how prompts are written, what data you have embedded, which agent framework you have chosen, which model your fine-tunes target. None of those show up on a bill until you try to leave.

Some lock-in is unavoidable in any meaningful AI deployment. The useful question is not whether you have lock-in. The useful question is which kind, how much, and whether you priced it before you signed. CloudZero’s 2025 analysis put it cleanly: AI lock-in is API-driven rather than infrastructure-driven, embedded inside product features rather than sitting as a standalone system, and that is what makes it harder to detect early.

Why it matters for your business

It matters because the cost of leaving compounds quietly across five vectors that owners rarely price up front. First, prompt engineering: prompts tuned for one model do not behave the same on another, and a 2025 academic study reported prompts dropping 68 percentage points in performance when moved between model families without re-tuning. Second, fine-tuning: fine-tunes are model-specific by design and do not transfer.

Third, RAG knowledge bases: the embedding space created by one vendor’s model is incompatible with another’s, so re-embedding a sizeable corpus is not optional if you switch. Fourth, agentic frameworks: each provider’s tool-calling format differs (OpenAI function calling, Anthropic tool_use, Google function declarations), and migration touches every agent. Fifth, contractual: multi-year commits, exit fees, and the cloud-credit nudge where Azure or AWS credits flow only to their hosted models.

The cloud-credit pattern is particularly subtle. SMEs with Microsoft Azure credits naturally gravitate to Azure OpenAI Service. Once embedded there, switching means losing credit value. Holland & Knight and Morgan Lewis have both flagged the same risk on the contractual side: exit mechanics in AI agreements are routinely thinner than in traditional outsourcing, leaving the customer stranded if pricing or roadmap changes.

Where you will meet it

You will meet vendor lock-in in three places that look different to the team and the same to the balance sheet. The first is the procurement conversation, where the lock-in is largely invisible. The vendor demos a working tool, the team gets excited, and nobody asks the migration questions because the system is not yet live. The architectural commitment is made here. It is also the cheapest moment to price the lock-in honestly.

The second is the eighteen-month review. By that point the prompt library has been tuned, the RAG corpus is embedded, agents are wired to one provider’s tool-calling format, and the team has built genuine expertise in one model’s quirks. A cheaper or better alternative appears, the team estimates the migration, and the answer is uncomfortable. Cost-savings of a few hundred pounds a month do not pay back forty thousand pounds of engineering effort. The decision is made for you.

The third is the renewal. The contract is up, the vendor proposes a price increase or a new minimum commit, and your bargaining position is gone. Without a credible alternative inside the firm’s stack, the negotiation is short. This is where the contractual layer (multi-year auto-renewals, 90-day notice windows, remaining-balance termination fees) does its real work. RMOK Legal has documented several mid-market firms locked into a third year of pricing they would never sign today, simply because the renewal clock ticked past their notice window.

When to care about it, when to ignore it

Care about it whenever your AI spend crosses 25,000 pounds a year, or when the deployment is touching a regulated process or a customer-facing product. At that scale the switching cost is real money and the architectural choices have compounded enough to matter. Architect for partial portability from day one: an abstraction layer between your code and the vendor API, an open-weight option as a credible alternative, a re-embeddable RAG design, and a tool-calling abstraction.

Ignore it when the deployment is small, the spend is under a few hundred pounds a month, and the system can be rebuilt in a fortnight if the vendor disappears. A pilot, a personal-productivity tool, a small internal helper. Trying to architect those for portability slows the team down and costs more than the lock-in you are avoiding. The honest answer for a 5,000-pound-a-year deployment is to accept the lock-in, ship the value, and revisit at year three.

The middle ground is where many firms sit, and it is where the procurement questions earn their keep. Before you sign anything material, ask the vendor what migration tools and documentation they provide, what their model-deprecation policy is, whether they support open-weight alternatives on their infrastructure, and what the exit terms actually look like. Vendors who answer cleanly are signalling commercial maturity. Vendors who deflect are telling you that lock-in is part of how the product works.

Total cost of ownership, or TCO, is where vendor lock-in shows up on the spreadsheet. Switching cost is a TCO line, even if your finance system has no place to record it. A clean TCO model for an AI deployment includes a switching-cost estimate alongside running cost, integration cost, and ongoing tuning. Without that line, the procurement conversation is structurally optimistic.

Hybrid pricing interacts with lock-in through multi-year commits and minimum spend tiers. The vendor offers a discount in exchange for predictability, and the predictability is yours to give. A three-year commit at 25 percent discount looks attractive in month one and looks expensive in month fourteen when a better model ships. Negotiate the exit terms before you accept the discount.

Open-weight models (Llama, Mistral, DeepSeek, Qwen) are the structural hedge. Because the weights are public, you can host the same model on AWS Bedrock, on Together, on Replicate, or on your own infrastructure. That removes lock-in at the model layer. It does not remove prompt-engineering or RAG lock-in, and it shifts operational burden onto your team. For many SMEs the right pattern is a tiered approach: a proprietary API for fast-moving small workloads, an open-weight fallback for high-volume or data-sensitive ones.

The Model Context Protocol (MCP) is the emerging standard for connecting tools and data to language models. As an open protocol it is designed to reduce lock-in. In practice each vendor’s implementation has its own quirks, so optimising hard for one provider’s MCP behaviour can still create stickiness. Treat MCP as a hedge worth having, not a guarantee of portability.

The honest test of any AI procurement decision is the migration question. If you cannot answer it cleanly today, you have not yet priced the lock-in. The work is to price it before you sign, not after.

Sources

CloudZero (2025). AI vendor lock-in: what it is and how to avoid it. Plain-English overview of the technical, data, and contractual layers of AI lock-in. https://www.cloudzero.com/blog/ai-vendor-lock-in/ TechTarget (2024). Best practices to avoid AI vendor lock-in. Cited for the API-portability and abstraction-layer recommendations in the body. https://www.techtarget.com/searchenterpriseai/tip/Best-practices-to-avoid-AI-vendor-lock-in Particle41 (2024). What CTOs actually worry about with AI vendor lock-in. Source for the prompt-engineering and institutional-knowledge lock-in framing. https://particle41.com/insights/ctos-worried-about-ai-vendor-lock-in/ Featured.com (2024). Hidden costs of migrating an LLM ecosystem. Cited for the named migration-cost categories (prompt rework, re-embedding, security, integration, knowledge loss). https://featured.com/questions/hidden-costs-migrating-llm-ecosystem itnext (2024). Vendor lock-in in the embedding layer: a migration story. Source for the RAG embedding-space incompatibility point. https://itnext.io/vendor-lock-in-in-the-embedding-layer-a-migration-story-183ea58e3668 Morgan Lewis (2026). Negotiating AI provisions in commercial and technology contracts. Cited for AI-specific exit and data-portability contract terms. https://www.morganlewis.com/blogs/sourcingatmorganlewis/2026/04/negotiating-ai-provisions-in-commercial-and-technology-contracts-where-the-market-is-heading RMOK Legal (2026). AI outsourcing contracts: what to fix before August 2026. Cited for EU AI Act compliance allocation and exit-mechanic risks. https://www.rmoklegal.com/news/ai-outsourcing-contracts-what-to-fix-before-august-2026 Holland & Knight (2026). US companies face the EU AI Act's August 2026 compliance deadline. Cited for the regulatory compliance lock-in dimension. https://www.hklaw.com/en/insights/publications/2026/04/us-companies-face-eu-ai-acts-possible-august-2026-compliance-deadline Anthropic (2026). Model Context Protocol connector documentation. Cited for the MCP integration lock-in section. https://platform.claude.com/docs/en/agents-and-tools/mcp-connector Understanding AI (2026). The best Chinese open-weight models. Source for the 2026 open-weight landscape (Qwen, DeepSeek, Llama 4, Mistral). https://www.understandingai.org/p/the-best-chinese-open-weight-models

Frequently asked questions

How do I know how much vendor lock-in I have already built up?

Walk through five questions. How many prompts have been tuned to your current model and how long would it take to re-tune them? Have you fine-tuned anything? How big is your RAG corpus and which embedding model is it on? Are your agents using one provider's tool-calling format? What does your contract say about early termination? The rough sum of those answers is your switching cost. For a 12-month deployment in a 30-staff firm it usually lands between 45,000 and 85,000 pounds.

Are open-weight models like Llama and Mistral the answer to lock-in?

They help, they do not solve it. Open weights remove lock-in at the model layer because you can host the same model with several providers or yourself. They do not remove the prompt-engineering, RAG-embedding, or tool-calling work you have done on top. They also shift operational burden onto your team. For many SMEs they make sense as a hedge for high-volume or sensitive workloads, not as a default.

What should I ask a vendor before I sign?

Five things. What format will my data, prompts, fine-tunes, and embeddings come out in if I leave? Do you support open-weight models on your infrastructure? What is the early termination fee, and is it flat or remaining-balance? What is your model deprecation policy and how much migration help do you fund? Can I bring my own observability stack? Vendors who answer these cleanly are signalling maturity. Vendors who deflect are telling you lock-in is part of their commercial model.

This post is general information and education only, not legal, regulatory, financial, or other professional advice. Regulations evolve, fee benchmarks shift, and every situation is different, so please take qualified professional advice before acting on anything you read here. See the Terms of Use for the full position.

Ready to talk it through?

Book a free 30 minute conversation. No pitch, no pressure, just a useful chat about where AI fits in your business.

Book a conversation

Related reading

If any of this sounds familiar, let's talk.

The next step is a conversation. No pitch, no pressure. Just an honest discussion about where you are and whether I can help.

Book a conversation