ThinkTech
Risk Entry6 min read

Vendor Concentration

When AI deployment depends on a small number of providers, organizations face pricing, availability, and policy risks that cannot be mitigated by technical controls alone.

By ThinkTech Research|Published May 12, 2026

You are evaluating or operating AI systems in production, and you need to understand vendor concentration: why it creates risk that technical architecture alone cannot address, how to measure it, and what organizational controls reduce exposure.

Definition

Vendor concentration in AI refers to organizational dependence on a small number of AI providers for capabilities that are critical to operations, products, or services. As of mid-2026, three providers (OpenAI, Anthropic, and Google) serve an estimated 90% of paid frontier AI inference by API revenue. The Herfindahl-Hirschman Index for this market is approximately 2,900 to 3,100, which the US Department of Justice classifies as highly concentrated.

Concentration risk is distinct from technical risk. An AI system may perform well, pass all safety evaluations, and meet every compliance requirement while still exposing the organization to unmanageable vendor dependency.

Documented risk patterns

Four risk patterns appear consistently in concentrated vendor relationships across technology markets. All four apply to AI inference:

Risk PatternDescriptionPrecedent
Pricing powerWhen alternatives are limited, the provider can raise prices with little competitive pressure. Switching costs increase over time as integrations deepen.Cloud computing pricing escalation after vendor lock-in (documented across AWS, Azure, and GCP enterprise accounts)
Availability dependencyIf the provider experiences an outage, every downstream product using that API is affected. No technical redundancy exists unless a secondary provider is actively maintained.AWS us-east-1 outages in 2021 and 2023 disrupted thousands of businesses simultaneously
Policy riskThe provider can change terms of service, acceptable use policies, content policies, or data handling practices unilaterally. Downstream organizations must comply or find an alternative.OpenAI has revised its usage policies multiple times since 2023, restricting and expanding permitted use cases
Data lock-inFine-tuned models, accumulated conversation history, custom system prompts, and evaluation pipelines all create switching costs that increase with time. Migrating between providers requires revalidation of every AI-dependent workflow.Enterprise SaaS vendor lock-in patterns (Salesforce, Oracle, SAP) apply in amplified form to AI because model behavior is not portable

Risk assessment by deployment type

The severity of vendor concentration risk varies by organizational context. Smaller organizations face higher relative impact because they have fewer resources to maintain alternatives:

Deployment TypeConcentration Risk LevelPrimary Concerns
Startup (single product on one AI API)CriticalA pricing change, policy change, or outage directly threatens the core product. No negotiating leverage. No redundancy budget.
Enterprise (AI integrated across business units)HighMultiple business units share the same vendor dependency. A single policy change affects legal, customer service, engineering, and product simultaneously. Procurement contracts may lock in terms for years.
Government (AI for public services)HighSovereign dependency on a foreign commercial provider. Policy changes by the provider can conflict with regulatory obligations. Outages affect public services with no fallback.
Critical infrastructure (healthcare, finance, energy)CriticalRegulatory requirements (NIST AI RMF, EU AI Act) may conflict with sole-source dependency. Availability requirements cannot be met with a single provider. Data residency requirements may limit provider options further.

Detection methods

Provider dependency ratio

Calculate the percentage of AI-dependent workflows that rely on a single provider. If more than 70% of AI functionality routes through one API, the organization is over-concentrated. Track this metric quarterly.

Pricing change impact modeling

Model the financial impact of a 2x and 5x price increase from your primary AI provider. If a 2x increase would force product changes or margin compression, the dependency is material. Run this analysis annually or when renewing contracts.

Switching cost estimation

Estimate the time and cost required to migrate your primary AI workloads to an alternative provider. Include: API integration changes, prompt rewriting and revalidation, fine-tuned model recreation, evaluation pipeline updates, and staff retraining. If the estimated migration exceeds three months, switching costs are high enough to constitute lock-in.

Outage impact assessment

Document what happens to your products and services if your primary AI provider is unavailable for four hours, 24 hours, and one week. If a four-hour outage degrades customer-facing functionality with no fallback, availability dependency is critical.

Mitigation strategies

Multi-provider architecture

Maintain active integrations with at least two AI providers for production workloads. Route traffic to both during normal operations (even if at different ratios) so that failover is tested continuously, not just theoretically. This adds cost. The cost is the price of resilience.

**Limitations:** Different providers' models behave differently. Prompts that work well on Claude may need rewriting for GPT-4o. Maintaining parity across providers requires ongoing evaluation effort.

Open-weight fallback models

Deploy an open-weight model (such as Llama 3.1 or Mistral Large) on your own infrastructure or a secondary cloud provider as a degraded-mode fallback. The fallback does not need to match frontier quality. It needs to keep critical functions operational during a primary provider outage or policy change.

**Limitations:** Self-hosting requires infrastructure expertise and ongoing maintenance. The fallback model must be tested regularly against production workloads.

Inference cost monitoring

Track per-unit AI costs monthly. Set alerts for cost increases that exceed 10% month-over-month. Pricing changes by providers are often announced with 30 to 90 days notice. Early detection allows time to evaluate alternatives.

**Limitations:** Some pricing changes are structural (new model generations at different price points) rather than direct increases on existing models. Monitor effective cost per task, not just per-token pricing.

Contractual protections

For enterprise and government deployments, negotiate pricing commitments, SLA guarantees, data handling commitments, and change notification periods into contracts. Standard API terms of service provide minimal protection. Custom agreements are necessary for critical deployments.

**Limitations:** Contractual protections are only as strong as the enforcement mechanism. A provider that changes its model architecture or discontinues a model line may technically comply with contract terms while fundamentally altering the service.

Abstraction layers

Build an internal abstraction layer between your application code and the AI provider API. This allows swapping providers without rewriting application logic. The abstraction should normalize request formats, response parsing, error handling, and cost tracking across providers.

**Limitations:** Abstraction adds latency, complexity, and a new failure mode. Provider-specific features (function calling formats, system prompt behavior, context window sizes) may not abstract cleanly.

Decision checklist

Assess whether your organization is over-concentrated on a single AI provider:

  • [ ] More than 70% of AI-dependent workflows use a single provider
  • [ ] No tested fallback exists for the primary AI provider
  • [ ] A provider outage of four hours or more would degrade customer-facing services
  • [ ] A 2x pricing increase would materially affect margins or product viability
  • [ ] Estimated migration time to an alternative provider exceeds three months
  • [ ] No contractual protections exist beyond standard API terms of service
  • [ ] Fine-tuned models or accumulated context would be lost in a provider switch
  • [ ] The provider's acceptable use policy has not been reviewed in the last six months

If three or more items are checked, the organization has material vendor concentration risk that should be addressed through the mitigation strategies above.

Key takeaways

  1. Vendor concentration in AI inference is a structural market condition, not a temporary phase. Three providers control approximately 90% of frontier inference revenue, with an HHI above 2,500 (highly concentrated by DOJ standards).
  2. Technical quality does not eliminate vendor risk. A well-performing AI system from a sole provider still exposes the organization to pricing, availability, policy, and lock-in risks.
  3. Open-weight models provide a partial mitigation path but require infrastructure investment and operational capability that many organizations lack.
  4. The regulatory gap is real. No current framework (EU AI Act, NIST AI RMF) explicitly addresses market concentration in AI inference as a governance concern.
  5. Organizations should treat AI vendor concentration the same way they treat cloud vendor concentration: as a procurement and resilience issue that requires active management, not passive acceptance.

Sources

  1. [1]
  2. [2]
  3. [3]
    EU AI ActLegal Source

Related