Back to all posts

Anthropic stopped covering OpenClaw in Claude subscriptions. Here's how AI pricing actually works.

Ian Zein
Anthropic stopped covering OpenClaw in Claude subscriptions. Here's how AI pricing actually works.

Anthropic stopped covering OpenClaw usage under Claude subscriptions. OpenClaw still works via the API, but you now pay per token instead of having it included in your flat monthly fee. A good moment to explain how AI pricing actually works, and why it matters for your organisation.

If you were using OpenClaw through your $20/month Claude Pro plan, that's no longer covered. You can still use it via the API, but depending on usage, your bill could jump from $20 to $500 a month.

This matters beyond OpenClaw. The way AI is priced determines who controls your AI strategy, what it costs at scale, and how vulnerable you are to changes you didn't see coming.

The gym membership model

AI subscriptions work like gym memberships. Most people pay monthly and barely use it. A few power users push the limits every day. The light users subsidise the heavy ones. That's the business model.

Anthropic charges $20/month for Claude Pro. Microsoft charges $30/user/month for Copilot (reportedly increasing to $60). OpenAI charges $20 for ChatGPT Plus and $200 for Pro. These are flat rates regardless of how much you actually use.

On top of that, Anthropic gives you better rates when you stay inside their ecosystem: Claude Desktop, Claude Web, Claude Code, Claude Cowork. So you're subsidised twice: once by low-usage subscribers, and once by Anthropic itself, who wants you using their apps instead of someone else's.

OpenClaw broke that model. It's an agent framework that runs hundreds of API calls per task. This kind of usage was already against Anthropic's terms of service since February 2026, but enforcement was loose until now. Boris Cherny, Head of Claude Code, announced the change on X, describing capacity as "a resource we manage thoughtfully." Anthropic gave about two weeks' notice and offered a one-time credit equal to each user's monthly subscription price (redeemable by April 17), up to 30% off pre-purchased Extra Usage bundles, and a full refund option. They pointed people toward Claude Cowork as an alternative. But the underlying message was clear: if your usage doesn't fit the subscription economics, the terms will change.

The per-seat problem

For individuals, subscriptions are simple. For organisations, it gets more complicated.

Larger organisations typically negotiate enterprise agreements: Claude for Enterprise, ChatGPT Enterprise, Microsoft 365 Copilot. These come with volume discounts, SSO, admin controls, and negotiated rates. But they're still per-seat. You're still paying for everyone, whether they use AI ten hours a day or ten minutes a week. And you're committing to one provider's ecosystem for the duration of the contract.

The reality in most organisations? Maybe 10 people use AI intensively. Another 20 use it occasionally. The rest barely touch it. Enterprise licensing softens the per-seat cost, but it doesn't solve the fundamental mismatch between flat-rate pricing and wildly uneven usage.

And you're locked in. If Anthropic has the best model today but Google or OpenAI leapfrogs them next quarter, you're still committed. Enterprise agreements typically run 12 months or longer.

What pay-per-token actually means

The alternative is API pricing: you pay per token. A token is roughly a word, though not exactly. It's the smallest unit an AI model processes. Short common words like "the" or "is" are one token. Longer or less common words get split into pieces: "organisation" might be two tokens, "pseudonymisation" might be four. A typical page of English text is around 500 tokens. Not a perfect measure, but close enough for cost estimates.

You pay separately for what goes in (your prompt) and what comes out (the AI's response). Output tokens always cost more because generating new text requires more compute than reading existing text.

Here's what the major providers charge for a selection of their models, per million tokens (pricing verified April 4, 2026). There are many more variants available, but these cover the flagship and budget options from each provider:

ProviderModelInputOutput
AnthropicClaude Opus 4.6$5$25
AnthropicClaude Sonnet 4.6$3$15
AnthropicClaude Haiku 4.5$1$5
OpenAIGPT-5.4$2.50$15
OpenAIGPT-5.3$1.75$14
OpenAIGPT-5.4 Nano$0.20$1.25
GoogleGemini 3.1 Pro$2$12
GoogleGemini 2.5 Flash$0.30$2.50

A few things jump out. First, the price range is enormous. The cheapest models cost literally 100x less than the most expensive ones. This matters because most questions don't need the most powerful model. Asking "summarise this email" doesn't require the same firepower as "analyse this contract for legal risk." Second, older models can be more expensive than newer ones. Anthropic's Opus 4.1 costs $15/$75 per million tokens. The newer Opus 4.6 costs $5/$25. Three times cheaper for a better model. Third, there are discounts layered on top. Prompt caching gives you a 90% discount on cached tokens. Batch processing halves your bill. These matter at scale. Verify these rates yourself at platform.claude.com/docs/en/about-claude/pricing, openai.com/api/pricing, and ai.google.dev/pricing.

The hidden complexity

Beyond basic token pricing, there are additional costs that catch people off guard. Extended thinking (where the model reasons step-by-step before responding) generates tokens billed at output rates. For agentic workloads, this is where costs really explode: the model "thinks" extensively before each action, and you pay full output pricing for all of it. OpenAI charges 2x input rates beyond 272K tokens; Anthropic's current models include the full 1M context window at standard pricing. If you require data to stay within a specific geographic region, most providers add a 10% surcharge. Some providers charge extra for specific features (Anthropic charges $0.01 per web search). Tool use adds extra tokens to every request. For an agent like OpenClaw that chains dozens of steps, reads documents, calls tools, and triggers extended thinking on each step, these costs compound quickly. That's why Anthropic's $20 subscription couldn't sustain it. Beyond cost, third-party agent frameworks also raise security questions: OpenClaw had several critical vulnerabilities disclosed in early 2026, including a remote code execution flaw that exposed over 40,000 instances.

Why this is actually an opportunity

Here's the part most people miss: per-token pricing, done right, can save organisations serious money compared to per-seat subscriptions.

Here's a rough calculation. 50 people, averaging 30 AI interactions per day, 22 working days per month. Assume each interaction uses about 1,000 input tokens and 500 output tokens (a short question and a paragraph-length answer). That's 33 million input tokens and 16.5 million output tokens per month. On Claude Opus 4.6 ($5/$25), that comes to about $580/month for all 50 people. For comparison: 50 seats of Microsoft Copilot at $30/user is $1,500/month today, or $3,000/month at the reported $60 rate. But most interactions don't need the flagship model. If you route 70% to Haiku ($1/$5), 25% to Sonnet ($3/$15), and only 5% to Opus ($5/$25), the total drops to under $200/month. The exact number depends on your usage patterns, but the contrast is clear: under $200 with smart routing versus $1,500+ in per-seat licensing. And you're not locked into one provider.

The bigger picture

This week's OpenClaw decision is a small example of a larger pattern. AI providers are figuring out their business models in real time. While Anthropic was restricting third-party subscription access, OpenAI was doing the opposite, offering free ChatGPT Pro to open-source maintainers and explicitly naming OpenClaw as eligible. Different providers, different strategies, shifting in real time. What's included today might not be tomorrow. The organisations that navigate this well are the ones that don't tie their AI strategy to one provider's pricing page. They separate the intelligence (the models) from the infrastructure (the platform that manages access, routing, cost, and rules).

This is exactly why we built Aimable. We use Anthropic's models ourselves and think Claude is excellent. This isn't about picking sides. It's about architecture. Any organisation that routes all AI through a single provider's subscription is exposed to exactly this kind of shift. A platform layer that sits between your teams and the models, handling routing, access, and cost management, gives you the flexibility to adapt when pricing changes. That's the approach we've taken, and we think it's worth considering regardless of which platform you choose.