AI Coding Tool Pricing: Type A vs Type B (2026)

By VibeDex ResearchOriginally published: April 20, 2026Updated: 20 April 2026

TL;DR

Who this is for: buyers of AI coding tools who need to predict whether bills scale linearly or compound structurally.Bolt is the only tested AI coding tool with Type A structural pricing pathology[4] — its WebContainer has no persistent index, so the platform re-reads the entire codebase every turn. A 20-component project burns around 100k tokens per minor edit. Every other platform in the set is Type B (usage-metered, persistent-index, caching-aware). The distinction matters because Type A cannot be fixed by smaller diffs or prompt engineering; Type B can. For predictable spend in 2026, the answers are narrow: Claude Code flat caps, Aider[6] bring-your-own-API, or Cline[7] per-file token budgeting (v0.8, Mar 2026). Everything else is a bet on the vendor not repricing you next quarter.

The Type A / Type B Split: Why the Split Matters

We split AI coding tool pricing into two types so buyers can predict cost shape. Scorers must cite which type of cost behaviour applies to each platform. The old anchor averaged two very different cost pathologies into one score. The new anchor does not.

Type A — Structural token-cost pathology

Cost compounds with project size regardless of edit scope. Caused by architecture, not plan choice. Cannot be fixed with prompt engineering. In this benchmark, Bolt is the only cleanly Type A platform. WebContainer has no persistent index, so every turn re-reads the codebase into context.

Anchor: score 1 on Pricing regardless of headline plan price. A 20-component project = ~100k tokens per minor-edit prompt. One documented case drained a Pro 100 plan (55M tokens) in 8 days.

Type B — Standard usage pricing

Per-token or per-credit with persistent-index caching. Small edits cost small amounts. Scales with actual work done. Includes Cursor[1], Claude Code, Lovable[2], Replit[3], Manus[5], Aider, Cline and Windsurf.

Type B can still produce big bills — Cursor, Replit and Manus all show this. But Type B pathologies come from pricing model choices (credit granularization, tokenizer changes, quota caps) that vendors can fix. Type A is architectural.

The practical implication: if your iteration workflow scales with project size rather than edit size, a Type A platform will bankrupt you at the same rate it would have at month one. A Type B platform at least lets you respond — smaller contexts, tighter diffs, better caching, flat-cap tiers.

Cost Reality: Marketing Price vs What Users Actually Pay

Headline subscription prices hide the consumption layer. Every tool in this table advertises a $20–50 entry tier; real-world invoices tell a different story. The right-most column applies the pricing-type classification.

PlatformMarketing priceReal-world cost patternPricing Type
Cursor$20 Pro / $200 UltraCredit-metered tier structure dates to the June 2025 restructure (CEO apology July 2025); headline plan prices stable since. Cursor 3.0 (April 2026) is the current baselineType B

Persistent-index caching; scales with edit size

Claude Code$20 Pro / $100 Max / $200 Max-20xFlat caps, no $ overage — but weekly limits hit in 19 minutes on Max after the 12 Feb 2026 tokenizer change (80x more API requests, 64x more output tokens for the same outcomes)Type B

Flat-cap; most predictable $-spend in the set

Lovable$25 Pro (100 credits) / $50 Business~73 credits in two weeks on a medium app with auth + DB; debug-loop credit-burn is the #1 community complaint. Current credit structure dates to the July 2025 granularization (9 months old; since then Lovable has rolled out Opus 4.5 in Dec 2025, reducing retry loops ~20% per their changelog)Type B

Credit-metered; structure stable since Jul 2025

Bolt$25 Pro (1M tokens) / $50 Pro+Pro 100 plan (55M tokens) drained in 8 days documented; ~100k tokens per minor-edit prompt on a 20-component project (WebContainer re-reads the entire codebase every turn)Type A

Structural — cost scales with project size regardless of edit scope

Replit$20–35 Core / TeamsEffort-based pricing (introduced Jun 2025, reinforced with Agent 3 Sept 2025) is the current model. The Register reported $1,000-in-one-week bills and $100–250/mo accounts jumping to $800–1,500/mo; community reports of $70/night Agent 3 burn persist. Agent 4 launched 11 Mar 2026 on the same structural model. Cost dynamic is live as of April 2026Type B

Effort-based consolidation — high volatility

Manus$20 Standard (4K credits) / $40 Pro / $200 Extended$39 lost in 10 minutes; $440 unauthorised charge; $5,000/mo over 3 months with ~40% output usable (Trustpilot)Type B

Credit-metered; mid-execution upsell friction

AiderFree (MIT) — pays API of choice ($22–50/mo typical solo dev)4.2x fewer tokens than Claude Code on 47-file refactor per Morphllm benchmark (71% first-pass success vs 78%)Type B

Bring-your-own-API; best price/control in the set

ClineFree (OSS) — pays API of choicev0.8 added per-file token budgeting (Mar 2026) in response to Feb billing complaints; file-edit retry loops doubled API costs for some usersType B

Bring-your-own-API; budgeting granularity added 2026

Windsurf$20 Pro (up from $15 Mar 2026)March 2026 credits-to-quota restructure + $5/mo price hike eliminated prior cost-advantage moat; quota-refresh caps block burst usageType B

Quota-refresh caps replacing credit pools

Real-world figures triangulated across platform pricing pages, The Register coverage of the Replit effort-based pricing complaints, Trustpilot Manus cases, and community-reported HN / Reddit threads during the Cursor June 2025 and Lovable July 2025 repricing events.

Four Pricing Failure Patterns (Taxonomy)

Beyond the Type A / Type B split, the 2025–2026 AI coding market produced four distinct usage-pricing failure modes. Each has at least one dated incident behind it. Bolt is cleanly structural (#0 in the list below); the other four patterns all fall inside Type B but produce the same user outcome — a bill several multiples of what the marketing page implied.

0. Structural per-turn re-cost (Type A) — Bolt

Incident: Pro 100 plan (55M tokens) drained in 8 days on a 20-component project; ~100k tokens per minor-edit prompt documented. Mechanism: WebContainer runtime has no persistent codebase index, so every agent turn re-reads the full project into context. This is not a bug — it is the design constraint of running Node inside the browser. Scoring-relevant because no combination of prompt engineering, smaller diffs, or plan upgrade materially improves the economics.

1. Token-redaction inflation — Claude Code, 12 February 2026

Incident: the redact-thinking-2026-02-12 rollout changed internal token accounting: post-change, users reported 80x more API requests and 64x more output tokens for functionally identical outcomes. Some Max-tier users hit weekly limits in 19 minutes. Mechanism: platform reduces internal efficiency invisibly; headline plan price unchanged but effective throughput drops. Detection difficulty: high — users only notice when the cap bites. Visible in GitHub issue #24727 and in the $53.65 overage incident reported there.

2. Credit granularization as stealth hike — Windsurf Mar 2026 (fresh); Lovable Jul 2025 (dated)

Current pattern: Lovable's current credit weighting[2](structure introduced July 2025, now 9 months old and simply the current model) weights previously single-credit operations as N-credit-weighted work; the stealth-hike framing only applies if compared to the pre-July 2025 structure. Windsurf's March 2026 restructure (fresh) eliminated credit pools entirely in favour of quota-refresh caps, concurrent with a $5/mo headline hike (Pro $15 to $20). Mechanism: what was “1 operation” becomes a variable-weight cost; total plan allocation appears unchanged but real work per dollar falls. Detection difficulty: high — requires before/after measurement on the same workflow.

3. Effort-based consolidation (current structural model) — Replit

Current state: Replit's Standard / Teams plans consolidate many small events into larger “effort” units — the structural model introduced June 2025 and reinforced with Agent 3 in September 2025, now simply the way Replit prices Agent usage. Dated proof point (10 months old, weight accordingly): The Register, September 2025 — “In the last week alone it charged me $1K, whereas before it was never more than $180–200/mo”[8]. Multiple $100–250/mo accounts jumped to $800–1,500/mo overnight at launch. Since then, $70/night Agent 3 burn reports have continued in community forums; Agent 4 launched 11 March 2026 on the same effort-unit structural model, and no public pricing-model revision has been located in the 20 January–20 April 2026 research window. Mechanism: pricing consolidates many small events into larger “effort” units, amplifying variance rather than dampening it. The structure remains load-bearing for any buyer picking Replit today — budget cap compute per project.

4. Quota-refresh caps replacing credit pools — Windsurf Mar 2026

Incident: March 2026 restructure replaced consumable credit pools with quota-refresh caps (fixed-rate refill) — blocks burst usage patterns that previously absorbed rollover. Mechanism: protects vendor unit economics on heavy days but punishes irregular workloads (weekend-warriors, pre-demo crunches). Concurrent $5/mo price hike eliminated the cost-advantage moat that had differentiated Windsurf in 2025. Community reception notably negative.

Why Bolt Is Uniquely Type A (and Cursor Isn't)

Every “Bolt is expensive” take in 2025 missed the architectural reason. The token problem is not about Bolt's plan tiering — it is about the WebContainer runtime itself. WebContainer is Node running inside the browser via WebAssembly, with no persistent server-side index of project state. When the agent takes a turn, the planner needs context about the full project to reason about what to edit. Without a persistent index, that context has to be reconstructed from the file tree each turn.

Cursor, Claude Code, Aider, Cline — all maintain persistent indices. Edit a single component and only that component's context (plus a small dependency graph) is included. The cost scales with the edit, not the project. The June 2025 Cursor repricing (a June 2025 incident that exhausted $7,000 annual subscriptions in a day) remains a useful illustration: Cursor users were hitting unit-economics pathology — which Cursor adjusted via plan restructure — not structural pathology. That is the difference between something a vendor can fix and something a vendor cannot.

The diagnostic: if your per-prompt token cost is roughly constant across a project's lifecycle, you're on Type B. If your per-prompt cost tracks project size — if editing one line in a 200-file project costs 10x what it cost in a 20-file project — you're on Type A. In our scoring, Bolt's Pricing score is pinned at 1 regardless of plan choice, and its Rate Limits marker is downgraded to 2 based on the 55M-tokens-in-8-days pattern.

Framework note: earlier versions of the Vibedex anchor did not distinguish architectural from economic pricing pathology. Scorers must cite which applies, because the advice to a buyer is fundamentally different: mitigate Type B via prompt engineering and flat-cap tiers; avoid Type A entirely if you expect the project to grow.

The Single Question Every Buyer Should Ask

“Has this vendor changed their billing model in the past 12 months?”

If yes — assume another change is coming. Recent pricing history in the category in recent history: Cursor repriced June 2025. Lovable granularized credits July 2025. Replit launched effort-based pricing June 2025 then Agent 3 September 2025. Windsurf restructured March 2026 (fresh). Claude Code's tokenizer change February 2026 (fresh) behaved like a price hike even though the plan prices didn't move. Five repricings across eight vendors in under 12 months is the base rate. Only Claude Code (flat caps) and Aider (bring-your-own-API) are structurally insulated from this pattern.

The base rate for pricing change inside 12 months in this category is roughly 1. Treat any paid-tier commitment as a 90-day decision, not an annual one. Per Vibedex our scoring, rising-hard platforms (Claude Code, Cursor, Lovable) carry a 60-day expiry — explicitly because these vendors ship pricing changes faster than quarterly benchmarks can track.

Best Pricing Transparency: Three Picks

If your constraint is bill predictability rather than absolute capability, three tools stand out in the set. All three win by removing the vendor's degree of freedom to reprice you invisibly.

Aider — bring-your-own-API, $22–50/mo typical solo spend[6]

MIT-licensed CLI. You pay Anthropic, OpenAI, or whichever provider you choose directly — Aider has no subscription layer between you and the model bill. Morphllm benchmark: 4.2x fewer tokens than Claude Code on 47-file refactor with 71% first-pass success (vs 78% for Claude Code). Zero CVEs in the 2025–2026 window. Paul Gauthier's engagement is GitHub-first, not marketing-first. Strongest overall pick for a cost-sensitive working engineer.

Claude Code — flat caps, $20 / $100 / $200

No dollar overage, ever. You hit weekly usage caps and work stops until refresh. The Feb 2026 tokenizer change compressed effective throughput inside those caps, but the dollar line on the invoice stayed flat. For anyone who needs finance-predictable tooling (solo founders, small teams without procurement cover), flat caps beat every variable-consumption model in this article. Excluded from Vibedex segment rankings per self-bias policy; included here because the pricing signal is buyer-relevant regardless.

Cline — OSS + per-file token budgeting (v0.8, Mar 2026)[7]

GitHub Octoverse 2025 named Cline the fastest-growing AI OSS coding project. v0.8 (March 2026) added per-file token budgeting — explicit response to February 2026 user complaints about file-edit retry loops doubling API costs. Free to install; you pay the model provider directly. Caveats: the Clinejection prompt-injection vulnerability (Dec 2025–Feb 2026) and the unauthorised @cline/cli npm publish are real security signals — Aider does not carry the same concerns.

Decision Framework: Project-Scope vs Edit-Scope

The single most useful question for picking an AI coding tool on price is: do your iterations scale with project size, or with edit size?

Edit-scope iterations

You mostly edit a file or two per turn. Refactors are occasional. Project size grows slowly. Any Type B tool works — pick on quality, not pricing type. Cursor + Claude Code is the $40/mo meta for working engineers; Aider is the cost-efficient alternative.

Project-scope iterations

You rely on the agent to reason across many files per turn. Refactors are the daily workflow. Project grows fast. A Type A tool will bankrupt you. Pick Type B with persistent indexing — or better, pick flat-cap Claude Code so the maximum monthly spend is bounded by plan, not by project size.

App-builder platforms (Lovable, Replit, Manus, Base44) conflate these because the agent ranges across the whole project by design. Any app-builder with a Type A runtime is acutely dangerous; any app-builder with volatile Type B pricing (Replit) is moderately dangerous. A deliberate Lovable user rotating out after V1 ships is fine; a founder who keeps iterating inside Lovable / Replit / Bolt for 3–6 months will hit one of the four failure patterns above.

Related Reading

For the full 10-platform benchmark with persona-specific winners see Best AI Coding Tool 2026: 10 Platforms Ranked by Persona.

For non-technical founders specifically, including the Lovable CVE-2025-48757 caveat, see Best AI Coding Tool for Non-Technical Founders (2026).

Sources & References

All external sources were verified as of April 2026. Ratings and metrics reflect the most recent data available at time of review.

  1. Cursor — Pricing (official)(cursor.com)
  2. Lovable — Pricing & Plans (official)(lovable.dev)
  3. Replit — Pricing (official)(replit.com)
  4. Bolt.new (official)(bolt.new)
  5. Manus (official)(manus.im)
  6. Aider — AI pair programming in your terminal(aider.chat)
  7. Cline — Autonomous coding agent (official)(cline.bot)
  8. The Register — Replit customers infuriated by surprise cost overruns (Sept 2025)(theregister.com)
  9. Answer.AI — Thoughts on a Month of Devin (Husain, Flath, Whitaker, 8 Jan 2025)(answer.ai)
  10. Manus blog — Manus joins Meta(manus.im)
  11. Mindgard — Manus Rubra Full Browser Remote Control (Rich Smith, 1 Dec 2025)(mindgard.ai)
  12. TechCrunch — Lovable raises $330M at $6.6B valuation (Dec 2025)(techcrunch.com)

Related Vibedex Benchmarks

Methodology: Rankings and scores in this article are based on VibeDex's independent benchmarks. Models are evaluated by AI-powered judges across multiple quality dimensions with scores weighted by prompt intent. See our full methodology

FAQ

Why does Bolt cost more at scale than Cursor?

Bolt's WebContainer runtime has no persistent codebase index — every turn the platform re-reads the entire project into context. On a 20-component app that's roughly 100k tokens per minor-edit prompt. Cursor uses persistent indexing with caching, so a small edit costs tokens proportional to the edit, not the project. One documented Bolt user drained a 55M-token Pro 100 plan in 8 days. Bolt is Type A (structural pathology); Cursor is Type B (standard per-token). The difference is architectural, not a plan-tier choice.

What is Type A vs Type B pricing?

Type A is structural token-cost pathology: platform cost compounds with project size regardless of edit scope. Bolt is the only tool in this set that's cleanly Type A — its WebContainer has no persistent index, so every turn re-reads the whole codebase. Type B is standard per-token or per-credit pricing where small edits cost small amounts (Cursor, Claude Code, Lovable, Replit, Manus, Aider, Cline, Windsurf). Both types can produce big bills, but only Type A is structurally unfixable via prompt engineering. A Type B platform can at least be tamed by smaller context, smaller diffs, better caching. A Type A platform cannot.

Which AI coding tool has the most predictable pricing?

Claude Code, on flat caps ($20 Pro / $100 Max / $200 Max-20x). There is no dollar overage — you hit weekly limits and stop. The tradeoff is that the 12 Feb 2026 internal tokenizer change (the redact-thinking rollout) made those caps hit much faster for the same work: some Max users reported hitting weekly limits in 19 minutes post-change. Aider is second — runs on the API of your choice with no subscription surprise, and uses 4.2x fewer tokens than Claude Code on benchmark refactors. If you genuinely cannot tolerate surprise bills, Claude Code's flat caps or Aider's bring-your-own-API model are the only answers.

What's the cheapest AI coding tool for solo devs?

Aider for most solo devs — MIT-licensed, runs on whichever API you want, benchmarked at 4.2x fewer tokens than Claude Code on 47-file refactors. Typical spend is $22–50/mo via direct Anthropic or OpenAI billing. Cline is the free-tier alternative (OSS, VS Code extension, GitHub Octoverse 2025 fastest-growing AI OSS project) — but Cline has security concerns (the Clinejection prompt-injection vulnerability in Dec 2025–Feb 2026 and the unauthorised @cline/cli npm publish) that Aider does not. For anyone on a strict budget, Aider is the lower-risk choice.

How do I protect myself from surprise bills?

Four checks before committing to any paid AI coding tool. First, ask whether the platform uses persistent indexing or re-reads on every turn (Type A vs B). Second, ask whether the vendor has changed its billing model in the past 12 months — if yes, assume another change is coming. Third, prefer flat-cap plans (Claude Code) or bring-your-own-API (Aider, Cline) over credit-metered consumption for predictable ceilings. Fourth, test on a small project first and extrapolate before committing a team. Replit, Lovable, Windsurf and Make all repriced inside 12 months; the base rate for surprise change is high.

Find the best model for your prompt

VibeDex analyzes your prompt and recommends the best AI image model based on what your specific image demands.

Try VibeDex