guide

AI Pricing Models Explained (2026)

Per-token, subscription, batch, and self-hosted AI pricing explained with examples for OpenAI, Claude, Gemini, DeepSeek, and team seats.

By AI Pricing Guru Editorial Team

AI pricing looks simple until you try to forecast a real bill.

One vendor charges per million input and output tokens. Another sells $20/month seats. A third offers batch discounts, cached-token discounts, or image prices per generation. Open-source models may be free to download but expensive to host. Enterprise plans might look like subscriptions, but the contract still has usage limits underneath.

That is why the right question is not just “which model is cheapest?”

The better question is:

Which pricing model matches the way your team actually uses AI?

This guide explains the main AI pricing models in 2026, where each one works, where each one gets dangerous, and how to compare them without fooling yourself.

The five AI pricing models you will see most often

Most AI products use one or more of these pricing models:

Pricing modelCommon forBest whenMain risk
Per-token API pricingOpenAI, Anthropic, Google, DeepSeek, xAI, MistralUsage is measurable and workload volume mattersOutput tokens, long context, and agent loops can spike cost
Subscription seatsChatGPT, Claude, Perplexity, Cursor, CopilotHumans use the product interactivelyHeavy users may hit hidden limits; light users overpay
Batch pricingOffline API jobs, enrichment, classification, media pipelinesLatency is flexibleJobs scale silently if you do not cap rows and retries
Per-generation pricingImages, video, speech, embeddings in some productsEach unit is easy to countQuality settings and retries make “one image” or “one clip” misleading
Self-hosting / open-source pricingLlama, Mistral, DeepSeek, Qwen-style deploymentsVolume is high and infrastructure skills existGPUs, ops time, reliability, and utilization dominate the economics

The confusing part is that vendors often combine these. An enterprise AI coding tool may charge per seat, route usage through a per-token model, apply rate limits, and sell an add-on for higher usage. A cloud provider may expose the same model through a managed AI platform with separate billing and governance controls.

So before comparing vendors, identify the actual billing unit.

1. Per-token pricing: best for APIs, risky for long answers

Per-token pricing is the standard model for AI APIs.

A token is a chunk of text. In rough English terms, 1,000 tokens is about 750 words, though code, JSON, logs, and non-English text can behave differently.

Most API bills split tokens into three buckets:

  1. Input tokens — what you send to the model
  2. Cached input tokens — repeated prompt or context that gets discounted
  3. Output tokens — what the model writes back

A simple cost formula looks like this:

monthly cost =
(input tokens / 1,000,000 × input price)
+ (cached input tokens / 1,000,000 × cached input price)
+ (output tokens / 1,000,000 × output price)

Here are example public API rates from our current pricing data:

ModelInput / 1MCached input / 1MOutput / 1M
GPT-5.5$5.00$0.50$30.00
GPT-5.4 mini$0.75$0.075$4.50
Claude Sonnet 4.6$3.00$0.30$15.00
Gemini 2.5 Flash$0.30$0.03$2.50
DeepSeek V4 Flash$0.14$0.0028$0.28
Grok 4.1 Fast$0.20$0.05$0.50

The big pattern: output is usually much more expensive than input. GPT-5.5 output costs $30 per million tokens, while its input costs $5. Claude Sonnet 4.6 output is $15 per million versus $3 input. That means a chatty app can cost far more than a concise one even if both receive the same prompts.

For live rate cards, compare the OpenAI pricing page, Anthropic pricing page, Google AI pricing page, and DeepSeek pricing page. To estimate your own workload, use the token cost calculator.

When per-token pricing works well

Per-token pricing is best when:

  • you can measure every request
  • usage varies by customer or feature
  • you need model routing across cheap and premium models
  • you want cost to scale with actual consumption
  • you are building a product where margin matters per task

It is especially good for SaaS products, support bots, enrichment jobs, retrieval-augmented generation, classification, and agent workflows where you need precise unit economics.

Where per-token pricing goes wrong

Per-token pricing gets dangerous when teams estimate by prompt count instead of token volume.

Common mistakes include:

  • ignoring output length
  • sending too much conversation history
  • adding long documents to every request
  • using a premium model for easy tasks
  • letting agents retry, browse, inspect files, or call tools without budgets
  • forgetting that logs and code can tokenize heavily

The fix is to log input, cached input, and output separately. If you only track total requests, you cannot explain the bill.

For a detailed calculation walkthrough, read How to Calculate AI API Costs.

2. Subscription pricing: simple for humans, blurry for companies

Subscription pricing is what most people see first: $20/month for a consumer AI assistant, a higher team plan for collaboration, or a per-seat coding assistant plan.

Subscriptions are appealing because they are predictable. Finance teams like them. Individual buyers understand them. Procurement can count seats.

But subscriptions are not magic. They usually hide one or more constraints:

  • message limits
  • rate limits
  • model access tiers
  • context window restrictions
  • file upload limits
  • workspace or admin features
  • fair-use policies
  • separate API billing

This matters because a $20/month plan can be cheap or expensive depending on the user.

A researcher who uses a frontier assistant all day may get excellent value from a fixed monthly plan. A casual user who asks ten questions per month is probably overpaying. A company that buys 200 seats but has only 60 active weekly users has a utilization problem, not a model-pricing problem.

Seat pricing vs API pricing

A useful rule:

  • Use subscriptions for human productivity.
  • Use API pricing for product workflows.

If an employee uses ChatGPT, Claude, Perplexity, Cursor, or Copilot to write, research, code, or analyze documents, a subscription is usually easier to manage than per-token reimbursement.

If your product calls a model on behalf of customers, per-token billing is usually the better mental model. It lets you calculate margin per user, per task, per ticket, per document, or per workflow.

Do not compare a $20/month assistant plan directly against API token rates unless you know the actual usage limits. They solve different budget problems.

3. Batch pricing: cheaper when latency can wait

Batch pricing is a discount model for work that does not need an immediate response.

Instead of sending one request and waiting synchronously, you upload a file of jobs, let the provider process them asynchronously, and retrieve the results later. This is useful for:

  • document enrichment
  • product catalog cleanup
  • support ticket tagging
  • content moderation backfills
  • embeddings refreshes
  • nightly analytics
  • evaluation runs
  • image or video generation queues

Batch can be cheaper because providers can schedule the work more efficiently. For buyers, the tradeoff is latency. If the job can finish in minutes or hours instead of seconds, batch pricing can improve unit economics.

The danger is scale. A chat UI makes cost visible because a person initiates each request. A batch job can process 5,000 rows today, 500,000 rows next week, and retry failures twice without anyone noticing until the invoice arrives.

Good batch hygiene:

  • cap rows per job
  • estimate tokens before upload
  • sample outputs before running the full dataset
  • separate batch budgets from interactive budgets
  • log retries as their own cost bucket
  • use cheaper models for easy classification or extraction

Batch pricing is not just a discount. It is an operational pattern. Treat it like one.

4. Per-generation pricing: easy to understand, hard to normalize

Image, video, speech, and some embedding products use per-unit pricing. Instead of tokens, you may pay per image, per minute, per file, per character, or per generated clip.

This feels simpler than token pricing:

  • one image costs X
  • one video costs Y
  • one transcription minute costs Z

But quality settings matter. A 1024×1024 standard image is not the same cost as an HD image. A short voice clip is not the same as a long voice session. A failed image generation that you retry three times still affects real cost.

The right way to compare per-generation tools is by accepted output, not attempted output.

For example, if a marketing team needs 100 usable images and each provider has a different retry rate, the sticker price per image is only the starting point. A tool that costs twice as much but produces usable results with half the retries may be cheaper per accepted asset.

That is why image and video pricing should be tracked with workflow metrics:

  • cost per accepted image
  • cost per edited image
  • cost per campaign asset
  • cost per minute of usable audio or video
  • cost per approved creative variation

If you are comparing image tools, see our AI image generation pricing guide.

5. Self-hosting: cheap models do not mean cheap operations

Open-source and open-weight models can look nearly free at first glance. You can download the model. There may be no per-token vendor bill. At high volume, self-hosting can absolutely win.

But the real cost is infrastructure:

  • GPUs or specialized inference hardware
  • cloud instances or reserved capacity
  • inference servers
  • monitoring and alerting
  • scaling and queueing
  • security updates
  • model evaluation
  • reliability work
  • engineering time

Self-hosting is usually strongest when you have high, predictable volume and enough technical depth to keep utilization high. A GPU sitting idle is not cheap. A GPU running near capacity on stable workloads can be very competitive.

A simple way to think about it:

self-hosted unit cost =
(monthly infrastructure + engineering + ops overhead)
/ successful model outputs

If you only send a small number of requests, hosted APIs are usually cheaper because you avoid fixed infrastructure and operational complexity. If you process millions or billions of tokens with predictable traffic, self-hosting or managed open-model hosting becomes more attractive.

This is why many teams use a hybrid approach: hosted frontier APIs for hard tasks, cheaper hosted models for routine work, and self-hosted models only where the workload is stable enough to justify the fixed cost.

Worked example: API vs subscription vs self-hosting

Imagine a small company with three AI use cases:

  1. 20 employees using AI assistants
  2. a support bot answering customer questions
  3. a nightly product-data cleanup job

The cheapest pricing model is probably different for each one.

Employee assistants

A per-seat subscription is easiest. The company can buy seats, monitor active usage, and cancel inactive licenses. Token-level accounting would create more overhead than value.

Best fit: subscription seats.

Support bot

The support bot needs margin discipline. Each customer question has an input, context, and output cost. The company should route simple questions to a cheap model, reserve premium models for complex escalations, and track cost per resolved ticket.

Best fit: per-token API pricing, with caching and model routing.

Product-data cleanup

The nightly cleanup job does not need instant answers. It can run asynchronously and tolerate delayed results. The company should test a sample, estimate token volume, then run in batches with job-size caps.

Best fit: batch pricing.

Could self-hosting help? Maybe, but only if the support bot or cleanup job reaches enough predictable volume to justify fixed GPU and operations costs.

How to choose the right pricing model

Use this decision table:

If your workload is…Start with…Why
Human productivitySubscription seatsPredictable budget and simple admin
Customer-facing product featurePer-token APILets you calculate margin per user or task
Offline enrichment or backfillsBatch pricingLower cost when latency is flexible
Creative image/video/audio workflowsPer-generation pricingEasier to track cost per accepted asset
Huge stable inference volumeSelf-hosting or managed open modelsFixed infrastructure can beat API markup at scale
Mixed complexityModel routingCheap models handle easy work; premium models handle hard work

The best AI budget is rarely one model or one vendor. It is usually a routing system:

  • cheap model for simple tasks
  • mid-tier model for normal work
  • premium model for hard tasks
  • cached prompts wherever possible
  • batch jobs when latency does not matter
  • subscriptions only where humans need interactive tools

FAQ

Is per-token pricing always cheaper than subscriptions?

No. Per-token pricing is better for measurable product usage, but subscriptions can be cheaper for heavy human users because the monthly fee smooths out variable consumption. The hard part is knowing the hidden limits and active-seat utilization.

Should startups choose the cheapest model by token price?

Not blindly. Cheap models are great for routing, extraction, classification, and high-volume simple tasks. For complex reasoning, coding, or support answers, a more expensive model may be cheaper per successful outcome if it reduces retries and human review.

When does self-hosting become worth it?

Usually when usage is high, predictable, and technically mature enough to keep GPUs utilized. If your traffic is small or spiky, hosted APIs often win because you avoid fixed infrastructure and operations work.

What is the easiest way to reduce an AI bill?

Start with output control, caching, and routing. Shorter answers, cached system prompts, and cheaper default models often reduce spend faster than vendor negotiation.

Bottom line

AI pricing is not one market. It is several billing models stacked together.

Per-token APIs are best for measurable product workflows. Subscriptions are best for human productivity. Batch pricing is best when work can wait. Per-generation pricing works for creative assets if you track accepted outputs. Self-hosting can win at scale, but only when infrastructure utilization and operations are under control.

If you want a practical next step, take one real workflow and price it three ways: subscription, API, and batch or self-hosted where relevant. Then compare cost per successful outcome, not just sticker price.

That is the metric that actually protects your AI budget.