I've reviewed more than 30 AI SaaS pitch decks in the last year. Every single one shows gross margin of 70–80%. Every single one is wrong — not because the founders are being dishonest, but because they're using a traditional SaaS cost model that breaks the moment LLM API spend becomes significant.

The error is consistent: LLM costs are modeled as infrastructure — fixed, amortized, scaling sublinearly with users. They're not. LLM API spend is variable COGS that scales directly with feature usage. When a user engages more, the cost goes up. When a feature becomes core to the product, the cost grows with every interaction. Traditional gross margin calculations that bury this in "hosting" or exclude it from COGS are showing investors a number that doesn't reflect reality.

TL;DR

1. LLM API costs are variable COGS — they scale with engagement, not with servers. AI SaaS gross margins are typically 40–60%, not 70–80%, once LLM spend is correctly classified.
2. The five metrics that matter — LLM cost per active user, gross margin by tier, power user margin, feature-level contribution, and LLM intensity ratio — aren't in standard SaaS dashboards.
3. The gap between unoptimized and optimized LLM costs is 40–60%. Most AI SaaS companies are presenting unoptimized margins as their steady-state.

Why Traditional SaaS Economics Break for AI Products

Traditional SaaS gross margin analysis is built on a stable assumption: infrastructure costs are largely fixed and decline as a percentage of revenue as you scale. AWS costs don't go up because a user spends more time in your app. Database costs don't double when usage doubles (at most scales). So you model COGS as a roughly fixed percentage and forecast healthy 70–80% gross margins at scale.

LLM API costs violate every part of this model:

None of this is fatal — it's a different cost structure, not a broken one. But it requires different metrics, different modeling, and different conversations with investors than traditional SaaS requires.

The Real Waterfall: From $29 MRR to Actual Gross Margin

Take a concrete example: an AI writing assistant at $29/month. Standard SaaS COGS math shows ~85% gross margin. Here's what it actually looks like when LLM API costs are modeled correctly as variable COGS:

Gross margin waterfall — AI writing assistant, $29/mo plan, per-user
Monthly revenue
$29.00
− Hosting & infra
$27.00 after −$2.00
− Support allocation
$25.50 after −$1.50
Scenario A — Unoptimized (avg 50 queries/day, gpt-4.1)
− LLM API costs
$12.63 after −$12.00
− Payment processing
$11.76
Gross profit
$11.76 = 41% gross margin

Scenario B — Optimized (routing + caching, same usage)
− LLM API costs
$18.63 after −$6.50
− Payment processing
$17.76
Gross profit
$17.76 = 61% gross margin

Same product, same users, same engagement. Unoptimized: 41% gross margin. Optimized with routing and caching: 61%. The 20-point gap is the difference between an AI SaaS that passes investor due diligence and one that doesn't.

Unit Economics by AI SaaS Category

LLM cost as a COGS component varies significantly by product type. Here are realistic unoptimized gross margins by category — these assume typical pricing and average (not power) user behavior:

AI SaaS Category Typical Price LLM Cost/User/Mo (Unopt.) Gross Margin
AI writing assistant $19–49/mo $6–$18 35–60%
AI coding assistant $19–99/mo $5–$25 45–70%
AI customer support $200+/mo per seat $30–$80 55–75%
Agentic workflow tools $99–299/mo $40–$120 20–60% (wide variance)
AI document analysis $49–199/mo $15–$60 40–65%
AI research tools $29–99/mo $12–$40 40–65%

Agentic tools show the widest variance because each user action can trigger 5–15 LLM calls. At high engagement — which is what product success looks like — agentic features can push individual users into negative margin territory without feature-level cost caps.

The Five Metrics Your Pitch Deck Is Missing

Standard SaaS metrics (MRR, churn, CAC, LTV) don't surface the AI-specific economics. Here's what you need to add:

1
LLM cost per daily active user (not per seat)
Seat-based cost averaging masks the actual cost of engaged users. A monthly active user who opens your product once costs near zero. A daily active user running 50 AI queries costs $10–30. Track cost per DAU session to understand your true unit economics at scale.
LLM cost per DAU = (monthly LLM spend) ÷ (sum of daily active sessions in the month)
2
Gross margin by pricing tier
Blended gross margin is almost always misleading for AI SaaS. Your free tier and entry paid tier may have very different usage patterns from your enterprise tier. Calculate gross margin per tier separately — you'll often find one tier subsidizing another.
Tier gross margin = (tier MRR − tier COGS) ÷ tier MRR
3
Power user LLM spend (top 10% by usage)
In most AI SaaS products, the top 10% of users by feature usage account for 40–60% of LLM costs. If those users are on your lowest pricing tier (often the case — heavy users explore before upgrading), you have a systematic negative-margin cohort. Identify it before investors do.
Power user margin = (tier price − P90 LLM cost − other COGS) ÷ tier price
4
Feature-level gross contribution
Different AI features have vastly different cost profiles. An AI chat feature at 500 tokens per exchange costs near nothing. An agentic task runner at 8 LLM calls and 2,000 tokens per task is 30x more expensive per user action. Without per-feature cost attribution, you can't know which features are margin-positive or how new feature launches affect your unit economics.
Feature gross contribution = (feature revenue allocation) − (feature-attributed LLM cost)
5
LLM intensity ratio (LLM cost as % of MRR)
This is your leading indicator of margin compression. Track it monthly. When LLM spend grows faster than MRR — which happens when engagement grows faster than pricing — your margin is compressing in real time. An LLM intensity ratio above 30% of MRR is a flag; above 50% is a crisis.
LLM intensity = (monthly LLM API spend) ÷ (monthly MRR) × 100

Four Scenarios: From Healthy to Crisis

To make this concrete, here's how the unit economics play out across four real-world usage patterns for a $99/month AI tool:

Scenario A
Light AI usage, cheap model
Usage: 10 queries/day
Model: gpt-4.1-nano
LLM cost/user/mo: ~$0.75
Gross margin: ~89%
Healthy — traditional SaaS-like
Scenario B
Core AI usage, capable model
Usage: 50 queries/day, 1K tokens avg
Model: gpt-4.1
LLM cost/user/mo: ~$18
Gross margin: ~73%
Viable — watch carefully as usage grows
Scenario C
Power user, unoptimized agentic
Usage: 20 tasks/day × 8 calls each
Model: gpt-4.1, 1.5K tokens/call
LLM cost/user/mo: ~$144
Gross margin: Negative
Crisis — paying to serve this user
Scenario D
Same as C, optimized
Routing: 70% nano, 30% gpt-4.1
Caching: 20% hit rate on duplicates
LLM cost/user/mo: ~$55
Gross margin: ~38%
Manageable — viable at this tier

Scenario C is the one that kills companies quietly. The power users who trigger it are often your most engaged, most likely to refer, most vocal about product value — and each one is generating a loss. Without per-user cost attribution, you don't see it until the aggregate margin is already compressed.

Want to calculate your real gross margin per pricing tier?

Plug your pricing, daily query volume, and model mix into the AI Unit Economics Calculator. See your per-tier margin, power user break-even threshold, and optimization impact in five minutes.

Get the AI Unit Economics Calculator

Free interactive tool. Enter your email and start calculating.

What the Pitch Deck Should Actually Show

Investors who understand AI unit economics are starting to ask the right questions. Here's what a defensible AI SaaS unit economics slide looks like in 2026:

Show gross margin by tier, not blended. "Our enterprise tier runs 72% gross margin; our starter tier runs 48%" is a far more credible and useful number than a blended "65% gross margin." Blended averages hide cross-tier subsidy and don't tell you where margin improvement dollars should go.

Show the margin improvement trajectory. Current unoptimized state → planned state with routing and caching → target state at scale. The gap between current and optimized is credible evidence that the business model works — you just haven't deployed the optimizations yet. This is the AI equivalent of traditional SaaS infrastructure cost curves.

Show power user economics explicitly. "Our top decile of users by usage breaks even at $X/month. We've implemented usage caps at that level on the starter tier, and our enterprise pricing is designed to capture this segment profitably." This demonstrates you've thought through the problem — which most AI founders haven't.

Show LLM cost per unit of delivered value, not just per user. Cost per document processed, per task completed, per query answered — these tie LLM spend to business outcomes rather than presenting it as a raw cost item.

The Optimization Path: From Current to Defensible

Most AI SaaS companies are presenting their current, unoptimized LLM cost structure as the steady state. They're not. The optimization path is well-understood and the improvements are significant:

Teams that deploy all four typically move from unoptimized to optimized within two to four weeks. The resulting margin profile is what you should be building your unit economics model around, not the current baseline.

Frequently Asked Questions

What is a realistic gross margin for an AI SaaS product?
It depends on product category and whether costs are optimized. AI writing and coding tools typically run 40–65% gross margin unoptimized, improving to 55–75% with routing and caching. Agentic tools run wider variance — 25–65% — because power users can trigger negative per-user margin. The key point: SaaS gross margin calculations that exclude LLM API costs from COGS are overstating margins. LLM API spend is variable COGS, not infrastructure.
How do you calculate LLM cost per user per month?
Multiply: (average daily active sessions) × (average tokens per session) × (blended cost per token across your model mix) × 30. The most common error is using monthly active users as the denominator — that understates cost per engaged user. Segment by pricing tier: your free tier power users likely have very different economics than your paid tier.
What is the biggest mistake AI SaaS founders make on unit economics?
Modeling LLM API costs as fixed infrastructure rather than variable COGS. Traditional SaaS infrastructure cost per user decreases as scale grows. LLM API costs grow linearly with usage — sometimes superlinearly if agentic features trigger multiple calls per user action. Engagement growth that looks great on product metrics can quietly compress gross margins in the same month.
How does usage-based pricing change AI SaaS unit economics?
Usage-based pricing aligns revenue with cost — revenue grows when LLM usage grows. Fixed subscription pricing is the risky model: if a user's usage grows 4x but revenue stays flat, margin erodes. The practical middle ground: tiered pricing with usage limits enforced at the proxy layer. Hard limits prevent negative-margin power users while keeping the simplicity of subscription billing.
What is the difference between gross margin and contribution margin for AI SaaS?
Gross margin excludes sales, marketing, and R&D — it's revenue minus COGS (hosting, support, LLM API). Contribution margin subtracts variable costs that scale with customer count, including CAC amortization and customer success. Both should include LLM API costs as COGS. Gross margin is right for SaaS benchmarking; contribution margin by cohort tells you whether growth is actually value-creating.

Calculate your real AI margin per user.

The AI Unit Economics Calculator lets you plug in your pricing tiers, daily query volume, and model mix to see your actual gross margin, power user break-even threshold, and the impact of routing and caching. Takes five minutes.

Get the AI Unit Economics Calculator

Free interactive tool. Enter your email and start calculating.

Gaurav Dagade
Gaurav Dagade

Founder of Preto.ai. 11 years engineering leadership. Previously Engineering Manager at Bynry. Building the cost intelligence layer for AI infrastructure.

LinkedIn · Twitter