Back to Blog

Why AI Pricing Requires Three Layers: Architecture, Observability, and Trust

A Clay Case Study

TL;DR

AI pricing works when buyers can predict value, usage, and cost before committing.
This article analyzes Clay's new pricing architecture using the AVS Trust Rubric and shows why AI pricing systems now require three layers: economic architecture, pricing observability, and trust surface distribution.

Clay announced a major pricing overhaul.

The architecture improved.

But when the AVS Trust Rubric evaluated the new pricing page, the score dropped from 81% to 75%.

That drop is not a failure signal. It is a buyer signal, because the score reflects the first place buyers look: the pricing page.


What the AVS Trust Rubric Measures

The Adaptive Value System (AVS) Rubric evaluates whether an AI SaaS product communicates predictable value, usage, and cost signals to buyers using only publicly observable signals.

It scores eight dimensions across a four-layer trust stack:

LayerCategoryDimensions
Layer 1FoundationProduct and Audience Clarity
Layer 2Value-Cost AlignmentValue Unit, Cost Driver Mapping
Layer 3Risk ManagementPools and Packaging, Overages and Risk Allocation, Safety Rails and Trust Surfaces
Layer 4Buyer AlignmentBuyer and Budget Alignment

Each dimension scores 0–2, with a confidence label reflecting how much public evidence the rubric can find.

The score is not the destination. It reveals which signals buyers can easily reach and which signals require effort to find or even understand.

The score does not measure product quality. It measures how easily buyers can infer value, usage, and cost before committing.

The AVS rubric evaluates trust infrastructure across four internal layers. But when buyers experience pricing during evaluation, those signals typically appear through three observable stages.


The Three Stages of The AI Pricing System

Clay's pricing change also highlights a broader structural shift happening across AI software.

The AI pricing system now operates across three stages:

1. Trust Surface Distribution: Limited Visibility

  • Pricing page
  • Cost examples
  • Spend controls

2. Pricing Observability: Partially Visible

  • Cost drivers
  • Usage visibility
  • Guardrails (alerts, caps)

3. Economic Architecture: Improved

  • Data Credits
  • Actions
  • Value unit design

Most AI startups solve the economic architecture stage first.

A smaller number build strong pricing observability.

Very few solve the final stage well: trust surface distribution, where pricing signals appear during evaluation.


What Changed in Clay's Pricing

Clay's pricing overhaul improved architecture and observability. But the third stage still shapes how buyers experience the new pricing model.

Clay's previous model priced everything through credits — data enrichment lookups, AI research steps, automation logic, outbound execution. One unit for every job the platform performed.

That worked when Clay was primarily a data enrichment tool.

It stopped working when Clay evolved into a full GTM automation system.

The credit model was capturing value from the data infrastructure, less from the automated GTM workflows where Clay now creates most of its value.

The new architecture separates two things that were always economically distinct:

  • Data Credits cover the cost of external data providers. Clay reduced marketplace data costs 50–90% and passes AI model usage through at 0% markup.
  • Actions measure the automation workflows Clay executes — research steps, logic, orchestration, and outbound motion. This becomes the monetized value layer.

The economic logic is sound. Data infrastructure becomes a pass-through. Automation becomes the durable revenue engine.

Clay acknowledged an expected 10% short-term revenue drop during the transition — a deliberate trade for a pricing architecture that scales with product usage over time.


Run One of the AVS Rubric: February Pricing — 81%

Before the pricing change, Clay scored 81%.

2/2

Value Unit — 80% confidence

2/2

Pools & Packaging — 70% confidence

The foundation was strong: clear target audiences, legible pricing tiers, and defined usage boundaries.

The gaps were specific and solvable. Cost driver transparency and safety rail documentation were the main weaknesses holding the score at only 60%.


Run Two of the AVS Rubric: March Pricing — 75%

After the pricing announcement, the rubric scored Clay 75%.

The architecture improved. Yet the score dropped six points.

That gap is worth reading closely.

It is not a failure signal. It is a discoverability signal.

When Actions became the value unit, the observability bar rose with it.

Under the old credit model, buyers could map credits directly to data lookups. The unit was concrete.

Under the new model, buyers need to understand how GTM workflows translate into Actions consumption before they can forecast spend confidently.

At the surface level of the new pricing page, that mapping was not visible.

Current Page Analysis

The Value Unit and Pools and Packaging dimensions held at 2/2, with confidence rising to 90% due to the cleaner architecture.

But two dimensions weakened:

1/2

Cost Driver Mapping — 60% confidence

1/2

Safety Rails & Trust Surfaces — 50% confidence

The 75% score reveals something precise. Clay's pricing structure is coherent. But the evidence buyers need to forecast workflow costs requires more than a standard pricing page review to find.


Run Three: When the Right Material Is Surfaced — 94%

The pricing FAQ at the bottom of the page contains a deep link many buyers would not find on a first pass.

That link points to Clay University's documentation on Actions & Data Credits — a detailed breakdown of how Data Credits and Actions interact, how workflows consume each resource, rollover policies, and the guardrails available to manage spend.

Cost Driver Mapping

1/2 at 60% → 2/2 at 80% confidence

The documentation clearly explains how workflows translate into usage.

Safety Rails and Trust Surfaces

1/2 at 50% → 2/2 at 80% confidence

Usage controls, rollover policies, and limit behavior become visible.

Overall observability moves from 68% to 79%.

The final score becomes 94%.

The product did not change. The evidence did.


What the Score Arc Actually Tells Founders

The most useful score in this sequence is 75%, not 94%.

94% tells you Clay's trust infrastructure is strong when a buyer performs deep research.

75% tells you where buyers lose confidence during the first evaluation pass.

Clay did not have a documentation problem.

Clay had a trust surface distribution problem.

The university.clay.com documentation is detailed and technically precise. But it lives several levels below where procurement teams, finance approvers, and internal champions usually evaluate pricing.

Buyers typically encounter pricing signals in four places:

  1. The pricing page
  2. The FAQ
  3. A buyer's guide / cost calculator (on the pricing page)
  4. User reviews

If trust signals live deeper than that, they function as support documentation, not decision infrastructure.


Clay's new architecture separates infrastructure costs (Data Credits) from automation value (Actions). Clay's documentation explains how Data Credits and Actions behave.

What's missing is helping buyers translate workflows into expected cost before running them at scale.

For usage-based AI pricing, buyers need three operational signals:

  1. What drives cost
  2. How workflows translate into usage
  3. How spend can be controlled

The Clay University documentation addresses these questions, but most buyers encounter pricing signals first on the pricing page, not in deep documentation.

Making those signals visible earlier would significantly strengthen cost predictability.


First Two Fixes

1. Surface Workflow Cost Examples on the Pricing Page

The documentation explains how Data Credits and Actions behave, but buyers still need to translate that into real GTM workflows.

Adding a simple "Workflow Cost Examples" section on the pricing page would bridge that gap.

Example WorkflowData CreditsActions
Enrich 1,000 leads with company + contact data1,000–3,000Minimal
Run AI research on 500 accountsMinimal500–1,500
Automated outbound prospecting workflow1,000–2,000500–1,200
Multi-source enrichment + AI research2,000–5,0001,000–3,000

These ranges do not need to be perfectly precise. Their purpose is to help buyers answer the question they ask during evaluation: "What does a typical workflow cost?"

Without this translation layer, buyers must infer cost behavior themselves.

2. Surface Spend Controls Directly in the Pricing Experience

The documentation describes how usage behaves, but buyers also want to know how spend can be controlled once automation is running.

Explicitly surfacing available controls would reduce perceived risk. Examples include:

a) Budget Caps

Allow teams to set monthly spending limits:

  • Total workspace budget
  • Action usage caps
  • Data Credit caps

b) Usage Alerts

Configurable alerts at thresholds:

  • 50% of monthly budget
  • 75% of monthly budget
  • 90% of monthly budget

c) Pre-Run Cost Estimates

Show estimated consumption before executing:

Estimated run cost:
1,200 Actions
2,300 Data Credits

These fixes do more than clarify Clay's pricing page. They illustrate a broader shift happening across AI software.


What This Signals for AI SaaS Pricing

Clay is not alone in this dynamic. PostHog, Replicate, and Modal have followed similar arcs. First usage pricing appears. Then predictability tooling improves. Then observability surfaces mature enough to change buyer behavior.

Three structural shifts are emerging across AI software:

  1. Usage pricing is becoming the default. Companies that once priced data are increasingly pricing workflows.
  2. Value units are moving up the stack. Companies that once priced data are now pricing workflows.
  3. Discoverability is becoming part of the pricing system. Documentation that exists but is not surfaced cannot build buyer confidence.

The architecture now reflects where Clay creates value. The observability infrastructure largely exists.

But the final stage still determines how buyers experience the system: where pricing signals appear during evaluation.

In Clay's case, the difference between the 75% score and the 94% score is not a product gap. It is the distance between documentation and discovery.

Many AI-native companies are operating inside that same gap today.

Most would score somewhere between 40% and 70% on these signals.


The AVS Trust Rubric

The AVS Trust Rubric evaluates AI SaaS products using publicly observable signals across eight trust dimensions.

It measures how easily buyers can infer value, usage, and cost before committing.

See Your Trust Gaps

Get your AVS assessment in 60 seconds