πŸ€– Agent Readiness Scorer

How ready is this product for AI agent commerce?

What is this?

When AI agents start buying things, which businesses will they be able to work with? That turns out to be a surprisingly hard question to answer by just looking at a website.

Some things are obvious. If a product has a public API with documentation, that's a good sign. If the only way to buy is "contact sales," that's a bad sign. But most companies fall somewhere in the messy middle β€” they have an API but no machine-readable pricing, or they offer self-serve signup but gate it behind a CAPTCHA.

We wanted a way to be more systematic about it. Not because a score can capture everything, but because the exercise of defining what "agent-ready" means forces you to be specific about what actually matters.

The Rubric

We score four categories. Each is worth 10 points, for a total of 40.

πŸ” Discovery

Can an agent find this product and understand what it does?

  • Public API documentation
  • OpenAPI / machine-readable specs
  • Machine-readable pricing
  • Listings in directories and marketplaces

If an agent can't figure out what you sell and what it costs without parsing marketing copy, you're invisible to the machine economy.

πŸ’³ Purchase

Can an agent buy this without a human?

  • Programmatic signup (no CAPTCHAs)
  • API-based billing
  • Free tier or self-serve trial
  • No mandatory sales calls

Every "contact us for pricing" page is a door that's closed to agents.

πŸ”Œ Integration

Can an agent actually use this?

  • SDKs and client libraries
  • Webhooks and event streams
  • Sandbox / test environments
  • Standard authentication (OAuth, API keys)

The difference between a product an agent can use and one it can't often comes down to whether there's a test environment.

πŸ›‘οΈ Trust

Can an agent verify this is reliable?

  • Published SLAs
  • Public status page
  • Spend limits and usage controls
  • Security documentation

When agents are spending money autonomously, they need programmatic ways to assess risk β€” not a logo wall of enterprise customers.

Is this score definitive?

No. A 40-point automated crawl is a blunt instrument β€” more conversation starter than definitive assessment. But even a rough score surfaces patterns that are hard to see otherwise.

We're putting this out there not because we think the scoring is perfect, but because we think the conversation matters. What should agent-readiness mean? Are we weighting the right things?

Tell us what we got wrong β€” on X or in the Substack comments.