Community-maintained pricing data for AI infrastructure services beyond the LLM layer — vector databases, inference hosts, managed agents, embedding services, MCP tool calls, fine-tuning, and evaluation platforms. YAML-first, schema-validated, supply-chain-hardened.
For LLM pricing specifically, we recommend LiteLLM's model_prices_and_context_window.json. This project complements that file; it does not replace it.
Modern AI cost attribution needs more than LLM pricing. An observability or FinOps tool running cost queries today needs pricing for:
- Vector databases — Pinecone per-query + storage-GB, Weaviate cluster pricing, Qdrant tier pricing
- Inference hosts — Replicate per-second, Modal per-second, Together per-token-per-model, Groq per-token, Fireworks per-token
- Managed agents — OpenAI Assistants API, Anthropic Assistants, agent orchestration services
- Embedding services — many now priced separately from the parent provider
- MCP tool calls — when metered per-invocation
- Fine-tuning — per-token training cost + per-inference host cost
- Evaluation platforms — per-eval-run cost
LiteLLM's file is excellent for LLM pricing and we inherit it. This repo covers the rest.
In March 2026, two LiteLLM PyPI releases were compromised via a poisoned GitHub Action. 40,000 packages were downloaded in 40 minutes before PyPI quarantined them. The compromise wasn't in source code — it was in the build pipeline.
If pricing data is a single supply-chain surface, a compromise means every downstream cost dashboard shows the wrong numbers. Customers reconciling against invoices find discrepancies weeks later. Trust takes years to rebuild.
This project treats pricing data as security-critical infrastructure. Practices we've adopted from the outset:
- Every GitHub Action SHA-pinned (never tag-pinned)
- SLSA Level 3 provenance on releases — build attestation signed by GitHub's OIDC identity
- SBOM published with every release — human + machine readable dependency inventory
- Signed commits required on the
mainbranch - CodeQL + dependency review on every PR
- Reproducible builds — verify locally that what's released matches what's in the repo
- Public audit log — every price change is a signed PR with human review
See SECURITY.md and THREAT_MODEL.md for the full posture.
- Vector databases: Pinecone, Weaviate, Qdrant, Chroma, Milvus/Zilliz
- Inference hosts: Replicate, Modal, Together, Groq, Fireworks, Anyscale
- Managed agent platforms: OpenAI Assistants pricing, Anthropic managed agents, crew/swarm platforms
- Embedding services (where priced separately): Cohere, Voyage, Jina
- MCP server monetization platforms (vend.sh compatible)
- Fine-tuning pricing across providers
- Evaluation platforms: Langfuse, Arize, Phoenix, Weights & Biases
- Core LLM inference pricing (GPT-4o, Claude, Gemini, Mistral, etc.)
- LLM context-window and mode metadata
- LLM rate-limit data
- Cloud infrastructure pricing (AWS, GCP, Azure) — those providers have rich pricing APIs; reinventing isn't useful
- Historical pricing archive — this is current pricing only, with git history providing audit trail
- Usage data — this repo is pricing data; usage belongs in observability tools
- Benchmarks — use Artificial Analysis or similar
Every provider's pricing lives in a single YAML file under data/. Files follow schema/pricing-schema.yaml. Example:
provider: pinecone
homepage: https://www.pinecone.io/pricing
pricing_url: https://www.pinecone.io/pricing/
last_reviewed: 2026-04-15
reviewed_by: '@jeffyaw'
services:
- id: serverless-reads
category: vector_db
description: Pinecone serverless read units
pricing:
read_units:
unit: per_1m
price_usd: 8.25
notes: Billed per million read units consumed
- id: serverless-storage
category: vector_db
description: Pinecone serverless storage
pricing:
storage_gb:
unit: per_gb_month
price_usd: 0.33The schema validates: required fields, category enum (defined in pricing-schema.yaml), pricing unit enum, ISO-8601 date on last_reviewed, and cross-references.
# Pin to a specific commit SHA — never a tag or branch
git submodule add \
https://github.com/YawLabs/ai-pricing \
vendor/ai-pricing
cd vendor/ai-pricing
git checkout <specific-commit-sha># Download from the GitHub release for the version you want:
curl -LO https://github.com/YawLabs/ai-pricing/releases/download/v0.1.4/ai-pricing-0.1.4.tar.gz
curl -LO https://github.com/YawLabs/ai-pricing/releases/download/v0.1.4/ai-pricing.intoto.jsonl
# Verify with slsa-verifier (see Provenance verification section below), then:
tar -xzf ai-pricing-0.1.4.tar.gzThe YAML files under data/ are designed to be human-inspectable and loaded directly from whatever runtime reads them — no package wrapper needed.
Every release ships with a signed SLSA provenance attestation. To verify a release yourself:
# Install slsa-verifier (pinned to v2.5.1 SHA)
curl -sSL https://github.com/slsa-framework/slsa-verifier/releases/download/v2.5.1/slsa-verifier-linux-amd64 \
-o slsa-verifier && chmod +x slsa-verifier
# Verify release tarball
./slsa-verifier verify-artifact \
ai-pricing-v1.0.0.tar.gz \
--provenance-path ai-pricing-v1.0.0.intoto.jsonl \
--source-uri github.com/YawLabs/ai-pricing \
--source-tag v1.0.0SLSA provenance binds the released artifact to the exact GitHub Actions workflow that built it, signed by GitHub's OIDC-issued certificate. A compromised build pipeline can't produce a valid attestation.
See CONTRIBUTING.md. Short version:
- Signed commits only on
main.git commit -S -m "..."or set commit signing as default. - One provider per PR. Don't mix.
- Include
last_revieweddate + verifier handle in the YAML. - Link to the provider's pricing page in the PR description.
- Screenshots welcome for nonstandard pricing structures.
- CI will validate schema, lint YAML, and fail on any unvalidated entries.
Vulnerability reports: open a private advisory (preferred) or email security@mcp.hosting. Details in SECURITY.md.
Never open a public issue for a vulnerability — use the private channel, coordinate disclosure.
- Pricing data (
data/): CC BY 4.0 — use it anywhere; attribution appreciated. - Tooling (
scripts/, validation code): MIT.
- Yaw Labs (YawLabs on GitHub) —
@jeffyawprimary, plus community contributors as recognized.
@yawlabs/mcp-compliance— open methodology for grading MCP server spec compliance. Same supply-chain practices.- Spend — AI spend tracking, cost estimation, and provider comparison. Uses this repo + LiteLLM's file as its pricing data sources, with both sources independently maintained for supply-chain resilience. On the reconciliation path (e.g. "list this month's usage," "fetch invoice," "get quota remaining") Spend is MCP-native — each provider's admin API is wrapped as its own MCP server and Spend consumes them as a client, so the same protocol it sells to users is the one it uses to reason about provider billing.
- vend.sh — license-key-native billing for builders. Includes MCP tool-call pricing pattern that informs the
mcp_tool_callcategory here.