Pay-per-call
AI gateway.
Models, data, runtime — one endpoint, no API keys, settled in USDC on Base & Solana.
Cut AI costs
by up to 78%.
A smart router for LLM calls. Routes to the cheapest capable model, compresses prompts, and caches responses. Drop-in OpenAI-compatible.
// route by task type → cost ceiling export default defineRouter({ fallback: "deepseek-v3.2", rules: [ { task: "code", model: "claude-sonnet-4.6", max: 0.02 }, { task: "reasoning", model: "gpt-5.4", max: 0.03 }, { task: "summarize", model: "gemini-2.5-flash", max: 0.005 }, ], compress: true, cache: true, });
The AI agent with a wallet.
Other agents write code. Franklin writes code and spends money to get things done. It holds your USDC, picks the best model per task, buys trading data, generates images, searches the web — and decides what's worth paying for. You set the budget. It runs it.
- Claude, GPT-5, Gemini, DeepSeek, Grok, and more
- USDC on Base + Solana · budget caps enforced on-chain
- Auto-selects best model per task — up to 89% savings vs Opus
- Learns your preferences locally across sessions
- `franklin migrate` imports Claude Code sessions
One install.
8 agent tools.
Drop BlockRun into Claude Code, Cursor, or any MCP client. Your agent gets models, live data, media generation, and a secure runtime — all paid with USDC.
55+ models.
One endpoint.
Chat, image, video, and music — one OpenAI-compatible API. Provider cost + 5% margin at settlement. No subscriptions, no minimum spend, no vendor lock-in.
AI buys data & runtime.
Any API. Any agent. Pay per request with USDC via x402 — no keys, no subscriptions, no accounts.
Neural search, find similar pages, extract content, grounded answers.
Polymarket, Kalshi, dFlow, Binance Futures data for agents.
Isolated code runtime for agents. Sandbox sessions, execute, inspect.
On-chain data, financial feeds, niche APIs — we onboard weekly.
AI infrastructure
for teams.
Budget controls, usage analytics, team management, and dedicated support for production AI deployments.
- Per-key spend caps, hard limits, alerts
- Per-team, per-model, per-call attribution
- SSO, roles, audit log, key rotation
- Slack channel, SLA, named CSM
We don't share
your data.
Your prompt goes to the AI provider you picked. Nothing else, nowhere else. No training, no retention beyond the request, no profile linking.
- No training, no retention beyond the request. Your prompt is forwarded only to the AI provider you select.
- Wallet in, prompt out. Pseudonymous by default — no email, no phone number, no identity documents.
- Read the code, audit the wire format, run it yourself. @blockrun/llm and blockrun-llm on npm and PyPI.
Questions we get
before you ask them.
- Do you train on my prompts?
- No. We forward your prompt only to the AI provider you select to generate the response. We don't store it beyond the request, and we don't share it with anyone else.
- What about "free" models?
- We don't ship free models that require sharing your data with the upstream provider. If a free tier's terms allow training on user prompts, we don't list it.
- How is this different from OpenRouter or Portkey?
- Pay-per-call USDC on Base or Solana via x402. No account, no API key, no subscription. Auto-fallback when a model is down. SDKs are OpenAI-compatible.
- Why USDC instead of a subscription?
- Agents can hold USDC; agents can't hold a credit card. Settlement happens on-chain in the same request via the x402 protocol — no prepaid credits, no minimum spend.
- What happens if a model is down?
- We auto-fallback to a comparable model and surface which one served you via the X-Fallback-Model response header, so your application never sees a hard failure.
- Are the SDKs OpenAI-compatible?
- Yes. Change base_url and you're done — every endpoint matches OpenAI's shape. @blockrun/llm for TypeScript, blockrun-llm for Python, both MIT-licensed.
- Can I use this in production?
- Yes. We run on Base mainnet with 1M+ calls per month. Use testnet.blockrun.ai for development against Base Sepolia.