Changelog
All notable changes to BlockRun.
[2026-04-25]
Models
- Added
openai/gpt-5.5— released 2026-04-23, first fully retrained base since GPT-4.5. 1M context, 128K output, native agent + computer use. Pricing: $5.00 / $30.00 per 1M tokens. - Promoted
openai/gpt-5.5toFEATURED_MODEL_IDS(replacesopenai/gpt-5.4as the OpenAI flagship surfaced on the homepage and discover pages).
SDK Updates
- Python SDK (
blockrun-llm) 0.17.0:PREMIUM_TIERS["MEDIUM"]primary nowopenai/gpt-5.5;gpt-5.4demoted to first fallback.estimate_costbaseline rebased from $2.50/$15 to $5.00/$30 to match the new flagship.__version__andVERSIONreconciled (was 0.16.1 vs 0.15.0). - TypeScript SDK (
@blockrun/llm) 1.11.0: catalog refreshed withopenai/gpt-5.5.VERSIONandpackage.jsonreconciled (was 1.9.0 vs 1.10.1). - Go SDK (
blockrun-llm-go) 0.4.0:RoutingPremium.TierMediumprimary nowopenai/gpt-5.5(wasgpt-5.4).
[2026-04-17]
Added — xAI Grok Imagine
- Video generation — new endpoint
POST /api/v1/videos/generationswithxai/grok-imagine-video($0.050/sec, 8-second clips, text or image-to-video). The endpoint handles xAI's async job + polling internally so callers get the MP4 URL in a single request. See Video Generation. - Grok Imagine image models —
xai/grok-imagine-image($0.02/image) andxai/grok-imagine-image-pro($0.07/image) added toPOST /api/v1/images/generations. - Grok 4.20 chat models (hidden, API-routable) —
xai/grok-4.20-reasoning,xai/grok-4.20-non-reasoning,xai/grok-4.20-multi-agent. 2M context, vision-capable, $2/M in · $6/M out.
Added — Permanent media URLs via GCS backup
- Every generated image and video is now mirrored to BlockRun's Google Cloud Storage bucket and returned through a blockrun-hosted proxy URL (
/api/media/...). Upstream URLs fromimgen.x.aiandvidgen.x.aiare temporary — the returnedurlis permanent. - Image/video responses now include:
url— permanent proxy URL (falls back to upstream if the backup step fails)source_url— original upstream URL (for debugging / comparison)backed_up— boolean indicating whether GCS mirroring succeeded
API Changes
GET /api/v1/modelsnow includes video models withbilling_mode: "per_second"andpricing: { per_second, default_duration_seconds, max_duration_seconds }.
[2026-03-24]
API Changes
/api/v1/modelsnow returns extended metadata for each model:name,description,context_window,max_output, andcategories- Model categories indicate capabilities:
chat,reasoning,coding,vision
Models
- 33 visible models across 9 providers (Base), 31 on Solana
- Added GPT-5.4 Mini ($0.75/$4.50, 400K context)
- Added GPT-5.4 Nano, GPT-5.3, GPT-5.3 Codex
- Added Claude Opus 4.6 (1M context), Claude Sonnet 4.6
- Added Gemini 3.1 Pro, Gemini 3 Pro Preview, Gemini 3.1 Flash Lite
- Added NVIDIA Kimi K2.5 (free tier)
- Removed xAI/Grok, Qwen, and legacy GPT-4 family from visible lineup (hidden, still routable via API)
SDK Updates
- TypeScript SDK (
@blockrun/llm) v1.4.3: Model type now includescategories,contextWindow,maxOutput, properly mapped from API - MCP Server (
@blockrun/mcp) v0.4.2:blockrun_modelsshows context window and categories, smart routing updated to current models
MCP Smart Routing
fast: Gemini Flash, GPT-5 Minibalanced: GPT-5.4, Claude Sonnet 4.6powerful: GPT-5.4, Claude Opus 4.6cheap: NVIDIA free models, DeepSeek, Gemini Flashreasoning: o3, o1, DeepSeek Reasoner
[Unreleased]
Added
- Continue Integration — ClawRouter merged as a native provider in Continue (32K+ ⭐), the open-source AI code assistant. Users can now select ClawRouter directly in Continue's provider settings for cost-optimized model routing. (PR #11751)
Coming Soon
- Streaming support
- Additional AI providers
[1.0.0] - 2025-12
Added
- Initial public release
- x402 v2 protocol support
- OpenAI models (GPT-4o, GPT-4o Mini, o1, o1-mini)
- Anthropic models (Claude Opus 4, Sonnet 4, Haiku 4.5)
- Google models (Gemini 2.5 Pro, Gemini 2.5 Flash)
- xAI models (Grok 4)
- DeepSeek models (V3, R1)
- Meta models via Together.ai (Llama 3.3 70B)
- Qwen models via Together.ai (Qwen 2.5 72B)
- Mistral models (Mistral Large)
- Python SDK (
blockrun-llm) - TypeScript SDK (
@blockrun/llm) - Base network support (USDC)
- CDP Facilitator integration
Technical
- EIP-3009 TransferWithAuthorization for gasless payments
- EIP-712 typed data signing
- Automatic retry with payment on 402 response
- OpenAI-compatible Chat Completions API
SDK Versions
Python SDK (blockrun-llm)
[0.1.0] - 2025-12
- Initial release
LLMClientwithchat()andchat_completion()methodsAsyncLLMClientfor async usage- x402 v2 protocol support
- Error handling with
APIErrorandPaymentError
TypeScript SDK (@blockrun/llm)
[0.1.0] - 2025-12
- Initial release
LLMClientclass with full TypeScript typeschat()andchatCompletion()methodslistModels()for available models- x402 v2 protocol support
- Error handling with
APIErrorandPaymentError
Version Policy
BlockRun follows Semantic Versioning:
- MAJOR - Breaking API changes
- MINOR - New features, backward compatible
- PATCH - Bug fixes, backward compatible
Release Notes
For detailed release notes, see the GitHub releases page for each repository.