AI APIs
70%+ cheaper

ChatGPT, Claude, Gemini — one endpoint, one key. Over 70% cheaper than direct access. Subscribe for even lower rates.

Instant
setup
Zero
logging
Crypto
payments
api.void-api.com — live
streaming
Connecting...

Unified API

One endpoint for GPT, Claude, and Gemini. Switch models with a single parameter.

Crypto Payments

Pay with cryptocurrency. Fast, private, no intermediaries.

Smart Caching

Repeated contexts cached automatically. Up to 90% savings on system prompts.

Zero Logging

We never store your prompts or completions. Only token counts for billing.

Auto Failover

If a cluster goes down, we switch to another instantly. Zero configuration.

Drop-in Replacement

Works with OpenAI, Anthropic, Google SDKs. Just change the base URL.

Models & Pricing

Per 1M tokens. All models on all plans.

ProviderModelInputOutputCachedContext
OpenAIgpt-5.2$0.60$1.75$4.50$14.00$0.06$0.17400k
OpenAIgpt-5.2-codex$0.60$1.75$4.50$14.00$0.06$0.17400k
OpenAIgpt-5.3-codex$0.60$1.75$4.50$14.00$0.06$0.17400k
OpenAIgpt-5.1-codex-mini$0.08$0.25$0.65$2.00$0.01$0.03400k
Anthropicclaude-4.6-sonnet$1.00$3.00$5.00$15.00$0.10 / 1.25$0.30 / 3.75200k
Anthropicclaude-4.6-opus$1.75$5.00$8.50$25.00$0.17 / 2.15$0.50 / 6.25200k
Googlegemini-3.1-pro$0.70$2.00$4.00$12.00$0.07$0.201M
Googlegemini-3-flash$0.15$0.50$1.00$3.00$0.01$0.051M

Pricing

Pay per token or subscribe for maximum savings.

Crypto payments
One-time

Weekend Pass

5 days of full access

$5
  • $35 API credits
  • 3 concurrent requests
  • All models
  • Priority Queue
Get started

Starter

For individual developers

$9/mo
  • $70 API credits
  • 1 concurrent request
  • All models
  • 50% off cached
  • Standard Queue
Get started
Popular

Pro

For teams & production

$19/mo
  • $200 API credits
  • 3 concurrent requests
  • All models
  • 65% off cached
  • Priority Queue
Get started

Agency

Pay as you go

PAYG
  • $10 minimum
  • Unlimited concurrency
  • All models
  • Express Queue
  • Volume discounts
Get started

Optimized for cost, not speed. Latency may be higher than direct API calls. Best for async workloads, batch processing, and non-time-critical tasks.

Integration Guide

Recommended setup for most apps: unified key + OpenAI-compatible endpoint.

Base URL: https://api.void-api.com/v1
Endpoint: /chat/completions
Model name picks provider: claude*, gemini*, gpt*

API key headers supported

Authorization: Bearer <key>, x-api-key, x-goog-api-key, query key / api_key.

Quick Start (Unified)
from openai import OpenAI

client = OpenAI(
    api_key="sk-va-unified-...",
    base_url="https://api.void-api.com/v1"
)

resp = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[{"role": "user", "content": "Hello"}]
)
Provider Routes
Unified (recommended):
  Base URL:    https://api.void-api.com/v1
  Request URL: https://api.void-api.com/v1/chat/completions

Anthropic (native):
  Base URL:    https://api.void-api.com
  Request URL: https://api.void-api.com/v1/messages

OpenAI-compatible:
  Base URL:    https://api.void-api.com/v1
  Request URL: https://api.void-api.com/v1/chat/completions

Google (native):
  Base URL:    https://api.void-api.com/google
  Request URL: https://api.void-api.com/google/v1beta/models/{model}:generateContent

Key compatibility:
- sk-va-unified-... works on all routes
- sk-va-anthropic-... works on Anthropic + /v1
- sk-va-google-... works on Google + /v1
- sk-va-openai-... works on OpenAI + /v1
Example: Cline
OpenAI-compatible mode:
  Provider: OpenAI Compatible
  Base URL: https://api.void-api.com/v1
  (Cline sends /chat/completions itself)
  Key: sk-va-unified-...

Anthropic mode:
  Provider: Anthropic
  Base URL: https://api.void-api.com
  (Cline sends /v1/messages itself)
  Key: sk-va-anthropic-... (or unified)

FAQ

Is it as fast as direct APIs?

No. We optimize for cost, not latency. Response times may vary and can be slower than direct API calls — sometimes significantly during high load. If you need consistent sub-1s latency, use official APIs directly.

Do you store my prompts?

No. We never log or store your prompts, completions, or any request content. We only track token counts for billing purposes. Your data passes through but is never persisted.

How does billing work?

Two options: Pay-as-you-go — no subscription, pay only for tokens used at standard rates. Subscription plans — monthly fee unlocks lower per-token rates and free cached tokens. Check the pricing table above for exact rates.

Can I use it in production?

Yes, but with caveats. We provide no SLA or uptime guarantees. Service may experience downtime or degraded performance without notice. Best suited for cost-sensitive workloads where occasional delays are acceptable — not for latency-critical or mission-critical systems.

Ready to save?

Get your API key in 30 seconds. Start with pay-as-you-go, upgrade when needed.

Create account