Relay.
[ China-direct · No VPN required ]

One key,
all major AI models.

Claude · GPT · DeepSeek · Doubao · Qwen.
Swap one base_url and you are in. Compatible with both Anthropic and OpenAI SDKs. Pay-as-you-go.

See integration

// Get an API key on signup. Pay-as-you-go, no upfront fee.

// FEATURES

Why Relay

Use the world's top AI models from China as if they were native APIs.

01

China-direct

Servers run on routes reachable from China. No VPN, no proxy tools, low latency, stable connection.

02

Dual protocol

Supports both Anthropic SDK (/v1/messages) and OpenAI SDK (/v1/chat/completions). ChatBox, LobeChat, Cursor — just change the base URL.

03

Multi-vendor models

Covers Claude Sonnet 4.5 / Opus 4.7 / Haiku 4.5, GPT-5.5 / 5.4 series, DeepSeek V4 Flash/Pro, Doubao, Qwen full series. Routed by model name.

04

Reliable

Multi-key load balancing on the upstream side. Auto-failover when a single key has issues.

// INTEGRATION

One line to integrate

Use whichever SDK fits — Relay routes to the right upstream automatically. One key, one base_url change, the rest of your code stays untouched.

  • Dual protocol: Anthropic /v1/messages + OpenAI /v1/chat/completions
  • Native support for major SDKs: anthropic / openai / langchain / autogen
  • Works with ChatBox, LobeChat, Cherry Studio, Cursor and more
  • Streaming (SSE) fully supported, error format matches the official one

// CLIENTS

Using a client app?

In your client settings, change the API endpoint to Relay and paste your Relay key. No coding required.

anthropic_sdk.py
# Anthropic SDK — call Claude models
import anthropic

client = anthropic.Anthropic(
    base_url="https://your-relay.com",
    api_key="rk-your-relay-key",
)
message = client.messages.create(
    model="claude-sonnet-4-5-20250929",
    max_tokens=1024,
    messages=[{"role": "user", "content": "hello!"}]
)
openai_sdk.py
# OpenAI SDK — call GPT / DeepSeek / Doubao
from openai import OpenAI

client = OpenAI(
    base_url="https://your-relay.com/v1",
    api_key="rk-your-relay-key",
)
resp = client.chat.completions.create(
    model="deepseek-v4-flash", # or gpt-5.4, doubao-seed-2-0-pro-260215
    messages=[{"role": "user", "content": "hello!"}]
)
// PRICING

Transparent usage-based pricing

Pay for what you use. No monthly fee, no minimum spend.

DeepSeek V4 Flash Best value
DeepSeek
Input $0.17
Output $0.34
Qwen Turbo CN model
Qwen
Input $0.05
Output $0.11
Claude Sonnet 4.5 Recommended
Anthropic
Input $3.60
Output $18.00
GPT-5.4 General
OpenAI
Input $3.00
Output $18.00
Claude Opus 4.7 Flagship
Anthropic
Input $6.00
Output $30.00

// Sample × 1.3. MTok = million tokens. Actual price applies at billing time.

  • Sign up and start, no prepayment
  • One key works across multiple model vendors
  • Real-time usage view, multi-dimensional analytics