Files
openrelay/README.md
2026-05-01 12:57:58 +08:00

18 KiB
Raw Permalink Blame History

OpenRelay

OpenRelay

聚合本机与直连 AI 配额,一键接入本地项目

Latest Release License Platform Community Forum Telegram Chat

中文 | English


痛点

你的 AI 订阅,各自为政。

Claude Pro 只能在 Claude Desktop 用。Kiro 配额只能在 Kiro 用。Groq 免费但每个工具都要手动配置。Cursor 500 次用完了,你只能停下来。

OpenRelay 打破这道墙。

  • 帮你发现和接入更多可用 AI 模型配额Groq、Cerebras、SambaNova、Gemini 等,按供应商账户实际可用)
  • 帮你把免费或收费的配额接入你正在使用的 AI 工具
  • 一键配置 Claude Code、OpenClaw、Aider、Goose 等所有工具的模型

演示

OpenRelay 演示

Provider 面板
自动发现的 Provider 和配额状态
Work — CLI 工具配置
一键配置 Claude Code、OpenClaw、Aider...
IDE RPC 代理
IDE 代理——Cursor、Windsurf、VS Code Copilot
自定义模型组
自定义模型组,自动故障转移

OpenRelay 能做什么

1. 自动发现你所有的 AI 配额

启动 OpenRelay它会接入你已经拥有的 AI 来源 — Claude Desktop、Claude Code、Kiro、Windsurf、Antigravity、OpenCode、VS Code Copilot、OpenAI Codex、Gemini CLI、Rovo Dev、QClaw 等。已发现的本地配额可在面板中统一管理;部分来源需要先登录原应用或配置 API Key。

支持 34 个直连 API 或本地端点Groq、Gemini API、DeepSeek、Mistral、OpenRouter、LongCat、千帆、七牛、Anthropic API、Ollama 等)— 按供应商要求配置 API Key 或端点后可复用。

45 个非虚拟提供商。一个面板。一个端点。

2. 任意配额用在任意工具

你的 Claude Pro 订阅现在可以驱动 Claude Code、Aider、Continue、Goose、Amp或任何支持 Anthropic/OpenAI API 的工具:

# Windows (PowerShell)
$env:ANTHROPIC_BASE_URL="http://localhost:18765"
$env:ANTHROPIC_API_KEY="unused"
# macOS / Linux
export ANTHROPIC_BASE_URL=http://localhost:18765
export ANTHROPIC_API_KEY=unused

搞定。Claude Code 现在使用你的 Claude Desktop 配额。

想在 Aider 里用 Kiro 的免费 Claude Sonnet改一下 URL

ANTHROPIC_BASE_URL=http://localhost:18765/kiro

3. 一键配置所有 CLI 工具

不再手动编辑 .zshrc,不再来回倒腾环境变量。打开 Web 面板,选择 Provider点一下开关

  • Claude Code → 走 Kiro按你的 Kiro 账户额度)
  • Aider → 走 Groq低延迟推理
  • Goose → 走 Gemini API大上下文模型
  • OpenCode → 走 DeepSeek最便宜的编程模型

重开终端,完事。每个工具都配好了。

4. 给你的 IDE 无缝接入外部配额

Cursor 配额烧完了Windsurf 额度用光了?别停下编码 — 无缝插入任何其他配额来源:

IDE 接入方式 效果
Cursor RPC 代理 (ConnectRPC, HTTP/2) 在 Cursor 里用 Claude/Kiro/Groq/任意 Provider
Windsurf RPC 代理 (ConnectRPC) 用任意 Provider 替换 Windsurf 内置模型
VS Code Copilot Ollama BYOK 桥接 用任意模型作为 Copilot 后端
Antigravity Gemini REST 代理 通过任意 Provider 路由

在面板启动代理IDE 无感切换。

5. 组合配额,减少手动切换

把多个 Provider 的配额合并成一个虚拟模型:

"fast-group" = Groq (Llama 90B) + Cerebras (Llama 70B) + SambaNova (Llama 405B)

Groq 额度不可用 → 自动切到 Cerebras → 再切 SambaNova。跨 Provider 轮询和故障转移会继续使用你配置的可用额度,减少手动切换。


安装 & 快速上手

直接下载可执行文件运行,无需 npm install / node 环境。

Windows点击下载 openrelay-windows-x64.exe,双击运行。

macOSIntel / Apple Silicon 通用):点击下载 openrelay-macos,然后终端执行:

chmod +x openrelay-macos
xattr -d com.apple.quarantine openrelay-macos
./openrelay-macos

xattr 命令用于解除 macOS 对未签名程序的安全限制,否则会提示"无法打开"。

Linux

然后终端执行对应文件:

# x64
chmod +x openrelay-linux-x64
./openrelay-linux-x64

# ARM64 / aarch64
chmod +x openrelay-linux-arm64
./openrelay-linux-arm64

Linux 支持的本地/CLI ProviderClaude Code、Kiro、Windsurf、OpenCode、VS Code Copilot、OpenAI Codex、Gemini CLI、Rovo Dev。Claude Desktop 和 Antigravity 目前无 Linux 版本。QClaw 取决于桌面应用和本地 gateway可降级运行。凭证存取通过 secret-toolgnome-keyring或文件缓存。

浏览器打开 http://localhost:18765 — 一切在 Web 面板中管理,支持中英双语。

觉得有用?给个 Star 是对我们最大的支持!


安全

凭据留在本机 — 应用 token/cookie 从你的机器读取,只用于连接原供应商。通过 OpenRelay 添加的 API Key 存储在本机 ~/.openrelay/ 配置中。

直连 AI 后端 — AI 请求从你的机器直接发送到所选 ProviderOpenRelay 官方服务器不在请求链路中。

默认不记录提示词内容 — 消息内容默认不会写入日志、缓存或持久化;只有你显式开启本地 request-shape 调试时才会输出排障信息。

最小产品网络请求 — 许可证和更新检查可能访问 OpenRelay/GitHub 端点,但不会携带 Provider 凭据或对话内容。

可审计 — 凭据处理代码(cookie.ts)可查看审计。

详见 DISCLAIMER.mdPRIVACY.md

常见问题

遇到问题?查看 常见问题 (中文) | FAQ (English)

社区

点个 Star + 加入 社区Telegram 群,有机会领 1 个月 Pro 体验码

许可证

Open Core 模式:

  • 框架部分(代理、格式转换、配置):MIT
  • Pro 功能(模型组合、更高请求量上限):商业授权

中文 | English

The Problem

Your AI subscriptions are locked in silos.

Claude Pro only works in Claude Desktop. Kiro quota only works in Kiro. Groq is free but you have to configure every tool manually. Cursor burned through your 500 requests and now you're stuck.

OpenRelay breaks the silos.

  • Find and connect more usable AI quota (Groq, Cerebras, SambaNova, Gemini, and others, depending on your provider accounts)
  • Connect any quota to any tool you're already using
  • One-click configure Claude Code, OpenClaw, Aider, Goose, and more

Demo

OpenRelay Demo

Provider Dashboard
Auto-discovered providers & quota status
Work — CLI tool config
One-click configure Claude Code, OpenClaw, Aider...
IDE RPC Proxies
IDE proxies — Cursor, Windsurf, VS Code Copilot
Custom Model Groups
Custom model groups with auto-failover

What OpenRelay Actually Does

1. Auto-discover all your AI quotas

Launch OpenRelay and it connects the AI sources already available to you — Claude Desktop, Claude Code, Kiro, Windsurf, Antigravity, OpenCode, VS Code Copilot, OpenAI Codex, Gemini CLI, Rovo Dev, QClaw, and more. Discovered local quotas are managed from one dashboard; some sources require logging in to the original app or configuring an API key first.

Plus 34 direct API or local providers (Groq, Gemini API, DeepSeek, Mistral, OpenRouter, LongCat, Qianfan, Qiniu, Anthropic API, Ollama, and more) — enter an API key or endpoint once and it can be reused across tools.

45 non-virtual providers. One dashboard. One endpoint.

2. Use any quota in any tool

Your Claude Pro subscription can now power Claude Code, Aider, Continue, Goose, Amp, or any tool that speaks Anthropic/OpenAI API:

# Windows (PowerShell)
$env:ANTHROPIC_BASE_URL="http://localhost:18765"
$env:ANTHROPIC_API_KEY="unused"
# macOS / Linux
export ANTHROPIC_BASE_URL=http://localhost:18765
export ANTHROPIC_API_KEY=unused

That's it. Claude Code now uses your Claude Desktop quota.

Want to use Kiro's free Claude Sonnet quota in Aider? Just change the URL:

ANTHROPIC_BASE_URL=http://localhost:18765/kiro

3. One-click setup for every CLI tool

No more editing .zshrc or juggling environment variables. Open the Web dashboard, pick a provider for each tool, flip a switch:

  • Claude Code → route through Kiro (using your Kiro account quota)
  • Aider → route through Groq (low-latency inference)
  • Goose → route through Gemini API (large-context models)
  • OpenCode → route through DeepSeek (cheapest coding model)

Reopen your terminal. Done. Every tool is configured.

4. Supercharge your IDE with external quotas

Cursor quota burned through? Windsurf credits gone? Don't stop coding — seamlessly plug in any other quota source:

IDE How it works What you get
Cursor RPC proxy (ConnectRPC, HTTP/2) Use Claude/Kiro/Groq/any provider inside Cursor
Windsurf RPC proxy (ConnectRPC) Replace Windsurf's built-in models
VS Code Copilot Ollama BYOK bridge Use any model as a Copilot backend
Antigravity Gemini REST proxy Route through any provider

Start the proxy from the dashboard. Your IDE doesn't know the difference.

5. Combine quotas with failover

Take quotas from multiple providers and merge them into a single virtual model:

"fast-group" = Groq (Llama 90B) + Cerebras (Llama 70B) + SambaNova (Llama 405B)

When Groq quota is unavailable → automatic failover to Cerebras → then SambaNova. Round-robin and failover keep using the providers you configured while quota remains, reducing manual switching.


Supported Providers (45 non-virtual)

Local / CLI / IDE Providers (11)

These use local app sessions, CLI auth files, or local gateways when available.

Provider Credential source Notes
Claude Desktop Local Claude Desktop session Claude Pro/Max account quota
Claude Code Claude Code credentials Claude account quota
Kiro (AWS) Kiro app session Kiro account quota
Windsurf (Codeium) Windsurf session IDE quota and models
Antigravity Antigravity app session Gemini-compatible route
OpenCode OpenCode local config Built-in route
VS Code Copilot VS Code / GitHub Copilot session Copilot account quota
OpenAI Codex Codex local auth REST + WebSocket transport
Gemini CLI ~/.gemini/oauth_creds.json Gemini CLI OAuth
Rovo Dev Atlassian / Rovo Dev config or env Rovo account quota
QClaw QClaw local gateway Agent gateway; best for QClaw workflows

API / Local Providers (34)

These use your provider API key, provider account, or local endpoint. Quotas and free tiers are controlled by each upstream provider and can change.

Provider Type
Groq OpenAI-compatible API
Cerebras OpenAI-compatible API
OpenRouter OpenAI-compatible API
SambaNova OpenAI-compatible API
Gemini API OpenAI-compatible API
Mistral OpenAI-compatible API
xAI OpenAI-compatible API
SiliconFlow OpenAI-compatible API
Zhipu / GLM OpenAI-compatible API
Together AI OpenAI-compatible API
DashScope OpenAI-compatible API
DeepSeek OpenAI-compatible API
NVIDIA NIM OpenAI-compatible API
GitHub Models OpenAI-compatible API
Fireworks OpenAI-compatible API
Volcengine OpenAI-compatible API
Qianfan OpenAI-compatible API
Qiniu OpenAI-compatible API
Moonshot OpenAI-compatible API
Baichuan OpenAI-compatible API
Stepfun OpenAI-compatible API
MiniMax OpenAI-compatible API
Hunyuan OpenAI-compatible API
Cloudflare AI OpenAI-compatible API
HuggingFace OpenAI-compatible API
LongCat OpenAI-compatible API
Kilo OpenAI-compatible API
LLM7 OpenAI-compatible API
Vercel AI Gateway OpenAI-compatible API
BlazeAPI OpenAI-compatible API
Pollinations OpenAI-compatible API
BazaarLink OpenAI-compatible API
Anthropic API Native Anthropic API
Ollama Local endpoint

Install & Quick Start

Download and run — no npm install or Node.js required.

Windows: Download openrelay-windows-x64.exe, double-click to run.

macOS (Intel / Apple Silicon universal): Download openrelay-macos, then run in terminal:

chmod +x openrelay-macos
xattr -d com.apple.quarantine openrelay-macos
./openrelay-macos

xattr removes macOS Gatekeeper quarantine flag — without it, macOS will block the unsigned binary.

Linux:

Then run the matching file in terminal:

# x64
chmod +x openrelay-linux-x64
./openrelay-linux-x64

# ARM64 / aarch64
chmod +x openrelay-linux-arm64
./openrelay-linux-arm64

Supported local / CLI providers on Linux: Claude Code, Kiro, Windsurf, OpenCode, VS Code Copilot, OpenAI Codex, Gemini CLI, and Rovo Dev. Claude Desktop and Antigravity have no Linux builds. QClaw depends on its desktop app and local gateway, with degraded behavior where unavailable. Credentials are stored via secret-tool (gnome-keyring) or file-based cache.

Open http://localhost:18765 → everything is managed from the Web dashboard.

Found it useful? Give us a Star — it helps a lot!


Security

Credentials stay local. App tokens/cookies are read from your machine and used only to authenticate with their original provider. API keys you add in OpenRelay are stored locally under ~/.openrelay/.

Direct provider connections. AI requests go from your machine to the selected AI provider. OpenRelay's own servers are not in the request path.

No prompt logging by default. Message content is not logged, cached, or persisted unless you explicitly enable request-shape debugging for local troubleshooting.

Minimal product network calls. License and update checks may contact OpenRelay/GitHub endpoints, but they do not include provider credentials or conversation content.

Auditable. The credential handling code (cookie.ts) is available for security review.

See DISCLAIMER.md and PRIVACY.md for details.

FAQ

Having trouble? Check the FAQ (English) | 常见问题 (中文) for solutions to common issues.

Community

Star this repo + join forum or Telegram group → chance to get 1-2 months Pro free.

License

Open Core model:

  • Framework (proxy, format translation, config): MIT
  • Pro features (custom model groups, higher request limits): Commercial