Praveen Thirumurugan a34aa17481 Update README.md
Signed-off-by: Praveen Thirumurugan <mail@praveent.com>
2025-12-23 12:48:44 +05:30
2025-12-17 05:25:55 +05:30
2025-10-18 01:22:02 +05:30
2025-12-23 12:48:44 +05:30
2025-10-18 01:22:02 +05:30

OpenProxy: Proxy for OpenAI/Anthropic-compatible APIs

OpenProxy is a lightweight LLM proxy server that seamlessly forwards API requests to OpenAI and Anthropic compatible endpoints with comprehensive logging, cost tracking, and PostgreSQL integration.

openproxy_cover-image

How to configure?

Setting up

I'd recommend forking this repository or cloning it directly. Once done, you should be able to get OpenProxy running with minimal configuration.

pnpm install

Set your environment variables:

export PORT=3007
export OPENAI_UPSTREAM_URL="https://api.example.com/v1"
export ANTHROPIC_UPSTREAM_URL="https://api.example.com/api/anthropic/v1"
export DATABASE_URL="postgresql://user:password@localhost:5432/database_name"

Start the server:

# Development mode with auto-reload
pnpm dev

# Production build
pnpm build && pnpm start

Configuration

Environment Variable Description Default
PORT Server port 3007
OPENAI_UPSTREAM_URL OpenAI-compatible endpoint URL -
ANTHROPIC_UPSTREAM_URL Anthropic-compatible endpoint URL -
DATABASE_URL PostgreSQL connection string Required

OpenProxy uses path prefixes for clean provider detection:

Proxy Path Routes To Auth Header
/openai/* OPENAI_UPSTREAM_URL/* Authorization: Bearer <key>
/anthropic/* ANTHROPIC_UPSTREAM_URL/* x-api-key: <key> or Authorization: Bearer <key>

PostgreSQL Logging

Every request is logged with comprehensive details to the PostgreSQL database. The table schema is as follows:

CREATE TABLE IF NOT EXISTS llm_proxy (
  timestamp TIMESTAMP WITH TIME ZONE DEFAULT CURRENT_TIMESTAMP,
  request_method VARCHAR(10) NOT NULL,
  request_path VARCHAR(255) NOT NULL,
  provider TEXT,
  model VARCHAR(50) NOT NULL,
  completion_tokens INTEGER,
  prompt_tokens INTEGER,
  total_tokens INTEGER,
  cached_tokens INTEGER,
  total_cost NUMERIC,
  response_time INTEGER,
  request_body JSONB,
  response_body JSONB,
  response_status INTEGER,
  provider_url VARCHAR(500),
  client_ip INET,
  user_agent TEXT,
  request_size INTEGER,
  response_size INTEGER,
  stream BOOLEAN,
  temperature REAL,
  max_tokens INTEGER,
  request_id UUID
);

Cost Calculation

OpenProxy automatically calculates costs based on model and token usage using the Helicone API. You can customize the costs for your own models in cost.ts.

How to use?

Using with Claude Code

For example, to use Z.AI or other Anthropic-compatible providers with Claude Code:

export ANTHROPIC_UPSTREAM_URL="https://api.z.ai/api/anthropic"
export DATABASE_URL="postgresql://user:password@localhost:5432/database_name"
pnpm dev

# Configure Claude Code to use:
# API Base URL: http://localhost:3007/anthropic

Using with OpenAI-compatible clients

For example, to use Z.AI or other OpenAI-compatible providers with OpenAI-compatible clients:

export OPENAI_UPSTREAM_URL="https://api.z.ai/api/coding/paas/v4"
export DATABASE_URL="postgresql://user:password@localhost:5432/database_name"
pnpm dev

# Configure your client to use:
# API Base URL: http://localhost:3007/openai

Metrics Dashboard

OpenProxy includes a lightweight Next.js dashboard for real-time metrics visualization. The dashboard is accessible at http://localhost:3008. To run the dashboard, run the following command:

pnpm --filter dashboard dev

Final notes

Get started today and for assistance, you can reach me on GitHub or X. PRs are always welcome if you'd like to add features or fix bugs!

This project is open source and available under the MIT License.

Description
Languages
TypeScript 98.4%
Dockerfile 1.1%
JavaScript 0.5%