Quick Start
Connect to Majordomo in minutes. Change one line of code and start logging requests.
How It Works
Learn how Majordomo proxies your LLM requests and what data goes where.
Connect Your Code
Integration examples for OpenAI, Anthropic, Gemini, and more SDKs.
Replay & Evals
Test a model switch against real production traffic before you commit.
What Majordomo gives you
Cost & Usage Tracking
Every request logged with model, tokens, cost, and latency. Break down spend by team, feature, environment, or user with custom metadata headers.
Replay
Run real production traffic against a candidate model. Get actual cost, latency, and quality numbers on your workload before you switch.
Evals
Build test suites from logged requests. Define scoring criteria. Run scored evaluations against any model before a change ships.
Multi-provider Routing
OpenAI, Anthropic, and Gemini from a single endpoint. Switch models in the dashboard without touching application code.
Get up and running
Create an account
Sign up at app.gomajordomo.com and create your first API key from the API Keys section.
Point your SDK at the gateway
Change the base URL to
https://gateway.gomajordomo.com/v1 and add your X-Majordomo-Key header. No other changes required.Verify in the dashboard
Send a request. It appears in the dashboard immediately with model, token counts, cost, and latency.
Two deployment modes
Managed — Majordomo runs the gateway. Point your SDK at the gateway endpoint, create an API key, and you’re logging requests within minutes. No infrastructure to operate. Self-hosted Steward — You run Steward inside your own VPC. Your prompts and completions never leave your infrastructure. Majordomo receives only metadata — token counts, cost, latency, model name. The right choice for teams with data residency requirements.Self-hosted Setup
Run Steward in your own VPC. Step-by-step setup with Docker, Postgres, and optional S3/GCS body storage.