Skip to main content
The Inference CLI is the fastest way to connect your app to Catalyst. It scans your codebase, finds your LLM clients, and updates them to route through the gateway with minimal manual work.
Install with AI works with OpenAI, Anthropic, Gemini, Groq, Cerebras, OpenRouter, LangChain and more.
1

Install the CLI

Install the Inference CLI globally.
npm install -g @inference/cli
2

Sign in

Sign in with your Inference account. Your browser will open to authenticate.
inf auth login
Running in CI or another headless environment? Use inf auth set-key instead of browser login.
3

Run instrumentation in your project

Navigate to your project root and run instrumentation.
cd /path/to/your/project && inf instrument
The command guides you through the following workflow:
  • Select a coding agent to use: Claude Code, OpenCode, or Codex.
  • Scan your codebase for LLM clients such as OpenAI, Anthropic, LangChain,etc
  • Redirect base URLs to the gateway
  • Add routing headers so requests are authenticated, forwarded, and traced
  • Add task IDs so each call site is grouped automatically in the dashboard
  • Review the generated changes before applying them
Run inf instrument --dry-run to preview changes without modifying any files.
4

Run your app

Run your application how you normally would to produce inference requests. Requests from your application are now routed through the gateway and will appear in the dashboard.
5

Verify it worked

Open the dashboard to see request details, traces, and analytics. You can also verify from the CLI:
inf inference list
Your app continues using the same provider SDKs you already have; the command updates your existing provider clients.
Add INFERENCE_API_KEY to your .env file so the instrumentation works across environments. Find your key in the dashboard under API Keys.

Supported AI coding agents

AgentBinary
Claude Codeclaude
OpenCodeopencode
Codexcodex

Supported providers

Built-in: OpenAI, Anthropic OpenAI-compatible via x-inference-provider-url: Google Gemini, Together AI, Groq, Fireworks AI, Mistral AI, Cerebras, Perplexity, DeepSeek, OpenRouter, Azure OpenAI, and any OpenAI-compatible endpoint.