Skip to main content
Route your Groq requests through the Inference Catalyst gateway to get cost tracking, latency monitoring, and analytics. Groq is OpenAI-compatible, so you use the OpenAI SDK with the x-inference-provider-url header to specify Groq’s base URL.
Prefer automatic setup? Run inf instrument to instrument your codebase in seconds. Learn more

Setup

1

Get your API keys

You need two keys:
  • Inference Catalyst project API key — from your dashboard under API Keys
  • Groq API key — from your Groq console
2

Set environment variables

export INFERENCE_API_KEY=<your-project-api-key>
export GROQ_API_KEY=<your-groq-api-key>
3

Update your code

Point the SDK at the gateway. Your project API key goes in apiKey, and the x-inference-provider-url header tells the gateway to forward requests to Groq.
import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://api.inference.net/v1",
  apiKey: process.env.INFERENCE_API_KEY,
  defaultHeaders: {
    "x-inference-provider-api-key": process.env.GROQ_API_KEY,
    "x-inference-provider-url": "https://api.groq.com/openai/v1",
    "x-inference-environment": process.env.NODE_ENV,
  },
});

const response = await client.chat.completions.create({
  model: "llama-3.3-70b-versatile",
  messages: [{ role: "user", content: "Hello" }],
}, {
  headers: { "x-inference-task-id": "default" },
});