What you’ll have when you finish
- one real request routed through Inference.net
- provider, model, cost, and latency visible in the dashboard
- enough metadata to start building datasets, evals, and training workflows
Before you start
- an existing app that already calls OpenAI, Anthropic, or another OpenAI-compatible provider
- your upstream provider API key
- an Inference.net project and observability project key
Step 1: choose the setup path
CLI path
Use
inf install if you want the fastest self-serve setup.Manual path
Keep full control over the SDK config and proxy headers.
Step 2: route one request through Inference.net
CLI path
- Install the CLI with /cli/install.
- Authenticate with
inf auth login. - Run
inf installinside your application. - Send one normal request from your app.
Manual path
For manual integration, the only essential changes are:- point your client at
https://api.inference.net/v1 - keep your upstream provider auth exactly as it is today
- add
x-inference-provider - add
x-inference-observability-api-key
Step 3: verify the first observed request
Open the dashboard and confirm that the request shows:- upstream provider
- model name
- environment and task
- duration
- total, input, and output tokens
- cost
Step 4: add just enough metadata to make the traffic useful
Start with these headers:x-inference-environmentx-inference-task-idx-inference-metadata-*
What to do next
Create Datasets from Observed Traffic
Turn live traffic into the eval and training datasets that power the rest of the platform.
Meet with Us
Talk to our team if you want help migrating a larger production workflow.