model value in normal Inference.net API requests.
What stays the same
Calling a deployed model uses the same core API shape as the direct serverless API:- base URL:
https://api.inference.net/v1 - auth:
Authorization: Bearer ... - standard request bodies for chat completions and related endpoints
What changes
The main difference is the model name. Instead of a catalog model identifier, you use the deployment’s public model identifier from the deployment overview page.Source-backed API examples
The dashboard generates these examples frominference/apps/web/src/components/deployments/DeploymentApiExamples.tsx using the same snippet generator as the direct quickstart.
Recommended workflow
- open the deployment overview
- copy the public model identifier or API example
- send a smoke test against the normal API endpoint using that model value
- inspect the request in the deployment’s inferences tab
Best practices
- keep a simple smoke test for each deployment
- verify both correctness and latency before routing real traffic
- continue observing production traffic after rollout