Deploy any supported Hugging Face model on Inference.net without managing your own infrastructure. This guide walks through selecting a model, creating a deployment, and verifying the endpoint.Documentation Index
Fetch the complete documentation index at: https://docs.inference.net/llms.txt
Use this file to discover all available pages before exploring further.
What you’ll have when you finish
- one dedicated deployment serving a Hugging Face model
- a public model identifier you can use with the standard API shape
- a verified smoke test confirming the model is responding
Before you start
- an Inference.net account with deployment access
- the Hugging Face model ID you want to deploy (e.g.,
meta-llama/Llama-3.2-1B-Instruct)
Step 1: Choose the model
Browse the model catalog or use a Hugging Face model ID directly when creating a deployment.Step 2: Create the deployment
Follow the same deployment creation flow described in Deploy a Trained Model. The key difference is that you specify a Hugging Face model ID instead of a trained model reference.Step 3: Verify the endpoint
Send a smoke test request using the deployment’s public model identifier:Next steps
API Quickstart
Learn the full API surface for calling deployed models.
Manage and Monitor
Understand deployment operations, billing, and scaling.