Inference.net Documentation home page
Search...
⌘K
Ask AI
Support
Dashboard
Dashboard
Search...
Navigation
Resources
Rate Limits
Documentation
View Dashboard
Search Models
Get Started
Introduction
Quickstart
Features
Embeddings API
Batch API
Structured Outputs
Vision
Background Inference (Asynchronous API)
Fine-Tuning
Fine‑Tuning & Distillation
Use Cases
Classification
Image Captioning
Translation
Resources
Rate Limits
FAQ
Partner Program
On this page
API Rate Limits
Resources
Rate Limits
Rate limits for the Inference.net API
API Rate Limits
Please
contact us
or use the support chat to request a higher rate limit.
Language Models: 500 requests per minute
Image Models: 100 requests per minute
Translation
FAQ
Assistant
Responses are generated using AI and may contain mistakes.