Resources
Rate Limits
Rate limits for the Inference.net API
Note: Inference Cloud is currently in beta. Some features may change.
API Rate Limits
Please contact us or use the support chat to request a higher rate limit.
- Language Models: 500 requests per minute
- Image Models: 100 requests per minute