Skip to main content
Inference.net is designed to support production workloads without treating captured request data casually.

Core principles

  • Request data is not used for model training by default
  • Secrets and similar sensitive values are stripped where possible
  • Platform data is encrypted in transit and at rest
  • Retention should match the operational need of the workflow

Direct API vs Observe

The direct API and Observe are different product paths, but the same general rule applies: only keep what is operationally useful, and use project-level controls and data curation intentionally. For the workflow-first entry point into traffic capture, start with Integrate with Your LLM Provider.
  • Use environments and task IDs to segment traffic
  • Create long-lived datasets only for the examples you want to preserve
  • Review retention expectations before broad production rollout

Need a specific retention policy?

If you need a specific policy, no-retention handling, or help mapping the platform into your internal compliance requirements, meet with our team.