Back

Fair Use Policy

How Standard Compute keeps unlimited AI compute reliable, fast, and fair for every customer on the platform.

What Fair Use Means

Standard Compute offers unlimited AI compute so you can build and run agent automations without worrying about per-token billing.

To keep the platform stable and responsive for everyone, Standard Compute applies a fair use system based on intelligent request batching, LLM routing, prompt compaction, priority scheduling, and adaptive throttling during heavy load.

These systems allow us to offer predictable pricing while maintaining high model quality and platform stability.

These plans are intended for personal use by individual developers. Typical use cases include experimentation, personal automations, prototype development, and learning projects. They are not intended to operate commercial services, SaaS products, or customer-facing applications at scale. Workloads supporting commercial services, external users, or large-scale integrations may require a custom enterprise agreement.

Platform Architecture

To make unlimited compute sustainable, Standard Compute uses several infrastructure optimizations.

Intelligent Batching

Requests from multiple users may be batched together when appropriate. This significantly improves GPU utilization and allows the platform to deliver unlimited compute economically.

Batching is more aggressive on lower-tier plans and reduced on higher-tier plans, resulting in faster response times.

LLM Routing

Standard Compute uses an internal LLM routing algorithm that dynamically selects the most efficient model configuration for each request while preserving output quality.

This allows the platform to deliver top-tier models at predictable cost.

Smart Prompt Compaction

Our Smart Compaction algorithm reduces unnecessary prompt tokens and optimizes request structure before execution. This improves throughput and reduces compute waste across the platform.

Adaptive Throttling

During periods of high demand, the system may apply temporary throttling to ensure fair distribution of shared resources.

Requests are normally not dropped and may instead be queued or experience increased latency depending on platform load and plan priority. However, in cases of sustained usage that violates fair use guidelines, excess requests may be throttled or dropped to protect platform stability.

Plan Tiers and Infrastructure Access

Each subscription tier is designed around a different execution profile and infrastructure priority.

Higher tiers receive faster scheduling, reduced batching latency, and larger execution pools, ensuring reliable performance for demanding agent workloads.

Unless otherwise agreed in writing, these plans are intended for personal developer use and non-commercial experimentation. Production deployments serving external customers may require an enterprise agreement.

Starter — $9 / month

Simple access for experimenting with agents. Includes a 3-day free trial, unlimited LLM compute, top-tier LLM models, slower execution speed, shared execution pool, heavily optimized batching, dynamic performance under load, and 1 API key.

Starter is optimized for personal experimentation and lightweight automation. Requests run in the shared execution pool with aggressive batching, which may increase latency during periods of high demand.

Standard — $39 / month

Balanced performance for everyday agent workflows. Includes a 3-day free trial, unlimited LLM compute, top-tier LLM models, standard execution speed, shared execution pool, optimized batching for efficiency, dynamic performance under load, and 1 API key.

Standard provides more consistent performance for regular personal workflows while still operating within the shared execution pool.

Fast — $99 / month

Faster execution for active and complex agent workflows. Includes a 3-day free trial, unlimited LLM compute, top-tier LLM models, faster execution speed, priority scheduling, higher-capacity execution pool, reduced batching latency, dynamic performance optimization under load, and 1 API key.

Fast users benefit from priority scheduling, meaning their requests are processed ahead of lower-tier traffic when the system experiences high load.

Turbo — $399 / month

Maximum responsiveness for demanding agent automation. Includes a 3-day free trial, unlimited LLM compute, top-tier LLM models, maximum execution speed, highest priority scheduling, high-capacity execution pool, minimal batching latency, optimized for sustained agent workloads, and 1 API key.

Turbo provides the highest scheduling priority and minimal batching delay, ensuring stable performance for high-volume workloads and advanced automation pipelines.

Fair Resource Sharing

Because Standard Compute operates a shared AI infrastructure, resources must be distributed fairly.

During periods of elevated demand, higher-tier plans receive priority scheduling, lower-tier plans may experience additional batching or queueing, and sustained excessive usage may trigger temporary throttling.

These mechanisms ensure that no single account can degrade platform performance for others.

Occasional bursts above normal usage are expected and supported. Fair use enforcement only applies when sustained activity significantly exceeds the intended usage profile for a plan.

Prohibited Usage

The following activities are not permitted on any plan: using the API as a general-purpose compute platform unrelated to AI inference (e.g., cryptocurrency mining), launching or coordinating denial-of-service attacks, large-scale scraping or data harvesting where the LLM is used only to extract third-party content, attempting to reverse engineer or benchmark upstream model providers in ways that violate their terms, and generating illegal content including fraud, harassment, or CSAM.

Violations may result in account suspension or termination.

Monitoring

To maintain system stability we monitor request volume, token throughput, and error rates.

Monitoring focuses on aggregate usage metrics. Standard Compute does not read or analyze prompt content.

Policy Updates

This policy may evolve as the platform grows. Any significant changes will be communicated to customers at least 14 days before taking effect.

For questions about this policy or specific use cases: contact@standardcompute.com