Skip to main content
At io.net, we believe AI access should be simple, transparent, and scalable. Whether you are experimenting with ideas, running daily creative workflows, or deploying large-scale automation. Our platform connects every capability, this includes text, images, code, data, and voice, all in one workspace. You can start free, upgrade as your usage grows, and scale on demand without worrying about token tracking or hidden fees. Each plan provides access to the same advanced models and tools. Usage is measured in credits, which refresh automatically. Pay-as-you-go (PAYG) is available when you exceed your daily credits quota.
  • Standard (default): Explore and learn with free, light daily access. Pay-as-you-go (PAYG) pricing is applied for usage beyond the daily limit.
  • Professional: Includes $15 in monthly usage credits, refreshed daily to provide steady, predictable access for light coding projects and applications.
  • Developer: Includes $150 in monthly usage credits, refreshed every 8 hours to support continuous development and production workloads.

Plan Overview

Each plan includes a fixed daily or hourly allowance that refreshes automatically, so you can focus on your work instead of tracking tokens or costs.
PlanUsageRefresh cycleIdeal for
StandardContinuous access with PAYG.No refreshes, pay only for what you use.Teams that need flexible scaling.
Professional$15 in monthly usage credits.Once every 24 hours.Coding projects or light applications.
Developer$150 in monthly usage credits.Every 8 hours (3 x per day).Builders, teams, and automations.

How Usage Works

  • Professional plans refresh once every 24 hours for predictable, worry-free access.
  • Developer plans refresh every 8 hours, designed for continuous work or API usage.
  • If you hit your allowance, your access will pause until the next refresh, unless you have IO Credits.
  • IO Credits (Pay-As-You-Go) allow instant continuation beyond limits, charging per request through your connected payment method.

More on Pay-As-You-Go

  • Billed directly to your IO Credits balance.
  • Includes the same tools and models as subscription plans.
  • The Developer plan offers roughly a 10% discount compared to PAYG for consistent high-volume users.
  • PAYG stops when your plan refreshes.
  • Enables precise accounting of model-level usage and costs.
For reference, see the Model Rate Overview below for the latest per-model pricing structure.

Model Rate Overview

Choose the model that best fits your workflow, from lightweight reasoning to complex multimodal generation. Each model consumes credits based on its scale, capability, and computational cost. The table below shows the current per-model usage rates available across all IO Intelligence plans. These rates apply to both Unified Chat and API requests and reflect the approximate cost.
All prices are subject to change as model performance, infrastructure efficiency, and market rates evolve.
ModelInput/M($)Output/M($)Note
deepseek-ai/DeepSeek-V3.2$0.28$0.42Per 1M tokens.
moonshotai/Kimi-K2-Thinking$0.45$2.35Per 1M tokens.
zai-org/GLM-4.6$0.40$1.75Per 1M tokens.
zai-org/GLM-4.7$0.16$0.80Per 1M tokens.
moonshotai/Kimi-K2-Instruct-0905$0.39$1.90Per 1M tokens.
deepseek-ai/DeepSeek-R1-0528$0.40$1.75Per 1M tokens.
meta-llama/Llama-3.3-70B-Instruct$0.13$0.38Per 1M tokens.
mistralai/Devstral-Small-2505$0.05$0.22Per 1M tokens.
Qwen/Qwen2.5-VL-32B-Instruct$0.05$0.22Per 1M tokens.
meta-llama/Llama-3.2-90B-Vision-Instruct$0.35$0.40Per 1M tokens.
openai/gpt-oss-120b$0.04$0.40Per 1M tokens.
Qwen/Qwen3-235B-A22B-Thinking-2507$0.11$0.60Per 1M tokens.
openai/gpt-oss-20b$0.03$0.14Per 1M tokens.
mistralai/Mistral-Nemo-Instruct-2407$0.02$0.04Per 1M tokens.
Qwen/Qwen3-Next-80B-A3B-Instruct$0.10$0.80Per 1M tokens.
meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8$0.15$0.60Per 1M tokens.
mistralai/Mistral-Large-Instruct-2411$2.00$6.00Per 1M tokens.
Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar$0.22$0.95Per 1M tokens.
BAAI/bge-multilingual-gemma2$0.01Input based cost only.Per 1M tokens.
speaches-ai/piper-en_US-lessac-medium$0.62Input based cost.Per 1M tokens.
speaches-ai/piper-en_US-amy-low$0.62Input based cost.Per 1M tokens.
speaches-ai/Kokoro-82M-v1.0-ONNX-fp16$0.62Input based cost.Per 1M tokens.
speaches-ai/piper-en_US-ryan-high$0.62Input based cost.Per 1M tokens.
deepdml/faster-whisper-large-v3-turbo-ct2$0.04Input based cost.Per 1 hour.
Systran/faster-whisper-large-v3$0.04Input based cost.Per 1 hour.
guillaumekln/faster-whisper-small.en$0.04Input based cost.Per 1 hour.
guillaumekln/faster-whisper-medium.en$0.04Input based cost.Per 1 hour.
To verify the latest model pricing, use the GET /models API endpoint. The response includes detailed pricing information for each available model. The fields "input_token_price" and "output_token_price" represent the respective costs per token for input and output usage. For implementation details and the full endpoint specification, refer to: GET /models API Documentation.

FAQs

You can buy IO Credits or wait for your daily limit to refresh.
If you already have credits, they will automatically cover additional usage with no interruptions.
IO Intelligence uses a shared credit pool system.
Credits can be spent on any model, with each consuming credits at a different rate depending on the complexity.
Yes. Chat interactions count toward your API quota and contribute to your daily limit.