Distributed Inference Network

Earn from your GPU.
Power AI inference.

Ynnova is an open inference network. Contribute your idle GPU resources and earn per token served — or consume fast, cheap LLM inference via our OpenAI-compatible API.

Become a Provider API Documentation
High throughput
Tokens per second
Low latency
Time to first token
OpenAI-compatible
Drop-in API
24 / 7
Uptime

For GPU Providers

Have an idle GPU? Connect it to the Ynnova network and start earning per token your hardware serves.

🔌

For Developers

Access distributed GPU inference with a single API key. Drop-in replacement for the OpenAI SDK.

How it works
Your GPU resources, routed intelligently across the network.
01

Provider connects

GPU owners deploy a worker agent that tunnels into the Ynnova network. No port forwarding required.

02

Request is routed

Incoming API calls are load-balanced to the best available node based on latency and capacity.

03

Provider earns

Token throughput is metered and settled. Payouts are calculated per million tokens served.