Models
The model catalog is served live by the Router. You can browse it two ways:
- Web UI — pc.0g.ai shows every model with current pricing, the number of healthy providers, and capability badges (streaming, tool calling, vision, etc.).
- API —
GET /v1/modelsreturns the same data in OpenAI's list format. No authentication required.
Listing Models
curl https://router-api.0g.ai/v1/models
{
"object": "list",
"data": [
{
"id": "zai-org/GLM-5-FP8",
"object": "model",
"owned_by": "0G Foundation",
"name": "zai-org/GLM-5-FP8",
"context_length": 131072,
"pricing": {
"prompt": "100000000000",
"completion": "320000000000"
},
"provider_count": 3
}
]
}
Prices are in neuron per token (1e18 neuron = 1 0G). Multiply by input_tokens / output_tokens to estimate cost.
Capability Flags
Not every model supports every feature. Before relying on tool calling, vision input, or JSON mode, check the model's entry in the Web UI or in the /v1/models response — capability flags are shown on each model card and in the API payload.
If you send a tools field to a model that doesn't support it, the Router returns 400 Bad Request rather than silently dropping the parameter.
Listing Providers for a Model
curl "https://router-api.0g.ai/v1/providers?model=zai-org/GLM-5-FP8"
Returns every TEE-acknowledged provider serving that model, with on-chain address, observed latency, and TEE attestation info. Use these addresses with the provider.address field if you want deterministic routing.
Query parameters:
| Field | Description |
|---|---|
model | Filter to providers serving a specific model ID |
service_type | Filter by service type (e.g. chatbot, text-to-image, speech-to-text) |