Skip to main content
GET
/
v1
/
models
List available models
curl --request GET \
  --url http://localhost:8080/v1/models
{
  "data": [
    {
      "id": "<string>",
      "canonical_slug": "<string>",
      "name": "<string>",
      "deployment": "<string>",
      "created": 123,
      "context_length": 123,
      "max_input_tokens": 123,
      "max_output_tokens": 123,
      "architecture": {
        "modality": "<string>",
        "tokenizer": "<string>",
        "instruct_type": "<string>",
        "input_modalities": [
          "<string>"
        ],
        "output_modalities": [
          "<string>"
        ]
      },
      "pricing": {
        "prompt": "<string>",
        "completion": "<string>",
        "request": "<string>",
        "image": "<string>",
        "web_search": "<string>",
        "internal_reasoning": "<string>",
        "input_cache_read": "<string>",
        "input_cache_write": "<string>"
      },
      "top_provider": {
        "is_moderated": true,
        "context_length": 123,
        "max_completion_tokens": 123
      },
      "per_request_limits": {
        "prompt_tokens": 123,
        "completion_tokens": 123
      },
      "supported_parameters": [
        "<string>"
      ],
      "default_parameters": {
        "temperature": 123,
        "top_p": 123,
        "frequency_penalty": 123
      },
      "hugging_face_id": "<string>",
      "description": "<string>",
      "owned_by": "<string>",
      "supported_methods": [
        "<string>"
      ]
    }
  ],
  "extra_fields": {
    "request_type": "<string>",
    "provider": "openai",
    "model_requested": "<string>",
    "model_deployment": "<string>",
    "latency": 123,
    "chunk_index": 123,
    "raw_request": {},
    "raw_response": {},
    "cache_debug": {
      "cache_hit": true,
      "cache_id": "<string>",
      "hit_type": "<string>",
      "provider_used": "<string>",
      "model_used": "<string>",
      "input_tokens": 123,
      "threshold": 123,
      "similarity": 123
    }
  },
  "next_page_token": "<string>"
}

Query Parameters

provider
enum<string>

Filter by provider (e.g., openai, anthropic, bedrock) AI model provider identifier

Available options:
openai,
azure,
anthropic,
bedrock,
cohere,
vertex,
mistral,
ollama,
groq,
sgl,
parasail,
perplexity,
cerebras,
gemini,
openrouter,
elevenlabs,
huggingface,
nebius
page_size
integer

Maximum number of models to return

Required range: x >= 0
page_token
string

Token for pagination

Response

Successful response

data
object[]
extra_fields
object

Additional fields included in responses

next_page_token
string