
Setup
-
Open Cursor Settings
Press
Cmd +, (macOS) orCtrl +, (Windows/Linux) and navigate to Models. - Enter your API key In the OpenAI API Key field, enter your Bifrost virtual key or provider API key.
-
Override the base URL
Toggle Override OpenAI Base URL to ON and enter your Bifrost endpoint:
For deployed instances, use your Bifrost deployment URL (e.g.,
https://bifrost.example.com/cursor). -
Add custom models (optional)
Type a model name in the Add or search model field using the
provider/model-nameformat: Examples:
anthropic/claude-sonnet-4-5-20250929,openai/gpt-5,gemini-2.5-proProvider Format Example Anthropic anthropic/model-nameanthropic/claude-sonnet-4-5-20250929Gemini model-namegemini-2.5-proOpenAI openai/model-nameopenai/gpt-5Bedrock bedrock/model-namebedrock/anthropic.claude-3Vertex (non-Gemini) vertex/model-namevertex/text-bisonOther providers provider/model-namegroq/llama-3.3-70b-versatile
Using Virtual Keys
Bifrost Virtual Keys can be used as the OpenAI API Key in Cursor. Virtual keys let you enforce budgets, rate limits, and provider access controls for each user or team.Model Selection
Cursor assigns models to different features — Chat, Agent, Inline Edit, and Tab Completion. After configuring Bifrost, you can assign anyprovider/model-name to each feature for optimal cost and performance:
- Use a powerful model like
openai/gpt-5oranthropic/claude-sonnet-4-5-20250929for Agent mode - Use a fast model like
groq/llama-3.3-70b-versatilefor Tab completion
Using Multiple Providers
Bifrost routes requests to the correct provider based on the model name. Use theprovider/model-name format to access any configured provider through the single OpenAI-compatible endpoint:
Supported Providers
Bifrost supports the following providers with theprovider/model-name format:
openai, anthropic, azure, gemini, vertex, bedrock, mistral, groq, cerebras, cohere, perplexity, xai, ollama, openrouter, huggingface, nebius, parasail, replicate, vllm, sgl
Cursor’s “Override OpenAI Base URL” is a global setting that applies to all OpenAI-compatible models. This works well with Bifrost since Bifrost handles routing to the correct provider based on the model name.
Observability
All Cursor requests through Bifrost are logged. Monitor them athttp://localhost:8080/logs — filter by provider, model, or search through conversation content.
