Fastest enterprise AI gateway (50x faster than LiteLLM) with adaptive load balancer, cluster mode, guardrails, 1000+ models support & <100 µs overhead at 5k RPS.
{
"mcpServers": {
"bifrost": {
"args": [
"-y",
"@maximhq/bifrost"
],
"command": "npx"
}
}
}Are you the author?
Add this badge to your README to show your security score and help users find safe servers.
Is it safe?
No known CVEs for @maximhq/bifrost.
No authentication — any process on your machine can connect.
Apache-2.0. View license →
Is it maintained?
Last commit 0 days ago. 3,591 stars. 784 weekly downloads.
Will it work with my client?
Transport: stdio, sse, http. Works with Claude Desktop, Cursor, Claude Code, and most MCP clients.
This server supports HTTP transport. Be the first to test it — help the community know if it works.
No known vulnerabilities.
Have you used this server?
Share your experience — it helps other developers decide.
Sign in to write a review.
MCP Server for GCP environment for interacting with various Observability APIs.
MCP Server for kubernetes management commands
Model Context Protocol (MCP) server for Kubernetes and OpenShift
The power of Claude Code / GeminiCLI / CodexCLI + [Gemini / OpenAI / OpenRouter / Azure / Grok / Ollama / Custom Model / All Of The Above] working as one.
MCP Security Weekly
Get CVE alerts and security updates for Bifrost and similar servers.
Start a conversation
Ask a question, share a tip, or report an issue.
Sign in to join the discussion.