Go MCP Gateways for BSD

Browse free open source Go MCP Gateways for BSD and projects below. Use the toggles on the left to filter open source Go MCP Gateways for BSD by OS, license, language, programming language, and project status.

  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Host LLMs in Production With On-Demand GPUs Icon
    Host LLMs in Production With On-Demand GPUs

    NVIDIA L4 GPUs. 5-second cold starts. Scale to zero when idle.

    Deploy your model, get an endpoint, pay only for compute time. No GPU provisioning or infrastructure management required.
    Try Free
  • 1
    Obot MCP Gateway

    Obot MCP Gateway

    Hosting, Registry, Gateway, and Chat Client

    Obot is an open-source platform built to help organizations adopt and operate Model Context Protocol (MCP) capabilities in a centralized, production-friendly way. It combines multiple MCP building blocks into one system, including hosting for MCP servers, a registry for discovery, a gateway layer to route access, and a standards-compliant chat client experience. The project is aimed at solving common enterprise rollout problems such as reliably hosting servers for internal and external users, curating “approved” MCP servers for employees to find, and enforcing authentication, access control, and auditable activity. It also supports building richer agents and chatbots that can leverage MCP servers while keeping operations manageable for IT and platform teams. The platform is designed to work with a variety of workflows and clients, so MCP servers managed inside Obot can be used by automation/agent frameworks as well as popular chat clients that speak MCP.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Bifrost

    Bifrost

    The Fastest LLM Gateway with built in OTel observability

    Bifrost is an LLM gateway designed to provide a unified OpenAI-compatible API front for many different model providers. It abstracts away the complexity of working directly with multiple backend providers (OpenAI, Anthropic, AWS Bedrock, Google Vertex, etc.), enabling you to plug in providers and switch between them without touching your client code. It is built to be high performance: in benchmark tests at 5,000 requests per second, it reportedly adds only microseconds of overhead and achieves perfect success rates with no failed requests. Bifrost supports features such as automatic fallback (failover between providers), load balancing across API keys/providers, and semantic caching to reduce latency and cost. It also includes observability with built-in metrics, tracing, logging, and supports governance features like rate limiting, access control, and cost budgeting. The architecture is modular: there is a core engine, plugin layers, and transport layers (HTTP APIs).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    kgateway

    kgateway

    The Cloud-Native API Gateway and AI Gateway

    kgateway is a mature, cloud-native API and ingress gateway designed to provide unified API connectivity for services, microservices, serverless workloads, and AI-centric systems running on Kubernetes clusters. It implements the Kubernetes Gateway API and can operate as both a lightweight in-cluster microgateway and a centralized gateway capable of handling billions of API calls with high performance and low latency. By integrating with Envoy and advanced data planes, it handles modern ingress concerns such as traffic routing, authentication, authorization, rate limiting, and observability for traditional HTTP/gRPC services and AI workloads alike. Beyond standard API traffic, kgateway also supports gateway patterns tailored for large language model (LLM) consumption, inference routing, and Model Context Protocol (MCP) orchestration, enabling secure access to models, tools, and agent interactions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Docker MCP Gateway

    Docker MCP Gateway

    Docker mcp CLI plugin / MCP Gateway

    Docker’s MCP Gateway project is a Docker CLI plugin and supporting gateway system designed to run, manage, and securely expose MCP servers using container isolation. It underpins the MCP Toolkit experience in Docker Desktop, but it can also be used independently as a general-purpose MCP operational layer. The core idea is to treat MCP servers like containerized services, giving each server controlled privileges and a lifecycle you can inspect, enable/disable, and reset as needed. Instead of having each AI client manage its own MCP server configuration, the gateway provides a unified interface so multiple clients can connect consistently to the same configured tool surface. The project emphasizes security and operational hygiene by supporting secrets management (to avoid leaking credentials via plain environment variables) and providing built-in OAuth flows for MCP servers that require authenticated service access.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB