Open Source Go Artificial Intelligence Software for BSD

Go Artificial Intelligence Software for BSD

Browse free open source Go Artificial Intelligence Software for BSD and projects below. Use the toggles on the left to filter open source Go Artificial Intelligence Software for BSD by OS, license, language, programming language, and project status.

  • Ship AI Apps Faster with Vertex AI Icon
    Ship AI Apps Faster with Vertex AI

    Go from idea to deployed AI app without managing infrastructure. Vertex AI offers one platform for the entire AI development lifecycle.

    Ship AI apps and features faster with Vertex AI—your end-to-end AI platform. Access Gemini 3 and 200+ foundation models, fine-tune for your needs, and deploy with enterprise-grade MLOps. Build chatbots, agents, or custom models. New customers get $300 in free credit.
    Try Vertex AI Free
  • Easily Host LLMs and Web Apps on Cloud Run Icon
    Easily Host LLMs and Web Apps on Cloud Run

    Run everything from popular models with on-demand NVIDIA L4 GPUs to web apps without infrastructure management.

    Run frontend and backend services, batch jobs, host LLMs, and queue processing workloads without the need to manage infrastructure. Cloud Run gives you on-demand GPU access for hosting LLMs and running real-time AI—with 5-second cold starts and automatic scale-to-zero so you only pay for actual usage. New customers get $300 in free credit to start.
    Try Cloud Run Free
  • 1
    kagent

    kagent

    Kubernetes native framework for building AI agents

    Kagent is a Kubernetes-native framework for building, deploying, and operating AI agents as first-class cloud-native workloads. It models core agent concepts declaratively using Kubernetes custom resources, so teams can manage agents similarly to other platform components via YAML, controllers, and standard cluster workflows. In kagent’s design, an “Agent” represents a system prompt plus a set of tools and other agents, along with an LLM configuration, making the agent definition portable and repeatable across environments. It supports multiple model providers through a dedicated configuration resource, allowing teams to switch providers or run mixed environments while keeping the agent spec stable. A major focus is tool integration via MCP: agents can connect to MCP servers for tool access, and kagent includes an MCP server with tools for common Kubernetes and platform engineering systems.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 2
    Memobase

    Memobase

    Fast backend for long-term AI user memory via structured profiles

    Memobase is an open source backend system that enables long-term user memory functionality for AI applications by capturing and structuring information about users across interactions. Its design centers on creating user profiles and recording event timelines, allowing AI systems to remember, understand, and evolve in their behaviour toward individual users over time. Instead of relying purely on traditional embedding-based retrieval or RAG systems, Memobase uses profile and timeline structures to deliver memory that reflects user context efficiently and meaningfully. The system focuses on three principal performance metrics: high search performance, reduced large language model (LLM) costs through batch processing techniques, and low latency with minimal SQL operations. Memobase supports integration with existing LLM workflows via APIs and SDKs (including Python, Node, and Go), making it easy to adopt within diverse application stacks.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    Docker MCP Gateway

    Docker MCP Gateway

    Docker mcp CLI plugin / MCP Gateway

    Docker’s MCP Gateway project is a Docker CLI plugin and supporting gateway system designed to run, manage, and securely expose MCP servers using container isolation. It underpins the MCP Toolkit experience in Docker Desktop, but it can also be used independently as a general-purpose MCP operational layer. The core idea is to treat MCP servers like containerized services, giving each server controlled privileges and a lifecycle you can inspect, enable/disable, and reset as needed. Instead of having each AI client manage its own MCP server configuration, the gateway provides a unified interface so multiple clients can connect consistently to the same configured tool surface. The project emphasizes security and operational hygiene by supporting secrets management (to avoid leaking credentials via plain environment variables) and providing built-in OAuth flows for MCP servers that require authenticated service access.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    GitHub Agentic Workflows

    GitHub Agentic Workflows

    GitHub Agentic Workflows

    GitHub Agentic Workflows is an experimental CLI extension and framework for the gh GitHub CLI that lets developers author automation driven by natural language specifications instead of hand-written code, compiling those descriptions into GitHub Actions workflows that run AI agents (like Copilot, Claude Code, or Codex) on schedule or in response to repository events. By writing intent in markdown files, a developer can quickly generate .yml Actions workflows that perform tasks such as summarizing issues, automating triage, generating reports, or maintaining documentation, all without manually crafting YAML logic from scratch. The system emphasizes safety and guardrails, running agents in sandboxed environments with minimal permissions by default, and using “safe outputs” to constrain what the workflow can write back into the repository. It includes tooling for compiling, testing, and iterating on agentic workflows locally and integrates with GitHub’s existing Actions ecosystem.
    Downloads: 4 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Bifrost

    Bifrost

    The Fastest LLM Gateway with built in OTel observability

    Bifrost is an LLM gateway designed to provide a unified OpenAI-compatible API front for many different model providers. It abstracts away the complexity of working directly with multiple backend providers (OpenAI, Anthropic, AWS Bedrock, Google Vertex, etc.), enabling you to plug in providers and switch between them without touching your client code. It is built to be high performance: in benchmark tests at 5,000 requests per second, it reportedly adds only microseconds of overhead and achieves perfect success rates with no failed requests. Bifrost supports features such as automatic fallback (failover between providers), load balancing across API keys/providers, and semantic caching to reduce latency and cost. It also includes observability with built-in metrics, tracing, logging, and supports governance features like rate limiting, access control, and cost budgeting. The architecture is modular: there is a core engine, plugin layers, and transport layers (HTTP APIs).
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    CyberStrikeAI

    CyberStrikeAI

    CyberStrikeAI is an AI-native security testing platform built in Go

    CyberStrikeAI is an AI-native security testing platform built in Go that brings autonomous penetration testing, vulnerability discovery, and attack chain analysis into a unified interface. The platform integrates over 100 security tools out of the box and pairs them with an intelligent orchestration engine that can be directed via natural language or policy definitions, allowing users to automate reconnaissance, scanning, exploitation, and reporting without manual sequencing of tools. It supports role-based testing, letting teams define security roles with tailored tool access and prompts, and includes a skills system that encapsulates specialized testing strategies that the AI can incorporate into its planning. Through comprehensive lifecycle management, results are tracked, aggregated, and visualized, with support for versioned persistence, search, and risk severity scoring.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Acontext

    Acontext

    Context data platform for building observable, self-learning AI agents

    Acontext is a cloud-native context data platform designed to support the development and operation of advanced AI agents. It provides a unified system to store and manage contexts, multimodal messages, artifacts, and task workflows, enabling developers to engineer context effectively for their agent products. The platform observes agent tasks and user feedback in real time, offering robust observability into workflows and helping teams understand how agents perform over time. Acontext also supports agent self-learning by distilling structured skills and experiences from previously completed tasks, which can later be reused or searched to improve future performance. It includes tools to interact with session data, background agents that monitor progress, and a dashboard that visualizes success rates, artifacts, and learned skills. By combining persistent storage, observability, and learning capabilities, Acontext aims to make AI agents more scalable, reliable, and capable.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Crush

    Crush

    The glamourous AI CLI coding agent for your favourite terminal 💘

    Crush is a next-generation, terminal-based AI coding assistant developed by Charm, designed to seamlessly integrate with your tools, workflows, and preferred LLMs. It provides developers with an intuitive, session-based experience where multiple contexts can be managed across projects. With flexible model switching, Crush allows you to change providers mid-session while retaining conversation history. It enhances productivity by combining LSP (Language Server Protocol) support with extensible MCP (Model Context Protocol) integrations for richer coding context and external tool connectivity. Built for portability, it offers first-class support across macOS, Linux, Windows (PowerShell and WSL), and BSD systems. Backed by the Charm ecosystem, Crush is a stable, actively maintained evolution of the original OpenCode project.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Obot MCP Gateway

    Obot MCP Gateway

    Hosting, Registry, Gateway, and Chat Client

    Obot is an open-source platform built to help organizations adopt and operate Model Context Protocol (MCP) capabilities in a centralized, production-friendly way. It combines multiple MCP building blocks into one system, including hosting for MCP servers, a registry for discovery, a gateway layer to route access, and a standards-compliant chat client experience. The project is aimed at solving common enterprise rollout problems such as reliably hosting servers for internal and external users, curating “approved” MCP servers for employees to find, and enforcing authentication, access control, and auditable activity. It also supports building richer agents and chatbots that can leverage MCP servers while keeping operations manageable for IT and platform teams. The platform is designed to work with a variety of workflows and clients, so MCP servers managed inside Obot can be used by automation/agent frameworks as well as popular chat clients that speak MCP.
    Downloads: 2 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    kMCP

    kMCP

    Kubernetes Controller for building, testing and deploying MCP servers

    KMCP is a companion toolchain for building, testing, and deploying MCP servers with a workflow that spans local development through Kubernetes production deployments. It includes a CLI for day-to-day development tasks like scaffolding new MCP projects, managing tools, building container images, and running an MCP server locally for validation. For cluster operations, it includes a Kubernetes controller that manages MCP server lifecycles using a dedicated Custom Resource Definition (CRD), allowing MCP servers to be represented as native Kubernetes objects you can operate with familiar kubectl-driven patterns. A key component is the transport adapter, which fronts MCP servers to provide routing and multi-transport support without requiring code changes in your server implementation. The project is geared toward consistency, aiming to reduce the “glue work” of writing Dockerfiles, hand-rolling manifests, and manually wiring networking and deployment details for each MCP server.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    kgateway

    kgateway

    The Cloud-Native API Gateway and AI Gateway

    kgateway is a mature, cloud-native API and ingress gateway designed to provide unified API connectivity for services, microservices, serverless workloads, and AI-centric systems running on Kubernetes clusters. It implements the Kubernetes Gateway API and can operate as both a lightweight in-cluster microgateway and a centralized gateway capable of handling billions of API calls with high performance and low latency. By integrating with Envoy and advanced data planes, it handles modern ingress concerns such as traffic routing, authentication, authorization, rate limiting, and observability for traditional HTTP/gRPC services and AI workloads alike. Beyond standard API traffic, kgateway also supports gateway patterns tailored for large language model (LLM) consumption, inference routing, and Model Context Protocol (MCP) orchestration, enabling secure access to models, tools, and agent interactions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Qwen2.5-Coder

    Qwen2.5-Coder

    Qwen2.5-Coder is the code version of Qwen2.5, the large language model

    Qwen2.5-Coder, developed by QwenLM, is an advanced open-source code generation model designed for developers seeking powerful and diverse coding capabilities. It includes multiple model sizes—ranging from 0.5B to 32B parameters—providing solutions for a wide array of coding needs. The model supports over 92 programming languages and offers exceptional performance in generating code, debugging, and mathematical problem-solving. Qwen2.5-Coder, with its long context length of 128K tokens, is ideal for a variety of use cases, from simple code assistants to complex programming scenarios, matching the capabilities of models like GPT-4o.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    ChatGPT Proxy

    ChatGPT Proxy

    Simple Cloudflare bypass for ChatGPT

    ChatGPTProxy is an open-source project that creates a lightweight proxy server to intermediate between client applications and the ChatGPT web endpoints, allowing developers to integrate ChatGPT-style functionality into their software without using an official API or embedding web UI code directly. This tool works by accepting requests in a defined format, forwarding them through the proxy to ChatGPT’s backend services, and returning responses to the caller, abstracting away direct browser automation or scraping concerns from the application layer. By consolidating the traffic through a proxy, developers can centralize logging, throttling, authentication, and caching in one place, making it easier to build consistent and controlled AI workflows. The proxy can also be customized to enforce usage policies, attach additional metadata, or translate request/response formats for compatibility with other tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ChatGPT-to-API

    ChatGPT-to-API

    Scalable unofficial ChatGPT API for production

    ChatGPT-to-API is an open-source project that exposes an interface intended to wrap ChatGPT interactions in an API-like experience, so that tools and services built around traditional API calls can work with ChatGPT even though no official programmatic API is provided in the same way. It functions as a translation layer, converting standardized API requests into ChatGPT-compatible prompts and then converting responses back into machine-friendly JSON objects that resemble API outputs. This makes it possible to plug ChatGPT into automated systems, serverless functions, or backend services that expect REST or JSON RPC interfaces without needing to modify each consumer to speak a browser protocol. Developers can deploy ChatGPT-to-API on a server, send it requests with structured parameters (like messages, metadata, or configuration flags), and receive structured replies in return.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    ENScan Go

    ENScan Go

    ENScan_GO is an enterprise information reconnaissance tool

    ENScan_GO is an enterprise information reconnaissance tool focused on Chinese corporate data sources. It aggregates official and third-party APIs to pull records like ICP filings, affiliated/holding companies, apps, mini-programs, and WeChat official accounts, then exports merged results for analysis. The tool targets analysts who need one-click collection and normalized output to reduce manual lookups across registries and platforms. Recent releases added a reworked task model with queueing, resumable searches via cached progress, export format options, and a public API surface for custom keyword strategies. Documentation and issues discuss operational concerns such as rate limits, verification challenges, and use of proxies to reduce bans. The project is maintained under Apache-2.0 and is positioned for both single-shot queries and batch investigations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MCP Language Server

    MCP Language Server

    mcp-language-server gives MCP enabled clients access semantic tools

    mcp-language-server gives MCP-enabled clients semantic code-navigation powers—go-to-definition, find references, rename, and diagnostics—by brokering requests to language servers. It is not “a language server for MCP,” but an MCP server that exposes language-server–style capabilities to agents and chat IDEs through typed tools. The README demonstrates a streamlined setup: install the Go server, plug in one or more language servers per language, and the MCP client gains editor-grade navigation across the workspace. This helps agents reason precisely about symbols and files instead of guessing via grep-like prompts, enabling safer edits and better refactoring proposals. The project maintains active issues and PRs, indicating ongoing polish around multi-language routing and robustness. It’s also listed in community indexes, reflecting adoption across MCP clients.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    godlp

    godlp

    Sensitive information protection toolkit

    godlp appears to be another software project from ByteDance — however, as of the most recent checks, there’s very little publicly available information about it: the repository exists under ByteDance’s GitHub, but its documentation, README, and metadata are minimal (or not human-readable), and the project seems to have limited community visibility compared to their other major tools. Because of that opacity, one must infer that godlp is likely a specialized internal or early-stage tool, possibly related to internal optimization, data processing, or platform-specific functionality (given ByteDance’s historical patterns). The minimal public footprint suggests it may be experimental, unmaintained, or only partially open-sourced, which reduces its immediate practicality for external developers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB