Showing 104 open source projects for "server local"

View related business solutions
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Last9 MCP Server

    Last9 MCP Server

    Last9 MCP Server

    The Last9 MCP Server is a Model Context Protocol server implementation for Last9, enabling AI agents to seamlessly bring real-time production context—logs, metrics, and traces—into local environments to auto-fix code faster. ​
    Downloads: 15 This Week
    Last Update:
    See Project
  • 2
    WhatsApp MCP Server

    WhatsApp MCP Server

    WhatsApp MCP server enabling AI access to chats and messaging

    whatsapp-mcp is an open source Model Context Protocol (MCP) server that enables AI agents to interact directly with a user’s WhatsApp account through a structured interface. It acts as a bridge between WhatsApp and large language models, allowing controlled access to messages, chats, and contacts. whatsapp-mcp is composed of two main components: a Go-based bridge that connects to the WhatsApp Web API and stores data locally, and a Python-based MCP server that exposes tools for AI interaction. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    OpenVINO Model Server

    OpenVINO Model Server

    A scalable inference server for models optimized with OpenVINO

    OpenVINO™ Model Server is a high-performance inference serving system designed to host and serve machine learning models that have been optimized with the OpenVINO toolkit. It’s implemented in C++ for scalability and efficiency, making it suitable for both edge and cloud deployments where inference workloads must be reliable and high throughput. The server exposes model inference via standard network protocols like REST and gRPC, allowing any client that speaks those protocols to request predictions remotely, abstracting away the complexity of where and how the model runs. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 4
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    OpenJarvis is an open-source framework designed to build personal AI agents that run primarily on local devices rather than relying on cloud infrastructure. Developed as part of the Intelligence Per Watt research initiative, it focuses on improving the efficiency and practicality of on-device AI systems. The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and...
    Downloads: 131 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 6
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. ...
    Downloads: 22 This Week
    Last Update:
    See Project
  • 7
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    mcpo

    mcpo

    A simple, secure MCP-to-OpenAPI proxy server

    mcpo is a minimal bridge that exposes any MCP tool as an OpenAPI-compatible HTTP server. Instead of writing glue code, you point mcpo at an MCP server command and it generates REST endpoints and an OpenAPI spec that other systems (or LLM agent frameworks) can call immediately. This design lets you reuse a growing library of MCP servers with platforms that only understand HTTP+OpenAPI, unifying tool access across ecosystems.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 9
    Anyquery

    Anyquery

    Query anything (GitHub, Notion, +40 more) with SQL and let LLMs

    ...Built on top of SQLite, the engine uses a plugin architecture that allows it to extend support to dozens of external services and data sources. Users can query structured files such as CSV, JSON, and Parquet as well as remote data sources like SaaS APIs, cloud storage services, and local applications. The platform also supports querying multiple data sources simultaneously and joining them together within a single SQL query, enabling powerful cross-system analysis. In addition to operating as a local query engine, the system can run as a MySQL-compatible server so that traditional database tools can connect to it.
    Downloads: 37 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    kMCP

    kMCP

    Kubernetes Controller for building, testing and deploying MCP servers

    KMCP is a companion toolchain for building, testing, and deploying MCP servers with a workflow that spans local development through Kubernetes production deployments. It includes a CLI for day-to-day development tasks like scaffolding new MCP projects, managing tools, building container images, and running an MCP server locally for validation. For cluster operations, it includes a Kubernetes controller that manages MCP server lifecycles using a dedicated Custom Resource Definition (CRD), allowing MCP servers to be represented as native Kubernetes objects you can operate with familiar kubectl-driven patterns. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 11
    web-eval-agent MCP Server

    web-eval-agent MCP Server

    An MCP server that autonomously evaluates web applications

    ...Marketing and README material emphasize supercharging local debugging loops by combining live browser execution with LLM-driven hypotheses and fixes. Activity on the repo shows steady iteration, with issues and PRs centered on reliability and developer experience. In short, it wraps autonomous, in-editor web testing and diagnosis behind a predictable MCP interface.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    XcodeBuildMCP

    XcodeBuildMCP

    A Model Context Protocol (MCP) server

    ...The server aims to be “agent-ready,” surfacing capabilities (build, clean, test, archive, install, run, log collection) with explicit schemas instead of brittle prompt instructions. It supports MCP transports suitable for local IDEs and service deployments, and pairs with a public website that positions it as a bridge between Xcode and autonomous assistants.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 13
    Shadcn UI v4 MCP Server

    Shadcn UI v4 MCP Server

    A mcp server to allow LLMS gain context about shadcn ui component

    ...The server supports multiple frontend frameworks including React, Svelte, Vue, and React Native, making it highly versatile for cross-platform development. It includes smart caching and efficient GitHub API usage to optimize performance and handle rate limits during component retrieval. The system also supports multiple transport modes such as standard input/output and Server-Sent Events, enabling both local and distributed deployments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Colab-MCP

    Colab-MCP

    An MCP server for interacting with Google Colab

    ...This approach bridges the gap between local AI agents and remote high-performance compute environments, allowing users to offload heavy workloads such as machine learning training, data analysis, and dependency-heavy tasks to Colab’s GPU and TPU resources. By exposing Colab as an MCP server, the tool enables seamless integration with a wide range of AI assistants and agent frameworks, creating a standardized interface for tool use and execution.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Jan.ai

    Jan.ai

    Open source alternative to ChatGPT that runs 100% offline

    Jan.ai is an open-source, privacy-focused AI assistant that serves as an alternative to ChatGPT, running completely locally on your device. It allows you to download and run LLMs (local language models) offline while also offering optional integration with cloud-based model providers—giving you full control over your data and AI interactions. Download and run LLMs (Llama, Gemma, Qwen, GPT-oss etc.) from HuggingFace. Connect to GPT models via OpenAI, Claude models via Anthropic, Mistral,...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 16
    Groq Desktop

    Groq Desktop

    Local Groq Desktop chat app with MCP support

    Groq Desktop is a cross-platform (Windows / macOS / Linux) local desktop application that provides a graphical chat interface for interacting with Groq-hosted, function-call-capable models. It bundles a built-in MCP (Model Context Protocol) server enabling smart function calling, letting users chat with an AI, send images, or interact with richer multimodal inputs — all from a friendly desktop UI.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 17
    LMCache

    LMCache

    Supercharge Your LLM with the Fastest KV Cache Layer

    LMCache is an extension layer for LLM serving engines that accelerates inference, especially with long contexts, by storing and reusing key-value (KV) attention caches across requests. Instead of rebuilding KV states for repeated or shared text segments, LMCache persists and retrieves them from multiple tiers—GPU memory, CPU DRAM, and local disk—then injects them into subsequent requests to reduce TTFT and increase throughput. Its design supports reuse beyond strict prefix matching and enables sharing across serving instances, improving efficiency under real multi-tenant traffic. The broader project includes examples, tests, a server component, and public posts describing cross-engine sharing and inter-GPU KV transfers. ...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 18
    5ire

    5ire

    5ire is a cross-platform desktop AI assistant, MCP client

    5ire is a sleek, cross‑platform desktop AI assistant and MCP client that connects to major service providers, supports a local knowledge base and tool integration via MCP servers, enabling robust RAG and assistant features. These components are required as they constitute the runtime environment for the MCP Server. If you don't anticipate using the tools feature immediately, you may choose to skip this installation step and complete it later when the need arises.
    Downloads: 21 This Week
    Last Update:
    See Project
  • 19
    OctoMind MCP

    OctoMind MCP

    An MCP server for octomind tools, resources and prompts

    The Octomind MCP Server is designed to integrate Octomind's end-to-end testing tools and resources into local development environments. It enables AI-powered interfaces to create, execute, and manage e2e tests, enhancing the testing workflow. ​
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 21
    PowerInfer

    PowerInfer

    High-speed Large Language Model Serving for Local Deployment

    ...PowerInfer incorporates specialized algorithms and sparse operators to manage neuron activation patterns and minimize data transfers between hardware components. As a result, it enables powerful language models to run on consumer hardware while achieving performance comparable to more expensive server-grade systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    OpenClaw Studio

    OpenClaw Studio

    A clean web dashboard for OpenClaw

    ...The platform provides real-time chat capabilities, approval management, and job configuration tools for agent operations. Built with a control-plane architecture, it handles communication between the browser and the gateway through server-managed connections. OpenClaw Studio supports both local and cloud deployments, making it flexible for different development and production environments. It is ideal for developers and teams looking to streamline agent management and improve visibility into autonomous AI workflows.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    MiniMax-MCP

    MiniMax-MCP

    Official MiniMax Model Context Protocol (MCP) server

    ...The server is written in Python and distributed under the MIT license, with a pyproject.toml and uv-based workflow that makes installation and execution reproducible. Configuration is handled through JSON files that tell MCP clients how to launch the server (typically via uvx minimax-mcp) and which environment variables to use for the API key, host, and output directory.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Deep Research

    Deep Research

    Use any LLMs (Large Language Models) for Deep Research

    Deep Research is a local-first research agent that orchestrates multiple LLMs to generate in-depth reports in minutes. It combines “thinking” and “task” model roles with live internet access to plan, search, read, and synthesize findings into structured outputs. The project emphasizes privacy: processing and storage happen locally, avoiding server-side retention of your queries and notes.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    ChatMCP

    ChatMCP

    ChatMCP is an AI chat client implementing the Model Context Protocol

    ChatMCP is a cross‑platform AI chat client that implements the Model Context Protocol (MCP) to provide unified chat experiences across environments—including desktop, mobile, and web—with synchronization and protocol support tailored for MCP. You can install MCP Server from MCP Server Market, MCP Server Market is a collection of MCP Server, you can use it to chat with different data. Tested on major distributions: Ubuntu, Fedora, Arch Linux, openSUSE. Improved Experience: Latest versions...
    Downloads: 17 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB