Showing 1084 open source projects for "python programming language"

View related business solutions
  • Dragonfly | An In-Memory Data Store without Limits Icon
    Dragonfly | An In-Memory Data Store without Limits

    Dragonfly Cloud is engineered to handle the heaviest data workloads with the strictest security requirements.

    Dragonfly is a drop-in Redis replacement that is designed for heavy data workloads running on modern cloud hardware. Migrate in less than a day and experience up to 25X the performance on half the infrastructure.
    Learn More
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Learn More
  • 1
    LlamaIndex

    LlamaIndex

    Central interface to connect your LLM's with external data

    LlamaIndex (GPT Index) is a project that provides a central interface to connect your LLM's with external data. LlamaIndex is a simple, flexible interface between your external data and LLMs. It provides the following tools in an easy-to-use fashion. Provides indices over your unstructured and structured data for use with LLM's. These indices help to abstract away common boilerplate and pain points for in-context learning. Dealing with prompt limitations (e.g. 4096 tokens for Davinci) when...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Auto-Deep-Research

    Auto-Deep-Research

    Your Fully-Automated Personal AI Assistant

    Auto-Deep-Research is a system designed to fully automate deep research workflows using language models, retrieval, planning, and multi-stage reasoning to produce structured research artifacts such as surveys, benchmarks, reports, and even prototypes without heavy human intervention. Users provide a research topic or multifaceted goal, and the system autonomously breaks the objective down into subtasks like literature collection, critical summarization, cross-comparison, citation extraction,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Claude Code Plugins

    Claude Code Plugins

    Intelligent automation and multi-agent orchestration for Claude Code

    Claude Code Plugins is a lightweight framework designed to define, manage, and execute AI agents in a modular and extensible way, typically focusing on orchestrating tasks using large language models and tool integrations. The project provides abstractions for building agents that can interpret instructions, execute commands, and interact with external systems in a structured workflow. It emphasizes simplicity and composability, allowing developers to define agent behaviors through reusable...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 4
    ADX MCP Server

    ADX MCP Server

    A Model Context Protocol (MCP) server that enables AI assistants

    The Azure Data Explorer MCP Server is a Model Context Protocol (MCP) server that enables AI assistants to query and analyze Azure Data Explorer databases through standardized interfaces. It allows the execution of Kusto Query Language (KQL) queries and exploration of data within Azure Data Explorer clusters. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • Composable, Open Source Payments Platform Icon
    Composable, Open Source Payments Platform

    Build or enhance your payments stack, while maintaining control with an open-source, full-stack and modular infrastructure.

    Juspay's Payments Orchestration Platform offers a comprehensive product suite for businesses, including open-source payment orchestration, global payouts, seamless authentication, payment tokenization, fraud & risk management, end-to-end reconciliation, unified payment analytics & more. The company’s offerings also include end-to-end white label payment gateway solutions & real-time payments infrastructure for banks. These solutions help businesses achieve superior conversion rates, reduce fraud, optimize costs, and deliver seamless customer experiences at scale.
    Learn More
  • 5
    Intel Extension for Transformers

    Intel Extension for Transformers

    Build your chatbot within minutes on your favorite device

    Intel Extension for Transformers is an innovative toolkit designed to accelerate Transformer-based models on Intel platforms, including CPUs and GPUs. It offers state-of-the-art compression techniques for Large Language Models (LLMs) and provides tools to build chatbots within minutes on various devices. The extension aims to optimize the performance of Transformer-based models, making them more efficient and accessible.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    TapeAgents

    TapeAgents

    A framework that facilitates all stages of LLM development

    TapeAgents is a framework that facilitates all stages of the Large Language Model (LLM) agent development lifecycle, providing tools for building, testing, and deploying AI agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    GLM-4

    GLM-4

    GLM-4 series: Open Multilingual Multimodal Chat LMs

    GLM-4 is a family of open models from ZhipuAI that spans base, chat, and reasoning variants at both 32B and 9B scales, with long-context support and practical local-deployment options. The GLM-4-32B-0414 models are trained on ~15T high-quality data (including substantial synthetic reasoning data), then post-trained with preference alignment, rejection sampling, and reinforcement learning to improve instruction following, coding, function calling, and agent-style behaviors. The...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    ReMe

    ReMe

    Memory Management Kit for Agents

    ReMe is a memory management kit for AI agents that gives them structured, persistent memory capabilities, enabling agents to extract, store, and reuse information across sessions, tasks, and interactions. It is designed to support long-running agent workflows where context matters and working memory alone isn’t enough, helping agents remember user preferences, task histories, and relevant past observations. The toolkit provides APIs to offload large, ephemeral outputs to external storage and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    files-to-prompt

    files-to-prompt

    Concatenate a directory full of files into a single prompt

    files-to-prompt is a Python command-line tool that takes one or more files or entire directories and concatenates their contents into a single, LLM-friendly prompt. It walks the directory tree, outputting each file preceded by its relative path and a separator, so a model can understand which content came from where. The tool is aimed at workflows where you want to ask an LLM questions about a whole codebase, documentation set, or notes folder without manually copying files together. It...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Manage and optimise Google, Facebook and Microsoft Ads faster and gain a competitive advantage with our digital advertising platform. Icon
    Manage and optimise Google, Facebook and Microsoft Ads faster and gain a competitive advantage with our digital advertising platform.

    Smarter, more effective advertising

    Slash the time it takes to manage and optimize your Google, Microsoft Advertising or Facebook Ads campaigns to just minutes a day. Adzooma's AI and machine learning based PPC platform offers stress free campaign management, state of the art 24/7 optimization and advanced automation, all in a simple to use interface. Scan for 50+ improvement 'opportunities', many of which can be actioned with a single click, track PPC performance and highlight over/under spending to improve your quality score, conversions and ROI. These trying times are tough for all. So we're giving away our whole award-winning platform for free until June 1st 2020. That's automated PPC ads, one-click optimisations, and world-class reporting - at zero cost. No strings attached. No credit card required.
    Free until June 1st 2020
  • 10
    Gemini Fullstack LangGraph Quickstart

    Gemini Fullstack LangGraph Quickstart

    Get started w/ building Fullstack Agents using Gemini 2.5 & LangGraph

    gemini-fullstack-langgraph-quickstart is a fullstack reference application from Google DeepMind’s Gemini team that demonstrates how to build a research-augmented conversational AI system using LangGraph and Google Gemini models. The project features a React (Vite) frontend and a LangGraph/FastAPI backend designed to work together seamlessly for real-time research and reasoning tasks. The backend agent dynamically generates search queries based on user input, retrieves information via the...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Everywhere

    Everywhere

    Context-aware desktop AI assistant that understands screen content

    Everywhere is a context-aware desktop AI assistant designed to interact directly with the content displayed on a user’s screen. It distinguishes itself from traditional AI tools by eliminating the need for manual input methods such as copying text or taking screenshots, instead allowing users to invoke assistance instantly through a shortcut. It can analyze on-screen information in real time and provide contextual responses, making it useful for tasks like troubleshooting errors, summarizing...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    Google Research

    Google Research

    This repository contains code released by Google Research

    Google Research is a massive monorepo that hosts a wide range of research code released by Google Research teams across machine learning, artificial intelligence, robotics, natural language processing, and other advanced domains. Rather than being a single framework, the repository serves as a centralized collection of experimental projects, reference implementations, and reproducible research artifacts. It is intended primarily for researchers and advanced practitioners who want to explore...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    CowAgent

    CowAgent

    AI assistant based on large models that can actively think and plan

    CowAgent, based on the chatgpt-on-wechat project, is an open-source AI agent framework that integrates large language models into the WeChat ecosystem to create intelligent conversational assistants. It enables automated message handling by connecting WeChat accounts with AI models that can generate contextual replies, process voice messages, and produce images directly inside chats. The platform has evolved beyond a simple chatbot into a more autonomous agent capable of planning complex...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    FireRedTTS-2

    FireRedTTS-2

    Long-form streaming TTS system for multi-speaker dialogue generation

    FireRedTTS2 is a next-generation open-source text-to-speech (TTS) system focused on long-form, streaming speech synthesis for multi-speaker dialogue, delivering stable natural speech with context-aware prosody and reliable speaker transitions that support real-time and conversational applications. It features a specialized streaming speech tokenizer and a dual-transformer architecture that enables low latency and high-quality synthesis, making it suitable for interactive systems like...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Spark TTS

    Spark TTS

    Spark-TTS Inference Code

    Spark TTS is an open-source, PyTorch-based text-to-speech inference system that leverages large language models to produce highly natural, intelligible speech from text input. It uses an efficient single-stream architecture where speech tokens are directly reconstructed from the predictions of an LLM, removing the need for external acoustic models or complex vocoders and making the generation pipeline cleaner and faster. The project supports zero-shot voice cloning, meaning it can imitate a...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    AutoCoder

    AutoCoder

    A long-running autonomous coding agent powered by the Claude Agent

    Autocoder is an experimental auto-generation engine that transforms high-level prompts or structured descriptions into functioning source code, models, or systems with minimal manual intervention. Rather than hand-writing boilerplate or repetitive patterns, users supply a specification—such as a description of a feature, a function prototype, or a module outline—and Autocoder fills in complete implementations that compile and run. It is built to support iterative refinement: after generating...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    TensorFlow Quantum is an open-source software framework designed for building and training hybrid quantum-classical machine learning models within the TensorFlow ecosystem. The framework enables researchers and developers to represent quantum circuits as data and integrate them directly into machine learning workflows. By combining classical deep learning techniques with quantum algorithms, the platform allows experimentation with quantum machine learning methods that may offer advantages...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    BeeAI Framework

    BeeAI Framework

    Build production-ready AI agents in both Python and Typescript

    ...It goes beyond simple prompt-based interactions by introducing rule-based governance and constraint enforcement, enabling developers to create agents with predictable and controllable behavior while still preserving advanced reasoning capabilities. The framework supports both Python and TypeScript with full feature parity, making it accessible to a wide range of developers and teams. It includes a unified backend layer that connects seamlessly to multiple large language model providers, allowing flexible deployment across different AI infrastructures without vendor lock-in. BeeAI also provides orchestration tools for designing dynamic workflows, enabling multiple agents to coordinate tasks through structured execution flows, retries, and parallel processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    VibeTensor

    VibeTensor

    Our first fully AI generated deep learning system

    ...It implements a PyTorch-style eager tensor library with a modern C++20 core that supports both CPU and CUDA backends, giving it the ability to manage tensors, automatic differentiation (autograd), and complex computation flows similar to mainstream frameworks. What makes VibeTensor remarkable is that every major component, from core libraries and dispatch systems to CUDA runtime support, caching allocators, and language bindings, was created and validated by coding agents using automated builds and tests rather than manual line-by-line human coding. The system includes both a Python frontend via a torch-like API and an experimental Node.js/TypeScript interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Agent Reinforcement Trainer

    Agent Reinforcement Trainer

    Train multi-step agents for real-world tasks using GRPO

    Agent Reinforcement Trainer, or ART is an open-source reinforcement learning framework tailored to training large language model agents through experience, making them more reliable and performant on multi-turn, multi-step tasks. Instead of just manually crafting prompts or relying on supervised fine-tuning, ART uses techniques like Group Relative Policy Optimization (GRPO) to let agents learn from environmental feedback and reward signals. The framework is designed to integrate easily with Python applications, abstracting much of the RL infrastructure so developers can train agents without deep RL expertise or heavy infrastructure overhead. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    MiniMax-M1

    MiniMax-M1

    Open-weight, large-scale hybrid-attention reasoning model

    MiniMax-M1 is presented as the world’s first open-weight, large-scale hybrid-attention reasoning model, designed to push the frontier of long-context, tool-using, and deeply “thinking” language models. It is built on the MiniMax-Text-01 foundation and keeps the same massive parameter budget, but reworks the attention and training setup for better reasoning and test-time compute scaling. Architecturally, it combines Mixture-of-Experts layers with lightning attention, enabling the model to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Jittor

    Jittor

    Jittor is a high-performance deep learning framework

    ...Jittor also contains a wealth of high-performance model libraries, including image recognition, detection, segmentation, generation, differentiable rendering, geometric learning, reinforcement learning, etc. The front-end language is Python. Module Design and Dynamic Graph Execution is used in the front-end, which is the most popular design for deep learning framework interface. The back-end is implemented by high-performance languages, such as CUDA, C++. Jittor'op is similar to NumPy. Let's try some operations. We create Var a and b via operation jt.float32, and add them. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Orpheus TTS

    Orpheus TTS

    Towards Human-Sounding Speech

    ...Inference is provided through a Python package that uses vLLM under the hood for high-throughput, low-latency generation, including streaming examples that show how to generate audio chunks in real time. The maintainers provide Colab notebooks, a standardized prompting format, and one-click deployment via Baseten for production-grade, FP8/FP16 optimized inference with ~200 ms streaming latency.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    RealtimeTTS

    RealtimeTTS

    Converts text to speech in realtime

    RealtimeTTS is a low-latency text-to-speech library built for real-time applications such as voice chat with LLMs, assistants, and interactive tools. It is designed around a streaming model: you can feed it text incrementally (for example, as an LLM responds) and get audio output almost immediately, which keeps end-to-end latency very low. The library is engine-agnostic and plugs into a wide range of cloud and local TTS systems, including OpenAI, ElevenLabs, Azure, Coqui, Piper, StyleTTS2,...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    ContextGem

    ContextGem

    ContextGem: Effortless LLM extraction from documents

    ContextGem is an open-source framework designed to simplify the extraction of structured data and insights from documents using large language models (LLMs). It provides a flexible, intuitive API that minimizes boilerplate code, enabling developers to build complex extraction workflows efficiently. ContextGem supports various document formats and integrates with multiple LLM providers, making it a versatile tool for tasks like contract analysis, anomaly detection, and information retrieval.​
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB