142 projects for "train" with 1 filter applied:

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    OpenClaw-RL

    OpenClaw-RL

    Train any agents simply by 'talking'

    OpenClaw-RL is an open-source reinforcement learning framework designed to train and personalize AI agents built on the OpenClaw ecosystem. The project focuses on enabling agents to improve their behavior through interactive learning rather than relying solely on static prompts or predefined skills. One of its key ideas is allowing users to train an AI agent simply by interacting with it conversationally, using natural language feedback to guide the learning process.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    ...Quality is a recurring theme: examples and utilities help filter low-value samples, enforce length limits, and split train/validation consistently so results are comparable. Licensing and provenance are surfaced to encourage compliant usage and to guide dataset selection in commercial settings. For practitioners, the repo is a practical “starting pantry” that accelerates experimentation and helps keep data wrangling from dominating the project timeline.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 3
    Agent Reinforcement Trainer

    Agent Reinforcement Trainer

    Train multi-step agents for real-world tasks using GRPO

    ...Instead of just manually crafting prompts or relying on supervised fine-tuning, ART uses techniques like Group Relative Policy Optimization (GRPO) to let agents learn from environmental feedback and reward signals. The framework is designed to integrate easily with Python applications, abstracting much of the RL infrastructure so developers can train agents without deep RL expertise or heavy infrastructure overhead. ART also supports scalable training patterns, observability tools, and integration with hosted platforms like Weights & Biases, and it provides notebooks that demonstrate training on standard benchmarks and tasks.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 4
    WeChatMsg

    WeChatMsg

    Project aimed at extracting, exporting, and analyzing chat records

    ...Beyond simple export, the project includes mechanisms for analyzing chat histories and generating annual reports or visual summaries about messaging trends, interaction patterns, and more. The original README communicates a guiding philosophy about owning personal data and using it responsibly to train personalized AI agents or preserve memories. Although the repository has seen periods of inactivity and may not receive frequent updates, its widespread use indicates community interest in preserving chat logs and understanding conversation data outside of the WeChat interface.
    Downloads: 311 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Claude Code

    Claude Code

    Claude Code is an agentic coding tool that lives in your terminal

    ...The tool is designed to simplify development by automating repetitive work and providing instant clarifications on code behavior. User feedback and usage data are collected responsibly, with strict privacy safeguards and limited retention, ensuring no feedback is used to train generative models. Claude Code is open and actively maintained with community-driven bug reporting and feature requests. Its natural language interface makes advanced coding workflows accessible without leaving your coding environment.
    Downloads: 245 This Week
    Last Update:
    See Project
  • 6
    verl-agent

    verl-agent

    Designed for training LLM/VLM agents via RL

    ...This step-wise interaction model makes it possible to train agents to operate in long-horizon scenarios where decisions depend on cumulative context and previous outcomes. Developers can configure memory modules that determine how historical information is stored and incorporated into each step of the reasoning process.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    SurvivalManual

    SurvivalManual

    Libre Survival Manual for Android with offline in mind

    ...But it doesn't have to be used only in emergency situations, it can also be useful for outdoor trips, walks, camps, and learning about nature and yourself truly. This is not only fun, but you can also train skills (fire, build shelter, ...) that you may need in a catastrophe. Some things work best with practice in a relaxed environment, so you also have time for some experiments. The refugees also are welcome to use this application to prepare and guide you for your dangerous journey. Although I hope that we as humans will come to feel and stop wars and end climate injustice so that people do not have to flee and be afraid.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    TextWorld

    TextWorld

    ​TextWorld is a sandbox learning environment for the training

    TextWorld is a learning environment designed to train reinforcement learning agents to play text-based games, where actions and observations are entirely in natural language. Developed by Microsoft Research, TextWorld focuses on language understanding, planning, and interaction in complex, narrative-driven environments. It generates games procedurally, enabling scalable testing of agents’ natural language processing and decision-making abilities.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 9
    AutoTrain Advanced

    AutoTrain Advanced

    Faster and easier training and deployments

    AutoTrain Advanced is an open-source machine learning training framework developed by Hugging Face that simplifies the process of training and fine-tuning state-of-the-art AI models. The project provides a no-code and low-code interface that allows users to train models using custom datasets without needing extensive expertise in machine learning engineering. It supports a wide range of tasks including text classification, sequence-to-sequence modeling, token classification, sentence embedding training, and large language model fine-tuning. The system integrates closely with the Hugging Face ecosystem and allows developers to train models using datasets hosted on the Hugging Face Hub. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • 10
    ReCall

    ReCall

    Learning to Reason with Search for LLMs via Reinforcement Learning

    ...Instead of relying purely on static knowledge stored inside the model, ReCall allows the language model to dynamically decide when it should retrieve information or invoke external capabilities during the reasoning process. The framework uses reinforcement learning to train models to perform these tool calls effectively while solving multi-step reasoning tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MaxText

    MaxText

    A simple, performant and scalable Jax LLM

    MaxText is a high-performance, highly scalable open-source framework designed to train and fine-tune large language models using the JAX ecosystem. The project acts as both a reference implementation and a practical training library that demonstrates best practices for building and scaling transformer-based language models on modern accelerator hardware. It is optimized to run efficiently on Google Cloud TPUs and GPUs, enabling researchers and engineers to train models ranging from small experiments to extremely large distributed workloads. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    RL Baselines3 Zoo

    RL Baselines3 Zoo

    Training framework for Stable Baselines3 reinforcement learning agents

    rl-baselines3-zoo is a collection of pre-trained models, benchmarks, and hyperparameter tuning tools built on top of Stable Baselines3, a reinforcement learning library. It provides an easy way to test, evaluate, and train RL agents across a wide variety of environments.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 13
    DeepSpeed

    DeepSpeed

    Deep learning optimization library: makes distributed training easy

    DeepSpeed is an easy-to-use deep learning optimization software suite that enables unprecedented scale and speed for Deep Learning Training and Inference. With DeepSpeed you can: 1. Train/Inference dense or sparse models with billions or trillions of parameters 2. Achieve excellent system throughput and efficiently scale to thousands of GPUs 3. Train/Inference on resource constrained GPU systems 4. Achieve unprecedented low latency and high throughput for inference 5. Achieve extreme compression for an unparalleled inference latency and model size reduction with low costs DeepSpeed offers a confluence of system innovations, that has made large scale DL training effective, and efficient, greatly improved ease of use, and redefined the DL training landscape in terms of scale that is possible. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    Datumaro

    Datumaro

    Dataset Management Framework, a Python library and a CLI tool to build

    Datumaro is a flexible Python-based dataset management framework and command-line tool for building, analyzing, transforming, and converting computer vision datasets in many popular formats. It supports importing and exporting annotations and images across a wide variety of standards like COCO, PASCAL VOC, YOLO, ImageNet, Cityscapes, and many more, enabling easy integration with different training pipelines and tools. Datumaro makes it easy to merge datasets, split them into...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 15
    spacy-transformers

    spacy-transformers

    Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

    ...You can convert word vectors from popular tools like FastText and Gensim, or you can load in any pre trained transformer model if you install spacy-transformers. You can also do your own language model pretraining via the spacy pre train command. You can even share your transformer or another contextual embedding model across multiple components, which can make long pipelines several times more efficient. To use transfer learning, you’ll need at least a few annotated examples for what you’re trying to predict.
    Downloads: 24 This Week
    Last Update:
    See Project
  • 16
    LLM101n

    LLM101n

    LLM101n: Let's build a Storyteller

    ...The materials favor compact, readable code and incremental steps, so learners can verify each concept before moving on. You’ll see how data pipelines, batching, masking, and positional encodings fit together to train a small GPT-style model end to end. The repo often complements explanations with runnable notebooks or scripts, encouraging experimentation and modification. By the end, the focus is less on polishing a production system and more on internalizing how LLM components interact to produce coherent text.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    Matcha-TTS

    Matcha-TTS

    A fast TTS architecture with conditional flow matching

    ...The repository provides an end-to-end TTS pipeline: a PyTorch/Lightning training stack, configuration files, pre-trained checkpoints, a command-line interface, and a Gradio app for interactive testing. Users can train on standard datasets like LJSpeech or plug in their own corpora, with helper tools for computing dataset statistics, extracting phoneme durations, and running multi-GPU training.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 18
    Habitat-Sim

    Habitat-Sim

    A flexible, high-performance 3D simulator for Embodied AI research

    ...Determinism and reproducibility are first-class goals, which is critical for benchmarking agents and comparing algorithms. Thanks to its speed and modular design, Habitat-Sim is widely used to prototype embodied agents, train at scale, and evaluate in standardized environments with consistent metrics.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    MiniMind-V is an experimental open-source project that aims to train a very small multimodal vision–language model (VLM) from scratch with extremely low compute and cost, making research and experimentation accessible to more people. The repository showcases training workflows and code designed to produce a 26-million parameter model—including both image and text capabilities—using minimal resources in very little time, reflecting a trend toward democratizing AI research.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    RF-DETR

    RF-DETR

    RF-DETR is a real-time object detection and segmentation

    ...RF-DETR emphasizes strong performance across both accuracy and latency benchmarks, allowing developers to deploy high-quality detection models in applications that require immediate processing such as robotics, autonomous systems, and industrial inspection. The repository includes Python packages, training scripts, and model configurations that enable researchers and engineers to train and deploy detection models on custom datasets.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    verl

    verl

    Volcano Engine Reinforcement Learning for LLMs

    VERL is a reinforcement-learning–oriented toolkit designed to train and align modern AI systems, from language models to decision-making agents. It brings together supervised fine-tuning, preference modeling, and online RL into one coherent training stack so teams can move from raw data to aligned policies with minimal glue code. The library focuses on scalability and efficiency, offering distributed training loops, mixed precision, and replay/buffering utilities that keep accelerators busy. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    FlagEmbedding

    FlagEmbedding

    Retrieval and Retrieval-augmented LLMs

    ...FlagEmbedding includes a family of models known as BGE (BAAI General Embedding), which are designed to achieve strong performance across multilingual and cross-lingual retrieval benchmarks. The toolkit provides infrastructure for inference, fine-tuning, evaluation, and dataset preparation, enabling developers to train custom embedding models for specific domains or applications. It also includes reranker models that refine search results by re-evaluating candidate documents using cross-encoder architectures, improving retrieval accuracy in complex queries.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    Humanoid-Gym

    Humanoid-Gym

    Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real

    Humanoid-Gym is a reinforcement learning framework designed to train locomotion and control policies for humanoid robots using high-performance simulation environments. The system is built on top of NVIDIA Isaac Gym, which allows large-scale parallel simulation of robotic environments directly on GPU hardware. Its primary goal is to enable efficient training of humanoid robots in simulation while enabling policies to transfer effectively to real-world hardware without additional training. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Xtuner

    Xtuner

    A Next-Generation Training Engine Built for Ultra-Large MoE Models

    ...The engine supports training models with hundreds of billions of parameters and enables long-context training with sequence lengths reaching tens of thousands of tokens. Its architecture incorporates memory-efficient optimizations that allow researchers to train large models even when computational resources are limited. XTuner is also designed to integrate with modern AI ecosystems, supporting multimodal training, reinforcement learning optimization, and instruction tuning pipelines.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    Megatron-LM

    Megatron-LM

    Ongoing research training transformer models at scale

    Megatron-LM is a GPU-optimized deep learning framework from NVIDIA designed to train extremely large transformer-based language models efficiently at scale. The repository provides both a reference training implementation and Megatron Core, a composable library of high-performance building blocks for custom large-model pipelines. It supports advanced parallelism strategies including tensor, pipeline, data, expert, and context parallelism, enabling training across massive multi-GPU and multi-node clusters. ...
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB