283 projects for "computer based training" with 2 filters applied:

  • SoftCo: Enterprise Invoice and P2P Automation Software Icon
    SoftCo: Enterprise Invoice and P2P Automation Software

    For companies that process over 20,000 invoices per year

    SoftCo Accounts Payable Automation processes all PO and non-PO supplier invoices electronically from capture and matching through to invoice approval and query management. SoftCoAP delivers unparalleled touchless automation by embedding AI across matching, coding, routing, and exception handling to minimize the number of supplier invoices requiring manual intervention. The result is 89% processing savings, supported by a context-aware AI Assistant that helps users understand exceptions, answer questions, and take the right action faster.
    Learn More
  • Network Management Software and Tools for Businesses and Organizations | Auvik Networks Icon
    Network Management Software and Tools for Businesses and Organizations | Auvik Networks

    Mapping, inventory, config backup, and more.

    Reduce IT headaches and save time with a proven solution for automated network discovery, documentation, and performance monitoring. Choose Auvik because you'll see value in minutes, and stay with us to improve your IT for years to come.
    Learn More
  • 1
    Torch Pruning

    Torch Pruning

    DepGraph: Towards Any Structural Pruning

    ...Torch-Pruning physically removes parameters rather than masking them, which results in smaller and faster models during both training and inference. The toolkit supports a wide variety of architectures used in computer vision and large language models, making it a flexible solution for model compression tasks.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    slime LLM

    slime LLM

    slime is an LLM post-training framework for RL Scaling

    slime is an open-source large language model (LLM) post-training framework developed to support reinforcement learning (RL)-based scaling and high-performance training workflows for advanced LLMs, blending training and rollout modules into an extensible system. It offers a flexible architecture that connects high-throughput training (e.g., via Megatron-LM) with a customizable data generation pipeline, enabling researchers and engineers to iterate on new RL training paradigms effectively. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 3
    imgclsmob Deep learning networks

    imgclsmob Deep learning networks

    Sandbox for training deep learning networks

    imgclsmob is a deep learning research repository focused on implementing and experimenting with convolutional neural networks for computer vision tasks. The project serves as a sandbox for training and evaluating a wide variety of neural network architectures used in image analysis. It includes implementations of models used for tasks such as image classification, object detection, semantic segmentation, and pose estimation. The repository also contains scripts that help train models, evaluate performance, and convert trained networks between different frameworks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Kubeflow Trainer

    Kubeflow Trainer

    Distributed AI Model Training and LLM Fine-Tuning on Kubernetes

    ...The platform supports a wide range of machine learning frameworks, including PyTorch, JAX, Hugging Face, DeepSpeed, and XGBoost, making it highly flexible for different AI use cases. One of its key innovations is the integration of MPI-based distributed computing within Kubernetes, allowing efficient communication between nodes for high-performance training. It also includes advanced scheduling capabilities through integrations with tools like Kueue and Volcano, enabling topology-aware resource allocation and multi-cluster job orchestration.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Comet Backup - Fast, Secure Backup Software for MSPs Icon
    Comet Backup - Fast, Secure Backup Software for MSPs

    Fast, Secure Backup Software for Businesses and IT Providers

    Comet is a flexible backup platform, giving you total control over your backup environment and storage destinations.
    Learn More
  • 5
    RF-DETR

    RF-DETR

    RF-DETR is a real-time object detection and segmentation

    RF-DETR is an open-source computer vision framework that implements a real-time object detection and instance segmentation model based on transformer architectures. Developed by Roboflow, the project builds upon modern vision transformer backbones such as DINOv2 to achieve strong accuracy while maintaining efficient inference speeds suitable for real-time applications.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Vision Transformer Pytorch

    Vision Transformer Pytorch

    Implementation of Vision Transformer, a simple way to achieve SOTA

    This repository provides a from-scratch, minimalist implementation of the Vision Transformer (ViT) in PyTorch, focusing on the core architectural pieces needed for image classification. It breaks down the model into patch embedding, positional encoding, multi-head self-attention, feed-forward blocks, and a classification head so you can understand each component in isolation. The code is intentionally compact and modular, which makes it easy to tinker with hyperparameters, depth, width, and...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 7
    SwanLab

    SwanLab

    An open-source, modern-design AI training tracking and visualization

    SwanLab is an open-source experiment tracking and visualization platform designed to help machine learning engineers monitor, compare, and analyze the training of artificial intelligence models. The tool records training metrics, hyperparameters, model outputs, and experiment configurations so that developers can easily understand how different experiments perform over time. It provides a modern user interface for visualizing results, enabling teams to compare runs, track model performance...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    face.evoLVe

    face.evoLVe

    High-Performance Face Recognition Library on PaddlePaddle & PyTorch

    face.evoLVe is a high-performance face recognition library designed for research and real-world applications in computer vision. The project provides a comprehensive framework for building and training modern face recognition models using deep learning architectures. It includes components for face alignment, landmark localization, data preprocessing, and model training pipelines that allow developers to construct end-to-end facial recognition systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Ultralytics

    Ultralytics

    Ultralytics YOLO

    Ultralytics is a comprehensive computer vision framework that provides state-of-the-art implementations of the YOLO (You Only Look Once) family of models, enabling developers to perform tasks such as object detection, segmentation, classification, tracking, and pose estimation within a unified system. It is designed to be fast, accurate, and easy to use, offering both command-line and Python-based interfaces for training, validation, and deployment of machine learning models. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Get full visibility and control over your tasks and projects with Wrike. Icon
    Get full visibility and control over your tasks and projects with Wrike.

    A cloud-based collaboration, work management, and project management software

    Wrike offers world-class features that empower cross-functional, distributed, or growing teams take their projects from the initial request stage all the way to tracking work progress and reporting results.
    Learn More
  • 10
    VGGT

    VGGT

    [CVPR 2025 Best Paper Award] VGGT

    VGGT is a transformer-based framework aimed at unifying classic visual geometry tasks—such as depth estimation, camera pose recovery, point tracking, and correspondence—under a single model. Rather than training separate networks per task, it shares an encoder and leverages geometric heads/decoders to infer structure and motion from images or short clips. The design emphasizes consistent geometric reasoning: outputs from one head (e.g., correspondences or tracks) reinforce others (e.g., pose or depth), making the system more robust to challenging viewpoints and textures. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Matcha-TTS

    Matcha-TTS

    A fast TTS architecture with conditional flow matching

    Matcha-TTS is a non-autoregressive neural text-to-speech architecture that uses conditional flow matching to generate speech quickly while maintaining natural quality. It models speech as an ODE-based generative process, and conditional flow matching lets it reach high-quality audio in only a few synthesis steps, which greatly reduces latency compared to score-matching diffusion approaches. The model is fully probabilistic, so it can generate diverse realizations of the same text while still sounding stable and intelligible. The repository provides an end-to-end TTS pipeline: a PyTorch/Lightning training stack, configuration files, pre-trained checkpoints, a command-line interface, and a Gradio app for interactive testing. ...
    Downloads: 16 This Week
    Last Update:
    See Project
  • 12
    JiT

    JiT

    PyTorch implementation of JiT

    JiT is an open-source PyTorch implementation of a state-of-the-art image diffusion model designed around a minimalist yet powerful architecture for pixel-level generative modeling, based on the paper Back to Basics: Let Denoising Generative Models Denoise. Rather than predicting noise, JiT models directly predict clean image data, which the research suggests aligns better with the manifold structure of natural images and leads to stronger generative performance at high resolution. This implementation supports training on large datasets like ImageNet with configurable model variants, and practical scripts for setup, training, and evaluation on GPUs are included, leveraging PyTorch’s ecosystem for real-world experimentation. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    MaxText

    MaxText

    A simple, performant and scalable Jax LLM

    MaxText is a high-performance, highly scalable open-source framework designed to train and fine-tune large language models using the JAX ecosystem. The project acts as both a reference implementation and a practical training library that demonstrates best practices for building and scaling transformer-based language models on modern accelerator hardware. It is optimized to run efficiently on Google Cloud TPUs and GPUs, enabling researchers and engineers to train models ranging from small experiments to extremely large distributed workloads. The framework focuses on simplicity while still supporting advanced techniques such as model sharding, distributed computation, and high-throughput training pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    dLLM

    dLLM

    dLLM: Simple Diffusion Language Modeling

    dLLM is an open-source framework designed to simplify the development, training, and evaluation of diffusion-based large language models. Unlike traditional autoregressive models that generate text sequentially token by token, diffusion language models generate text through an iterative denoising process that refines masked tokens over multiple steps. This approach allows models to reason over the entire sequence simultaneously and potentially produce more coherent outputs with bidirectional context. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    OpenClaw-RL

    OpenClaw-RL

    Train any agents simply by 'talking'

    OpenClaw-RL is an open-source reinforcement learning framework designed to train and personalize AI agents built on the OpenClaw ecosystem. The project focuses on enabling agents to improve their behavior through interactive learning rather than relying solely on static prompts or predefined skills. One of its key ideas is allowing users to train an AI agent simply by interacting with it conversationally, using natural language feedback to guide the learning process. The system incorporates...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    TextWorld

    TextWorld

    ​TextWorld is a sandbox learning environment for the training

    TextWorld is a learning environment designed to train reinforcement learning agents to play text-based games, where actions and observations are entirely in natural language. Developed by Microsoft Research, TextWorld focuses on language understanding, planning, and interaction in complex, narrative-driven environments. It generates games procedurally, enabling scalable testing of agents’ natural language processing and decision-making abilities.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 17
    Synthetic Data Generator

    Synthetic Data Generator

    SDG is a specialized framework

    ...The platform enables developers and data scientists to create artificial datasets that preserve important relationships between variables without containing sensitive personal information. This makes the generated data suitable for tasks such as machine learning model training, testing software systems, sharing datasets across organizations, and conducting research without violating privacy regulations. The system supports multiple generation methods including statistical models, generative adversarial networks, and large language model–based synthesis. It also includes a data processing module capable of handling different data types, preprocessing columns, managing missing values, and converting formats automatically before model training.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 18
    RL Games

    RL Games

    RL implementations

    rl_games is a high-performance reinforcement learning framework optimized for GPU-based training, particularly in environments like robotics and continuous control tasks. It supports advanced algorithms and is built with PyTorch.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 19
    SimpleTuner

    SimpleTuner

    A general fine-tuning kit geared toward image/video/audio diffusion

    SimpleTuner is an open-source toolkit designed to simplify the fine-tuning of modern diffusion models for generating images, video, and audio. The project focuses on providing a clear and understandable training environment for researchers, developers, and artists who want to customize generative AI models without navigating complex machine learning pipelines. It supports fine-tuning workflows for models such as Stable Diffusion variants and other diffusion architectures, enabling users to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    SAM 2

    SAM 2

    The repository provides code for running inference with SAM 2

    SAM2 is a next-generation version of the Segment Anything Model (SAM), designed to improve performance, generalization, and efficiency in promptable image segmentation tasks. It retains the core promptable interface—accepting points, boxes, or masks—but incorporates architectural and training enhancements to produce higher-fidelity masks, better boundary adherence, and robustness to complex scenes. The updated model is optimized for faster inference and lower memory use, enabling real-time...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 22
    RAG-Retrieval

    RAG-Retrieval

    Unify Efficient Fine-tuning of RAG Retrieval, including Embedding

    ...It includes implementations of modern embedding architectures designed to map documents and queries into vector spaces for efficient similarity search. The framework also supports reranking models that refine retrieved results using large language models or lightweight transformer architectures. Additional training techniques such as preference-based supervised fine-tuning and embedding distillation are included to improve retrieval quality.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    EnvPool

    EnvPool

    C++-based high-performance parallel environment execution engine

    EnvPool is a fast, asynchronous, and parallel RL environment library designed for scaling reinforcement learning experiments. Developed by SAIL at Singapore, it leverages C++ backend and Python frontend for extremely high-speed environment interaction, supporting thousands of environments running in parallel on a single machine. It's compatible with Gymnasium API and RLlib, making it suitable for scalable training pipelines.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 24
    NVIDIA PhysicsNeMo

    NVIDIA PhysicsNeMo

    Open-source deep-learning framework for building and training

    ...The framework focuses on the emerging field of physics-informed machine learning, where neural networks are used alongside physical equations to model complex scientific systems. PhysicsNeMo provides modular Python components that allow developers to create scalable training and inference pipelines for models that combine data-driven learning with physics-based constraints. It is built on top of the PyTorch ecosystem and integrates with GPU-accelerated computing environments to handle computationally demanding simulations and datasets. The framework supports a wide range of scientific applications, including computational fluid dynamics, climate modeling, weather prediction, and engineering simulations.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 25
    MetaCLIP

    MetaCLIP

    ICLR2024 Spotlight: curation/training code, metadata, distribution

    MetaCLIP is a research codebase that extends the CLIP framework into a meta-learning / continual learning regime, aiming to adapt CLIP-style models to new tasks or domains efficiently. The goal is to preserve CLIP’s strong zero-shot transfer capability while enabling fast adaptation to domain shifts or novel class sets with minimal data and without catastrophic forgetting. The repository provides training logic, adaptation strategies (e.g. prompt tuning, adapter modules), and evaluation...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB