Open Source Machine Learning Software - Page 16

Machine Learning Software

View 446 business solutions
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    Simd Library

    Simd Library

    C++ image processing and machine learning library with using of SIMD

    The Simd Library is a free open-source image processing and machine learning library, designed for C and C++ programmers. It provides many useful high-performance algorithms for image processing such as pixel format conversion, image scaling and filtration, extraction of statistical information from images, motion detection, object detection and classification, neural networks. The algorithms are optimized with using of different SIMD CPU extensions. In particular, the library supports the following CPU extensions: SSE, AVX, AVX-512, and AMX for x86/x64, and NEON for ARM. The Simd Library has C API and also contains useful C++ classes and functions to facilitate access to C API. The library supports dynamic and static linking, 32-bit and 64-bit Windows and Linux, MSVS, G++ and Clang compilers, MSVS projects, and CMake build systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    SimpleTuner

    SimpleTuner

    A general fine-tuning kit geared toward image/video/audio diffusion

    SimpleTuner is an open-source toolkit designed to simplify the fine-tuning of modern diffusion models for generating images, video, and audio. The project focuses on providing a clear and understandable training environment for researchers, developers, and artists who want to customize generative AI models without navigating complex machine learning pipelines. It supports fine-tuning workflows for models such as Stable Diffusion variants and other diffusion architectures, enabling users to adapt pretrained models to specialized datasets or creative tasks. The system includes configuration-driven training processes that allow users to define datasets, model paths, and training parameters with minimal setup. SimpleTuner also emphasizes experimentation and academic collaboration, encouraging contributions and iterative improvements from the open-source community.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Spotlight

    Spotlight

    Deep recommender models using PyTorch

    Spotlight uses PyTorch to build both deep and shallow recommender models. By providing both a slew of building blocks for loss functions (various pointwise and pairwise ranking losses), representations (shallow factorization representations, deep sequence models), and utilities for fetching (or generating) recommendation datasets, it aims to be a tool for rapid exploration and prototyping of new recommender models. Spotlight offers a slew of popular datasets, including Movielens 100K, 1M, 10M, and 20M. It also incorporates utilities for creating synthetic datasets. For example, generate_sequential generates a Markov-chain-derived interaction dataset, where the next item a user chooses is a function of their previous interactions. Recommendations can be seen as a sequence prediction task: given the items a user has interacted with in the past, what will be the next item they will interact with? Spotlight provides a range of models.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    The Unsplash Dataset

    The Unsplash Dataset

    Unsplash images made available for research and machine learning

    The Unsplash Dataset is made up of over 350,000+ contributing global photographers and data sourced from hundreds of millions of searches across a nearly unlimited number of uses and contexts. Due to the breadth of intent and semantics contained within the Unsplash dataset, it enables new opportunities for research and learning.
    Downloads: 2 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 5
    TurboPilot

    TurboPilot

    Open source large-language-model based code completion engine

    TurboPilot is a self-hosted copilot clone that uses the library behind llama.cpp to run the 6 Billion Parameter Salesforce Codegen model in 4GiB of RAM. It is heavily based and inspired by on the fauxpilot project. This is a proof of concept right now rather than a stable tool. Autocompletion is quite slow in this version of the project. Feel free to play with it, but your mileage may vary.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Unity ML-Agents Toolkit

    Unity ML-Agents Toolkit

    Unity machine learning agents toolkit

    Train and embed intelligent agents by leveraging state-of-the-art deep learning technology. Creating responsive and intelligent virtual players and non-playable game characters is hard. Especially when the game is complex. To create intelligent behaviors, developers have had to resort to writing tons of code or using highly specialized tools. With Unity Machine Learning Agents (ML-Agents), you are no longer “coding” emergent behaviors, but rather teaching intelligent agents to “learn” through a combination of deep reinforcement learning and imitation learning. Using ML-Agents allows developers to create more compelling gameplay and an enhanced game experience. Advancement of artificial intelligence (AI) research depends on figuring out tough problems in existing environments using current benchmarks for training AI models. Using Unity and the ML-Agents toolkit, you can create AI environments that are physically, visually, and cognitively rich.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Velocity server

    Velocity server

    The modern, next-generation Minecraft server proxy

    Velocity is a high-performance Minecraft proxy server developed by the PaperMC project that allows multiple Minecraft servers to be linked together into a single network. Acting as an intermediary between players and backend servers, Velocity manages player connections and routes them to different game servers within a network. This architecture allows large Minecraft communities to run multiple servers for different game modes while presenting them as a unified system to players. The software is designed with a focus on performance, scalability, and modern architecture, allowing it to handle thousands of simultaneous players efficiently. Velocity also includes a plugin API that allows developers to extend the proxy with custom functionality and integrate it with existing server tools. Compared with older proxy solutions, the project emphasizes improved performance, reliability, and cleaner software design.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    WaveFunctionCollapse

    WaveFunctionCollapse

    Bitmap & tilemap generation from a single example

    This program generates bitmaps that are locally similar to the input bitmap. WFC initializes output bitmap in a completely unobserved state, where each pixel value is in superposition of colors of the input bitmap (so if the input was black & white then the unobserved states are shown in different shades of grey). The coefficients in these superpositions are real numbers, not complex numbers, so it doesn't do the actual quantum mechanics, but it was inspired by QM. Then the program goes into the observation-propagation cycle. It may happen that during propagation all the coefficients for a certain pixel become zero. That means that the algorithm has run into a contradiction and can not continue. The problem of determining whether a certain bitmap allows other nontrivial bitmaps satisfying condition (C1) is NP-hard, so it's impossible to create a fast solution that always finishes. In practice, however, the algorithm runs into contradictions surprisingly rarely.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    captcha_break

    captcha_break

    Identification codes

    This project will use Keras to build a deep convolutional neural network to identify the captcha verification code. It is recommended to use a graphics card to run the project. The following visualization codes are jupyter notebookall done in . If you want to write a python script, you can run it normally with a little modification. Of course, you can also remove these visualization codes. captcha is a library written in python to generate verification codes. It supports image verification codes and voice verification codes. We use its function of generating image verification codes. First, we set our verification code format to numbers and capital letters, and generate a string of verification codes. It is well known that tensorflow occupies all video memory by default, which is not conducive to us conducting multiple experiments at the same time, so we can use the following code when tensorflow uses the video memory it needs instead of directly occupying all video memory.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    python-small-examples

    python-small-examples

    Focus on creating classic Python small examples and cases

    python-small-examples is an open-source educational repository that contains hundreds of concise Python programming examples designed to illustrate practical coding techniques. The project focuses on teaching programming concepts through small, focused scripts that demonstrate common tasks in data processing, visualization, and general programming. Each example highlights a specific function or programming pattern so that learners can quickly understand how to apply Python features in real-world scenarios. The repository includes examples covering topics such as file processing, JSON manipulation, data visualization, and library usage. The examples are intentionally short and easy to read, making them useful for beginners who want to understand Python syntax and programming logic step by step. The repository is organized as a large collection of small scripts and notes that can be browsed individually without needing to study a full project.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    spaGO

    spaGO

    Self-contained Machine Learning and Natural Language Processing lib

    A Machine Learning library written in pure Go designed to support relevant neural architectures in Natural Language Processing. Spago is self-contained, in that it uses its own lightweight computational graph both for training and inference, easy to understand from start to finish. The core module of Spago relies only on testify for unit testing. In other words, it has "zero dependencies", and we are committed to keeping it that way as much as possible. Spago uses a multi-module workspace to ensure that additional dependencies are downloaded only when specific features (e.g. persistent embeddings) are used. A good place to start is by looking at the implementation of built-in neural models, such as the LSTM. Except for a few linear algebra operations written in assembly for optimal performance (a bit of copying from Gonum), it's straightforward Go code, so you don't have to worry.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    talos

    talos

    Hyperparameter Optimization for TensorFlow, Keras and PyTorch

    Talos radically changes the ordinary Keras, TensorFlow (tf.keras), and PyTorch workflow by fully automating hyperparameter tuning and model evaluation. Talos exposes Keras and TensorFlow (tf.keras) and PyTorch functionality entirely and there is no new syntax or templates to learn. Talos is made for data scientists and data engineers that want to remain in complete control of their TensorFlow (tf.keras) and PyTorch models, but are tired of mindless parameter hopping and confusing optimization solutions that add complexity instead of reducing it. Within minutes, without learning any new syntax, Talos allows you to configure, perform, and evaluate hyperparameter optimization experiments that yield state-of-the-art results across a wide range of prediction tasks. Talos provides the simplest and yet most powerful available method for hyperparameter optimization with TensorFlow (tf.keras) and PyTorch.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    jMIR

    jMIR

    Music research software

    jMIR is an open-source software suite implemented in Java for use in music information retrieval (MIR) research. It can be used to study music in the form of audio recordings, symbolic encodings and lyrical transcriptions, and can also mine cultural information from the Internet. It also includes tools for managing and profiling large music collections and for checking audio for production errors. jMIR includes software for extracting features, applying machine learning algorithms, applying heuristic error error checkers, mining metadata and analyzing metadata.
    Downloads: 29 This Week
    Last Update:
    See Project
  • 14
    This project contains weka packages of neural networks algorithms implementations like Learning Vector Quantizer (LVQ) and Self-organizing Maps (SOM). For more information about weka, please visit http://www.cs.waikato.ac.nz/~ml/weka/
    Leader badge
    Downloads: 38 This Week
    Last Update:
    See Project
  • 15
    openEAR is the Munich Open-Source Emotion and Affect Recognition Toolkit developed at the Technische Universität München (TUM). It provides efficient (audio) feature extraction algorithms implemented in C++, classfiers, and pre-trained models on well-known emotion databases. It is now maintained and supported by audEERING. Updates will follow soon.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 16
    Scikit Learn
    Machine Learning framework in Python
    Downloads: 10 This Week
    Last Update:
    See Project
  • 17
    OpenNN - Open Neural Networks Library

    OpenNN - Open Neural Networks Library

    Machine learning algorithms for advanced analytics

    OpenNN is a software library written in C++ for advanced analytics. It implements neural networks, the most successful machine learning method. Some typical applications of OpenNN are business intelligence (customer segmentation, churn prevention…), health care (early diagnosis, microarray analysis…) and engineering (performance optimization, predictive maitenance…). OpenNN does not deal with computer vision or natural language processing. The main advantage of OpenNN is its high performance. This library outstands in terms of execution speed and memory allocation. It is constantly optimized and parallelized in order to maximize its efficiency. The documentation is composed by tutorials and examples to offer a complete overview about the library. OpenNN is developed by Artelnics, a company specialized in artificial intelligence.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 18
    AI Cheatsheets

    AI Cheatsheets

    Essential Cheat Sheets for deep learning and machine learning research

    cheatsheets-ai is an open-source repository that collects essential cheat sheets covering many tools and concepts used in machine learning, deep learning, and data science. The project aims to provide quick-reference materials that help engineers, researchers, and students review key techniques and frameworks without reading extensive documentation. It compiles cheat sheets for widely used libraries and technologies such as TensorFlow, Keras, NumPy, Pandas, Scikit-learn, Matplotlib, and PySpark. These materials summarize common functions, workflows, and best practices in a concise visual format that makes them easy to consult during development or study sessions. The repository functions as a centralized library where users can quickly access reference materials for both machine learning theory and practical programming tools. Many of the cheat sheets are available as downloadable PDFs and images, allowing learners to keep them as quick references while working on projects.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Ai-Learn

    Ai-Learn

    The artificial intelligence learning roadmap compiles 200 cases

    Ai-Learn is an open-source artificial intelligence learning roadmap that aggregates educational materials, tutorials, and practical projects designed to help beginners study AI and machine learning systematically. The repository was created to help learners start self-study programs in artificial intelligence without getting overwhelmed by the large number of available resources. It organizes topics such as Python programming, mathematics for machine learning, data analysis, deep learning, computer vision, and natural language processing into a structured learning path. The project also provides a large collection of practical exercises and case studies that allow learners to apply theoretical knowledge through real projects. According to the repository description, it includes nearly two hundred hands-on AI examples developed through years of teaching experience.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    AliceMind

    AliceMind

    ALIbaba's Collection of Encoder-decoders from MinD

    This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba's MinD (Machine IntelligeNce of Damo) Lab. Pre-trained models for natural language understanding (NLU). We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the word and sentence levels, respectively. Pre-trained models for natural language generation (NLG). We propose a novel scheme that jointly pre-trains an autoencoding and autoregressive language model on a large unlabeled corpus, specifically designed for generating new text conditioned on context. It achieves new SOTA results in several downstream tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    AlphaTree

    AlphaTree

    DNN && GAN && NLP && BIG DATA

    AlphaTree is an educational repository that provides a visual roadmap of deep learning models and related artificial intelligence technologies. The project focuses on explaining the historical development and relationships between major neural network architectures used in modern machine learning. It presents diagrams and documentation describing the evolution of models such as LeNet, AlexNet, VGG, ResNet, DenseNet, and Inception networks. The repository organizes these architectures into a structured learning path that helps learners understand how deep learning models improved over time through changes in depth, architectural complexity, and training techniques. In addition to neural networks used for image classification, the project also references broader AI fields such as generative adversarial networks, natural language processing, and graph neural networks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Apache MXNet (incubating)

    Apache MXNet (incubating)

    A flexible and efficient library for deep learning

    Apache MXNet is an open source deep learning framework designed for efficient and flexible research prototyping and production. It contains a dynamic dependency scheduler that automatically parallelizes both symbolic and imperative operations. On top of this is a graph optimization layer, overall making MXNet highly efficient yet still portable, lightweight and scalable.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Asteroid

    Asteroid

    The PyTorch-based audio source separation toolkit for researchers

    The PyTorch-based audio source separation toolkit for researchers. Pytorch-based audio source separation toolkit that enables fast experimentation on common datasets. It comes with a source code thats supports a large range of datasets and architectures, and a set of recipes to reproduce some important papers. Building blocks are thought and designed to be seamlessly plugged together. Filterbanks, encoders, maskers, decoders and losses are all common building blocks that can be combined in a flexible way to create new systems. Extending the toolkit with new features is simple. Add a new filterbank, separator architecture, dataset or even recipe very easily. Recipes provide an easy way to reproduce results with data preparation, system design, training and evaluation in a single script. This is an essential tool for the community! The default logger is TensorBoard in all the recipes. From the recipe folder, you can run the following to visualize the logs of all your runs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    BERTScore

    BERTScore

    BERT score for text generation

    Automatic Evaluation Metric described in the paper BERTScore: Evaluating Text Generation with BERT (ICLR 2020). We now support about 130 models (see this spreadsheet for their correlations with human evaluation). Currently, the best model is Microsoft/debate-large-online, please consider using it instead of the default roberta-large in order to have the best correlation with human evaluation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    BentoML

    BentoML

    Unified Model Serving Framework

    BentoML simplifies ML model deployment and serves your models at a production scale. Support multiple ML frameworks natively: Tensorflow, PyTorch, XGBoost, Scikit-Learn and many more! Define custom serving pipeline with pre-processing, post-processing and ensemble models. Standard .bento format for packaging code, models and dependencies for easy versioning and deployment. Integrate with any training pipeline or ML experimentation platform. Parallelize compute-intense model inference workloads to scale separately from the serving logic. Adaptive batching dynamically groups inference requests for optimal performance. Orchestrate distributed inference graph with multiple models via Yatai on Kubernetes. Easily configure CUDA dependencies for running inference with GPU. Automatically generate docker images for production deployment.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB