Showing 25 open source projects for "kernel-br"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    tt-metal

    tt-metal

    TT-NN operator library, and TT-Metalium low level kernel programming

    tt-metal, also referred to in its documentation as TT-Metalium, is Tenstorrent’s low-level software development kit for programming applications on Tenstorrent AI accelerators. The project is designed for developers who need direct access to the company’s Tensix processor architecture, exposing a programming model that is closer to hardware control than high-level inference frameworks. Instead of following a traditional GPU model centered on massive thread parallelism, the platform is built...
    Downloads: 31 This Week
    Last Update:
    See Project
  • 2
    DeepGEMM

    DeepGEMM

    Clean and efficient FP8 GEMM kernels with fine-grained scaling

    DeepGEMM is a specialized CUDA library for efficient, high-performance general matrix multiplication (GEMM) operations, with particular focus on low-precision formats such as FP8 (and experimental support for BF16). The library is designed to work cleanly and simply, avoiding overly templated or heavily abstracted code, while still delivering performance that rivals expert-tuned libraries. It supports both standard and “grouped” GEMMs, which is useful for architectures like Mixture of...
    Downloads: 167 This Week
    Last Update:
    See Project
  • 3
    Compute Library

    Compute Library

    The Compute Library is a set of computer vision and machine learning

    The Compute Library is a set of computer vision and machine learning functions optimized for both Arm CPUs and GPUs using SIMD technologies. The library provides superior performance to other open-source alternatives and immediate support for new Arm® technologies e.g. SVE2.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 4
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    FlashMLA is a high-performance decoding kernel library designed especially for Multi-Head Latent Attention (MLA) workloads, targeting NVIDIA Hopper GPU architectures. It provides optimized kernels for MLA decoding, including support for variable-length sequences, helping reduce latency and increase throughput in model inference systems using that attention style. The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to efficiently manage memory during decoding. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 5
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for deep learning inference applications. TensorRT-based applications perform up to 40X faster than CPU-only platforms during inference. With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers,...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 6
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    dlib

    dlib

    Toolkit for making machine learning and data analysis applications

    Dlib is a modern C++ toolkit containing machine learning algorithms and tools for creating complex software in C++ to solve real world problems. It is used in both industry and academia in a wide range of domains including robotics, embedded devices, mobile phones, and large high performance computing environments. Dlib's open source licensing allows you to use it in any application, free of charge. Good unit test coverage, the ratio of unit test lines of code to library lines of code is...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 8
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    This software was previously known as Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) and Deep Neural Network Library (DNNL). oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN is part of oneAPI. The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the following architectures: Arm* 64-bit Architecture (AArch64), NVIDIA* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V. oneDNN is intended for deep learning applications and framework developers interested in improving application performance on Intel CPUs and GPUs. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    Proximus for Ryzen AI

    Proximus for Ryzen AI

    Runtime extension of Proximus enabling Deployment on AMD Ryzen™ AI

    This project extends the Proximus development environment to support deployment of AI workloads on next-generation AMD Ryzen™ AI processors, such as the Ryzen™ AI 7 PRO 7840U featured in the Lenovo ThinkPad T14s Gen 4 ,one of the first true AI PCs with an onboard Neural Processing Unit (NPU) capable of 16 TOPS (trillion operations per second). Originally designed for use with Windows 11 Pro, this runtime was further enhanced to work under Linux environments, allowing developers and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Flashlight library

    Flashlight library

    A C++ standalone library for machine learning

    Flashlight is a fast, flexible machine learning library written entirely in C++ by Facebook AI Research and the creators of Torch, TensorFlow, Eigen, and Deep Speech. Native support in C++ and simple extensibility make Flashlight a powerful research framework that's hackable to its core and enables fast iteration on new experimental setups and algorithms with little unopinionated and without sacrificing performance. In a single repository, Flashlight provides apps for research across...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    MACE

    MACE

    Deep learning inference framework optimized for mobile platforms

    ...Chip-dependent power options like big.LITTLE scheduling, Adreno GPU hints are included as advanced APIs. UI responsiveness guarantee is sometimes obligatory when running a model. Mechanism like automatically breaking OpenCL kernel into small units is introduced to allow better preemption for the UI rendering task. Graph level memory allocation optimization and buffer reuse are supported. The core library tries to keep minimum external dependencies to keep the library footprint small.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    TurboTransformers

    TurboTransformers

    Fast and user-friendly runtime for transformer inference

    TurboTransformers is a high-performance inference framework optimized for running Transformer models efficiently on CPUs and GPUs. It improves latency and throughput for NLP applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    PyML is an interactive object oriented framework for machine learning written in python. PyML focuses on kernel classifiers, providing tools for feature selection, model selection, and methods for assessing classifier performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Adaptive Gaussian Filtering

    Adaptive Gaussian Filtering

    Machine learning with Gaussian kernels.

    Libagf is a machine learning library that includes adaptive kernel density estimators using Gaussian kernels and k-nearest neighbours. Operations include statistical classification, interpolation/non-linear regression and pdf estimation. For statistical classification there is a borders training feature for creating fast and general pre-trained models that nonetheless return the conditional probabilities.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    Graphlet kernel framework

    Calculates similarity between neighborhoods of two vertices in a graph

    This software package provides a framework for calculating similarity between neighborhoods rooted at two vertices of interest in a labeled graph (undirected or directed). The list of available similarity functions includes: cumulative random walk, standard random walk, standard graphlet kernel, edit distance graphlet kernel, label substitution graphlet kernel and edge indel graphlet kernel. The graphlet kernel framework can be used for vertex (node) classification in graphs, kernel-based clustering, or community detection. If you use this framework, please cite the following paper: Lugo-Martinez J, Radivojac P. Generalized graphlet kernels for probabilistic inference in sparse graphs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Azul OS

    Azul OS

    Azul OS version dev(Linux) IA

    ...Disponible # Azul dev rev 0.4.1 . Disponible [changelog] software added : php5-mysql gcc-c++ php5-gd php5-ctype perl-HTML-Tagset php5-zip php5-curl kernel-source mysql-connector-java php5-pear php5-mcrypt php5-ftp devel_C_C++ gimp gedit recode libreoffice MozillaFirefox wireshark audacity nano This work is licensed under a Creative Commons Attribution-NonCommercial 3.0 Unported License. #Blog : http://azul0.wordpress.com/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18

    Linux for Beagleboard-xm

    A Tailored Small Linux for Beagleboard-xm

    Beagleboard-xm is a powerful chip with a cortex-A8 CPU and a DSP. I have the plan to build an OCR gadget using it with Linux. As a by product I will post my tailored Linux kernel and u-boot, and all relevant stuff here, from now on. I was shocked by the blocking of Chinese citizens from accessing some of the contents on sourceforge. I deeply regret the outrageous action initiated, even though I fully understand the reasoning behind it.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    QVision: Computer Vision Library for Qt

    QVision: Computer Vision Library for Qt

    Computer vision and image processing library for Qt.

    ...Thanks to a large number of wrapper functions these objects can be used with highly efficient functionality from third party libraries such as OpenCV, GNU Scientific Library, Computational Geometry Algorithms Library, Intel's Math Kernel Library and Integrated Performance Primitives, the Octave library, etc...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    GenetiK is a generic framework that supports evolutionary algorithms, in particular Genetic Algorithms and Genetic Programming (and also Strongly Typed GP).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    EverCrack cryptanalysis engine. The goal is cryptanalysis of any class of cipher. The focus now is the kernel - cryptanalysis of unilateral monalpha substitution ciphers extremely quickly, and later, cryptanalysis of complex ciphers for the kernel.
    Leader badge
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Maiden-core is a Linux kernel robotics API module, extensible in Erc. The Maiden Project also documents and develops the Maiden architecture for mobile robots, a set of (optional) designs and standards for robots using the Maiden-core API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    This Meshula Kernel contains a reactive AI core, a high quality physics simulator, a math library, and assorted utility routines
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    This Project would create an automaton that could interact with humans or other automaton in the net. It is developed by creating a kernel and a plugin system tha add more capacity to interact with the enviroment that it is in.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Data Mining Platform is a platform for data mining and analysis. It contains many of the new and sophisticated methods such as kernel-based classification, two-way clustering, bayesian networks, pattern recognition for time series analysis and many other
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB