Showing 383 open source projects for "cuda-gpumemtest"

View related business solutions
  • Get full visibility and control over your tasks and projects with Wrike. Icon
    Get full visibility and control over your tasks and projects with Wrike.

    A cloud-based collaboration, work management, and project management software

    Wrike offers world-class features that empower cross-functional, distributed, or growing teams take their projects from the initial request stage all the way to tracking work progress and reporting results.
    Learn More
  • Jscrambler: Pioneering Client-Side Protection Platform Icon
    Jscrambler: Pioneering Client-Side Protection Platform

    Jscrambler offers an exclusive blend of cutting-edge first-party JavaScript obfuscation and state-of-the-art third-party tag protection.

    Jscrambler is the leader in Client-Side Protection and Compliance. We were the first to merge advanced polymorphic JavaScript obfuscation with fine-grained third-party tag protection in a unified Client-Side Protection and Compliance Platform. Our integrated solution ensures a robust defense against current and emerging client-side cyber threats, data leaks, and IP theft, empowering software development and digital teams to innovate securely. With Jscrambler, businesses adopt a unified, future-proof client-side security policy all while achieving compliance with emerging security standards including PCI DSS v4.0. Trusted by digital leaders worldwide, Jscrambler gives businesses the freedom to innovate securely.
    Learn More
  • 1
    CV-CUDA

    CV-CUDA

    CV-CUDA™ is an open-source, GPU accelerated library

    CV-CUDA is an open-source project that enables building efficient cloud-scale Artificial Intelligence (AI) imaging and computer vision (CV) applications. It uses graphics processing unit (GPU) acceleration to help developers build highly efficient pre- and post-processing pipelines. CV-CUDA originated as a collaborative effort between NVIDIA and ByteDance.
    Downloads: 34 This Week
    Last Update:
    See Project
  • 2
    CUDA-QX

    CUDA-QX

    Accelerated libraries for quantum-classical computing built on CUDA-Q

    CUDA-QX is a collection of accelerated libraries built on top of the CUDA-Q platform, designed to enable rapid development of hybrid quantum-classical applications. It extends the CUDA-Q programming model by providing optimized implementations of domain-specific quantum computing primitives and workflows. The libraries are intended to help researchers and developers leverage GPUs, CPUs, and quantum processing units together in a unified computational model.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 3
    CUDA Python

    CUDA Python

    Performance meets Productivity

    CUDA Python is a unified Python interface for accessing and working with the NVIDIA CUDA platform, enabling developers to build GPU-accelerated applications entirely in Python. It acts as a metapackage composed of multiple submodules that provide both high-level and low-level access to CUDA functionality, including runtime APIs, driver APIs, and JIT compilation tools.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    CUDA-Q

    CUDA-Q

    C++ and Python support for the CUDA Quantum programming model

    CUDA-Q is an open-source platform for developing hybrid quantum-classical applications using a unified programming model across CPUs, GPUs, and quantum processing units. It provides a full toolchain that includes compilers, runtimes, and libraries for writing quantum programs in both C++ and Python. The platform is designed to be hardware-agnostic, allowing developers to run applications on different quantum backends or simulate them efficiently using GPU acceleration when physical quantum hardware is unavailable. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Budgyt Is The Highest Rated Business Budgeting Software In The Market. Icon
    Budgyt Is The Highest Rated Business Budgeting Software In The Market.

    Affordable budgeting software for companies with multiple users and multiple departments.

    Budgyt is an easy to use, intuitive platform with a clean simple interface that makes budgeting multiple P&L’s easy to do without needing Excel.
    Book a Demo
  • 5
    CUDA Agent

    CUDA Agent

    Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

    CUDA Agent is a research-driven agentic reinforcement learning system designed to automatically generate and optimize high-performance CUDA kernels for GPU workloads. The project addresses the long-standing challenge that efficient CUDA programming typically requires deep hardware expertise by training an autonomous coding agent capable of iterative improvement through execution feedback.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    CUDA.jl

    CUDA.jl

    CUDA programming in Julia

    High-performance GPU programming in a high-level language. JuliaGPU is a GitHub organization created to unify the many packages for programming GPUs in Julia. With its high-level syntax and flexible compiler, Julia is well-positioned to productively program hardware accelerators like GPUs without sacrificing performance. The latest development version of CUDA.jl requires Julia 1.8 or higher. If you are using an older version of Julia, you need to use a previous version of CUDA.jl. This will...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Numba CUDA Target

    Numba CUDA Target

    The CUDA target for Numba

    Numba CUDA Target is NVIDIA’s maintained CUDA backend for the Numba JIT compiler, enabling developers to write GPU-accelerated code directly in Python. It allows users to define CUDA kernels using Python syntax, which are then compiled into efficient GPU code at runtime using LLVM-based toolchains. This approach significantly lowers the barrier to entry for GPU programming by eliminating the need to write CUDA C++ while still delivering high performance. ...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 8
    CUDA API Wrappers

    CUDA API Wrappers

    Thin, unified, C++-flavored wrappers for the CUDA APIs

    ...In a nutshell - making CUDA API work more fun.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    CUDA Core Compute Libraries (CCCL)

    CUDA Core Compute Libraries (CCCL)

    CUDA Core Compute Libraries

    CCCL, or CUDA Core Compute Libraries, is a unified repository that consolidates several foundational CUDA C++ libraries into a single, cohesive development platform. It brings together Thrust, CUB, and libcudacxx, which collectively provide high-level abstractions, low-level performance primitives, and a CUDA-compatible standard library for GPU programming.
    Downloads: 13 This Week
    Last Update:
    See Project
  • Resco toolkit for building mobile apps Icon
    Resco toolkit for building mobile apps

    A no-code toolkit for building responsive and resilient mobile business applications for Microsoft Power Platform, Dynamics 365, Dataverse and Salesfo

    Deploying mobile apps with Resco takes days, not months—all without writing a single line of code. Workers can download the Resco app from AppStore, Google Play, or Windows Store, log into your company environment, and instantly use the app you have published on any device.
    Learn More
  • 10
    CuPy

    CuPy

    A NumPy-compatible array library accelerated by CUDA

    CuPy is an open source implementation of NumPy-compatible multi-dimensional array accelerated with NVIDIA CUDA. It consists of cupy.ndarray, a core multi-dimensional array class and many functions on it. CuPy offers GPU accelerated computing with Python, using CUDA-related libraries to fully utilize the GPU architecture. According to benchmarks, it can even speed up some operations by more than 100X. CuPy is highly compatible with NumPy, serving as a drop-in replacement in most cases. ...
    Downloads: 22 This Week
    Last Update:
    See Project
  • 11
    how-to-optim-algorithm-in-cuda

    how-to-optim-algorithm-in-cuda

    How to optimize some algorithm in cuda

    how-to-optim-algorithm-in-cuda is an open educational repository focused on teaching developers how to optimize algorithms for high-performance execution on GPUs using CUDA. The project combines technical notes, code examples, and practical experiments that demonstrate how common computational kernels can be optimized to improve speed and memory efficiency. Instead of presenting only theoretical explanations, the repository includes hand-written CUDA implementations of fundamental operations such as reductions, element-wise computations, softmax, and attention mechanisms. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Tiny CUDA Neural Networks

    Tiny CUDA Neural Networks

    Lightning fast C++/CUDA neural network framework

    ...It will likely only work on an RTX 3090, an RTX 2080 Ti, or high-end enterprise GPUs. Lower-end cards must reduce the n_neurons parameter or use the CutlassMLP (better compatibility but slower) instead. tiny-cuda-nn comes with a PyTorch extension that allows using the fast MLPs and input encodings from within a Python context. These bindings can be significantly faster than full Python implementations; in particular for the multiresolution hash encoding.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Numbast

    Numbast

    Build an automated pipeline that converts CUDA APIs into Numba

    Numbast is an automated toolchain that bridges CUDA C++ and Python by generating Numba-compatible bindings directly from CUDA header files. Its primary goal is to eliminate the manual effort required to expose CUDA libraries to Python, enabling developers to use GPU-accelerated functionality in Python environments more easily. The system parses CUDA C++ declarations and converts them into Python bindings that can be used within Numba, allowing seamless integration with Python-based GPU workflows. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    CUDA Containers for Edge AI & Robotics

    CUDA Containers for Edge AI & Robotics

    Machine Learning Containers for NVIDIA Jetson and JetPack-L4T

    CUDA Containers for Edge AI & Robotics is an open-source project that provides a modular container build system designed for running machine learning and AI workloads on NVIDIA Jetson devices. The repository contains container configurations that package the latest AI frameworks and dependencies optimized for Jetson hardware. These containers simplify the deployment of complex machine learning environments by bundling libraries such as CUDA, TensorRT, and deep learning frameworks into reproducible container images. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    NVIDIA Warp

    NVIDIA Warp

    A Python framework for accelerated simulation, data generation

    NVIDIA Warp is a high-performance Python framework developed by NVIDIA for building and accelerating simulation, graphics, and physics-based workloads using GPU computing. It enables developers to write kernel-level code in Python that is automatically compiled into efficient CUDA kernels, combining ease of use with near-native performance. The framework is designed for applications such as robotics, reinforcement learning, physical simulation, and differentiable computing, where performance and flexibility are critical. Warp provides a set of primitives for working with arrays, geometry, and physics operations, allowing users to implement complex simulations without writing low-level CUDA code directly. ...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 16
    XMRig

    XMRig

    RandomX, KawPow, CryptoNight, AstroBWT and GhostRider unified miner

    High performance, open-source, cross-platform RandomX, KawPow, CryptoNight, and AstroBWT CPU/GPU miner, RandomX benchmark, and stratum proxy. XMRig is a high-performance, open-source, cross-platform RandomX, KawPow, CryptoNight, and AstroBWT unified CPU/GPU miner and RandomX benchmark. Official binaries are available for Windows, Linux, macOS, and FreeBSD. The preferred way to configure the miner is the JSON config file as it is more flexible and human-friendly. The command-line interface...
    Downloads: 183 This Week
    Last Update:
    See Project
  • 17
    AIMr

    AIMr

    The best AI Aimbot for Fortnite, Valorant, CS2, R6, COD, Apex, & more

    ...AIMr also provides visual customization options like field-of-view displays and detection indicators, allowing players to tailor their interface. The system is compatible with games that use human-shaped models, and although it functions effectively out of the box, optimizing it with CUDA-accelerated OpenCV is recommended for maximum performance.
    Downloads: 311 This Week
    Last Update:
    See Project
  • 18
    CUTLASS

    CUTLASS

    CUDA Templates for Linear Algebra Subroutines

    CUTLASS is a collection of CUDA C++ template abstractions for implementing high-performance matrix-multiplication (GEMM) and related computations at all levels and scales within CUDA. It incorporates strategies for hierarchical decomposition and data movement similar to those used to implement cuBLAS and cuDNN. CUTLASS decomposes these "moving parts" into reusable, modular software components abstracted by C++ template classes.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    KeyKiller-Cuda

    KeyKiller-Cuda

    Solving the Satoshi Puzzle

    KeyKiller is a GPU-accelerated version of the KeyKiller project, designed to achieve extreme performance in solving Satoshi Nakamoto's puzzles using modern NVIDIA GPUs. KeyKiller CUDA pushes the limits of cryptographic key search performance by leveraging CUDA, thread-beam parallelism, and batch EC operations. The command-line version is open-source and free to use. For the paid advanced graphics version, please visit: https://gitlab.com/8891689/KeyKiller-Cuda/
    Downloads: 13 This Week
    Last Update:
    See Project
  • 20
    GPU Puzzles

    GPU Puzzles

    Solve puzzles. Learn CUDA

    ...Instead of presenting traditional lecture-style explanations, the project immerses learners directly in hands-on programming tasks that demonstrate how GPU computation works. The exercises are implemented using Python with the Numba CUDA interface, which allows Python code to compile into GPU kernels that run on CUDA-enabled hardware. By solving progressively more complex puzzles, learners gain a practical understanding of how parallel algorithms operate on graphics processing units. The project emphasizes experimentation and problem solving, encouraging learners to discover GPU programming techniques through trial and exploration. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    TensorRT Backend For ONNX

    TensorRT Backend For ONNX

    ONNX-TensorRT: TensorRT backend for ONNX

    ...For building within docker, we recommend using and setting up the docker containers as instructed in the main (TensorRT repository). Note that this project has a dependency on CUDA. By default the build will look in /usr/local/cuda for the CUDA toolkit installation. If your CUDA path is different, overwrite the default path. ONNX models can be converted to serialized TensorRT engines using the onnx2trt executable.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    DeepGEMM

    DeepGEMM

    Clean and efficient FP8 GEMM kernels with fine-grained scaling

    DeepGEMM is a specialized CUDA library for efficient, high-performance general matrix multiplication (GEMM) operations, with particular focus on low-precision formats such as FP8 (and experimental support for BF16). The library is designed to work cleanly and simply, avoiding overly templated or heavily abstracted code, while still delivering performance that rivals expert-tuned libraries.
    Downloads: 92 This Week
    Last Update:
    See Project
  • 23
    Taichi

    Taichi

    Productive, portable, and performant GPU programming in Python

    ...It uses JIT compilation (via LLVM and its runtime TiRT) to offload compute-heavy code to CPUs, GPUs, mobile devices, and embedded systems. With built-in support for sparse data structures (SNode), automatic differentiation, AOT deployment, and compatibility with CUDA, Vulkan, Metal, and OpenGL ES, it empowers disciplines like simulation, graphics, AI, and robotics
    Downloads: 6 This Week
    Last Update:
    See Project
  • 24
    ImplicitGlobalGrid.jl

    ImplicitGlobalGrid.jl

    Distributed parallelization of stencil-based GPU and CPU applications

    ...It renders the distributed parallelization of stencil-based GPU and CPU applications on a regular staggered grid almost trivial and enables close to ideal weak scaling of real-world applications on thousands of GPUs [1, 2, 3]. ImplicitGlobalGrid relies on the Julia MPI wrapper (MPI.jl) to perform halo updates close to hardware limit and leverages CUDA-aware or ROCm-aware MPI for GPU-applications. The communication can straightforwardly be hidden behind computation [1, 3] (how this can be done automatically when using ParallelStencil.jl is shown in; a general approach particularly suited for CUDA C applications is explained in.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Triton

    Triton

    Development repository for the Triton language and compiler

    Triton is a programming language and compiler framework specifically designed for writing highly efficient custom deep learning operations, particularly for GPUs. It aims to bridge the gap between low-level GPU programming, such as CUDA, and higher-level abstractions by providing a more productive and flexible environment for developers. Triton enables users to write optimized kernels for machine learning workloads while maintaining readability and control over performance-critical aspects like memory access patterns and parallel execution. The project leverages LLVM and MLIR to compile code into efficient GPU instructions, supporting both NVIDIA and AMD hardware. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB