Showing 42 open source projects for "open_gapps-arm-7.1-nano-20211008"

View related business solutions
  • Apify is a full-stack web scraping and automation platform helping anyone get value from the web. Icon
    Apify is a full-stack web scraping and automation platform helping anyone get value from the web.

    Get web data. Build automations.

    Actors are serverless cloud programs that extract data, automate web tasks, and run AI agents. Developers build them using JavaScript, Python, or Crawlee, Apify's open-source library. Build once, publish to Store, and earn when others use it. Thousands of developers do this - Apify handles infrastructure, billing, and monthly payouts.
    Learn More
  • Next-generation security awareness training. Built for AI email phishing, vishing, smishing, and deepfakes. Icon
    Next-generation security awareness training. Built for AI email phishing, vishing, smishing, and deepfakes.

    Track your GenAI risk, run multichannel deepfake simulations, and engage employees with incredible security training.

    Assess how your company's digital footprint can be leveraged by cybercriminals. Identify the most at-risk individuals using thousands of public data points and take steps to proactively defend them.
    Learn More
  • 1
    nano-graphrag

    nano-graphrag

    A simple, easy-to-hack GraphRAG implementation

    nano-graphrag is a lightweight implementation of the GraphRAG approach designed to simplify experimentation with graph-based retrieval-augmented generation systems. GraphRAG expands traditional RAG pipelines by constructing knowledge graphs from documents and using relationships between entities to improve the quality and reasoning of AI responses.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 2
    Nano-vLLM

    Nano-vLLM

    A lightweight vLLM implementation built from scratch

    ...Despite its compact design, nano-vllm incorporates advanced optimization techniques such as prefix caching, tensor parallelism, and CUDA graph execution to achieve high performance during model inference. The engine is intended primarily for educational use, experimentation, and lightweight deployments where a full production-grade inference stack may be unnecessary.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Porcupine

    Porcupine

    On-device wake word detection powered by deep learning

    ...It enables building always-listening voice-enabled applications. It is using deep neural networks trained in real-world environments. Compact and computationally-efficient. It is perfect for IoT. Cross-platform. Arm Cortex-M, STM32, PSoC, Arduino, and i.MX RT. Raspberry Pi, NVIDIA Jetson Nano, and BeagleBone. Android and iOS. Chrome, Safari, Firefox, and Edge. Linux (x86_64), macOS (x86_64, arm64), and Windows (x86_64). Scalable. It can detect multiple always-listening voice commands with no added runtime footprint. Self-service. Developers can train custom wake word models using Picovoice Console. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 4
    PicoClaw

    PicoClaw

    Ultra-Efficient AI Assistant in Go

    ...Inspired by earlier AI assistant projects like “nanobot,” it was refactored to emphasize resource efficiency while still supporting meaningful AI-driven interactions such as conversational workflows, planning tasks, and automation. PicoClaw can run on hardware costing as little as $10 and on resource-constrained environments like RISC-V or ARM boards, with cross-architecture portability achieved through a single self-contained binary. The project’s goals include broad platform support (including Linux, macOS, and multiple CPU architectures), rapid startup times that make the assistant feel responsive, and integration with popular messaging platforms via gateways or bots.
    Downloads: 111 This Week
    Last Update:
    See Project
  • Unimus makes Network Automation and Configuration Management easy. Icon
    Unimus makes Network Automation and Configuration Management easy.

    Deploying Unimus to manage your entire network requires only minutes, allowing for rapid deployment without headaches.

    We aim to make automation, disaster recovery, change management and configuration auditing painless and affordable for a network of any size.
    Learn More
  • 5
    Banana Slides

    Banana Slides

    A native AI PPT generation application based on nano banana pro

    Banana Slides is an open-source application designed to automatically generate presentation slides using artificial intelligence. Built on top of the Nano Banana Pro framework, the software enables users to transform simple prompts or outlines into complete slide decks without manually formatting content. Instead of relying on traditional slide editing workflows, the system allows users to describe the desired presentation in natural language and have the AI generate structured slides, including titles, bullet points, and layout suggestions. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 6
    Compute Library

    Compute Library

    The Compute Library is a set of computer vision and machine learning

    The Compute Library is a set of computer vision and machine learning functions optimized for both Arm CPUs and GPUs using SIMD technologies. The library provides superior performance to other open-source alternatives and immediate support for new Arm® technologies e.g. SVE2.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 7
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 372 This Week
    Last Update:
    See Project
  • 8
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing large models to run within the constraints of mobile hardware. It supports a wide range of AI tasks including text generation, speech-to-text, vision processing, and retrieval-augmented workflows through a unified API interface. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 9
    Distributed Llama

    Distributed Llama

    Connect home devices into a powerful cluster to accelerate LLM

    ...By leveraging tensor parallelism and high-speed synchronization over Ethernet, it allows for faster performance as more devices are added to the cluster. The system supports various operating systems, including Linux, macOS, and Windows, and is optimized for both ARM and x86_64 AVX2 CPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • The top-rated AI recruiting platform for faster, smarter hiring. Icon
    The top-rated AI recruiting platform for faster, smarter hiring.

    Humanly is an AI recruiting platform that automates candidate conversations, screening, and scheduling.

    Humanly is an AI-first recruiting platform that helps talent teams hire in days, not months—without adding headcount. Our intuitive CRM pairs with powerful agentic AI to engage and screen every candidate instantly, surfacing top talent fast. Built on insights from over 4 million candidate interactions, Humanly delivers speed, structure, and consistency at scale—engaging 100% of interested candidates and driving pipeline growth through targeted outreach and smart re-engagement. We integrate seamlessly with all major ATSs to reduce manual work, improve data flow, and enhance recruiter efficiency and candidate experience. Independent audits ensure our AI remains fair and bias-free, so you can hire confidently.
    Learn More
  • 10
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    ...The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the following architectures: Arm* 64-bit Architecture (AArch64), NVIDIA* GPU, OpenPOWER* Power ISA (PPC64), IBMz* (s390x), and RISC-V. oneDNN is intended for deep learning applications and framework developers interested in improving application performance on Intel CPUs and GPUs. Deep learning practitioners should use one of the applications enabled with oneDNN.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 11
    MNN

    MNN

    MNN is a blazing fast, lightweight deep learning framework

    MNN is a highly efficient and lightweight deep learning framework. It supports inference and training of deep learning models, and has industry leading performance for inference and training on-device. At present, MNN has been integrated in more than 20 apps of Alibaba Inc, such as Taobao, Tmall, Youku, Dingtalk, Xianyu and etc., covering more than 70 usage scenarios such as live broadcast, short video capture, search recommendation, product searching by image, interactive marketing, equity...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 12
    Netron

    Netron

    Visualizer for neural network, deep learning, machine learning models

    ...Netron supports ONNX, Keras, TensorFlow Lite, Caffe, Darknet, Core ML, MNN, MXNet, ncnn, PaddlePaddle, Caffe2, Barracuda, Tengine, TNN, RKNN, MindSpore Lite, and UFF. Netron has experimental support for TensorFlow, PyTorch, TorchScript, OpenVINO, Torch, Arm NN, BigDL, Chainer, CNTK, Deeplearning4j, MediaPipe, ML.NET, scikit-learn, TensorFlow.js. There is an extense variety of sample model files to download or open using the browser version. It is supported by macOS, Windows, Linux, Python Server and browser.
    Downloads: 67 This Week
    Last Update:
    See Project
  • 13
    mllm

    mllm

    Fast Multimodal LLM on Mobile Devices

    ...Implemented primarily in C and C++, it is designed to operate with minimal external dependencies while taking advantage of hardware-specific acceleration technologies such as ARM NEON and x86 AVX2 instructions. The system supports multiple optimization techniques including quantization, pruning, and speculative decoding to improve performance while reducing computational overhead. It also provides tools to convert models from popular formats like PyTorch checkpoints into optimized runtime formats that can be executed on supported hardware platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    PHP Telegram Bot

    PHP Telegram Bot

    PHP Telegram Bot based on the official Telegram Bot API

    A Telegram Bot based on the official Telegram Bot API. Telegram announced official support for a Bot API, allowing integrators of all sorts to bring automated interactions to the mobile platform. This Bot aims to provide a platform where one can simply write a bot and have interactions in a matter of minutes. This is a pure PHP Telegram Bot, fully extensible via plugins.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    PhantomBot

    PhantomBot

    PhantomBot is an actively developed open source interactive Twitch bot

    PhantomBot is an actively developed open source interactive Twitch bot with a vibrant community that provides entertainment and moderation for your channel, allowing you to focus on what matters the most to you, your game and your viewers. PhantomBot is a Twitch chat bot powered by Java. PhantomBot has many modern features out of the box such as a built-in webpanel, enhanced moderation, games, a point system, raffles, custom commands, a music player, and more. PhantomBot can also be...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 16
    ncnn

    ncnn

    High-performance neural network inference framework for mobile

    ncnn is a high-performance neural network inference computing framework designed specifically for mobile platforms. It brings artificial intelligence right at your fingertips with no third-party dependencies, and speeds faster than all other known open source frameworks for mobile phone cpu. ncnn allows developers to easily deploy deep learning algorithm models to the mobile platform and create intelligent APPs. It is cross-platform and supports most commonly used CNN networks, including...
    Downloads: 29 This Week
    Last Update:
    See Project
  • 17
    NullClaw

    NullClaw

    Fastest, smallest, and fully autonomous AI assistant infrastructure

    NullClaw is the smallest fully autonomous AI assistant infrastructure, built entirely in Zig as a single static binary with zero runtime dependencies. At just 678 KB with ~1 MB peak RAM usage, it boots in under 2 milliseconds and runs on virtually any hardware, including low-cost ARM boards. Despite its size, it delivers a complete AI stack with 22+ model providers, 18+ communication channels, integrated tools, hybrid memory, and sandboxed runtime support. Its architecture is fully modular, using vtable interfaces that allow providers, channels, tools, memory backends, and runtimes to be swapped without code changes. ...
    Downloads: 30 This Week
    Last Update:
    See Project
  • 18
    Triton Inference Server

    Triton Inference Server

    The Triton Inference Server provides an optimized cloud

    ...Triton enables teams to deploy any AI model from multiple deep learning and machine learning frameworks, including TensorRT, TensorFlow, PyTorch, ONNX, OpenVINO, Python, RAPIDS FIL, and more. Triton supports inference across cloud, data center, edge, and embedded devices on NVIDIA GPUs, x86 and ARM CPU, or AWS Inferentia. Triton delivers optimized performance for many query types, including real-time, batched, ensembles, and audio/video streaming. Provides Backend API that allows adding custom backends and pre/post-processing operations. Model pipelines using Ensembling or Business Logic Scripting (BLS). HTTP/REST and GRPC inference protocols based on the community-developed KServe protocol. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 19
    Simd Library

    Simd Library

    C++ image processing and machine learning library with using of SIMD

    ...The algorithms are optimized with using of different SIMD CPU extensions. In particular, the library supports the following CPU extensions: SSE, AVX, AVX-512, and AMX for x86/x64, and NEON for ARM. The Simd Library has C API and also contains useful C++ classes and functions to facilitate access to C API. The library supports dynamic and static linking, 32-bit and 64-bit Windows and Linux, MSVS, G++ and Clang compilers, MSVS projects, and CMake build systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    MegEngine

    MegEngine

    Easy-to-use deep learning framework with 3 key features

    MegEngine is a fast, scalable and easy-to-use deep learning framework with 3 key features. You can represent quantization/dynamic shape/image pre-processing and even derivation in one model. After training, just put everything into your model and inference it on any platform at ease. Speed and precision problems won't bother you anymore due to the same core inside. In training, GPU memory usage could go down to one-third at the cost of only one additional line, which enables the DTR...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    Meta-World

    Meta-World

    Collections of robotics environments

    Meta-World is an open-source benchmark suite of robotic manipulation environments focused on multi-task and meta reinforcement learning. It provides a large collection of continuous-control tasks, such as reaching, pushing, opening doors, and manipulating objects with a simulated robot arm. The library defines standardized benchmarks like MT1, MT10, and MT50 for multi-task learning, where a single policy is trained across different numbers of tasks. It also offers meta-learning benchmarks (ML1, ML10, ML45) that evaluate few-shot adaptation to new goals or entirely new tasks. The environments adhere to the Gymnasium API, which makes them easy to plug into existing RL pipelines, and they support both synchronous and asynchronous vectorized execution for running many environments in parallel. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Dead Deer 3.14.82.2025

    Dead Deer 3.14.82.2025

    3D modeler, 3D game maker, 3D demo maker

    ...Intel x86/64, ARMv7/ARM64, RISCV. Linux (Ubuntu/wxWidgets(Gtk3)). iOS /iPasOS (with XCode) (GLES20/METAL) Windows Phone Windows VR (Steam/Oculus) WebAsm/WebGL UWP Windows/XBOX SDL2 Linux ARM 32/64 RISCV OpenXR (Quest?/Pico) 3.14.82.2025
    Leader badge
    Downloads: 20 This Week
    Last Update:
    See Project
  • 23
    GnoppixNG

    GnoppixNG

    Gnoppix Linux

    Gnoppix is a Linux distribution based on Debian Linux available in for amd64 and ARM architectures. Gnoppix is a great choice for users who want a lightweight and easy-to-use with security in mind. Gnoppix was first announced in June 2003. Currently we're working on a Gnoppix version for WSL, Mobile devices like smartphones and tablets as well.
    Leader badge
    Downloads: 70 This Week
    Last Update:
    See Project
  • 24
    Fractal Neuro Oscillator

    Fractal Neuro Oscillator

    A fractal neural network

    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    NanoDet-Plus

    NanoDet-Plus

    Lightweight anchor-free object detection model

    Super fast and high accuracy lightweight anchor-free object detection model. Real-time on mobile devices. NanoDet is a FCOS-style one-stage anchor-free object detection model which using Generalized Focal Loss as classification and regression loss. In NanoDet-Plus, we propose a novel label assignment strategy with a simple assign guidance module (AGM) and a dynamic soft label assigner (DSLA) to solve the optimal label assignment problem in lightweight model training. We also introduce a...
    Downloads: 9 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB