Showing 2 open source projects for "android arm"

View related business solutions
  • Zendesk: The Complete Customer Service Solution Icon
    Zendesk: The Complete Customer Service Solution

    Discover AI-powered, award-winning customer service software trusted by 200k customers

    Equip your agents with powerful AI tools and workflows that boost efficiency and elevate customer experiences across every channel.
    Learn More
  • Build a Custom Ad Server in Just Weeks Icon
    Build a Custom Ad Server in Just Weeks

    Adzerk's ad serving APIs allow developers to build and scale innovative, server-side ad platforms without reinventing the wheel.

    Adzerk's APIs make it easy for engineers and PMs to build their own server-side, fully-customized ad server. Top e-retailers and user communities use Adzerk to build innovative ad servers to promote anything from native ads to internal content to sponsored listings (where vendors and sellers pay for their organic listing to be promoted in search and browsing results). Engineers reliably see a 90%+ reduction in dev time using Adzerk’s APIs versus doing it entirely from scratch. Adzerk’s customer list includes Fortune 500 brands, public companies, and unicorn startups, including Bed Bath & Beyond, LiveNation/TicketMaster, Wattpad, TradingView, imgur, Strava, and many more. Our Ad.Product community makes it easy for product managers, engineers, ad ops, and others to discover and discuss how to build innovative, user-first ad platforms.
    Start Now
  • 1
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 365 This Week
    Last Update:
    See Project
  • 2
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB