Alternatives to Apache ServiceMix
Compare Apache ServiceMix alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Apache ServiceMix in 2026. Compare features, ratings, user reviews, pricing, and more from Apache ServiceMix competitors and alternatives in order to make an informed decision for your business.
-
1
Apache Geronimo
Apache
Apache Geronimo is an open-source set of projects that are focused on providing JavaEE/JakartaEE libraries and Microprofile implementations. We are actively delivering reusable Java EE components though. They are widely used and still actively maintained! Apache Geronimo provides libraries for the implementations of the Java EE and Jakarta EE specifications. The implementations are also focused on providing OSGi bundle metadata. The goal of XBean project is to create a plugin-based server analogous to Eclipse is a plugin-based IDE. XBean will be able to discover, download and install server plugins from an Internet-based repository. In addition, we include support for multiple IoC systems, support for running with no IoC system, JMX without JMX code, lifecycle and class loader management, and rock-solid Spring integration. Apache Geronimo hosts several Microprofile implementations. Apache Geronimo Arthur is an effort to build a thin layer on top of Oracle GraalVM.Starting Price: Free -
2
Apache Camel
Apache Software Foundation
Camel is an Open Source integration framework that empowers you to quickly and easily integrate various systems consuming or producing data. Camel supports most of the Enterprise Integration Patterns from the excellent book by Gregor Hohpe and Bobby Woolf, and newer integration patterns from microservice architectures to help you solve your integration problem by applying best practices out of the box. Apache Camel is standalone, and can be embedded as a library within Spring Boot, Quarkus, Application Servers, and in the clouds. Camel subprojects focus on making your work easy. Packed with several hundred components that are used to access databases, message queues, APIs or basically anything under the sun. Helping you integrate with everything. Camel supports around 50 data formats, allowing to translate messages in multiple formats, and with support from industry standard formats from finance, telco, health-care, and more. -
3
BMC middleware management software provides real-time monitoring and administration for messaging-oriented middleware environments including IBM® MQ, Integration Bus (IIB), App Connect Enterprise (ACE), Apache ActiveMQ, and DataPower and TIBCO Enterprise Message Service (EMS). Automate alerts and gain insight into a wide array of middleware technologies with a single, intuitive solution. MainView Middleware Monitor securely provides real-time monitoring and automatic notification of potential problems to ensure an optimally performing middleware layer. Analyze historical data to identify patterns, predict trends, and solve reoccurring problems. Maximize application availability and mitigate risk with proactive problem detection and automated resolutions. Improve productivity and efficiency for managing, administering, and troubleshooting with flexible dashboards for infrastructure and application views.
-
4
Red Hat AMQ
Red Hat
Red Hat AMQ is a flexible messaging platform that delivers information reliably, enabling real-time integration and connecting the Internet of Things (IoT). Based on open source communities like Apache ActiveMQ and Apache Kafka, it supports various messaging patterns to integrate applications, endpoints, and devices quickly and efficiently, enhancing enterprise responsiveness and agility. AMQ facilitates data sharing between microservices and other applications with high throughput and low latency. AMQ supports connectivity from client programs written in multiple languages. It defines an open-wire protocol for messaging interoperability, allowing enterprises to deploy various distributed messaging solutions to meet evolving business requirements. Backed by Red Hat's award-winning support and services, AMQ has a track record of supporting mission-critical applications. -
5
Apache Synapse
Apache Software Foundation
Apache Synapse is a lightweight and high-performance Enterprise Service Bus (ESB). Powered by a fast and asynchronous mediation engine, Apache Synapse provides exceptional support for XML, Web Services and REST. In addition to XML and SOAP, Apache Synapse supports several other content interchange formats, such as plain text, binary, Hessian and JSON. The wide range of transport adapters available for Synapse, enables it to communicate over many application and transport layer protocols. As of now, Apache Synapse supports HTTP/S, Mail (POP3, IMAP, SMTP), JMS, TCP, UDP, VFS, SMS, XMPP and FIX. High-performing PassThrough HTTP transport support for all mediation scenarios. Ultra-fast, low latency mediation of HTTP requests. Supporting a very large number of inbound (client -> ESB) and outbound (ESB -> server) connections concurrently. Intelligently handle message content and content awareness built into the engine with shared buffer for handling data. -
6
Eclipse Jetty
Eclipse Foundation
Jetty provides a web server and servlet container, additionally providing support for HTTP/2, WebSocket, OSGi, JMX, JNDI, JAAS and many other integrations. These components are open source and are freely available for commercial use and distribution. Jetty is used in a wide variety of projects and products, both in development and production. Jetty has long been loved by developers due to its long history of being easily embedded in devices, tools, frameworks, application servers, and modern cloud services. Full-featured and standards-based. Open source and commercially usable, flexible and extensible, small footprint, embeddable, asynchronous, enterprise scalable, and dual-licensed under Apache and Eclipse. Large clusters, such as Facebook Presto. Cloud computing, such as Google AppEngine. With the direction of Java and the JakartaEE project (formerly JavaEE) in 2020, the current recommended version of Jetty for use depends upon the servlet API version, desired licensing, etc. -
7
ActiveMQ
Apache Software Foundation
Apache ActiveMQ® is the most popular open source, multi-protocol, Java-based message broker. It supports industry standard protocols so users get the benefits of client choices across a broad range of languages and platforms. Connect from clients written in JavaScript, C, C++, Python, .Net, and more. Integrate your multi-platform applications using the ubiquitous AMQP protocol. Exchange messages between your web applications using STOMP over websockets. Manage your IoT devices using MQTT. Support your existing JMS infrastructure and beyond. ActiveMQ offers the power and flexibility to support any messaging use-case. There are currently two "flavors" of ActiveMQ available - the well-known "classic" broker and the "next generation" broker code-named Artemis. Once Artemis reaches a sufficient level of feature parity with the "Classic" code-base it will become the next major version of ActiveMQ. Initial migration documentation is available as well as a development roadmap for Artemis. -
8
Amazon MQ
Amazon
Amazon MQ is a managed message broker service for Apache ActiveMQ that makes it easy to set up and operate message brokers in the cloud. Message brokers allow different software systems–often using different programming languages, and on different platforms–to communicate and exchange information. Amazon MQ reduces your operational load by managing the provisioning, setup, and maintenance of ActiveMQ, a popular open-source message broker. Connecting your current applications to Amazon MQ is easy because it uses industry-standard APIs and protocols for messaging, including JMS, NMS, AMQP, STOMP, MQTT, and WebSocket. Using standards means that in most cases, there’s no need to rewrite any messaging code when you migrate to AWS. With a few clicks in the Amazon MQ Console, Amazon MQ provisions your broker with support for version upgrades, so you can always use the latest version that Amazon MQ supports. Once you configure your broker, your applications can produce and consume messages. -
9
Apache Lucene
Apache Software Foundation
The Apache Lucene™ project develops open-source search software. The project releases a core search library, named Lucene™ core, as well as PyLucene, a python binding for Lucene. Lucene Core is a Java library providing powerful indexing and search features, as well as spellchecking, hit highlighting and advanced analysis/tokenization capabilities. The PyLucene sub project provides Python bindings for Lucene Core. The Apache Software Foundation provides support for the Apache community of open-source software projects. Apache Lucene is distributed under a commercially friendly Apache Software license. Apache Lucene set the standard for search and indexing performance. Lucene is the search core of both Apache Solr™ and Elasticsearch™. Our core algorithms along with the Solr search server power applications the world over, ranging from mobile devices to sites like Twitter, Apple and Wikipedia. The goal of Apache Lucene is to provide world class search capabilities. -
10
Amazon EMR
Amazon
Amazon EMR is the industry-leading cloud big data platform for processing vast amounts of data using open-source tools such as Apache Spark, Apache Hive, Apache HBase, Apache Flink, Apache Hudi, and Presto. With EMR you can run Petabyte-scale analysis at less than half of the cost of traditional on-premises solutions and over 3x faster than standard Apache Spark. For short-running jobs, you can spin up and spin down clusters and pay per second for the instances used. For long-running workloads, you can create highly available clusters that automatically scale to meet demand. If you have existing on-premises deployments of open-source tools such as Apache Spark and Apache Hive, you can also run EMR clusters on AWS Outposts. Analyze data using open-source ML frameworks such as Apache Spark MLlib, TensorFlow, and Apache MXNet. Connect to Amazon SageMaker Studio for large-scale model training, analysis, and reporting. -
11
Amazon MSK
Amazon
Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.Starting Price: $0.0543 per hour -
12
Apache APISIX
Apache APISIX
Apache APISIX provides rich traffic management features like Load Balancing, Dynamic Upstream, Canary Release, Circuit Breaking, Authentication, Observability, etc. Apache APISIX provides open source API Gateway to help you manage microservices, delivering the ultimate performance, security, and scalable platform for all your APIs and microservices. Apache APISIX is the first open-source API Gateway that includes a built-in low-code Dashboard, which offers a powerful and flexible UI for developers to use. The Apache APISIX Dashboard is designed to make it as easy as possible for users to operate Apache APISIX through a frontend interface. It’s open-source and ever evolving, feel free to contribute. The Apache APISIX dashboard is flexible to User demand, providing option to create custom modules through code matching your requirements, alongside the existing no-code toolchain. -
13
E-MapReduce
Alibaba
EMR is an all-in-one enterprise-ready big data platform that provides cluster, job, and data management services based on open-source ecosystems, such as Hadoop, Spark, Kafka, Flink, and Storm. Alibaba Cloud Elastic MapReduce (EMR) is a big data processing solution that runs on the Alibaba Cloud platform. EMR is built on Alibaba Cloud ECS instances and is based on open-source Apache Hadoop and Apache Spark. EMR allows you to use the Hadoop and Spark ecosystem components, such as Apache Hive, Apache Kafka, Flink, Druid, and TensorFlow, to analyze and process data. You can use EMR to process data stored on different Alibaba Cloud data storage service, such as Object Storage Service (OSS), Log Service (SLS), and Relational Database Service (RDS). You can quickly create clusters without the need to configure hardware and software. All maintenance operations are completed on its Web interface. -
14
Apache TomEE
Apache
Apache TomEE, pronounced “Tommy”, is an all-Apache Jakarta EE 9.1 certified application server that extends Apache Tomcat that is assembled from a vanilla Apache Tomcat zip file. We start with Apache Tomcat, add our jars, and zip up the rest. The result is Tomcat plus EE features, TomEE. Stable and ready for production, Apache TomEE 8.0 implements Java EE 8/Jakarta EE 8 and supports the javax namespace. Runs on Java 8 or higher. Mostly Jakarta EE 9.1 web profile compliant and supports the new jakarta namespace. Runs on Java 11 or higher. Apache TomEE comes in four different flavors, web profile, MicroProfile, Plus and Plume. Apache TomEE web profile delivers servlets, JSP, JSF, JTA, JPA, CDI, bean validation and EJB Lite. Apache TomEE MicroProfile adds support for MicroProfile. Apache TomEE Plus and Plume add support for JMS, JAX-WS, and more. Mostly Jakarta EE 9.1 Web Profile compliant and supports the new jakarta namespace.Starting Price: Free -
15
Conduktor
Conduktor
We created Conduktor, the all-in-one friendly interface to work with the Apache Kafka ecosystem. Develop and manage Apache Kafka with confidence. With Conduktor DevTools, the all-in-one Apache Kafka desktop client. Develop and manage Apache Kafka with confidence, and save time for your entire team. Apache Kafka is hard to learn and to use. Made by Kafka lovers, Conduktor best-in-class user experience is loved by developers. Conduktor offers more than just an interface over Apache Kafka. It provides you and your teams the control of your whole data pipeline, thanks to our integration with most technologies around Apache Kafka. Provide you and your teams the most complete tool on top of Apache Kafka. -
16
AtlasMap
AtlasMap
AtlasMap is a data mapping solution with interactive web based user interface, that simplifies configuring integrations between Java, XML, CSV and JSON data sources. You can design your data mapping with AtlasMap Data Mapper UI canvas, and then run that data mapping via runtime engine. In addition to plain Java API provided by runtime engine, camel-atlasmap Component is also available to perform data mapping as a part of Apache Camel route. There is also a Camel Quarkus extension available. The easiest way to use AtlasMap Data Mapper UI is the standalone mode. Or you can use it through a VS Code plugin. Historically, the AtlasMap Data Mapper UI was designed to work within Syndesis UI and it’s still a best way to experience full benefits of integrated typed data mapping with UI. You can install and run Syndesis by following the Syndesis Developer Handbook. You will find the AtlasMap Data Mapper UI under the integrations panel after selecting or adding an integration with a Data Mapper. -
17
Apache Storm
Apache Software Foundation
Apache Storm is a free and open source distributed realtime computation system. Apache Storm makes it easy to reliably process unbounded streams of data, doing for realtime processing what Hadoop did for batch processing. Apache Storm is simple, can be used with any programming language, and is a lot of fun to use! Apache Storm has many use cases: realtime analytics, online machine learning, continuous computation, distributed RPC, ETL, and more. Apache Storm is fast: a benchmark clocked it at over a million tuples processed per second per node. It is scalable, fault-tolerant, guarantees your data will be processed, and is easy to set up and operate. Apache Storm integrates with the queueing and database technologies you already use. An Apache Storm topology consumes streams of data and processes those streams in arbitrarily complex ways, repartitioning the streams between each stage of the computation however needed. Read more in the tutorial. -
18
IBM Analytics for Apache Spark is a flexible and integrated Spark service that empowers data science professionals to ask bigger, tougher questions, and deliver business value faster. It’s an easy-to-use, always-on managed service with no long-term commitment or risk, so you can begin exploring right away. Access the power of Apache Spark with no lock-in, backed by IBM’s open-source commitment and decades of enterprise experience. A managed Spark service with Notebooks as a connector means coding and analytics are easier and faster, so you can spend more of your time on delivery and innovation. A managed Apache Spark services gives you easy access to the power of built-in machine learning libraries without the headaches, time and risk associated with managing a Sparkcluster independently.
-
19
Amazon MWAA
Amazon
Amazon Managed Workflows for Apache Airflow (MWAA) is a managed orchestration service for Apache Airflow that makes it easier to set up and operate end-to-end data pipelines in the cloud at scale. Apache Airflow is an open-source tool used to programmatically author, schedule, and monitor sequences of processes and tasks referred to as “workflows.” With Managed Workflows, you can use Airflow and Python to create workflows without having to manage the underlying infrastructure for scalability, availability, and security. Managed Workflows automatically scales its workflow execution capacity to meet your needs, and is integrated with AWS security services to help provide you with fast and secure access to data.Starting Price: $0.49 per hour -
20
Astra Streaming
DataStax
Responsive applications keep users engaged and developers inspired. Rise to meet these ever-increasing expectations with the DataStax Astra Streaming service platform. DataStax Astra Streaming is a cloud-native messaging and event streaming platform powered by Apache Pulsar. Astra Streaming allows you to build streaming applications on top of an elastically scalable, multi-cloud messaging and event streaming platform. Astra Streaming is powered by Apache Pulsar, the next-generation event streaming platform which provides a unified solution for streaming, queuing, pub/sub, and stream processing. Astra Streaming is a natural complement to Astra DB. Using Astra Streaming, existing Astra DB users can easily build real-time data pipelines into and out of their Astra DB instances. With Astra Streaming, avoid vendor lock-in and deploy on any of the major public clouds (AWS, GCP, Azure) compatible with open-source Apache Pulsar. -
21
Apache Tomcat
Apache
The Apache Tomcat® software is an open source implementation of the Jakarta Servlet, Jakarta Server Pages, Jakarta Expression Language, Jakarta WebSocket, Jakarta Annotations and Jakarta Authentication specifications. These specifications are part of the Jakarta EE platform. Apache Tomcat software powers numerous large-scale, mission-critical web applications across a diverse range of industries and organizations. Some of these users and their stories are listed on the PoweredBy wiki page. The Apache Tomcat Project is proud to announce the release of version 10.0.10 of Apache Tomcat. This release implements specifications that are part of the Jakarta EE 9 platform. -
22
Thousands of customers use Amazon Managed Service for Apache Flink to run stream processing applications. With Amazon Managed Service for Apache Flink, you can transform and analyze streaming data in real-time using Apache Flink and integrate applications with other AWS services. There are no servers and clusters to manage, and there is no computing and storage infrastructure to set up. You pay only for the resources you use. Build and run Apache Flink applications, without setting up infrastructure and managing resources and clusters. Process gigabytes of data per second with subsecond latencies and respond to events in real-time. Deploy highly available and durable applications with Multi-AZ deployments and APIs for application lifecycle management. Develop applications that transform and deliver data to Amazon Simple Storage Service (Amazon S3), Amazon OpenSearch Service, and more.Starting Price: $0.11 per hour
-
23
Apache Mahout
Apache Software Foundation
Apache Mahout is a powerful, scalable, and versatile machine learning library designed for distributed data processing. It offers a comprehensive set of algorithms for various tasks, including classification, clustering, recommendation, and pattern mining. Built on top of the Apache Hadoop ecosystem, Mahout leverages MapReduce and Spark to enable data processing on large-scale datasets. Apache Mahout(TM) is a distributed linear algebra framework and mathematically expressive Scala DSL designed to let mathematicians, statisticians, and data scientists quickly implement their own algorithms. Apache Spark is the recommended out-of-the-box distributed back-end or can be extended to other distributed backends. Matrix computations are a fundamental part of many scientific and engineering applications, including machine learning, computer vision, and data analysis. Apache Mahout is designed to handle large-scale data processing by leveraging the power of Hadoop and Spark. -
24
PDFBox
Apache Software Foundation
The Apache PDFBox® library is an open-source Java tool for working with PDF documents. This project allows the creation of new PDF documents, manipulation of existing documents and the ability to extract content from documents. Apache PDFBox also includes several command-line utilities. Apache PDFBox is published under the Apache License v2.0. Extract Unicode text from PDF files. Split a single PDF into many files or merge multiple PDF files. Extract data from PDF forms or fill a PDF form. Validate PDF files against the PDF/A-1b standard. Print a PDF file using the standard Java printing API. Create a PDF from scratch, with embedded fonts and images. Save PDFs as image files, such as PNG or JPEG and digitally sign PDF files. See also the export control information related to the encryption features included in Apache PDFBox. -
25
Apache James
The Apache Software Foundation
James stands for Java Apache Mail Enterprise Server. It has a modular architecture based on a rich set of modern and efficient components which provides at the end complete, stable, secure, and extendable Mail Servers running on the JVM. Create your own personal solution of email treatment by assembling the components you need thanks to the Inversion of Control mail platform offered and go further customizing filtering and routing rules using James Mailet Container. The Apache James project wires together the different libraries composing James to provide running services, ready to download on the Apache mirrors.Starting Price: Free -
26
Apache Sentry
Apache Software Foundation
Apache Sentry™ is a system for enforcing fine grained role based authorization to data and metadata stored on a Hadoop cluster. Apache Sentry has successfully graduated from the Incubator in March of 2016 and is now a Top-Level Apache project. Apache Sentry is a granular, role-based authorization module for Hadoop. Sentry provides the ability to control and enforce precise levels of privileges on data for authenticated users and applications on a Hadoop cluster. Sentry currently works out of the box with Apache Hive, Hive Metastore/HCatalog, Apache Solr, Impala and HDFS (limited to Hive table data). Sentry is designed to be a pluggable authorization engine for Hadoop components. It allows you to define authorization rules to validate a user or application’s access requests for Hadoop resources. Sentry is highly modular and can support authorization for a wide variety of data models in Hadoop. -
27
iDempiere
iDempiere
iDempiere. Community Powered Enterprise, also known as OSGi + ADempiere, Business Suite ERP/CRM/SCM done the community way. Focus is on the Community that includes Subject Matter Specialists, Implementors and End-Users. iDempiere is based on original Compiere/Adempiere plus a new architecture to use state-of-the-art technologies like OSGi, Buckminster, zk, Jetty -
28
Apache Gump
Apache Software Foundation
The Apache Gump continuous integration tool was the first one developed at the Apache Software Foundation. It is written in Python and fully supports Apache Ant, Apache Maven (1.x to 3.x) and other build tools. Gump is unique in that it builds and compiles software against the latest development versions of those projects. This allows Gump to detect potentially incompatible changes to that software just a few hours after those changes are checked into the version control system. Notifications are sent to the project team as soon as such a change is detected, referencing more detailed reports available online. You can set up and run Gump on your own machine and run it on your own projects, however it is currently most famous for building many of Apache's projects and their dependencies. For this purpose, the Gump project maintains its own dedicated server. -
29
OpenMeetings
Apache Software Foundation
Openmeetings provides video conferencing, instant messaging, white board, collaborative document editing and other groupware tools. It uses API functions of Media Server for Remoting and Streaming Kurento. OpenMeetings is a project of The Apache Software Foundation, the old project website at GoogleCode will receive no updates anymore. The website at Apache is the only place that receives updates. -
30
Apache Pulsar
Apache Software Foundation
Apache Pulsar is a cloud-native, distributed messaging and streaming platform originally created at Yahoo! and now a top-level Apache Software Foundation project. Easy to deploy, lightweight compute process, developer-friendly APIs, no need to run your own stream processing engine. Run in production at Yahoo! scale for over 5 years, with millions of messages per second across millions of topics. Built from the ground up as a multi-tenant system. Supports isolation, authentication, authorization and quotas. Configurable replication between data centers across multiple geographic regions. Persistent message storage based on Apache BookKeeper. IO-level isolation between write and read operations. Rest admin API for provisioning, administration, tools and monitoring. -
31
Apache Axiom
The Apache Software Foundation
The Apache Axiom™ library provides an XML Infoset compliant object model implementation which supports on-demand building of the object tree. It supports a novel "pull-through" model which allows one to turn off the tree building and directly access the underlying pull event stream using the StAX API. It also has built in support for XML Optimized Packaging (XOP) and MTOM, the combination of which allows XML to carry binary data efficiently and in a transparent manner. The combination of these is an easy to use API with a very high performant architecture! Developed as part of Apache Axis2, Apache Axiom is the core of Apache Axis2. However, it is a pure standalone XML Infoset model with novel features and can be used independently of Apache Axis2. -
32
Apache Spark
Apache Software Foundation
Apache Spark™ is a unified analytics engine for large-scale data processing. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. Spark offers over 80 high-level operators that make it easy to build parallel apps. And you can use it interactively from the Scala, Python, R, and SQL shells. Spark powers a stack of libraries including SQL and DataFrames, MLlib for machine learning, GraphX, and Spark Streaming. You can combine these libraries seamlessly in the same application. Spark runs on Hadoop, Apache Mesos, Kubernetes, standalone, or in the cloud. It can access diverse data sources. You can run Spark using its standalone cluster mode, on EC2, on Hadoop YARN, on Mesos, or on Kubernetes. Access data in HDFS, Alluxio, Apache Cassandra, Apache HBase, Apache Hive, and hundreds of other data sources. -
33
Apache Subversion
Apache Software Foundation
Welcome to subversion, the online home of the Apache® Subversion® software project. Subversion is an open-source version control system. Founded in 2000 by CollabNet, Inc., the Subversion project and software have seen incredible success over the past decade. Subversion has enjoyed and continues to enjoy widespread adoption in both the open-source arena and the corporate world. Subversion is developed as a project of the Apache Software Foundation, and as such is part of a rich community of developers and users. We're always in need of individuals with a wide range of skills, and we invite you to participate in the development of Apache Subversion. Subversion exists to be universally recognized and adopted as an open-source, centralized version control system characterized by its reliability as a safe haven for valuable data; the simplicity of its model and usage; and its ability to support the needs of a wide variety of users and projects. -
34
DataStax
DataStax
The Open, Multi-Cloud Stack for Modern Data Apps. Built on open-source Apache Cassandra™. Global-scale and 100% uptime without vendor lock-in. Deploy on multi-cloud, on-prem, open-source, and Kubernetes. Elastic and pay-as-you-go for improved TCO. Start building faster with Stargate APIs for NoSQL, real-time, reactive, JSON, REST, and GraphQL. Skip the complexity of multiple OSS projects and APIs that don’t scale. Ideal for commerce, mobile, AI/ML, IoT, microservices, social, gaming, and richly interactive applications that must scale-up and scale-down with demand. Get building modern data applications with Astra, a database-as-a-service powered by Apache Cassandra™. Use REST, GraphQL, JSON with your favorite full-stack framework Richly interactive apps that are elastic and viral-ready from Day 1. Pay-as-you-go Apache Cassandra DBaaS that scales effortlessly and affordably. -
35
ZSegment
314e Corporation
ZSegment is a cloud-native interface engine built by 314e Corporation, designed to seamlessly connect and integrate clinical, financial, and administrative systems across healthcare organizations. It offers more than 300 components, supports real-time metrics based on open telemetry standards, and is constructed on a Kubernetes-native architecture using Apache Camel to enable scalability, flexibility and developer-friendly customization including version control (Git), native custom schema support for HL7 v2, FHIR, X12 and CCD, visual tools for message routing, Groovy scripting for transformations, and secure message delivery over multiple protocols (File, FTP, HTTP(S), TCP). It also features message indexing, tracing, editing/resubmission, and custom job definitions for data hygiene and operational monitoring. ZSegment is positioned as a modern alternative to legacy on-premise interface engines like Mirth Connect, alleviating hardware procurement costs, expensive licensing, etc. -
36
Stack Console
Stack Console Cloud Solutions
Stack Console is a cloud management platform built on top of open source cloud computing platform i.e. Apache Cloudstack. It helps datacenters, large web hosting companies & telco's to launch their own public cloud business. StackConsole is a fully-integrated, transparent and single SLA-driven cloud management portal powered by Apache CloudStack™ software. It offers enterprise cloud solutions for datacenter, large web hosting companies, telecom operators and cloud service providers with first-class support for all technical, management and service operations.Starting Price: $599 per month -
37
Ververica
Ververica
Ververica Platform enables every enterprise to take advantage and derive immediate insight from its data in real-time. Powered by Apache Flink's robust streaming runtime, Ververica Platform makes this possible by providing an integrated solution for stateful stream processing and streaming analytics at scale. Powered by Apache Flink, Ververica Platform provides high throughput, low latency data processing, powerful abstractions and the operational flexibility trusted by some of the world’s largest and most successful data-driven enterprises such as Alibaba, Netflix and Uber. Ververica Platform brings the accumulated knowledge of our experience working with some of these large and innovative, data-driven companies into an easily-accessible, cost-effective and secure enterprise-ready platform. -
38
Apache Derby
Apache
Apache Derby, an Apache DB subproject, is an open source relational database implemented entirely in Java and available under the Apache License, Version 2.0. Derby has a small footprint - about 3.5 megabytes for the base engine and embedded JDBC driver. Derby provides an embedded JDBC driver that lets you embed Derby in any Java-based solution. Derby also supports the more familiar client/server mode with the Derby Network Client JDBC driver and Derby Network Server. -
39
Apache PredictionIO
Apache
Apache PredictionIO® is an open-source machine learning server built on top of a state-of-the-art open-source stack for developers and data scientists to create predictive engines for any machine learning task. It lets you quickly build and deploy an engine as a web service on production with customizable templates. Respond to dynamic queries in real-time once deployed as a web service, evaluate and tune multiple engine variants systematically, and unify data from multiple platforms in batch or in real-time for comprehensive predictive analytics. Speed up machine learning modeling with systematic processes and pre-built evaluation measures, support machine learning and data processing libraries such as Spark MLLib and OpenNLP. Implement your own machine learning models and seamlessly incorporate them into your engine. Simplify data infrastructure management. Apache PredictionIO® can be installed as a full machine learning stack, bundled with Apache Spark, MLlib, HBase, Akka HTTP, etc.Starting Price: Free -
40
IOMETE
IOMETE
IOMETE is a self-hosted data lakehouse platform built on Apache Iceberg, Apache Spark, and Kubernetes. Run it on-premises or in your private cloud — your infrastructure, your data, your control. Built for enterprises in regulated industries, IOMETE eliminates third-party ICT risk at the data layer by architecture — not by contract. No SaaS dependencies. No data leaving your perimeter. Compliance with GDPR, DORA, and NIS2 is structural, not contractual. Included in one platform: - Data Lakehouse(s) - Data Catalog - SQL Editor - Apache Spark Jobs - ML Notebooks - Orchestration Engine - Spark Connect Key capabilities: Apache Iceberg-native storage, Kubernetes-native deployment (K8s + OpenShift), row/column/tag-based access control, Data Mesh support, air-gapped and zero-trust compatible. Transparent pricing — CPU-based, no per-query fees, no billing surprises.Starting Price: Free -
41
ApacheBooster
NdimensionZ
ApacheBooster has been specifically designed to enhance the working of web servers based on cPanel. ApacheBooster as the name suggests boosts the working ability of the Apache web server, which is according to the census the most used web server in the world! Nginx and varnish has been fused together in ApacheBooster to make it effectively efficient in its working. Nginx is a super quality high performing web server software that speeds up the working of the web server. The best feature of Nginx is that it is very fast in its working i.e in retrieving static files and also helps in saving memory by using less memory for processing of concurrent requests. It is very efficient in handling traffic requests. With the less amount of memory used, it is capable of handling more requests/clients when compared to Apache. Nginx is a reverse proxy server of open source type that smartly balances the load, a web server and web cache (also known as HTTP cache). -
42
Apache Hive
Apache Software Foundation
The Apache Hive data warehouse software facilitates reading, writing, and managing large datasets residing in distributed storage using SQL. Structure can be projected onto data already in storage. A command line tool and JDBC driver are provided to connect users to Hive. Apache Hive is an open source project run by volunteers at the Apache Software Foundation. Previously it was a subproject of Apache® Hadoop®, but has now graduated to become a top-level project of its own. We encourage you to learn about the project and contribute your expertise. Traditional SQL queries must be implemented in the MapReduce Java API to execute SQL applications and queries over distributed data. Hive provides the necessary SQL abstraction to integrate SQL-like queries (HiveQL) into the underlying Java without the need to implement queries in the low-level Java API. -
43
Airy Messenger
Airy
From Conversational AI assistants to customer service experiences, own your own conversational platform with open-source Airy. Airy Core is an open source, fully-featured, production-ready conversational platform. With Airy you can process conversational data from a variety of sources. Since Airy's infrastructure is built around Apache Kafka, it can process a large amount of conversations and messages simultaneously and stream the relevant conversational data to wherever you need it. Connect anything from our free open-source live chat plugin to Facebook Messenger & Google's Business Messages to your Airy Core. This is all possible through an ingestion platform that heavily relies on Apache Kafka to process incoming webhook data from different sources. We make sense of the data and reshape it into source-independent contacts, conversations, and messages. -
44
Apache Giraph
Apache Software Foundation
Apache Giraph is an iterative graph processing system built for high scalability. For example, it is currently used at Facebook to analyze the social graph formed by users and their connections. Giraph originated as the open-source counterpart to Pregel, the graph processing architecture developed at Google and described in a 2010 paper. Both systems are inspired by the Bulk Synchronous Parallel model of distributed computation introduced by Leslie Valiant. Giraph adds several features beyond the basic Pregel model, including master computation, sharded aggregators, edge-oriented input, out-of-core computation, and more. With a steady development cycle and a growing community of users worldwide, Giraph is a natural choice for unleashing the potential of structured datasets at a massive scale. Apache Giraph is an iterative graph processing framework, built on top of Apache Hadoop. -
45
Red Hat OpenShift Streams
Red Hat
Red Hat® OpenShift® Streams for Apache Kafka is a managed cloud service that provides a streamlined developer experience for building, deploying, and scaling new cloud-native applications or modernizing existing systems. Red Hat OpenShift Streams for Apache Kafka makes it easy to create, discover, and connect to real-time data streams no matter where they are deployed. Streams are a key component for delivering event-driven and data analytics applications. The combination of seamless operations across distributed microservices, large data transfer volumes, and managed operations allows teams to focus on team strengths, speed up time to value, and lower operational costs. OpenShift Streams for Apache Kafka includes a Kafka ecosystem and is part of a family of cloud services—and the Red Hat OpenShift product family—which helps you build a wide range of data-driven solutions. -
46
Apache Ivy
Apache Software Foundation
Apache Ivy™ is a popular dependency manager focusing on flexibility and simplicity. Find out more about its unique enterprise features, what people say about it, and how it can improve your build system! Ivy is a tool for managing (recording, tracking, resolving, and reporting) project dependencies. Ivy is essentially process agnostic and is not tied to any methodology or structure. Instead, it provides the necessary flexibility and reconfigurability to be adapted to a broad range of dependency management and build processes. While available as a standalone tool, Ivy works particularly well with Apache Ant providing a number of powerful Ant tasks ranging from dependency resolution to dependency reporting and publication. Ivy has a lot of powerful features, the most popular and useful being its flexibility, integration with Ant, and strong transitive dependencies management engine. Ivy is open source and released under a very permissive Apache License.Starting Price: Free -
47
Apache Phoenix
Apache Software Foundation
Apache Phoenix enables OLTP and operational analytics in Hadoop for low-latency applications by combining the best of both worlds. The power of standard SQL and JDBC APIs with full ACID transaction capabilities and the flexibility of late-bound, schema-on-read capabilities from the NoSQL world by leveraging HBase as its backing store. Apache Phoenix is fully integrated with other Hadoop products such as Spark, Hive, Pig, Flume, and Map Reduce. Become the trusted data platform for OLTP and operational analytics for Hadoop through well-defined, industry-standard APIs. Apache Phoenix takes your SQL query, compiles it into a series of HBase scans, and orchestrates the running of those scans to produce regular JDBC result sets. Direct use of the HBase API, along with coprocessors and custom filters, results in performance on the order of milliseconds for small queries, or seconds for tens of millions of rows.Starting Price: Free -
48
ODFToEPub
Pincette
With ODFToEPub everyone can write an e-book with full control over how it will look. All you need is a word processor that can produce a document in a format suitable for Apache OpenOffice or LibreOffice. That is, of course, Apache OpenOffice and LibreOffice, but also Microsoft Word, iWork, WordPerfect, Zoho, Google Docs, etc. In Apache OpenOffice and LibreOffice you choose the export function to convert an ODT-file to ePub. With this tool self publishers have immediate feedback about what their e-book will look like. Publishers can give their authors a standard template and build the tool into their systems in order to streamline their ePub production process. Companies can cut down printing by publishing their internal documents as e-books. ODFToEPub is an Apache OpenOffice and LibreOffice extension as well as a stand-alone program. After receiving the license.xml file by e-mail, you should save it on your computer and install it.Starting Price: $52.00/one-time/user -
49
Deeplearning4j
Deeplearning4j
DL4J takes advantage of the latest distributed computing frameworks including Apache Spark and Hadoop to accelerate training. On multi-GPUs, it is equal to Caffe in performance. The libraries are completely open-source, Apache 2.0, and maintained by the developer community and Konduit team. Deeplearning4j is written in Java and is compatible with any JVM language, such as Scala, Clojure, or Kotlin. The underlying computations are written in C, C++, and Cuda. Keras will serve as the Python API. Eclipse Deeplearning4j is the first commercial-grade, open-source, distributed deep-learning library written for Java and Scala. Integrated with Hadoop and Apache Spark, DL4J brings AI to business environments for use on distributed GPUs and CPUs. There are a lot of parameters to adjust when you're training a deep-learning network. We've done our best to explain them, so that Deeplearning4j can serve as a DIY tool for Java, Scala, Clojure, and Kotlin programmers. -
50
MXNet
The Apache Software Foundation
A hybrid front-end seamlessly transitions between Gluon eager imperative mode and symbolic mode to provide both flexibility and speed. Scalable distributed training and performance optimization in research and production is enabled by the dual parameter server and Horovod support. Deep integration into Python and support for Scala, Julia, Clojure, Java, C++, R and Perl. A thriving ecosystem of tools and libraries extends MXNet and enables use-cases in computer vision, NLP, time series and more. Apache MXNet is an effort undergoing incubation at The Apache Software Foundation (ASF), sponsored by the Apache Incubator. Incubation is required of all newly accepted projects until a further review indicates that the infrastructure, communications, and decision-making process have stabilized in a manner consistent with other successful ASF projects. Join the MXNet scientific community to contribute, learn, and get answers to your questions.