Alternatives to CodeT5
Compare CodeT5 alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to CodeT5 in 2026. Compare features, ratings, user reviews, pricing, and more from CodeT5 competitors and alternatives in order to make an informed decision for your business.
-
1
Amazon CodeWhisperer
Amazon
Build apps faster with ML-powered coding companion. Accelerate application development with automatic code recommendations based on the code and comments in your IDE. Empower developers to use artificial intelligence (AI) responsibly to create syntactically correct and secure applications. Generate entire functions and logical code blocks without having to search and customize code snippets from the web. Stay focused and never leave the IDE, with real-time customized code recommendations for all your Java, Python, and JavaScript projects. Amazon CodeWhisperer is a machine learning (ML)–powered service that helps improve developer productivity by generating code recommendations based on their comments in natural language and code in the integrated development environment (IDE). Accelerate frontend and backend development by empowering developers with automatic code recommendations. Save time and effort by using CodeWhisperer to generate code to build and train your ML models. -
2
GLM-OCR
Z.ai
GLM-OCR is a multimodal optical character recognition model and open source repository that provides accurate, efficient, and comprehensive document understanding by combining text and visual modalities into a unified encoder–decoder architecture derived from the GLM-V family. Built with a visual encoder pre-trained on large-scale image–text data and a lightweight cross-modal connector feeding into a GLM-0.5B language decoder, the model supports layout detection, parallel region recognition, and structured output for text, tables, formulas, and complicated real-world document formats. It introduces Multi-Token Prediction (MTP) loss and stable full-task reinforcement learning to improve training efficiency, recognition accuracy, and generalization, achieving state-of-the-art benchmarks on major document understanding tasks.Starting Price: Free -
3
Mu
Microsoft
Mu is a 330-million-parameter encoder–decoder language model designed to power the agent in Windows settings by mapping natural-language queries to Settings function calls, running fully on-device via NPUs at over 100 tokens per second while maintaining high accuracy. Drawing on Phi Silica optimizations, Mu’s encoder–decoder architecture reuses a fixed-length latent representation to cut computation and memory overhead, yielding 47 percent lower first-token latency and 4.7× higher decoding speed on Qualcomm Hexagon NPUs compared to similar decoder-only models. Hardware-aware tuning, including a 2/3–1/3 encoder–decoder parameter split, weight sharing between input and output embeddings, Dual LayerNorm, rotary positional embeddings, and grouped-query attention, enables fast inference at over 200 tokens per second on devices like Surface Laptop 7 and sub-500 ms response times for settings queries. -
4
KamuSEO
KamuSEO
It's a complete visitor and SEO analytics, a great tool to analyze your site's visitors and analyze any site's information. It has the ability to analyze your own website's information. It has the ability to analyze any other website's information. It has a native API by which developers can integrate its facilities with another app. KamuSEO is an app to analyze your site visitors and analyze any site's information such as Alexa data, similar web data, whois data, social media data, Moz check, search engine index, Google page rank, IP analysis, malware check, etc. Input a domain name and you will get a js code. Copy the embedded js code and paste it into your web page. You will get a daily report about your website. You will get some bonus utility tools such as email encoder/decoder, metatag generator, tag generator, plagiarism check, valid email check, duplicate email filter, URL encoder/decoder, etc.Starting Price: $29 per month -
5
OpenAI Whisper
OpenAI
Whisper is an automatic speech recognition (ASR) system developed by OpenAI for converting spoken language into text. It is trained on 680,000 hours of multilingual and multitask audio data collected from the web. The model is designed to handle diverse accents, background noise, and technical language with high accuracy. Whisper supports transcription in multiple languages as well as translation into English. It uses an encoder-decoder Transformer architecture to process audio inputs and generate text outputs. The system can also perform tasks like language identification and timestamp generation. Overall, Whisper enables developers to build robust voice-enabled applications with ease. -
6
CodeQwen
Alibaba
CodeQwen is the code version of Qwen, the large language model series developed by the Qwen team, Alibaba Cloud. It is a transformer-based decoder-only language model pre-trained on a large amount of data of codes. Strong code generation capabilities and competitive performance across a series of benchmarks. Supporting long context understanding and generation with the context length of 64K tokens. CodeQwen supports 92 coding languages and provides excellent performance in text-to-SQL, bug fixes, etc. You can just write several lines of code with transformers to chat with CodeQwen. Essentially, we build the tokenizer and the model from pre-trained methods, and we use the generate method to perform chatting with the help of the chat template provided by the tokenizer. We apply the ChatML template for chat models following our previous practice. The model completes the code snippets according to the given prompts, without any additional formatting.Starting Price: Free -
7
Qwen-7B
Alibaba
Qwen-7B is the 7B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Alibaba Cloud. Qwen-7B is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc. Additionally, based on the pretrained Qwen-7B, we release Qwen-7B-Chat, a large-model-based AI assistant, which is trained with alignment techniques. The features of the Qwen-7B series include: Trained with high-quality pretraining data. We have pretrained Qwen-7B on a self-constructed large-scale high-quality dataset of over 2.2 trillion tokens. The dataset includes plain texts and codes, and it covers a wide range of domains, including general domain data and professional domain data. Strong performance. In comparison with the models of the similar model size, we outperform the competitors on a series of benchmark datasets, which evaluates natural language understanding, mathematics, coding, etc. And more.Starting Price: Free -
8
OPT
Meta
Large language models, which are often trained for hundreds of thousands of compute days, have shown remarkable capabilities for zero- and few-shot learning. Given their computational cost, these models are difficult to replicate without significant capital. For the few that are available through APIs, no access is granted to the full model weights, making them difficult to study. We present Open Pre-trained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M to 175B parameters, which we aim to fully and responsibly share with interested researchers. We show that OPT-175B is comparable to GPT-3, while requiring only 1/7th the carbon footprint to develop. We are also releasing our logbook detailing the infrastructure challenges we faced, along with code for experimenting with all of the released models. -
9
Llama 2
Meta
The next generation of our open source large language model. This release includes model weights and starting code for pretrained and fine-tuned Llama language models — ranging from 7B to 70B parameters. Llama 2 pretrained models are trained on 2 trillion tokens, and have double the context length than Llama 1. Its fine-tuned models have been trained on over 1 million human annotations. Llama 2 outperforms other open source language models on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests. Llama 2 was pretrained on publicly available online data sources. The fine-tuned model, Llama-2-chat, leverages publicly available instruction datasets and over 1 million human annotations. We have a broad range of supporters around the world who believe in our open approach to today’s AI — companies that have given early feedback and are excited to build with Llama 2.Starting Price: Free -
10
CodeGemma
Google
CodeGemma is a collection of powerful, lightweight models that can perform a variety of coding tasks like fill-in-the-middle code completion, code generation, natural language understanding, mathematical reasoning, and instruction following. CodeGemma has 3 model variants, a 7B pre-trained variant that specializes in code completion and generation from code prefixes and/or suffixes, a 7B instruction-tuned variant for natural language-to-code chat and instruction following; and a state-of-the-art 2B pre-trained variant that provides up to 2x faster code completion. Complete lines, and functions, and even generate entire blocks of code, whether you're working locally or using Google Cloud resources. Trained on 500 billion tokens of primarily English language data from web documents, mathematics, and code, CodeGemma models generate code that's not only more syntactically correct but also semantically meaningful, reducing errors and debugging time. -
11
Olmo 2
Ai2
Olmo 2 is a family of fully open language models developed by the Allen Institute for AI (AI2), designed to provide researchers and developers with transparent access to training data, open-source code, reproducible training recipes, and comprehensive evaluations. These models are trained on up to 5 trillion tokens and are competitive with leading open-weight models like Llama 3.1 on English academic benchmarks. Olmo 2 emphasizes training stability, implementing techniques to prevent loss spikes during long training runs, and utilizes staged training interventions during late pretraining to address capability deficiencies. The models incorporate state-of-the-art post-training methodologies from AI2's Tülu 3, resulting in the creation of Olmo 2-Instruct models. An actionable evaluation framework, the Open Language Modeling Evaluation System (OLMES), was established to guide improvements through development stages, consisting of 20 evaluation benchmarks assessing core capabilities. -
12
Keepsake
Replicate
Keepsake is an open-source Python library designed to provide version control for machine learning experiments and models. It enables users to automatically track code, hyperparameters, training data, model weights, metrics, and Python dependencies, ensuring that all aspects of the machine learning workflow are recorded and reproducible. Keepsake integrates seamlessly with existing workflows by requiring minimal code additions, allowing users to continue training as usual while Keepsake saves code and weights to Amazon S3 or Google Cloud Storage. This facilitates the retrieval of code and weights from any checkpoint, aiding in re-training or model deployment. Keepsake supports various machine learning frameworks, including TensorFlow, PyTorch, scikit-learn, and XGBoost, by saving files and dictionaries in a straightforward manner. It also offers features such as experiment comparison, enabling users to analyze differences in parameters, metrics, and dependencies across experiments.Starting Price: Free -
13
Amazon SageMaker JumpStart
Amazon
Amazon SageMaker JumpStart is a machine learning (ML) hub that can help you accelerate your ML journey. With SageMaker JumpStart, you can access built-in algorithms with pretrained models from model hubs, pretrained foundation models to help you perform tasks such as article summarization and image generation, and prebuilt solutions to solve common use cases. In addition, you can share ML artifacts, including ML models and notebooks, within your organization to accelerate ML model building and deployment. SageMaker JumpStart provides hundreds of built-in algorithms with pretrained models from model hubs, including TensorFlow Hub, PyTorch Hub, HuggingFace, and MxNet GluonCV. You can also access built-in algorithms using the SageMaker Python SDK. Built-in algorithms cover common ML tasks, such as data classifications (image, text, tabular) and sentiment analysis. -
14
Hugging Face Transformers
Hugging Face
Transformers is a library of pretrained natural language processing, computer vision, audio, and multimodal models for inference and training. Use Transformers to train models on your data, build inference applications, and generate text with large language models. Explore the Hugging Face Hub today to find a model and use Transformers to help you get started right away. Simple and optimized inference class for many machine learning tasks like text generation, image segmentation, automatic speech recognition, document question answering, and more. A comprehensive trainer that supports features such as mixed precision, torch.compile, and FlashAttention for training and distributed training for PyTorch models. Fast text generation with large language models and vision language models. Every model is implemented from only three main classes (configuration, model, and preprocessor) and can be quickly used for inference or training.Starting Price: $9 per month -
15
yarl
Python Software Foundation
All URL parts, scheme, user, password, host, port, path, query, and fragment are accessible by properties. All URL manipulations produce a new URL object. Strings passed to constructor and modification methods are automatically encoded giving canonical representation as result. Regular properties are percent-decoded, use raw_ versions for getting encoded strings. Human-readable representation of URL is available as .human_repr(). PyPI contains binary wheels for Linux, Windows and MacOS. If you want to install yarl on another operating system (like Alpine Linux, which is not manylinux-compliant because of the missing glibc and therefore, cannot be used with our wheels) the tarball will be used to compile the library from the source code. It requires a C compiler and Python headers installed. Please note that the pure-Python (uncompiled) version is much slower. However, PyPy always uses a pure-Python implementation, and, as such, it is unaffected by this variable.Starting Price: Free -
16
Codestral
Mistral AI
We introduce Codestral, our first-ever code model. Codestral is an open-weight generative AI model explicitly designed for code generation tasks. It helps developers write and interact with code through a shared instruction and completion API endpoint. As it masters code and English, it can be used to design advanced AI applications for software developers. Codestral is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash. It also performs well on more specific ones like Swift and Fortran. This broad language base ensures Codestral can assist developers in various coding environments and projects.Starting Price: Free -
17
StableCode
Stability AI
StableCode offers a unique way for developers to become more efficient by using three different models to help in their coding. The base model was first trained on a diverse set of programming languages from the stack-dataset (v1.2) from BigCode and then trained further with popular languages like Python, Go, Java, Javascript, C, markdown and C++. In total, we trained our models on 560B tokens of code on our HPC cluster. After the base model had been established, the instruction model was then tuned for specific use cases to help solve complex programming tasks. ~120,000 code instruction/response pairs in Alpaca format were trained on the base model to achieve this result. StableCode is the ideal building block for those wanting to learn more about coding, and the long-context window model is the perfect assistant to ensure single and multiple-line autocomplete suggestions are available for the user. This model is built to handle a lot more code at once. -
18
CodeGeeX
AMiner
We introduce CodeGeeX, a large-scale multilingual code generation model with 13 billion parameters, pre-trained on a large code corpus of more than 20 programming languages. Based on CodeGeeX, we develop a VS Code extension (search 'CodeGeeX' in the Extension Marketplace) that assists the programming of different programming languages. Besides the multilingual code generation/translation abilities, we turn CodeGeeX into a custom programming assistant using its few-shot ability. It means that when a few examples are provided as extra prompts in the input, CodeGeeX will imitate what are done by these examples and generate codes accordingly. Some cool features can be implemented using this ability, like code explanation, summarization, generation with specific coding style, and more. For example, one can add code snippets with his/her own coding style, and CodeGeeX will generate codes in a similar way. You can also try prompts with specific formats to inspire CodeGeeX for new skills.Starting Price: Free -
19
The CodeGround
The CodeGround
TheCodeground is an online integrated development environment that offers a suite of tools for real-time coding practice and collaboration. It supports multiple programming languages, including Rust, GoLang, Node.js, Python, Java, HTML, CSS, and JavaScript. Users can engage in live code sharing, conduct code interviews, and access insightful articles through the Reads section. The platform features an interface similar to Visual Studio Code, complete with autocomplete functionality, JSON differentiation, and a JWT decoder, enhancing the coding experience. TheCodeground is accessible via web browsers and can also be installed as a desktop application on Mac, Windows, and Linux. With The Code Ground, you can code from any device without the hassle of setup. Our cloud-based platform provides instant execution, rich tools, and a smooth coding experience. The CodeGround ensures you have everything you need for efficient development and accurate data handling. -
20
Granite Code
IBM
We introduce the Granite series of decoder-only code models for code generative tasks (e.g., fixing bugs, explaining code, documenting code), trained with code written in 116 programming languages. A comprehensive evaluation of the Granite Code model family on diverse tasks demonstrates that our models consistently reach state-of-the-art performance among available open source code LLMs. The key advantages of Granite Code models include: All-rounder Code LLM: Granite Code models achieve competitive or state-of-the-art performance on different kinds of code-related tasks, including code generation, explanation, fixing, editing, translation, and more. Demonstrating their ability to solve diverse coding tasks. Trustworthy Enterprise-Grade LLM: All our models are trained on license-permissible data collected following IBM's AI Ethics principles and guided by IBM’s Corporate Legal team for trustworthy enterprise usage.Starting Price: Free -
21
{CodeWhizz}
{CodeWhizz}
The AI-Powered Python and JavaScript Generator/Debugger/Tutor. Become a pro-coder in seconds. Generate pro-level code in an instant. Type what you need, run the program, and boom! The Whizzy AI model will compute your request and generate your code in an editable code window, so you can touch it up and personalize it however you need. Don't hassle with clunky and slow IDE's, the integrated CodeEngine will run your Python code and generate outputs, and plots, seamlessly. The ScriptRepo allows you to save your favorite creations with ease. We'll keep them secure so your can come back to them anytime. Limited availability. Request access now and secure your own personalized AI-Powered Python Code generation tool.Starting Price: $37.50 per month -
22
Qwen3-Coder
Qwen
Qwen3‑Coder is an agentic code model available in multiple sizes, led by the 480B‑parameter Mixture‑of‑Experts variant (35B active) that natively supports 256K‑token contexts (extendable to 1M) and achieves state‑of‑the‑art results comparable to Claude Sonnet 4. Pre‑training on 7.5T tokens (70 % code) and synthetic data cleaned via Qwen2.5‑Coder optimized both coding proficiency and general abilities, while post‑training employs large‑scale, execution‑driven reinforcement learning, scaling test‑case generation for diverse coding challenges, and long‑horizon RL across 20,000 parallel environments to excel on multi‑turn software‑engineering benchmarks like SWE‑Bench Verified without test‑time scaling. Alongside the model, the open source Qwen Code CLI (forked from Gemini Code) unleashes Qwen3‑Coder in agentic workflows with customized prompts, function calling protocols, and seamless integration with Node.js, OpenAI SDKs, and environment variables.Starting Price: Free -
23
Azure OpenAI Service
Microsoft
Apply advanced coding and language models to a variety of use cases. Leverage large-scale, generative AI models with deep understandings of language and code to enable new reasoning and comprehension capabilities for building cutting-edge applications. Apply these coding and language models to a variety of use cases, such as writing assistance, code generation, and reasoning over data. Detect and mitigate harmful use with built-in responsible AI and access enterprise-grade Azure security. Gain access to generative models that have been pretrained with trillions of words. Apply them to new scenarios including language, code, reasoning, inferencing, and comprehension. Customize generative models with labeled data for your specific scenario using a simple REST API. Fine-tune your model's hyperparameters to increase accuracy of outputs. Use the few-shot learning capability to provide the API with examples and achieve more relevant results.Starting Price: $0.0004 per 1000 tokens -
24
Qwen Code
Qwen
Qwen3‑Coder is an agentic code model available in multiple sizes, led by the 480B‑parameter Mixture‑of‑Experts variant (35B active) that natively supports 256K‑token contexts (extendable to 1M) and achieves state‑of‑the‑art results on Agentic Coding, Browser‑Use, and Tool‑Use tasks comparable to Claude Sonnet 4. Pre‑training on 7.5T tokens (70 % code) and synthetic data cleaned via Qwen2.5‑Coder optimized both coding proficiency and general abilities, while post‑training employs large‑scale, execution‑driven reinforcement learning and long‑horizon RL across 20,000 parallel environments to excel on multi‑turn software‑engineering benchmarks like SWE‑Bench Verified without test‑time scaling. Alongside the model, the open source Qwen Code CLI (forked from Gemini Code) unleashes Qwen3‑Coder in agentic workflows with customized prompts, function calling protocols, and seamless integration with Node.js, OpenAI SDKs, and more.Starting Price: Free -
25
BERT
Google
BERT is a large language model and a method of pre-training language representations. Pre-training refers to how BERT is first trained on a large source of text, such as Wikipedia. You can then apply the training results to other Natural Language Processing (NLP) tasks, such as question answering and sentiment analysis. With BERT and AI Platform Training, you can train a variety of NLP models in about 30 minutes.Starting Price: Free -
26
ERNIE 3.0 Titan
Baidu
Pre-trained language models have achieved state-of-the-art results in various Natural Language Processing (NLP) tasks. GPT-3 has shown that scaling up pre-trained language models can further exploit their enormous potential. A unified framework named ERNIE 3.0 was recently proposed for pre-training large-scale knowledge enhanced models and trained a model with 10 billion parameters. ERNIE 3.0 outperformed the state-of-the-art models on various NLP tasks. In order to explore the performance of scaling up ERNIE 3.0, we train a hundred-billion-parameter model called ERNIE 3.0 Titan with up to 260 billion parameters on the PaddlePaddle platform. Furthermore, We design a self-supervised adversarial loss and a controllable language modeling loss to make ERNIE 3.0 Titan generate credible and controllable texts. -
27
GLM-5
Zhipu AI
GLM-5 is Z.ai’s latest large language model built for complex systems engineering and long-horizon agentic tasks. It scales significantly beyond GLM-4.5, increasing total parameters and training data while integrating DeepSeek Sparse Attention to reduce deployment costs without sacrificing long-context capacity. The model combines enhanced pre-training with a new asynchronous reinforcement learning infrastructure called slime, improving training efficiency and post-training refinement. GLM-5 achieves best-in-class performance among open-source models across reasoning, coding, and agent benchmarks, narrowing the gap with leading frontier models. It ranks highly on evaluations such as Vending Bench 2, demonstrating strong long-term planning and operational capabilities. The model is open-sourced under the MIT License.Starting Price: Free -
28
Callstack.ai PR Reviewer
Callstack.ai
AI-powered pull request reviewer with deep contextual insights, tailored feedback, and one-click automated setup. Callstack.ai PR Reviewer saves you time and lowers error risk with automatic PR summaries, security & bug checks, and performance optimization suggestions. Automatic PR Summaries Understand code changes quickly with auto-generated summaries and diagram. Context-Aware Feedback Callstack.ai aligns with your team’s coding standards by understanding the core structure of your code for context-driven insights. Customizable Reviews Tailor Callstack.ai to provide feedback that aligns with your unique coding standards and requirements. Compatible with Major Programming Languages We currently support repositories written in the following languages: javascript, typescript, python, java, kotlin, php, go, ruby, rust, react & vue.Starting Price: $285/month (per 100 reviews) -
29
Kaywa
Kaywa
QR Codes are a proven and easy-to-understand technology to bridge the gap between the physical (aka meatspace) and the digital world. You can encode any type of textual information in a QR Code, for example your website's address, a Facebook page, a coupon, a contact. Once published on paper, any other physical support (or even on the web), people with a QR App can scan your QR Code. By scanning they decode the information and the app then renders the website, the Facebook page, the coupon or the contact. QR Codes come in two types, either static or dynamic. We always recommend to use dynamic codes. Static code simply encode the information. With Kaywa you can create an unlimited number of static QR Codes free of charge. More interesting are dynamic codes and our QR MGMT only stores these. dynamic codes can be changed, tracked and they are ideal for scanning with a mobile phone.Starting Price: $13.75 per month -
30
StarCoder
BigCode
StarCoder and StarCoderBase are Large Language Models for Code (Code LLMs) trained on permissively licensed data from GitHub, including from 80+ programming languages, Git commits, GitHub issues, and Jupyter notebooks. Similar to LLaMA, we trained a ~15B parameter model for 1 trillion tokens. We fine-tuned StarCoderBase model for 35B Python tokens, resulting in a new model that we call StarCoder. We found that StarCoderBase outperforms existing open Code LLMs on popular programming benchmarks and matches or surpasses closed models such as code-cushman-001 from OpenAI (the original Codex model that powered early versions of GitHub Copilot). With a context length of over 8,000 tokens, the StarCoder models can process more input than any other open LLM, enabling a wide range of interesting applications. For example, by prompting the StarCoder models with a series of dialogues, we enabled them to act as a technical assistant.Starting Price: Free -
31
Qwen3-Omni
Alibaba
Qwen3-Omni is a natively end-to-end multilingual omni-modal foundation model that processes text, images, audio, and video and delivers real-time streaming responses in text and natural speech. It uses a Thinker-Talker architecture with a Mixture-of-Experts (MoE) design, early text-first pretraining, and mixed multimodal training to support strong performance across all modalities without sacrificing text or image quality. The model supports 119 text languages, 19 speech input languages, and 10 speech output languages. It achieves state-of-the-art results: across 36 audio and audio-visual benchmarks, it hits open-source SOTA on 32 and overall SOTA on 22, outperforming or matching strong closed-source models such as Gemini-2.5 Pro and GPT-4o. To reduce latency, especially in audio/video streaming, Talker predicts discrete speech codecs via a multi-codebook scheme and replaces heavier diffusion approaches. -
32
Qwen3-Max
Alibaba
Qwen3-Max is Alibaba’s latest trillion-parameter large language model, designed to push performance in agentic tasks, coding, reasoning, and long-context processing. It is built atop the Qwen3 family and benefits from the architectural, training, and inference advances introduced there; mixing thinker and non-thinker modes, a “thinking budget” mechanism, and support for dynamic mode switching based on complexity. The model reportedly processes extremely long inputs (hundreds of thousands of tokens), supports tool invocation, and exhibits strong performance on benchmarks in coding, multi-step reasoning, and agent benchmarks (e.g., Tau2-Bench). While its initial variant emphasizes instruction following (non-thinking mode), Alibaba plans to bring reasoning capabilities online to enable autonomous agent behavior. Qwen3-Max inherits multilingual support and extensive pretraining on trillions of tokens, and it is delivered via API interfaces compatible with OpenAI-style functions.Starting Price: Free -
33
Kimi K2
Moonshot AI
Kimi K2 is a state-of-the-art open source large language model series built on a mixture-of-experts (MoE) architecture, featuring 1 trillion total parameters and 32 billion activated parameters for task-specific efficiency. Trained with the Muon optimizer on over 15.5 trillion tokens and stabilized by MuonClip’s attention-logit clamping, it delivers exceptional performance in frontier knowledge, reasoning, mathematics, coding, and general agentic workflows. Moonshot AI provides two variants, Kimi-K2-Base for research-level fine-tuning and Kimi-K2-Instruct pre-trained for immediate chat and tool-driven interactions, enabling both custom development and drop-in agentic capabilities. Benchmarks show it outperforms leading open source peers and rivals top proprietary models in coding tasks and complex task breakdowns, while its 128 K-token context length, tool-calling API compatibility, and support for industry-standard inference engines.Starting Price: Free -
34
Baichuan-13B
Baichuan Intelligent Technology
Baichuan-13B is an open source and commercially available large-scale language model containing 13 billion parameters developed by Baichuan Intelligent following Baichuan -7B . It has achieved the best results of the same size on authoritative Chinese and English benchmarks. This release contains two versions of pre-training ( Baichuan-13B-Base ) and alignment ( Baichuan-13B-Chat ). Larger size, more data : Baichuan-13B further expands the number of parameters to 13 billion on the basis of Baichuan -7B , and trains 1.4 trillion tokens on high-quality corpus, which is 40% more than LLaMA-13B. It is currently open source The model with the largest amount of training data in the 13B size. Support Chinese and English bilingual, use ALiBi position code, context window length is 4096.Starting Price: Free -
35
Gemma 2
Google
A family of state-of-the-art, light-open models created from the same research and technology that were used to create Gemini models. These models incorporate comprehensive security measures and help ensure responsible and reliable AI solutions through selected data sets and rigorous adjustments. Gemma models achieve exceptional comparative results in their 2B, 7B, 9B, and 27B sizes, even outperforming some larger open models. With Keras 3.0, enjoy seamless compatibility with JAX, TensorFlow, and PyTorch, allowing you to effortlessly choose and change frameworks based on task. Redesigned to deliver outstanding performance and unmatched efficiency, Gemma 2 is optimized for incredibly fast inference on various hardware. The Gemma family of models offers different models that are optimized for specific use cases and adapt to your needs. Gemma models are large text-to-text lightweight language models with a decoder, trained in a huge set of text data, code, and mathematical content. -
36
Forefront
Forefront.ai
Powerful language models a click away. Join over 8,000 developers building the next wave of world-changing applications. Fine-tune and deploy GPT-J, GPT-NeoX, Codegen, and FLAN-T5. Multiple models, each with different capabilities and price points. GPT-J is the fastest model, while GPT-NeoX is the most powerful—and more are on the way. Use these models for classification, entity extraction, code generation, chatbots, content generation, summarization, paraphrasing, sentiment analysis, and much more. These models have been pre-trained on a vast amount of text from the open internet. Fine-tuning improves upon this for specific tasks by training on many more examples than can fit in a prompt, letting you achieve better results on a wide number of tasks. -
37
RoBERTa
Meta
RoBERTa builds on BERT’s language masking strategy, wherein the system learns to predict intentionally hidden sections of text within otherwise unannotated language examples. RoBERTa, which was implemented in PyTorch, modifies key hyperparameters in BERT, including removing BERT’s next-sentence pretraining objective, and training with much larger mini-batches and learning rates. This allows RoBERTa to improve on the masked language modeling objective compared with BERT and leads to better downstream task performance. We also explore training RoBERTa on an order of magnitude more data than BERT, for a longer amount of time. We used existing unannotated NLP datasets as well as CC-News, a novel set drawn from public news articles.Starting Price: Free -
38
Xgen-small
Salesforce
Xgen-small is an enterprise-ready compact language model developed by Salesforce AI Research, designed to deliver long-context performance at a predictable, low cost. It combines domain-focused data curation, scalable pre-training, length extension, instruction fine-tuning, and reinforcement learning to meet the complex, high-volume inference demands of modern enterprises. Unlike traditional large models, Xgen-small offers efficient processing of extensive contexts, enabling the synthesis of information from internal documentation, code repositories, research reports, and real-time data streams. With sizes optimized at 4B and 9B parameters, it provides a strategic advantage by balancing cost efficiency, privacy safeguards, and long-context understanding, making it a sustainable and predictable solution for deploying Enterprise AI at scale. -
39
Olmo 3
Ai2
Olmo 3 is a fully open model family spanning 7 billion and 32 billion parameter variants that delivers not only high-performing base, reasoning, instruction, and reinforcement-learning models, but also exposure of the entire model flow, including raw training data, intermediate checkpoints, training code, long-context support (65,536 token window), and provenance tooling. Starting with the Dolma 3 dataset (≈9 trillion tokens) and its disciplined mix of web text, scientific PDFs, code, and long-form documents, the pre-training, mid-training, and long-context phases shape the base models, which are then post-trained via supervised fine-tuning, direct preference optimisation, and RL with verifiable rewards to yield the Think and Instruct variants. The 32 B Think model is described as the strongest fully open reasoning model to date, competitively close to closed-weight peers in math, code, and complex reasoning.Starting Price: Free -
40
Codey
Google
Codey accelerates software development with real-time code completion and generation, customizable to a customer’s own codebase. This code generation model supports 20+ coding languages, including Go, Google Standard SQL, Java, Javascript, Python, and Typescript. It enables a wide variety of coding tasks, helping developers to work faster and close skills gaps through: Code completion: Codey suggests the next few lines based on the context of code entered into the prompt. Code generation: Codey generates code based on natural language prompts from a developer. Code chat: Codey lets developers converse with a bot to get help with debugging, documentation, learning new concepts, and other code-related questions. -
41
MLflow
MLflow
MLflow is an open source platform to manage the ML lifecycle, including experimentation, reproducibility, deployment, and a central model registry. MLflow currently offers four components. Record and query experiments: code, data, config, and results. Package data science code in a format to reproduce runs on any platform. Deploy machine learning models in diverse serving environments. Store, annotate, discover, and manage models in a central repository. The MLflow Tracking component is an API and UI for logging parameters, code versions, metrics, and output files when running your machine learning code and for later visualizing the results. MLflow Tracking lets you log and query experiments using Python, REST, R API, and Java API APIs. An MLflow Project is a format for packaging data science code in a reusable and reproducible way, based primarily on conventions. In addition, the Projects component includes an API and command-line tools for running projects. -
42
Codespy
Codespy
Codespy AI Detector is a powerful tool designed to identify AI-generated code within software projects quickly and accurately. It supports popular programming languages such as Java, Python, JavaScript, C++, C#, and PHP. The platform helps developers find AI-written code from models like ChatGPT, Gemini, and Claude, which can introduce bugs or unexpected errors. Codespy integrates seamlessly with common development environments like Visual Studio Code and is available as a ChatGPT plugin. Its technology enables teams to create processes and guardrails around AI code usage to reduce risk and improve code quality. With simple pricing plans and no credit card required for the free tier, Codespy is accessible to individuals and businesses of all sizes.Starting Price: $27.98/month -
43
Codeanywhere
Codeanywhere
Our Cloud IDE saves you time by deploying a development environment in seconds, enabling you to code, learn, build, and collaborate on your projects. With our amazing web-based code editor in Codeanywhere, you will forget you ever used any other code editor. All major programming languages fully supported, including JavaScript/TypeScript, PHP, Python, Ruby, Go, Java, C/C++, C# any many other. Intelligent editing features such as auto-complete, code refactor, go to definition, rename symbol and many others. Debug code with break points, call stacks and interactive console. Fully featured Git client. Expandable with vast amount of existing extensions. You can also spin up powerful containers in seconds, that can be fully preconfigured for the programming environment of your choice. Develop and run your code on our infrastructure with full sudo access. Prebuilt development environments for all major programming languages, packed with tools and database preinstalled.Starting Price: $2.50 per user per month -
44
Spark NLP
John Snow Labs
Experience the power of large language models like never before, unleashing the full potential of Natural Language Processing (NLP) with Spark NLP, the open source library that delivers scalable LLMs. The full code base is open under the Apache 2.0 license, including pre-trained models and pipelines. The only NLP library built natively on Apache Spark. The most widely used NLP library in the enterprise. Spark ML provides a set of machine learning applications that can be built using two main components, estimators and transformers. The estimators have a method that secures and trains a piece of data to such an application. The transformer is generally the result of a fitting process and applies changes to the target dataset. These components have been embedded to be applicable to Spark NLP. Pipelines are a mechanism for combining multiple estimators and transformers in a single workflow. They allow multiple chained transformations along a machine-learning task.Starting Price: Free -
45
CodePal
CodePal.ai
CodePal is your ultimate coding companion. It is a comprehensive platform that offers a range of coding helpers and tools to assist developers. It is great for students, beginners, experienced developers, and companies who want to improve their development process. The free plan includes access to all the tools and helpers, but with some limitation. It's a great option for hobbyists who want to try out CodePal. AI Code generators can do a variety of tasks, such as programming, apply transformations and manipulations on the code, and many other tasks. AI Code Generators are very useful for learning, as they can give a good grasp of how the task at hand should be written in code. Code generation can be used with many different programming languages, including Java, C#, Python, and TypeScript, among others. The choice of programming language will often depend on the specific needs of the project, as well as the availability of code generation tools and frameworks for that language.Starting Price: $15 per month -
46
Figstack
Figstack
Reading and writing code is hard. Figstack helps you interpret the details so you can focus on building what matters. Powered by AI and trained with billions of lines of code, Figstack supercharges your ability to read and write code across languages. Having trouble understanding the program that you're looking at? Use the explain code function to explain it to you in natural language. Need to convert Python to Go or Ruby to Javascript? The language translator function lets you reliably swap from one language to another. Don't like writing documentation for your functions? Have Figstack's docstring writer craft detailed docstrings so your functions are always readable and maintainable. Measure the efficiency of your code in Big O notation using the time complexity function and see how you can optimize your program. Get the time complexity of your program. Automate documentation for your functions. Translate programming languages. -
47
ZZZ Code AI
ZZZ Code AI
ZZZ Code AI is an AI-powered coding assistant designed to support developers across various programming tasks. It offers a suite of tools, including AI Code Generator, AI Bug Detector, AI Code Explainer, AI Code Refactor, AI Code Review, AI Code Converter, and AI Code Documentation. It supports multiple programming languages such as Python, C#, C++, Java, JavaScript, HTML, CSS, SQL, and Excel formulas. Users can input their coding requirements or questions, and the AI provides instant responses, code snippets, explanations, or conversions as needed. Specialized tools are available for specific languages and frameworks, including Dapper and Entity Framework Core. It is accessible online without the need for account creation, although character limits apply to prevent abuse. ZZZ Code AI aims to enhance productivity and reduce errors for both novice and experienced developers by automating routine coding tasks and providing immediate assistance.Starting Price: Free -
48
Qwen3
Alibaba
Qwen3, the latest iteration of the Qwen family of large language models, introduces groundbreaking features that enhance performance across coding, math, and general capabilities. With models like the Qwen3-235B-A22B and Qwen3-30B-A3B, Qwen3 achieves impressive results compared to top-tier models, thanks to its hybrid thinking modes that allow users to control the balance between deep reasoning and quick responses. The platform supports 119 languages and dialects, making it an ideal choice for global applications. Its pre-training process, which uses 36 trillion tokens, enables robust performance, and advanced reinforcement learning (RL) techniques continue to refine its capabilities. Available on platforms like Hugging Face and ModelScope, Qwen3 offers a powerful tool for developers and researchers working in diverse fields.Starting Price: Free -
49
Code Llama
Meta
Code Llama is a large language model (LLM) that can use text prompts to generate code. Code Llama is state-of-the-art for publicly available LLMs on code tasks, and has the potential to make workflows faster and more efficient for current developers and lower the barrier to entry for people who are learning to code. Code Llama has the potential to be used as a productivity and educational tool to help programmers write more robust, well-documented software. Code Llama is a state-of-the-art LLM capable of generating code, and natural language about code, from both code and natural language prompts. Code Llama is free for research and commercial use. Code Llama is built on top of Llama 2 and is available in three models: Code Llama, the foundational code model; Codel Llama - Python specialized for Python; and Code Llama - Instruct, which is fine-tuned for understanding natural language instructions.Starting Price: Free -
50
DeepSeekMath
DeepSeek
DeepSeekMath is a specialized 7B parameter language model developed by DeepSeek-AI, designed to push the boundaries of mathematical reasoning in open-source language models. It starts from the DeepSeek-Coder-v1.5 7B model and undergoes further pre-training with 120B math-related tokens sourced from Common Crawl, alongside natural language and code data. DeepSeekMath has demonstrated remarkable performance, achieving a 51.7% score on the competition-level MATH benchmark without external tools or voting techniques, closely competing with the likes of Gemini-Ultra and GPT-4. The model's capabilities are enhanced by a meticulous data selection pipeline and the introduction of Group Relative Policy Optimization (GRPO), which optimizes both mathematical reasoning and memory usage. DeepSeekMath is available in base, instruct, and RL versions, supporting both research and commercial use, and is aimed at those looking to explore or apply advanced mathematical problem-solving in AI contexts.Starting Price: Free