Qwen3-Coder-Next
Qwen3-Coder-Next is an open-weight language model specifically designed for coding agents and local development that delivers advanced coding reasoning, complex tool usage, and robust performance on long-horizon programming tasks with high efficiency, using a mixture-of-experts architecture that balances powerful capabilities with resource-friendly operation. It provides enhanced agentic coding abilities that help software developers, AI system builders, and automated coding workflows generate, debug, and reason about code with deep contextual understanding while recovering from execution errors, making it well-suited for autonomous coding agents and development-oriented applications. By achieving strong performance comparable to much larger parameter models while requiring fewer active parameters, Qwen3-Coder-Next enables cost-effective deployment for dynamic and complex programming workloads in research and production environments.
Learn more
MiniMax M2.5
MiniMax M2.5 is a frontier AI model engineered for real-world productivity across coding, agentic workflows, search, and office tasks. Extensively trained with reinforcement learning in hundreds of thousands of real-world environments, it achieves state-of-the-art performance in benchmarks such as SWE-Bench Verified and BrowseComp. The model demonstrates strong architectural thinking, decomposing complex problems before generating code across more than ten programming languages. M2.5 operates at high throughput speeds of up to 100 tokens per second, enabling faster completion of multi-step tasks. It is optimized for efficient reasoning, reducing token usage and execution time compared to previous versions. With dramatically lower pricing than competing frontier models, it delivers powerful performance at minimal cost. Integrated into MiniMax Agent, M2.5 supports professional-grade office workflows, financial modeling, and autonomous task execution.
Learn more
SWE-1.5
SWE-1.5 is the latest agent-model release by Cognition, purpose-built for software engineering and characterized by a “frontier-size” architecture comprising hundreds of billions of parameters and optimized end-to-end (model, inference engine, and agent harness) for both speed and intelligence. It achieves near-state-of-the-art coding performance and sets a new benchmark in latency, delivering inference speeds up to 950 tokens/second, roughly six times faster than its predecessor Haiku 4.5 and thirteen times faster than Sonnet 4.5. The model was trained using extensive reinforcement learning in realistic coding-agent environments with multi-turn workflows, unit tests, quality rubrics, and browser-based agentic execution; it also benefits from tightly integrated software tooling and high-throughput hardware (including thousands of GB200 NVL72 chips and a custom hypervisor infrastructure).
Learn more
SWE-1.6
SWE-1.6 is an engineering–focused AI model developed by Cognition and integrated into the Windsurf environment, designed to optimize both raw intelligence and what the company calls “model UX,” or the overall feel and efficiency of interacting with an AI agent. It represents a new iteration in the SWE model family, improving performance on benchmarks such as SWE-Bench Pro by over 10% compared to SWE-1.5 while maintaining similar underlying capabilities. It was trained from scratch to jointly improve reasoning quality and user experience, addressing issues observed in earlier versions such as overthinking simple problems, taking too many steps, looping in repetitive reasoning, and relying excessively on terminal commands instead of specialized tools. SWE-1.6 introduces behavioral improvements such as more frequent parallel tool usage, faster context retrieval, and reduced need for user input, resulting in smoother and more efficient workflows.
Learn more