19 Integrations with Google Antigravity

View a list of Google Antigravity integrations and software that integrates with Google Antigravity below. Compare the best Google Antigravity integrations as well as features, ratings, user reviews, and pricing of software that integrates with Google Antigravity. Here are the current Google Antigravity integrations in 2026:

  • 1
    Google AI Studio
    Google AI Studio is a unified development platform that helps teams explore, build, and deploy applications using Google’s most advanced AI models, including Gemini 3. It brings text, image, audio, and video models together in one interactive playground. With vibe coding, developers can use natural language to quickly turn ideas into working AI applications. The platform reduces friction by generating functional apps that are ready for deployment with minimal setup. Built-in integrations like Google Search enhance real-world use cases. Google AI Studio also centralizes API key management, usage monitoring, and billing. It offers a fast, intuitive path from prompt to production powered by vibe coding workflows.
    Starting Price: Free
    View Software
    Visit Website
  • 2
    Gemini 3 Pro
    Gemini 3 Pro is Google’s most advanced multimodal AI model, built for developers who want to bring ideas to life with intelligence, precision, and creativity. It delivers breakthrough performance across reasoning, coding, and multimodal understanding—surpassing Gemini 2.5 Pro in both speed and capability. The model excels in agentic workflows, enabling autonomous coding, debugging, and refactoring across entire projects with long-context awareness. With superior performance in image, video, and spatial reasoning, Gemini 3 Pro powers next-generation applications in development, robotics, XR, and document intelligence. Developers can access it through the Gemini API, Google AI Studio, or Vertex AI, integrating seamlessly into existing tools and IDEs. Whether generating code, analyzing visuals, or building interactive apps from a single prompt, Gemini 3 Pro represents the future of intelligent, multimodal AI development.
    Starting Price: $19.99/month
  • 3
    Nano Banana Pro
    Nano Banana Pro is Google DeepMind’s advanced evolution of the original Nano Banana, designed to deliver studio-quality image generation with far greater accuracy, text rendering, and world knowledge. Built on Gemini 3 Pro, it brings improved reasoning capabilities that help users transform ideas into detailed visuals, diagrams, prototypes, and educational content. It produces highly legible multilingual text inside images, making it ideal for posters, logos, storyboards, and international designs. The model can also ground images in real-time information, pulling from Google Search to create infographics for recipes, weather data, or factual explanations. With powerful consistency controls, Nano Banana Pro can blend up to 14 images and maintain recognizable details across multiple people or elements. Its enhanced creative editing tools let users refine lighting, adjust focus, manipulate camera angles, and produce final outputs in up to 4K resolution.
  • 4
    Claude Opus 4.6
    Claude Opus 4.6 is an advanced AI model developed by Anthropic, designed for high-level reasoning, coding, and knowledge work tasks. It introduces significant improvements in coding, debugging, and code review capabilities. The model can handle long, complex workflows and sustain agentic tasks with greater reliability. It features a 1 million token context window in beta, enabling it to process and retain large amounts of information. Claude Opus 4.6 is optimized for tasks such as financial analysis, research, and document creation. It also integrates with tools like Excel and PowerPoint for enhanced productivity. Overall, it is a state-of-the-art AI model built for complex, real-world professional applications.
  • 5
    AppDeploy

    AppDeploy

    AppDeploy

    AppDeploy is chat-native deployment — go from AI chat to deployed app. Tell your current AI chat or agent what to build, AppDeploy.ai makes it real, without leaving the chat or touching infrastructure. Deploy a real, full-stack apps directly from ChatGPT, Claude, Cursor, Gemini, Claude Code, Codex, or any other AI assistant — and get a live URL in seconds, without leaving the chat. No Git, no CLI, no IDE required. Hosting, database, backend services, storage, auth, and AI integrations are handled automatically. Every deploy gives you a live application with a shareable URL — a real deployed app, not a prototype. AppDeploy works for builders of all levels - no setup screens, no technical decisions.
    Starting Price: $0
  • 6
    InsForge

    InsForge

    InsForge

    InsForge is an AI-native backend platform designed specifically for agentic development, providing everything needed to build, manage, and deploy full-stack applications through AI coding agents. It functions as a Backend-as-a-Service with built-in primitives such as a managed PostgreSQL database, authentication with OAuth and JWT, cloud storage, serverless functions, real-time updates, and AI integrations, all accessible through a structured, agent-friendly interface. Unlike traditional backends built for human developers, InsForge exposes its services through a semantic layer and an MCP server that allows AI agents to understand, reason about, and operate backend infrastructure end to end. This enables agents to configure databases, manage schemas, handle authentication flows, deploy logic, and maintain applications with minimal manual intervention.
    Starting Price: $25 per month
  • 7
    OpenSpec

    OpenSpec

    Fission AI

    OpenSpec is an open-source spec-driven development framework designed to bring structure and clarity to AI-assisted coding workflows. It introduces a lightweight specification layer that helps teams define requirements before writing code. The platform organizes each change into structured artifacts such as proposals, specifications, designs, and task lists. It integrates with over 20 AI coding tools, allowing developers to use their preferred assistants while maintaining consistency. OpenSpec emphasizes an iterative and flexible approach rather than rigid development phases. Its command-based workflow enables users to propose, implement, and archive features efficiently. Overall, OpenSpec helps developers align with AI systems, reduce ambiguity, and produce more predictable and reliable outcomes.
    Starting Price: Free
  • 8
    Gemini 3 Pro Image
    Gemini Image Pro is a high-capability, multimodal image-generation and editing system that enables users to create, transform, and refine visuals through natural-language prompts or by combining multiple input images, with support for consistent character and object appearance across edits, precise local transformations (such as background blur, object removal, style transfers or pose changes), and native world-knowledge understanding to ensure context-aware outcomes. It supports multi-image fusion, merging several photo inputs into a cohesive new image, and emphasizes design workflow features such as template-based outputs, brand-asset consistency, and repeated character/person-style appearances across scenes. It includes digital watermarking to tag AI-generated imagery and is available through the Gemini API, Google AI Studio, and Vertex AI platforms.
  • 9
    Gemini 3 Flash
    Gemini 3 Flash is Google’s latest AI model built to deliver frontier intelligence with exceptional speed and efficiency. It combines Pro-level reasoning with Flash-level latency, making advanced AI more accessible and affordable. The model excels in complex reasoning, multimodal understanding, and agentic workflows while using fewer tokens for everyday tasks. Gemini 3 Flash is designed to scale across consumer apps, developer tools, and enterprise platforms. It supports rapid coding, data analysis, video understanding, and interactive application development. By balancing performance, cost, and speed, Gemini 3 Flash redefines what fast AI can achieve.
  • 10
    gpt-oss-20b
    gpt-oss-20b is a 20-billion-parameter, text-only reasoning model released under the Apache 2.0 license and governed by OpenAI’s gpt-oss usage policy, built to enable seamless integration into custom AI workflows via the Responses API without reliance on proprietary infrastructure. Trained for robust instruction following, it supports adjustable reasoning effort, full chain-of-thought outputs, and native tool use (including web search and Python execution), producing structured, explainable answers. Developers must implement their own deployment safeguards, such as input filtering, output monitoring, and usage policies, to match the system-level protections of hosted offerings and mitigate risks from malicious or unintended behaviors. Its open-weight design makes it ideal for on-premises or edge deployments where control, customization, and transparency are paramount.
  • 11
    gpt-oss-120b
    gpt-oss-120b is a reasoning model engineered for deep, transparent thinking, delivering full chain-of-thought explanations, adjustable reasoning depth, and structured outputs, while natively invoking tools like web search and Python execution via the API. Built to slot seamlessly into self-hosted or edge deployments, it eliminates dependence on proprietary infrastructure. Although it includes default safety guardrails, its open-weight architecture allows fine-tuning that could override built-in controls, so implementers are responsible for adding input filtering, output monitoring, and governance measures to achieve enterprise-grade security. As a community–driven model card rather than a managed service spec, it emphasizes transparency, customization, and the need for downstream safety practices.
  • 12
    Claude Opus 4.1
    Claude Opus 4.1 is an incremental upgrade to Claude Opus 4 that boosts coding, agentic reasoning, and data-analysis performance without changing deployment complexity. It raises coding accuracy to 74.5 percent on SWE-bench Verified and sharpens in-depth research and detailed tracking for agentic search tasks. GitHub reports notable gains in multi-file code refactoring, while Rakuten Group highlights its precision in pinpointing exact corrections within large codebases without introducing bugs. Independent benchmarks show about a one-standard-deviation improvement on junior developer tests compared to Opus 4, mirroring major leaps seen in prior Claude releases. Opus 4.1 is available now to paid Claude users, in Claude Code, and via the Anthropic API (model ID claude-opus-4-1-20250805), as well as through Amazon Bedrock and Google Cloud Vertex AI, and integrates seamlessly into existing workflows with no additional setup beyond selecting the new model.
  • 13
    Claude Sonnet 4.5
    Claude Sonnet 4.5 is Anthropic’s latest frontier model, designed to excel in long-horizon coding, agentic workflows, and intensive computer use while maintaining safety and alignment. It achieves state-of-the-art performance on the SWE-bench Verified benchmark (for software engineering) and leads on OSWorld (a computer use benchmark), with the ability to sustain focus over 30 hours on complex, multi-step tasks. The model introduces improvements in tool handling, memory management, and context processing, enabling more sophisticated reasoning, better domain understanding (from finance and law to STEM), and deeper code comprehension. It supports context editing and memory tools to sustain long conversations or multi-agent tasks, and allows code execution and file creation within Claude apps. Sonnet 4.5 is deployed at AI Safety Level 3 (ASL-3), with classifiers protecting against inputs or outputs tied to risky domains, and includes mitigations against prompt injection.
  • 14
    Claude Opus 4.5
    Claude Opus 4.5 is Anthropic’s newest flagship model, delivering major improvements in reasoning, coding, agentic workflows, and real-world problem solving. It outperforms previous models and leading competitors on benchmarks such as SWE-bench, multilingual coding tests, and advanced agent evaluations. Opus 4.5 also introduces stronger safety features, including significantly higher resistance to prompt injection and improved alignment across sensitive tasks. Developers gain new controls through the Claude API—like effort parameters, context compaction, and advanced tool use—allowing for more efficient, longer-running agentic workflows. Product updates across Claude, Claude Code, the Chrome extension, and Excel integrations expand how users interact with the model for software engineering, research, and everyday productivity. Overall, Claude Opus 4.5 marks a substantial step forward in capability, reliability, and usability for developers, enterprises, and end users.
  • 15
    Nano Banana 2
    Nano Banana 2 is Google DeepMind’s latest image generation model, combining the advanced capabilities of Nano Banana Pro with the high-speed performance of Gemini Flash. It delivers improved world knowledge, enabling more accurate subject rendering and data-driven visuals grounded in real-time information. The model enhances precision text rendering and translation, making it ideal for marketing assets, infographics, and localized content. Users benefit from stronger instruction following, ensuring complex prompts are captured accurately. Nano Banana 2 supports subject consistency across multiple characters and objects within a single workflow. It offers production-ready output with customizable aspect ratios and resolutions up to 4K. Available across Gemini, Search, AI Studio, Google Cloud, and more, Nano Banana 2 brings high-quality visual generation at lightning-fast speed.
  • 16
    Gemini 3.1 Pro
    Gemini 3.1 Pro is Google’s upgraded core intelligence model designed for complex tasks that require advanced reasoning. Building on the Gemini 3 series, it delivers significant improvements in problem-solving performance and logical pattern recognition. On the ARC-AGI-2 benchmark, Gemini 3.1 Pro achieved a verified score of 77.1%, more than doubling the reasoning performance of Gemini 3 Pro. The model is engineered for challenges where simple answers are insufficient, enabling deeper analysis, synthesis, and creative output. It can generate practical outputs such as animated, website-ready SVGs directly from text prompts, combining intelligence with real-world usability. Gemini 3.1 Pro is rolling out in preview across consumer, developer, and enterprise platforms including the Gemini app, NotebookLM, Gemini API, Vertex AI, and Android Studio. With expanded access for Google AI Pro and Ultra users, 3.1 Pro sets a stronger baseline for ambitious agentic workflow & advanced applications.
  • 17
    Gemini 3.1 Flash Image
    Gemini 3.1 Flash Image is Google DeepMind’s latest image generation model, combining advanced Pro-level capabilities with lightning-fast performance. It delivers enhanced world knowledge, enabling more accurate subject rendering and data-informed visuals grounded in real-time information. The model improves precision text rendering and in-image translation, making it well-suited for marketing assets, infographics, and localized creative content. Stronger instruction following ensures complex prompts are executed with clarity and accuracy. Gemini 3.1 Flash Image maintains subject consistency across multiple characters and objects within a single workflow. It supports production-ready outputs with customizable aspect ratios and resolutions up to 4K. Available across Gemini, Search, AI Studio, Google Cloud, and more, it brings high-quality visual generation at Flash-level speed.
  • 18
    Gemini 3.1 Flash-Lite
    Gemini 3.1 Flash-Lite is Google’s fastest and most cost-efficient model in the Gemini 3 series, designed for high-volume developer workloads. It delivers strong performance at scale while maintaining affordability, with pricing set at $0.25 per million input tokens and $1.50 per million output tokens. The model significantly improves speed, offering a 2.5x faster time to first answer token and a 45% increase in output speed compared to Gemini 2.5 Flash. Despite its lower cost tier, it achieves high benchmark results, including an Elo score of 1432 and strong performance across reasoning and multimodal evaluations. Gemini 3.1 Flash-Lite supports adaptive “thinking levels,” allowing developers to control how much reasoning power is used for different tasks. It is suitable for large-scale applications such as translation, content moderation, user interface generation, and simulation building.
  • 19
    Claude Sonnet 4.6
    Claude Sonnet 4.6 is Anthropic’s most advanced Sonnet model to date, delivering significant upgrades across coding, computer use, long-context reasoning, agent planning, and knowledge work. It introduces a 1 million token context window in beta, allowing users to analyze entire codebases, lengthy contracts, or large research collections in a single session. The model demonstrates major improvements in instruction following, consistency, and reduced hallucinations compared to previous Sonnet versions. In developer testing, users strongly preferred Sonnet 4.6 over Sonnet 4.5 and even favored it over Opus 4.5 in many coding scenarios. Its enhanced computer-use capabilities enable it to interact with real software interfaces similarly to a human, improving automation for legacy systems without APIs. Sonnet 4.6 also performs strongly on major benchmarks, approaching Opus-level intelligence at a more accessible price point.
  • Previous
  • You're on page 1
  • Next