Nolan Carrow's avatar

Google Antigravity IDE first impressions & survival guide

By Nolan Carrow 0 comments • less than a minute ago

I’ve been reading docs, codelabs, blog posts, security write-ups, forum threads, X, and Reddit about Google Antigravity. Here’s a dev-focused first look plus the main gotchas.

  1. What Antigravity actually is

Antigravity is Google’s free, agent-first IDE built on a VS Code fork.

You get:

  • A normal Editor (VS Code style) with autocomplete and inline commands
  • A Manager view where you run agents that can touch:
    • the editor
    • a terminal
    • an embedded browser
  • Agents emit Artifacts: plans, task lists, diffs, screenshots, recordings

You describe tasks (“build a dashboard”, “refactor auth”), agents do multi-step work, and you review artifacts instead of digging through logs.

Short version: AI IDE + async agents + browser automation + artifact review, free public preview on macOS, Windows, and Linux. 2. Install & first-run: choices that matter

Useful entry points:

  • Site: https://antigravity.google
  • Docs: https://antigravity.google/docs
  • Official intro on the Google Developers Blog
  • “Getting Started” Codelab

Things worth deciding up front:

  1. Development mode

    On first launch you pick:

    • Agent-driven – most autonomy
    • Agent-assisted – agent works but checks in
    • Review-driven – you approve almost everything

    Default to Agent-assisted. You get real work done but still see key checkpoints.

  2. Review policy

    Common policies:

    • Always proceed – don’t use this at the start
    • Agent decides – fine once you trust it
    • Request review – safest on day one

    Start with Request review or Agent decides plus a careful eye.

  3. Separate workspaces

    Use a dedicated sandbox folder (e.g. antigravity-playground) while you learn it. Add real repos later.

  4. Install the browser extension early

    The browser agent needs the Chrome extension to click around, run UI checks, and capture screenshots / video. The codelab walks you through it.

  5. Mental model: not “just VS Code”

The high-level model:

  • Editor View
    Normal coding: files, inline completions, commands.

  • Manager View
    “Mission control”:

    • Start tasks / conversations
    • See an Inbox of jobs (Idle / Running / Blocked)
    • Switch between Planning and Fast
    • Review Artifacts
  • Planning vs Fast

    • Planning – more reasoning, task groups, richer artifacts; good for bigger work
    • Fast – fewer steps; good for small edits and quick shell jobs
  • Browser Agent

    Sub-agent that drives the embedded browser:

    • clicks, scrolls, forms
    • reads DOM
    • captures screenshots and recordings

Mental shortcut: Editor = hands, Manager = brain + task board, Browser agent = QA intern with a mouse. 4. What people are actually doing with it

From tutorials and early reviews, common real-world uses:

  • Spin up small apps

    • Next.js + Tailwind + shadcn dashboards
    • Simple games (e.g. snake in Pygame)
  • Multi-file refactors

    • “Refactor auth module”
    • “Generate tests for billing API”
    • “Update the dependency tree”
  • UI / landing pages

    • Generate a few variants, pick one, let it build the implementation plan and code.
  • Browser-driven testing

    • Run flows in the embedded browser
    • Attach screenshots for failing states

If you want a quick feel:

  • Build a simple game (like snake)
  • Build a basic dashboard app

These cover Manager View, Artifacts, and the browser agent in one go. 5. Patterns that seem to work

Pulled from codelabs, blog posts, and dev threads.

Prompting

  • Treat the agent like a senior contractor, not autocomplete:

    • goal
    • constraints
    • success criteria
  • Ask it to plan before touching files:

    “Create an implementation plan artifact first, wait for my review.”

Task breakdown

  • Use multiple conversations instead of one giant prompt:
    • “Set up project skeleton”
    • “Implement feature X”
    • “Add tests and browser checks”
  • Run several agents in parallel and use the Inbox to hop between them.

Use Artifacts like PRs

  • Skim plans, checklists, and walkthroughs as if they were PR descriptions:
    • adjust architecture early
    • only then let it execute
  • Stick with Agent decides or Request review until you understand its habits.

Model / mode choices

  • Use higher-quality models for complex, multi-step tasks.
  • Use cheaper / faster models for small edits.
  • If a config keeps failing or hits quotas, switch to a smaller model and narrow the prompt.
  1. Rough edges and “this is buggy” reports

This is clearly still a preview. A few recurring complaints:

Code corruption / replace issues

Some users report Antigravity:

  • corrupting files while “fixing” them
  • looping on bad edits
  • messing up large replace operations

Mitigations:

  • Commit to git before big runs
  • Ask for diffs / patches instead of blind rewrites
  • Be strict with plan review; reject weird plans early

Performance

Common issues:

  • Heavy RAM and CPU usage
  • Sluggish after long sessions
  • Battery drain on laptops

Treat it as a sidecar IDE, not your only editor.

Rate limits / quotas

  • Free preview, but quotas on the higher-tier models are tight.
  • You can hit limits fast if you poke it constantly.

Practical mindset:

  • Use fewer, better-defined tasks
  • Let runs finish instead of over-steering
  • Fall back to cheaper models when you hit limits
  1. Security: the big red flag

Security research (including Mindgard and others) points to serious risks.

Key problems:

  1. Prompt-based exfiltration

    • Malicious repo content can instruct the agent to read secrets (like .env)
    • It can encode them into URLs or artifacts that get exfiltrated
    • The agent may read files outside your repo and ignore .gitignore
  2. Persistent code execution via global config

    • Project-level config can cause writes to ~/.gemini/antigravity
    • That can lead to arbitrary code execution on future launches
    • Workspace access settings may not fully protect against this

Practical rules for now:

  • Do not open untrusted repos in Antigravity
  • Treat .agent / rules / workflows as code, not comments
  • Avoid Always proceed; keep human review in the loop
  • Keep secrets out of the workspace tree where possible
  • Use separate machines / VMs for sensitive work

If you work with production or regulated data: keep Antigravity in playground / prototype territory until the security story improves. 8. How it compares (Cursor & Claude Code)

High-level comparison:

  • Cursor

    • VS Code-like AI IDE
    • Great line-level help and small tasks
    • Low friction as a daily driver
  • Claude Code

    • Terminal-first coding agent
    • Strong reasoning and multi-file work
    • You mostly live in the repo + shell
  • Antigravity

    • “Max autonomy” agent IDE
    • Strong task-level orchestration across editor + terminal + browser
    • Weak on polish, stability, and security (for now)

Rough analogy:

Cursor = main editor, Claude Code = deep-dive buddy, Antigravity = experimental control room where you let robots work and read their reports. Just don’t point it at your production monolith yet.

  1. Simple experiments to try

If you want a structured test drive:

  1. Toy app session

    • New folder
    • Small game or playground app
    • Force yourself to approve plans and diffs like PRs.
  2. Refactor sprint

    • Clone a non-critical side project
    • Ask it to:
      • introduce a design system
      • extract a shared layout
      • add tests for one subsystem
    • Keep everything behind “implementation plan + diff review”.
  3. Browser agent as QA helper

    • Have it build a tiny full-stack app
    • Prompt it to run auth / form flows and attach screenshots for failures
    • Decide if you’d trust that style of report in real work.
  4. References

  1. Questions for other users

If you’ve used Antigravity for more than a weekend:

  • What has it actually shipped for you?
  • Any prompt patterns / review policies that made it much more reliable?
  • Have you found a defensive setup you’d trust on semi-sensitive repos?
  • How do you split work between Cursor / Claude Code / Antigravity day to day?

Real stories, workflows, and failure modes are the most useful part IMO.

Sign in to join the discussion.

No comments yet. Be the first to reply!