TL;DR: Cursor Background Agents (also called Bug Bot) are cloud AI agents that work on coding tasks while you're away. You describe a task, they clone your repo into a VM, do the work, and open a PR. You review it, merge it, done. Think of them as tireless junior devs you can clone and run in parallel — on your schedule, not theirs.
Why AI Coders Need to Know This
Here's something that probably sounds familiar: you're in the middle of building something, and you realize there are three other things that need doing before this feature is complete. You write a list. Then you pick one, work on it, and the other two just sit there — waiting for you to have time, energy, and context to switch.
That's the bottleneck Background Agents are designed to break.
Up until now, AI coding tools have been like having a really smart co-pilot in the passenger seat — useful, fast, but they need you driving. You prompt, they respond. You review, you iterate. The whole thing runs at the pace of your attention.
Background Agents flip that model. Instead of working with you in real time, they work for you in the background. You hand off a task, they disappear into a cloud VM, and when they're done — could be ten minutes, could be an hour — they surface with a pull request for you to review.
For a vibe coder, this is massive. It means the limit on how fast you can ship is no longer just your available hours. You can have multiple things in flight at once. You can assign work before you go to bed and wake up to completed code. You can context-switch freely, because the AI doesn't need you present to make progress.
This is what agentic coding looks like in practice — not a chatbot, but something closer to a team member with a task list.
What Background Agents Actually Do
Let's make this concrete, because "AI agent in the cloud" can sound vague until you see the actual mechanics.
When you start a Background Agent in Cursor, here's roughly what happens:
- You describe the task — in plain English, in Cursor's interface. "Fix the bug where the login form shows an error even when credentials are correct." Or: "Write unit tests for the functions in utils/payments.js." Or: "Add a dark mode toggle that saves the user's preference to localStorage."
- Cursor spins up a cloud VM — a virtual machine running in Cursor's infrastructure. It's isolated from your local machine and has its own compute, its own filesystem, its own environment.
- The VM clones your repository — it pulls down your code from GitHub or GitLab (you need one of these connected). It checks out a new branch specifically for this task.
- The agent works through the task — it reads your codebase, figures out what needs to change, writes the code, and (if your project has tests) runs them to verify the changes don't break anything.
- It opens a pull request — when the agent is satisfied with its work, it pushes the branch and opens a PR on your connected repository. The PR description explains what it changed and why.
- You review and merge — you come back whenever you're ready, look at the diff, decide if it looks right, and either merge it or leave feedback for another pass.
The key word is asynchronous. None of those steps require your attention while they're happening. You set the task and go do something else — work on a different feature, take a meeting, eat lunch, go to sleep.
The Bug Bot Name Internally, Cursor's team called this feature "Bug Bot" during development — a name that stuck informally. The public-facing name is Background Agents, but you'll see both terms used in Cursor's docs and community. They refer to the same thing.
Background Agents have access to a full development environment inside that VM. They can read files, create files, edit files across your entire codebase, run your build system, execute test suites, and install packages if needed. They're not limited to small single-file edits — they can make sweeping changes across a large project if the task requires it.
What they can't do (as of early 2026): make decisions that require visual judgment (they can't look at your UI and tell you if it's pretty), access external services that require authentication they don't have (your API keys, your production database), or handle tasks so open-ended that there's no clear definition of "done."
How to Use Background Agents (Step by Step)
Before you start, you need two things set up:
- Cursor Pro — Background Agents are not available on the free tier. You need a paid subscription.
- GitHub or GitLab connected — Agents need to clone your repo and open PRs. Go to Cursor Settings → Integrations and connect your account if you haven't already.
Once those are in place:
- Open the Background Agents panel — In Cursor, look for the Background Agents section. As of 2026, you can access it from the sidebar or via the command palette. The exact UI location has shifted across versions — if you don't see it, check Cursor's documentation for the current location in your version.
- Write a clear task description — This is the most important step. The agent only knows what you tell it. Bad: "fix the thing." Good: "The signup form at /auth/register throws a 500 error when the email field contains a plus sign (e.g., user+tag@example.com). Fix the email validation in the backend so plus signs are handled correctly. The validation logic is in server/auth/validate.js." The more specific, the better the result.
-
Choose your branch target — Decide which branch the agent should branch off from. Usually
mainordevelop, but you can point it at a feature branch if needed. - Start the agent — Hit run. Cursor will show the agent as "running" in your panel. You'll see it's working, but you don't need to watch.
- Go do something else — Seriously. That's the point. Start another task, go to a meeting, close your laptop. The agent is working in the cloud — your machine doesn't need to be on or connected.
- Get notified when the PR is ready — Cursor will notify you when the agent opens its PR. You'll also get the standard GitHub/GitLab PR notification if those are set up.
- Review the pull request — Open the PR like any other. Read the description the agent wrote, look at the diff, check if it makes sense. Run the branch locally if the change is significant. Merge if it's good. Leave a comment and re-run the agent if something needs adjusting.
Review Every PR Background Agents are capable but not infallible. Always review the diff before merging. Agents can misunderstand scope, introduce unrelated changes, or solve the symptom instead of the cause. A fast review habit protects you from shipping agent mistakes. Never auto-merge agent PRs without eyes on the code.
Tips for Getting Better Results
- Include the file path when you know which file needs changing. "The function is in
src/utils/dates.js" is much better than "somewhere in the utils folder." - Describe the expected behavior, not just the problem. "It should return null instead of throwing an error when the date string is empty" gives the agent a target.
- Reference existing patterns when you want consistency. "Add a new endpoint following the same pattern as
api/users.js" helps the agent match your existing style. - Tell the agent about your tests. If you have a test suite, mention it: "Run
npm testafter making changes to verify nothing breaks." - Start small your first few times. Single-file bugs or adding one function are good early tasks. Get a feel for what the agent produces before throwing complex multi-file features at it.
What They're Good At vs What They Struggle With
Background Agents are powerful, but they're not magic. Knowing where they excel — and where they fall flat — saves you time and frustration.
Background Agents Shine At:
- Well-defined bug fixes — "This function returns the wrong value when the input is an empty array" is a perfect agent task. Clear problem, clear expected behavior, bounded scope.
- Writing tests — Agents are excellent at reading existing code and generating a test suite for it. Point them at a file, tell them what test framework you're using, and they'll produce solid coverage.
- Repetitive refactoring — Renaming things consistently across 30 files, updating deprecated API calls throughout a codebase, converting a pattern from one style to another. Work that's tedious for humans but systematic for agents.
- Adding endpoints that follow existing patterns — If your API has ten endpoints and you need an eleventh that works like the others, agents can read the patterns and reproduce them accurately.
- Updating dependencies and fixing downstream breakage — "Update lodash to v5, fix any breaking changes" is a good agent task. Bounded, verifiable, tedious.
- Documentation generation — Writing JSDoc comments, README sections, or inline documentation for existing functions.
Background Agents Struggle With:
- Open-ended creative tasks — "Build me a dashboard" without specific requirements leads to something generic that doesn't match your vision. Agents need a clear definition of done.
- Anything visual — They can write CSS, but they can't look at your site and judge if it looks right. Visual iteration still needs you in the loop.
- Tasks requiring external credentials — If the fix requires hitting your production database, calling a third-party API with your keys, or reading from a private service, the agent can't do it from a VM it doesn't have access to.
- Architectural decisions — "Should we use Redux or Zustand for state management?" is not an agent task. Agents implement — they don't architect. Make the big decisions yourself, then hand off implementation.
- Long chains of dependent tasks — Agents do one task well. If the task requires "first do X, then based on what X produces do Y, then evaluate if Y needs Z," the dependencies create risk. Break these into separate agent runs with human checkpoints between them.
- Debugging production issues with live data — They can fix a reproducible bug. They can't investigate a flapping production issue that only appears under certain load conditions they can't replicate.
Background Agents vs Regular Cursor vs Claude Code
If you already use Cursor or have looked at other AI coding tools, here's how Background Agents fit into the picture:
| Feature | Cursor (inline) | Cursor Background Agents | Claude Code |
|---|---|---|---|
| Where it runs | Your machine, your editor | Cloud VM, isolated | Your machine, terminal |
| Requires you present? | Yes | No | Yes (mostly) |
| Output format | Inline diff, you accept | Pull request on GitHub/GitLab | Inline edits, commits |
| Parallelism | One task at a time | Multiple agents simultaneously | One session at a time |
| Best task size | Single feature or file edit | Bounded, well-defined tasks | Complex multi-file work |
| Needs Git integration | No | Yes (required) | Optional |
| Review step | In-editor diff | PR review | In-terminal diff |
| Learning curve | Low | Low-medium | Medium |
The practical summary: inline Cursor is for when you're actively building and want AI assistance in real time. Background Agents are for well-defined work you can delegate and check on later. Claude Code is for complex reasoning tasks, large codebase refactors, or when you want a powerful terminal-native agent with deep context.
Most experienced vibe coders end up using all three depending on what they're doing. They're not competing tools — they're different modes of the same shift toward AI-assisted development. Read How to Choose an AI Coding Tool for a fuller breakdown of when each approach makes sense.
The Real Workflow A common pattern: use inline Cursor to build a new feature, then hand the completed feature off to a Background Agent with the instruction "write a full test suite for what I just built." You keep moving. The agent handles the test coverage. When you check back, you have tests ready to review.
The Pricing Question
Background Agents cost more than standard Cursor use, and it's worth understanding why before you start spinning up agents for every task.
When you use inline Cursor — Tab autocomplete, Composer mode, chat — the compute is relatively lightweight. The model responds to your prompts and your session uses your plan's included credits.
Background Agents are heavier. Here's what's happening behind the scenes:
- A cloud VM is provisioned and run for the duration of the task
- The VM clones your full repository
- Multiple model calls happen as the agent plans, writes, and verifies
- If the agent runs tests, it's executing real compute
- All of this runs for however long the task takes — potentially 30-60+ minutes for complex work
That compute costs money, and Cursor passes some of that cost to users via usage credits that burn faster for agent runs than for standard interactions.
Verify Current Pricing Cursor's pricing model for Background Agents was still evolving as of early 2026 — both the cost per agent run and what's included in each plan tier. Before using them heavily, check the current pricing at cursor.com/pricing. What we describe here is the general structure; exact numbers may have changed.
Here's how to think about the cost-benefit math:
- If a task would take you 45 minutes of focused work, and an agent does it in 30 minutes while you're doing something else, the cost is almost always worth it — even at $5-10 per agent run.
- If you're spinning up agents for tiny two-line fixes you could do in two minutes, you're burning credits wastefully. Save agents for tasks with real time value.
- The parallelism multiplier changes the calculus too. Running three agents in parallel — one fixing a bug, one writing tests, one updating docs — in the time it used to take you to do one thing is a meaningful productivity gain even at a higher per-unit cost.
Think of Background Agents the way Chuck, who came from 20 years in construction, thinks about renting a concrete mixer: it costs money to rent, but if you're pouring a foundation, it pays for itself in hours saved. Don't rent it to mix a small batch of mortar.
What AI Gets Wrong About Background Agents
There's a lot of hype around Background Agents — and some of that hype creates expectations that don't match reality. Here are the most common misconceptions worth clearing up:
1. "They Can Build Entire Features Unattended"
Not reliably, no. Background Agents are excellent at bounded, well-defined work. "Add a dark mode toggle" is doable. "Build me a complete user authentication system" is not — that requires architectural decisions, security review, multiple interconnected pieces, and iteration based on what you actually want. Agents work best when the scope is tight and the success criteria are clear.
2. "You Can Just Merge Without Reviewing"
This is how you ship bugs. Agents make mistakes. They might fix the reported issue but introduce a subtle edge case. They might add code in a style that doesn't match the rest of your project. They might solve the wrong problem because your description was slightly ambiguous. Always read the diff. The review step is not optional safety theater — it's a real defense against real errors.
3. "They Work Well for Tasks You Haven't Fully Thought Through"
Actually the opposite is true. Background Agents reward clarity and punish vagueness more than inline AI does. With inline Cursor, you can iterate interactively — "not quite, try again, actually no, let's approach this differently." With Background Agents, the agent is off doing work for 30 minutes based on your initial description. If that description was fuzzy, you get 30 minutes of work in the wrong direction. Think clearly before you assign.
4. "They're Only for Big Codebases"
Small projects benefit just as much, sometimes more. If you're a solo vibe coder with a side project and your backlog has twelve small fixes, Background Agents let you knock several of those out in parallel while you focus on the one that needs your full attention. Size of project is not the constraint — clarity of task is.
5. "They Replace Learning to Code"
You still need to understand what the agent is producing well enough to review it. If you can't read a diff and have a basic intuition for whether the changes make sense, you're flying blind. Background Agents amplify your capabilities — but they work best when you have enough context to evaluate their output. Keep building your understanding of the codebase even as you delegate more work.
FAQ
Cursor Background Agents are AI agents that run in cloud virtual machines (VMs) and work on coding tasks without you watching. You give them a task — fix this bug, add this feature, write these tests — and they go off independently, clone your repo, make changes, run tests, and open a pull request when done. You review the PR and merge if it looks good. Also called Bug Bot internally at Cursor.
Regular Cursor Composer runs inside your editor while you're sitting there watching it. Background Agents run in the cloud, on their own timeline, without you present. You can close your laptop and come back later. Composer is like working with a contractor in the same room. Background Agents are like sending specs to a contractor who works overnight and leaves the result on your desk. The review step is different too: Composer shows you inline diffs; Background Agents produce a full GitHub/GitLab pull request.
Background Agents use more compute than standard Cursor sessions because they spin up cloud VMs and run for extended periods. They consume your plan's usage credits at a higher rate than inline editing or chat. The exact pricing model has evolved — verify the current cost per agent run at cursor.com/pricing before using them heavily. Pro plan users have access to background agents but should monitor usage to avoid unexpected costs.
Background Agents shine on well-defined, bounded tasks: fixing a specific bug with a clear description, writing tests for existing functions, updating dependencies, adding a documented feature (like a new API endpoint following existing patterns), or doing repetitive refactoring across many files. They struggle with open-ended creative tasks, anything requiring visual judgment, tasks needing external credentials, or work where the requirements aren't fully specified upfront. Clarity of task is more important than size of project.
Yes — one of the key advantages of Background Agents is parallelism. You can assign different tasks to multiple agents running simultaneously: one fixing a bug, one writing tests, one updating documentation. Each agent works in its own isolated VM with its own branch. The main risk is conflicting changes if agents touch the same files. Plan task assignments so agents work in different parts of the codebase when running in parallel, and review all PRs before merging to catch any conflicts.
What to Learn Next
Background Agents are one piece of a larger shift toward AI doing more of the work independently. These articles round out the picture: