unsubbed.co

Browser Use

Make websites accessible for AI agents — automate browsing, extraction, testing, and monitoring in natural language with Playwright and LLMs.

Open-source browser automation, honestly reviewed. What you actually get when you let an LLM loose on Chrome.

TL;DR

  • What it is: Open-source (MIT) Python library that connects LLMs to a real browser via Chrome DevTools Protocol — you describe a task in plain English, the agent executes it in a browser [3][4].
  • Who it’s for: AI engineers and Python developers building agents that need to interact with websites that don’t have APIs. Not for non-technical founders operating solo [3][5].
  • Cost reality: The library is free, but you still pay for LLM API calls — and for production reliability you’ll likely pay for cloud infra too. Self-hosting is viable but not free in practice [3][5].
  • Key strength: 81,136 GitHub stars, MIT license, works with any major LLM (GPT-4o, Claude, Gemini), and handles real-world browser tasks that brittle Selenium scripts cannot [merged profile][4].
  • Key weakness: It’s a developer tool, full stop. Requires Python, LLM API keys, and non-trivial infra setup. The self-hosted version measurably underperforms the cloud version on complex tasks, by their own benchmark [README].

What is Browser Use

Browser Use is a Python library that gives an LLM control over a real Chrome browser. You write a short task in natural language, and the agent — backed by whichever LLM you plug in — clicks, types, scrolls, fills forms, and extracts data without you writing a single CSS selector [3][4].

The pitch is a direct rebuttal to Selenium and Playwright: stop maintaining brittle scripts that break when a div ID changes, and let the AI figure out the UI. Instead of driver.find_element(By.ID, "submit-btn").click(), you write "Fill in the contact form and submit it" [4].

Under the hood it sits on top of Chrome DevTools Protocol (CDP) — the same low-level interface Playwright and Puppeteer use [3]. The innovation isn’t the browser layer; it’s the LLM loop on top. The agent takes a screenshot, reasons about what it sees, decides on the next action (click, type, scroll, navigate), executes it, and repeats until the task is done or it gives up.

The project hit 81,136 GitHub stars as of this writing, making it one of the fastest-growing developer tools of 2024–2025. The company has also shipped a managed cloud service (browser-use.com) that bundles stealth browsers, proxy rotation, captcha solving, and a purpose-built LLM — which is a meaningful upgrade over the open-source base [README][pricing page].


Why people choose it over Selenium, Playwright, and Operator

The reviews converge on one core argument: traditional automation is a maintenance tax, and Browser Use is an attempt to eliminate it [3][4].

Versus Selenium/Playwright. The comparison the Gologin review [3] makes is the most concrete: traditional tools require you to hunt for selectors by hand and rebuild scripts each time the page layout shifts. Browser Use changes this by combining LLM reasoning with real browser control — you describe the goal, not the steps. The same task that would require a 200-line Playwright script (“find the third job posting, check if it’s remote, click apply, fill the form…”) becomes a single English sentence. The tradeoff is reliability: LLM-based agents are non-deterministic and can hallucinate steps. Playwright is brittle but predictable; Browser Use is flexible but occasionally baffling.

Versus OpenAI Operator / Google Project Mariner. Both are closed-source enterprise products [1][2]. Operator requires human confirmation for significant actions and is locked to ChatGPT Pro. Project Mariner runs in Chrome but isn’t self-hostable. Browser Use is the MIT-licensed, run-on-your-own-infra alternative that lets you plug in any LLM and extend it with custom Python functions [3][5].

Versus Browserbase. Browserbase is infra, not AI — it provides managed headless browsers compatible with Selenium/Playwright but doesn’t add LLM reasoning [5]. Browser Use is the AI layer. They’re not quite competitors; some teams use Browser Use on top of Browserbase infrastructure. The Skyvern comparison [5] frames it clearly: Browserbase = managed browser hosting, Browser Use = the agent that drives them. Skyvern itself positions as the product that combines both, which is a fair characterization of the gap.

The honest caveat from their own data. The Browser Use README publishes a benchmark of 100 real-world browser tasks comparing open-source vs. cloud performance. The cloud version — using their purpose-built BU model and stealth infrastructure — outperforms the open-source setup on complex tasks. They publish this graph themselves [README]. If you self-host with a general-purpose LLM, you’re accepting a reliability ceiling that their own cloud clears. For simple scraping tasks the gap is small; for complex multi-step workflows it matters.


Features: what it actually does

Core agent loop:

  • Natural language task execution via any supported LLM [3][4]
  • Screenshot-based reasoning — the agent sees the page like a user would [3]
  • Click, type, scroll, navigate, extract, submit [3]
  • Multi-tab management [5]
  • Session persistence — cookies and auth states can be reused across runs [3]
  • Custom Python tool integration — extend the agent with your own functions that mix browser actions with data processing [3]

LLM support:

  • OpenAI GPT-4o via ChatBrowserUse() [README]
  • Google Gemini via ChatGoogle() [README]
  • Anthropic Claude via ChatAnthropic() [README]
  • Any LLM provider you can wire up through the Python API

Recent additions:

  • Workflow Use — a deterministic, self-healing automation layer for production workloads that need reliability over flexibility. This is Browser Use’s answer to the “LLMs are unpredictable” criticism for high-frequency enterprise tasks [3].
  • Skills API — record a workflow once, get a callable API endpoint from it. The cloud pricing charges $2.00 per skill creation and $0.02 per API call [pricing page]. The self-hosted version doesn’t include this.
  • Benchmark suite — open-sourced at browser-use/benchmark, 100 real-world task evaluations [README].

What it doesn’t do out of the box:

  • Visual builder or UI — this is code-first, always [3][4]
  • Built-in proxy rotation or captcha solving (cloud handles this; self-hosted doesn’t) [README]
  • Support for non-Python runtimes — JavaScript/TypeScript developers are not the target audience here

Pricing: cloud vs self-hosted math

Self-hosted (open source):

  • Browser Use library: $0 (MIT) [merged profile]
  • LLM API costs: real money — Claude Sonnet on the cloud tier is priced at ~$3.60/$18.00 per 1M tokens [pricing page]; self-hosted users pay their own API rates directly to Anthropic/OpenAI/Google
  • Browser infrastructure: your own machine or VPS; add $5–20/mo for a VPS if you’re running headless Chrome at any scale
  • Stealth/proxy/captcha: not included — you set this up yourself or go without

Browser Use Cloud:

  • Pay-as-you-go: $100 in credits, 25 concurrent sessions, basic stealth mode [pricing page]
  • Subscription: $75/mo, $100 in credits (save 25%), 100 concurrent sessions, advanced stealth, priority support [pricing page]
  • Session cost: $0.06/hour of browser time, billed per minute with unused time refunded [pricing page]
  • LLM cost per step: from $0.002 [pricing page]
  • Enterprise: custom SLAs, data retention, on-prem deployment [pricing page]

Concrete math for a typical agent workflow:

Say you’re running a daily data collection job — 50 tasks/day, each taking ~5 minutes of browser time, using Claude Sonnet with ~500 tokens per step and ~20 steps per task.

  • Browser time: 50 × (5/60) hours × $0.06 = $0.25/day = ~$7.50/month
  • LLM cost at $0.002/step × 20 steps × 50 tasks = $2/day = ~$60/month
  • Cloud total: ~$67–75/month on subscription tier

Self-hosted alternative: $0 for the library + direct API costs (Anthropic charges ~$3/1M input tokens for Sonnet — your actual spend depends on your usage pattern). At this volume, self-hosted LLM costs would be roughly similar, but you avoid the session fee and gain flexibility at the cost of stealth/captcha handling.

For scraping jobs that need residential proxies and captcha solving, the cloud math often wins despite the session fee. For internal automation that doesn’t face anti-bot measures, self-hosted is cheaper.

No free-tier for production work. The open-source library is free, but any non-trivial agent workflow has real LLM costs attached. There is no Zapier free tier equivalent here.


Deployment reality check

The Gologin review [3] is the most honest account of what setup actually looks like.

For a developer:

uv init && uv add browser-use && uv sync
# uvx browser-use install  # if Chromium isn't installed

Set your LLM API key, write 15 lines of Python, run it. Genuinely fast for a developer. Chromium is bundled or installable in one command [README].

For production:

This is where it gets harder. You need:

  • A Python 3.11+ environment [README]
  • Chrome/Chromium installed
  • LLM API credentials (and a budget for them)
  • Session state management if you’re reusing authenticated sessions [3]
  • Anti-bot handling if your targets block headless browsers — rolling your own proxy rotation is non-trivial [3][5]
  • Monitoring for agent failures — LLM-based agents fail silently in ways Selenium doesn’t

The XDA/Gologin review notes that the cloud version handles stealth and proxies transparently while the open-source version leaves all of that to you [3]. For scraping targets with any anti-bot protection (Cloudflare, shape.security, DataDome), self-hosted hits a wall fast.

Who should not self-host: Non-technical founders who want to automate tasks themselves. This is a Python developer tool. There is no GUI, no drag-and-drop interface, and no way to use it without writing code. The “Fortune 500 trusted” claim on the homepage [website] refers to enterprise teams with engineers, not solo operators.


Pros and cons

Pros

  • Actually MIT-licensed. You can embed it in your product, modify it, build a business on it — no lawyer needed [merged profile].
  • 81K+ GitHub stars. The momentum is real. Active development, fast issue resolution, large community [merged profile].
  • LLM-agnostic. Works with GPT-4o, Claude, Gemini, or any model you can wire through the API. Not locked to one provider [README][3].
  • Handles what scripts can’t. Dynamic pages, auth flows, multi-step interactions that change with page state — the LLM reasons about what it sees rather than matching a selector [3][4].
  • Workflow Use for determinism. The newer deterministic layer addresses the “LLMs are unreliable in production” criticism directly [3].
  • Active benchmark. They publish a real benchmark with 100 tasks and model comparisons, open-sourced. That’s more transparency than most tools in this space [README].

Cons

  • Python-only. No JavaScript/TypeScript SDK. If your stack is Next.js or Node.js, you’re writing a Python microservice or using the cloud API [README][3].
  • Self-hosted underperforms cloud on complex tasks — their own benchmark says so [README]. You pay for reliability when you go cloud.
  • LLM costs are the real cost. The library being free is misleading. Any serious usage has LLM API bills. Budget accordingly [3][5].
  • Non-deterministic failures. The agent can misinterpret a page, take the wrong action, get stuck in a loop, or hallucinate a successful completion. This requires monitoring infrastructure [3][4].
  • No stealth out of the box. Anti-bot detection is a real problem for scraping use cases, and the open-source version doesn’t help with it [3][5].
  • Requires technical setup for production. Not plug-and-play. Session handling, error recovery, retry logic — all yours to build [3].
  • “Trusted by Fortune 500” marketing oversells the self-hosted story. Those teams are using the cloud product with their engineering staff, not deploying the Python library on a $5 VPS [website].

Who should use this / who shouldn’t

Use Browser Use if:

  • You’re a Python developer building an AI agent that needs to interact with websites — competitors, job boards, internal tools, portals without APIs.
  • You want MIT-licensed browser automation you can embed in a commercial product.
  • You’re doing research scraping, lead gen automation, or QA testing where traditional scripts keep breaking.
  • You have an LLM API budget and want control over which model you use.

Use Browser Use Cloud if:

  • You need production reliability, stealth browsers, captcha solving, and proxy rotation without managing that yourself.
  • Your automation hits sites with anti-bot protection.
  • You want the best possible agent performance (their benchmark shows measurable gains) [README].

Skip it (stay with Playwright/Selenium) if:

  • Your automation target is stable and predictable — no need to pay LLM costs for tasks a deterministic script handles fine.
  • You need reproducible, bit-for-bit identical behavior across runs.
  • Your team doesn’t write Python.

Skip it (use Skyvern or similar) if:

  • You want an AI-native browser agent bundled with managed infrastructure, without stitching together Browser Use + Browserbase yourself [5].

Skip it entirely if:

  • You’re a non-technical founder hoping to automate your own workflows through a UI. This tool has no UI. It’s code.

Alternatives worth considering

  • Skyvern — combines AI-native browser automation with managed cloud infrastructure. More complete than Browser Use alone; less control over the underlying model [5].
  • Browserbase — managed headless browser infrastructure without the AI layer. Use it as the infra layer under Browser Use or traditional frameworks [5].
  • Playwright/Selenium — still the right choice for deterministic, selector-based automation on stable targets. Cheaper per run and predictable.
  • OpenAI Operator — closed-source enterprise browser agent, no self-hosting, requires ChatGPT Pro. Enterprise alternative [1][2].
  • Browser MCP — Model Context Protocol server for browser control; different architecture, interesting for Claude-centric workflows [3].
  • Stagehand — TypeScript-native alternative to Browser Use for Node.js teams.

Bottom line

Browser Use is the real thing for what it claims to be: an MIT-licensed, LLM-driven browser automation library with serious community traction. The 81K+ stars aren’t astroturfed — it solves a genuine problem (brittle browser scripts) with a genuinely different approach (LLM reasoning over screenshots). For an AI engineer building agents that need to touch the web, it’s the first thing to evaluate.

What it isn’t: a tool for non-technical founders, a free automation platform, or a self-hosted replacement for something like Browserbase. The library is free; the LLM API calls aren’t. Production-grade usage with stealth and reliability requirements will push you toward the cloud product or require significant engineering investment to replicate it yourself. If the engineering investment is the blocker, that’s precisely what unsubbed.co’s parent studio upready.dev helps with — deploy once, own the infrastructure, skip the recurring per-seat SaaS bill.


Sources

  1. O-Mega.ai“Top 10 Browser Automation Agents for Web Task Automation 2024”. https://o-mega.ai/articles/the-top-10-browser-automation-agents
  2. dev.to (Abhishek)“Top 10 AI Web Agents in 2025 — Ranked by Usage & Popularity (Free & Paid)”. https://dev.to/abhishek-nexgen-dev/top-10-ai-web-agents-in-2025-ranked-by-usage-popularity-free-paid-2nhe
  3. GoLogin Blog (Anthony Heath)“Browser Use: Technical Expert Review & Tests” (Nov 25, 2025). https://gologin.com/blog/browser-use-technical-expert-review-tests/
  4. Skywork.ai“Browser Use: An In-Depth Review and Guide for the AI-Powered Web” (Oct 6, 2025). https://skywork.ai/skypage/en/Browser-Use-An-In-Depth-Review-and-Guide-for-the-AI-Powered-Web/1972881756238442496
  5. Skyvern Blog“Browser Use vs Browserbase: Comparison, Reviews, and Alternatives”. https://www.skyvern.com/blog/browser-use-vs-browserbase-comparison-reviews-and-alternatives/

Primary sources: