Record your E2E tests in one chat. Replay them 100× cheaper.
Claude MCP test automation for Playwright-grade E2E. Use Claude where it shines — designing the test once. Replay from saved snapshots: far fewer LLM tokens, faster runs, full history and video reports. Self-healing on broken Playwright and Cypress selectors.
Computer Use cost = LLM tokens at ~$3.20 per 100-step replay. E2Easy LLM cost = $0.04 per authoring + occasional self-heal (~$0.01). CI compute on top: $0 local · ~$0.02/run in our cloud.
The toolset Claude gets
A focused MCP API. No prompt-engineered hacks.
10 deliberate verbs. Claude composes them into anything from one-liner to full suite.
inspect_page(url)→ Element[]// real browser, interactive elements
create_test(name, steps[])→ Test// save a new E2E test
list_tests(filter?)→ Test[]// search workspace
get_test_steps(test_id)→ Step[]// read step list
replace_test_steps(test_id, steps[])→ Test// patch one step or rewrite
update_test(test_id, patch)→ Test// rename, move, tag
delete_test(test_id)→ void// destructive — confirms first
run_test(test_id, env?)→ RunId// dispatch to runner
list_test_runs(test_id?)→ Run[]// run history
OAuth 2.1 + PKCE. Short-lived tokens. Revocable.
Destructive verbs require Claude to confirm first.
Side by side
The same intelligence, without the runtime tax.
Computer Use reasons. We make reasoning durable, deterministic, replayable.
Claude Computer Use
Agent drives the browser
Brilliant at reasoning
Pays LLM tokens every run
Probabilistic, slow per step
Cypress / Playwright
Engineer writes the code
Deterministic, fast
Hours of code per flow
Manual fixes on UI changes
This is the oneE2Easy + Claude
Author once, replay anywhere
One chat — saved as a recorded test
Replay locally, on your server, or in our cloud
100× fewer tokens per replay
Full history, video and run reports
Case study · Forge.io
From 4 hours per sprint to 22 minutes.
How a 12-engineer team at Forge.io ran their entire E2E maintenance off Claude MCP — and cut LLM spend 87× in the process.
E2E suite size
47 tests · ~2,800 steps
Replays / month
~9,400
Stack before
Cypress + manual fixes
Stack now
Claude Desktop + E2Easy MCP + Playwright
The problem
Every UI refactor broke 3-7 selectors. Two engineers on rotation spent ~4 hours per sprint hunting and patching them. Bug-fixing PRs queued behind selector PRs. They'd evaluated Claude Computer Use but the per-replay LLM bill was a non-starter for nightly + per-PR runs.
The flip
They switched in one afternoon: connected Claude Desktop to E2Easy via MCP, re-recorded the broken tests in chat, then pushed all 47 to the platform. The team kept Cypress for the unit-level UI tests and moved the cross-page flows to E2Easy.
The numbers · 90 days in
22 minselector maintenance per sprint (was 4 hr)
$3.10total LLM spend / month (Computer Use estimate: $270)
11 → 0selector-fix PRs per sprint blocking releases
100%runs with video + step logs available in dashboard
A selector breaks, Claude patches that one step. The rest of the suite keeps running. Self-healing without re-prompting the whole flow — and the run history actually means something now, because every replay is the same path.
Anthropic's Model Context Protocol — the open standard for plugging LLMs into real tools. E2Easy's MCP server gives Claude native verbs for recording, inspecting and replaying browser tests.
Does Claude run my tests every time?
No. Claude authors once. Playwright replays from the saved snapshot — locally, in CI, or in our cloud. Self-healing fires only on the broken step.
Is this the same as Claude Computer Use?
No. Computer Use calls the LLM on every step. E2Easy uses Claude to author the test once from the live DOM, then replays deterministically from snapshots — far fewer tokens per replay, with full video reports.
Do I need an Anthropic API key?
No. Bring your own Claude account (Desktop, Claude.ai, Cursor — anything MCP-compatible). OAuth-based, no static keys.
Will my tests still work without Claude?
Yes. Tests live on Playwright. Claude is the author, not a runtime dependency.
Can other MCP-compatible clients use this?
Yes. The E2Easy MCP server speaks standard MCP over OAuth. Claude Desktop, Cursor, Continue.dev, Cline and similar clients all connect with the same endpoint.
How does Claude know which selectors to use?
It doesn't guess. inspect_page and find_element visit the page in a real Playwright browser and return stable selectors, ARIA, xpath, locators — straight from the DOM.
How is this different from Microsoft Playwright MCP?
playwright-mcp is a raw browser bridge — every replay calls the LLM on each step. E2Easy stores tests as Playwright-grade recordings: replay is deterministic with run history and video reports, and self-healing fires only on the failing step.
Can I use this to fix broken Cypress selectors with Claude?
Yes. CSS, ARIA, xpath, role-based locators — all supported. After a UI refactor Claude reads the live DOM via MCP and produces a drop-in selector fix. Only the failing step is patched.
Does this count as AI test automation or self-healing test automation?
Both. AI test automation for authoring, self-healing test automation for maintenance. Claude patches only the failing step.
What about security?
OAuth 2.1 + PKCE. Short-lived tokens. Revocable any time. Test data stays in your workspace.
Use Claude to design your tests. Not to run them.
Two minutes to connect via MCP. Author once. Replay 100× cheaper, fully logged, with video reports.