Free SKILL.md scraped from GitHub. Clone the repo or copy the file directly into your Claude Code skills directory.
npx versuz@latest install ultroncore-claude-skill-vault-skills-integrations-firecrawl-browser-versions-v1git clone https://github.com/UltronCore/claude-skill-vault.gitcp claude-skill-vault/SKILL.MD ~/.claude/skills/ultroncore-claude-skill-vault-skills-integrations-firecrawl-browser-versions-v1/SKILL.md--- name: firecrawl-browser description: | DEPRECATED — use scrape + interact instead. Interact lets you scrape a page and then click, fill forms, and navigate without managing sessions manually. Use this skill when the user needs to interact with a webpage, log into a site, click buttons, fill forms, navigate multi-step flows, handle pagination, or when regular scraping fails because content requires JavaScript interaction. Triggers on "click", "fill out the form", "log in to", "paginated", "infinite scroll", "interact with the page", or "scrape failed". allowed-tools: - Bash(firecrawl *) - Bash(npx firecrawl *) --- # firecrawl interact (formerly browser) > **The `browser` command is deprecated.** Use `scrape` + `interact` instead. Interact lets you scrape a page and then click, fill forms, and navigate without managing sessions manually. Interact with scraped pages in a live browser session. Scrape a page first, then use natural language prompts or code to click, fill forms, navigate, and extract data. ## When to use - Content requires interaction: clicks, form fills, pagination, login - `scrape` failed because content is behind JavaScript interaction - You need to navigate a multi-step flow - Last resort in the [workflow escalation pattern](firecrawl-cli): search → scrape → map → crawl → **interact** - **Never use interact for web searches** — use `search` instead ## Quick start ```bash # 1. Scrape a page (scrape ID is saved automatically) firecrawl scrape "<url>" # 2. Interact with the page using natural language firecrawl interact --prompt "Click the login button" firecrawl interact --prompt "Fill in the email field with test@example.com" firecrawl interact --prompt "Extract the pricing table" # 3. Or use code for precise control firecrawl interact --code "agent-browser click @e5" --language bash firecrawl interact --code "agent-browser snapshot -i" --language bash # 4. Stop the session when done firecrawl interact stop ``` ## Options | Option | Description | | --------------------- | ------------------------------------------------- | | `--prompt <text>` | Natural language instruction (use this OR --code) | | `--code <code>` | Code to execute in the browser session | | `--language <lang>` | Language for code: bash, python, node | | `--timeout <seconds>` | Execution timeout (default: 30, max: 300) | | `--scrape-id <id>` | Target a specific scrape (default: last scrape) | | `-o, --output <path>` | Output file path | ## Profiles Use `--profile` on the scrape to persist browser state (cookies, localStorage) across scrapes: ```bash # Session 1: Login and save state firecrawl scrape "https://app.example.com/login" --profile my-app firecrawl interact --prompt "Fill in email with user@example.com and click login" # Session 2: Come back authenticated firecrawl scrape "https://app.example.com/dashboard" --profile my-app firecrawl interact --prompt "Extract the dashboard data" ``` Read-only reconnect (no writes to profile state): ```bash firecrawl scrape "https://app.example.com" --profile my-app --no-save-changes ``` ## Tips - Always scrape first — `interact` requires a scrape ID from a previous `firecrawl scrape` call - The scrape ID is saved automatically, so you don't need `--scrape-id` for subsequent interact calls - Use `firecrawl interact stop` to free resources when done - For parallel work, scrape multiple pages and interact with each using `--scrape-id` ## See also - [firecrawl-scrape](../firecrawl-scrape/SKILL.md) — try scrape first, escalate to interact only when needed - [firecrawl-search](../firecrawl-search/SKILL.md) — for web searches (never use interact for searching) - [firecrawl-agent](../firecrawl-agent/SKILL.md) — AI-powered extraction (less manual control)