Claude Agent Skill · by Firecrawl

Firecrawl Interact

When regular web scraping hits a wall because content is locked behind JavaScript interactions, login forms, or multi-step flows, this gives Claude control of a

Install
Terminal · npx
$npx skills add https://github.com/vercel-labs/agent-skills --skill vercel-react-best-practices
Works with Paperclip

How Firecrawl Interact fits into a Paperclip company.

Firecrawl Interact drops into any Paperclip agent that handles this kind of work. Assign it to a specialist inside a pre-configured PaperclipOrg company and the skill becomes available on every heartbeat — no prompt engineering, no tool wiring.

S
SaaS FactoryPaired

Pre-configured AI company — 18 agents, 18 skills, one-time purchase.

$27$59
Explore pack
Source file
SKILL.md83 lines
Expand
---name: firecrawl-interactdescription: |  Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on "interact", "click", "fill out the form", "log in to", "sign in", "submit", "paginated", "next page", "infinite scroll", "interact with the page", "navigate to", "open a session", or "scrape failed".allowed-tools:  - Bash(firecrawl *)  - Bash(npx firecrawl *)--- # firecrawl interact Interact with scraped pages in a live browser session. Scrape a page first, then use natural language prompts or code to click, fill forms, navigate, and extract data. ## When to use - Content requires interaction: clicks, form fills, pagination, login- `scrape` failed because content is behind JavaScript interaction- You need to navigate a multi-step flow- Last resort in the [workflow escalation pattern](firecrawl-cli): search → scrape → map → crawl → **interact**- **Never use interact for web searches** — use `search` instead ## Quick start ```bash# 1. Scrape a page (scrape ID is saved automatically)firecrawl scrape "<url>" # 2. Interact with the page using natural languagefirecrawl interact --prompt "Click the login button"firecrawl interact --prompt "Fill in the email field with test@example.com"firecrawl interact --prompt "Extract the pricing table" # 3. Or use code for precise controlfirecrawl interact --code "agent-browser click @e5" --language bashfirecrawl interact --code "agent-browser snapshot -i" --language bash # 4. Stop the session when donefirecrawl interact stop``` ## Options | Option                | Description                                       || --------------------- | ------------------------------------------------- || `--prompt <text>`     | Natural language instruction (use this OR --code) || `--code <code>`       | Code to execute in the browser session            || `--language <lang>`   | Language for code: bash, python, node             || `--timeout <seconds>` | Execution timeout (default: 30, max: 300)         || `--scrape-id <id>`    | Target a specific scrape (default: last scrape)   || `-o, --output <path>` | Output file path                                  | ## Profiles Use `--profile` on the scrape to persist browser state (cookies, localStorage) across scrapes: ```bash# Session 1: Login and save statefirecrawl scrape "https://app.example.com/login" --profile my-appfirecrawl interact --prompt "Fill in email with user@example.com and click login" # Session 2: Come back authenticatedfirecrawl scrape "https://app.example.com/dashboard" --profile my-appfirecrawl interact --prompt "Extract the dashboard data"``` Read-only reconnect (no writes to profile state): ```bashfirecrawl scrape "https://app.example.com" --profile my-app --no-save-changes``` ## Tips - Always scrape first — `interact` requires a scrape ID from a previous `firecrawl scrape` call- The scrape ID is saved automatically, so you don't need `--scrape-id` for subsequent interact calls- Use `firecrawl interact stop` to free resources when done- For parallel work, scrape multiple pages and interact with each using `--scrape-id` ## See also - [firecrawl-scrape](../firecrawl-scrape/SKILL.md) — try scrape first, escalate to interact only when needed- [firecrawl-search](../firecrawl-search/SKILL.md) — for web searches (never use interact for searching)- [firecrawl-agent](../firecrawl-agent/SKILL.md) — AI-powered extraction (less manual control)