npx skills add https://github.com/firecrawl/cli --skill firecrawl-agentHow Firecrawl Agent fits into a Paperclip company.
Firecrawl Agent drops into any Paperclip agent that handles this kind of work. Assign it to a specialist inside a pre-configured PaperclipOrg company and the skill becomes available on every heartbeat — no prompt engineering, no tool wiring.
Pre-configured AI company — 18 agents, 18 skills, one-time purchase.
SKILL.md57 linesExpandCollapse
---name: firecrawl-agentdescription: | AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on "extract structured data", "get all the products", "pull pricing info", "extract as JSON", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.allowed-tools: - Bash(firecrawl *) - Bash(npx firecrawl *)--- # firecrawl agent AI-powered autonomous extraction. The agent navigates sites and extracts structured data (takes 2-5 minutes). ## When to use - You need structured data from complex multi-page sites- Manual scraping would require navigating many pages- You want the AI to figure out where the data lives ## Quick start ```bash# Extract structured datafirecrawl agent "extract all pricing tiers" --wait -o .firecrawl/pricing.json # With a JSON schema for structured outputfirecrawl agent "extract products" --schema '{"type":"object","properties":{"name":{"type":"string"},"price":{"type":"number"}}}' --wait -o .firecrawl/products.json # Focus on specific pagesfirecrawl agent "get feature list" --urls "<url>" --wait -o .firecrawl/features.json``` ## Options | Option | Description || ---------------------- | ----------------------------------------- || `--urls <urls>` | Starting URLs for the agent || `--model <model>` | Model to use: spark-1-mini or spark-1-pro || `--schema <json>` | JSON schema for structured output || `--schema-file <path>` | Path to JSON schema file || `--max-credits <n>` | Credit limit for this agent run || `--wait` | Wait for agent to complete || `--pretty` | Pretty print JSON output || `-o, --output <path>` | Output file path | ## Tips - Always use `--wait` to get results inline. Without it, returns a job ID.- Use `--schema` for predictable, structured output — otherwise the agent returns freeform data.- Agent runs consume more credits than simple scrapes. Use `--max-credits` to cap spending.- For simple single-page extraction, prefer `scrape` — it's faster and cheaper. ## See also - [firecrawl-scrape](../firecrawl-scrape/SKILL.md) — simpler single-page extraction- [firecrawl-interact](../firecrawl-interact/SKILL.md) — scrape + interact for manual page interaction (more control)- [firecrawl-crawl](../firecrawl-crawl/SKILL.md) — bulk extraction without AIFirecrawl
This is autonomous web scraping that actually works for complex data extraction tasks. Instead of writing brittle scrapers that break when sites change, you des
Firecrawl Build Interact
When basic web scraping hits a wall because content only appears after clicking buttons, filling forms, or navigating through multi-step flows, this skill integ
Firecrawl Build Onboarding
Gets Firecrawl set up in your project by handling the API key setup, SDK installation, and initial configuration. Runs a browser auth flow if you need to sign u