firecrawl-web
Fetch web content, take screenshots, extract structured data, search the web, and crawl documentation sites. Use when the user needs current web information, asks to scrape a URL, wants a screenshot, needs to extract specific data from a page, or wants to learn about a framework or library.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/firecrawl-web/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How firecrawl-web Compares
| Feature / Agent | firecrawl-web | Standard Approach |
|---|---|---|
| Platform Support | multi | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Fetch web content, take screenshots, extract structured data, search the web, and crawl documentation sites. Use when the user needs current web information, asks to scrape a URL, wants a screenshot, needs to extract specific data from a page, or wants to learn about a framework or library.
Which AI agents support this skill?
This skill is compatible with multi.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Firecrawl Web Skill
This skill provides web access through Firecrawl's API.
## Script Location
All commands use the bundled script:
~/.claude/skills/firecrawl-web/fc.py
## Getting Page Content
Fetch any webpage as clean markdown:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py markdown "https://example.com"
```
For cleaner output without navigation and footers:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py markdown "https://example.com" --main-only
```
## Taking Screenshots
Capture a full-page screenshot:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py screenshot "https://example.com" -o page.png
```
## Extracting Structured Data
Extract specific data using a JSON schema. Create a schema file first:
```json
{
"type": "object",
"properties": {
"title": {"type": "string"},
"price": {"type": "number"},
"features": {"type": "array", "items": {"type": "string"}}
}
}
```
Then extract:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py extract "https://example.com/product" --schema schema.json
```
Add a prompt for better accuracy:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py extract "https://example.com/product" --schema schema.json --prompt "Extract the main product details"
```
## Searching the Web
Search for current information:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py search "Python 3.13 new features"
```
Limit results:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py search "latest React documentation" --limit 3
```
## Crawling Documentation
Crawl a documentation site to learn about a new framework:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py crawl "https://docs.newframework.dev" --limit 30
```
Save pages to a directory:
```bash
python3 ~/.claude/skills/firecrawl-web/fc.py crawl "https://docs.example.com" --limit 50 --output ./docs
```
Each page costs one credit. Set a reasonable limit to avoid burning through your quota.