multiAI Summary Pending
search-cluster
Aggregated search aggregator using Google CSE, GNews RSS, Wikipedia, Reddit, and Scrapling.
3,556 stars
byopenclaw
Installation
Claude Code / Cursor / Codex
$curl -o ~/.claude/skills/search-cluster/SKILL.md --create-dirs "https://raw.githubusercontent.com/openclaw/skills/main/skills/1999azzar/search-cluster/SKILL.md"
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/search-cluster/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How search-cluster Compares
| Feature / Agent | search-cluster | Standard Approach |
|---|---|---|
| Platform Support | multi | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Aggregated search aggregator using Google CSE, GNews RSS, Wikipedia, Reddit, and Scrapling.
Which AI agents support this skill?
This skill is compatible with multi.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Search Cluster (Industrial Standard v3.1) A multi-provider search aggregator designed for high-availability and security. ## Installation The scrapling provider requires a dedicated virtual environment. 1. Create a venv: python3 -m venv venv/scrapling 2. Install scrapling: venv/scrapling/bin/pip install scrapling 3. Provide the path to the venv binary in SCRAPLING_PYTHON_PATH. ## Security Posture - Subprocess Isolation: Query inputs are passed as arguments to stealth_fetch.py. - Strict TLS: Mandatory SSL verification on all providers. - Sanitization: Integrated native internal scrubber (Path Neutral). ## Requirements and Environment Declare these variables in your environment or vault: | Variable | Requirement | Description | |---|---|---| | GOOGLE_API_KEY | Optional | API Key for Google Custom Search. | | GOOGLE_CSE_ID | Optional | Search Engine ID for Google CSE. | | SCRAPLING_PYTHON_PATH | Optional | Path to the scrapling venv python binary. | | REDIS_HOST | Optional | Host for result caching. | | REDIS_PORT | Optional | Port for result caching (Default: 6379). | | SEARCH_USER_AGENT | Optional | Custom User-Agent string. | ## Providers - google: Official Google Custom Search. - wiki: Wikipedia OpenSearch API. - reddit: Reddit JSON search API. - gnews: Google News RSS aggregator. - scrapling: Headless stealth scraping (via DuckDuckGo). ## Included Scripts - scripts/search-cluster.py: Main entry point. - scripts/stealth_fetch.py: Scrapling fetcher (REQUIRED for scrapling provider). ## Workflow 1. Execute: scripts/search-cluster.py all "<query>" 2. Output is structured JSON with source, title, link, and sanitized snippet.