reconnaissance

Systematic technology and market reconnaissance for extracting actionable intelligence from repositories, papers, and competitive landscapes.

28 stars

Best use case

reconnaissance is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

Systematic technology and market reconnaissance for extracting actionable intelligence from repositories, papers, and competitive landscapes.

Teams using reconnaissance should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/reconnaissance/SKILL.md --create-dirs "https://raw.githubusercontent.com/DNYoussef/context-cascade/main/skills/security/reconnaissance/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/reconnaissance/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How reconnaissance Compares

Feature / AgentreconnaissanceStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

Systematic technology and market reconnaissance for extracting actionable intelligence from repositories, papers, and competitive landscapes.

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

---

## LIBRARY-FIRST PROTOCOL (MANDATORY)

**Before writing ANY code, you MUST check:**

### Step 1: Library Catalog
- Location: `.claude/library/catalog.json`
- If match >70%: REUSE or ADAPT

### Step 2: Patterns Guide
- Location: `.claude/docs/inventories/LIBRARY-PATTERNS-GUIDE.md`
- If pattern exists: FOLLOW documented approach

### Step 3: Existing Projects
- Location: `D:\Projects\*`
- If found: EXTRACT and adapt

### Decision Matrix
| Match | Action |
|-------|--------|
| Library >90% | REUSE directly |
| Library 70-90% | ADAPT minimally |
| Pattern exists | FOLLOW pattern |
| In project | EXTRACT |
| No match | BUILD (add to library after) |

---

## TIER 1: CRITICAL SECTIONS

### Overview

Reconnaissance is a systematic intelligence-gathering skill designed to extract actionable insights from diverse sources including GitHub repositories, academic papers, competitive products, and market landscapes. Unlike general research which synthesizes existing knowledge, reconnaissance actively probes, extracts, and structures novel intelligence for strategic decision-making.

The skill operates on the principle that raw information must be transformed through structured extraction into decision-ready artifacts. It employs a three-phase methodology: Discovery (identify sources), Extraction (pull structured data), and Analysis (synthesize intelligence).

Reconnaissance is essential for:
- Pre-project feasibility assessments
- Technology stack evaluation
- Competitive positioning
- Academic literature mining for implementation patterns
- Market opportunity identification

### Core Principles

Reconnaissance operates on 5 fundamental principles:

#### Principle 1: Source Diversification
Never rely on a single source type. Cross-reference findings across repositories, papers, documentation, and market signals.

**In Practice:**
- Query GitHub + arXiv + product pages for any technology assessment
- Compare stated claims against actual implementations
- Verify recency and maintenance status of all sources

#### Principle 2: Structured Extraction
Raw information is useless without structure. Every reconnaissance output must follow defined schemas.

**In Practice:**
- Use MANIFEST.md templates for repository extractions
- Create comparison charts for multi-source analysis
- Tag all findings with source, date, and confidence level

#### Principle 3: Evidence Grounding
Every claim must trace back to a verifiable source with explicit confidence ceiling.

**In Practice:**
- Link directly to code files, paper sections, or documentation pages
- Apply confidence ceilings: witnessed (0.95), reported (0.70), inferred (0.70)
- Mark gaps explicitly rather than inferring without evidence

#### Principle 4: Actionable Output
Intelligence must drive decisions. Abstract findings without recommendations waste effort.

**In Practice:**
- End every recon with specific recommendations
- Include effort estimates and risk assessments
- Prioritize findings by strategic impact

#### Principle 5: Temporal Awareness
Technology landscapes change rapidly. All reconnaissance has an expiration date.

**In Practice:**
- Date-stamp all outputs prominently
- Note last-commit dates for repositories
- Flag papers older than 2 years as potentially outdated

### When to Use

**Use Reconnaissance When:**
- Evaluating a new technology for adoption
- Analyzing competitor products or approaches
- Extracting implementation patterns from papers
- Assessing open-source project viability
- Mapping market landscape before product decisions
- Pre-mortem analysis for project risks

**Do NOT Use Reconnaissance When:**
- Synthesizing already-gathered sources (use `literature-synthesis`)
- Single-source deep reading (use `academic-reading-workflow`)
- General question answering (use `researcher`)
- Implementation planning (use `research-driven-planning`)

### Main Workflow

#### Phase 1: Scope Definition
**Agent:** intent-analyzer
**Purpose:** Clarify reconnaissance objectives and boundaries

**Inputs:**
- Target domain or technology
- Strategic questions to answer
- Time and depth constraints
- Output format requirements

**Outputs:**
- Scoped reconnaissance brief
- Source categories to probe
- Success criteria

#### Phase 2: Source Discovery
**Agent:** research-agent
**Purpose:** Identify and qualify sources

**Process:**
1. GitHub search for repositories matching domain
2. arXiv/papers search for academic sources
3. Product/company identification for competitive analysis
4. Documentation and blog discovery
5. Source qualification (recency, authority, relevance)

**Outputs:**
- Qualified source list with metadata
- Source-type distribution analysis
- Gap identification (what sources are missing?)

#### Phase 3: Structured Extraction
**Agent:** code-extractor or document-analyst
**Purpose:** Pull structured data from each source

**Process:**
1. Apply extraction template per source type:
   - Repository: MANIFEST.md (structure, deps, patterns)
   - Paper: KEY-FINDINGS.md (methods, results, code availability)
   - Product: FEATURE-MATRIX.md (capabilities, pricing, integration)
2. Normalize extracted data to common schema
3. Cross-reference overlapping claims

**Outputs:**
- Per-source extraction documents
- Normalized data tables
- Cross-reference matrix

#### Phase 4: Intelligence Synthesis
**Agent:** synthesis-agent
**Purpose:** Transform extractions into actionable intelligence

**Process:**
1. Identify patterns across sources
2. Surface contradictions and gaps
3. Generate comparison charts
4. Formulate recommendations with confidence ceilings
5. Create decision frameworks

**Outputs:**
- COMPREHENSIVE-ANALYSIS.md
- COMPARISON-CHART.md
- RECOMMENDATIONS.md with action items
- **GAP-INVENTORY.md** with specific missing capabilities

#### Phase 5: Gap-Filling Extraction (CRITICAL)
**Agent:** code-extractor + research-agent
**Purpose:** Find external repos/research to fill discovered gaps, extract ONLY applicable parts, delete everything else

**Process:**
1. **Gap Identification:** Take gaps from Phase 4 GAP-INVENTORY.md
2. **Source Discovery:** Search GitHub/arXiv/papers for solutions to each gap
3. **Pre-Mortem Analysis:** Before cloning, assess applicability:
   - Does target project already have this capability?
   - Is the architecture compatible?
   - What's the integration effort?
4. **Selective Clone:** Clone promising repos to temp directory
5. **Extraction Criteria:** For each cloned repo, evaluate:
   ```
   KEEP if:
   - Directly fills identified gap
   - Architecture compatible with target
   - Integration effort < gap severity

   DELETE if:
   - Target already has equivalent functionality
   - Architecture incompatible
   - Duplicates existing capabilities
   - Integration effort exceeds value
   ```
6. **Focused Extraction:** Extract ONLY applicable patterns:
   - Copy relevant source files to extraction folder
   - Create integration notes per extracted component
   - Map extracted code to target project structure
7. **Cleanup:** Delete cloned repos after extraction complete
8. **Extraction Report:** Document what was kept, what was deleted, and why

**Outputs:**
- `{target}-gap-extractions/` folder with:
  - Extracted source files (only applicable ones)
  - `EXTRACTION-MANIFEST.md` - what was extracted and why
  - `INTEGRATION-GUIDE.md` - how to integrate each component
  - `DELETED-SUMMARY.md` - what was NOT applicable (with reasons)
- Updated GAP-INVENTORY.md with extraction status

**Critical Rules:**
1. NEVER extract without pre-mortem applicability analysis
2. ALWAYS delete more than you keep (80%+ deletion rate is healthy)
3. Document every deletion decision for future reference
4. Map extractions to specific gaps they fill

#### Phase 6: Delivery and Storage
**Agent:** delivery-agent
**Purpose:** Package and preserve intelligence

**Process:**
1. Organize outputs in dated folder structure:
   ```
   {target}-recon-{date}/
     EXECUTIVE-SUMMARY.md
     MANIFEST.md
     COMPREHENSIVE-ANALYSIS.md
     COMPARISON-CHART.md
     RECOMMENDATIONS.md
     GAP-INVENTORY.md
     gap-extractions/
       EXTRACTION-MANIFEST.md
       INTEGRATION-GUIDE.md
       DELETED-SUMMARY.md
       extracted-files/
   ```
2. Update memory-mcp with key findings
3. Generate executive summary
4. Archive raw extractions for audit trail

**Outputs:**
- Complete reconnaissance package with gap extractions
- Memory-mcp entries for future retrieval
- Executive brief for stakeholders
- Actionable extraction folder ready for integration

---

## TIER 2: ESSENTIAL SECTIONS

### Pattern Recognition

Different reconnaissance targets require different approaches:

#### Technology Evaluation Pattern
**Triggers:** "evaluate", "assess", "should we use", "compare frameworks"
**Characteristics:**
- Focus on implementation feasibility
- Heavy repository analysis
- Performance and scalability concerns
**Key Focus:** Code quality, maintenance activity, community health
**Approach:** Repository-first with paper backup for theoretical grounding

#### Competitive Analysis Pattern
**Triggers:** "competitor", "market position", "alternative to"
**Characteristics:**
- Product feature comparison
- Pricing and positioning analysis
- Strategic differentiation needs
**Key Focus:** Feature gaps, market positioning, integration capabilities
**Approach:** Product documentation and public materials analysis

#### Academic Extraction Pattern
**Triggers:** "paper", "research", "implementation from", "replicate"
**Characteristics:**
- Method and algorithm focus
- Code availability critical
- Reproducibility concerns
**Key Focus:** Implementation details, performance claims, code artifacts
**Approach:** Paper-first with repository search for implementations

#### Market Landscape Pattern
**Triggers:** "landscape", "players in", "market for", "opportunity"
**Characteristics:**
- Broad coverage over depth
- Trend identification
- Segment analysis
**Key Focus:** Market size, key players, emerging trends, gaps
**Approach:** Multi-source sweep with structured aggregation

### Advanced Techniques

#### Multi-Model Reconnaissance
Route different extraction tasks to optimal models:
- **Gemini:** Large repository analysis (1M context), Google Search grounding
- **Codex:** Automated code extraction and pattern mining
- **Claude:** Synthesis and strategic analysis

```bash
# Example multi-model routing
bash -lc "gemini --all-files 'extract architecture patterns from this repo'"
bash -lc "codex exec 'analyze dependency graph and identify risks'"
```

#### Temporal Layering
Compare sources across time to identify trends:
1. Historical baseline (2+ years ago)
2. Recent developments (6-24 months)
3. Current state (last 6 months)
4. Emerging signals (preprints, announcements)

#### Confidence Cascading
Layer confidence based on source quality:
```
witnessed:code-analysis     -> conf:0.95
witnessed:doc-extraction    -> conf:0.90
reported:paper-claims       -> conf:0.85
reported:blog-posts         -> conf:0.70
inferred:trend-analysis     -> conf:0.65
```

### Common Anti-Patterns

| Anti-Pattern | Problem | Solution |
|--------------|---------|----------|
| **Single-Source Reliance** | Biased or incomplete view | Always cross-reference 3+ sources |
| **Recency Blindness** | Outdated recommendations | Check last-commit/publish dates prominently |
| **Claim Inflation** | Overstating source confidence | Apply strict ceiling discipline |
| **Feature List Syndrome** | Listing without analysis | Always include strategic implications |
| **Extraction Without Synthesis** | Raw data dumps | Require actionable conclusions |
| **Scope Creep** | Unbounded reconnaissance | Time-box and scope strictly upfront |

### Practical Guidelines

#### Full vs. Quick Reconnaissance

| Aspect | Quick Recon (2-4h) | Full Recon (1-2 days) |
|--------|-------------------|----------------------|
| Sources | 3-5 primary | 10-20 comprehensive |
| Depth | Surface extraction | Deep analysis |
| Output | Single summary | Full package |
| Use When | Time-sensitive decisions | Strategic planning |

#### Quality Checkpoints
- [ ] Every claim has source link
- [ ] Confidence ceilings applied consistently
- [ ] Date stamps on all documents
- [ ] Recommendations are actionable
- [ ] Gaps explicitly identified

#### Trade-offs
- **Breadth vs. Depth:** Quick decisions need breadth; implementation needs depth
- **Speed vs. Accuracy:** Higher confidence requires more cross-referencing
- **Comprehensiveness vs. Actionability:** Too much data obscures recommendations

---

## TIER 3: INTEGRATION SECTIONS

### Cross-Skill Coordination

#### Upstream Skills (provide input)
| Skill | When Used Before | What It Provides |
|-------|-----------------|------------------|
| intent-analyzer | Always | Clarified objectives and constraints |
| prompt-architect | Complex requests | Optimized reconnaissance brief |

#### Downstream Skills (use output)
| Skill | When Used After | What It Does |
|-------|----------------|--------------|
| literature-synthesis | Multi-paper recon | Synthesizes extracted papers |
| research-driven-planning | Tech evaluation | Plans implementation from findings |
| decision-framework | Strategic recon | Structures decision from intelligence |

#### Parallel Skills (work together)
| Skill | When Co-invoked | Coordination |
|-------|-----------------|--------------|
| code-extractor | Repository recon | Parallel extraction tasks |
| web-researcher | Market recon | Parallel web searches |

### MCP Requirements

#### Required MCPs
| MCP | Purpose | Why Needed |
|-----|---------|------------|
| memory-mcp | Store findings | Cross-session retrieval of reconnaissance |
| sequential-thinking | Complex analysis | Multi-step reasoning chains |

#### Optional MCPs
| MCP | Purpose | When to Enable |
|-----|---------|---------------|
| playwright | Web scraping | Product page extraction |
| github-mcp | Repo analysis | Deep repository intelligence |

#### MCP Tagging Protocol
```json
{
  "WHO": "reconnaissance-{session_id}",
  "WHEN": "ISO8601 timestamp",
  "PROJECT": "target-domain",
  "WHY": "reconnaissance|competitive-analysis|tech-eval"
}
```

### Input/Output Contracts

```yaml
inputs:
  target: string           # Required: domain, technology, or competitor name
  questions: list[string]  # Required: specific questions to answer
  depth: enum[quick, full] # Optional: default 'quick'
  sources: list[string]    # Optional: pre-specified sources to include
  constraints:
    time_limit: string     # Optional: e.g., "4 hours"
    exclusions: list[string] # Optional: sources to skip

outputs:
  reconnaissance_package:
    manifest: file         # Source inventory with metadata
    extractions: list[file] # Per-source structured extractions
    analysis: file         # COMPREHENSIVE-ANALYSIS.md
    comparison: file       # COMPARISON-CHART.md (if multi-source)
    recommendations: file  # Actionable next steps
    executive_summary: string # Brief for stakeholders

  memory_entries:
    key_findings: list     # Stored to memory-mcp
    source_index: object   # For future retrieval

  metadata:
    confidence: float      # Overall confidence (0.0-1.0)
    ceiling: string        # Confidence ceiling type
    duration: string       # Actual time spent
    gaps: list[string]     # Identified information gaps
```

### Recursive Improvement

#### Self-Application
```
reconnaissance.improve(reconnaissance)
```

**Improvement Dimensions:**
1. Extraction template coverage
2. Source discovery efficiency
3. Synthesis quality metrics
4. Recommendation actionability

#### Eval Harness Integration
- Track source discovery rate
- Measure extraction completeness
- Assess recommendation adoption
- Monitor confidence calibration

#### Memory Namespace
```
reconnaissance:{project}:{date}:{type}
```

---

## TIER 4: CLOSURE SECTIONS

### Examples

#### Example 1: Technology Evaluation
```
User: "Should we use FATE-LLM for our federated learning project?"

Task("reconnaissance", {
  target: "FATE-LLM",
  questions: [
    "What are the core capabilities?",
    "How active is development?",
    "What are the integration requirements?",
    "Are there production deployments?"
  ],
  depth: "full"
})

Output:
- FATE-LLM MANIFEST.md with architecture analysis
- Dependency and compatibility matrix
- Community health metrics
- Recommendation: Suitable for X, concerns about Y
```

#### Example 2: Competitive Analysis
```
User: "Map the edge inference market for our fog-compute pitch"

Task("reconnaissance", {
  target: "edge inference platforms",
  questions: [
    "Who are the major players?",
    "What are the pricing models?",
    "Where are the gaps we can fill?"
  ],
  depth: "quick"
})

Output:
- Player inventory with positioning
- Feature comparison matrix
- Gap analysis with opportunity sizing
- Pitch angle recommendations
```

#### Example 3: Paper Extraction
```
User: "Extract implementation patterns from these 3 federated learning papers"

Task("reconnaissance", {
  target: "federated learning papers",
  sources: ["arXiv:2504.00407", "arXiv:2411.16086", "arXiv:2503.18986"],
  questions: [
    "What are the key algorithms?",
    "Is code available?",
    "What are the performance claims?"
  ],
  depth: "full"
})

Output:
- Per-paper extraction documents
- Algorithm comparison chart
- Code availability matrix
- Integration roadmap for fog-compute
```

### Troubleshooting

| Problem | Cause | Solution |
|---------|-------|----------|
| No repositories found | Narrow search terms | Broaden keywords, try alternative names |
| Conflicting claims | Multiple sources disagree | Note conflict explicitly, investigate root cause |
| Stale sources only | Domain is new or niche | Expand to preprints, conference proceedings |
| Extraction too shallow | Time constraints | Prioritize highest-impact sources |
| Recommendations unclear | Insufficient synthesis | Re-run synthesis phase with explicit decision criteria |
| Memory storage fails | MCP not configured | Verify memory-mcp connection, use fallback file storage |

### Conclusion

Reconnaissance transforms raw information into strategic advantage through systematic discovery, structured extraction, and evidence-grounded synthesis. The skill's value lies not in information gathering alone, but in producing decision-ready intelligence with explicit confidence bounds.

Key success factors:
1. Scope strictly before starting
2. Diversify sources systematically
3. Apply consistent extraction templates
4. Synthesize with confidence ceilings
5. Deliver actionable recommendations

### Completion Verification

- [ ] Scope defined with clear questions and constraints
- [ ] Sources diversified across 3+ types (repos, papers, products)
- [ ] Structured extractions completed for all sources
- [ ] Cross-reference matrix identifies conflicts and gaps
- [ ] Synthesis includes confidence ceilings on all claims
- [ ] Recommendations are specific and actionable
- [ ] Memory-mcp entries created for key findings
- [ ] Executive summary provides decision-ready brief
- [ ] All outputs date-stamped

Confidence: 0.85 (ceiling: research 0.85) - skill definition based on observed reconnaissance patterns and skill-forge requirements.

---

## VCL COMPLIANCE APPENDIX (Internal Reference)

[[HON:teineigo]] [[MOR:root:R-C-N]] [[COM:Reconnaissance+Skill]] [[CLS:ge_skill]] [[EVD:-DI<gozlem>]] [[ASP:nesov.]] [[SPC:path:/skills/research/reconnaissance]]
[define|neutral] CONFIDENCE_CEILINGS := {inference:0.70, report:0.70, research:0.85, observation:0.95, definition:0.95} [conf:0.9] [state:confirmed]
[direct|emphatic] L2_LANGUAGE := English; user-facing outputs exclude VCL markers. [conf:0.99] [state:confirmed]
[commit|confident] <promise>RECONNAISSANCE_SKILL_VCL_V3.1.1_COMPLIANT</promise> [conf:0.85] [state:confirmed]

Related Skills

/*============================================================================*/

28
from DNYoussef/context-cascade

/* SKILL SKILL :: VERILINGUA x VERIX EDITION */

web-scraping

28
from DNYoussef/context-cascade

Structured data extraction from web pages using claude-in-chrome MCP with sequential-thinking planning. Focus on READ operations, data transformation, and pagination handling for multi-page extraction.

visual-testing

28
from DNYoussef/context-cascade

Screenshot-based visual comparison and regression testing using claude-in-chrome MCP. Captures, compares, and validates UI states to detect layout shifts, visual bugs, and design regressions across viewports.

reflect

28
from DNYoussef/context-cascade

Extract learnings from session corrections and patterns, update skill files with persistent memory. Implements Loop 1.5 - per-session micro-learning between execution and meta-optimization.

fix-bug

28
from DNYoussef/context-cascade

Fix bug command

e2e-test

28
from DNYoussef/context-cascade

End-to-end testing workflow for validating complete user journeys through web applications using claude-in-chrome MCP. Specializes in test assertions, suite organization, evidence collection, and pass/fail reporting.

build-feature

28
from DNYoussef/context-cascade

Build feature command

browser-automation

28
from DNYoussef/context-cascade

Complex browser automation workflow using claude-in-chrome MCP with mandatory sequential-thinking planning. Use when automating multi-step web interactions, form filling, navigation sequences, or web scraping.

image-gen

28
from DNYoussef/context-cascade

Modular image generation - supports local SDXL Lightning, OpenAI DALL-E, Replicate, or custom providers

reverse-engineering-quick-triage

28
from DNYoussef/context-cascade

Fast binary analysis with string reconnaissance and static disassembly\ \ (RE Levels 1-2). Use when triaging suspicious binaries, extracting IOCs quickly,\ \ or performing initial malware analysis. Completes in \u22642 hours with automated\ \ decision gates."

reverse-engineering-firmware-analysis

28
from DNYoussef/context-cascade

Firmware extraction and IoT security analysis (RE Level 5) for routers and embedded systems. Use when analyzing IoT firmware, extracting embedded filesystems (SquashFS/JFFS2/CramFS), finding hardcoded credentials, performing CVE scans, or auditing embedded system security. Handles encrypted firmware with known decryption schemes. Completes in 2-8 hours with binwalk+firmadyne+QEMU emulation.

reverse-engineering-deep-analysis

28
from DNYoussef/context-cascade

Advanced binary analysis with runtime execution and symbolic path exploration (RE Levels 3-4). Use when need runtime behavior, memory dumps, secret extraction, or input synthesis to reach specific program states. Completes in 3-7 hours with GDB+Angr.