perf-theory-gatherer
Use when generating performance hypotheses backed by git history and code evidence.
Best use case
perf-theory-gatherer is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
Use when generating performance hypotheses backed by git history and code evidence.
Teams using perf-theory-gatherer should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/perf-theory-gatherer/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How perf-theory-gatherer Compares
| Feature / Agent | perf-theory-gatherer | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Use when generating performance hypotheses backed by git history and code evidence.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# perf-theory-gatherer
Generate performance hypotheses for a specific scenario.
Follow `docs/perf-requirements.md` as the canonical contract.
## Required Steps
1. Review recent git history (scope to relevant paths when possible).
2. Identify code paths involved in the scenario (repo-map or grep).
3. Produce up to 5 hypotheses with evidence + confidence.
## Output Format
```
hypotheses:
- id: H1
hypothesis: <short description>
evidence: <file/path or git change>
confidence: low|medium|high
- id: H2
...
```
## Constraints
- MUST check git history before hypothesizing.
- No optimization suggestions; only hypotheses.
- Keep to 5 hypotheses maximum.Related Skills
perf-benchmarker
Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.
perf-baseline-manager
Use when managing perf baselines, consolidating results, or comparing versions. Ensures one baseline JSON per version.
python-performance-optimization
Profile and optimize Python code using cProfile, memory profilers, and performance best practices. Use when debugging slow Python code, optimizing bottlenecks, or improving application performance.
power-bi-performance-troubleshooting
Systematic Power BI performance troubleshooting prompt for identifying, diagnosing, and resolving performance issues in Power BI models, reports, and queries.
performance-profiling
Performance profiling principles. Measurement, analysis, and optimization techniques.
performance-engineer
Expert performance engineer specializing in modern observability,
application-performance-performance-optimization
Optimize end-to-end application performance with profiling, observability, and backend/frontend tuning. Use when coordinating performance optimization across the stack.
performance-testing-review-multi-agent-review
Use when working with performance testing review multi agent review
performance-testing-review-ai-review
You are an expert AI-powered code review specialist combining automated static analysis, intelligent pattern recognition, and modern DevOps practices. Leverage AI tools (GitHub Copilot, Qodo, GPT-5, C
repo-story-time
Generate a comprehensive repository summary and narrative story from commit history
release-notes
Generates structured release notes from git history between two references (tags, commits, branches). Groups changes by type (features, fixes, docs, breaking), extracts PR references, and produces a publish-ready document.
release-it
Build production-ready systems with stability patterns: circuit breakers, bulkheads, timeouts, and retry logic. Use when the user mentions "production outage", "circuit breaker", "timeout strategy", "deployment pipeline", or "chaos engineering". Covers capacity planning, health checks, and anti-fragility patterns. For data systems, see ddia-systems. For system architecture, see system-design.