perf-benchmarker

Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.

23 stars

Best use case

perf-benchmarker is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.

Teams using perf-benchmarker should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/perf-benchmarker/SKILL.md --create-dirs "https://raw.githubusercontent.com/christophacham/agent-skills-library/main/skills/game-dev/perf-benchmarker/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/perf-benchmarker/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How perf-benchmarker Compares

Feature / Agentperf-benchmarkerStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

Use when running performance benchmarks, establishing baselines, or validating regressions with sequential runs. Enforces 60s minimum runs (30s only for binary search) and no parallel benchmarks.

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

# perf-benchmarker

Run sequential benchmarks with strict duration rules.

Follow `docs/perf-requirements.md` as the canonical contract.

## Parse Arguments

```javascript
const args = '$ARGUMENTS'.split(' ').filter(Boolean);
const command = args.find(a => !a.match(/^\d+$/)) || '';
const duration = parseInt(args.find(a => a.match(/^\d+$/)) || '60', 10);
```

## Required Rules

- Benchmarks MUST run sequentially (never parallel).
- Minimum duration: 60s per run (30s only for binary search).
- Warmup: 10s minimum before measurement.
- Re-run anomalies.

## Output Format

```
command: <benchmark command>
duration: <seconds>
warmup: <seconds>
results: <metrics summary>
notes: <anomalies or reruns>
```

## Output Contract

Benchmarks MUST emit a JSON metrics block between markers:

```
PERF_METRICS_START
{"scenarios":{"low":{"latency_ms":120},"high":{"latency_ms":450}}}
PERF_METRICS_END
```

## Constraints

- No short runs unless binary-search phase.
- Do not change code while benchmarking.

Related Skills

perf-theory-gatherer

23
from christophacham/agent-skills-library

Use when generating performance hypotheses backed by git history and code evidence.

perf-baseline-manager

23
from christophacham/agent-skills-library

Use when managing perf baselines, consolidating results, or comparing versions. Ensures one baseline JSON per version.

python-performance-optimization

23
from christophacham/agent-skills-library

Profile and optimize Python code using cProfile, memory profilers, and performance best practices. Use when debugging slow Python code, optimizing bottlenecks, or improving application performance.

power-bi-performance-troubleshooting

23
from christophacham/agent-skills-library

Systematic Power BI performance troubleshooting prompt for identifying, diagnosing, and resolving performance issues in Power BI models, reports, and queries.

performance-profiling

23
from christophacham/agent-skills-library

Performance profiling principles. Measurement, analysis, and optimization techniques.

performance-engineer

23
from christophacham/agent-skills-library

Expert performance engineer specializing in modern observability,

application-performance-performance-optimization

23
from christophacham/agent-skills-library

Optimize end-to-end application performance with profiling, observability, and backend/frontend tuning. Use when coordinating performance optimization across the stack.

performance-testing-review-multi-agent-review

23
from christophacham/agent-skills-library

Use when working with performance testing review multi agent review

performance-testing-review-ai-review

23
from christophacham/agent-skills-library

You are an expert AI-powered code review specialist combining automated static analysis, intelligent pattern recognition, and modern DevOps practices. Leverage AI tools (GitHub Copilot, Qodo, GPT-5, C

microsoft-code-reference

23
from christophacham/agent-skills-library

Look up Microsoft API references, find working code samples, and verify SDK code is correct. Use when working with Azure SDKs, .NET libraries, or Microsoft APIs—to find the right method, check parameters, get working examples, or troubleshoot errors. Catches hallucinated methods, wrong signatures, and deprecated patterns by querying official docs.

eos-composition

23
from christophacham/agent-skills-library

Strunk & White composition review using the 11 principles from "Elements of Style" Chapter II. Use when analyzing structure, improving flow, or tightening prose.

enhance-cross-file

23
from christophacham/agent-skills-library

Use when checking cross-file consistency: tools vs frontmatter, agent references, duplicate rules, contradictions.