capacity-planner
Translates performance test results into infrastructure scaling recommendations with cost estimates. Use when someone has load test data and needs to know what infrastructure changes are required to handle target traffic. Trigger words: capacity planning, scaling plan, infrastructure sizing, how many pods, RDS sizing, handle more traffic, scale for launch.
Best use case
capacity-planner is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
Translates performance test results into infrastructure scaling recommendations with cost estimates. Use when someone has load test data and needs to know what infrastructure changes are required to handle target traffic. Trigger words: capacity planning, scaling plan, infrastructure sizing, how many pods, RDS sizing, handle more traffic, scale for launch.
Teams using capacity-planner should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/capacity-planner/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How capacity-planner Compares
| Feature / Agent | capacity-planner | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Translates performance test results into infrastructure scaling recommendations with cost estimates. Use when someone has load test data and needs to know what infrastructure changes are required to handle target traffic. Trigger words: capacity planning, scaling plan, infrastructure sizing, how many pods, RDS sizing, handle more traffic, scale for launch.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Capacity Planner ## Overview This skill takes performance test results (breaking points, latency data, error patterns) and produces actionable infrastructure scaling recommendations. It identifies bottlenecks, recommends specific resource changes, estimates costs, and generates verification criteria. ## Instructions ### Step 1: Gather Current State Ask or determine: - Current infrastructure specs (pod count, instance types, DB size, cache config) - Current sustainable RPS and latency at that level - Breaking point from load tests (RPS or VU count where SLOs break) - Target traffic (expressed as RPS, concurrent users, or multiplier like "3x") ### Step 2: Identify Bottleneck Layer From load test error patterns, determine the primary bottleneck: | Error Pattern | Likely Bottleneck | |---|---| | Timeouts increasing gradually | CPU saturation on application pods | | Sudden spike in 5xx errors | Database connection pool exhaustion | | Connection refused errors | Pod/instance limit reached | | Slow first byte, fast transfer | Database query latency | | Memory-related crashes | Application memory leak or undersized instances | | Intermittent 503s | Load balancer target limits | ### Step 3: Calculate Required Resources Use linear scaling as baseline, then apply correction factors: ``` Required pods = current_pods × (target_RPS / current_max_RPS) × 1.3 safety factor Required DB connections = current_pool × (target_RPS / current_max_RPS) × 1.2 ``` Adjust for: - Database scaling is NOT linear (add read replicas for read-heavy, scale vertically for write-heavy) - Cache hit rates improve with more traffic (warm cache) - Connection overhead per pod ### Step 4: Estimate Costs Use these AWS reference prices (us-east-1, on-demand): | Resource | Spec | Monthly Cost | |---|---|---| | EKS pod (500m/1Gi) | ~0.25 vCPU equiv | ~$35/mo | | RDS db.r5.large | 2 vCPU, 16 GiB | $274/mo | | RDS db.r5.xlarge | 4 vCPU, 32 GiB | $548/mo | | ElastiCache r6g.large | 2 vCPU, 13 GiB | $172/mo | | RDS Read Replica | Same as primary instance | Same cost | | ALB | Base + LCU | ~$25-50/mo | Note: actual costs vary by region, reserved instances, and savings plans. ### Step 5: Produce Scaling Plan Format as actionable items with: 1. What to change (specific resource, config parameter) 2. Current value → target value 3. Why (which bottleneck it addresses) 4. Cost impact 5. Implementation order (dependencies between changes) 6. Verification criteria (what to re-test after) ## Examples ### Example 1: Pre-Launch Scaling Plan **User prompt:** ``` We break at 400 RPS with 3 pods and RDS db.r5.large. Need to handle 1200 RPS for launch. ``` **Agent output:** ``` Capacity Plan: 400 → 1,200 RPS ─────────────────────────────── Step Change Cost Delta Addresses 1 DB pool_size: 20 → 80 $0 Connection exhaustion (primary bottleneck) 2 RDS: db.r5.large → db.r5.2xlarge +$822/mo Write throughput + connections 3 Add RDS read replica +$1,096/mo Read query offloading 4 API pods: 3 → 10 (HPA min=7) +$245/mo Request processing capacity 5 Redis: r6g.medium → r6g.xlarge +$258/mo Session/cache throughput Total additional cost: $2,421/month Implementation order: 1 → 2 → 3 → 4 → 5 (each step unblocks the next) Verification: Re-run load test after each step. Expected milestones: After step 1-2: Breaking point moves to ~700 RPS After step 3: Breaking point moves to ~1,000 RPS After step 4-5: Target 1,200 RPS with p95 < 500ms ``` ### Example 2: Cost-Optimized Scaling **User prompt:** ``` We need to handle 2x traffic but leadership wants minimal cost increase. ``` **Agent output:** ``` Cost-Optimized Scaling Plan (2x traffic) ──────────────────────────────────────── FREE WINS (configuration only) 1. Increase DB connection pool: 20 → 40 +0$/mo Impact: +30% throughput 2. Enable gzip compression on API responses +0$/mo Impact: -40% bandwidth 3. Add Redis caching for GET /api/products (TTL 60s) +0$/mo Impact: -50% DB reads LOW COST 4. Add 2 API pods (3 → 5) +$70/mo 5. Enable RDS query cache +$0/mo Expected result: 1.8-2.2x capacity at +$70/month Tradeoff: Less headroom than full scaling. If traffic exceeds 2.2x, latency degrades. ``` ## Guidelines - **Bottleneck first** — always identify and fix the bottleneck before adding horizontal capacity; more pods don't help if the database is the limit - **Implementation order matters** — scaling changes often have dependencies; present them in the right sequence - **Include verification criteria** — every change should have a "how to confirm it worked" test - **Cost transparency** — always show per-item cost changes and total monthly impact - **Safety margins** — recommend 30% headroom over target for unexpected spikes - **Distinguish read vs write scaling** — reads scale with replicas/cache; writes require vertical scaling or sharding - **Consider the free wins first** — configuration changes (pool sizes, caching, compression) often provide significant gains at zero cost
Related Skills
zustand
You are an expert in Zustand, the small, fast, and scalable state management library for React. You help developers manage global state without boilerplate using Zustand's hook-based stores, selectors for performance, middleware (persist, devtools, immer), computed values, and async actions — replacing Redux complexity with a simple, un-opinionated API in under 1KB.
zoho
Integrate and automate Zoho products. Use when a user asks to work with Zoho CRM, Zoho Books, Zoho Desk, Zoho Projects, Zoho Mail, or Zoho Creator, build custom integrations via Zoho APIs, automate workflows with Deluge scripting, sync data between Zoho apps and external systems, manage leads and deals, automate invoicing, build custom Zoho Creator apps, set up webhooks, or manage Zoho organization settings. Covers Zoho CRM, Books, Desk, Projects, Creator, and cross-product integrations.
zod
You are an expert in Zod, the TypeScript-first schema declaration and validation library. You help developers define schemas that validate data at runtime AND infer TypeScript types at compile time — eliminating the need to write types and validators separately. Used for API input validation, form validation, environment variables, config files, and any data boundary.
zipkin
Deploy and configure Zipkin for distributed tracing and request flow visualization. Use when a user needs to set up trace collection, instrument Java/Spring or other services with Zipkin, analyze service dependencies, or configure storage backends for trace data.
zig
Expert guidance for Zig, the systems programming language focused on performance, safety, and readability. Helps developers write high-performance code with compile-time evaluation, seamless C interop, no hidden control flow, and no garbage collector. Zig is used for game engines, operating systems, networking, and as a C/C++ replacement.
zed
Expert guidance for Zed, the high-performance code editor built in Rust with native collaboration, AI integration, and GPU-accelerated rendering. Helps developers configure Zed, create custom extensions, set up collaborative editing sessions, and integrate AI assistants for productive coding.
zeabur
Expert guidance for Zeabur, the cloud deployment platform that auto-detects frameworks, builds and deploys applications with zero configuration, and provides managed services like databases and message queues. Helps developers deploy full-stack applications with automatic scaling and one-click marketplace services.
zapier
Automate workflows between apps with Zapier. Use when a user asks to connect apps without code, automate repetitive tasks, sync data between services, or build no-code integrations between SaaS tools.
zabbix
Configure Zabbix for enterprise infrastructure monitoring with templates, triggers, discovery rules, and dashboards. Use when a user needs to set up Zabbix server, configure host monitoring, create custom templates, define trigger expressions, or automate host discovery and registration.
yup
Validate data with Yup schemas. Use when adding form validation, defining API request schemas, validating configuration, or building type-safe validation pipelines in JavaScript/TypeScript.
yt-dlp
Download video and audio from YouTube and other platforms with yt-dlp. Use when a user asks to download YouTube videos, extract audio from videos, download playlists, get subtitles, download specific formats or qualities, batch download, archive channels, extract metadata, embed thumbnails, download from social media platforms (Twitter, Instagram, TikTok), or build media ingestion pipelines. Covers format selection, audio extraction, playlists, subtitles, metadata, and automation.
youtube-transcription
Transcribe YouTube videos to text using OpenAI Whisper and yt-dlp. Use when the user wants to get a transcript from a YouTube video, generate subtitles, convert video speech to text, create SRT/VTT captions, or extract spoken content from YouTube URLs.