Product Analytics — Metrics, Funnels, and Growth
## Overview
Best use case
Product Analytics — Metrics, Funnels, and Growth is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
## Overview
Teams using Product Analytics — Metrics, Funnels, and Growth should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/product-analytics/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How Product Analytics — Metrics, Funnels, and Growth Compares
| Feature / Agent | Product Analytics — Metrics, Funnels, and Growth | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
## Overview
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
AI Agents for Marketing
Discover AI agents for marketing workflows, from SEO and content production to campaign research, outreach, and analytics.
AI Agents for Startups
Explore AI agent skills for startup validation, product research, growth experiments, documentation, and fast execution with small teams.
Best AI Agents for Marketing
A curated list of the best AI agents and skills for marketing teams focused on SEO, content systems, outreach, and campaign execution.
SKILL.md Source
# Product Analytics — Metrics, Funnels, and Growth ## Overview Product analytics, helping product teams define metrics, build funnels, analyze retention, run A/B tests, and make data-driven decisions. This skill applies frameworks for North Star metrics, pirate metrics (AARRR), cohort analysis, and experiment design. ## Instructions ### North Star Metric ```markdown ## Define Your North Star Metric The North Star is a single metric that captures the core value your product delivers to customers. It aligns every team around what matters most. ### Criteria for a Good North Star 1. Reflects customer value (not just revenue) 2. Leading indicator of revenue (not lagging) 3. Measurable and actionable 4. Every team can influence it ### Examples by Product Type - **Marketplace**: Weekly transactions (Airbnb: nights booked) - **SaaS Productivity**: Weekly active users completing core action (Slack: messages sent in channels with 3+ participants) - **Subscription Media**: Weekly engaged time (Spotify: listening hours) - **E-commerce**: Weekly purchases from repeat customers - **Developer Tool**: Weekly API calls in production (Stripe, Twilio) ### North Star Framework Your North Star has 3-5 input metrics that drive it: North Star: "Weekly active teams completing core workflow" Input metrics: 1. **Breadth**: New teams activated this week 2. **Depth**: Core workflows completed per team per week 3. **Frequency**: Days active per week per team 4. **Efficiency**: Time to complete core workflow 5. **Quality**: Workflow completion rate (started vs finished) Each team owns an input metric. Together they drive the North Star. ``` ### AARRR Pirate Metrics ```markdown ## Pirate Metrics Funnel ### Acquisition — How do users find you? Metrics: Website visitors, signup rate, CAC, channel attribution Questions: Which channels bring the highest-quality users? What's the CAC by channel? ### Activation — Do users experience core value? Metrics: Onboarding completion rate, time-to-first-value, "aha moment" reached Questions: What % of signups complete onboarding? What's the "aha moment"? ### Retention — Do users come back? Metrics: D1/D7/D30 retention, weekly active rate, churn rate Questions: What does the retention curve look like? Where does it flatten? ### Revenue — Do users pay? Metrics: Conversion rate (free → paid), ARPU, LTV, expansion revenue Questions: What triggers the upgrade? What's the payback period on CAC? ### Referral — Do users invite others? Metrics: Viral coefficient, referral rate, NPS, organic share rate Questions: Do users invite others? What's the average referrals per user? ### Identify Your Bottleneck Measure each stage. The stage with the worst drop-off is your bottleneck. Focus there — don't optimize acquisition if nobody activates. ``` ### Retention Analysis ```markdown ## Analyze Retention ### Retention Curve Plot % of users who return on Day 1, Day 7, Day 14, Day 30. - **Good**: Curve flattens (users who stay past day 7 stick around) - **Bad**: Curve approaches zero (no habitual users) ### Cohort Analysis Compare retention across weekly or monthly sign-up cohorts: Cohort | Week 1 | Week 2 | Week 3 | Week 4 | Week 8 Jan 1-7 | 100% | 42% | 31% | 28% | 25% Jan 8-14 | 100% | 45% | 35% | 32% | 29% Jan 15-21 | 100% | 51% | 40% | 38% | — Reading this: Jan 15-21 cohort retains better → what changed? Check: product changes, marketing channel mix, seasonality. ### Retention by Segment Break retention by: - **Acquisition channel**: Do SEO users retain better than paid? - **Plan tier**: Do Pro users retain better than Free? - **Activation actions**: Did they complete onboarding? Use feature X? - **Company size**: Do small teams churn more than large? ### Find the "Aha Moment" The aha moment is the action that predicts retention. Method: 1. List all actions a user can take in the first week 2. For each action, split users who did it vs didn't 3. Compare 30-day retention between the two groups 4. The action with the biggest retention gap is your aha moment Example findings: - Users who create 3+ projects in week 1: 72% D30 retention - Users who create 0-2 projects: 18% D30 retention → Aha moment: creating the 3rd project → Action: optimize onboarding to drive users to create 3 projects ``` ### A/B Testing ```markdown ## Run A/B Tests ### Before You Test 1. Define the hypothesis: "Changing X will improve Y by Z%" 2. Choose the primary metric (one!) 3. Calculate sample size: use a power calculator - Baseline conversion: current rate - Minimum detectable effect: smallest change worth detecting - Statistical power: 80% (standard) - Significance level: 95% (standard) 4. Estimate duration: sample size ÷ daily traffic ### Common Mistakes - ❌ Stopping early because results "look significant" - ❌ Testing too many variants (dilutes sample size) - ❌ Changing the test mid-experiment - ❌ Using the wrong metric (vanity vs actionable) - ❌ Not segmenting results (overall flat, but +30% for mobile users) ### Interpreting Results **Statistical significance ≠ practical significance** - 2% lift with p<0.05 might not be worth the engineering cost - Consider the confidence interval, not just the point estimate - Always check for novelty effects (run for 2+ full weeks) ### Sequential Testing For faster decisions, use sequential testing: - Define stopping rules before starting - Check daily, but only stop if the boundary is crossed - Avoids the "peeking problem" of traditional tests ### Post-Test - Document: hypothesis, variant, result, learnings - If winner: roll out to 100% - If flat: was the sample size large enough? Consider the learning. - If loser: document why and share the learning ``` ### Funnel Analysis ```markdown ## Build and Optimize Funnels ### Define the Funnel Map every step from entry to conversion: E-commerce: Visit → Product View → Add to Cart → Checkout → Purchase SaaS: Visit → Signup → Onboarding → Activation → Upgrade ### Measure Drop-off Step | Users | Conversion | Drop-off Visit | 10,000 | 100% | — Signup | 2,500 | 25% | 75% Complete onboarding | 1,000 | 40% | 60% Reach aha moment | 600 | 60% | 40% Still active at D30 | 300 | 50% | 50% Upgrade to paid | 90 | 30% | 70% ### Find the Biggest Lever The step with the biggest absolute drop-off has the most room for improvement. In the example above: 75% drop at Signup → fix the landing page first. ### Optimize Each Step - **Visit → Signup**: Value proposition clarity, social proof, friction reduction - **Signup → Onboarding**: Reduce form fields, add progress indicators - **Onboarding → Activation**: Guide to aha moment, remove unnecessary steps - **Activation → Retention**: Habit loops, notifications, value reminders - **Retention → Revenue**: Upgrade triggers, usage limits, feature gating ``` ## Guidelines 1. **One North Star** — Align the entire product team around a single metric that reflects customer value 2. **Input metrics per team** — Each team owns an input metric that drives the North Star; this creates autonomy with alignment 3. **Retention before acquisition** — Fix retention first; acquiring users into a leaky bucket wastes money 4. **Cohort everything** — Never look at aggregate metrics; always break by cohort (signup week, plan, channel) to find patterns 5. **Find the aha moment** — Identify the action that predicts retention; then optimize onboarding to drive users to that action 6. **A/B test with discipline** — Pre-register hypothesis, sample size, and duration; never peek and stop early 7. **Instrument early** — Add analytics from day one; retroactive instrumentation means lost data you can never recover 8. **Metrics are questions** — A metric tells you WHAT happened; you need qualitative research (interviews, session recordings) to understand WHY
Related Skills
collecting-infrastructure-metrics
Collect comprehensive infrastructure performance metrics across compute, storage, network, containers, load balancers, and databases. Use when monitoring system performance or troubleshooting infrastructure issues. Trigger with phrases like "collect infrastructure metrics", "monitor server performance", or "track system resources".
aggregating-performance-metrics
Aggregate and centralize performance metrics from applications, systems, databases, caches, and services. Use when consolidating monitoring data from multiple sources. Trigger with phrases like "aggregate metrics", "centralize monitoring", or "collect performance data".
gtm-technical-product-pricing
Pricing strategy for technical products. Use when choosing usage-based vs seat-based, designing freemium thresholds, structuring enterprise pricing conversations, deciding when to raise prices, or using price as a positioning signal.
gtm-product-led-growth
Build self-serve acquisition and expansion motions. Use when deciding PLG vs sales-led, optimizing activation, driving freemium conversion, building growth equations, or recognizing when product complexity demands human touch. Includes the parallel test where sales-led won 10x on revenue.
dataverse-python-production-code
Generate production-ready Python code using Dataverse SDK with error handling, optimization, and best practices
copilot-usage-metrics
Retrieve and display GitHub Copilot usage metrics for organizations and enterprises using the GitHub CLI and REST API.
using-dbt-for-analytics-engineering
Builds and modifies dbt models, writes SQL transformations using ref() and source(), creates tests, and validates results with dbt show. Use when doing any dbt work - building or modifying models, debugging errors, exploring unfamiliar data sources, writing tests, or evaluating impact of changes.
product-analysis
Multi-path parallel product analysis with cross-model test-time compute scaling. Spawns parallel agents (Claude Code agent teams + Codex CLI) to explore product from multiple perspectives, then synthesizes findings into actionable optimization plans. Can invoke competitors-analysis for competitive benchmarking. Use when "product audit", "self-review", "发布前审查", "产品分析", "analyze our product", "UX audit", or "信息架构审计".
apify-content-analytics
Track engagement metrics, measure campaign ROI, and analyze content performance across Instagram, Facebook, YouTube, and TikTok.
../../../marketing-skill/x-twitter-growth/SKILL.md
No description provided.
saas-metrics-coach
SaaS financial health advisor. Use when a user shares revenue or customer numbers, or mentions ARR, MRR, churn, LTV, CAC, NRR, or asks how their SaaS business is doing.
../../../product-team/roadmap-communicator/SKILL.md
No description provided.