analytics-metrics-kpi — community analytics-metrics-kpi, custom-plugin-product-manager, community, ide skills, Claude Code, Cursor, Windsurf

v2.0.0

About this Skill

Perfect for Data-Driven Agents needing advanced analytics and metrics capabilities for informed decision-making. Master metrics definition, KPI tracking, dashboarding, A/B testing, and data-driven decision making. Use data to guide product decisions.

pluginagentmarketplace pluginagentmarketplace
[0]
[0]
Updated: 3/12/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reference-Only Page Review Score: 7/11

This page remains useful for operators, but Killer-Skills treats it as reference material instead of a primary organic landing page.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Locale and body language aligned
Review Score
7/11
Quality Score
41
Canonical Locale
en
Detected Body Locale
en

Perfect for Data-Driven Agents needing advanced analytics and metrics capabilities for informed decision-making. Master metrics definition, KPI tracking, dashboarding, A/B testing, and data-driven decision making. Use data to guide product decisions.

Core Value

Empowers agents to define meaningful metrics, build dashboards, and run experiments using a metrics framework, driving data-driven decisions with North Star Metrics and Acquisition-to-Revenue analysis.

Ideal Agent Persona

Perfect for Data-Driven Agents needing advanced analytics and metrics capabilities for informed decision-making.

Capabilities Granted for analytics-metrics-kpi

Defining North Star Metrics for product success
Building dashboards for data visualization and insights
Running experiments to measure product improvements

! Prerequisites & Limits

  • Requires understanding of business success metrics
  • Driven by product improvements and revenue indicators

Why this page is reference-only

  • - The underlying skill quality score is below the review floor.

Source Boundary

The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is analytics-metrics-kpi?

Perfect for Data-Driven Agents needing advanced analytics and metrics capabilities for informed decision-making. Master metrics definition, KPI tracking, dashboarding, A/B testing, and data-driven decision making. Use data to guide product decisions.

How do I install analytics-metrics-kpi?

Run the command: npx killer-skills add pluginagentmarketplace/custom-plugin-product-manager/analytics-metrics-kpi. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for analytics-metrics-kpi?

Key use cases include: Defining North Star Metrics for product success, Building dashboards for data visualization and insights, Running experiments to measure product improvements.

Which IDEs are compatible with analytics-metrics-kpi?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for analytics-metrics-kpi?

Requires understanding of business success metrics. Driven by product improvements and revenue indicators.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add pluginagentmarketplace/custom-plugin-product-manager/analytics-metrics-kpi. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use analytics-metrics-kpi immediately in the current project.

! Reference-Only Mode

This page remains useful for installation and reference, but Killer-Skills no longer treats it as a primary indexable landing page. Read the review above before relying on the upstream repository instructions.

Imported Repository Instructions

The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.

Supporting Evidence

analytics-metrics-kpi

Install analytics-metrics-kpi, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly
Imported Repository Instructions
The section below is supporting source material from the upstream repository. Use the Killer-Skills review above as the primary decision layer.
Supporting Evidence

Analytics & Metrics Skill

Become data-driven. Define meaningful metrics, build dashboards, run experiments, and make decisions based on data, not intuition.

Metrics Framework (Acquisition → Revenue)

North Star Metric

Definition: One metric that best captures the value your product delivers.

Characteristics:

  • Directly tied to business success
  • Driven by product improvements
  • Leading indicator of revenue
  • Understandable to whole company

Examples:

  • Slack: Daily Active Users (DAU)
  • Airbnb: Booked Nights
  • YouTube: Watch Time
  • Uber: Rides Completed
  • Stripe: Payment Volume Processed

Funnel Metrics (Acquisition)

Total Visitors: 100,000/month
↓ 20% conversion
Free Signups: 20,000
↓ 10% free-to-paid
Paid Customers: 2,000

CAC: $50 (marketing + sales spend / customers acquired)
LTCAC: $100 (all customer acquisition costs)

Metrics to Track:

  • Traffic - Total visitors to website/app
  • Signup Rate - % who sign up (target: 10-15%)
  • Free-to-Paid Conversion - % free users who pay (target: 2-5%)
  • CAC - Cost per acquired customer
  • CAC Payback - Months to recover CAC from revenue (target: < 12 months)

Activation Metrics

Goal: New users become active users

Free Signups: 2,000
↓ 30% onboard successfully
Activated: 600
↓ 60% remain active Day 7
Day 7 Active: 360

Metrics to Track:

  • Onboarding Completion Rate - % who complete setup (target: 50-80%)
  • Time to First Value - Hours to first successful use
  • Feature Adoption - % who try key features
  • Day 1/7/30 Retention - % active those days (target: 40/25/15)

Engagement Metrics

Goal: Users regularly use product

Daily/Monthly Metrics:

  • DAU/MAU - Daily/Monthly Active Users
  • DAU/MAU Ratio - Stickiness (target: 20-30%)
  • Feature Usage - % using key features
  • Session Length - Minutes per session
  • Session Frequency - Times per week

Cohort Analysis Example:

Jan Cohort (1,000 signups):
- Day 1: 600 active (60%)
- Day 7: 360 active (36%)
- Day 30: 180 active (18%)
- Month 3: 90 active (9%)

Feb Cohort (1,500 signups):
- Day 1: 1050 active (70%) ← Improving!
- Day 7: 630 active (42%)
- Day 30: 300 active (20%)

Retention Metrics

Goal: Users stay and continue paying

Month 1: 1,000 customers
Month 2: 900 active (90% retained)
Month 3: 810 active (90% of month 2)
Month 12: 314 active (31% annual retention)

Churn Rate: % lost each period

  • Monthly churn: (Customers Lost / Month Start) × 100
  • Annual churn: 1 - (Ending / Starting)
  • Target for SaaS: < 5% monthly churn

NPS (Net Promoter Score)

  • Question: "How likely to recommend (0-10)?"
  • Score = % Promoters (9-10) - % Detractors (0-6)
  • Range: -100 to +100
  • Target: 50+ (world-class)

Revenue Metrics

Monthly Recurring Revenue (MRR)

MRR = (Total paid customers) × (average subscription price)
Growth MRR = New MRR + Expansion MRR - Churn MRR

Annual Run Rate (ARR)

ARR = MRR × 12

Average Revenue Per User (ARPU)

ARPU = MRR / Total Users

Customer Lifetime Value (LTV)

LTV = (ARPU × Gross Margin %) / Monthly Churn %

Example:
ARPU: $100
Gross Margin: 80%
Monthly Churn: 5%
LTV = ($100 × 80%) / 5% = $1,600

If CAC = $400: LTV/CAC = 4x ✓ (target: 3x+)

Dashboard Architecture

Executive Dashboard (C-Level)

Weekly Updates:

  • MRR / ARR (vs target, vs month ago)
  • New customers (weekly, monthly)
  • Churn rate (%)
  • NPS score
  • Engagement (DAU, MAU)
  • Key initiatives status

Frequency: Weekly

Product Dashboard (Product Team)

Daily/Weekly:

  • Funnel metrics (signup → paid)
  • Feature adoption
  • Engagement metrics
  • User feedback score
  • A/B test results
  • Support ticket volume

Frequency: Daily updates

Financial Dashboard (Finance/Operations)

Monthly:

  • MRR / ARR
  • Customer acquisition cost
  • Customer lifetime value
  • Gross margin
  • CAC payback period
  • Revenue by segment
  • Churn by cohort

Frequency: Monthly

Health Dashboard (Operations)

Realtime:

  • System uptime (%)
  • Error rate (%)
  • Response time (p95)
  • Database performance
  • Support ticket response time
  • Support backlog

Frequency: Realtime/hourly

A/B Testing (Experimentation)

Test Planning

Hypothesis: "If we change X, then Y will improve, because Z"

Example: "If we move signup button above the fold, then conversion will improve 15%, because users won't scroll."

Test Structure

Experiment Design:

  • Control: Keep current version
  • Treatment: New version
  • Sample size: Enough users to be statistical
  • Duration: 2-4 weeks minimum
  • Metric: Clear success metric

Statistical Significance

Confidence Level: 95% (industry standard)

  • Means 5% chance of false positive
  • Need enough samples (typically 1000-10K per variant)
  • Use calculator for exact sample size

P-Value: Probability result is random chance

  • P < 0.05: Statistically significant
  • P > 0.05: Not significant, inconclusive

Example A/B Test

Hypothesis: Moving signup button above fold increases conversion 15%

Setup:

  • Control: Current design
  • Treatment: Button moved above fold
  • Success metric: Conversion rate (signup / visit)
  • Sample size: 10,000 users per variant
  • Duration: 2 weeks
  • Confidence: 95%

Results:

  • Control: 2.0% conversion (200 signups from 10K visitors)
  • Treatment: 2.8% conversion (280 signups from 10K visitors)
  • Improvement: 40% increase (0.8% / 2% = 40%)
  • P-value: 0.02 (statistically significant!)
  • Decision: SHIP IT - Roll out to 100%

Test Ideas by Priority

High Priority (Start Here):

  • Signup flow optimization (biggest funnel)
  • Onboarding experience
  • Pricing page clarity
  • Feature discoverability

Medium Priority:

  • UI copy optimization
  • CTA button colors
  • Email subject lines
  • Notification triggers

Low Priority:

  • Micro-copy tweaks
  • Animation effects
  • Color scheme changes

Metric Pitfalls to Avoid

Vanity Metrics

❌ "We have 1M page views!" ✓ "We have 50K daily active users, growing 10% monthly"

Actionable vs Non-Actionable

❌ "User satisfaction increased" (what changed?) ✓ "Onboarding completion rate 65% → 78% (↑20%)" (clear action)

Correlation vs Causation

❌ "Ice cream sales correlate with drownings" ✓ Understand actual causation, not just correlation

Look-Alike Metrics

❌ Track MRR but not Customer LTV (can grow MRR by spending more on acquisition) ✓ Track both acquisition efficiency AND retention

Metrics Review Cadence

Daily:

  • System uptime
  • Error rates
  • Support response time

Weekly:

  • Funnel metrics
  • Feature adoption
  • Key engagement metrics
  • Test results

Monthly:

  • Revenue metrics
  • Cohort analysis
  • Churn breakdown
  • LTV/CAC trends

Quarterly:

  • Strategic metric review
  • Long-term trend analysis
  • Metric changes needed

Troubleshooting

Yaygın Hatalar & Çözümler

HataOlası SebepÇözüm
Vanity metrics focusWrong KPI selectionNorth Star alignment
Inconclusive A/B testLow sample sizeExtend duration
Data inconsistencyMultiple sourcesSingle source of truth
Dashboard unusedToo complexSimplify to 5-7 KPIs

Debug Checklist

[ ] North Star metric defined mi?
[ ] Metrics business goals'a aligned mi?
[ ] Data collection accurate mi?
[ ] Dashboard refreshed mi?
[ ] A/B test sample sufficient mi?
[ ] Statistical significance achieved mi?

Recovery Procedures

  1. Data Quality Issues → Flag affected metrics, exclude
  2. Inconclusive A/B → Extend test duration
  3. Misleading Metrics → Add context/segmentation

Master data-driven decision making and grow faster!

Related Skills

Looking for an alternative to analytics-metrics-kpi or another community skill for your workflow? Explore these related open-source skills.

View All

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
AI

widget-generator

Logo of f
f

Generate customizable widget plugins for the prompts.chat feed system

149.6k
0
AI

flags

Logo of vercel
vercel

The React Framework

138.4k
0
Browser

pr-review

Logo of pytorch
pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

98.6k
0
Developer