Claude Plugin

experiment-tracker

PROACTIVELY use this agent when experiments are started, modified, or when results need analysis. This agent specializes in tracking A/B tests, feature experiments, and iterative improvements within the 6-day development cycle. Should be triggered automatically when experimental code paths or feature flags are introduced. Examples:\n\n<example>\nContext: When implementing feature flags or A/B test variants\nuser: "Add a feature flag to test the new onboarding flow"\nassistant: "I've implemented the feature flag for the onboarding flow A/B test. Let me use the experiment-tracker agent to properly document this experiment and set up tracking."\n<commentary>\nEvery feature flag represents an experiment that needs proper tracking and success metrics.\n</commentary>\n</example>\n\n<example>\nContext: After deploying experimental features\nuser: "The new viral sharing feature is now live for 10% of users"\nassistant: "Great! The viral sharing feature is deployed. I'll use the experiment-tracker agent to monitor its performance and document initial results."\n<commentary>\nLive experiments need immediate tracking setup to capture valuable data from the start.\n</commentary>\n</example>\n\n<example>\nContext: When reaching experiment milestones\nuser: "It's been a week since we launched the TikTok integration test"\nassistant: "Perfect timing for analysis. Let me use the experiment-tracker agent to compile the week-one results and determine if we should expand the test."\n<commentary>\nRegular experiment checkpoints ensure data-driven decisions about feature rollouts.\n</commentary>\n</example>\n\n<example>\nContext: Before making product decisions\nuser: "Should we keep the AI avatar feature or remove it?"\nassistant: "Let me analyze the experiment data to inform this decision. I'll use the experiment-tracker agent to review all metrics and user feedback for the AI avatar feature test."\n<commentary>\nProduct decisions should be backed by experiment data, not gut feelings.\n</commentary>\n</example>

Editor's Note

PROACTIVELY use this agent when experiments are started, modified, or when results need analysis. This agent specializes in tracking A/B tests, feature experiments, and iterative improvements within the 6-day development cycle. Should be triggered...

Plugin Overview

This item is backed by a plugin manifest rather than a `SKILL.md` file, so the most useful fields are surfaced here first.

Plugin Name

experiment-tracker

Version

1.0.0

Author

Michael Galpert

Manifest Description

PROACTIVELY use this agent when experiments are started, modified, or when results need analysis. This agent specializes in tracking A/B tests, feature experiments, and iterative improvements within the 6-day development cycle. Should be triggered automatically when experimental code paths or feature flags are introduced. Examples:\n\n<example>\nContext: When implementing feature flags or A/B test variants\nuser: "Add a feature flag to test the new onboarding flow"\nassistant: "I've implemented the feature flag for the onboarding flow A/B test. Let me use the experiment-tracker agent to properly document this experiment and set up tracking."\n<commentary>\nEvery feature flag represents an experiment that needs proper tracking and success metrics.\n</commentary>\n</example>\n\n<example>\nContext: After deploying experimental features\nuser: "The new viral sharing feature is now live for 10% of users"\nassistant: "Great! The viral sharing feature is deployed. I'll use the experiment-tracker agent to monitor its performance and document initial results."\n<commentary>\nLive experiments need immediate tracking setup to capture valuable data from the start.\n</commentary>\n</example>\n\n<example>\nContext: When reaching experiment milestones\nuser: "It's been a week since we launched the TikTok integration test"\nassistant: "Perfect timing for analysis. Let me use the experiment-tracker agent to compile the week-one results and determine if we should expand the test."\n<commentary>\nRegular experiment checkpoints ensure data-driven decisions about feature rollouts.\n</commentary>\n</example>\n\n<example>\nContext: Before making product decisions\nuser: "Should we keep the AI avatar feature or remove it?"\nassistant: "Let me analyze the experiment data to inform this decision. I'll use the experiment-tracker agent to review all metrics and user feedback for the AI avatar feature test."\n<commentary>\nProduct decisions should be backed by experiment data, not gut feelings.\n</commentary>\n</example>

Raw Manifest

The structured plugin fields above are derived from the same upstream manifest shown below.

{ "name": "experiment-tracker", "description": "PROACTIVELY use this agent when experiments are started, modified, or when results need analysis. This agent specializes in tracking A/B tests, feature experiments, and iterative improvements within the 6-day development cycle. Should be triggered automatically when experimental code paths or feature flags are introduced. Examples:\\n\\n<example>\\nContext: When implementing feature flags or A/B test variants\\nuser: \"Add a feature flag to test the new onboarding flow\"\\nassistant: \"I've implemented the feature flag for the onboarding flow A/B test. Let me use the experiment-tracker agent to properly document this experiment and set up tracking.\"\\n<commentary>\\nEvery feature flag represents an experiment that needs proper tracking and success metrics.\\n</commentary>\\n</example>\\n\\n<example>\\nContext: After deploying experimental features\\nuser: \"The new viral sharing feature is now live for 10% of users\"\\nassistant: \"Great! The viral sharing feature is deployed. I'll use the experiment-tracker agent to monitor its performance and document initial results.\"\\n<commentary>\\nLive experiments need immediate tracking setup to capture valuable data from the start.\\n</commentary>\\n</example>\\n\\n<example>\\nContext: When reaching experiment milestones\\nuser: \"It's been a week since we launched the TikTok integration test\"\\nassistant: \"Perfect timing for analysis. Let me use the experiment-tracker agent to compile the week-one results and determine if we should expand the test.\"\\n<commentary>\\nRegular experiment checkpoints ensure data-driven decisions about feature rollouts.\\n</commentary>\\n</example>\\n\\n<example>\\nContext: Before making product decisions\\nuser: \"Should we keep the AI avatar feature or remove it?\"\\nassistant: \"Let me analyze the experiment data to inform this decision. I'll use the experiment-tracker agent to review all metrics and user feedback for the AI avatar feature test.\"\\n<commentary>\\nProduct decisions should be backed by experiment data, not gut feelings.\\n</commentary>\\n</example>", "version": "1.0.0", "author": { "name": "Michael Galpert" }, "homepage": "https://github.com/ccplugins/awesome-claude-code-plugins/tree/main/plugins/experiment-tracker" }

Related Items

Claude Plugin

prd-specialist

Use this agent when you need to create comprehensive Product Requirements Documents (PRDs) that combine business strategy, technical architecture, and user research. Examples: <example>Context: The user needs to create a PRD for a new feature or product launch. user: "I need to create a PRD for our new user authentication system that will support SSO and multi-factor authentication" assistant: "I'll use the prd-specialist agent to create a comprehensive PRD that covers the strategic foundation, technical requirements, and implementation blueprint for your authentication system."</example> <example>Context: The user is planning a major product initiative and needs strategic documentation. user: "We're launching a mobile app for our e-commerce platform and need a detailed PRD to guide development" assistant: "Let me engage the prd-specialist agent to develop a thorough PRD that includes market analysis, user research integration, technical architecture, and implementation roadmap for your mobile app initiative."</example>

Claude Plugin

problem-solver-specialist

Universal expert problem-solving agent specializing in complex debugging, mysterious runtime behavior, integration issues, and multi-layered technical challenges across any technology stack or project type. Uses advanced research methodologies including GitHub issues mining, Perplexity deep research, community solutions validation, browser automation testing, and multi-source documentation analysis. Proactively use for intricate bugs, cryptic error messages, performance anomalies, framework integration problems, legacy system compatibility issues, and any technical challenges requiring deep investigation across multiple knowledge sources. Project-agnostic and universally applicable.

Claude Plugin

support-responder

Use this agent when handling customer support inquiries, creating support documentation, setting up automated responses, or analyzing support patterns. This agent excels at maintaining high-quality support across all studio projects while identifying product improvement opportunities. Examples:\n\n<example>\nContext: Setting up support for a new app launch

Claude Plugin

test-results-analyzer

Use this agent for analyzing test results, synthesizing test data, identifying trends, and generating quality metrics reports. This agent specializes in turning raw test data into actionable insights that drive quality improvements. Examples:\n\n<example>\nContext: Analyzing test suite results

Deploy agents, MCP servers, and backends fast logo

Railway - Deploy agents and MCP servers fast

Try Railway