Build features guided by data insights, A/B testing, and continuous measurement using specialized agents for analysis, implementation, and experimentation.
Add this skill
npx mdskills install sickn33/data-engineering-data-driven-featureComprehensive data-driven workflow with detailed multi-phase orchestration and specialized agents
1---2name: data-engineering-data-driven-feature3description: "Build features guided by data insights, A/B testing, and continuous measurement using specialized agents for analysis, implementation, and experimentation."4---56# Data-Driven Feature Development78Build features guided by data insights, A/B testing, and continuous measurement using specialized agents for analysis, implementation, and experimentation.910[Extended thinking: This workflow orchestrates a comprehensive data-driven development process from initial data analysis and hypothesis formulation through feature implementation with integrated analytics, A/B testing infrastructure, and post-launch analysis. Each phase leverages specialized agents to ensure features are built based on data insights, properly instrumented for measurement, and validated through controlled experiments. The workflow emphasizes modern product analytics practices, statistical rigor in testing, and continuous learning from user behavior.]1112## Use this skill when1314- Working on data-driven feature development tasks or workflows15- Needing guidance, best practices, or checklists for data-driven feature development1617## Do not use this skill when1819- The task is unrelated to data-driven feature development20- You need a different domain or tool outside this scope2122## Instructions2324- Clarify goals, constraints, and required inputs.25- Apply relevant best practices and validate outcomes.26- Provide actionable steps and verification.27- If detailed examples are required, open `resources/implementation-playbook.md`.2829## Phase 1: Data Analysis and Hypothesis Formation3031### 1. Exploratory Data Analysis32- Use Task tool with subagent_type="machine-learning-ops::data-scientist"33- Prompt: "Perform exploratory data analysis for feature: $ARGUMENTS. Analyze existing user behavior data, identify patterns and opportunities, segment users by behavior, and calculate baseline metrics. Use modern analytics tools (Amplitude, Mixpanel, Segment) to understand current user journeys, conversion funnels, and engagement patterns."34- Output: EDA report with visualizations, user segments, behavioral patterns, baseline metrics3536### 2. Business Hypothesis Development37- Use Task tool with subagent_type="business-analytics::business-analyst"38- Context: Data scientist's EDA findings and behavioral patterns39- Prompt: "Formulate business hypotheses for feature: $ARGUMENTS based on data analysis. Define clear success metrics, expected impact on key business KPIs, target user segments, and minimum detectable effects. Create measurable hypotheses using frameworks like ICE scoring or RICE prioritization."40- Output: Hypothesis document, success metrics definition, expected ROI calculations4142### 3. Statistical Experiment Design43- Use Task tool with subagent_type="machine-learning-ops::data-scientist"44- Context: Business hypotheses and success metrics45- Prompt: "Design statistical experiment for feature: $ARGUMENTS. Calculate required sample size for statistical power, define control and treatment groups, specify randomization strategy, and plan for multiple testing corrections. Consider Bayesian A/B testing approaches for faster decision making. Design for both primary and guardrail metrics."46- Output: Experiment design document, power analysis, statistical test plan4748## Phase 2: Feature Architecture and Analytics Design4950### 4. Feature Architecture Planning51- Use Task tool with subagent_type="data-engineering::backend-architect"52- Context: Business requirements and experiment design53- Prompt: "Design feature architecture for: $ARGUMENTS with A/B testing capability. Include feature flag integration (LaunchDarkly, Split.io, or Optimizely), gradual rollout strategy, circuit breakers for safety, and clean separation between control and treatment logic. Ensure architecture supports real-time configuration updates."54- Output: Architecture diagrams, feature flag schema, rollout strategy5556### 5. Analytics Instrumentation Design57- Use Task tool with subagent_type="data-engineering::data-engineer"58- Context: Feature architecture and success metrics59- Prompt: "Design comprehensive analytics instrumentation for: $ARGUMENTS. Define event schemas for user interactions, specify properties for segmentation and analysis, design funnel tracking and conversion events, plan cohort analysis capabilities. Implement using modern SDKs (Segment, Amplitude, Mixpanel) with proper event taxonomy."60- Output: Event tracking plan, analytics schema, instrumentation guide6162### 6. Data Pipeline Architecture63- Use Task tool with subagent_type="data-engineering::data-engineer"64- Context: Analytics requirements and existing data infrastructure65- Prompt: "Design data pipelines for feature: $ARGUMENTS. Include real-time streaming for live metrics (Kafka, Kinesis), batch processing for detailed analysis, data warehouse integration (Snowflake, BigQuery), and feature store for ML if applicable. Ensure proper data governance and GDPR compliance."66- Output: Pipeline architecture, ETL/ELT specifications, data flow diagrams6768## Phase 3: Implementation with Instrumentation6970### 7. Backend Implementation71- Use Task tool with subagent_type="backend-development::backend-architect"72- Context: Architecture design and feature requirements73- Prompt: "Implement backend for feature: $ARGUMENTS with full instrumentation. Include feature flag checks at decision points, comprehensive event tracking for all user actions, performance metrics collection, error tracking and monitoring. Implement proper logging for experiment analysis."74- Output: Backend code with analytics, feature flag integration, monitoring setup7576### 8. Frontend Implementation77- Use Task tool with subagent_type="frontend-mobile-development::frontend-developer"78- Context: Backend APIs and analytics requirements79- Prompt: "Build frontend for feature: $ARGUMENTS with analytics tracking. Implement event tracking for all user interactions, session recording integration if applicable, performance metrics (Core Web Vitals), and proper error boundaries. Ensure consistent experience between control and treatment groups."80- Output: Frontend code with analytics, A/B test variants, performance monitoring8182### 9. ML Model Integration (if applicable)83- Use Task tool with subagent_type="machine-learning-ops::ml-engineer"84- Context: Feature requirements and data pipelines85- Prompt: "Integrate ML models for feature: $ARGUMENTS if needed. Implement online inference with low latency, A/B testing between model versions, model performance tracking, and automatic fallback mechanisms. Set up model monitoring for drift detection."86- Output: ML pipeline, model serving infrastructure, monitoring setup8788## Phase 4: Pre-Launch Validation8990### 10. Analytics Validation91- Use Task tool with subagent_type="data-engineering::data-engineer"92- Context: Implemented tracking and event schemas93- Prompt: "Validate analytics implementation for: $ARGUMENTS. Test all event tracking in staging, verify data quality and completeness, validate funnel definitions, ensure proper user identification and session tracking. Run end-to-end tests for data pipeline."94- Output: Validation report, data quality metrics, tracking coverage analysis9596### 11. Experiment Setup97- Use Task tool with subagent_type="cloud-infrastructure::deployment-engineer"98- Context: Feature flags and experiment design99- Prompt: "Configure experiment infrastructure for: $ARGUMENTS. Set up feature flags with proper targeting rules, configure traffic allocation (start with 5-10%), implement kill switches, set up monitoring alerts for key metrics. Test randomization and assignment logic."100- Output: Experiment configuration, monitoring dashboards, rollout plan101102## Phase 5: Launch and Experimentation103104### 12. Gradual Rollout105- Use Task tool with subagent_type="cloud-infrastructure::deployment-engineer"106- Context: Experiment configuration and monitoring setup107- Prompt: "Execute gradual rollout for feature: $ARGUMENTS. Start with internal dogfooding, then beta users (1-5%), gradually increase to target traffic. Monitor error rates, performance metrics, and early indicators. Implement automated rollback on anomalies."108- Output: Rollout execution, monitoring alerts, health metrics109110### 13. Real-time Monitoring111- Use Task tool with subagent_type="observability-monitoring::observability-engineer"112- Context: Deployed feature and success metrics113- Prompt: "Set up comprehensive monitoring for: $ARGUMENTS. Create real-time dashboards for experiment metrics, configure alerts for statistical significance, monitor guardrail metrics for negative impacts, track system performance and error rates. Use tools like Datadog, New Relic, or custom dashboards."114- Output: Monitoring dashboards, alert configurations, SLO definitions115116## Phase 6: Analysis and Decision Making117118### 14. Statistical Analysis119- Use Task tool with subagent_type="machine-learning-ops::data-scientist"120- Context: Experiment data and original hypotheses121- Prompt: "Analyze A/B test results for: $ARGUMENTS. Calculate statistical significance with confidence intervals, check for segment-level effects, analyze secondary metrics impact, investigate any unexpected patterns. Use both frequentist and Bayesian approaches. Account for multiple testing if applicable."122- Output: Statistical analysis report, significance tests, segment analysis123124### 15. Business Impact Assessment125- Use Task tool with subagent_type="business-analytics::business-analyst"126- Context: Statistical analysis and business metrics127- Prompt: "Assess business impact of feature: $ARGUMENTS. Calculate actual vs expected ROI, analyze impact on key business metrics, evaluate cost-benefit including operational overhead, project long-term value. Make recommendation on full rollout, iteration, or rollback."128- Output: Business impact report, ROI analysis, recommendation document129130### 16. Post-Launch Optimization131- Use Task tool with subagent_type="machine-learning-ops::data-scientist"132- Context: Launch results and user feedback133- Prompt: "Identify optimization opportunities for: $ARGUMENTS based on data. Analyze user behavior patterns in treatment group, identify friction points in user journey, suggest improvements based on data, plan follow-up experiments. Use cohort analysis for long-term impact."134- Output: Optimization recommendations, follow-up experiment plans135136## Configuration Options137138```yaml139experiment_config:140 min_sample_size: 10000141 confidence_level: 0.95142 runtime_days: 14143 traffic_allocation: "gradual" # gradual, fixed, or adaptive144145analytics_platforms:146 - amplitude147 - segment148 - mixpanel149150feature_flags:151 provider: "launchdarkly" # launchdarkly, split, optimizely, unleash152153statistical_methods:154 - frequentist155 - bayesian156157monitoring:158 - real_time_metrics: true159 - anomaly_detection: true160 - automatic_rollback: true161```162163## Success Criteria164165- **Data Coverage**: 100% of user interactions tracked with proper event schema166- **Experiment Validity**: Proper randomization, sufficient statistical power, no sample ratio mismatch167- **Statistical Rigor**: Clear significance testing, proper confidence intervals, multiple testing corrections168- **Business Impact**: Measurable improvement in target metrics without degrading guardrail metrics169- **Technical Performance**: No degradation in p95 latency, error rates below 0.1%170- **Decision Speed**: Clear go/no-go decision within planned experiment runtime171- **Learning Outcomes**: Documented insights for future feature development172173## Coordination Notes174175- Data scientists and business analysts collaborate on hypothesis formation176- Engineers implement with analytics as first-class requirement, not afterthought177- Feature flags enable safe experimentation without full deployments178- Real-time monitoring allows for quick iteration and rollback if needed179- Statistical rigor balanced with business practicality and speed to market180- Continuous learning loop feeds back into next feature development cycle181182Feature to develop with data-driven approach: $ARGUMENTS183
Full transparency — inspect the skill content before installing.