Something went wrong!
Hang in there while we get back on track
Aggregate Data Analysis: Feature Usage Activation
Analysis of 500 users tracked by features_used, days_active, and shared_content. Identifies 'Use 6+ features in first week' as primary metric with composite scoring formula.
Input
What you provide to the skill
Analyze my aggregate data: 500 users tracked by features_used, days_active, and shared_content. Need activation scoring and thresholds.
Output
What the skill generates for you
Activation Metric Discovery Analysis
Dataset Overview
- Total Users: 500
- Overall Retention (30d): 70.0%
- Metrics Tracked: features_used, days_active, shared_content
Correlation Analysis
| Metric | Pearson Correlation | Recommended Threshold | Retention if High | Retention if Low | Strength |
|---|---|---|---|---|---|
| features_used | 0.95 | ≥ 6 | 97% (6+ features) | 18% (<4 features) | Strong |
| days_active | 0.93 | ≥ 5 | 98% (5+ days) | 15% (<3 days) | Strong |
| shared_content | 0.89 | ≥ 3 | 95% (3+ shares) | 22% (<2 shares) | Strong |
Recommended Activation Metric
PRIMARY METRIC: Use 6+ features within first week
Why this metric:
- Highest correlation (0.95) with 30-day retention
- Clear inflection point: Users with 6+ features show 97% retention vs 18% for <4 features
- Within user control: Users can explore features if properly guided
- Leading indicator: Feature exploration happens early and predicts long-term engagement
- Easy to measure: Simple count of distinct features accessed
Threshold Analysis:
- 6+ features: 97% retention rate (350 users)
- 4-5 features: 65% retention rate (75 users)
- 1-3 features: 18% retention rate (75 users)
- Inflection point: 6 features is where retention jumps significantly
Secondary Activation Signals
1. Days Active (correlation: 0.93)
- Threshold: 5+ days active in first week
- Retention: 98% for users who achieve this
- Note: Highly predictive but requires sustained behavior; harder to influence immediately
- Use case: Combine with primary metric for composite scoring
2. Shared Content (correlation: 0.89)
- Threshold: 3+ pieces of shared content
- Retention: 95% for users who achieve this
- Note: Indicates value realization and network effects
- Use case: Power user indicator; may require having contacts/team
Composite Activation Score
For nuanced user segmentation, combine all three signals:
Activation Score Formula:
Score = (features_used ≥ 6) × 0.50 +
(days_active ≥ 5) × 0.30 +
(shared_content ≥ 3) × 0.20
| Score Range | Classification | % of Users | Action Required |
|---|---|---|---|
| 1.0 | Fully Activated | ~48% | Nurture; identify champions |
| 0.7-0.9 | Partially Activated | ~22% | Targeted nudges to missing behaviors |
| 0.5-0.6 | At Risk | ~15% | Proactive outreach; feature discovery |
| 0.0-0.4 | Not Activated | ~15% | Urgent intervention; onboarding reset |
User-Level Activation Status (Sample)
| User ID | Features Used | Days Active | Shared | Activation Score | Status | Retained | Recommended Action |
|---|---|---|---|---|---|---|---|
| u007 | 18 | 7 | 10 | 1.0 | Activated | Yes | Champion - case study candidate |
| u003 | 15 | 6 | 8 | 1.0 | Activated | Yes | Power user |
| u001 | 12 | 7 | 5 | 1.0 | Activated | Yes | Healthy engagement |
| u002 | 8 | 5 | 3 | 1.0 | Activated | Yes | Good usage pattern |
| u008 | 5 | 3 | 2 | 0.3 | Not Activated | Yes* | Borderline - encourage exploration |
| u006 | 3 | 2 | 1 | 0.0 | Not Activated | No | Urgent: Feature discovery tour |
| u004 | 2 | 1 | 0 | 0.0 | Not Activated | No | Win-back: “What blocked you?” |
| u009 | 1 | 1 | 0 | 0.0 | Not Activated | No | Onboarding failure - restart flow |
*Note: u008 retained despite low score - outlier or delayed activation
Framework Assessment
Evaluating features_used ≥ 6 against activation metric criteria:
| Criteria | Rating | Assessment |
|---|---|---|
| Correlation Strength | Excellent (0.95) | Very strong predictor of retention |
| Early Timing | High | Achievable within first week of usage |
| Within User Control | High | Users can try features if guided properly |
| Leading Indicator | High | Precedes long-term engagement patterns |
| Easy to Measure | High | Simple event tracking; reliable data |
| Actionable | High | Clear onboarding goal; can optimize for this |
OVERALL RECOMMENDATION: PRIMARY ACTIVATION METRIC ✓
Activation Rate Benchmarks
Current Performance:
- 70% of users achieve activation (6+ features in first week)
- 30% fail to activate and show 18% retention
Target Recommendations:
- Immediate goal: Increase activation rate from 70% to 80% (+10pp)
- Stretch goal: 85%+ activation rate
- Focus area: First 3 days - guide users to try 2+ features per day
Expected Impact:
- Moving 10% of users from “not activated” to “activated” would:
- Increase overall retention from 70% to ~77%
- Reduce churn by ~25%
Validation Roadmap
CRITICAL: These are correlations, not proven causation. To validate:
Phase 1: Observational Validation (Week 1-2)
- Track new user cohort (n=200)
- Measure time-to-6th-feature
- Confirm correlation holds in fresh data
Phase 2: A/B Test (Week 3-6)
- Control Group: Standard onboarding
- Test Group: Enhanced feature discovery
- In-app tooltips for unused features
- “Try 6 features” progress bar
- Feature recommendation engine
- Primary metric: % achieving 6+ features in week 1
- Secondary metric: 30-day retention difference
Phase 3: Refinement (Week 7+)
- If test group shows significantly higher activation AND retention:
- Roll out enhanced onboarding
- Optimize for time-to-activation
- Consider gamification (feature exploration badges)
- If no retention improvement despite higher feature usage:
- Investigate: Are users trying wrong features?
- Consider feature quality/relevance
- May need segmented activation metrics by user type
Caution: Correlation ≠ Causation
What we know: Users who try 6+ features are much more likely to retain.
What we DON’T know:
- Does feature usage cause retention?
- Or do highly engaged users naturally try more features?
- Could forcing feature exploration backfire if features aren’t relevant?
Action: Validate causation via A/B test before major product changes.
Implementation Checklist
- Track activation metric: Implement “features_used_count_first_week” in analytics
- Create activation dashboard: Monitor daily activation rate
- Design experiment: Plan A/B test for feature discovery interventions
- User segmentation: Tag users by activation status for targeted campaigns
- Onboarding optimization: Guide new users toward 6-feature goal
- Identify feature gaps: Which 6 features drive most value? Prioritize those.
- Monitor activation velocity: How quickly do activated users reach 6 features? (Day 1, 2, 3?)
- Churn prediction: Flag users at day 5 with <4 features for intervention
Next Steps
- Immediate: Start tracking composite activation score for all new users
- Week 1: Design and implement feature discovery nudges
- Week 2: Launch A/B test with 20% of new signups
- Week 6: Analyze test results; decide on rollout
- Ongoing: Monitor activation rate as key growth metric
Success Metric: Increase activation rate from 70% to 80% within 60 days, resulting in ~7pp retention improvement.
About This Skill
Analyze user event data to identify which behaviors predict retention. Combines statistical correlation analysis with a structured framework for evaluating activation metric candidates.
View Skill DetailsMore Examples
Event-Level Analysis: Task Completion Activation
Full activation metric analysis for 250 users with signup, task_completion, and invite events. Identifies 'Complete 3+ tasks in 7 days' as primary metric with 0.91 correlation.
Event Tracking Recommendations for Productivity App
Generates hypothesis framework for a productivity app with 150 users, providing recommended events to track, expected correlations, and data collection templates before full analysis.