Mixpanel is one of the most widely used product analytics platforms, powering event-based tracking for thousands of SaaS products, mobile apps, and digital businesses. Understanding the full landscape of Mixpanel's dimensions and metrics is essential for anyone building dashboards, analyzing user behavior, or optimizing product experiences based on data.
This guide provides a complete reference of every dimension and metric available in Mixpanel as of 2026. We cover event properties, user profiles, funnel analytics, retention curves, flow analysis, segmentation, A/B testing, and revenue tracking — with property names, formulas, and practical context on when and how to use each one.
What Are Dimensions vs Metrics in Mixpanel?
Before exploring the full reference, it's essential to understand how Mixpanel organizes its data model — which differs significantly from traditional web analytics tools like Google Analytics.
Dimensions in Mixpanel are properties that describe and categorize your data. These include event names, event properties (like button color, page URL, or plan type), user profile properties (like email, sign-up date, or subscription tier), and default properties collected automatically (like device type, OS, browser, city, and country). Dimensions answer: "How do I want to slice this data?"
Metrics are the quantitative measurements Mixpanel calculates from your events and user data. These include total event counts, unique user counts, conversion rates, retention percentages, session durations, and revenue totals. Metrics answer: "What happened in my product?"
Unlike traditional analytics that centers on page views and sessions, Mixpanel is built around events — discrete actions users take in your product. Every button click, form submission, purchase, or feature usage can be tracked as an event with associated properties. This event-centric model gives you far more granular data than session-based tools.
Event Dimensions
Events are the foundation of Mixpanel's data model. Every user interaction you track is an event, and each event carries a set of properties that describe it. These properties serve as your dimensions for filtering, grouping, and segmenting data.
Default Event Properties
Mixpanel automatically collects certain properties with every event, providing baseline context without any additional instrumentation. These are available immediately after you integrate the Mixpanel SDK.
| Dimension | Property Name | Description |
|---|---|---|
| Event Name | $event_name | The name of the tracked event (e.g., "Sign Up", "Purchase", "Page View") |
| Time | $time | Unix timestamp when the event was triggered |
| Distinct ID | $distinct_id | Unique identifier for the user (anonymous or authenticated) |
| Insert ID | $insert_id | Deduplication identifier to prevent counting duplicate events |
| Browser | $browser | User's browser: Chrome, Safari, Firefox, Edge, etc. |
| Browser Version | $browser_version | Version number of the browser |
| OS | $os | Operating system: iOS, Android, Windows, macOS, Linux |
| OS Version | $os_version | Version of the operating system |
| Device | $device | Device model: iPhone 15, Samsung Galaxy S24, iPad Pro, etc. |
| Device Type | $device_type | Category: Mobile, Tablet, Desktop |
| Screen Height | $screen_height | Screen height in pixels |
| Screen Width | $screen_width | Screen width in pixels |
| City | $city | City derived from IP address geolocation |
| Region | $region | State or province from IP geolocation |
| Country Code | $country_code | Two-letter ISO country code (US, FR, GB, DE, etc.) |
| Timezone | $timezone | User's timezone offset from UTC |
| Library | $lib | SDK used: javascript, ios, android, python, etc. |
| Library Version | $lib_version | Version of the Mixpanel SDK |
| App Version | $app_version_string | Your application version (mobile apps) |
| App Build Number | $app_build_number | Build number of your mobile application |
Marketing Attribution Properties
Mixpanel captures UTM parameters and referrer data automatically for web events, letting you attribute user actions to their acquisition source without custom instrumentation.
| Dimension | Property Name | Description |
|---|---|---|
| UTM Source | utm_source | Traffic source: google, facebook, newsletter, etc. |
| UTM Medium | utm_medium | Marketing medium: cpc, email, social, organic, etc. |
| UTM Campaign | utm_campaign | Campaign name or identifier |
| UTM Content | utm_content | Specific ad or content variation identifier |
| UTM Term | utm_term | Search keyword or targeting term |
| Referrer | $referrer | Full URL of the referring page |
| Referring Domain | $referring_domain | Domain of the referrer (e.g., google.com, twitter.com) |
| Initial Referrer | $initial_referrer | First referrer URL in the user's history (first-touch attribution) |
| Initial Referring Domain | $initial_referring_domain | Domain of the user's very first referrer |
| Search Engine | $search_engine | Search engine if traffic came from organic search: Google, Bing, etc. |
Page and Screen Properties
| Dimension | Property Name | Description |
|---|---|---|
| Current URL | $current_url | Full URL of the page where the event was triggered |
| Current Page Title | $title | HTML title of the current page |
| Current Path | $pathname | URL path without domain (e.g., /pricing, /dashboard) |
| Screen Name | $screen_name | Name of the current screen (mobile apps) |
User and People Dimensions
Mixpanel's user profiles (historically called "People") store persistent attributes about each user. These properties are set via the people.set() method and persist across sessions, making them ideal for segmenting users by who they are rather than what they did in a single event.
Default User Profile Properties
| Dimension | Property Name | Description |
|---|---|---|
| Name | $name | User's full name |
$email | User's email address | |
| Phone | $phone | User's phone number |
| Created | $created | Timestamp when the user profile was first created |
| Last Seen | $last_seen | Timestamp of the user's most recent event |
| First Seen | $first_seen | Timestamp of the user's very first tracked event |
| Avatar | $avatar | URL to the user's profile picture |
| City | $city | User's city from their most recent event geolocation |
| Region | $region | User's region/state from most recent geolocation |
| Country Code | $country_code | User's country from most recent geolocation |
Common Custom User Properties
While Mixpanel does not prescribe specific custom properties, most implementations include these commonly used profile dimensions for segmentation and analysis.
| Dimension | Typical Property Name | Description |
|---|---|---|
| Plan Type | plan | Subscription plan: free, starter, pro, enterprise |
| Company | company | User's organization or company name |
| Role | role | User role: admin, member, viewer, owner |
| Sign-Up Source | signup_source | How the user first arrived: organic, paid, referral, direct |
| Lifecycle Stage | lifecycle_stage | Current lifecycle stage: new, activated, engaged, dormant, churned |
| Cohort | $cohort | Behavioral cohort membership defined in Mixpanel's UI |
Core Metrics
Core metrics in Mixpanel measure the fundamental pulse of your product — how many events occur, how many users are active, and how frequently they engage. These form the foundation for every Mixpanel report.
| Metric | Calculation | Description |
|---|---|---|
| Total Events | Count of all events | Total number of tracked events in the selected time period |
| Unique Users | Count of distinct distinct_id | Number of unique users who triggered at least one event |
| Events Per User | Total Events ÷ Unique Users | Average number of events each user triggered |
| DAU (Daily Active Users) | Unique users per day | Distinct users who triggered any event on a given calendar day |
| WAU (Weekly Active Users) | Unique users in 7-day window | Distinct users active within a rolling 7-day period |
| MAU (Monthly Active Users) | Unique users in 28/30-day window | Distinct users active within a rolling 28 or 30-day period |
| Stickiness (DAU/MAU) | DAU ÷ MAU | Ratio indicating daily engagement intensity — higher means users return more often |
| Stickiness (DAU/WAU) | DAU ÷ WAU | Weekly stickiness ratio — a DAU/WAU of 0.5 means average users are active 3-4 days per week |
| Session Count | Count of sessions | Total sessions tracked (requires session tracking enabled) |
| Session Duration | Time between first and last event in session | Average or median time users spend per session |
| Sessions Per User | Sessions ÷ Unique Users | Average number of sessions per unique user |
| New Users | Users with first event in period | Count of users whose very first tracked event falls within the selected time range |
Funnel Metrics
Mixpanel's Funnels report tracks how users progress through a defined sequence of events. Funnel metrics reveal where users convert and where they drop off, enabling targeted optimization of your product flows. You can define funnels with 2 to 20 steps, set completion windows, and apply any property as a filter or breakdown.
| Metric | Calculation | Description |
|---|---|---|
| Overall Conversion Rate | (Users completing last step ÷ Users entering first step) × 100 | Percentage of users who completed the entire funnel from first to last step |
| Step Conversion Rate | (Users completing step N ÷ Users completing step N-1) × 100 | Percentage of users who advanced from one step to the next |
| Step Drop-Off Rate | 100 - Step Conversion Rate | Percentage of users who did not advance from one step to the next |
| Drop-Off Count | Users at step N-1 minus users at step N | Absolute number of users who dropped off between two steps |
| Median Time to Convert | Median of individual conversion times | Median time between first and last step completion across all converting users |
| Step-to-Step Time | Time between completing step N-1 and step N | How long users take between consecutive funnel steps |
| Funnel Re-Entry Count | Count of users who re-entered | Users who started the funnel again after completing or dropping off (when re-entry is enabled) |
| Frequency to Convert | Average attempts before converting | How many times users trigger the first step before completing the funnel |
| Holding Constant Conversion | Conversion filtered by property consistency | Conversion rate where a specified property must remain constant across all steps (e.g., same item_id throughout checkout) |
Retention Metrics
Retention is one of Mixpanel's most powerful report types. It measures whether users who performed a starting action return to perform a return action over subsequent time periods. Mixpanel supports three retention calculation methods, each revealing different aspects of user behavior.
N-Day Retention (Classic)
N-day retention measures the percentage of users who return on exactly day N after their initial action. This is the strictest retention metric — a user must be active on that specific day to count.
| Metric | Calculation | Description |
|---|---|---|
| Day 1 Retention | (Users active on day 1 ÷ Users in cohort) × 100 | Percentage who returned exactly 1 day after their initial action |
| Day 7 Retention | (Users active on day 7 ÷ Users in cohort) × 100 | Percentage who returned exactly 7 days after initial action |
| Day 14 Retention | (Users active on day 14 ÷ Users in cohort) × 100 | Percentage who returned exactly 14 days after initial action |
| Day 30 Retention | (Users active on day 30 ÷ Users in cohort) × 100 | Percentage who returned exactly 30 days after initial action |
| Day 90 Retention | (Users active on day 90 ÷ Users in cohort) × 100 | Long-term retention — percentage still active 3 months later |
Unbounded Retention (Rolling)
Unbounded retention measures users who returned on day N or any day after. This gives a more forgiving view of retention — users don't have to come back on the exact day, just at some point from that day onward.
| Metric | Calculation | Description |
|---|---|---|
| Unbounded Day 1 | (Users active on day 1+ ÷ Users in cohort) × 100 | Percentage who returned on day 1 or any subsequent day |
| Unbounded Day 7 | (Users active on day 7+ ÷ Users in cohort) × 100 | Percentage who returned on day 7 or later |
| Unbounded Day 30 | (Users active on day 30+ ÷ Users in cohort) × 100 | Percentage who returned on day 30 or later |
Frequency Retention
| Metric | Calculation | Description |
|---|---|---|
| Frequency Distribution | Users bucketed by return count | How many users returned 1x, 2x, 3x, etc. within the period |
| Average Return Frequency | Total return actions ÷ Returning users | Average number of times returning users came back |
| Power Users Percentage | (Users with N+ returns ÷ Total cohort) × 100 | Percentage of users who returned frequently (e.g., 10+ times in 30 days) |
Flow and Pathfinding Dimensions
Mixpanel's Flows report visualizes the paths users take through your product. Unlike funnels (which track a predefined sequence), flows reveal the actual sequences users follow — showing you the most common paths, unexpected detours, and where users go after specific actions.
| Dimension / Metric | Description |
|---|---|
| Starting Event | The anchor event from which paths are traced (forward or backward) |
| Path Step | Each subsequent or preceding event in the user's sequence |
| Path Depth | Number of steps shown in the flow (configurable, typically 3-10 steps) |
| Path Volume | Number of users who followed a specific path sequence |
| Path Percentage | Percentage of total users who followed each path branch |
| Drop-Off Node | Point where users left the product (no subsequent event within the session) |
| Top Paths | Most frequently traveled event sequences ranked by volume |
| Sankey Width | Visual width of each path branch proportional to its user volume |
Segmentation Dimensions
Segmentation is Mixpanel's primary analysis tool — it lets you break down any metric by any property dimension. You can group by event properties, user properties, or cohort membership to compare performance across segments.
Property Breakdown Types
| Breakdown Type | Description | Example |
|---|---|---|
| Event Property | Break down metrics by a property attached to the event | Total purchases broken down by payment_method |
| User Property | Break down metrics by a persistent user profile attribute | Sign-ups broken down by plan type |
| Cohort | Break down by behavioral cohort membership | DAU broken down by "Power Users" vs "Casual Users" cohort |
| Computed Property | Break down by a dynamically calculated property (custom formulas) | Revenue broken down by custom "LTV Tier" formula |
| Session Property | Break down by session-level attributes | Events broken down by $session_duration bucket |
Aggregation Methods
When building segmentation queries, you choose how to aggregate your data. Each method reveals different aspects of user behavior.
| Aggregation | Description |
|---|---|
| Total Events | Count every event occurrence (including repeats from the same user) |
| Unique Users | Count each user only once regardless of how many events they triggered |
| Total Per User (Average) | Average event count per user — Total Events ÷ Unique Users |
| Total Per User (Median) | Median event count per user — more robust to outliers than average |
| Sum of Property | Sum a numeric property across all events (e.g., sum of revenue) |
| Average of Property | Average value of a numeric property across events |
| Min / Max of Property | Minimum or maximum value of a numeric property in the selected period |
| Distinct Count of Property | Number of unique values a property takes (e.g., distinct page URLs visited) |
| DAU / WAU / MAU | Active user counts at daily, weekly, or monthly granularity |
A/B Test Metrics
Mixpanel integrates with experimentation platforms and supports native experiment analysis. When you pass an experiment variant as an event or user property, you can measure the causal impact of product changes on any metric.
| Metric | Description |
|---|---|
| Variant Distribution | Number of users assigned to each experiment variant (control vs. treatment) |
| Conversion Rate by Variant | Percentage of users in each variant who completed the target action |
| Lift | Percentage improvement of treatment over control: ((Treatment Rate - Control Rate) ÷ Control Rate) × 100 |
| Statistical Significance | Confidence level that the observed difference is not due to chance (typically 95% threshold) |
| P-Value | Probability of observing the result if there were no real difference — lower means more confident |
| Confidence Interval | Range within which the true lift likely falls (e.g., +5% to +12% improvement) |
| Sample Size | Total users included in each variant — affects statistical power |
| Metric Per Variant | Any Mixpanel metric (revenue, events, retention) calculated separately per variant |
Signal Metrics
Mixpanel's Signal report automatically identifies correlations between user actions and outcomes. It surfaces which events and properties most strongly predict a target behavior — like which onboarding actions predict long-term retention.
| Metric | Description |
|---|---|
| Correlation Score | Strength of association between an event/property and the target outcome (0 to 1) |
| Correlation Direction | Whether the relationship is positive (more action = more outcome) or negative |
| Users Who Did | Number of users who performed the correlated action AND the target outcome |
| Users Who Didn't | Number of users who performed the correlated action but NOT the target outcome |
| Conversion with Action | Target outcome rate among users who performed the correlated action |
| Conversion without Action | Target outcome rate among users who did not perform the correlated action |
| Lift from Action | Difference in outcome rate: Conversion with Action minus Conversion without Action |
Revenue Metrics
Revenue tracking in Mixpanel connects product behavior to business outcomes. By tracking purchase and subscription events with monetary properties, you can analyze revenue patterns across user segments, acquisition channels, and time cohorts.
| Metric | Calculation | Description |
|---|---|---|
| Total Revenue | Sum of amount property on revenue events | Total monetary value from all tracked revenue events in the period |
| ARPU (Avg Revenue Per User) | Total Revenue ÷ Total Users | Average revenue generated per user (including non-paying users) |
| ARPPU (Avg Revenue Per Paying User) | Total Revenue ÷ Paying Users | Average revenue from users who actually made a purchase |
| Paying User Count | Unique users with revenue events | Number of distinct users who triggered at least one revenue event |
| Conversion to Paid | (Paying Users ÷ Total Users) × 100 | Percentage of all users who became paying customers |
| Average Transaction Value | Total Revenue ÷ Transaction Count | Average monetary value per individual transaction |
| Transaction Count | Count of revenue events | Total number of purchase or payment events |
| LTV by Cohort | Cumulative revenue per user in a sign-up cohort | Lifetime value tracked over time for users who joined in the same period |
| MRR (Monthly Recurring Revenue) | Sum of monthly subscription amounts | Total recurring revenue from active subscriptions in a given month |
| Churn Rate | (Cancelled subscriptions ÷ Active at start) × 100 | Percentage of subscribers who cancelled within the period |
| Expansion Revenue | Revenue from upgrades and add-ons | Additional revenue from existing customers upgrading plans or buying add-ons |
| Revenue by Segment | Revenue broken down by any property | Revenue split by plan type, acquisition channel, geography, or any custom dimension |
How to Use These Metrics for Product Optimization
Having access to dozens of metrics is powerful, but selecting the right ones for your analysis determines whether you generate actionable insights or drown in data noise. Here's a practical framework for each analysis scenario.
For understanding product-market fit
Start with DAU/MAU stickiness ratio — a ratio above 0.25 indicates decent engagement, above 0.4 suggests strong product-market fit for most products. Combine with Day 7 and Day 30 retention to see if users stick around. Use the Signal report to identify which early actions predict long-term retention — these are your "aha moment" indicators.
For optimizing onboarding
Build a funnel from Sign Up through each key onboarding step to your activation event. Track step conversion rate and median time to convert at each step. Break down the funnel by acquisition source to see if certain channels produce users who complete onboarding more reliably. Use Flows to discover what users actually do right after sign-up — are they following your intended path or going elsewhere?
For reducing churn
Compare N-day retention curves between churned and active users to find where engagement diverges. Use Signal to identify events that predict churn (negative correlations with retention). Track frequency retention to understand how often your most loyal users engage. Segment retention by user property (plan, company size, acquisition source) to find which segments have the worst drop-off.
For revenue optimization
Track ARPU and ARPPU trends over time — rising ARPPU with flat ARPU means your existing payers spend more but you're not converting more free users. Build a conversion funnel from free to paid to identify where the paywall experience breaks. Use LTV by cohort to compare how much users acquired in different months generate over their lifetime. Break down revenue by acquisition sourceto invest in channels that produce the highest-value customers.
For feature adoption
Track unique users of a feature event as a percentage of your total MAU to measure adoption rate. Use funnels to understand the discovery path — how do users find and first use the feature? Compare retention between users who adopted the feature and those who didn't to measure its impact on engagement.
Common Mistakes in Mixpanel Analytics
Even experienced analysts make these errors when working with Mixpanel data. Avoiding them will produce more accurate insights and better decisions.
1. Tracking too many or too few events
Some teams track every micro-interaction (every button hover, scroll position, modal open) while others track only 2-3 events. The sweet spot is 15-50 meaningful events that map to user intent and business value. Too many events create noise and slow queries; too few leave blind spots in your analysis.
2. Not using user profiles for segmentation
If you only track event properties without setting user profile properties, you lose the ability to segment by persistent user attributes like plan type, company size, or lifecycle stage. Always set key user properties via people.set() so you can break down any metric by who the user is, not just what they did.
3. Comparing N-day and unbounded retention
N-day retention requires users to return on exactly that day. Unbounded retention counts users who returned on that day or later. Comparing a 20% N-day-7 retention with a 45% unbounded-day-7 retention from another product is misleading — they measure fundamentally different things. Always specify which retention method you are using.
4. Averaging ratios across segments
Conversion rates, ARPU, and other ratios cannot be averaged across segments to get accurate totals. A segment with 1,000 users and 5% conversion rate should not be averaged with a segment of 100 users and 20% conversion. Calculate totals from the base numbers: total conversions divided by total users.
5. Ignoring identity management
If you don't properly call identify() when users log in, you end up with duplicate profiles — the same person counted as two users (anonymous and authenticated). This inflates unique user counts, deflates retention, and makes funnel analysis unreliable. Implement Mixpanel's ID Merge correctly from the start.
6. Not setting a funnel completion window
Mixpanel funnels default to a 30-day completion window. If your product's typical conversion path takes 2 hours (e.g., e-commerce checkout), a 30-day window will include conversions that span weeks — mixing urgent intent with casual browsing. Set the completion window to match your product's realistic conversion timeframe.
