Article
Reporting lift against control groups
Lift reports are designed to answer one of the hardest questions in marketing measurement:
Did exposure to my brand, creative, or message cause people to think or behave differently?
Many reports can tell you what people think or how groups differ. Lift goes a step further. It isolates the effect of exposure by comparing people who were exposed to a stimulus with a carefully matched group who were not. The result is a clear, defensible view of effectiveness that moves beyond simple correlation.
Lift reports are most useful when you want to:
- Prove that a campaign or creative worked
- Compare the effectiveness of different stimuli
- Understand which outcomes were genuinely influenced by exposure
What a Lift Report Compares
At its core, every Lift report compares two groups:
- Exposed respondents People whose responses indicate they were exposed to a stimulus (for example, they saw an ad or recognized a brand).
- Control respondents People who were not exposed, but who are otherwise similar in terms of key demographics.
The difference between these two groups is the lift.
Step 1: Select Report Type
When creating a new report, choose Lift.

A Lift report is specifically designed to:
- Compare outcomes against a control group
- Apply demographic matching automatically
- Test whether observed differences are statistically meaningful
Step 2: Create Lift Report
This step defines the rules of the analysis: how exposure is identified, how strict the statistics should be, and how stable the results need to be.

Name
The name is for internal reference and reporting.
Example:
Brand Lift – Spring Campaign
Choose something that clearly reflects the stimulus and timeframe being measured.
Project
Select the project this report belongs to. This determines which surveys and responses are available for analysis.
Matching iterations
Lift reports do not rely on a single control group. Instead, they repeatedly rebuild a balanced control group to improve stability.
- Each iteration creates a new matched control group
- Results are averaged across all iterations
- Higher numbers reduce random variation
Typical guidance:
- 3–5 iterations for early exploration
- 5–10 iterations for final or client-facing results
Default is 5.
P value threshold
The P value threshold controls how strict the statistical filter is.
Only results with a P value at or below this threshold are shown.
Common interpretations:
- 0.05 → strong statistical evidence
- 0.10 → directional but less certain
- 0.50 (default) → exploratory, surfaces emerging signals
Lower thresholds mean fewer results, but higher confidence.
Include Negative Lift
By default, the report only shows positive lift.
Turning this on will also show:
- Outcomes where exposed respondents performed worse than control
This is useful for:
- Diagnosing problematic creatives
- Understanding unintended effects
Positive indicator values
This defines how exposure is recognized.
Any response matching one of these values is treated as exposed.
Examples:
True Yes Seen
Values are:
- Case-insensitive
- Compared against the Indicator Tag (configured later)
Step 3: Select Report Questions
Next, choose which questions from the survey will be used in the Lift report.

At this stage, you are simply selecting questions.
Their roles are defined in the next step.
Typical selection includes:
- Outcome questions (awareness, consideration, intent, recall)
- Demographic questions (used for matching)
Step 4: Configure Analysis
This is where the structure of the Lift analysis is defined: which questions are measured, which are used for matching, and how stimuli are interpreted.

Question Configuration
Each selected question must be assigned a role.
At least one **Outcome ** and one Control Demo are required.
Outcome
Outcome questions are where lift is measured. These represent the behaviors or attitudes you want to influence.
Examples:
- Brand Awareness
- Brand Consideration
- Recommend Likelihood
- Message Recall
Each Outcome requires two additional settings:
- Type-** Stimulus Topic**
Control Demo
Control Demo questions are used only to create a balanced control group.
Examples:
- Gender
- Age Group
- Region
- Prior category usage
These questions are not measured for lift.
They ensure exposed and control respondents are comparable.
Outcome Types
Value
Used for categorical responses.
Examples:
- Yes / No
- Brand selected
- Option chosen
Lift is calculated separately for each response value.
Number
Used for numeric or scale-based responses.
Examples:
- Likelihood (0–10)
- Rating (1–5)
Lift is calculated at thresholds (for example, 7+,8+,9+).
Stimulus Topic
The Stimulus Topic defines what the outcome is being lifted by.
Examples:
- brand
- creative
- message
This allows the same outcome to be analyzed separately for each brand, creative, or message in the data.
Indicator Tag
The Indicator Tag identifies where exposure is stored.
Example:
exposed
This tag is combined with Positive indicator values to classify respondents as:
- Exposed-** Control**
How Lift Is Calculated
For each stimulus value (for example, each brand):
- Respondents are split into exposed and control candidates
- A matched control group is built using Control Demo questions
- Outcome rates are calculated for both groups
- Lift is calculated as the difference between those rates
- Statistical significance is tested
- Results are averaged across matching iterations
- Only results passing the P value threshold are shown
Worked Example: Brand Lift
Scenario
You want to understand whether exposure to a brand campaign increased key brand metrics.
Indicator Tag
exposed
Positive indicator values
True
Stimulus Topic
brand
Outcomes
- Brand Awareness (Value)
- Brand Consideration (Number)
- Recommend Likelihood (Number)
Control Demo
- Gender
Example Result
| Brand | Metric | Exposed | Control | Lift |
|---|---|---|---|---|
| Brand A | Awareness | 62% | 54% | +8% |
| Brand A | Consideration 7+ | 38% | 33% | +5% |
Interpretation
Exposure to Brand A increased both awareness and consideration. The lift indicates a genuine effect of the campaign, not just demographic differences.
Worked Example: Creative Testing
Scenario
You are testing multiple creatives within the same campaign.
Stimulus Topic
creative
Outcomes
- Message Recall (Value)
- Brand Favorability (Number)
Example Result

Strengths of Lift Reports
Lift reports provide:
- A causal view of effectiveness
- Automatic demographic control
- Clear exposed vs control logic
- Statistically filtered results
- Comparability across brands or creatives
Limitations to Keep in Mind
- Matching reduces usable sample size
- Exposure must be measured reliably
- Too many Control Demo variables can reduce power
- Results apply to the matched population, not the full sample
When to Use Lift
Lift reports are best when you need to:
- Prove effectiveness
- Compare creatives or messages
- Defend decisions with evidence
They are not designed for:
- Media mix modeling
- Budget optimization
- Very small samples
Summary
Lift reports isolate the impact of exposure by comparing exposed respondents to a carefully matched control group. When configured correctly, they provide one of the clearest and most defensible ways to demonstrate whether a stimulus genuinely worked.