Cascading Position Bias
Definition
Cascading Position Bias is a refinement of Position Bias that models more realistic user behavior: users examine items sequentially (top-to-bottom) and the probability of examining an item at position depends on whether they found satisfactory items at positions .
Formally:
Examination at position = product of irrelevance of all previous items.
Intuition
Sequential Scanning
Real users don’t independently evaluate every item. They scan top-to-bottom and stop when satisfied:
User sees ranking: [A, B, C, D, E]
Step 1: Examine A
Is A relevant?
YES → Click & Leave
NO → Continue
Step 2: Examine B (only if A was irrelevant)
Is B relevant?
YES → Click & Leave
NO → Continue
Step 3: Examine C (only if A & B irrelevant)
...
This creates dependence: whether you examine item depends on items .
Contrast with Position-Based Model
PBM (Position-Based Model):
- Examination probability is independent of item relevance
- regardless of what’s above
Even if A, B, C are all irrelevant,
C still has 0.8 chance of examination
Cascade Model:
- Examination probability depends on previous items
- If A & B are relevant, user stops; C never examined
- If A & B are irrelevant, C has high exam probability
Mathematical Formulation
The Cascade Model (CM)
Click probability under cascading:
where:
Key: Examination at position is item-specific and ranking-specific (not just position-specific).
Example
Ranking: [A, B, C, D]
True relevances: , , ,
Examination probabilities:
- (always examined first)
- (20% reach B)
- (8% reach C)
- (2.4% reach D)
Click probabilities:
Note: Items lower in ranking have much lower click probabilities because of cascading, not just position bias.
Dependent Click Model (DCM)
A specific cascade model parameterization:
Additional factor:
- = probability the user is satisfied after clicking (decides to leave)
Parameter = rank-dependent satisfaction probability.
Dynamic Bayesian Network (DBN) Model
Another cascade variant:
where:
- = probability of session abandonment (user leaves without finding anything)
Key Differences from PBM
| Aspect | PBM | Cascade Model |
|---|---|---|
| Exam depends on position | Yes | Yes |
| Exam depends on relevance of previous items | No | Yes |
| Ranking-specific exams | No | Yes |
| Identifiability | Easier | Harder |
| Empirical fit | Good for some queries | Better overall |
| Computational complexity | Low | Higher |
Empirical Evidence
When Does Cascading Dominate?
Cascading behavior is observed more in:
- Navigational queries: “find this specific item” → user stops once found
- Informational queries in some contexts: “learn about X” → user might scan exhaustively
Less pronounced in:
- Exploratory queries: “show me options” → user might browse many items
Empirical Studies
Research shows:
- Navigational queries: Cascade model fit better than PBM
- Commercial queries: Mixed—sometimes cascade, sometimes PBM dominates
- Informational queries: Varies by context and result quality
Challenges: Session-Dependent Propensities
The critical problem: Under cascading, examination propensities are ranking-specific.
depends on , not just on .
Example Problem
Scenario 1: Ranking [Relevant, Irrelevant, ...]
P(Exam_2) = 1 - P(Rel_1) = small (user likely satisfied)
Scenario 2: Ranking [Irrelevant, Relevant, ...]
P(Exam_2) = 1 - P(Rel_1) = large (user continues scanning)
Same position (position 2), but very different examination probability!
This breaks the standard IPS formula which assumes position-specific propensities.
Solution: Session-Dependent Probabilities
Use clicks in the current session to estimate propensities:
This requires:
- Logging which positions users examined
- Per-session adaptation
- More data per session
Cascade Model Estimation
EM for Cascade Models
Classic approach: Expectation-Maximization with latent examination variables.
E-step: Given current parameters, infer which items were examined.
M-step: Update parameters to maximize likelihood.
Complication: Must jointly estimate relevance and satisfaction parameters.
RegressionEM with Cascade
Extend cascade model to use features:
Fits regression model to cascade-inferred relevance.
IPS Breaks Down with Cascading
The Problem
If you use PBM-based IPS on cascade-generated data:
where P(\text{Exam}_k^{\text{PBM})}} is position-specific (ignoring cascading):
Result: Severe bias.
Why?
- PBM assumes high examination at rank 3
- Cascade model implies low examination at rank 3 (if ranks 1-2 were relevant)
- IPS weights rank-3 clicks too lightly
- Underestimates relevance of rank-3 items
Empirical Impact
Studies show:
- IPS with PBM on cascade data: biased, noisy estimates
- Specially designed cascade-based IPS: better, but still challenging
Solutions for Cascading
Solution 1: Cascade-Based Propensity Estimation
Estimate propensities accounting for cascading:
Then apply IPS with cascade propensities.
Advantage: Theoretically sound
Disadvantage: Requires accurate relevance estimation first (circular dependency)
Solution 2: Click Models Instead of IPS
Fit a cascade-based click model directly:
Advantage: Avoids IPS variance
Disadvantage: Identifiability issues (multiple solutions possible)
Solution 3: Doubly Robust with Cascade Model
Combine cascade-based DM + IPS:
Advantage: Low variance + unbiased if either component correct
Disadvantage: Complex to implement
Solution 4: Online Randomization
The nuclear option: randomize with random ranking probabilities to break cascading.
(not confounded by previous relevance)
Advantage: Clean propensity estimates
Disadvantage: Harms user experience; requires new data collection
Real-World Implications
When You Should Care
- Web search: Moderate cascading (users often scan multiple results)
- E-commerce search: Strong cascading (users stop after finding good product)
- Recommendations: Moderate (depends on context)
- Ads: Less cascading (users might ignore ad regardless)
When You Can Ignore It
- Item-level feedback (not position-based)
- Conversion data (not click data)
- Systems where users always examine all items
Related Models
Variants of Cascade
- Dependent Click Model: Adds satisfaction probability
- Dynamic Bayesian Network: Adds abandonment probability
- Cascade with attractiveness: Items attract examination independent of position
Different User Behaviors
- Trust Bias: Position affects relevance perception (different from cascade)
- Item Selection Bias: Hard cutoff (items below fold never seen)
- Outlier Bias: Distinctive items break cascade (attract examination out of order)
Connections
- Extends: Position Bias (PBM is a special case)
- Alternative: Click Models provide joint estimation
- Related: Examination Hypothesis (same assumption but different model)
- Impact on: Inverse Propensity Weighting (breaks standard IPS)
- Solution: Doubly Robust Estimation, cascade-based click models
Appears In
- Click Models
- Unbiased Learning to Rank
- User behavior modeling in Information Retrieval