Sensitivity Analysis: Deficiency vs. E-values

Deniz Akdemir

2026-03-26

Introduction

Sensitivity analysis answers: “How strong would unmeasured confounding need to be to change our conclusions?”

Several frameworks exist:

Framework Key Metric Interpretation
E-value Risk ratio needed “Unmeasured confounder must have RR = X”
Partial R² Variance explained “Confounder must explain X% of variance”
Le Cam Deficiency (δ) Information loss “Transfer penalty bounded by \(M\delta\)

This vignette shows how deficiency-based sensitivity analysis relates to and extends traditional approaches.


Conceptual Translation

The E-value Perspective

The E-value (VanderWeele & Ding, 2017) asks:

“To explain away the observed effect, an unmeasured confounder would need to have a risk ratio of at least E with both treatment and outcome.”

For an observed risk ratio RR, the E-value is:

\[E = RR + \sqrt{RR \times (RR - 1)}\]

The Deficiency Perspective

Deficiency (δ) takes a decision-theoretic view:

“Given the information gap between observational and interventional data, the worst-case regret inflation term is bounded by \(M\delta\) (and there is a minimax floor \((M/2)\delta\)).”

Key insight: δ directly quantifies policy consequences, not just statistical associations.

Conceptual Mapping

E-value Concept Deficiency Equivalent
“Effect explained away” δ → 1 (maximal deficiency)
“Effect robust” δ → 0 (zero deficiency)
E-value = 2 Moderate unmeasured confounding
E-value = 5 Strong unmeasured confounding

Practical Example

Setup

Deficiency Estimation

Confounding Frontier

The confounding frontier maps deficiency as a function of confounding strength:

Reading the Plot:

Policy Regret Bound


Comparison with E-values

Computing an Approximate E-value

For comparison, we can compute the E-value for our effect estimate:

Deficiency vs. E-value: Key Differences

Aspect E-value Deficiency (δ)
Scale Risk ratio Total variation distance
Interpretation Strength needed to “explain away” Information loss for decisions
Decision utility Abstract Direct (via \(M\delta\) transfer penalty)
Multi-method Single estimate Compares strategies
Negative controls Not integrated Built-in diagnostics

When to Use Each

Use E-values when: - Communicating to epidemiologists/clinicians familiar with RR - Binary outcomes with clear risk ratio interpretation - Want a single summary number

Use Deficiency (δ) when: - Need decision-theoretic bounds (policy regret) - Comparing multiple adjustment strategies - Have negative control outcomes available - Working with non-binary outcomes (continuous, survival) - Need to combine with sensitivity frontiers


Extended Sensitivity Analysis

Benchmarking Observed Covariates

A key advantage of the frontier approach is benchmarking: we can see where observed covariates fall on the confounding map.

Using Benchmarks:

The benchmarks show the inferred confounding strength of each observed covariate. If an unmeasured confounder would need to be “stronger than W3” (which we know explains 81% of U’s variance), conclusions are robust.

Combining with Negative Controls

# Add negative control
df$Y_nc <- U + rnorm(n, sd = 0.5) # Affected by U, not by A

spec_full <- causal_spec(
    df, "A", "Y", c("W", "W2", "W3"),
    negative_control = "Y_nc"
)
#> ✔ Created causal specification: n=500, 3 covariate(s)

# Complete analysis
def_full <- estimate_deficiency(
    spec_full,
    methods = c("unadjusted", "iptw"),
    n_boot = 100
)
#> ℹ Estimating deficiency: unadjusted
#> ℹ Estimating deficiency: iptw

nc_full <- nc_diagnostic(spec_full, method = "iptw", n_boot = 100)
#> ℹ Using kappa = 1 (conservative). Consider domain-specific estimation or sensitivity analysis via kappa_range.
#> ✔ No evidence against causal assumptions (p = 0.74257 )

print(def_full)
#> 
#> -- Deficiency Proxy Estimates (PS-TV) ------
#> 
#>      Method  Delta     SE               CI            Quality
#>  unadjusted 0.1033 0.0233 [0.1016, 0.1825] Insufficient (Red)
#>        iptw 0.0204 0.0089  [0.016, 0.0482]  Excellent (Green)
#> Note: delta is a propensity-score TV proxy (overlap/balance diagnostic).
#> 
#> Best method: iptw (delta = 0.0204 )
print(nc_full)
#> 
#> -- Negative Control Diagnostic ----------------------------------------
#> 
#> * screening statistic (weighted corr): 0.0173 
#> * delta_NC (association proxy): 0.0173 
#> * delta bound (under kappa alignment): 0.0173 (kappa = 1 )
#> * screening p-value: 0.74257 
#> * screening method: weighted_permutation_correlation 
#> 
#> RESULT: NOT REJECTED. This is a screening result, not proof that confounding is absent.
#> NOTE: Your effect estimate must exceed the Noise Floor (delta_bound) to be meaningful.

Summary: Unified Sensitivity Analysis

The causaldef approach provides a unified framework:

┌─────────────────────────────────────────────────────────────────┐
│                     SENSITIVITY ANALYSIS                         │
├─────────────────────────────────────────────────────────────────┤
│  confounding_frontier()                                          │
│    → Maps δ as function of confounding strength (α, γ)          │
│    → Benchmarks observed covariates as reference points         │
├─────────────────────────────────────────────────────────────────┤
│  nc_diagnostic()                                                 │
│    → Empirical falsification test                               │
│    → Bounds δ using observable negative control                 │
├─────────────────────────────────────────────────────────────────┤
│  policy_regret_bound()                                           │
│    → Translates δ into decision-theoretic consequences          │
│    → Transfer penalty = Mδ; minimax floor = (M/2)δ              │
└─────────────────────────────────────────────────────────────────┘

Key Advantages:

  1. Decision-theoretic meaning: δ bounds actual regret, not just association strength
  2. Multi-method comparison: See which adjustment does best
  3. Empirical validation: Negative controls test assumptions
  4. Visual sensitivity: Frontiers show robustness at a glance

References

  1. Akdemir, D. (2026). Constraints on Causal Inference as Experiment Comparison. DOI: 10.5281/zenodo.18367347

  2. VanderWeele, T. J., & Ding, P. (2017). Sensitivity Analysis in Observational Research: Introducing the E-value. Annals of Internal Medicine.

  3. Cinelli, C., & Hazlett, C. (2020). Making Sense of Sensitivity: Extending Omitted Variable Bias. JRSS-B.

  4. Torgersen, E. (1991). Comparison of Statistical Experiments. Cambridge University Press.