Sinkhorn Treatment Effects: A Causal Optimal Transport Measure

Medha Agarwal1, Alex Luedtke2
1University of Washington (Statistics), 2Harvard Medical School (Department of Health Care Policy)
Sinkhorn_vs_MMD

MMD vs. Sinkhorn divergence across increasing separation $\theta$ between counterfactual outcome distribution under control $P_0 = \mathcal{N}(\mathbf{0}_2, I_2)$ and under treatment $P_{1, \theta} = \tfrac{1}{2} \mathcal{N}(-\theta\mathbf{1}_2, I_2) + \tfrac{1}{2} \mathcal{N}(\theta\mathbf{1}_2, I_2)$. The average treatment effect is zero for all $\theta > 0$. Here $D_\theta$ denotes either MMD or Sinkhorn divergence between $P_0$ and $P_{1, \theta}$. As $\theta$ increases, the distributions diverge and MMD saturates, failing to distinguish between `far' and `very far' distributions, whereas the Sinkhorn divergence continues to grow..

Highlights

Why not only compare means?

In the binary treatment setting, we observe $Z=(X,A,Y) \in \mathcal{Z}$, where $X \in \mathcal{X}$ are confounders, $A\in\{0,1\}$ is treatment, and $Y \in \mathcal{Y}$ is the observed outcome. Under standard identification conditions, the observed law $P$ identifies the counterfactual outcome laws $P_1$ and $P_0$. The ATE compares only $\mathbb E_{P_1}[Y] - \mathbb E_{P_0}[Y]$.

While convenient for computation and inference, mean-based summaries can mask important effects that occur away from the center of the distribution or saturate, failing to distinguish between far and very far distributions. Distributional treatment effects address this by comparing the entire laws $P_1$ and $P_0$.

Kernel-based methods estimate counterfactual distributions via kernel mean embeddings (KME) into characteristic reproducing kernel Hilbert spaces (RKHS) and quantify DTE using maximum mean discrepancy (MMD), termed as MMD treatment effect (MTE). Although MMD metrizes weak convergence, it induces a flat geometry on the space of probability measures and may fail to capture meaningful geometric discrepancies between counterfactual distributions, especially when their supports weakly overlap or separate.

Temporal Smoothing Animation

MMD vs Sinkhorn: MMD saturates as the difference between means of $P$ and $Q_\theta$ increases, whereas Sinkhorn divergence continues to distinguish between distributions.

The estimand

The STE at a fixed regularization level $\varepsilon>0$ is $\mathcal{S}(P) = S_\varepsilon(P_1,P_0)$, where $P_a$ is the counterfactual law of $Y_a$ identified from the observed law $P$, and $S_\varepsilon$ is the debiased entropic optimal transport cost known as the Sinkhorn divergence.

Functional Representation under Identifiability

The key statistical object is the map from the observed law to the Sinkhorn divergence between identified counterfactual laws. We write this composition as

$$\mathcal{S}(P) = S_\varepsilon \circ J \circ \Psi (P),$$

  1. Counterfactual Mean Embedding: $\Psi(P) = (\psi^1(P), \psi^0(P))$ maps the observed law to the two counterfactual kernel mean embeddings.
  2. Canonical Isometric Embedding: $J$ embeds kernel embeddings of distributions into the Banach space in which the Sinkhorn divergence is analyzed.
  3. Sinkhorn Divergence: $S_\varepsilon$ maps the pair of embedded counterfactual distributions to the Sinkhorn divergence between them.

Smoothness Properties

  1. First-order pathwise differentiability: The functional $P \mapsto \mathcal{S}(P)$ is first-order pathwise differentiable at a general $P$ with efficient influence function $\dot{\mathcal{S}}_P: \mathcal{Z} \to \mathbb{R}$ (Lemma 3.1).
  2. Second-order pathwise differentiability: At the null hypothesis of no DTE, when $P_1 = P_0$, the first-order derivative of the Sinkhorn divergence vanishes. This degeneracy enables a second-order expansion, yielding a second-order efficient influence function $\ddot{\mathcal{S}}_P: \mathcal{Z} \times \mathcal{Z} \to \mathbb{R}$ (Theorem 3.2).

One-Step Estimator and Inference

Observed law maps to counterfactual laws, Sinkhorn divergence, and one-step inference

Estimation pipeline. The observed data are split to construct an initial estimator $\widehat{P}$ and an empirical distribution $P_n$. The initial estimator is used to estimate nuisance quantities. The plug-in estimator $S_\varepsilon(\widehat{P}_1,\widehat{P}_0)$ is then debiased using the ones-step estimation. First-order debiasing yields an asymptotically linear estimator away from the null; second-order debiasing yields the test statistic for the null hypothesis with nominal type-I error control.

Aggregating over regularization levels

A single value of $\varepsilon$ may be poorly matched to the alternative: small $\varepsilon$ emphasizes fine geometric structure, while larger $\varepsilon$ yields a smoother discrepancy closer to kernel-based comparisons. We therefore consider a finite grid $\Xi$ of regularization values and combine evidence across scales using a max-type statistic. The aggregated test achieves a nominal type-I error asymptotically.

Simulations

To probe the sensitivity of MTE and STE, we consider two distinct regimes governing the separation between the counterfactual outcome distributions. In Exp (i) (mean shift), the counterfactual outcome laws differ through their means: $P_0=\mathcal{N}(0,\Sigma)$ and $P_1=\mathcal{N}(\theta\mathbf{1}_2,\Sigma)$. In Exp (ii) (covariance shift), the counterfactual outcome laws have the same mean but different covariance: $P_0=\mathcal{N}(0,\Sigma)$ and $P_1=\mathcal{N}(\mathbf{0}_2,\Sigma+\theta\Delta)$. In both settings, $\theta=0$ corresponds to the null and increasing $\theta$ corresponds to stronger alternatives.

PCam image outcomes embedded into feature space and compared by Sinkhorn treatment effects

Simulations. I: Type~I error of the MTE and STE under null (\(\theta=0.0\)); II: Power of MTE and STE under increasing separation between counterfactual distributions (increasing \(\theta\)) under Exp (i); III: Power of MTE and STE under increasing separation between counterfactual distributions (increasing \(\theta\)) under Exp (ii); IV: Mean squared error of plugin vs one-step STE for \(\theta = 1.6\) from Exp (i); V: Coverage of Wald-type $95\%$ confidence intervals for \(\theta = 1.6\) from Exp (i).

Citation

@article{agarwal2026sinkhorn,
        title={Sinkhorn Treatment Effects: A Causal Optimal Transport Measure},
        author={Agarwal, Medha and Luedtke, Alex},
        journal={arXiv preprint arXiv:2605.08485},
        year={2026}
      }