Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Jul;44(15-17):e70185.
doi: 10.1002/sim.70185.

Performance of Cross-Validated Targeted Maximum Likelihood Estimation

Affiliations

Performance of Cross-Validated Targeted Maximum Likelihood Estimation

Matthew J Smith et al. Stat Med. 2025 Jul.

Abstract

Background: Advanced methods for causal inference, such as targeted maximum likelihood estimation (TMLE), require specific convergence rates and the Donsker class condition for valid statistical estimation and inference. In situations where there is no differentiability due to data sparsity or near-positivity violations, the Donsker class condition is violated. In such instances, the bias of the targeted estimand is inflated, and its variance is anti-conservative, leading to poor coverage. Cross-validation of the TMLE algorithm (CVTMLE) is a straightforward, yet effective way to ensure efficiency, especially in settings where the Donsker class condition is violated, such as random or near-positivity violations. We aim to investigate the performance of CVTMLE compared to TMLE in various settings.

Methods: We utilized the data-generating mechanism described in Leger et al. (2022) to run a Monte Carlo experiment under different Donsker class violations. Then, we evaluated the respective statistical performances of TMLE and CVTMLE with different super learner libraries, with and without regression tree methods.

Results: We found that CVTMLE vastly improves confidence interval coverage without adversely affecting bias, particularly in settings with small sample sizes and near-positivity violations. Furthermore, incorporating regression trees using standard TMLE with ensemble super learner-based initial estimates increases bias and reduces variance, leading to invalid statistical inference.

Conclusions: We show through simulations that CVTMLE is much less sensitive to the choice of the super learner library and thereby provides better estimation and inference in cases where the super learner library uses more flexible candidates and is prone to overfitting.

Keywords: Donsker class condition; causal inference; data sparsity; epidemiology; near‐positivity violation; observational studies; targeted maximum likelihood estimation.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflicts of interest.

Figures

FIGURE 1
FIGURE 1
Process map of cross‐validated targeted maximum likelihood estimation.
FIGURE 2
FIGURE 2
Relative bias and coverage of all TMLE and CVTMLE approaches under data‐generating mechanisms 1–4.
FIGURE 3
FIGURE 3
Relative bias and coverage of all TMLE and CVTMLE approaches under data‐generating mechanisms 5–8.
FIGURE 4
FIGURE 4
Decision tree for the appropriate choice of method given the scenarios (i.e., near‐positivity violation, sample size) that can cause the lack of differentiability of the influence curve and potentially violate the Donsker class condition.
FIGURE A1
FIGURE A1
Probability of the outcome given the exposure and Z1 (variable creating near‐positivity violations), stratified by prevalence of the exposure (i.e., 50% or 80%) and presence of extrapolation issue (i.e., none or high). (A) is 50% prevalence of the exposure with no extrapolation issue. (B) is 80% prevalence of the exposure with no extrapolation issue. (C) is a 50% prevalence of the exposure with an extrapolation issue. (D) is an 80% prevalence of the exposure with an extrapolation issue.

References

    1. Smith M. J., Mansournia M. A., Maringe C., et al., “Introduction to Computational Causal Inference Using Reproducible Stata, R, and Python Code: A Tutorial,” Statistics in Medicine 41, no. 2 (2022): 407–432, 10.1002/sim.9234. - DOI - PMC - PubMed
    1. Austin P. C., “An Introduction to Propensity Score Methods for Reducing the Effects of Confounding in Observational Studies,” Multivariate Behavioral Research 46, no. 3 (2011): 399–424, 10.1080/00273171.2011.568786. - DOI - PMC - PubMed
    1. Williamson E., Morley R., Lucas A., and Carpenter J., “Propensity Scores: From Naïve Enthusiasm to Intuitive Understanding,” Statistical Methods in Medical Research 21, no. 3 (2011): 273–293, 10.1177/0962280210394483. - DOI - PubMed
    1. Robins J. M., Hernán M. Á., and Brumback B., “Marginal Structural Models and Causal Inference in Epidemiology,” Epidemiology 11, no. 5 (2000): 550–560, 10.1097/00001648-200009000-00011. - DOI - PubMed
    1. Rosenbaum P. R. and Rubin D. B., “The Central Role of the Propensity Score in Observational Studies for Causal Effects,” Biometrika 70, no. 1 (1983): 41–55, 10.1093/biomet/70.1.41. - DOI

LinkOut - more resources