Fast and powerful conditional randomization testing via distillation
- PMID: 37416628
- PMCID: PMC10323874
- DOI: 10.1093/biomet/asab039
Fast and powerful conditional randomization testing via distillation
Abstract
We consider the problem of conditional independence testing: given a response and covariates , we test the null hypothesis that . The conditional randomization test was recently proposed as a way to use distributional information about to exactly and nonasymptotically control Type-I error using any test statistic in any dimensionality without assuming anything about . This flexibility, in principle, allows one to derive powerful test statistics from complex prediction algorithms while maintaining statistical validity. Yet the direct use of such advanced test statistics in the conditional randomization test is prohibitively computationally expensive, especially with multiple testing, due to the requirement to recompute the test statistic many times on resampled data. We propose the distilled conditional randomization test, a novel approach to using state-of-the-art machine learning algorithms in the conditional randomization test while drastically reducing the number of times those algorithms need to be run, thereby taking advantage of their power and the conditional randomization test's statistical guarantees without suffering the usual computational expense. In addition to distillation, we propose a number of other tricks, like screening and recycling computations, to further speed up the conditional randomization test without sacrificing its high power and exact validity. Indeed, we show in simulations that all our proposals combined lead to a test that has similar power to the most powerful existing conditional randomization test implementations, but requires orders of magnitude less computation, making it a practical tool even for large datasets. We demonstrate these benefits on a breast cancer dataset by identifying biomarkers related to cancer stage.
Keywords: Conditional independence test; Conditional randomization test; High-dimensional inference; Machine learning; Model-X.
Figures
References
-
- Barber RF & Candès EJ (2015). Controlling the false discovery rate via knockoffs. Ann. Statist 43, 2055–85.
-
- Bellot A & van der Schaar M (2019). Conditional independence testing using generative adversarial networks. Proc. Adv. Neural Inf. Proc. Syst 32, 2199–208.
-
- Benjamini Y & Hochberg Y (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Statist. Soc. B 57, 289–300.
-
- Benjamini Y & Yekutieli D (2001). The control of the false discovery rate in multiple testing under dependency. Ann. Statist 29, 1165–88.
