A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks
- PMID: 36299529
- PMCID: PMC9587964
- DOI: 10.1007/s11222-022-10169-0
A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks
Abstract
To fit sparse linear associations, a LASSO sparsity inducing penalty with a single hyperparameter provably allows to recover the important features (needles) with high probability in certain regimes even if the sample size is smaller than the dimension of the input vector (haystack). More recently learners known as artificial neural networks (ANN) have shown great successes in many machine learning tasks, in particular fitting nonlinear associations. Small learning rate, stochastic gradient descent algorithm and large training set help to cope with the explosion in the number of parameters present in deep neural networks. Yet few ANN learners have been developed and studied to find needles in nonlinear haystacks. Driven by a single hyperparameter, our ANN learner, like for sparse linear associations, exhibits a phase transition in the probability of retrieving the needles, which we do not observe with other ANN learners. To select our penalty parameter, we generalize the universal threshold of Donoho and Johnstone (Biometrika 81(3):425-455, 1994) which is a better rule than the conservative (too many false detections) and expensive cross-validation. In the spirit of simulated annealing, we propose a warm-start sparsity inducing algorithm to solve the high-dimensional, non-convex and non-differentiable optimization problem. We perform simulated and real data Monte Carlo experiments to quantify the effectiveness of our approach.
Keywords: Model selection; Neural networks; Phase transition; Sparsity; Universal threshold.
© The Author(s) 2022.
Figures
References
-
- Adcock, B., Brugiapaglia, S., Dexter, N., Morage, S.: Deep neural networks are effective at learning high-dimensional Hilbert-valued functions from limited data. In: Proceedings of the 2nd Mathematical and Scientific Machine Learning Conference, vol. 145, pp. 1–36. PMLR (2022)
-
- Adcock B, Dexter N. The gap between theory and practice in function approximation with deep neural networks. SIAM J. Math. Data Sci. 2021;3(2):624–655. doi: 10.1137/20M131309X. - DOI
-
- Arlot S, Celisse A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010;4:40–79. doi: 10.1214/09-SS054. - DOI
-
- Bach F, Jenatton R, Mairal J, Obozinski G. Optimization with sparsity-inducing penalties. Found. Trends Mach. Learn. 2011;4(1):1–106. doi: 10.1561/2200000015. - DOI
LinkOut - more resources
Full Text Sources