Algebraic geometrical methods for hierarchical learning machines
- PMID: 11681750
- DOI: 10.1016/s0893-6080(01)00069-7
Algebraic geometrical methods for hierarchical learning machines
Abstract
Hierarchical learning machines such as layered perceptrons, radial basis functions, Gaussian mixtures are non-identifiable learning machines, whose Fisher information matrices are not positive definite. This fact shows that conventional statistical asymptotic theory cannot be applied to neural network learning theory, for example either the Bayesian a posteriori probability distribution does not converge to the Gaussian distribution, or the generalization error is not in proportion to the number of parameters. The purpose of this paper is to overcome this problem and to clarify the relation between the learning curve of a hierarchical learning machine and the algebraic geometrical structure of the parameter space. We establish an algorithm to calculate the Bayesian stochastic complexity based on blowing-up technology in algebraic geometry and prove that the Bayesian generalization error of a hierarchical learning machine is smaller than that of a regular statistical model, even if the true distribution is not contained in the parametric model.
Similar articles
-
Algebraic analysis for nonidentifiable learning machines.Neural Comput. 2001 Apr;13(4):899-933. doi: 10.1162/089976601300014402. Neural Comput. 2001. PMID: 11255575
-
Singularities in mixture models and upper bounds of stochastic complexity.Neural Netw. 2003 Sep;16(7):1029-38. doi: 10.1016/S0893-6080(03)00005-4. Neural Netw. 2003. PMID: 14692637
-
Stochastic complexities of reduced rank regression in Bayesian estimation.Neural Netw. 2005 Sep;18(7):924-33. doi: 10.1016/j.neunet.2005.03.014. Neural Netw. 2005. PMID: 15993036
-
Gaussian processes for machine learning.Int J Neural Syst. 2004 Apr;14(2):69-106. doi: 10.1142/S0129065704001899. Int J Neural Syst. 2004. PMID: 15112367 Review.
-
Bottom-up learning of explicit knowledge using a Bayesian algorithm and a new Hebbian learning rule.Neural Netw. 2011 Apr;24(3):219-32. doi: 10.1016/j.neunet.2010.12.002. Epub 2010 Dec 16. Neural Netw. 2011. PMID: 21239141 Review.
Cited by
-
Identifying the Best Approximating Model in Bayesian Phylogenetics: Bayes Factors, Cross-Validation or wAIC?Syst Biol. 2023 Jun 17;72(3):616-638. doi: 10.1093/sysbio/syad004. Syst Biol. 2023. PMID: 36810802 Free PMC article.
-
Development of spectral decomposition based on Bayesian information criterion with estimation of confidence interval.Sci Technol Adv Mater. 2020 Jul 2;21(1):402-419. doi: 10.1080/14686996.2020.1773210. Sci Technol Adv Mater. 2020. PMID: 32939165 Free PMC article.
-
Learning Coefficient of Vandermonde Matrix-Type Singularities in Model Selection.Entropy (Basel). 2019 Jun 4;21(6):561. doi: 10.3390/e21060561. Entropy (Basel). 2019. PMID: 33267275 Free PMC article.