Algebraic Analysis for Nonidentifiable Learning Machines

  • Sumio Watanabe
    P&I Laboratory, Tokyo Institute of Technology, Yokohama, 226-8503 Japan

抄録

<jats:p> This article clarifies the relation between the learning curve and the algebraic geometrical structure of a nonidentifiable learning machine such as a multilayer neural network whose true parameter set is an analytic set with singular points. By using a concept in algebraic analysis, we rigorously prove that the Bayesian stochastic complexity or the free energy is asymptotically equal to λ<jats:sub>1</jats:sub> logn − (m<jats:sub>1</jats:sub> − 1) loglogn + constant, where n is the number of training samples and λ<jats:sub>1</jats:sub> and m<jats:sub>1</jats:sub> are the rational number and the natural number, which are determined as the birational invariant values of the singularities in the parameter space. Also we show an algorithm to calculate λ<jats:sub>1</jats:sub> and m<jats:sub>1</jats:sub> based on the resolution of singularities in algebraic geometry. In regular statistical models, 2λ<jats:sub>1</jats:sub> is equal to the number of parameters and m<jats:sub>1</jats:sub> = 1, whereas in nonregular models, such as multilayer networks, 2λ<jats:sub>1</jats:sub> is not larger than the number of parameters and m<jats:sub>1</jats:sub> ≥ 1. Since the increase of the stochastic complexity is equal to the learning curve or the generalization error, the nonidentifiable learning machines are better models than the regular ones if Bayesian ensemble learning is applied. </jats:p>

収録刊行物

被引用文献 (113)*注記

もっと見る

詳細情報 詳細情報について

問題の指摘

ページトップへ