Analytic Optimization of Adaptive Ridge Parameters Based on Regularized Subspace Information Criterion

  • GOKITA Shun
    Department of Computer Science, Tokyo Institute of Technology
  • SUGIYAMA Masashi
    Department of Computer Science, Tokyo Institute of Technology
  • SAKURAI Keisuke
    Department of Computational Intelligence and Systems Science, Tokyo Institute of Technology

Search this article

Abstract

In order to obtain better learning results in supervised learning, it is important to choose model parameters appropriately. Model selection is usually carried out by preparing a finite set of model candidates, estimating a generalization error for each candidate, and choosing the best one from the candidates. If the number of candidates is increased in this procedure, the optimization quality may be improved. However, this in turn increases the computational cost. In this paper, we focus on a generalization error estimator called the regularized subspace information criterion and derive an analytic form of the optimal model parameter over a set of infinitely many model candidates. This allows us to maximize the optimization quality while the computational cost is kept moderate.

Journal

References(12)*help

See more

Details 詳細情報について

  • CRID
    1573950402335474944
  • NII Article ID
    110007537996
  • NII Book ID
    AA10826239
  • ISSN
    09168508
  • Text Lang
    en
  • Data Source
    • CiNii Articles

Report a problem

Back to top