Conjugate gradient algorithms in nonconvex optimization
著者
書誌事項
Conjugate gradient algorithms in nonconvex optimization
(Nonconvex optimization and its applications, v. 89)
Springer, c2009
大学図書館所蔵 全8件
  青森
  岩手
  宮城
  秋田
  山形
  福島
  茨城
  栃木
  群馬
  埼玉
  千葉
  東京
  神奈川
  新潟
  富山
  石川
  福井
  山梨
  長野
  岐阜
  静岡
  愛知
  三重
  滋賀
  京都
  大阪
  兵庫
  奈良
  和歌山
  鳥取
  島根
  岡山
  広島
  山口
  徳島
  香川
  愛媛
  高知
  福岡
  佐賀
  長崎
  熊本
  大分
  宮崎
  鹿児島
  沖縄
  韓国
  中国
  タイ
  イギリス
  ドイツ
  スイス
  フランス
  ベルギー
  オランダ
  スウェーデン
  ノルウェー
  アメリカ
注記
Formerly CIP
Includes bibliographical references and index
内容説明・目次
内容説明
Conjugate direction methods were proposed in the early 1950s. When high speed digital computing machines were developed, attempts were made to lay the fo- dations for the mathematical aspects of computations which could take advantage of the ef?ciency of digital computers. The National Bureau of Standards sponsored the Institute for Numerical Analysis, which was established at the University of California in Los Angeles. A seminar held there on numerical methods for linear equationswasattendedbyMagnusHestenes, EduardStiefel andCorneliusLanczos. This led to the ?rst communication between Lanczos and Hestenes (researchers of the NBS) and Stiefel (of the ETH in Zurich) on the conjugate direction algorithm. The method is attributed to Hestenes and Stiefel who published their joint paper in 1952 [101] in which they presented both the method of conjugate gradient and the conjugate direction methods including conjugate Gram-Schmidt processes. A closelyrelatedalgorithmwasproposedbyLanczos[114]whoworkedonalgorithms for determiningeigenvalues of a matrix.
His iterative algorithm yields the similarity transformation of a matrix into the tridiagonal form from which eigenvalues can be well approximated.Thethree-termrecurrencerelationofthe Lanczosprocedurecan be obtained by eliminating a vector from the conjugate direction algorithm scheme. Initially the conjugate gradient algorithm was called the Hestenes-Stiefel-Lanczos method [86].
目次
Conjugate Direction Methods for Quadratic Problems.- Conjugate Gradient Methods for Nonconvex Problems.- Memoryless Quasi-Newton Methods.- Preconditioned Conjugate Gradient Algorithms.- Limited Memory Quasi-Newton Algorithms.- The Method of Shortest Residuals and Nondifferentiable Optimization.- The Method of Shortest Residuals for Differentiable Problems.- The Preconditioned Shortest Residuals Algorithm.- Optimization on a Polyhedron.- Conjugate Gradient Algorithms for Problems with Box Constraints.- Preconditioned Conjugate Gradient Algorithms for Problems with Box Constraints.- Preconditioned Conjugate Gradient Based Reduced-Hessian Methods.
「Nielsen BookData」 より