E.T. Jaynes : papers on probability, statistics and statistical physics
著者
書誌事項
E.T. Jaynes : papers on probability, statistics and statistical physics
(Synthese library, v. 158)
D. Reidel , Sold and distributed in the U.S.A. and Canada by Kluwer Boston, c1983
- タイトル別名
-
Papers on probability, statistics and statistical physics
大学図書館所蔵 全36件
  青森
  岩手
  宮城
  秋田
  山形
  福島
  茨城
  栃木
  群馬
  埼玉
  千葉
  東京
  神奈川
  新潟
  富山
  石川
  福井
  山梨
  長野
  岐阜
  静岡
  愛知
  三重
  滋賀
  京都
  大阪
  兵庫
  奈良
  和歌山
  鳥取
  島根
  岡山
  広島
  山口
  徳島
  香川
  愛媛
  高知
  福岡
  佐賀
  長崎
  熊本
  大分
  宮崎
  鹿児島
  沖縄
  韓国
  中国
  タイ
  イギリス
  ドイツ
  スイス
  フランス
  ベルギー
  オランダ
  スウェーデン
  ノルウェー
  アメリカ
注記
Bibliography: p. 425-429
Includes index
内容説明・目次
内容説明
The first six chapters of this volume present the author's 'predictive' or information theoretic' approach to statistical mechanics, in which the basic probability distributions over microstates are obtained as distributions of maximum entropy (Le. , as distributions that are most non-committal with regard to missing information among all those satisfying the macroscopically given constraints). There is then no need to make additional assumptions of ergodicity or metric transitivity; the theory proceeds entirely by inference from macroscopic measurements and the underlying dynamical assumptions. Moreover, the method of maximizing the entropy is completely general and applies, in particular, to irreversible processes as well as to reversible ones. The next three chapters provide a broader framework - at once Bayesian and objective - for maximum entropy inference. The basic principles of inference, including the usual axioms of probability, are seen to rest on nothing more than requirements of consistency, above all, the requirement that in two problems where we have the same information we must assign the same probabilities.
Thus, statistical mechanics is viewed as a branch of a general theory of inference, and the latter as an extension of the ordinary logic of consistency. Those who are familiar with the literature of statistics and statistical mechanics will recognize in both of these steps a genuine 'scientific revolution' - a complete reversal of earlier conceptions - and one of no small significance.
目次
1. Introductory Remarks.- 2. Information Theory and Statistical Mechanics, I (1957).- 3. Information Theory and Statistical Mechanics, II (1957).- 4. Brandeis Lectures (1963).- 5. Gibbs vs Boltzmann Entropies (1965).- 6. Delaware Lecture (1967).- 7. Prior Probabilities (1968).- 8. The Well-Posed Problem (1973).- 9. Confidence Intervals vs Bayesian Intervals (1976).- 10. Where Do We Stand on Maximum Entropy? (1978).- 11. Concentration of Distributions at Entropy Maxima (1979).- 12. Marginalization and Prior Probabilities (1980).- 13. What is the Question? (1981).- 14. The Minimum Entropy Production Principle (1980).- Supplementary Bibliography.
「Nielsen BookData」 より