Bibliographic Information

From statistical physics to statistical inference and back

edited by Peter Grassberger and Jean-Pierre Nadal

(NATO ASI series, ser. C . Mathematical and physical sciences ; v. 428)

Kluwer Academic, c1994

Available at  / 18 libraries

Search this Book/Journal

Note

"Published in cooperation with NATO Scientific Affairs Division"

"Proceedings of the NATO Advanced Study Institute on From Statistical Physics to Statistical Inference and Back, Cargèse (Corsica), France, August 31-September 12, 1992" -- T.p. verso

Includes bibliographical references and index

Description and Table of Contents

Description

Physicists, when modelling physical systems with a large number of degrees of freedom, and statisticians, when performing data analysis, have developed their own concepts and methods for making the `best' inference. But are these methods equivalent, or not? What is the state of the art in making inferences? The physicists want answers. More: neural computation demands a clearer understanding of how neural systems make inferences; the theory of chaotic nonlinear systems as applied to time series analysis could profit from the experience already booked by the statisticians; and finally, there is a long-standing conjecture that some of the puzzles of quantum mechanics are due to our incomplete understanding of how we make inferences. Matter enough to stimulate the writing of such a book as the present one. But other considerations also arise, such as the maximum entropy method and Bayesian inference, information theory and the minimum description length. Finally, it is pointed out that an understanding of human inference may require input from psychologists. This lively debate, which is of acute current interest, is well summarized in the present work.

Table of Contents

  • Preface. In Place of an Introduction. Concluding Remarks
  • G. Toulouse. Principles for Inference. Statistical mechanics and the maximum entropy method
  • R. Balian. Irreversibility, probability and entropy
  • A.J.M. Garrett. Maximum entropy for random cellular structures
  • N. Rivier. Minimal Description Length modeling: an introduction
  • J. Rissanen. An introduction to learning and generalization
  • G. Parisi. Information geometry and manifolds of neural networks
  • S.I. Amari. Uncertainty as a resource for managing complexity
  • G.J. Klir. Coding and Statistical Physics of Disordered Systems. The development of Information Theory
  • S. Verdu. Statistical inference, zero-knowledge and proofs of identity
  • J. Stern. Spin glasses: an introduction
  • M. Mezard. Statistical Mechanics and error-correcting codes
  • N. Sourlas. Learning. Learning and generalization with undetermined architecture
  • N. Tishby. Confronting neural network and human behavior in a quasiregular environment
  • M.A. Virasoro. Sensory processing and information theory
  • R. Linsker. The formation of representations in the visual cortex
  • H.-U. Bauer, T. Geisel, K. Pawelzik, F. Wolf. Classifier systems: models for learning agents
  • D.A. Lane. Dynamical Systems. Space time dynamics and biorthogonal analysis: mementum
  • R. Lima. Symbolic encoding in dynamical systems
  • A. Politi. Topological organization of (low-dimensional) chaos
  • N.B. Tufillaro. Noise separation and MDL modeling of chaotic processes
  • J. Rissanen. Quantum Mechanics. Inference in Quantum Mechanics
  • R. Omnes. Decoherence and the existential interpretation ofquantum theory or `no information without representation'
  • W.H. Zurek. Index.

by "Nielsen BookData"

Related Books: 1-1 of 1

Details

Page Top