Handbook of Markov decision processes : methods and applications
Author(s)
Bibliographic Information
Handbook of Markov decision processes : methods and applications
(International series in operations research & management science, 40)
Kluwer Academic, c2002
Available at 34 libraries
  Aomori
  Iwate
  Miyagi
  Akita
  Yamagata
  Fukushima
  Ibaraki
  Tochigi
  Gunma
  Saitama
  Chiba
  Tokyo
  Kanagawa
  Niigata
  Toyama
  Ishikawa
  Fukui
  Yamanashi
  Nagano
  Gifu
  Shizuoka
  Aichi
  Mie
  Shiga
  Kyoto
  Osaka
  Hyogo
  Nara
  Wakayama
  Tottori
  Shimane
  Okayama
  Hiroshima
  Yamaguchi
  Tokushima
  Kagawa
  Ehime
  Kochi
  Fukuoka
  Saga
  Nagasaki
  Kumamoto
  Oita
  Miyazaki
  Kagoshima
  Okinawa
  Korea
  China
  Thailand
  United Kingdom
  Germany
  Switzerland
  France
  Belgium
  Netherlands
  Sweden
  Norway
  United States of America
Note
Includes bibliographical references and index
Description and Table of Contents
Description
Eugene A. Feinberg Adam Shwartz This volume deals with the theory of Markov Decision Processes (MDPs) and their applications. Each chapter was written by a leading expert in the re spective area. The papers cover major research areas and methodologies, and discuss open questions and future research directions. The papers can be read independently, with the basic notation and concepts ofSection 1.2. Most chap ters should be accessible by graduate or advanced undergraduate students in fields of operations research, electrical engineering, and computer science. 1.1 AN OVERVIEW OF MARKOV DECISION PROCESSES The theory of Markov Decision Processes-also known under several other names including sequential stochastic optimization, discrete-time stochastic control, and stochastic dynamic programming-studiessequential optimization ofdiscrete time stochastic systems. The basic object is a discrete-time stochas tic system whose transition mechanism can be controlled over time. Each control policy defines the stochastic process and values of objective functions associated with this process. The goal is to select a "good" control policy. In real life, decisions that humans and computers make on all levels usually have two types ofimpacts: (i) they cost orsavetime, money, or other resources, or they bring revenues, as well as (ii) they have an impact on the future, by influencing the dynamics. In many situations, decisions with the largest immediate profit may not be good in view offuture events. MDPs model this paradigm and provide results on the structure and existence of good policies and on methods for their calculation.
Table of Contents
- 1. Introduction
- E.A. Feinberg, A. Shwartz. Part I: Finite State and Action Models. 2. Finite State and Action MDPs
- L. Kallenberg. 3. Bias Optimality
- M.E. Lewis, M.L. Puterman. 4. Singular Perturbations of Markov Chains and Decision Processes
- K.E. Avrachenkov, et al. Part II: Infinite State Models. 5. Average Reward Optimization Theory for Denumerable State Spaces
- L.I. Sennott. 6. Total Reward Criteria
- E.A. Feinberg. 7. Mixed Criteria
- E.A. Feinberg, A. Shwartz. 8. Blackwell Optimality
- A. Hordijk, A.A. Yushkevich. 9. The Poisson Equation for Countable Markov Chains: Probabilistic Methods and Interpretations
- A.M. Makowski, A. Shwartz. 10. Stability, Performance Evaluation, and Optimization
- S.P. Meyn. 11. Convex Analytic Methods in Markov Decision Processes
- V.S. Borkar. 12. The Linear Programming Approach
- O. Hernandez-Lerma, J.B. Lasserre. 13. Invariant Gambling Problems and Markov Decision Processes
- L.E. Dubins, et al. Part III: Applications. 14. Neuro-Dynamic Programming: Overview and Recent Trends
- B. Van Roy. 15. Markov Decision Processes in Finance and Dynamic Options
- M. Schal. 16. Applications of Markov Decision Processes in Communication Networks
- E. Altman. 17. Water Reservoir Applications of Markov Decision Processes
- B.F. Lamond, A. Boukhtouta. Index.
by "Nielsen BookData"