Handbook of learning and approximate dynamic programming
Author(s)
Bibliographic Information
Handbook of learning and approximate dynamic programming
(IEEE series on computational intelligence / David B. Fogel, series editor)
Wiley-Interscience , IEEE Press, c2004
Available at 11 libraries
  Aomori
  Iwate
  Miyagi
  Akita
  Yamagata
  Fukushima
  Ibaraki
  Tochigi
  Gunma
  Saitama
  Chiba
  Tokyo
  Kanagawa
  Niigata
  Toyama
  Ishikawa
  Fukui
  Yamanashi
  Nagano
  Gifu
  Shizuoka
  Aichi
  Mie
  Shiga
  Kyoto
  Osaka
  Hyogo
  Nara
  Wakayama
  Tottori
  Shimane
  Okayama
  Hiroshima
  Yamaguchi
  Tokushima
  Kagawa
  Ehime
  Kochi
  Fukuoka
  Saga
  Nagasaki
  Kumamoto
  Oita
  Miyazaki
  Kagoshima
  Okinawa
  Korea
  China
  Thailand
  United Kingdom
  Germany
  Switzerland
  France
  Belgium
  Netherlands
  Sweden
  Norway
  United States of America
Note
Includes bibliographical references and index
Description and Table of Contents
Description
A complete resource to Approximate Dynamic Programming (ADP), including on-line simulation code
Provides a tutorial that readers can use to start implementing the learning algorithms provided in the book
Includes ideas, directions, and recent results on current research issues and addresses applications where ADP has been successfully implemented
The contributors are leading researchers in the field
Table of Contents
Foreword. 1. ADP: goals, opportunities and principles.
Part I: Overview.
2. Reinforcement learning and its relationship to supervised learning.
3. Model-based adaptive critic designs.
4. Guidance in the use of adaptive critics for control.
5. Direct neural dynamic programming.
6. The linear programming approach to approximate dynamic programming.
7. Reinforcement learning in large, high-dimensional state spaces.
8. Hierarchical decision making.
Part II: Technical advances.
9. Improved temporal difference methods with linear function approximation.
10. Approximate dynamic programming for high-dimensional resource allocation problems.
11. Hierarchical approaches to concurrency, multiagency, and partial observability.
12. Learning and optimization - from a system theoretic perspective.
13. Robust reinforcement learning using integral-quadratic constraints.
14. Supervised actor-critic reinforcement learning.
15. BPTT and DAC - a common framework for comparison.
Part III: Applications.
16. Near-optimal control via reinforcement learning.
17. Multiobjective control problems by reinforcement learning.
18. Adaptive critic based neural network for control-constrained agile missile.
19. Applications of approximate dynamic programming in power systems control.
20. Robust reinforcement learning for heating, ventilation, and air conditioning control of buildings.
21. Helicopter flight control using direct neural dynamic programming.
22. Toward dynamic stochastic optimal power flow.
23. Control, optimization, security, and self-healing of benchmark power systems.
by "Nielsen BookData"