Coverart for item
The Resource A Concise Introduction to Decentralized POMDPs

A Concise Introduction to Decentralized POMDPs

Label
A Concise Introduction to Decentralized POMDPs
Title
A Concise Introduction to Decentralized POMDPs
Creator
Contributor
Subject
Language
eng
Member of
Cataloging source
MiAaPQ
Literary form
non fiction
Nature of contents
dictionaries
Series statement
SpringerBriefs in Intelligent Systems Ser.
Series volume
v.0
A Concise Introduction to Decentralized POMDPs
Label
A Concise Introduction to Decentralized POMDPs
Link
http://libproxy.rpi.edu/login?url=https://ebookcentral.proquest.com/lib/rpi/detail.action?docID=4537987
Publication
Copyright
Related Contributor
Related Location
Related Agents
Related Authorities
Related Subjects
Related Items
Carrier category
online resource
Carrier category code
cr
Carrier MARC source
rdacarrier
Color
multicolored
Content category
text
Content type code
txt
Content type MARC source
rdacontent
Contents
  • Preface -- Acknowledgments -- Contents -- Acronyms -- List of Symbols -- 1 Multiagent Systems Under Uncertainty -- 1.1 Motivating Examples -- 1.2 Multiagent Systems -- 1.3 Uncertainty -- 1.4 Applications -- 2 The Decentralized POMDP Framework -- 2.1 Single-Agent Decision Frameworks -- 2.1.1 MDPs -- 2.1.2 POMDPs -- 2.2 Multiagent Decision Making: Decentralized POMDPs -- 2.3 Example Domains -- 2.3.1 Dec-Tiger -- 2.3.2 Multirobot Coordination: Recycling and Box-Pushing -- 2.3.3 Network Protocol Optimization -- 2.3.4 Efficient Sensor Networks -- 2.4 Special Cases, Generalizations and Related Models -- 2.4.1 Observability and Dec-MDPs -- 2.4.2 Factored Models -- 2.4.3 Centralized Models: MMDPs and MPOMDPs -- 2.4.4 Multiagent Decision Problems -- 2.4.5 Partially Observable Stochastic Games -- 2.4.6 Interactive POMDPs -- 3 Finite-Horizon Dec-POMDPs -- 3.1 Optimality Criteria -- 3.2 Policy Representations: Histories and Policies -- 3.2.1 Histories -- 3.2.2 Policies -- 3.3 Multiagent Beliefs -- 3.4 Value Functions for Joint Policies -- 3.5 Complexity -- 4 Exact Finite-Horizon Planning Methods -- 4.1 Backwards Approach: Dynamic Programming -- 4.1.1 Growing Policies from Subtree Policies -- 4.1.2 Dynamic Programming for Dec-POMDPs -- 4.2 Forward Approach: Heuristic Search -- 4.2.1 Temporal Structure in Policies: Decision Rules -- 4.2.2 Multiagent A* -- 4.3 Converting to a Non-observable MDP -- 4.3.1 The Plan-Time MDP and Optimal Value Function -- 4.3.2 Plan-Time Sufficient Statistics -- 4.3.3 An NOMDP Formulation -- 4.4 Other Finite-Horizon Methods -- 4.4.1 Point-Based DP -- 4.4.2 Optimization -- 5 Approximate and Heuristic Finite-Horizon Planning Methods -- 5.1 Approximation Methods -- 5.1.1 Bounded Dynamic Programming -- 5.1.2 Early Stopping of Heuristic Search -- 5.1.3 Application of POMDP Approximation Algorithms -- 5.2 Heuristic Methods
  • 5.2.1 Alternating Maximization -- 5.2.2 Memory-Bounded Dynamic Programming -- 5.2.3 Approximate Heuristic-Search Methods -- 5.2.4 Evolutionary Methods and Cross-Entropy Optimization -- 6 Infinite-Horizon Dec-POMDPs -- 6.1 Optimality Criteria -- 6.1.1 Discounted Cumulative Reward -- 6.1.2 Average Reward -- 6.2 Policy Representation -- 6.2.1 Finite-State Controllers: Moore and Mealy -- 6.2.2 An Example Solution for DEC-TIGER -- 6.2.3 Randomization -- 6.2.4 Correlation Devices -- 6.3 Value Functions for Joint Policies -- 6.4 Undecidability, Alternative Goals and Their Complexity -- 7 Infinite-Horizon Planning Methods: Discounted Cumulative Reward -- 7.1 Policy Iteration -- 7.2 Optimizing Fixed-Size Controllers -- 7.2.1 Best-First Search -- 7.2.2 Bounded Policy Iteration -- 7.2.3 Nonlinear Programming -- 7.2.4 Expectation Maximization -- 7.2.5 Reduction to an NOMDP -- 8 Further Topics -- 8.1 Exploiting Structure in Factored Models -- 8.1.1 Exploiting Constraint Optimization Methods -- 8.1.1.1 Coordination (Hyper-)Graphs -- 8.1.1.2 ND-POMDPs -- 8.1.1.3 Factored Dec-POMDPs -- 8.1.2 Exploiting Influence-Based Policy Abstraction -- 8.2 Hierarchical Approaches and Macro-Actions -- 8.3 Communication -- 8.3.1 Implicit Communication and Explicit Communication -- 8.3.1.1 Explicit Communication Frameworks -- 8.3.1.2 Updating of Information States and Semantics -- 8.3.2 Delayed Communication -- 8.3.2.1 One-Step Delayed Communication -- 8.3.2.2 k-Steps Delayed Communication -- 8.3.3 Communication with Costs -- 8.3.4 Local Communication -- 8.4 Reinforcement Learning -- 9 Conclusion -- References
http://library.link/vocab/cover_art
https://contentcafe2.btol.com/ContentCafe/Jacket.aspx?Return=1&Type=S&Value=9783319289298&userID=ebsco-test&password=ebsco-test
Dimensions
unknown
http://library.link/vocab/discovery_link
{'f': 'http://opac.lib.rpi.edu/record=b4385443'}
Extent
1 online resource (146 pages)
Form of item
online
Isbn
9783319289298
Media category
computer
Media MARC source
rdamedia
Media type code
c
Sound
unknown sound
Specific material designation
remote

Library Locations

    • Folsom LibraryBorrow it
      110 8th St, Troy, NY, 12180, US
      42.729766 -73.682577
Processing Feedback ...