Sciweavers

948 search results - page 97 / 190
» Modelling Agents as Observable Sources
Sort
View
AIPS
2008
15 years 6 months ago
Multiagent Planning Under Uncertainty with Stochastic Communication Delays
We consider the problem of cooperative multiagent planning under uncertainty, formalized as a decentralized partially observable Markov decision process (Dec-POMDP). Unfortunately...
Matthijs T. J. Spaan, Frans A. Oliehoek, Nikos A. ...
CORR
2008
Springer
150views Education» more  CORR 2008»
15 years 4 months ago
A Local Mean Field Analysis of Security Investments in Networks
Getting agents in the Internet, and in networks in general, to invest in and deploy security features and protocols is a challenge, in particular because of economic reasons arisi...
Marc Lelarge, Jean Bolot
IAT
2005
IEEE
15 years 9 months ago
Decomposing Large-Scale POMDP Via Belief State Analysis
Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing ...
Xin Li, William K. Cheung, Jiming Liu
ATAL
2004
Springer
15 years 9 months ago
Communication for Improving Policy Computation in Distributed POMDPs
Distributed Partially Observable Markov Decision Problems (POMDPs) are emerging as a popular approach for modeling multiagent teamwork where a group of agents work together to joi...
Ranjit Nair, Milind Tambe, Maayan Roth, Makoto Yok...
AAAI
2007
15 years 6 months ago
Scaling Up: Solving POMDPs through Value Based Clustering
Partially Observable Markov Decision Processes (POMDPs) provide an appropriately rich model for agents operating under partial knowledge of the environment. Since finding an opti...
Yan Virin, Guy Shani, Solomon Eyal Shimony, Ronen ...