Sciweavers

334 search results - page 61 / 67
» How to Dynamically Merge Markov Decision Processes
Sort
View
BTW
2009
Springer
168views Database» more  BTW 2009»
15 years 27 days ago
Embedded Analytics in Front Office Applications
: Today, decision making by users of front office applications happens without analytical information supporting this process. We propose as solution Embedded Analytics (EA) making...
Martin Oberhofer, Erik Nijkamp
AAAI
2006
14 years 11 months ago
Action Selection in Bayesian Reinforcement Learning
My research attempts to address on-line action selection in reinforcement learning from a Bayesian perspective. The idea is to develop more effective action selection techniques b...
Tao Wang
JMLR
2010
157views more  JMLR 2010»
14 years 4 months ago
Why are DBNs sparse?
Real stochastic processes operating in continuous time can be modeled by sets of stochastic differential equations. On the other hand, several popular model families, including hi...
Shaunak Chatterjee, Stuart Russell
HICSS
2003
IEEE
207views Biometrics» more  HICSS 2003»
15 years 2 months ago
Formalizing Multi-Agent POMDP's in the context of network routing
This paper uses partially observable Markov decision processes (POMDP’s) as a basic framework for MultiAgent planning. We distinguish three perspectives: first one is that of a...
Bharaneedharan Rathnasabapathy, Piotr J. Gmytrasie...
ICDCS
2010
IEEE
15 years 1 months ago
Stochastic Steepest-Descent Optimization of Multiple-Objective Mobile Sensor Coverage
—We propose a steepest descent method to compute optimal control parameters for balancing between multiple performance objectives in stateless stochastic scheduling, wherein the ...
Chris Y. T. Ma, David K. Y. Yau, Nung Kwan Yip, Na...