Sciweavers

129 search results - page 12 / 26
» Automatic Recovery Using Bounded Partially Observable Markov...
Sort
View
ICMLA
2009
15 years 1 months ago
Sensitivity Analysis of POMDP Value Functions
In sequential decision making under uncertainty, as in many other modeling endeavors, researchers observe a dynamical system and collect data measuring its behavior over time. The...
Stéphane Ross, Masoumeh T. Izadi, Mark Merc...
145
Voted
AROBOTS
2008
166views more  AROBOTS 2008»
15 years 1 months ago
User-adapted plan recognition and user-adapted shared control: A Bayesian approach to semi-autonomous wheelchair driving
Abstract Many elderly and physically impaired people experience difficulties when maneuvering a powered wheelchair. In order to provide improved maneuvering, powered wheelchairs ha...
Eric Demeester, Alexander Hüntemann, Dirk Van...
CORR
2010
Springer
105views Education» more  CORR 2010»
15 years 1 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
ICTAI
2005
IEEE
15 years 9 months ago
Planning with POMDPs Using a Compact, Logic-Based Representation
Partially Observable Markov Decision Processes (POMDPs) provide a general framework for AI planning, but they lack the structure for representing real world planning problems in a...
Chenggang Wang, James G. Schmolze
ATAL
2007
Springer
15 years 7 months ago
Interactive dynamic influence diagrams
This paper extends the framework of dynamic influence diagrams (DIDs) to the multi-agent setting. DIDs are computational representations of the Partially Observable Markov Decisio...
Kyle Polich, Piotr J. Gmytrasiewicz