In sequential decision-making problems formulated as Markov decision processes, state-value function approximation using domain features is a critical technique for scaling up the...
Recent research has shown the benefit of framing problems of imitation learning as solutions to Markov Decision Problems. This approach reduces learning to the problem of recoveri...
Brian Ziebart, Andrew L. Maas, J. Andrew Bagnell, ...
Partially observable Markov decision processes (POMDPs) are an intuitive and general way to model sequential decision making problems under uncertainty. Unfortunately, even approx...
Tao Wang, Pascal Poupart, Michael H. Bowling, Dale...
Revelation policies in an e-marketplace differ in terms of the level of competitive information disseminated to participating sellers. Since sellers who repeatedly compete against...
Amy R. Greenwald, Karthik Kannan, Ramayya Krishnan
We present an application of hierarchical Bayesian estimation to robot map building. The revisiting problem occurs when a robot has to decide whether it is seeing a previously-bui...
Benjamin Stewart, Jonathan Ko, Dieter Fox, Kurt Ko...