Sciweavers

14 search results - page 1 / 3
» Learning Options for an MDP from Demonstrations
Sort
View
AAAI
2006
13 years 6 months ago
Using Homomorphisms to Transfer Options across Continuous Reinforcement Learning Domains
We examine the problem of Transfer in Reinforcement Learning and present a method to utilize knowledge acquired in one Markov Decision Process (MDP) to bootstrap learning in a mor...
Vishal Soni, Satinder P. Singh
ATAL
2010
Springer
13 years 5 months ago
Combining manual feedback with subsequent MDP reward signals for reinforcement learning
As learning agents move from research labs to the real world, it is increasingly important that human users, including those without programming skills, be able to teach agents de...
W. Bradley Knox, Peter Stone
ECAI
2006
Springer
13 years 8 months ago
Learning by Automatic Option Discovery from Conditionally Terminating Sequences
Abstract. This paper proposes a novel approach to discover options in the form of conditionally terminating sequences, and shows how they can be integrated into reinforcement learn...
Sertan Girgin, Faruk Polat, Reda Alhajj
ICML
2007
IEEE
14 years 5 months ago
Constructing basis functions from directed graphs for value function approximation
Basis functions derived from an undirected graph connecting nearby samples from a Markov decision process (MDP) have proven useful for approximating value functions. The success o...
Jeffrey Johns, Sridhar Mahadevan
JMLR
2010
161views more  JMLR 2010»
12 years 11 months ago
Accuracy-Rejection Curves (ARCs) for Comparing Classification Methods with a Reject Option
Data extracted from microarrays are now considered an important source of knowledge about various diseases. Several studies based on microarray data and the use of receiver operat...
Malik Sajjad Ahmed Nadeem, Jean-Daniel Zucker, Bla...