Sciweavers

238 search results - page 32 / 48
» Value-Function Approximations for Partially Observable Marko...
Sort
View
AAAI
2010
15 years 1 months ago
Robust Policy Computation in Reward-Uncertain MDPs Using Nondominated Policies
The precise specification of reward functions for Markov decision processes (MDPs) is often extremely difficult, motivating research into both reward elicitation and the robust so...
Kevin Regan, Craig Boutilier
ICC
2008
IEEE
169views Communications» more  ICC 2008»
15 years 6 months ago
Optimality of Myopic Sensing in Multi-Channel Opportunistic Access
—We consider opportunistic communications over multiple channels where the state (“good” or “bad”) of each channel evolves as independent and identically distributed Mark...
Tara Javidi, Bhaskar Krishnamachari, Qing Zhao, Mi...
ICC
2007
IEEE
121views Communications» more  ICC 2007»
15 years 6 months ago
Structure and Optimality of Myopic Sensing for Opportunistic Spectrum Access
We consider opportunistic spectrum access for secondary users over multiple channels whose occupancy by primary users is modeled as discrete-time Markov processes. Due to hardware...
Qing Zhao, Bhaskar Krishnamachari
GLOBECOM
2009
IEEE
14 years 9 months ago
Dogfight in Spectrum: Jamming and Anti-Jamming in Multichannel Cognitive Radio Systems
Primary user emulation attack in multichannel cognitive radio systems is discussed. An attacker is assumed to be able to send primary-user-like signals during spectrum sensing peri...
Husheng Li, Zhu Han
ECSQARU
2001
Springer
15 years 4 months ago
Space-Progressive Value Iteration: An Anytime Algorithm for a Class of POMDPs
Abstract. Finding optimal policies for general partially observable Markov decision processes (POMDPs) is computationally difficult primarily due to the need to perform dynamic-pr...
Nevin Lianwen Zhang, Weihong Zhang