Abstract-Reliable and accurate localization of mobile objects is essential for many applications in wireless networks. In rangebased localization, the position of the object can be inferred using the distance measurements from wireless signals exchanged with active objects or reflected by passive ones. Power allocation for ranging signals is important since it affects not only network lifetime and throughput but also localization accuracy. In this paper, we establish a unifying optimization framework for power allocation in both active and passive localization networks. In particular, we first determine the functional properties of the localization accuracy metric, which enable us to transform the power allocation problems into second-order cone programs (SOCPs). We then propose the robust counterparts of the problems in the presence of parameter uncertainty and develop asymptotically optimal and efficient near-optimal SOCP-based algorithms. Our simulation results validate the efficiency and robustness of the proposed algorithms.
In the classic Bayesian restless multi-armed bandit (RMAB) problem, there are N arms, with rewards on all arms evolving at each time as Markov chains with known parameters. A player seeks to activate K ≥ 1 arms at each time in order to maximize the expected total reward obtained over multiple plays. RMAB is a challenging problem that is known to be PSPACE-hard in general. We consider in this work the even harder non-Bayesian RMAB, in which the parameters of the Markov chain are assumed to be unknown a priori. We develop an original approach to this problem that is applicable when the corresponding Bayesian problem has the structure that, depending on the known parameter values, the optimal solution is one of a prescribed finite set of policies. In such settings, we propose to learn the optimal policy for the non-Bayesian RMAB by employing a suitable meta-policy which treats each policy from this finite set as an arm in a different non-Bayesian multi-armed bandit problem for which a single-arm selection policy is optimal. We demonstrate this approach by developing a novel sensing policy for opportunistic spectrum access over unknown dynamic channels. We prove that our policy achieves near-logarithmic regret (the difference in expected reward compared to a modelaware genie), which leads to the same average reward that can be achieved by the optimal policy under a known model. This is the first such result in the literature for a non-Bayesian RMAB. For our proof, we also develop a novel generalization of the Chernoff-Hoeffding bound.
Wideband ranging is essential for numerous emerging applications that rely on accurate location awareness. The quality of range information, which depends on network intrinsic properties and signal processing techniques, affects the localization accuracy. A popular class of ranging techniques is based on energy detection owing to its low complexity implementation. This paper establishes a tractable model for the range information as a function of wireless environment, signal features, and energy detection techniques. Such a model serves as a corner- stone for the design and analysis of wideband ranging systems. Based on the proposed model, we develop practical soft-decision and hard-decision algorithms. A case study for ranging and local- ization systems operating in a wireless environment is presented. Sample-level simulations validate our theoretical results
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.