We propose in this White Paper a concept for a space experiment using cold atoms to search for ultra-light dark matter, and to detect gravitational waves in the frequency range between the most sensitive ranges of LISA and the terrestrial LIGO/Virgo/KAGRA/INDIGO experiments. This interdisciplinary experiment, called Atomic Experiment for Dark Matter and Gravity Exploration (AEDGE), will also complement other planned searches for dark matter, and exploit synergies with other gravitational wave detectors. We give examples of the extended range of sensitivity to ultra-light dark matter offered by AEDGE, and how its gravitational-wave measurements could explore the assembly of super-massive black holes, first-order phase transitions in the early universe and cosmic strings. AEDGE will be based upon technologies now being developed for terrestrial experiments using cold atoms, and will benefit from the space experience obtained with, e.g., LISA and cold atom experiments in microgravity.KCL-PH-TH/2019-65, CERN-TH-2019-126
We introduce a general model for a network of quantum sensors, and we use this model to consider the following question: When can entanglement between the sensors, and/or global measurements, enhance the precision with which the network can measure a set of unknown parameters? We rigorously answer this question by presenting precise theorems proving that for a broad class of problems there is, at most, a very limited intrinsic advantage to using entangled states or global measurements. Moreover, for many estimation problems separable states and local measurements are optimal, and can achieve the ultimate quantum limit on the estimation uncertainty. This immediately implies that there are broad conditions under which simultaneous estimation of multiple parameters cannot outperform individual, independent estimations. Our results apply to any situation in which spatially localized sensors are unitarily encoded with independent parameters, such as when estimating multiple linear or nonlinear optical phase shifts in quantum imaging, or when mapping out the spatial profile of an unknown magnetic field. We conclude by showing that entangling the sensors can enhance the estimation precision when the parameters of interest are global properties of the entire network.
We discuss a scheme for using entangled Bose-Einstein condensates to detect phase differences with a resolution better than the standard quantum limit. To date, schemes have shown that the enhancement in phase resolution gained by entangling condensates is lost when dissipation is present. Here we show how this can be overcome by using number correlated condensates, as have been produced recently in the laboratory. We also outline a scheme for measuring this phase that is not destroyed when the effects of finite detector efficiency are considered.
Many results in the quantum metrology literature use the Cramér-Rao bound and the Fisher information to compare different quantum estimation strategies. However, there are several assumptions that go into the construction of these tools, and these limitations are sometimes not taken into account. While a strategy that utilizes this method can considerably simplify the problem and is valid asymptotically, to have a rigorous and fair comparison we need to adopt a more general approach. In this work we use a methodology based on Bayesian inference to understand what happens when the Cramér-Rao bound is not valid. In particular we quantify the impact of these restrictions on the overall performance of a wide range of schemes including those commonly employed for the estimation of optical phases. We calculate the number of observations and the minimum prior knowledge that are needed such that the Cramér-Rao bound is a valid approximation. Since these requirements are state-dependent, the usual conclusions that can be drawn from the standard methods do not always hold when the analysis is more carefully performed. These results have important implications for the analysis of theory and experiments in quantum metrology.
Quantum metrology protocols are typically designed around the assumption that we have an abundance of measurement data, but recent practical applications are increasingly driving interest in cases with very limited data. In this regime the best approach involves an interesting interplay between the amount of data and the prior information. Here we propose a new way of optimising these schemes based on the practically-motivated assumption that we have a sequence of identical and independent measurements. For a given probe state we take our measurement to be the best one for a single shot and we use this sequentially to study the performance of different practical states in a Mach-Zehnder interferometer when we have moderate prior knowledge of the underlying parameter. We find that we recover the quantum Cramér-Rao bound asymptotically, but for low data counts we find a completely different structure. Despite the fact that intra-mode correlations are known to be the key to increasing the asymptotic precision, we find evidence that these could be detrimental in the low data regime and that entanglement between the paths of the interferometer may play a more important role. Finally, we analyse how close realistic measurements can get to the bound and find that measuring quadratures can improve upon counting photons, though both strategies converge asymptotically. These results may prove to be important in the development of quantum enhanced metrology applications where practical considerations mean that we are limited to a small number of trials.the experiment enough times and that we have certain prior knowledge about the unknown parameter 2 [4,5,18,19] (see footnote 2), and this simplifies the optimisation of the error considerably. Furthermore, the Fisher information has a certain fundamental character. In particular, it can be seen as a distinguishability metric [20] that arises in the expansion of the Bures distance between two infinitesimally close states [3,18]. Moreover, its reciprocal gives us the asymptotic limit for the Bayesian mean square error as a function of the number of repetitions under some fairly general assumptions 3 [5] (see footnote 3), and this is also the case for other approaches that are more conservative than the Cramér-Rao bound too [21,22].Nevertheless, the fact that this technique normally requires many repetitions to be useful is an important drawback to study realistic physical systems such as those previously mentioned. This problem has already been acknowledged in the literature (e.g. in [4, 5, 18, 27]), and several solutions have been proposed. A conceptually simple and straightforward approach consists in using a general measure of uncertainty and estimating how many measurements are needed such that the results predicted by the asymptotic theory are valid, which can always be done numerically [5,28]. In addition, we can rely on numerical techniques such as Monte Carlo simulations [29] or machine learning [30] to perform the optimisation directly, or can simply examine the behavi...
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.