3,171 research outputs found
Comparing compact binary parameter distributions I: Methods
Being able to measure each merger's sky location, distance, component masses,
and conceivably spins, ground-based gravitational-wave detectors will provide a
extensive and detailed sample of coalescing compact binaries (CCBs) in the
local and, with third-generation detectors, distant universe. These
measurements will distinguish between competing progenitor formation models. In
this paper we develop practical tools to characterize the amount of
experimentally accessible information available, to distinguish between two a
priori progenitor models. Using a simple time-independent model, we demonstrate
the information content scales strongly with the number of observations. The
exact scaling depends on how significantly mass distributions change between
similar models. We develop phenomenological diagnostics to estimate how many
models can be distinguished, using first-generation and future instruments.
Finally, we emphasize that multi-observable distributions can be fully
exploited only with very precisely calibrated detectors, search pipelines,
parameter estimation, and Bayesian model inference
How `hot' are mixed quantum states?
Given a mixed quantum state of a qudit, we consider any observable
as a kind of `thermometer' in the following sense. Given a source which emits
pure states with these or those distributions, we select such distributions
that the appropriate average value of the observable is equal to the
average Tr of in the stare . Among those distributions we find
the most typical one, namely, having the highest differential entropy. We call
this distribution conditional Gibbs ensemble as it turns out to be a Gibbs
distribution characterized by a temperature-like parameter . The
expressions establishing the liaisons between the density operator and
its temperature parameter are provided. Within this approach, the
uniform mixed state has the highest `temperature', which tends to zero as the
state in question approaches to a pure state.Comment: Contribution to Quantum 2006: III workshop ad memoriam of Carlo
Novero: Advances in Foundations of Quantum Mechanics and Quantum Information
with atoms and photons. 2-5 May 2006 - Turin, Ital
zCap: a zero configuration adaptive paging and mobility management mechanism
Today, cellular networks rely on fixed collections of cells (tracking areas) for user equipment localisation. Locating users within these areas involves broadcast search (paging), which consumes radio bandwidth but reduces the user equipment signalling required for mobility management. Tracking areas are today manually configured, hard to adapt to local mobility and influence the load on several key resources in the network. We propose a decentralised and self-adaptive approach to mobility management based on a probabilistic model of local mobility. By estimating the parameters of this model from observations of user mobility collected online, we obtain a dynamic model from which we construct local neighbourhoods of cells where we are most likely to locate user equipment. We propose to replace the static tracking areas of current systems with neighbourhoods local to each cell. The model is also used to derive a multi-phase paging scheme, where the division of neighbourhood cells into consecutive phases balances response times and paging cost. The complete mechanism requires no manual tracking area configuration and performs localisation efficiently in terms of signalling and response times. Detailed simulations show that significant potential gains in localisation effi- ciency are possible while eliminating manual configuration of mobility management parameters. Variants of the proposal can be implemented within current (LTE) standards
Universality of optimal measurements
We present optimal and minimal measurements on identical copies of an unknown
state of a qubit when the quality of measuring strategies is quantified with
the gain of information (Kullback of probability distributions). We also show
that the maximal gain of information occurs, among isotropic priors, when the
state is known to be pure. Universality of optimal measurements follows from
our results: using the fidelity or the gain of information, two different
figures of merits, leads to exactly the same conclusions. We finally
investigate the optimal capacity of copies of an unknown state as a quantum
channel of information.Comment: Revtex, 5 pages, no figure
Gradient Flows in Filtering and Fisher-Rao Geometry
Uncertainty propagation and filtering can be interpreted as gradient flows
with respect to suitable metrics in the infinite dimensional manifold of
probability density functions. Such a viewpoint has been put forth in recent
literature, and a systematic way to formulate and solve the same for linear
Gaussian systems has appeared in our previous work where the gradient flows
were realized via proximal operators with respect to Wasserstein metric arising
in optimal mass transport. In this paper, we derive the evolution equations as
proximal operators with respect to Fisher-Rao metric arising in information
geometry. We develop the linear Gaussian case in detail and show that a
template two step optimization procedure proposed earlier by the authors still
applies. Our objective is to provide new geometric interpretations of known
equations in filtering, and to clarify the implication of different choices of
metric
On a comparative study between dependence scales determined by linear and non-linear measures
In this manuscript we present a comparative study about the determination of
the relaxation (\textit{i.e.}, independence) time scales obtained from the
correlation function, the mutual information, and a criterion based on the
evaluation of a nonextensive generalisation of mutual entropy. Our results show
that, for systems with a small degree of complexity, standard mutual
information and the criterion based on its nonextensive generalisation provide
the same scale, whereas for systems with a higher complex dynamics the standard
mutual information presents a time scale consistently smaller.Comment: 14 pages. To appear in Physica
Pairwise Confusion for Fine-Grained Visual Classification
Fine-Grained Visual Classification (FGVC) datasets contain small sample
sizes, along with significant intra-class variation and inter-class similarity.
While prior work has addressed intra-class variation using localization and
segmentation techniques, inter-class similarity may also affect feature
learning and reduce classification performance. In this work, we address this
problem using a novel optimization procedure for the end-to-end neural network
training on FGVC tasks. Our procedure, called Pairwise Confusion (PC) reduces
overfitting by intentionally {introducing confusion} in the activations. With
PC regularization, we obtain state-of-the-art performance on six of the most
widely-used FGVC datasets and demonstrate improved localization ability. {PC}
is easy to implement, does not need excessive hyperparameter tuning during
training, and does not add significant overhead during test time.Comment: Camera-Ready version for ECCV 201
WIMP astronomy and particle physics with liquid-noble and cryogenic direct-detection experiments
Once weakly-interacting massive particles (WIMPs) are unambiguously detected
in direct-detection experiments, the challenge will be to determine what one
may infer from the data. Here, I examine the prospects for reconstructing the
local speed distribution of WIMPs in addition to WIMP particle-physics
properties (mass, cross sections) from next-generation cryogenic and
liquid-noble direct-detection experiments. I find that the common method of
fixing the form of the velocity distribution when estimating constraints on
WIMP mass and cross sections means losing out on the information on the speed
distribution contained in the data and may lead to biases in the inferred
values of the particle-physics parameters. I show that using a more general,
empirical form of the speed distribution can lead to good constraints on the
speed distribution. Moreover, one can use Bayesian model-selection criteria to
determine if a theoretically-inspired functional form for the speed
distribution (such as a Maxwell-Boltzmann distribution) fits better than an
empirical model. The shape of the degeneracy between WIMP mass and cross
sections and their offset from the true values of those parameters depends on
the hypothesis for the speed distribution, which has significant implications
for consistency checks between direct-detection and collider data. In addition,
I find that the uncertainties on theoretical parameters depends sensitively on
the upper end of the energy range used for WIMP searches. Better constraints on
the WIMP particle-physics parameters and speed distribution are obtained if the
WIMP search is extended to higher energy (~ 1 MeV).Comment: 25 pages, 27 figures, matches published versio
Compressing Probability Distributions
We show how to store good approximations of probability distributions in
small space
- …
