6,971 research outputs found
Precision voltage regulator
Balanced positive and negative voltage output circuit, in which error voltage for control is developed from difference in absolute value of positive and negative voltages referenced to a common point, regulates voltage for use with inertial reference unit. Fast-acting, temperature-compensated, high-gain operational amplifier circuits maintain common point
Navigation and interaction in a real-scale digital mock-up using natural language and user gesture
This paper tries to demonstrate a very new real-scale 3D system and sum up some firsthand and cutting edge results concerning multi-modal navigation and interaction interfaces. This work is part of the CALLISTO-SARI collaborative project. It aims at constructing an immersive room, developing a set of software tools and some navigation/interaction interfaces. Two sets of interfaces will be introduced here: 1) interaction devices, 2) natural language (speech processing) and user gesture. The survey on this system using subjective observation (Simulator Sickness Questionnaire, SSQ) and objective measurements (Center of Gravity, COG) shows that using natural languages and gesture-based interfaces induced less cyber-sickness comparing to device-based interfaces. Therefore, gesture-based is more efficient than device-based interfaces.FUI CALLISTO-SAR
Recommended from our members
FutureGRID: A Program for long-term research into GRID systems architecture
Proceedings of the 2003 UK e-Science All Hands Meeting, 31st August - 3rd September, Nottingham UKThis is a project to carry out research into long-term GRID architecture, in the University of Cambridge
Computer Laboratory and the Cambridge eScience Center, with support from the Microsoft Research
Laboratory, Cambridge.
It is part of a larger vision for future systems architectures for public computing platforms, including
both scientitic GRID and commodity level computing such as games, peer2peer computing and storage
services and so forth, based on work in the laboratories in recent years into massively scaleable distributed systems for storage, computation, content distribution and collaboration[26]
Measurement Invariance, Entropy, and Probability
We show that the natural scaling of measurement for a particular problem
defines the most likely probability distribution of observations taken from
that measurement scale. Our approach extends the method of maximum entropy to
use measurement scale as a type of information constraint. We argue that a very
common measurement scale is linear at small magnitudes grading into logarithmic
at large magnitudes, leading to observations that often follow Student's
probability distribution which has a Gaussian shape for small fluctuations from
the mean and a power law shape for large fluctuations from the mean. An inverse
scaling often arises in which measures naturally grade from logarithmic to
linear as one moves from small to large magnitudes, leading to observations
that often follow a gamma probability distribution. A gamma distribution has a
power law shape for small magnitudes and an exponential shape for large
magnitudes. The two measurement scales are natural inverses connected by the
Laplace integral transform. This inversion connects the two major scaling
patterns commonly found in nature. We also show that superstatistics is a
special case of an integral transform, and thus can be understood as a
particular way in which to change the scale of measurement. Incorporating
information about measurement scale into maximum entropy provides a general
approach to the relations between measurement, information and probability
Electronic integrator for gyro rate output voltages
Circuit which integrates spacecraft gyro output voltages to provide analog position signals has been developed. Accurate integration is provided by all solid state system which uses no choppers and takes advantage of commercially available flight qualified components
Q-based design equations for resonant metamaterials and experimental validation
Practical design parameters of resonant metamaterials, such as loss tangent,
are derived in terms of the quality factor of the resonant effective medium
permeability or permittivity. Through electromagnetic simulations of loop-based
resonant particles, it is also shown that the of the effective medium
response is essentially equal to the of an individual resonant particle.
Thus, by measuring the of a single fabricated metamaterial particle, the
effective permeability or permittivity of a metamaterial can be calculated
simply and accurately without requiring complex simulations, fabrication, or
measurements. Experimental validation shows that the complex permeability
analytically estimated from the measured of a single fabricated
self-resonant loop agrees with the complex permeability extracted from
parameter measurements of a metamaterial slab to better than 20%. This
equivalence reduces the design of a metamaterial to meet a given loss
constraint to the simpler problem of the design of a resonant particle to meet
a specific constraint. This analysis also yields simple analytical
expressions for estimating the loss tangent of a planar loop magnetic
metamaterial due to ohmic losses. It is shown that
is a strong lower bound for magnetic loss tangents for frequencies not too far
from 1 GHz. The ohmic loss of the metamaterial varies inversely with the
electrical size of the metamaterial particle, indicating that there is a loss
penalty for reducing the particle size at a fixed frequency
Investigation of the Gravitational Potential Dependence of the Fine-Structure Constant Using Atomic Dysprosium
Radio-frequency E1 transitions between nearly degenerate, opposite parity
levels of atomic dysprosium were monitored over an eight month period to search
for a variation in the fine-structure constant. During this time period, data
were taken at different points in the gravitational potential of the Sun. The
data are fitted to the variation in the gravitational potential yielding a
value of for the fit parameter . This
value gives the current best laboratory limit. In addition, our value of
combined with other experimental constraints is used to extract
the first limits on k_e and k_q. These coefficients characterize the variation
of m_e/m_p and m_q/m_p in a changing gravitational potential, where m_e, m_p,
and m_q are electron, proton, and quark masses. The results are and .Comment: 6 pages, 3 figure
Bayesian anomaly detection methods for social networks
Learning the network structure of a large graph is computationally demanding,
and dynamically monitoring the network over time for any changes in structure
threatens to be more challenging still. This paper presents a two-stage method
for anomaly detection in dynamic graphs: the first stage uses simple, conjugate
Bayesian models for discrete time counting processes to track the pairwise
links of all nodes in the graph to assess normality of behavior; the second
stage applies standard network inference tools on a greatly reduced subset of
potentially anomalous nodes. The utility of the method is demonstrated on
simulated and real data sets.Comment: Published in at http://dx.doi.org/10.1214/10-AOAS329 the Annals of
Applied Statistics (http://www.imstat.org/aoas/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Efficient estimation of AUC in a sliding window
In many applications, monitoring area under the ROC curve (AUC) in a sliding
window over a data stream is a natural way of detecting changes in the system.
The drawback is that computing AUC in a sliding window is expensive, especially
if the window size is large and the data flow is significant.
In this paper we propose a scheme for maintaining an approximate AUC in a
sliding window of length . More specifically, we propose an algorithm that,
given , estimates AUC within , and can maintain this
estimate in time, per update, as the window slides.
This provides a speed-up over the exact computation of AUC, which requires
time, per update. The speed-up becomes more significant as the size of
the window increases. Our estimate is based on grouping the data points
together, and using these groups to calculate AUC. The grouping is designed
carefully such that () the groups are small enough, so that the error stays
small, () the number of groups is small, so that enumerating them is not
expensive, and () the definition is flexible enough so that we can
maintain the groups efficiently.
Our experimental evaluation demonstrates that the average approximation error
in practice is much smaller than the approximation guarantee ,
and that we can achieve significant speed-ups with only a modest sacrifice in
accuracy
Noticing for Equity to Sustain Multilingual Literacies
This department explores how teachers can sustain students’ multilingual literacies and reimagine literacy learning across multiple contexts in conversation with researchers, practitioners, and communities
- …
