48,938 research outputs found

    Hard X-ray emission cutoff in anomalous X-ray pulsar 4U 0142+61 detected by INTEGRAL

    Full text link
    The anomalous X-ray pulsar 4U 0142+61 was studied by the INTEGRAL observations. The hard X-ray spectrum of 18 -- 500 keV for 4U 0142+61 was derived using near 9 years of INTEGRAL/IBIS data. We obtained the average hard X-ray spectrum of 4U 0142+61 with all available data. The spectrum of 4U 0142+61 can be fitted with a power-law with an exponential high energy cutoff. This average spectrum is well fitted with a power-law of Γ0.51±0.11\Gamma\sim 0.51\pm 0.11 plus a cutoff energy at 128.6±17.2128.6\pm 17.2 keV. The hard X-ray flux of the source from 20 -- 150 keV showed no significant variations (within 20%\%) from 2003 -- 2011. The spectral profiles have some variability in nine years: photon index varied from 0.3 -- 1.5, and cutoff energies of 110 -- 250 keV. The detection of the high energy cutoff around 130 keV shows some constraints on the radiation mechanisms of magnetars and possibly probes the differences between magnetar and accretion models for these special class of neutron stars. Future HXMT observations could provide stronger constraints on the hard X-ray spectral properties of this source and other magnetar candidates.Comment: 9 pages, 5 figures, 2 tables, figures are updated, new data are added, conclusion does not change, to be published in RA

    Superfluidity enhanced by spin-flip tunnelling in the presence of a magnetic field

    Full text link
    It is well-known that when the magnetic field is stronger than a critical value, the spin imbalance can break the Cooper pairs of electrons and hence hinder the superconductivity in a spin-singlet channel. In a bilayer system of ultra-cold Fermi gases, however, we demonstrate that the critical value of the magnetic field at zero temperature can be significantly increased by including a spin-flip tunnelling, which opens a gap in the spin-triplet channel near the Fermi surface and hence reduces the influence of the effective magnetic field on the superfluidity. The phase transition also changes from first order to second order when the tunnelling exceeds a critical value. Considering a realistic experiment, this mechanism can be implemented by applying an intralayer Raman coupling between the spin states with a phase difference between the two layers.Comment: 10+4 pages, 8 figure

    LERC: Coordinated Cache Management for Data-Parallel Systems

    Full text link
    Memory caches are being aggressively used in today's data-parallel frameworks such as Spark, Tez and Storm. By caching input and intermediate data in memory, compute tasks can witness speedup by orders of magnitude. To maximize the chance of in-memory data access, existing cache algorithms, be it recency- or frequency-based, settle on cache hit ratio as the optimization objective. However, unlike the conventional belief, we show in this paper that simply pursuing a higher cache hit ratio of individual data blocks does not necessarily translate into faster task completion in data-parallel environments. A data-parallel task typically depends on multiple input data blocks. Unless all of these blocks are cached in memory, no speedup will result. To capture this all-or-nothing property, we propose a more relevant metric, called effective cache hit ratio. Specifically, a cache hit of a data block is said to be effective if it can speed up a compute task. In order to optimize the effective cache hit ratio, we propose the Least Effective Reference Count (LERC) policy that persists the dependent blocks of a compute task as a whole in memory. We have implemented the LERC policy as a memory manager in Spark and evaluated its performance through Amazon EC2 deployment. Evaluation results demonstrate that LERC helps speed up data-parallel jobs by up to 37% compared with the widely employed least-recently-used (LRU) policy

    LRC: Dependency-Aware Cache Management for Data Analytics Clusters

    Full text link
    Memory caches are being aggressively used in today's data-parallel systems such as Spark, Tez, and Piccolo. However, prevalent systems employ rather simple cache management policies--notably the Least Recently Used (LRU) policy--that are oblivious to the application semantics of data dependency, expressed as a directed acyclic graph (DAG). Without this knowledge, memory caching can at best be performed by "guessing" the future data access patterns based on historical information (e.g., the access recency and/or frequency), which frequently results in inefficient, erroneous caching with low hit ratio and a long response time. In this paper, we propose a novel cache replacement policy, Least Reference Count (LRC), which exploits the application-specific DAG information to optimize the cache management. LRC evicts the cached data blocks whose reference count is the smallest. The reference count is defined, for each data block, as the number of dependent child blocks that have not been computed yet. We demonstrate the efficacy of LRC through both empirical analysis and cluster deployments against popular benchmarking workloads. Our Spark implementation shows that, compared with LRU, LRC speeds up typical applications by 60%.Comment: 9 page
    corecore