302 research outputs found

    EIT ground-state cooling of long ion strings

    Get PDF
    Electromagnetically-induced-transparency (EIT) cooling is a ground-state cooling technique for trapped particles. EIT offers a broader cooling range in frequency space compared to more established methods. In this work, we experimentally investigate EIT cooling in strings of trapped atomic ions. In strings of up to 18 ions, we demonstrate simultaneous ground state cooling of all radial modes in under 1 ms. This is a particularly important capability in view of emerging quantum simulation experiments with large numbers of trapped ions. Our analysis of the EIT cooling dynamics is based on a novel technique enabling single-shot measurements of phonon numbers, by rapid adiabatic passage on a vibrational sideband of a narrow transition

    Ask Language Model to Clean Your Noisy Translation Data

    Get PDF
    TTransformer models have demonstrated remarkable performance in neural machine translation (NMT). However, their vulnerability to noisy input poses a significant challenge in practical implementation, where generating clean output from noisy input is crucial. The MTNT dataset is widely used as a benchmark for evaluating the robustness of NMT models against noisy input. Nevertheless, its utility is limited due to the presence of noise in both the source and target sentences. To address this limitation, we focus on cleaning the noise from the target sentences in MTNT, making it more suitable as a benchmark for noise evaluation. Leveraging the capabilities of large language models (LLMs), we observe their impressive abilities in noise removal. For example, they can remove emojis while considering their semantic meaning. Additionally, we show that LLM can effectively rephrase slang, jargon, and profanities. The resulting datasets, called C-MTNT, exhibit significantly less noise in the target sentences while preserving the semantic integrity of the original sentences. Our human and GPT-4 evaluations also lead to a consistent conclusion that LLM performs well on this task. Lastly, experiments on C-MTNT showcased its effectiveness in evaluating the robustness of NMT models, highlighting the potential of advanced language models for data cleaning and emphasizing C-MTNT as a valuable resource

    Rank-based model selection for multiple ions quantum tomography

    Get PDF
    The statistical analysis of measurement data has become a key component of many quantum engineering experiments. As standard full state tomography becomes unfeasible for large dimensional quantum systems, one needs to exploit prior information and the "sparsity" properties of the experimental state in order to reduce the dimensionality of the estimation problem. In this paper we propose model selection as a general principle for finding the simplest, or most parsimonious explanation of the data, by fitting different models and choosing the estimator with the best trade-off between likelihood fit and model complexity. We apply two well established model selection methods -- the Akaike information criterion (AIC) and the Bayesian information criterion (BIC) -- to models consising of states of fixed rank and datasets such as are currently produced in multiple ions experiments. We test the performance of AIC and BIC on randomly chosen low rank states of 4 ions, and study the dependence of the selected rank with the number of measurement repetitions for one ion states. We then apply the methods to real data from a 4 ions experiment aimed at creating a Smolin state of rank 4. The two methods indicate that the optimal model for describing the data lies between ranks 6 and 9, and the Pearson χ2\chi^{2} test is applied to validate this conclusion. Additionally we find that the mean square error of the maximum likelihood estimator for pure states is close to that of the optimal over all possible measurements.Comment: 24 pages, 6 figures, 3 table

    Ask Language Model to Clean Your Noisy Translation Data

    Get PDF
    TTransformer models have demonstrated remarkable performance in neural machine translation (NMT). However, their vulnerability to noisy input poses a significant challenge in practical implementation, where generating clean output from noisy input is crucial. The MTNT dataset is widely used as a benchmark for evaluating the robustness of NMT models against noisy input. Nevertheless, its utility is limited due to the presence of noise in both the source and target sentences. To address this limitation, we focus on cleaning the noise from the target sentences in MTNT, making it more suitable as a benchmark for noise evaluation. Leveraging the capabilities of large language models (LLMs), we observe their impressive abilities in noise removal. For example, they can remove emojis while considering their semantic meaning. Additionally, we show that LLM can effectively rephrase slang, jargon, and profanities. The resulting datasets, called C-MTNT, exhibit significantly less noise in the target sentences while preserving the semantic integrity of the original sentences. Our human and GPT-4 evaluations also lead to a consistent conclusion that LLM performs well on this task. Lastly, experiments on C-MTNT showcased its effectiveness in evaluating the robustness of NMT models, highlighting the potential of advanced language models for data cleaning and emphasizing C-MTNT as a valuable resource

    Low-dimensional quite noisy bound entanglement with cryptographic key

    Full text link
    We provide a class of bound entangled states that have positive distillable secure key rate. The smallest state of this kind is 4 \bigotimes 4. Our class is a generalization of the class presented in [1] (IEEE Trans. Inf. Theory 54, 2621 (2008); arXiv:quant-ph/0506203). It is much wider, containing, in particular, states from the boundary of PPT entangled states (all of the states in the class in [1] were of this kind) but also states inside the set of PPT entangled states, even, approaching the separable states. This generalization comes with a price: for the wider class a positive key rate requires, in general, apart from the one-way Devetak-Winter protocol (used in [1]) also the recurrence preprocessing and thus effectively is a two-way protocol. We also analyze the amount of noise that can be admixtured to the states of our class without losing key distillability property which may be crucial for experimental realization. The wider class contains key-distillable states with higher entropy (up to 3.524, as opposed to 2.564 for the class in [1]).Comment: 10 pages, final version for J. Phys. A: Math. Theo

    Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning

    Get PDF
    Parameter-efficient fine-tuning (PEFT) of pre-trained language models (PLMs) has emerged as a highly successful approach, with training only a small number of parameters without sacrificing performance and becoming the de-facto learning paradigm with the increasing size of PLMs. However, existing PEFT methods are not memory-efficient, because they still require caching most of the intermediate activations for the gradient calculation, akin to fine-tuning. One effective way to reduce the activation memory is to apply a reversible model, so the intermediate activations are not necessary to be cached and can be recomputed. Nevertheless, modifying a PLM to its reversible variant is not straightforward, since the reversible model has a distinct architecture from the currently released PLMs. In this paper, we first investigate what is a key factor for the success of existing PEFT methods, and realize that it's essential to preserve the PLM's starting point when initializing a PEFT method. With this finding, we propose memory-efficient fine-tuning (MEFT) that inserts adapters into a PLM, preserving the PLM's starting point and making it reversible without additional pre-training. We evaluate MEFT on the GLUE benchmark and five question-answering tasks with various backbones, BERT, RoBERTa, BART and OPT. MEFT significantly reduces the activation memory up to 84% of full fine-tuning with a negligible amount of trainable parameters. Moreover, MEFT achieves the same score on GLUE and a comparable score on the question-answering tasks as full fine-tuning. A similar finding is also observed for the image classification task

    Parameter-Efficient Fine-Tuning without Introducing New Latency

    Get PDF
    Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently demonstrated remarkable achievements, effectively matching the performance of full fine-tuning while utilizing significantly fewer trainable parameters, and consequently addressing the storage and communication constraints. Nonetheless, various PEFT methods are limited by their inherent characteristics. In the case of sparse fine-tuning, which involves modifying only a small subset of the existing parameters, the selection of fine-tuned parameters is task- and domain-specific, making it unsuitable for federated learning. On the other hand, PEFT methods with adding new parameters typically introduce additional inference latency. In this paper, we demonstrate the feasibility of generating a sparse mask in a task-agnostic manner, wherein all downstream tasks share a common mask. Our approach, which relies solely on the magnitude information of pre-trained parameters, surpasses existing methodologies by a significant margin when evaluated on the GLUE benchmark. Additionally, we introduce a novel adapter technique that directly applies the adapter to pre-trained parameters instead of the hidden representation, thereby achieving identical inference speed to that of full fine-tuning. Through extensive experiments, our proposed method attains a new state-of-the-art outcome in terms of both performance and storage efficiency, storing only 0.03% parameters of full fine-tuning

    Additivity and non-additivity of multipartite entanglement measures

    Full text link
    We study the additivity property of three multipartite entanglement measures, i.e. the geometric measure of entanglement (GM), the relative entropy of entanglement and the logarithmic global robustness. First, we show the additivity of GM of multipartite states with real and non-negative entries in the computational basis. Many states of experimental and theoretical interests have this property, e.g. Bell diagonal states, maximally correlated generalized Bell diagonal states, generalized Dicke states, the Smolin state, and the generalization of D\"{u}r's multipartite bound entangled states. We also prove the additivity of other two measures for some of these examples. Second, we show the non-additivity of GM of all antisymmetric states of three or more parties, and provide a unified explanation of the non-additivity of the three measures of the antisymmetric projector states. In particular, we derive analytical formulae of the three measures of one copy and two copies of the antisymmetric projector states respectively. Third, we show, with a statistical approach, that almost all multipartite pure states with sufficiently large number of parties are nearly maximally entangled with respect to GM and relative entropy of entanglement. However, their GM is not strong additive; what's more surprising, for generic pure states with real entries in the computational basis, GM of one copy and two copies, respectively, are almost equal. Hence, more states may be suitable for universal quantum computation, if measurements can be performed on two copies of the resource states. We also show that almost all multipartite pure states cannot be produced reversibly with the combination multipartite GHZ states under asymptotic LOCC, unless relative entropy of entanglement is non-additive for generic multipartite pure states.Comment: 45 pages, 4 figures. Proposition 23 and Theorem 24 are revised by correcting a minor error from Eq. (A.2), (A.3) and (A.4) in the published version. The abstract, introduction, and summary are also revised. All other conclusions are unchange
    corecore