302 research outputs found
EIT ground-state cooling of long ion strings
Electromagnetically-induced-transparency (EIT) cooling is a ground-state
cooling technique for trapped particles. EIT offers a broader cooling range in
frequency space compared to more established methods. In this work, we
experimentally investigate EIT cooling in strings of trapped atomic ions. In
strings of up to 18 ions, we demonstrate simultaneous ground state cooling of
all radial modes in under 1 ms. This is a particularly important capability in
view of emerging quantum simulation experiments with large numbers of trapped
ions. Our analysis of the EIT cooling dynamics is based on a novel technique
enabling single-shot measurements of phonon numbers, by rapid adiabatic passage
on a vibrational sideband of a narrow transition
Ask Language Model to Clean Your Noisy Translation Data
TTransformer models have demonstrated remarkable performance in neural machine translation (NMT). However, their vulnerability to noisy input poses a significant challenge in practical implementation, where generating clean output from noisy input is crucial. The MTNT dataset is widely used as a benchmark for evaluating the robustness of NMT models against noisy input. Nevertheless, its utility is limited due to the presence of noise in both the source and target sentences. To address this limitation, we focus on cleaning the noise from the target sentences in MTNT, making it more suitable as a benchmark for noise evaluation. Leveraging the capabilities of large language models (LLMs), we observe their impressive abilities in noise removal. For example, they can remove emojis while considering their semantic meaning. Additionally, we show that LLM can effectively rephrase slang, jargon, and profanities. The resulting datasets, called C-MTNT, exhibit significantly less noise in the target sentences while preserving the semantic integrity of the original sentences. Our human and GPT-4 evaluations also lead to a consistent conclusion that LLM performs well on this task. Lastly, experiments on C-MTNT showcased its effectiveness in evaluating the robustness of NMT models, highlighting the potential of advanced language models for data cleaning and emphasizing C-MTNT as a valuable resource
Rank-based model selection for multiple ions quantum tomography
The statistical analysis of measurement data has become a key component of
many quantum engineering experiments. As standard full state tomography becomes
unfeasible for large dimensional quantum systems, one needs to exploit prior
information and the "sparsity" properties of the experimental state in order to
reduce the dimensionality of the estimation problem. In this paper we propose
model selection as a general principle for finding the simplest, or most
parsimonious explanation of the data, by fitting different models and choosing
the estimator with the best trade-off between likelihood fit and model
complexity. We apply two well established model selection methods -- the Akaike
information criterion (AIC) and the Bayesian information criterion (BIC) -- to
models consising of states of fixed rank and datasets such as are currently
produced in multiple ions experiments. We test the performance of AIC and BIC
on randomly chosen low rank states of 4 ions, and study the dependence of the
selected rank with the number of measurement repetitions for one ion states. We
then apply the methods to real data from a 4 ions experiment aimed at creating
a Smolin state of rank 4. The two methods indicate that the optimal model for
describing the data lies between ranks 6 and 9, and the Pearson test
is applied to validate this conclusion. Additionally we find that the mean
square error of the maximum likelihood estimator for pure states is close to
that of the optimal over all possible measurements.Comment: 24 pages, 6 figures, 3 table
Ask Language Model to Clean Your Noisy Translation Data
TTransformer models have demonstrated remarkable performance in neural machine translation (NMT). However, their vulnerability to noisy input poses a significant challenge in practical implementation, where generating clean output from noisy input is crucial. The MTNT dataset is widely used as a benchmark for evaluating the robustness of NMT models against noisy input. Nevertheless, its utility is limited due to the presence of noise in both the source and target sentences. To address this limitation, we focus on cleaning the noise from the target sentences in MTNT, making it more suitable as a benchmark for noise evaluation. Leveraging the capabilities of large language models (LLMs), we observe their impressive abilities in noise removal. For example, they can remove emojis while considering their semantic meaning. Additionally, we show that LLM can effectively rephrase slang, jargon, and profanities. The resulting datasets, called C-MTNT, exhibit significantly less noise in the target sentences while preserving the semantic integrity of the original sentences. Our human and GPT-4 evaluations also lead to a consistent conclusion that LLM performs well on this task. Lastly, experiments on C-MTNT showcased its effectiveness in evaluating the robustness of NMT models, highlighting the potential of advanced language models for data cleaning and emphasizing C-MTNT as a valuable resource
Low-dimensional quite noisy bound entanglement with cryptographic key
We provide a class of bound entangled states that have positive distillable
secure key rate. The smallest state of this kind is 4 \bigotimes 4. Our class
is a generalization of the class presented in [1] (IEEE Trans. Inf. Theory 54,
2621 (2008); arXiv:quant-ph/0506203). It is much wider, containing, in
particular, states from the boundary of PPT entangled states (all of the states
in the class in [1] were of this kind) but also states inside the set of PPT
entangled states, even, approaching the separable states. This generalization
comes with a price: for the wider class a positive key rate requires, in
general, apart from the one-way Devetak-Winter protocol (used in [1]) also the
recurrence preprocessing and thus effectively is a two-way protocol. We also
analyze the amount of noise that can be admixtured to the states of our class
without losing key distillability property which may be crucial for
experimental realization. The wider class contains key-distillable states with
higher entropy (up to 3.524, as opposed to 2.564 for the class in [1]).Comment: 10 pages, final version for J. Phys. A: Math. Theo
Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
Parameter-efficient fine-tuning (PEFT) of pre-trained language models (PLMs) has emerged as a highly successful approach, with training only a small number of parameters without sacrificing performance and becoming the de-facto learning paradigm with the increasing size of PLMs. However, existing PEFT methods are not memory-efficient, because they still require caching most of the intermediate activations for the gradient calculation, akin to fine-tuning. One effective way to reduce the activation memory is to apply a reversible model, so the intermediate activations are not necessary to be cached and can be recomputed. Nevertheless, modifying a PLM to its reversible variant is not straightforward, since the reversible model has a distinct architecture from the currently released PLMs. In this paper, we first investigate what is a key factor for the success of existing PEFT methods, and realize that it's essential to preserve the PLM's starting point when initializing a PEFT method. With this finding, we propose memory-efficient fine-tuning (MEFT) that inserts adapters into a PLM, preserving the PLM's starting point and making it reversible without additional pre-training. We evaluate MEFT on the GLUE benchmark and five question-answering tasks with various backbones, BERT, RoBERTa, BART and OPT. MEFT significantly reduces the activation memory up to 84% of full fine-tuning with a negligible amount of trainable parameters. Moreover, MEFT achieves the same score on GLUE and a comparable score on the question-answering tasks as full fine-tuning. A similar finding is also observed for the image classification task
Parameter-Efficient Fine-Tuning without Introducing New Latency
Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently demonstrated remarkable achievements, effectively matching the performance of full fine-tuning while utilizing significantly fewer trainable parameters, and consequently addressing the storage and communication constraints. Nonetheless, various PEFT methods are limited by their inherent characteristics. In the case of sparse fine-tuning, which involves modifying only a small subset of the existing parameters, the selection of fine-tuned parameters is task- and domain-specific, making it unsuitable for federated learning. On the other hand, PEFT methods with adding new parameters typically introduce additional inference latency. In this paper, we demonstrate the feasibility of generating a sparse mask in a task-agnostic manner, wherein all downstream tasks share a common mask. Our approach, which relies solely on the magnitude information of pre-trained parameters, surpasses existing methodologies by a significant margin when evaluated on the GLUE benchmark. Additionally, we introduce a novel adapter technique that directly applies the adapter to pre-trained parameters instead of the hidden representation, thereby achieving identical inference speed to that of full fine-tuning. Through extensive experiments, our proposed method attains a new state-of-the-art outcome in terms of both performance and storage efficiency, storing only 0.03% parameters of full fine-tuning
Additivity and non-additivity of multipartite entanglement measures
We study the additivity property of three multipartite entanglement measures,
i.e. the geometric measure of entanglement (GM), the relative entropy of
entanglement and the logarithmic global robustness. First, we show the
additivity of GM of multipartite states with real and non-negative entries in
the computational basis. Many states of experimental and theoretical interests
have this property, e.g. Bell diagonal states, maximally correlated generalized
Bell diagonal states, generalized Dicke states, the Smolin state, and the
generalization of D\"{u}r's multipartite bound entangled states. We also prove
the additivity of other two measures for some of these examples. Second, we
show the non-additivity of GM of all antisymmetric states of three or more
parties, and provide a unified explanation of the non-additivity of the three
measures of the antisymmetric projector states. In particular, we derive
analytical formulae of the three measures of one copy and two copies of the
antisymmetric projector states respectively. Third, we show, with a statistical
approach, that almost all multipartite pure states with sufficiently large
number of parties are nearly maximally entangled with respect to GM and
relative entropy of entanglement. However, their GM is not strong additive;
what's more surprising, for generic pure states with real entries in the
computational basis, GM of one copy and two copies, respectively, are almost
equal. Hence, more states may be suitable for universal quantum computation, if
measurements can be performed on two copies of the resource states. We also
show that almost all multipartite pure states cannot be produced reversibly
with the combination multipartite GHZ states under asymptotic LOCC, unless
relative entropy of entanglement is non-additive for generic multipartite pure
states.Comment: 45 pages, 4 figures. Proposition 23 and Theorem 24 are revised by
correcting a minor error from Eq. (A.2), (A.3) and (A.4) in the published
version. The abstract, introduction, and summary are also revised. All other
conclusions are unchange
- …
