330 research outputs found
Modeling Adversarial Attack on Pre-trained Language Models as Sequential Decision Making
Pre-trained language models (PLMs) have been widely used to underpin various
downstream tasks. However, the adversarial attack task has found that PLMs are
vulnerable to small perturbations. Mainstream methods adopt a detached
two-stage framework to attack without considering the subsequent influence of
substitution at each step. In this paper, we formally model the adversarial
attack task on PLMs as a sequential decision-making problem, where the whole
attack process is sequential with two decision-making problems, i.e., word
finder and word substitution. Considering the attack process can only receive
the final state without any direct intermediate signals, we propose to use
reinforcement learning to find an appropriate sequential attack path to
generate adversaries, named SDM-Attack. Extensive experimental results show
that SDM-Attack achieves the highest attack success rate with a comparable
modification rate and semantic similarity to attack fine-tuned BERT.
Furthermore, our analyses demonstrate the generalization and transferability of
SDM-Attack. The code is available at https://github.com/fduxuan/SDM-Attack
Evolving Connectivity for Recurrent Spiking Neural Networks
Recurrent spiking neural networks (RSNNs) hold great potential for advancing
artificial general intelligence, as they draw inspiration from the biological
nervous system and show promise in modeling complex dynamics. However, the
widely-used surrogate gradient-based training methods for RSNNs are inherently
inaccurate and unfriendly to neuromorphic hardware. To address these
limitations, we propose the evolving connectivity (EC) framework, an
inference-only method for training RSNNs. The EC framework reformulates
weight-tuning as a search into parameterized connection probability
distributions, and employs Natural Evolution Strategies (NES) for optimizing
these distributions. Our EC framework circumvents the need for gradients and
features hardware-friendly characteristics, including sparse boolean
connections and high scalability. We evaluate EC on a series of standard
robotic locomotion tasks, where it achieves comparable performance with deep
neural networks and outperforms gradient-trained RSNNs, even solving the
complex 17-DoF humanoid task. Additionally, the EC framework demonstrates a two
to three fold speedup in efficiency compared to directly evolving parameters.
By providing a performant and hardware-friendly alternative, the EC framework
lays the groundwork for further energy-efficient applications of RSNNs and
advances the development of neuromorphic devices
Speak It Out: Solving Symbol-Related Problems with Symbol-to-Language Conversion for Language Models
Symbols (or more broadly, non-natural language textual representations) such
as numerical sequences, molecular formulas, and table delimiters widely exist,
playing important roles in various tasks such as abstract reasoning, chemical
property prediction, and table question answering. Despite the impressive
natural language comprehension capabilities of large language models (LLMs),
their reasoning abilities for symbols remain inadequate, which could attributed
to the difference between symbol representations and general natural languages.
We propose symbol-to-language (S2L), a tuning-free method that enables large
language models to solve symbol-related problems with information expressed in
natural language. Specifically, S2L first converts the symbols involved to
language-based representations, which can be implemented by prompting LLMs or
leveraging external tools, then these language-based representations are
integrated into the original problem via direct substitution or concatenation,
serving as useful input information for LLMs. We evaluate the S2L method using
both API-based (GPT-4, ChatGPT) and open-source (OpenChat) models over eight
symbol-related tasks, ranging from symbol-only abstract reasoning to sentiment
analysis in social media. Experimental results show that S2L consistently leads
to superior performance. For example, by employing S2L for GPT-4, there can be
average significant improvements of +21.9% and +9.5% for subtasks in 1D-ARC and
Dyck language, respectively. Codes and data are available at
https://github.com/THUNLP-MT/symbol2language.Comment: ICLR AGI Workshop 202
Exercise training improved body composition, cardiovascular function, and physical fitness of 5-year-old children with obesity or normal body mass
Objectives: To explore the effects of exercise training on body composition, cardiovascular function, and physical fitness in 5-year-old obese and lean children.
Methods: 42 obese and 62 lean children were randomly allocated into exercise and control groups separately. Body composition, cardiovascular function, and physical fitness were measured at baseline and the end of the intervention. The exercise groups participated in 10 weeks of supervised moderate intensity exercise training (at 50% of heart rate reserve), 50 training sessions in total.
Results: The physical activity program was successfully completed and no sport injury occurred. Exercise training decreased BMI, waist circumference, body fat%, and fat mass; and slowed down the growth speed of body mass of both trained obese and lean children. Exercise training significantly decreased systolic blood pressure of obese children and decreased their heart rate responses during exercise. Trained obese children improved the performances of long jump, 10-m x 4 shuttle run, and 3-m balance beam walk; while trained lean children improved more items of physical fitness.
Conclusions: 10 weeks of moderate intensity exercise training is an effective and safe treatment for children aged five years, either obese or with normal body mass
DEEM: Dynamic Experienced Expert Modeling for Stance Detection
Recent work has made a preliminary attempt to use large language models
(LLMs) to solve the stance detection task, showing promising results. However,
considering that stance detection usually requires detailed background
knowledge, the vanilla reasoning method may neglect the domain knowledge to
make a professional and accurate analysis. Thus, there is still room for
improvement of LLMs reasoning, especially in leveraging the generation
capability of LLMs to simulate specific experts (i.e., multi-agents) to detect
the stance. In this paper, different from existing multi-agent works that
require detailed descriptions and use fixed experts, we propose a Dynamic
Experienced Expert Modeling (DEEM) method which can leverage the generated
experienced experts and let LLMs reason in a semi-parametric way, making the
experts more generalizable and reliable. Experimental results demonstrate that
DEEM consistently achieves the best results on three standard benchmarks,
outperforms methods with self-consistency reasoning, and reduces the bias of
LLMs.Comment: Accepted by LREC-COLING 2024, Oral presentatio
OpenChat: Advancing Open-source Language Models with Mixed-Quality Data
Nowadays, open-source large language models like LLaMA have emerged. Recent
developments have incorporated supervised fine-tuning (SFT) and reinforcement
learning fine-tuning (RLFT) to align these models with human goals. However,
SFT methods treat all training data with mixed quality equally, while RLFT
methods require high-quality pairwise or ranking-based preference data. In this
study, we present a novel framework, named OpenChat, to advance open-source
language models with mixed-quality data. Specifically, we consider the general
SFT training data, consisting of a small amount of expert data mixed with a
large proportion of sub-optimal data, without any preference labels. We propose
the C(onditioned)-RLFT, which regards different data sources as coarse-grained
reward labels and learns a class-conditioned policy to leverage complementary
data quality information. Interestingly, the optimal policy in C-RLFT can be
easily solved through single-stage, RL-free supervised learning, which is
lightweight and avoids costly human preference labeling. Through extensive
experiments on three standard benchmarks, our openchat-13b fine-tuned with
C-RLFT achieves the highest average performance among all 13b open-source
language models. Moreover, we use AGIEval to validate the model generalization
performance, in which only openchat-13b surpasses the base model. Finally, we
conduct a series of analyses to shed light on the effectiveness and robustness
of OpenChat. Our code, data, and models are publicly available at
https://github.com/imoneoi/openchat
Unsupervised Explanation Generation via Correct Instantiations
While large pre-trained language models (PLM) have shown their great skills
at solving discriminative tasks, a significant gap remains when compared with
humans for explanation-related tasks. Among them, explaining the reason why a
statement is wrong (e.g., against commonsense) is incredibly challenging. The
major difficulty is finding the conflict point, where the statement contradicts
our real world. This paper proposes Neon, a two-phrase, unsupervised
explanation generation framework. Neon first generates corrected instantiations
of the statement (phase I), then uses them to prompt large PLMs to find the
conflict point and complete the explanation (phase II). We conduct extensive
experiments on two standard explanation benchmarks, i.e., ComVE and e-SNLI.
According to both automatic and human evaluations, Neon outperforms baselines,
even for those with human-annotated instantiations. In addition to explaining a
negative prediction, we further demonstrate that Neon remains effective when
generalizing to different scenarios.Comment: Accepted to AAAI-2
DDT: Dual-branch Deformable Transformer for Image Denoising
Transformer is beneficial for image denoising tasks since it can model
long-range dependencies to overcome the limitations presented by inductive
convolutional biases. However, directly applying the transformer structure to
remove noise is challenging because its complexity grows quadratically with the
spatial resolution. In this paper, we propose an efficient Dual-branch
Deformable Transformer (DDT) denoising network which captures both local and
global interactions in parallel. We divide features with a fixed patch size and
a fixed number of patches in local and global branches, respectively. In
addition, we apply deformable attention operation in both branches, which helps
the network focus on more important regions and further reduces computational
complexity. We conduct extensive experiments on real-world and synthetic
denoising tasks, and the proposed DDT achieves state-of-the-art performance
with significantly fewer computational costs.Comment: The code is avaliable at: https://github.com/Merenguelkl/DD
Offline Reinforcement Learning with Imbalanced Datasets
The prevalent use of benchmarks in current offline reinforcement learning
(RL) research has led to a neglect of the imbalance of real-world dataset
distributions in the development of models. The real-world offline RL dataset
is often imbalanced over the state space due to the challenge of exploration or
safety considerations. In this paper, we specify properties of imbalanced
datasets in offline RL, where the state coverage follows a power law
distribution characterized by skewed policies. Theoretically and empirically,
we show that typically offline RL methods based on distributional constraints,
such as conservative Q-learning (CQL), are ineffective in extracting policies
under the imbalanced dataset. Inspired by natural intelligence, we propose a
novel offline RL method that utilizes the augmentation of CQL with a retrieval
process to recall past related experiences, effectively alleviating the
challenges posed by imbalanced datasets. We evaluate our method on several
tasks in the context of imbalanced datasets with varying levels of imbalance,
utilizing the variant of D4RL. Empirical results demonstrate the superiority of
our method over other baselines
- …
