318 research outputs found
SEE: Syntax-aware Entity Embedding for Neural Relation Extraction
Distant supervised relation extraction is an efficient approach to scale
relation extraction to very large corpora, and has been widely used to find
novel relational facts from plain text. Recent studies on neural relation
extraction have shown great progress on this task via modeling the sentences in
low-dimensional spaces, but seldom considered syntax information to model the
entities. In this paper, we propose to learn syntax-aware entity embedding for
neural relation extraction. First, we encode the context of entities on a
dependency tree as sentence-level entity embedding based on tree-GRU. Then, we
utilize both intra-sentence and inter-sentence attentions to obtain sentence
set-level entity embedding over all sentences containing the focus entity pair.
Finally, we combine both sentence embedding and entity embedding for relation
classification. We conduct experiments on a widely used real-world dataset and
the experimental results show that our model can make full use of all
informative instances and achieve state-of-the-art performance of relation
extraction.Comment: 8 pages, AAAI-201
Quantum dot-based thermometry uncovers decreased myosin efficiency in an experimental intensive care unit model
Critical illness myopathy (CIM) detrimentally affects muscle function in ICU patients, with a dramatic loss of muscle mass and function where the loss in specific force exceeds the loss in muscle mass (maximum force normalized to muscle cross-sectional area). The preferential loss of the molecular motor protein myosin, representing the hallmark of CIM, exhibiting a significant negative impact on the specific force generation by the muscle. Interestingly however, the preferential myosin loss is a relatively late event, and a specific loss in force generation capacity, is observed prior to the myosin loss. In the current study, employing an optimized cadmium telluride quantum dots (QD) mediated-thermometry approach to assess the efficiency of the myosin, we were able to determine the loss in specific force generated by the muscle, prior to the preferential loss of myosin. Reduction in QD fluorescent intensity correlates with greater heat loss, reflecting inefficient myosin function (less mechanical work performed and more heat loss on ATP hydrolysis by myosin). A significant decrease in myosin efficiency was observed in rats subjected to the ICU condition (immobilization and mechanical ventilation) for 5 days using an established experimental ICU model not limited by early mortality. Thus, qualitative myosin changes preceding quantitative myosin loss offer a mechanism underlying the early loss in specific force generation capacity associated with CIM and opens a venue for future CIM intervention strategies
On the Robustness of Aspect-based Sentiment Analysis: Rethinking Model, Data, and Training
Aspect-based sentiment analysis (ABSA) aims at automatically inferring the
specific sentiment polarities toward certain aspects of products or services
behind the social media texts or reviews, which has been a fundamental
application to the real-world society. Since the early 2010s, ABSA has achieved
extraordinarily high accuracy with various deep neural models. However,
existing ABSA models with strong in-house performances may fail to generalize
to some challenging cases where the contexts are variable, i.e., low robustness
to real-world environments. In this study, we propose to enhance the ABSA
robustness by systematically rethinking the bottlenecks from all possible
angles, including model, data, and training. First, we strengthen the current
best-robust syntax-aware models by further incorporating the rich external
syntactic dependencies and the labels with aspect simultaneously with a
universal-syntax graph convolutional network. In the corpus perspective, we
propose to automatically induce high-quality synthetic training data with
various types, allowing models to learn sufficient inductive bias for better
robustness. Last, we based on the rich pseudo data perform adversarial training
to enhance the resistance to the context perturbation and meanwhile employ
contrastive learning to reinforce the representations of instances with
contrastive sentiments. Extensive robustness evaluations are conducted. The
results demonstrate that our enhanced syntax-aware model achieves better
robustness performances than all the state-of-the-art baselines. By
additionally incorporating our synthetic corpus, the robust testing results are
pushed with around 10% accuracy, which are then further improved by installing
the advanced training strategies. In-depth analyses are presented for revealing
the factors influencing the ABSA robustness.Comment: Accepted in ACM Transactions on Information System
LasUIE: Unifying Information Extraction with Latent Adaptive Structure-aware Generative Language Model
Universally modeling all typical information extraction tasks (UIE) with one
generative language model (GLM) has revealed great potential by the latest
study, where various IE predictions are unified into a linearized hierarchical
expression under a GLM. Syntactic structure information, a type of effective
feature which has been extensively utilized in IE community, should also be
beneficial to UIE. In this work, we propose a novel structure-aware GLM, fully
unleashing the power of syntactic knowledge for UIE. A heterogeneous structure
inductor is explored to unsupervisedly induce rich heterogeneous structural
representations by post-training an existing GLM. In particular, a structural
broadcaster is devised to compact various latent trees into explicit high-order
forests, helping to guide a better generation during decoding. We finally
introduce a task-oriented structure fine-tuning mechanism, further adjusting
the learned structures to most coincide with the end-task's need. Over 12 IE
benchmarks across 7 tasks our system shows significant improvements over the
baseline UIE system. Further in-depth analyses show that our GLM learns rich
task-adaptive structural bias that greatly resolves the UIE crux, the
long-range dependence issue and boundary identifying. Source codes are open at
https://github.com/ChocoWu/LasUIE.Comment: NeurIPS2022 conference pape
Recognizing Everything from All Modalities at Once: Grounded Multimodal Universal Information Extraction
In the field of information extraction (IE), tasks across a wide range of
modalities and their combinations have been traditionally studied in isolation,
leaving a gap in deeply recognizing and analyzing cross-modal information. To
address this, this work for the first time introduces the concept of grounded
Multimodal Universal Information Extraction (MUIE), providing a unified task
framework to analyze any IE tasks over various modalities, along with their
fine-grained groundings. To tackle MUIE, we tailor a multimodal large language
model (MLLM), Reamo, capable of extracting and grounding information from all
modalities, i.e., recognizing everything from all modalities at once. Reamo is
updated via varied tuning strategies, equipping it with powerful capabilities
for information recognition and fine-grained multimodal grounding. To address
the absence of a suitable benchmark for grounded MUIE, we curate a
high-quality, diverse, and challenging test set, which encompasses IE tasks
across 9 common modality combinations with the corresponding multimodal
groundings. The extensive comparison of Reamo with existing MLLMs integrated
into pipeline approaches demonstrates its advantages across all evaluation
dimensions, establishing a strong benchmark for the follow-up research. Our
resources are publicly released at https://haofei.vip/MUIE
Retrieval-style In-Context Learning for Few-shot Hierarchical Text Classification
Hierarchical text classification (HTC) is an important task with broad
applications, while few-shot HTC has gained increasing interest recently. While
in-context learning (ICL) with large language models (LLMs) has achieved
significant success in few-shot learning, it is not as effective for HTC
because of the expansive hierarchical label sets and extremely-ambiguous
labels. In this work, we introduce the first ICL-based framework with LLM for
few-shot HTC. We exploit a retrieval database to identify relevant
demonstrations, and an iterative policy to manage multi-layer hierarchical
labels. Particularly, we equip the retrieval database with HTC label-aware
representations for the input texts, which is achieved by continual training on
a pretrained language model with masked language modeling (MLM), layer-wise
classification (CLS, specifically for HTC), and a novel divergent contrastive
learning (DCL, mainly for adjacent semantically-similar labels) objective.
Experimental results on three benchmark datasets demonstrate superior
performance of our method, and we can achieve state-of-the-art results in
few-shot HTC.Comment: 17 page
Syntax-aware Neural Semantic Role Labeling
Semantic role labeling (SRL), also known as shallow semantic parsing, is an
important yet challenging task in NLP. Motivated by the close correlation
between syntactic and semantic structures, traditional discrete-feature-based
SRL approaches make heavy use of syntactic features. In contrast,
deep-neural-network-based approaches usually encode the input sentence as a
word sequence without considering the syntactic structures. In this work, we
investigate several previous approaches for encoding syntactic trees, and make
a thorough study on whether extra syntax-aware representations are beneficial
for neural SRL models. Experiments on the benchmark CoNLL-2005 dataset show
that syntax-aware SRL approaches can effectively improve performance over a
strong baseline with external word representations from ELMo. With the extra
syntax-aware representations, our approaches achieve new state-of-the-art 85.6
F1 (single model) and 86.6 F1 (ensemble) on the test data, outperforming the
corresponding strong baselines with ELMo by 0.8 and 1.0, respectively. Detailed
error analysis are conducted to gain more insights on the investigated
approaches.Comment: AAAI 201
- …
