787 research outputs found

    DualSMC: Tunneling Differentiable Filtering and Planning under Continuous POMDPs

    Full text link
    A major difficulty of solving continuous POMDPs is to infer the multi-modal distribution of the unobserved true states and to make the planning algorithm dependent on the perceived uncertainty. We cast POMDP filtering and planning problems as two closely related Sequential Monte Carlo (SMC) processes, one over the real states and the other over the future optimal trajectories, and combine the merits of these two parts in a new model named the DualSMC network. In particular, we first introduce an adversarial particle filter that leverages the adversarial relationship between its internal components. Based on the filtering results, we then propose a planning algorithm that extends the previous SMC planning approach [Piche et al., 2018] to continuous POMDPs with an uncertainty-dependent policy. Crucially, not only can DualSMC handle complex observations such as image input but also it remains highly interpretable. It is shown to be effective in three continuous POMDP domains: the floor positioning domain, the 3D light-dark navigation domain, and a modified Reacher domain.Comment: IJCAI 202

    Lobectomy with Bronchoplasty and Reconstruction of Pulmonary Artery by Minitrauma-technique for Lung Cancer

    Get PDF
    Background and objective To research the effect and practicalbility of lobectomy with bronchoplasty and reconstruction of pulmonary artery by minitrauma-technique for lung cancer. Methods We retrospectibely reviewed our experience on 61 cases being lobectomy with bronchoplasty and bronchoplasty with or without video assisted thoracic small incision surgery for lung cancer from July 2005 to June 2009 from Shandong Provincal Hospital and 46 cases simultaneously by routine posterolateral incision. All patients whose bronchus and/or pulmonary artery were involved underwent the operation and experienced the bronchial sleeve/wedge resection or reconstruction of the pulmonary artery. Results All patients were done operation successfully and there were no operative mortality and no occurrence of anastomosis stenosis as well as fistula. The small incisions’ length was from 8 cm-15 cm while the routine posterolateral incision’s length was 25 cm-35 cm. The patients done the operation of small incision had less postoperative shoulder joint dysfunction and had better quality of life compaired to the patients done the routine posterolateral incision. Conclusion Lobectomy with bronchoplasty and reconstruction of pulmonary artery by minitrauma-technique for lung cancer could finished the same work with the traditional thoracic lateral incision and had less trauma, less pain, less recovery time

    MoDS: Model-oriented Data Selection for Instruction Tuning

    Full text link
    Instruction tuning has become the de facto method to equip large language models (LLMs) with the ability of following user instructions. Usually, hundreds of thousands or millions of instruction-following pairs are employed to fine-tune the foundation LLMs. Recently, some studies show that a small number of high-quality instruction data is enough. However, how to select appropriate instruction data for a given LLM is still an open problem. To address this problem, in this paper we present a model-oriented data selection (MoDS) approach, which selects instruction data based on a new criteria considering three aspects: quality, coverage and necessity. First, our approach utilizes a quality evaluation model to filter out the high-quality subset from the original instruction dataset, and then designs an algorithm to further select from the high-quality subset a seed instruction dataset with good coverage. The seed dataset is applied to fine-tune the foundation LLM to obtain an initial instruction-following LLM. Finally, we develop a necessity evaluation model to find out the instruction data which are performed badly in the initial instruction-following LLM and consider them necessary instructions to further improve the LLMs. In this way, we can get a small high-quality, broad-coverage and high-necessity subset from the original instruction datasets. Experimental results show that, the model fine-tuned with 4,000 instruction pairs selected by our approach could perform better than the model fine-tuned with the full original dataset which includes 214k instruction data

    Unlink to Unlearn: Simplifying Edge Unlearning in GNNs

    Full text link
    As concerns over data privacy intensify, unlearning in Graph Neural Networks (GNNs) has emerged as a prominent research frontier in academia. This concept is pivotal in enforcing the \textit{right to be forgotten}, which entails the selective removal of specific data from trained GNNs upon user request. Our research focuses on edge unlearning, a process of particular relevance to real-world applications. Current state-of-the-art approaches like GNNDelete can eliminate the influence of specific edges yet suffer from \textit{over-forgetting}, which means the unlearning process inadvertently removes excessive information beyond needed, leading to a significant performance decline for remaining edges. Our analysis identifies the loss functions of GNNDelete as the primary source of over-forgetting and also suggests that loss functions may be redundant for effective edge unlearning. Building on these insights, we simplify GNNDelete to develop \textbf{Unlink to Unlearn} (UtU), a novel method that facilitates unlearning exclusively through unlinking the forget edges from graph structure. Our extensive experiments demonstrate that UtU delivers privacy protection on par with that of a retrained model while preserving high accuracy in downstream tasks, by upholding over 97.3\% of the retrained model's privacy protection capabilities and 99.8\% of its link prediction accuracy. Meanwhile, UtU requires only constant computational demands, underscoring its advantage as a highly lightweight and practical edge unlearning solution.Comment: Accepted by WWW 2024 as a Short Research Pape

    Lost in the Source Language: How Large Language Models Evaluate the Quality of Machine Translation

    Full text link
    This study investigates how Large Language Models (LLMs) leverage source and reference data in machine translation evaluation task, aiming to better understand the mechanisms behind their remarkable performance in this task. We design the controlled experiments across various input modes and model types, and employ both coarse-grained and fine-grained prompts to discern the utility of source versus reference information. We find that reference information significantly enhances the evaluation accuracy, while surprisingly, source information sometimes is counterproductive, indicating LLMs' inability to fully leverage the cross-lingual capability when evaluating translations. Further analysis of the fine-grained evaluation and fine-tuning experiments show similar results. These findings also suggest a potential research direction for LLMs that fully exploits the cross-lingual capability of LLMs to achieve better performance in machine translation evaluation tasks.Comment: Accepted by ACL2024 Finding

    An intelligent video fire detection approach based on object detection technology

    Get PDF
    PresentationFire that is one of the most serious accidents in chemical factories, may lead to considerable product losses, equipment damages and casualties. With the rapid development of computer vision technology, intelligent fire detection has been proposed and applied in various scenarios. This paper presents a new intelligent video fire detection approach based on object detection technology using convolutional neural networks (CNN). First, a CNN model is trained for the fire detection task which is framed as a regression problem to predict bounding boxes and associated probabilities. In the application phase, videos from surveillance cameras are detected frame by frame. Once fire appears in the current frame, the model will output the coordinates of the fire region. Simultaneously, the frame where the fire region is localized will be immediately sent to safety supervisors as a fire alarm. This will help detect fire at the early stage, prevent fire spreading and improve the emergency response

    Constraining interacting dark energy models with the halo concentration - mass relation

    Full text link
    The interacting dark energy (IDE) model is a promising alternative cosmological model which has the potential to solve the fine-tuning and coincidence problems by considering the interaction between dark matter and dark energy. Previous studies have shown that the energy exchange between the dark sectors in this model can significantly affect the dark matter halo properties. In this study, utilising a large set of cosmological NN-body simulations, we analyse the redshift evolution of the halo concentration - mass (cc - MM) relation in the IDE model, and show that the cc - MM relation is a sensitive proxy of the interaction strength parameter ξ2\xi_2, especially at lower redshifts. Furthermore, we construct parametrized formulae to quantify the dependence of the cc - MM relation on ξ2\xi_2 at redshifts ranging from z=0z=0 to 0.60.6. Our parametrized formulae provide a useful tool in constraining ξ2\xi_2 with the observational cc - MM relation. As a first attempt, we use the data from X-ray, gravitational lensing, and galaxy rotational curve observations and obtain a tight constraint on ξ2\xi_2, i.e. ξ2=0.071±0.034\xi_2 = 0.071 \pm 0.034. Our work demonstrates that the halo cc - MM relation, which reflects the halo assembly history, is a powerful probe to constrain the IDE model.Comment: 9 pages, 5 figures, 5 table
    corecore