367 research outputs found

    Piperidinium bis­(2-oxidobenzoato-κ2 O 1,O 2)borate

    Get PDF
    The asymmetric unit of the title compound, C5H12N+·C14H8BO6 − or [C5H12N][BO4(C7H4O)2], contains two piperidinium cations and two bis­(salicylato)borate anions. The coordination geometries around the B atoms are distorted tetra­hedral. In the two mol­ecules, the aromatic rings are oriented at dihedral angles of 76.27 (3) and 83.86 (3)°. The rings containing B atoms have twist-boat conformations, while the two cations adopt chair conformations. In the crystal, the component species are linked by N—H⋯O hydrogen bonds. In the crystal structure, intra- and inter­molecular N—H⋯O hydrogen bonds link the mol­ecules

    Object-based attention mechanism for color calibration of UAV remote sensing images in precision agriculture.

    Get PDF
    Color calibration is a critical step for unmanned aerial vehicle (UAV) remote sensing, especially in precision agriculture, which relies mainly on correlating color changes to specific quality attributes, e.g. plant health, disease, and pest stresses. In UAV remote sensing, the exemplar-based color transfer is popularly used for color calibration, where the automatic search for the semantic correspondences is the key to ensuring the color transfer accuracy. However, the existing attention mechanisms encounter difficulties in building the precise semantic correspondences between the reference image and the target one, in which the normalized cross correlation is often computed for feature reassembling. As a result, the color transfer accuracy is inevitably decreased by the disturbance from the semantically unrelated pixels, leading to semantic mismatch due to the absence of semantic correspondences. In this article, we proposed an unsupervised object-based attention mechanism (OBAM) to suppress the disturbance of the semantically unrelated pixels, along with a further introduced weight-adjusted Adaptive Instance Normalization (AdaIN) (WAA) method to tackle the challenges caused by the absence of semantic correspondences. By embedding the proposed modules into a photorealistic style transfer method with progressive stylization, the color transfer accuracy can be improved while better preserving the structural details. We evaluated our approach on the UAV data of different crop types including rice, beans, and cotton. Extensive experiments demonstrate that our proposed method outperforms several state-of-the-art methods. As our approach requires no annotated labels, it can be easily embedded into the off-the-shelf color transfer approaches. Relevant codes and configurations will be available at https://github.com/huanghsheng/object-based-attention-mechanis

    The emerging role of DOT1L in cell proliferation and differentiation: Friend or foe

    Get PDF
    Cell proliferation and differentiation are the basic physiological activities of cells. Mistakes in these processes may affect cell survival, or cause cell cycle dysregulation, such as tumorigenesis, birth defects and degenerative diseases. In recent years, it has been found that histone methyltransferase DOT1L is the only H3 lysine 79 methyltransferase, which plays an important role in the process of cell fate determination through monomethylation, dimethylation and trimethylation of H3K79. DOT1L has a pro-proliferative effect in leukemia cells; however, loss of heart-specific DOT1L leads to increased proliferation of cardiac tissue. Additionally, DOT1L has carcinogenic or tumor suppressive effects in different neoplasms. At present, some DOT1L inhibitors for the treatment of MLL-driven leukemia have achieved promising results in clinical trials, but completely blocking DOT1L will also bring some side effects. Thus, this uncertainty suggests that DOT1L has a unique function in cell physiology. In this review, we summarize the primary findings of DOT1L in regulating cell proliferation and differentiation. Correlations between DOT1L and cell fate specification might suggest DOT1L as a therapeutic target for diseases

    Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models

    Full text link
    We introduce Generalized Instruction Tuning (called GLAN), a general and scalable method for instruction tuning of Large Language Models (LLMs). Unlike prior work that relies on seed examples or existing datasets to construct instruction tuning data, GLAN exclusively utilizes a pre-curated taxonomy of human knowledge and capabilities as input and generates large-scale synthetic instruction data across all disciplines. Specifically, inspired by the systematic structure in human education system, we build the taxonomy by decomposing human knowledge and capabilities to various fields, sub-fields and ultimately, distinct disciplines semi-automatically, facilitated by LLMs. Subsequently, we generate a comprehensive list of subjects for every discipline and proceed to design a syllabus tailored to each subject, again utilizing LLMs. With the fine-grained key concepts detailed in every class session of the syllabus, we are able to generate diverse instructions with a broad coverage across the entire spectrum of human knowledge and skills. Extensive experiments on large language models (e.g., Mistral) demonstrate that GLAN excels in multiple dimensions from mathematical reasoning, coding, academic exams, logical reasoning to general instruction following without using task-specific training data of these tasks. In addition, GLAN allows for easy customization and new fields or skills can be added by simply incorporating a new node into our taxonomy.Comment: Work in progres

    Citrus fruit detection based on an improved YOLOv5 under natural orchard conditions.

    Get PDF
    Accurate detection of citrus can be easily affected by adjacent branches and overlapped fruits in natural orchard conditions, where some specific information of citrus might be lost due to the resultant complex occlusion. Traditional deep learning models might result in lower detection accuracy and detection speed when facing occluded targets. To solve this problem, an improved deep learning algorithm based on YOLOv5, named IYOLOv5, was proposed for accurate detection of citrus fruits. An innovative Res-CSPDarknet network was firstly employed to both enhance feature extraction performance and minimize feature loss within the backbone network, which aims to reduce the miss detection rate. Subsequently, the BiFPN module was adopted as the new neck net to enhance the function for extracting deep semantic features. A coordinate attention mechanism module was then introduced into the network's detection layer. The performance of the proposed model was evaluated on a home-made citrus dataset containing 2000 optical images. The results show that the proposed IYOLOv5 achieved the highest mean average precision (93.5%) and F1-score (95.6%), compared to the traditional deep learning models including Faster R-CNN, CenterNet, YOLOv3, YOLOv5, and YOLOv7. In particular, the proposed IYOLOv5 obtained a decrease of missed detection rate (at least 13.1%) on the specific task of detecting heavily occluded citrus, compared to other models. Therefore, the proposed method could be potentially used as part of the vision system of a picking robot to identify the citrus fruits accurately

    Detection and localization of citrus fruit based on improved You Only Look Once v5s and binocular vision in the orchard

    Get PDF
    Intelligent detection and localization of mature citrus fruits is a critical challenge in developing an automatic harvesting robot. Variable illumination conditions and different occlusion states are some of the essential issues that must be addressed for the accurate detection and localization of citrus in the orchard environment. In this paper, a novel method for the detection and localization of mature citrus using improved You Only Look Once (YOLO) v5s with binocular vision is proposed. First, a new loss function (polarity binary cross-entropy with logit loss) for YOLO v5s is designed to calculate the loss value of class probability and objectness score, so that a large penalty for false and missing detection is applied during the training process. Second, to recover the missing depth information caused by randomly overlapping background participants, Cr-Cb chromatic mapping, the Otsu thresholding algorithm, and morphological processing are successively used to extract the complete shape of the citrus, and the kriging method is applied to obtain the best linear unbiased estimator for the missing depth value. Finally, the citrus spatial position and posture information are obtained according to the camera imaging model and the geometric features of the citrus. The experimental results show that the recall rates of citrus detection under non-uniform illumination conditions, weak illumination, and well illumination are 99.55%, 98.47%, and 98.48%, respectively, approximately 2–9% higher than those of the original YOLO v5s network. The average error of the distance between the citrus fruit and the camera is 3.98 mm, and the average errors of the citrus diameters in the 3D direction are less than 2.75 mm. The average detection time per frame is 78.96 ms. The results indicate that our method can detect and localize citrus fruits in the complex environment of orchards with high accuracy and speed. Our dataset and codes are available at https://github.com/AshesBen/citrus-detection-localization

    Multimodal fusion model for diagnosing mild cognitive impairment in unilateral middle cerebral artery steno-occlusive disease

    Get PDF
    ObjectivesTo propose a multimodal functional brain network (FBN) and structural brain network (SBN) topological feature fusion technique based on resting-state functional magnetic resonance imaging (rs-fMRI), diffusion tensor imaging (DTI), 3D-T1-weighted imaging (3D-T1WI), and demographic characteristics to diagnose mild cognitive impairment (MCI) in patients with unilateral middle cerebral artery (MCA) steno-occlusive disease.MethodsThe performances of different algorithms on the MCI dataset were evaluated using 5-fold cross-validation. The diagnostic results of the multimodal performance were evaluated using t-distributed stochastic neighbor embedding (t-SNE) analysis. The four-modal analysis method proposed in this study was applied to identify brain regions and connections associated with MCI, thus confirming its validity.ResultsBased on the fusion of the topological features of the multimodal FBN and SBN, the accuracy for the diagnosis of MCI in patients with unilateral MCA steno-occlusive disease reached 90.00%. The accuracy, recall, sensitivity, and F1-score were higher than those of the other methods, as was the diagnostic efficacy (AUC = 0.9149).ConclusionThe multimodal FBN and SBN topological feature fusion technique, which incorporates rs-fMRI, DTI, 3D-T1WI, and demographic characteristics, obtains the most discriminative features of MCI in patients with unilateral MCA steno-occlusive disease and can effectively identify disease-related brain areas and connections. Efficient automated diagnosis facilitates the early and accurate detection of MCI and timely intervention and treatment to delay or prevent disease progression

    Tool Learning with Foundation Models

    Full text link
    Humans possess an extraordinary ability to create and utilize tools, allowing them to overcome physical limitations and explore new frontiers. With the advent of foundation models, AI systems have the potential to be equally adept in tool use as humans. This paradigm, i.e., tool learning with foundation models, combines the strengths of specialized tools and foundation models to achieve enhanced accuracy, efficiency, and automation in problem-solving. Despite its immense potential, there is still a lack of a comprehensive understanding of key challenges, opportunities, and future endeavors in this field. To this end, we present a systematic investigation of tool learning in this paper. We first introduce the background of tool learning, including its cognitive origins, the paradigm shift of foundation models, and the complementary roles of tools and models. Then we recapitulate existing tool learning research into tool-augmented and tool-oriented learning. We formulate a general tool learning framework: starting from understanding the user instruction, models should learn to decompose a complex task into several subtasks, dynamically adjust their plan through reasoning, and effectively conquer each sub-task by selecting appropriate tools. We also discuss how to train models for improved tool-use capabilities and facilitate the generalization in tool learning. Considering the lack of a systematic tool learning evaluation in prior works, we experiment with 18 representative tools and show the potential of current foundation models in skillfully utilizing tools. Finally, we discuss several open problems that require further investigation for tool learning. In general, we hope this paper could inspire future research in integrating tools with foundation models
    corecore