40 research outputs found

    Unsupervised Feature Selection with Adaptive Structure Learning

    Full text link
    The problem of feature selection has raised considerable interests in the past decade. Traditional unsupervised methods select the features which can faithfully preserve the intrinsic structures of data, where the intrinsic structures are estimated using all the input features of data. However, the estimated intrinsic structures are unreliable/inaccurate when the redundant and noisy features are not removed. Therefore, we face a dilemma here: one need the true structures of data to identify the informative features, and one need the informative features to accurately estimate the true structures of data. To address this, we propose a unified learning framework which performs structure learning and feature selection simultaneously. The structures are adaptively learned from the results of feature selection, and the informative features are reselected to preserve the refined structures of data. By leveraging the interactions between these two essential tasks, we are able to capture accurate structures and select more informative features. Experimental results on many benchmark data sets demonstrate that the proposed method outperforms many state of the art unsupervised feature selection methods

    An experimental study of the intrinsic stability of random forest variable importance measures

    Get PDF
    BACKGROUND: The stability of Variable Importance Measures (VIMs) based on random forest has recently received increased attention. Despite the extensive attention on traditional stability of data perturbations or parameter variations, few studies include influences coming from the intrinsic randomness in generating VIMs, i.e. bagging, randomization and permutation. To address these influences, in this paper we introduce a new concept of intrinsic stability of VIMs, which is defined as the self-consistence among feature rankings in repeated runs of VIMs without data perturbations and parameter variations. Two widely used VIMs, i.e., Mean Decrease Accuracy (MDA) and Mean Decrease Gini (MDG) are comprehensively investigated. The motivation of this study is two-fold. First, we empirically verify the prevalence of intrinsic stability of VIMs over many real-world datasets to highlight that the instability of VIMs does not originate exclusively from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. Second, through Spearman and Pearson tests we comprehensively investigate how different factors influence the intrinsic stability. RESULTS: The experiments are carried out on 19 benchmark datasets with diverse characteristics, including 10 high-dimensional and small-sample gene expression datasets. Experimental results demonstrate the prevalence of intrinsic stability of VIMs. Spearman and Pearson tests on the correlations between intrinsic stability and different factors show that #feature (number of features) and #sample (size of sample) have a coupling effect on the intrinsic stability. The synthetic indictor, #feature/#sample, shows both negative monotonic correlation and negative linear correlation with the intrinsic stability, while OOB accuracy has monotonic correlations with intrinsic stability. This indicates that high-dimensional, small-sample and high complexity datasets may suffer more from intrinsic instability of VIMs. Furthermore, with respect to parameter settings of random forest, a large number of trees is preferred. No significant correlations can be seen between intrinsic stability and other factors. Finally, the magnitude of intrinsic stability is always smaller than that of traditional stability. CONCLUSION: First, the prevalence of intrinsic stability of VIMs demonstrates that the instability of VIMs not only comes from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. This finding gives a better understanding of VIM stability, and may help reduce the instability of VIMs. Second, by investigating the potential factors of intrinsic stability, users would be more aware of the risks and hence more careful when using VIMs, especially on high-dimensional, small-sample and high complexity datasets

    Extracting scientific trends by mining topics from Call for Papers

    Get PDF
    © 2019, Emerald Publishing Limited. Purpose: The purpose of this paper is to present a novel approach for mining scientific trends using topics from Call for Papers (CFP). The work contributes a valuable input for researchers, academics, funding institutes and research administration departments by sharing the trends to set directions of research path. Design/methodology/approach: The authors procure an innovative CFP data set to analyse scientific evolution and prestige of conferences that set scientific trends using scientific publications indexed in DBLP. Using the Field of Research code 804 from Australian Research Council, the authors identify 146 conferences (from 2006 to 2015) into different thematic areas by matching the terms extracted from publication titles with the Association for Computing Machinery Computing Classification System. Furthermore, the authors enrich the vocabulary of terms from the WordNet dictionary and Growbag data set. To measure the significance of terms, the authors adopt the following weighting schemas: probabilistic, gram, relative, accumulative and hierarchal. Findings: The results indicate the rise of “big data analytics” from CFP topics in the last few years. Whereas the topics related to “privacy and security” show an exponential increase, the topics related to “semantic web” show a downfall in recent years. While analysing publication output in DBLP that matches CFP indexed in ERA Core A* to C rank conference, the authors identified that A* and A tier conferences not merely set publication trends, since B or C tier conferences target similar CFP. Originality/value: Overall, the analyses presented in this research are prolific for the scientific community and research administrators to study research trends and better data management of digital libraries pertaining to the scientific literature

    Global Perspectives on Task Shifting and Task Sharing in Neurosurgery.

    Get PDF
    BACKGROUND: Neurosurgical task shifting and task sharing (TS/S), delegating clinical care to non-neurosurgeons, is ongoing in many hospital systems in which neurosurgeons are scarce. Although TS/S can increase access to treatment, it remains highly controversial. This survey investigated perceptions of neurosurgical TS/S to elucidate whether it is a permissible temporary solution to the global workforce deficit. METHODS: The survey was distributed to a convenience sample of individuals providing neurosurgical care. A digital survey link was distributed through electronic mailing lists of continental neurosurgical societies and various collectives, conference announcements, and social media platforms (July 2018-January 2019). Data were analyzed by descriptive statistics and univariate regression of Likert Scale scores. RESULTS: Survey respondents represented 105 of 194 World Health Organization member countries (54.1%; 391 respondents, 162 from high-income countries and 229 from low- and middle-income countries [LMICs]). The most agreed on statement was that task sharing is preferred to task shifting. There was broad consensus that both task shifting and task sharing should require competency-based evaluation, standardized training endorsed by governing organizations, and maintenance of certification. When perspectives were stratified by income class, LMICs were significantly more likely to agree that task shifting is professionally disruptive to traditional training, task sharing should be a priority where human resources are scarce, and to call for additional TS/S regulation, such as certification and formal consultation with a neurosurgeon (in person or electronic/telemedicine). CONCLUSIONS: Both LMIC and high-income countries agreed that task sharing should be prioritized over task shifting and that additional recommendations and regulations could enhance care. These data invite future discussions on policy and training programs

    Radiology Community Attitude in Saudi Arabia about the Applications of Artificial Intelligence in Radiology

    No full text
    Artificial intelligence (AI) is a broad, umbrella term that encompasses the theory and development of computer systems able to perform tasks normally requiring human intelligence. The aim of this study is to assess the radiology community’s attitude in Saudi Arabia toward the applications of AI. Methods: Data for this study were collected using electronic questionnaires in 2019 and 2020. The study included a total of 714 participants. Data analysis was performed using SPSS Statistics (version 25). Results: The majority of the participants (61.2%) had read or heard about the role of AI in radiology. We also found that radiologists had statistically different responses and tended to read more about AI compared to all other specialists. In addition, 82% of the participants thought that AI must be included in the curriculum of medical and allied health colleges, and 86% of the participants agreed that AI would be essential in the future. Even though human–machine interaction was considered to be one of the most important skills in the future, 89% of the participants thought that it would never replace radiologists. Conclusion: Because AI plays a vital role in radiology, it is important to ensure that radiologists and radiographers have at least a minimum understanding of the technology. Our finding shows an acceptable level of knowledge regarding AI technology and that AI applications should be included in the curriculum of the medical and health sciences colleges.</jats:p

    Radiology Community Attitude in Saudi Arabia about the Applications of Artificial Intelligence in Radiology

    No full text
    Artificial intelligence (AI) is a broad, umbrella term that encompasses the theory and development of computer systems able to perform tasks normally requiring human intelligence. The aim of this study is to assess the radiology community’s attitude in Saudi Arabia toward the applications of AI. Methods: Data for this study were collected using electronic questionnaires in 2019 and 2020. The study included a total of 714 participants. Data analysis was performed using SPSS Statistics (version 25). Results: The majority of the participants (61.2%) had read or heard about the role of AI in radiology. We also found that radiologists had statistically different responses and tended to read more about AI compared to all other specialists. In addition, 82% of the participants thought that AI must be included in the curriculum of medical and allied health colleges, and 86% of the participants agreed that AI would be essential in the future. Even though human–machine interaction was considered to be one of the most important skills in the future, 89% of the participants thought that it would never replace radiologists. Conclusion: Because AI plays a vital role in radiology, it is important to ensure that radiologists and radiographers have at least a minimum understanding of the technology. Our finding shows an acceptable level of knowledge regarding AI technology and that AI applications should be included in the curriculum of the medical and health sciences colleges

    Joint Dictionary Learning for Unsupervised Feature Selection

    No full text

    A Novel Computational Model for Detecting the Severity of Inflammation in Confirmed COVID-19 Patients Using Chest X-ray Images

    No full text
    Since late 2019, Coronavirus Disease 2019 (COVID-19) has spread all over the world. The disease is highly contagious, and it may lead to acute respiratory distress (ARD). Medical imaging can play an important role in classifying, detecting, and measuring the severity of the virus. This study aims to provide a novel auto-detection tool that can detect abnormal changes in conventional X-ray images for confirmed COVID-19 cases. X-ray images from patients diagnosed with COVID-19 were converted into 19 different colored layers. Each layer represented objects with similar contrast that could be defined as a specific color. The objects with similar contrasts were formed in a single layer. All the objects from all the layers were extracted as a single-color image. Based on the differentiation of colors, the prototype model was able to recognize a wide spectrum of abnormal changes in the image texture. This was true even if there was minimal variation of the contrast values of the detected uncleared abnormalities. The results indicate that the proposed novel method can detect and determine the degree of lung infection from COVID-19 with an accuracy of 91%, compared to the opinions of three experienced radiologists. The method can also efficiently determine the sites of infection and the severity of the disease by classifying the X-rays into five levels of severity. Thus, the proposed COVID-19 autodetection method can identify locations and indicate the degree of severity of the disease by comparing affected tissue with healthy tissue, and it can predict where the disease may spread.</jats:p
    corecore