754 research outputs found
Impact of managed clinical networks on neonatal care in England : a population-based study
Objective: To assess the impact of reorganisation of neonatal specialist care services in England after a UK Department of Health report in 2003.
Design: A population-wide observational comparison of outcomes over two epochs, before and after the establishment of managed clinical neonatal networks.
Setting: Epoch one: 294 maternity and neonatal units in England, Wales, and Northern Ireland, 1 September 1998 to 31 August 2000, as reported by the Confidential Enquiry into Stillbirths and Sudden Deaths in Infancy Project 27/28. Epoch two: 146 neonatal units in England contributing data to the National Neonatal Research Database at the Neonatal Data Analysis Unit, 1 January 2009 to 31 December 2010.
Participants: Babies born at a gestational age of 27+0-28+6 (weeks+days): 3522 live births in epoch one; 2919 babies admitted to a neonatal unit within 28 days of birth in epoch two.
Intervention: The national reorganisation of neonatal services into managed clinical networks.
Main outcome measures: The proportion of babies born at hospitals providing the highest volume of neonatal specialist care (≥2000 neonatal intensive care days annually), having an acute transfer (within the first 24 hours after birth) and/or a late transfer (between 24 hours and 28 days after birth) to another hospital, assessed by change in distribution of transfer category (“none,” “acute,” “late”), and babies from multiple births separated by transfer. For acute transfers in epoch two, the level of specialist neonatal care provided at the destination hospital (British Association of Perinatal Medicine criteria).
Results: After reorganisation, there were increases in the proportions of babies born at 27-28 weeks’ gestation in hospitals providing the highest volume of neonatal specialist care (18% (631/3495) v 49% (1325/2724); odds ratio 4.30, 95% confidence interval 3.83 to 4.82; P<0.001) and in acute and late postnatal transfers (7% (235) v 12% (360) and 18% (579) v 22% (640), respectively; P<0.001). There was no significant change in the proportion of babies from multiple births separated by transfer (33% (39) v 29% (38); 0.86, 0.50 to 1.46; P=0.57). In epoch two, 32% of acute transfers were to a neonatal unit providing either an equivalent (n=87) or lower (n=26) level of specialist care.
Conclusions: There is evidence of some improvement in the delivery of neonatal specialist care after reorganisation. The increase in acute transfers in epoch two, in conjunction with the high proportion transferred to a neonatal unit providing an equivalent or lower level of specialist care, and the continued separation of babies from multiple births, are indicative of poor coordination between maternity and neonatal services to facilitate in utero transfer before delivery, and continuing inadequacies in capacity of intensive care cots. Historical data representing epoch one are available only in aggregate form, preventing examination of temporal trends or confounding factors. This limits the extent to which differences between epochs can be attributed to reorganisation and highlights the importance of routine, prospective data collection for evaluation of future health service reorganisations
Automated Discrimination of Pathological Regions in Tissue Images: Unsupervised Clustering vs Supervised SVM Classification
Recognizing and isolating cancerous cells from non pathological tissue areas (e.g. connective stroma) is crucial for fast and objective immunohistochemical analysis of tissue images. This operation allows the further application of fully-automated techniques for quantitative evaluation of protein activity, since it avoids the necessity of a preventive manual selection of the representative pathological areas in the image, as well as of taking pictures only in the pure-cancerous portions of the tissue. In this paper we present a fully-automated method based on unsupervised clustering that performs tissue segmentations highly comparable with those provided by a skilled operator, achieving on average an accuracy of 90%. Experimental results on a heterogeneous dataset of immunohistochemical lung cancer tissue images demonstrate that our proposed unsupervised approach overcomes the accuracy of a theoretically superior supervised method such as Support Vector Machine (SVM) by 8%
Generative Invertible Networks (GIN): Pathophysiology-Interpretable Feature Mapping and Virtual Patient Generation
Machine learning methods play increasingly important roles in pre-procedural
planning for complex surgeries and interventions. Very often, however,
researchers find the historical records of emerging surgical techniques, such
as the transcatheter aortic valve replacement (TAVR), are highly scarce in
quantity. In this paper, we address this challenge by proposing novel
generative invertible networks (GIN) to select features and generate
high-quality virtual patients that may potentially serve as an additional data
source for machine learning. Combining a convolutional neural network (CNN) and
generative adversarial networks (GAN), GIN discovers the pathophysiologic
meaning of the feature space. Moreover, a test of predicting the surgical
outcome directly using the selected features results in a high accuracy of
81.55%, which suggests little pathophysiologic information has been lost while
conducting the feature selection. This demonstrates GIN can generate virtual
patients not only visually authentic but also pathophysiologically
interpretable
Automated segmentation of tissue images for computerized IHC analysis
This paper presents two automated methods for the segmentation ofimmunohistochemical tissue images that overcome the limitations of themanual approach aswell as of the existing computerized techniques. The first independent method, based on unsupervised color clustering, recognizes automatically the target cancerous areas in the specimen and disregards the stroma; the second method, based on colors separation and morphological processing, exploits automated segmentation of the nuclear membranes of the cancerous cells. Extensive experimental results on real tissue images demonstrate the accuracy of our techniques compared to manual segmentations; additional experiments show that our techniques are more effective in immunohistochemical images than popular approaches based on supervised learning or active contours. The proposed procedure can be exploited for any applications that require tissues and cells exploration and to perform reliable and standardized measures of the activity of specific proteins involved in multi-factorial genetic pathologie
MidA is a putative methyltransferase that is required for mitochondrial complex I function
10 páginas, 6 figuras.-- et al.Dictyostelium and human MidA are homologous proteins that belong to a family of proteins of unknown function called DUF185. Using yeast two-hybrid screening and pull-down experiments, we showed that both proteins interact with the mitochondrial complex I subunit NDUFS2. Consistent with this, Dictyostelium cells lacking MidA showed a specific defect in complex I activity, and knockdown of human MidA in HEK293T cells resulted in reduced levels of assembled complex I. These results indicate a role for MidA in complex I assembly or stability. A structural bioinformatics analysis suggested the presence of a methyltransferase domain; this was further supported by site-directed mutagenesis of specific residues from the putative catalytic site. Interestingly, this complex I deficiency in a Dictyostelium midA- mutant causes a complex phenotypic outcome, which includes phototaxis and thermotaxis defects. We found that these aspects of the phenotype are mediated by a chronic activation of AMPK, revealing a possible role of AMPK signaling in complex I cytopathology.This work was supported by grants BMC2006-00394 and BMC2009-09050 to R.E. from the Spanish Ministerio de Ciencia e Innovación; to P.R.F. from the Thyne Reid Memorial Trusts and the Australian Research Council; to A.V. and O.G. from the Spanish National Bioinformatics Institute (www.inab.org), a platform of Genome Spain; to R.G. from the Fondo de Investigaciones Sanitarias, Instituto de Salud Carlos III, Spain (PI070167) and from the Comunidad de Madrid (GEN-0269/2006). S.C. is supported by a research contract from Consejería de Educación de la Comunidad de Madrid y del Fondo Social Europeo (FSE).Peer Reviewe
Challenges in the Analysis of Mass-Throughput Data: A Technical Commentary from the Statistical Machine Learning Perspective
Sound data analysis is critical to the success of modern molecular medicine research that involves collection and interpretation of mass-throughput data. The novel nature and high-dimensionality in such datasets pose a series of nontrivial data analysis problems. This technical commentary discusses the problems of over-fitting, error estimation, curse of dimensionality, causal versus predictive modeling, integration of heterogeneous types of data, and lack of standard protocols for data analysis. We attempt to shed light on the nature and causes of these problems and to outline viable methodological approaches to overcome them
Effects of Environment, Genetics and Data Analysis Pitfalls in an Esophageal Cancer Genome-Wide Association Study
The development of new high-throughput genotyping technologies has allowed fast evaluation of single nucleotide polymorphisms (SNPs) on a genome-wide scale. Several recent genome-wide association studies employing these technologies suggest that panels of SNPs can be a useful tool for predicting cancer susceptibility and discovery of potentially important new disease loci.In the present paper we undertake a careful examination of the relative significance of genetics, environmental factors, and biases of the data analysis protocol that was used in a previously published genome-wide association study. That prior study reported a nearly perfect discrimination of esophageal cancer patients and healthy controls on the basis of only genetic information. On the other hand, our results strongly suggest that SNPs in this dataset are not statistically linked to the phenotype, while several environmental factors and especially family history of esophageal cancer (a proxy to both environmental and genetic factors) have only a modest association with the disease.The main component of the previously claimed strong discriminatory signal is due to several data analysis pitfalls that in combination led to the strongly optimistic results. Such pitfalls are preventable and should be avoided in future studies since they create misleading conclusions and generate many false leads for subsequent research
Assessing quality and completeness of human transcriptional regulatory pathways on a genome-wide scale
Abstract Background Pathway databases are becoming increasingly important and almost omnipresent in most types of biological and translational research. However, little is known about the quality and completeness of pathways stored in these databases. The present study conducts a comprehensive assessment of transcriptional regulatory pathways in humans for seven well-studied transcription factors: MYC, NOTCH1, BCL6, TP53, AR, STAT1, and RELA. The employed benchmarking methodology first involves integrating genome-wide binding with functional gene expression data to derive direct targets of transcription factors. Then the lists of experimentally obtained direct targets are compared with relevant lists of transcriptional targets from 10 commonly used pathway databases. Results The results of this study show that for the majority of pathway databases, the overlap between experimentally obtained target genes and targets reported in transcriptional regulatory pathway databases is surprisingly small and often is not statistically significant. The only exception is MetaCore pathway database which yields statistically significant intersection with experimental results in 84% cases. Additionally, we suggest that the lists of experimentally derived direct targets obtained in this study can be used to reveal new biological insight in transcriptional regulation and suggest novel putative therapeutic targets in cancer. Conclusions Our study opens a debate on validity of using many popular pathway databases to obtain transcriptional regulatory targets. We conclude that the choice of pathway databases should be informed by solid scientific evidence and rigorous empirical evaluation. Reviewers This article was reviewed by Prof. Wing Hung Wong, Dr. Thiago Motta Venancio (nominated by Dr. L Aravind), and Prof. Geoff J McLachlan.</p
The FAST-AIMS Clinical Mass Spectrometry Analysis System
Within clinical proteomics, mass spectrometry analysis of biological samples is emerging as an important high-throughput technology, capable of producing powerful diagnostic and prognostic models and identifying important disease biomarkers. As interest in this area grows, and the number of such proteomics datasets continues to increase, the need has developed for efficient, comprehensive, reproducible methods of mass spectrometry data analysis by both experts and nonexperts. We have designed and implemented a stand-alone software system, FAST-AIMS, which seeks to meet this need through automation of data preprocessing, feature selection, classification model generation, and performance estimation. FAST-AIMS is an efficient and user-friendly stand-alone software for predictive analysis of mass spectrometry data. The present resource review paper will describe the features and use of the FAST-AIMS system. The system is freely available for download for noncommercial use
An experimental study of the intrinsic stability of random forest variable importance measures
BACKGROUND: The stability of Variable Importance Measures (VIMs) based on random forest has recently received increased attention. Despite the extensive attention on traditional stability of data perturbations or parameter variations, few studies include influences coming from the intrinsic randomness in generating VIMs, i.e. bagging, randomization and permutation. To address these influences, in this paper we introduce a new concept of intrinsic stability of VIMs, which is defined as the self-consistence among feature rankings in repeated runs of VIMs without data perturbations and parameter variations. Two widely used VIMs, i.e., Mean Decrease Accuracy (MDA) and Mean Decrease Gini (MDG) are comprehensively investigated. The motivation of this study is two-fold. First, we empirically verify the prevalence of intrinsic stability of VIMs over many real-world datasets to highlight that the instability of VIMs does not originate exclusively from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. Second, through Spearman and Pearson tests we comprehensively investigate how different factors influence the intrinsic stability. RESULTS: The experiments are carried out on 19 benchmark datasets with diverse characteristics, including 10 high-dimensional and small-sample gene expression datasets. Experimental results demonstrate the prevalence of intrinsic stability of VIMs. Spearman and Pearson tests on the correlations between intrinsic stability and different factors show that #feature (number of features) and #sample (size of sample) have a coupling effect on the intrinsic stability. The synthetic indictor, #feature/#sample, shows both negative monotonic correlation and negative linear correlation with the intrinsic stability, while OOB accuracy has monotonic correlations with intrinsic stability. This indicates that high-dimensional, small-sample and high complexity datasets may suffer more from intrinsic instability of VIMs. Furthermore, with respect to parameter settings of random forest, a large number of trees is preferred. No significant correlations can be seen between intrinsic stability and other factors. Finally, the magnitude of intrinsic stability is always smaller than that of traditional stability. CONCLUSION: First, the prevalence of intrinsic stability of VIMs demonstrates that the instability of VIMs not only comes from data perturbations or parameter variations, but also stems from the intrinsic randomness of VIMs. This finding gives a better understanding of VIM stability, and may help reduce the instability of VIMs. Second, by investigating the potential factors of intrinsic stability, users would be more aware of the risks and hence more careful when using VIMs, especially on high-dimensional, small-sample and high complexity datasets
- …
