1,514 research outputs found
Gene set bagging for estimating replicability of gene set analyses
Background: Significance analysis plays a major role in identifying and
ranking genes, transcription factor binding sites, DNA methylation regions, and
other high-throughput features for association with disease. We propose a new
approach, called gene set bagging, for measuring the stability of ranking
procedures using predefined gene sets. Gene set bagging involves resampling the
original high-throughput data, performing gene-set analysis on the resampled
data, and confirming that biological categories replicate. This procedure can
be thought of as bootstrapping gene-set analysis and can be used to determine
which are the most reproducible gene sets. Results: Here we apply this approach
to two common genomics applications: gene expression and DNA methylation. Even
with state-of-the-art statistical ranking procedures, significant categories in
a gene set enrichment analysis may be unstable when subjected to resampling.
Conclusions: We demonstrate that gene lists are not necessarily stable, and
therefore additional steps like gene set bagging can improve biological
inference of gene set analysis.Comment: 3 Figure
Significance Analysis of Time Course Microarray Experiments
Characterizing the genome-wide dynamic regulation of gene expression is important and will be of much interest in the future. However, there is currently no established method for identifying differentially expressed genes in a time course study. Here we propose a significance method for analyzing time course microarray studies that can be applied to the typical types of comparisons and sampling schemes. This method is applied to two studies on humans. In one study, genes are identified that show differential expression over time in response to in vivo endotoxin administration. Using our method 7409 genes are called significant at a 1% FDR level, whereas several existing approaches fail to identify any genes. In another study, 417 genes are identified at a 10% FDR level that show expression changing with age in the kidney cortex. Here it is also shown that as many as 47% of the genes change with age in a manner more complex than simple exponential growth or decay. The methodology proposed here has been implemented in the freely distributed and open-source EDGE software package
MLP: a MATLAB toolbox for rapid and reliable auditory threshold estimation
In this paper, we present MLP, a MATLAB toolbox enabling auditory
thresholds estimation via the adaptive Maximum Likelihood procedure proposed
by David Green (1990, 1993). This adaptive procedure is particularly appealing for
those psychologists that need to estimate thresholds with a good degree of accuracy
and in a short time. Together with a description of the toolbox, the current text
provides an introduction to the threshold estimation theory and a theoretical
explanation of the maximum likelihood adaptive procedure. MLP comes with a
graphical interface and it is provided with several built-in, classic psychoacoustics
experiments ready to use at a mouse click
Having Fun in Learning Formal Specifications
There are many benefits in providing formal specifications for our software.
However, teaching students to do this is not always easy as courses on formal
methods are often experienced as dry by students. This paper presents a game
called FormalZ that teachers can use to introduce some variation in their
class. Students can have some fun in playing the game and, while doing so, also
learn the basics of writing formal specifications in the form of pre- and
post-conditions. Unlike existing software engineering themed education games
such as Pex and Code Defenders, FormalZ takes the deep gamification approach
where playing gets a more central role in order to generate more engagement.
This short paper presents our work in progress: the first implementation of
FormalZ along with the result of a preliminary users' evaluation. This
implementation is functionally complete and tested, but the polishing of its
user interface is still future work
Microcavity controlled coupling of excitonic qubits
Controlled non-local energy and coherence transfer enables light harvesting
in photosynthesis and non-local logical operations in quantum computing. The
most relevant mechanism of coherent coupling of distant qubits is coupling via
the electromagnetic field. Here, we demonstrate the controlled coherent
coupling of spatially separated excitonic qubits via the photon mode of a solid
state microresonator. This is revealed by two-dimensional spectroscopy of the
sample's coherent response, a sensitive and selective probe of the coherent
coupling. The experimental results are quantitatively described by a rigorous
theory of the cavity mediated coupling within a cluster of quantum dots
excitons. Having demonstrated this mechanism, it can be used in extended
coupling channels - sculptured, for instance, in photonic crystal cavities - to
enable a long-range, non-local wiring up of individual emitters in solids
Differential expression analysis with global network adjustment
<p>Background: Large-scale chromosomal deletions or other non-specific perturbations of the transcriptome can alter the expression of hundreds or thousands of genes, and it is of biological interest to understand which genes are most profoundly affected. We present a method for predicting a gene’s expression as a function of other genes thereby accounting for the effect of transcriptional regulation that confounds the identification of genes differentially expressed relative to a regulatory network. The challenge in constructing such models is that the number of possible regulator transcripts within a global network is on the order of thousands, and the number of biological samples is typically on the order of 10. Nevertheless, there are large gene expression databases that can be used to construct networks that could be helpful in modeling transcriptional regulation in smaller experiments.</p>
<p>Results: We demonstrate a type of penalized regression model that can be estimated from large gene expression databases, and then applied to smaller experiments. The ridge parameter is selected by minimizing the cross-validation error of the predictions in the independent out-sample. This tends to increase the model stability and leads to a much greater degree of parameter shrinkage, but the resulting biased estimation is mitigated by a second round of regression. Nevertheless, the proposed computationally efficient “over-shrinkage” method outperforms previously used LASSO-based techniques. In two independent datasets, we find that the median proportion of explained variability in expression is approximately 25%, and this results in a substantial increase in the signal-to-noise ratio allowing more powerful inferences on differential gene expression leading to biologically intuitive findings. We also show that a large proportion of gene dependencies are conditional on the biological state, which would be impossible with standard differential expression methods.</p>
<p>Conclusions: By adjusting for the effects of the global network on individual genes, both the sensitivity and reliability of differential expression measures are greatly improved.</p>
- …
