6,955 research outputs found

    The covert set-cover problem with application to Network Discovery

    Full text link
    We address a version of the set-cover problem where we do not know the sets initially (and hence referred to as covert) but we can query an element to find out which sets contain this element as well as query a set to know the elements. We want to find a small set-cover using a minimal number of such queries. We present a Monte Carlo randomized algorithm that approximates an optimal set-cover of size OPTOPT within O(logN)O(\log N) factor with high probability using O(OPTlog2N)O(OPT \cdot \log^2 N) queries where NN is the input size. We apply this technique to the network discovery problem that involves certifying all the edges and non-edges of an unknown nn-vertices graph based on layered-graph queries from a minimal number of vertices. By reducing it to the covert set-cover problem we present an O(log2n)O(\log^2 n)-competitive Monte Carlo randomized algorithm for the covert version of network discovery problem. The previously best known algorithm has a competitive ratio of Ω(nlogn)\Omega (\sqrt{n\log n}) and therefore our result achieves an exponential improvement

    Low Degree Metabolites Explain Essential Reactions and Enhance Modularity in Biological Networks

    Get PDF
    Recently there has been a lot of interest in identifying modules at the level of genetic and metabolic networks of organisms, as well as in identifying single genes and reactions that are essential for the organism. A goal of computational and systems biology is to go beyond identification towards an explanation of specific modules and essential genes and reactions in terms of specific structural or evolutionary constraints. In the metabolic networks of E. coli, S. cerevisiae and S. aureus, we identified metabolites with a low degree of connectivity, particularly those that are produced and/or consumed in just a single reaction. Using FBA we also determined reactions essential for growth in these metabolic networks. We find that most reactions identified as essential in these networks turn out to be those involving the production or consumption of low degree metabolites. Applying graph theoretic methods to these metabolic networks, we identified connected clusters of these low degree metabolites. The genes involved in several operons in E. coli are correctly predicted as those of enzymes catalyzing the reactions of these clusters. We independently identified clusters of reactions whose fluxes are perfectly correlated. We find that the composition of the latter `functional clusters' is also largely explained in terms of clusters of low degree metabolites in each of these organisms. Our findings mean that most metabolic reactions that are essential can be tagged by one or more low degree metabolites. Those reactions are essential because they are the only ways of producing or consuming their respective tagged metabolites. Furthermore, reactions whose fluxes are strongly correlated can be thought of as `glued together' by these low degree metabolites.Comment: 12 pages main text with 2 figures and 2 tables. 16 pages of Supplementary material. Revised version has title changed and contains study of 3 organisms instead of 1 earlie

    Linear Coding Schemes for the Distributed Computation of Subspaces

    Full text link
    Let X1,...,XmX_1, ..., X_m be a set of mm statistically dependent sources over the common alphabet Fq\mathbb{F}_q, that are linearly independent when considered as functions over the sample space. We consider a distributed function computation setting in which the receiver is interested in the lossless computation of the elements of an ss-dimensional subspace WW spanned by the elements of the row vector [X1,,Xm]Γ[X_1, \ldots, X_m]\Gamma in which the (m×s)(m \times s) matrix Γ\Gamma has rank ss. A sequence of three increasingly refined approaches is presented, all based on linear encoders. The first approach uses a common matrix to encode all the sources and a Korner-Marton like receiver to directly compute WW. The second improves upon the first by showing that it is often more efficient to compute a carefully chosen superspace UU of WW. The superspace is identified by showing that the joint distribution of the {Xi}\{X_i\} induces a unique decomposition of the set of all linear combinations of the {Xi}\{X_i\}, into a chain of subspaces identified by a normalized measure of entropy. This subspace chain also suggests a third approach, one that employs nested codes. For any joint distribution of the {Xi}\{X_i\} and any WW, the sum-rate of the nested code approach is no larger than that under the Slepian-Wolf (SW) approach. Under the SW approach, WW is computed by first recovering each of the {Xi}\{X_i\}. For a large class of joint distributions and subspaces WW, the nested code approach is shown to improve upon SW. Additionally, a class of source distributions and subspaces are identified, for which the nested-code approach is sum-rate optimal.Comment: To appear in IEEE Journal of Selected Areas in Communications (In-Network Computation: Exploring the Fundamental Limits), April 201

    Parameterized Lower Bound and Improved Kernel for Diamond-free Edge Deletion

    Get PDF
    A diamond is a graph obtained by removing an edge from a complete graph on four vertices. A graph is diamond-free if it does not contain an induced diamond. The Diamond-free Edge Deletion problem asks to find whether there exist at most k edges in the input graph whose deletion results in a diamond-free graph. The problem was proved to be NP-complete and a polynomial kernel of O(k^4) vertices was found by Fellows et. al. (Discrete Optimization, 2011). In this paper, we give an improved kernel of O(k^3) vertices for Diamond-free Edge Deletion. We give an alternative proof of the NP-completeness of the problem and observe that it cannot be solved in time 2^{o(k)} * n^{O(1)}, unless the Exponential Time Hypothesis fails
    corecore