2,006 research outputs found

    Evaluating testing methods by delivered reliability

    Get PDF
    There are two main goals in testing software: (1) to achieve adequate quality (debug testing), where the objective is to probe the software for defects so that these can be removed, and (2) to assess existing quality (operational testing), where the objective is to gain confidence that the software is reliable. Debug methods tend to ignore random selection of test data from an operational profile, while for operational methods this selection is all-important. Debug methods are thought to be good at uncovering defects so that these can be repaired, but having done so they do not provide a technically defensible assessment of the reliability that results. On the other hand, operational methods provide accurate assessment, but may not be as useful for achieving reliability. This paper examines the relationship between the two testing goals, using a probabilistic analysis. We define simple models of programs and their testing, and try to answer the question of how to attain program reliability: is it better to test by probing for defects as in debug testing, or to assess reliability directly as in operational testing? Testing methods are compared in a model where program failures are detected and the software changed to eliminate them. The “better” method delivers higher reliability after all test failures have been eliminated. Special cases are exhibited in which each kind of testing is superior. An analysis of the distribution of the delivered reliability indicates that even simple models have unusual statistical properties, suggesting caution in interpreting theoretical comparisons

    Two-Source Dispersers for Polylogarithmic Entropy and Improved Ramsey Graphs

    Full text link
    In his 1947 paper that inaugurated the probabilistic method, Erd\H{o}s proved the existence of 2logn2\log{n}-Ramsey graphs on nn vertices. Matching Erd\H{o}s' result with a constructive proof is a central problem in combinatorics, that has gained a significant attention in the literature. The state of the art result was obtained in the celebrated paper by Barak, Rao, Shaltiel and Wigderson [Ann. Math'12], who constructed a 22(loglogn)1α2^{2^{(\log\log{n})^{1-\alpha}}}-Ramsey graph, for some small universal constant α>0\alpha > 0. In this work, we significantly improve the result of Barak~\etal and construct 2(loglogn)c2^{(\log\log{n})^c}-Ramsey graphs, for some universal constant cc. In the language of theoretical computer science, our work resolves the problem of explicitly constructing two-source dispersers for polylogarithmic entropy

    A Characterization of Mixed Unit Interval Graphs

    Full text link
    We give a complete characterization of mixed unit interval graphs, the intersection graphs of closed, open, and half-open unit intervals of the real line. This is a proper superclass of the well known unit interval graphs. Our result solves a problem posed by Dourado, Le, Protti, Rautenbach and Szwarcfiter (Mixed unit interval graphs, Discrete Math. 312, 3357-3363 (2012)).Comment: 17 pages, referees' comments adde

    Quantum Weakly Nondeterministic Communication Complexity

    Full text link
    We study the weakest model of quantum nondeterminism in which a classical proof has to be checked with probability one by a quantum protocol. We show the first separation between classical nondeterministic communication complexity and this model of quantum nondeterministic communication complexity for a total function. This separation is quadratic.Comment: 12 pages. v3: minor correction

    Non locality, closing the detection loophole and communication complexity

    Get PDF
    It is shown that the detection loophole which arises when trying to rule out local realistic theories as alternatives for quantum mechanics can be closed if the detection efficiency η\eta is larger than ηd1/220.0035d\eta \geq d^{1/2} 2^{-0.0035d} where dd is the dimension of the entangled system. Furthermore it is argued that this exponential decrease of the detector efficiency required to close the detection loophole is almost optimal. This argument is based on a close connection that exists between closing the detection loophole and the amount of classical communication required to simulate quantum correlation when the detectors are perfect.Comment: 4 pages Latex, minor typos correcte

    Diagonally Neighbour Transitive Codes and Frequency Permutation Arrays

    Get PDF
    Constant composition codes have been proposed as suitable coding schemes to solve the narrow band and impulse noise problems associated with powerline communication. In particular, a certain class of constant composition codes called frequency permutation arrays have been suggested as ideal, in some sense, for these purposes. In this paper we characterise a family of neighbour transitive codes in Hamming graphs in which frequency permutation arrays play a central rode. We also classify all the permutation codes generated by groups in this family

    Synchronizing Automata on Quasi Eulerian Digraph

    Full text link
    In 1964 \v{C}ern\'{y} conjectured that each nn-state synchronizing automaton posesses a reset word of length at most (n1)2(n-1)^2. From the other side the best known upper bound on the reset length (minimum length of reset words) is cubic in nn. Thus the main problem here is to prove quadratic (in nn) upper bounds. Since 1964, this problem has been solved for few special classes of \sa. One of this result is due to Kari \cite{Ka03} for automata with Eulerian digraphs. In this paper we introduce a new approach to prove quadratic upper bounds and explain it in terms of Markov chains and Perron-Frobenius theories. Using this approach we obtain a quadratic upper bound for a generalization of Eulerian automata.Comment: 8 pages, 1 figur

    Submonolayer Epitaxy Without A Critical Nucleus

    Full text link
    The nucleation and growth of two--dimensional islands is studied with Monte Carlo simulations of a pair--bond solid--on--solid model of epitaxial growth. The conventional description of this problem in terms of a well--defined critical island size fails because no islands are absolutely stable against single atom detachment by thermal bond breaking. When two--bond scission is negligible, we find that the ratio of the dimer dissociation rate to the rate of adatom capture by dimers uniquely indexes both the island size distribution scaling function and the dependence of the island density on the flux and the substrate temperature. Effective pair-bond model parameters are found that yield excellent quantitative agreement with scaling functions measured for Fe/Fe(001).Comment: 8 pages, Postscript files (the paper and Figs. 1-3), uuencoded, compressed and tarred. Surface Science Letters, in press
    corecore