176,649 research outputs found

    Tolerating Correlated Failures in Massively Parallel Stream Processing Engines

    Full text link
    Fault-tolerance techniques for stream processing engines can be categorized into passive and active approaches. A typical passive approach periodically checkpoints a processing task's runtime states and can recover a failed task by restoring its runtime state using its latest checkpoint. On the other hand, an active approach usually employs backup nodes to run replicated tasks. Upon failure, the active replica can take over the processing of the failed task with minimal latency. However, both approaches have their own inadequacies in Massively Parallel Stream Processing Engines (MPSPE). The passive approach incurs a long recovery latency especially when a number of correlated nodes fail simultaneously, while the active approach requires extra replication resources. In this paper, we propose a new fault-tolerance framework, which is Passive and Partially Active (PPA). In a PPA scheme, the passive approach is applied to all tasks while only a selected set of tasks will be actively replicated. The number of actively replicated tasks depends on the available resources. If tasks without active replicas fail, tentative outputs will be generated before the completion of the recovery process. We also propose effective and efficient algorithms to optimize a partially active replication plan to maximize the quality of tentative outputs. We implemented PPA on top of Storm, an open-source MPSPE and conducted extensive experiments using both real and synthetic datasets to verify the effectiveness of our approach

    Incremental learning with respect to new incoming input attributes

    Get PDF
    Neural networks are generally exposed to a dynamic environment where the training patterns or the input attributes (features) will likely be introduced into the current domain incrementally. This paper considers the situation where a new set of input attributes must be considered and added into the existing neural network. The conventional method is to discard the existing network and redesign one from scratch. This approach wastes the old knowledge and the previous effort. In order to reduce computational time, improve generalization accuracy, and enhance intelligence of the learned models, we present ILIA algorithms (namely ILIA1, ILIA2, ILIA3, ILIA4 and ILIA5) capable of Incremental Learning in terms of Input Attributes. Using the ILIA algorithms, when new input attributes are introduced into the original problem, the existing neural network can be retained and a new sub-network is constructed and trained incrementally. The new sub-network and the old one are merged later to form a new network for the changed problem. In addition, ILIA algorithms have the ability to decide whether the new incoming input attributes are relevant to the output and consistent with the existing input attributes or not and suggest to accept or reject them. Experimental results show that the ILIA algorithms are efficient and effective both for the classification and regression problems

    Cooperative Data Exchange based on MDS Codes

    Full text link
    The cooperative data exchange problem is studied for the fully connected network. In this problem, each node initially only possesses a subset of the KK packets making up the file. Nodes make broadcast transmissions that are received by all other nodes. The goal is for each node to recover the full file. In this paper, we present a polynomial-time deterministic algorithm to compute the optimal (i.e., minimal) number of required broadcast transmissions and to determine the precise transmissions to be made by the nodes. A particular feature of our approach is that {\it each} of the KdK-d transmissions is a linear combination of {\it exactly} d+1d+1 packets, and we show how to optimally choose the value of d.d. We also show how the coefficients of these linear combinations can be chosen by leveraging a connection to Maximum Distance Separable (MDS) codes. Moreover, we show that our method can be used to solve cooperative data exchange problems with weighted cost as well as the so-called successive local omniscience problem.Comment: 21 pages, 1 figur

    Bootstrapping Mixed Correlators in the Five Dimensional Critical O(N) Models

    Get PDF
    We use the conformal bootstrap approach to explore 5D5D CFTs with O(N)O(N) global symmetry, which contain NN scalars ϕi\phi_i transforming as O(N)O(N) vector. Specifically, we study multiple four-point correlators of the leading O(N)O(N) vector ϕi\phi_i and the O(N)O(N) singlet σ\sigma. The crossing symmetry of the four-point functions and the unitarity condition provide nontrivial constraints on the scaling dimensions (Δϕ\Delta_\phi, Δσ\Delta_\sigma) of ϕi\phi_i and σ\sigma. With reasonable assumptions on the gaps between scaling dimensions of ϕi\phi_i (σ\sigma) and the next O(N)O(N) vector (singlet) scalar, we are able to isolate the scaling dimensions (Δϕ(\Delta_\phi, Δσ)\Delta_\sigma) in small islands. In particular, for large N=500N=500, the isolated region is highly consistent with the result obtained from large NN expansion. We also study the interacting O(N)O(N) CFTs for 1N1001\leqslant N\leqslant100. Isolated regions on (Δϕ,Δσ)(\Delta_\phi,\Delta_\sigma) plane are obtained using conformal bootstrap program with lower order of derivatives Λ\Lambda; however, they disappear after increasing Λ\Lambda. We think these islands are corresponding to interacting but nonunitary O(N)O(N) CFTs. Our results provide a lower bound on the critical value Nc>100N_c>100, below which the interacting O(N)O(N) CFTs turn into nonunitary. The critical value is unexpectedly large comparing with previous estimations.Comment: 28 pages, 4 figure
    corecore