1,473 research outputs found

    Fooling intersections of low-weight halfspaces

    Full text link
    A weight-tt halfspace is a Boolean function f(x)=f(x)=sign(w1x1++wnxnθ)(w_1 x_1 + \cdots + w_n x_n - \theta) where each wiw_i is an integer in {t,,t}.\{-t,\dots,t\}. We give an explicit pseudorandom generator that δ\delta-fools any intersection of kk weight-tt halfspaces with seed length poly(logn,logk,t,1/δ)(\log n, \log k,t,1/\delta). In particular, our result gives an explicit PRG that fools any intersection of any quasipoly(n)(n) number of halfspaces of any polylog(n)\log(n) weight to any 1/1/polylog(n)\log(n) accuracy using seed length polylog(n).\log(n). Prior to this work no explicit PRG with non-trivial seed length was known even for fooling intersections of nn weight-1 halfspaces to constant accuracy. The analysis of our PRG fuses techniques from two different lines of work on unconditional pseudorandomness for different kinds of Boolean functions. We extend the approach of Harsha, Klivans and Meka \cite{HKM12} for fooling intersections of regular halfspaces, and combine this approach with results of Bazzi \cite{Bazzi:07} and Razborov \cite{Razborov:09} on bounded independence fooling CNF formulas. Our analysis introduces new coupling-based ingredients into the standard Lindeberg method for establishing quantitative central limit theorems and associated pseudorandomness results.Comment: 27 page

    Testing probability distributions using conditional samples

    Full text link
    We study a new framework for property testing of probability distributions, by considering distribution testing algorithms that have access to a conditional sampling oracle.* This is an oracle that takes as input a subset S[N]S \subseteq [N] of the domain [N][N] of the unknown probability distribution DD and returns a draw from the conditional probability distribution DD restricted to SS. This new model allows considerable flexibility in the design of distribution testing algorithms; in particular, testing algorithms in this model can be adaptive. We study a wide range of natural distribution testing problems in this new framework and some of its variants, giving both upper and lower bounds on query complexity. These problems include testing whether DD is the uniform distribution U\mathcal{U}; testing whether D=DD = D^\ast for an explicitly provided DD^\ast; testing whether two unknown distributions D1D_1 and D2D_2 are equivalent; and estimating the variation distance between DD and the uniform distribution. At a high level our main finding is that the new "conditional sampling" framework we consider is a powerful one: while all the problems mentioned above have Ω(N)\Omega(\sqrt{N}) sample complexity in the standard model (and in some cases the complexity must be almost linear in NN), we give poly(logN,1/ε)\mathrm{poly}(\log N, 1/\varepsilon)-query algorithms (and in some cases poly(1/ε)\mathrm{poly}(1/\varepsilon)-query algorithms independent of NN) for all these problems in our conditional sampling setting. *Independently from our work, Chakraborty et al. also considered this framework. We discuss their work in Subsection [1.4].Comment: Significant changes on Section 9 (detailing and expanding the proof of Theorem 16). Several clarifications and typos fixed in various place

    Efficient deterministic approximate counting for low-degree polynomial threshold functions

    Full text link
    We give a deterministic algorithm for approximately counting satisfying assignments of a degree-dd polynomial threshold function (PTF). Given a degree-dd input polynomial p(x1,,xn)p(x_1,\dots,x_n) over RnR^n and a parameter ϵ>0\epsilon> 0, our algorithm approximates Prx{1,1}n[p(x)0]\Pr_{x \sim \{-1,1\}^n}[p(x) \geq 0] to within an additive ±ϵ\pm \epsilon in time Od,ϵ(1)poly(nd)O_{d,\epsilon}(1)\cdot \mathop{poly}(n^d). (Any sort of efficient multiplicative approximation is impossible even for randomized algorithms assuming NPRPNP\not=RP.) Note that the running time of our algorithm (as a function of ndn^d, the number of coefficients of a degree-dd PTF) is a \emph{fixed} polynomial. The fastest previous algorithm for this problem (due to Kane), based on constructions of unconditional pseudorandom generators for degree-dd PTFs, runs in time nOd,c(1)ϵcn^{O_{d,c}(1) \cdot \epsilon^{-c}} for all c>0c > 0. The key novel contributions of this work are: A new multivariate central limit theorem, proved using tools from Malliavin calculus and Stein's Method. This new CLT shows that any collection of Gaussian polynomials with small eigenvalues must have a joint distribution which is very close to a multidimensional Gaussian distribution. A new decomposition of low-degree multilinear polynomials over Gaussian inputs. Roughly speaking we show that (up to some small error) any such polynomial can be decomposed into a bounded number of multilinear polynomials all of which have extremely small eigenvalues. We use these new ingredients to give a deterministic algorithm for a Gaussian-space version of the approximate counting problem, and then employ standard techniques for working with low-degree PTFs (invariance principles and regularity lemmas) to reduce the original approximate counting problem over the Boolean hypercube to the Gaussian version

    An average-case depth hierarchy theorem for Boolean circuits

    Full text link
    We prove an average-case depth hierarchy theorem for Boolean circuits over the standard basis of AND\mathsf{AND}, OR\mathsf{OR}, and NOT\mathsf{NOT} gates. Our hierarchy theorem says that for every d2d \geq 2, there is an explicit nn-variable Boolean function ff, computed by a linear-size depth-dd formula, which is such that any depth-(d1)(d-1) circuit that agrees with ff on (1/2+on(1))(1/2 + o_n(1)) fraction of all inputs must have size exp(nΩ(1/d)).\exp({n^{\Omega(1/d)}}). This answers an open question posed by H{\aa}stad in his Ph.D. thesis. Our average-case depth hierarchy theorem implies that the polynomial hierarchy is infinite relative to a random oracle with probability 1, confirming a conjecture of H{\aa}stad, Cai, and Babai. We also use our result to show that there is no "approximate converse" to the results of Linial, Mansour, Nisan and Boppana on the total influence of small-depth circuits, thus answering a question posed by O'Donnell, Kalai, and Hatami. A key ingredient in our proof is a notion of \emph{random projections} which generalize random restrictions

    Efficiency versus Convergence of Boolean Kernels for On-Line Learning Algorithms

    Full text link
    The paper studies machine learning problems where each example is described using a set of Boolean features and where hypotheses are represented by linear threshold elements. One method of increasing the expressiveness of learned hypotheses in this context is to expand the feature set to include conjunctions of basic features. This can be done explicitly or where possible by using a kernel function. Focusing on the well known Perceptron and Winnow algorithms, the paper demonstrates a tradeoff between the computational efficiency with which the algorithm can be run over the expanded feature space and the generalization ability of the corresponding learning algorithm. We first describe several kernel functions which capture either limited forms of conjunctions or all conjunctions. We show that these kernels can be used to efficiently run the Perceptron algorithm over a feature space of exponentially many conjunctions; however we also show that using such kernels, the Perceptron algorithm can provably make an exponential number of mistakes even when learning simple functions. We then consider the question of whether kernel functions can analogously be used to run the multiplicative-update Winnow algorithm over an expanded feature space of exponentially many conjunctions. Known upper bounds imply that the Winnow algorithm can learn Disjunctive Normal Form (DNF) formulae with a polynomial mistake bound in this setting. However, we prove that it is computationally hard to simulate Winnows behavior for learning DNF over such a feature set. This implies that the kernel functions which correspond to running Winnow for this problem are not efficiently computable, and that there is no general construction that can run Winnow with kernels
    corecore