1,473 research outputs found
Fooling intersections of low-weight halfspaces
A weight- halfspace is a Boolean function sign where each is an integer in We give
an explicit pseudorandom generator that -fools any intersection of
weight- halfspaces with seed length poly. In
particular, our result gives an explicit PRG that fools any intersection of any
quasipoly number of halfspaces of any poly weight to any
poly accuracy using seed length poly Prior to this work
no explicit PRG with non-trivial seed length was known even for fooling
intersections of weight-1 halfspaces to constant accuracy.
The analysis of our PRG fuses techniques from two different lines of work on
unconditional pseudorandomness for different kinds of Boolean functions. We
extend the approach of Harsha, Klivans and Meka \cite{HKM12} for fooling
intersections of regular halfspaces, and combine this approach with results of
Bazzi \cite{Bazzi:07} and Razborov \cite{Razborov:09} on bounded independence
fooling CNF formulas. Our analysis introduces new coupling-based ingredients
into the standard Lindeberg method for establishing quantitative central limit
theorems and associated pseudorandomness results.Comment: 27 page
Testing probability distributions using conditional samples
We study a new framework for property testing of probability distributions,
by considering distribution testing algorithms that have access to a
conditional sampling oracle.* This is an oracle that takes as input a subset of the domain of the unknown probability distribution
and returns a draw from the conditional probability distribution restricted
to . This new model allows considerable flexibility in the design of
distribution testing algorithms; in particular, testing algorithms in this
model can be adaptive.
We study a wide range of natural distribution testing problems in this new
framework and some of its variants, giving both upper and lower bounds on query
complexity. These problems include testing whether is the uniform
distribution ; testing whether for an explicitly
provided ; testing whether two unknown distributions and
are equivalent; and estimating the variation distance between and the
uniform distribution. At a high level our main finding is that the new
"conditional sampling" framework we consider is a powerful one: while all the
problems mentioned above have sample complexity in the
standard model (and in some cases the complexity must be almost linear in ),
we give -query algorithms (and in some
cases -query algorithms independent of ) for
all these problems in our conditional sampling setting.
*Independently from our work, Chakraborty et al. also considered this
framework. We discuss their work in Subsection [1.4].Comment: Significant changes on Section 9 (detailing and expanding the proof
of Theorem 16). Several clarifications and typos fixed in various place
Efficient deterministic approximate counting for low-degree polynomial threshold functions
We give a deterministic algorithm for approximately counting satisfying
assignments of a degree- polynomial threshold function (PTF). Given a
degree- input polynomial over and a parameter
, our algorithm approximates to within an additive in time . (Any sort of efficient multiplicative approximation is
impossible even for randomized algorithms assuming .) Note that the
running time of our algorithm (as a function of , the number of
coefficients of a degree- PTF) is a \emph{fixed} polynomial. The fastest
previous algorithm for this problem (due to Kane), based on constructions of
unconditional pseudorandom generators for degree- PTFs, runs in time
for all .
The key novel contributions of this work are: A new multivariate central
limit theorem, proved using tools from Malliavin calculus and Stein's Method.
This new CLT shows that any collection of Gaussian polynomials with small
eigenvalues must have a joint distribution which is very close to a
multidimensional Gaussian distribution. A new decomposition of low-degree
multilinear polynomials over Gaussian inputs. Roughly speaking we show that (up
to some small error) any such polynomial can be decomposed into a bounded
number of multilinear polynomials all of which have extremely small
eigenvalues. We use these new ingredients to give a deterministic algorithm for
a Gaussian-space version of the approximate counting problem, and then employ
standard techniques for working with low-degree PTFs (invariance principles and
regularity lemmas) to reduce the original approximate counting problem over the
Boolean hypercube to the Gaussian version
An average-case depth hierarchy theorem for Boolean circuits
We prove an average-case depth hierarchy theorem for Boolean circuits over
the standard basis of , , and gates.
Our hierarchy theorem says that for every , there is an explicit
-variable Boolean function , computed by a linear-size depth- formula,
which is such that any depth- circuit that agrees with on fraction of all inputs must have size This
answers an open question posed by H{\aa}stad in his Ph.D. thesis.
Our average-case depth hierarchy theorem implies that the polynomial
hierarchy is infinite relative to a random oracle with probability 1,
confirming a conjecture of H{\aa}stad, Cai, and Babai. We also use our result
to show that there is no "approximate converse" to the results of Linial,
Mansour, Nisan and Boppana on the total influence of small-depth circuits, thus
answering a question posed by O'Donnell, Kalai, and Hatami.
A key ingredient in our proof is a notion of \emph{random projections} which
generalize random restrictions
Efficiency versus Convergence of Boolean Kernels for On-Line Learning Algorithms
The paper studies machine learning problems where each example is described
using a set of Boolean features and where hypotheses are represented by linear
threshold elements. One method of increasing the expressiveness of learned
hypotheses in this context is to expand the feature set to include conjunctions
of basic features. This can be done explicitly or where possible by using a
kernel function. Focusing on the well known Perceptron and Winnow algorithms,
the paper demonstrates a tradeoff between the computational efficiency with
which the algorithm can be run over the expanded feature space and the
generalization ability of the corresponding learning algorithm. We first
describe several kernel functions which capture either limited forms of
conjunctions or all conjunctions. We show that these kernels can be used to
efficiently run the Perceptron algorithm over a feature space of exponentially
many conjunctions; however we also show that using such kernels, the Perceptron
algorithm can provably make an exponential number of mistakes even when
learning simple functions. We then consider the question of whether kernel
functions can analogously be used to run the multiplicative-update Winnow
algorithm over an expanded feature space of exponentially many conjunctions.
Known upper bounds imply that the Winnow algorithm can learn Disjunctive Normal
Form (DNF) formulae with a polynomial mistake bound in this setting. However,
we prove that it is computationally hard to simulate Winnows behavior for
learning DNF over such a feature set. This implies that the kernel functions
which correspond to running Winnow for this problem are not efficiently
computable, and that there is no general construction that can run Winnow with
kernels
- …
