23,282 research outputs found

    A light metric spanner

    Full text link
    It has long been known that dd-dimensional Euclidean point sets admit (1+ϵ)(1+\epsilon)-stretch spanners with lightness WE=ϵO(d)W_E = \epsilon^{-O(d)}, that is total edge weight at most WEW_E times the weight of the minimum spaning tree of the set [DHN93]. Whether or not a similar result holds for metric spaces with low doubling dimension has remained an important open problem, and has resisted numerous attempts at resolution. In this paper, we resolve the question in the affirmative, and show that doubling spaces admit (1+ϵ)(1+\epsilon)-stretch spanners with lightness WD=(ddim/ϵ)O(ddim)W_D = (ddim/\epsilon)^{O(ddim)}. Important in its own right, our result also implies a much faster polynomial-time approximation scheme for the traveling salesman problemin doubling metric spaces, improving upon the bound presented in [BGK-12]

    Efficient Classification for Metric Data

    Full text link
    Recent advances in large-margin classification of data residing in general metric spaces (rather than Hilbert spaces) enable classification under various natural metrics, such as string edit and earthmover distance. A general framework developed for this purpose by von Luxburg and Bousquet [JMLR, 2004] left open the questions of computational efficiency and of providing direct bounds on generalization error. We design a new algorithm for classification in general metric spaces, whose runtime and accuracy depend on the doubling dimension of the data points, and can thus achieve superior classification performance in many common scenarios. The algorithmic core of our approach is an approximate (rather than exact) solution to the classical problems of Lipschitz extension and of Nearest Neighbor Search. The algorithm's generalization performance is guaranteed via the fat-shattering dimension of Lipschitz classifiers, and we present experimental evidence of its superiority to some common kernel methods. As a by-product, we offer a new perspective on the nearest neighbor classifier, which yields significantly sharper risk asymptotics than the classic analysis of Cover and Hart [IEEE Trans. Info. Theory, 1967].Comment: This is the full version of an extended abstract that appeared in Proceedings of the 23rd COLT, 201

    Near-optimal sample compression for nearest neighbors

    Full text link
    We present the first sample compression algorithm for nearest neighbors with non-trivial performance guarantees. We complement these guarantees by demonstrating almost matching hardness lower bounds, which show that our bound is nearly optimal. Our result yields new insight into margin-based nearest neighbor classification in metric spaces and allows us to significantly sharpen and simplify existing bounds. Some encouraging empirical results are also presented
    corecore