• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Geometric Discrepancy; An Illustrated Guide (1999)

by J Matousek
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 77
Next 10 →

Quasi-Randomized Path Planning

by Michael S. Branicky, Steven M. Lavalle, Kari Olson, Libo Yang - In Proc. IEEE Int’l Conf. on Robotics and Automation , 2001
"... We propose the use of quasi-random sampling techniques for path planning in high-dimensional conguration spaces. Following similar trends from related numerical computation elds, we show several advantages oered by these techniques in comparison to random sampling. Our ideas are evaluated in the con ..."
Abstract - Cited by 74 (8 self) - Add to MetaCart
We propose the use of quasi-random sampling techniques for path planning in high-dimensional conguration spaces. Following similar trends from related numerical computation elds, we show several advantages oered by these techniques in comparison to random sampling. Our ideas are evaluated in the context of the probabilistic roadmap (PRM) framework. Two quasi-random variants of PRM-based planners are proposed: 1) a classical PRM with quasi-random sampling, and 2) a quasi-random Lazy-PRM. Both have been implemented, and are shown through experiments to oer some performance advantages in comparison to their randomized counterparts. 1 Introduction Over two decades of path planning research have led to two primary trends. In the 1980s, deterministic approaches provided both elegant, complete algorithms for solving the problem, and also useful approximate or incomplete algorithms. The curse of dimensionality due to high-dimensional conguration spaces motivated researchers from the 199...
(Show Context)

Citation Context

...crepancy for rectangle R. The discrepancy of point set P is obtained bysnding the maximum such dierence over all possible rectangles. Thus, one measure of the discrepancy that has been proposed (see [=-=19-=-] for a litany of others plus discussion) is the L1 -discrepancy of thesrst N points of a d-dimensional sequence, S, in [0; 1] d : sup [a;b)[0;1] d jfs 1 ; : : : ; s N g \ [a; b)j N d Y i=1 (b i a i )...

Range counting over multidimensional data streams

by Subhash Suri, Csaba D. T'oth, Yunhong Zhou - Discrete & Computational Geometry , 2004
"... \Lambda \Lambda Abstract We consider the problem of approximate range counting over streams of d-dimensional points. In the data stream model, the algorithm makes a single scan of the data, which is presented in an arbitrary order, and computes a compact summary (called a sketch). The sketch, whose ..."
Abstract - Cited by 30 (0 self) - Add to MetaCart
\Lambda \Lambda Abstract We consider the problem of approximate range counting over streams of d-dimensional points. In the data stream model, the algorithm makes a single scan of the data, which is presented in an arbitrary order, and computes a compact summary (called a sketch). The sketch, whose size depends on the approximation parameter ", can be used to count the number of points inside a query range within additive error "n, where n is the size of the stream. We present several results, deterministic and randomized, for both rectangle and halfplane ranges. 1 Introduction Data streams have emerged as an important paradigm for processing data that arrives and needs to be processed continuously. For instance, telecom service providers routinely monitor packet flows through their networks to infer usage patterns and signs of attack, or to optimize their routing tables. Financial markets, banks, web servers, and news organizations also generate rapid and continuous data streams.
(Show Context)

Citation Context

...wer simplex queries in O( √ n) time. An interested reader should consult one of the surveys [1, 19] for many related results and extensions to d dimensions. ε-approximations and geometric discrepancy =-=[10, 18]-=-. An ε-approximation of a point set P for a range space Q is a set S ⊂ P such that, for any range Q ∈ Q, |P | |S| · |S ∩Q| − |P ∩Q| ≤ ε|P |. A range query is answered by counting how many points of th...

Constructive algorithms for discrepancy minimization

by Nikhil Bansal - In FOCS , 2010
"... ar ..."
Abstract - Cited by 29 (1 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...cal integration, derandomization, communication complexity, machine learning, optimization and so on. We shall not attempt to describe these connections and applications here, but refer the reader to =-=[6, 9, 12]-=-. ∗IBM T. J. Watson Research Center, Yorktown Heights, NY 10598. E-mail: nikhil@us.ibm.com 1 1.1 Discrepancy of General Set Systems What is the discrepancy of an arbitrary set system with n elements a...

A randomized quasi-Monte Carlo simulation method for Markov chains

by Christian Lécot, Bruno Tuffin - Operations Research , 2007
"... Abstract. We introduce and study a randomized quasi-Monte Carlo method for estimating the state distribution at each step of a Markov chain. The number of steps in the chain can be random and unbounded. The method simulates n copies of the chain in parallel, using a (d + 1)-dimensional highly-unifor ..."
Abstract - Cited by 27 (9 self) - Add to MetaCart
Abstract. We introduce and study a randomized quasi-Monte Carlo method for estimating the state distribution at each step of a Markov chain. The number of steps in the chain can be random and unbounded. The method simulates n copies of the chain in parallel, using a (d + 1)-dimensional highly-uniform point set of cardinality n, randomized independently at each step, where d is the number of uniform random numbers required at each transition of the Markov chain. This technique is effective in particular to obtain a low-variance unbiased estimator of the expected total cost up to some random stopping time, when state-dependent costs are paid at each step. It is generally more effective when the state space has a natural order related to the cost function. We provide numerical illustrations where the variance reduction with respect to standard Monte Carlo is substantial. The variance can be reduced by factors of several thousands in some cases. We prove bounds on the convergence rate of the worst-case error and variance for special situations. In line with what is typically observed in randomized quasi-Monte Carlo contexts, our empirical results indicate much better convergence than what these bounds guarantee.
(Show Context)

Citation Context

...ption 2; for example, a digitally shifted version of the twodimensional Hammersley point set (or Sobol’ net, or Faure net) in prime base b, or an affine matrix scramble of it, when n is a power of b (=-=Matoušek 1999-=-, Owen 2003b), or even a stratified sample of n points in the n subsquares. Assumption 3. Let Assumption 2 hold, and let the random variables vi = uimod n−1/2 be pairwise independent and uniformly dis...

Mergeable Summaries

by Pankaj K. Agarwal, Jeff M. Phillips, Graham Cormode, Zhewei Wei, Zengfeng Huang, Ke Yi
"... We study the mergeability of data summaries. Informally speaking, mergeability requires that, given two summaries on two data sets, there is a way to merge the two summaries into a single summary on the union of the two data sets, while preserving the error and size guarantees. This property means t ..."
Abstract - Cited by 22 (7 self) - Add to MetaCart
We study the mergeability of data summaries. Informally speaking, mergeability requires that, given two summaries on two data sets, there is a way to merge the two summaries into a single summary on the union of the two data sets, while preserving the error and size guarantees. This property means that the summaries can be merged in a way like other algebraic operators such as sum and max, which is especially useful for computing summaries on massive distributed data. Several data summaries are trivially mergeable by construction, most notably all the sketches that are linear functions of the data sets. But some other fundamental ones like those for heavy hitters and quantiles, are not (known to be) mergeable. In this paper, we demonstrate that these summaries are indeed mergeable or can be made mergeable after appropriate modifications. Specifically, we show that for ε-approximate heavy hitters, there is a deterministic mergeable summary of size O(1/ε); for ε-approximate quantiles, there is a deterministic summary of size O ( 1 log(εn)) that has a restricted form of mergeability, ε and a randomized one of size O ( 1 1 log3/2) with full merge-ε ε ability. We also extend our results to geometric summaries such as ε-approximations and ε-kernels. We also achieve two results of independent interest: (1) we provide the best known randomized streaming bound for ε-approximate quantiles that depends only on ε, of size O ( 1 1 log3/2), and (2) we demonstrate that the MG and the ε ε SpaceSaving summaries for heavy hitters are isomorphic. Supported by NSF under grants CNS-05-40347, IIS-07-
(Show Context)

Citation Context

...ε)) [27], and an ε-approximation of size O((1/ε) log 2d (1/ε)) [38] can be computed efficiently. More generally, an ε-approximation of size O(1/ε 2ν/(ν+1) ) exists for a range space of VC-dimension ν =-=[34]-=-. Furthermore, such an ε-approximation can be constructed using Bansal’s algorithm [5]; see also [11, 34]. These algorithms for constructing ε-approximations are not known to be mergeable. Although th...

From Discrepancy to Declustering: Near-optimal multidimensional declustering strategies for range queries (Extended Abstract)

by Chung-Min Chen, Christine T. Cheng , 2001
"... Declustering schemes allocate data blocks among multiple disks to enable parallel retrieval. Given a declustering scheme D, its response time with respect to a query Q, rt(Q), is defined to be the maximum number of disk blocks of the query stored by the scheme in any one of the disks. If |Q| is the ..."
Abstract - Cited by 22 (2 self) - Add to MetaCart
Declustering schemes allocate data blocks among multiple disks to enable parallel retrieval. Given a declustering scheme D, its response time with respect to a query Q, rt(Q), is defined to be the maximum number of disk blocks of the query stored by the scheme in any one of the disks. If |Q| is the number of data blocks in Q and M is the number of disks then rt(Q) is at least |Q|/M. One way to evaluate the performance of D with respect to a set of queries Q is to measure its additive error- the maximum difference between rt(Q) from |Q|/M over all range queries Q ∈ Q. In this paper, we consider the problem of designing declustering schemes for uniform multidimensional data arranged in a d-dimensional grid so that their additive errors with respect to range queries are as small as possible. It has been shown that such declustering schemes will have an additive error of Ω(log M) when d = 2 and Ω(log d−1 2 M) when d> 2 with respect to range queries. Asymptotically optimal declustering schemes exist for 2dimensional data. For data in larger dimensions, however, the best bound for additive errors is O(M d−1), which is extremely large. In this paper, we propose the two declustering schemes based on low discrepancy points in d-dimensions. When d is fixed, both schemes have an additive error of O(log d−1 M) with respect to range queries provided certain conditions are satisfied: the first scheme requires d ≥ 3 and M to be a power of a prime where the prime is at least d while the second scheme requires the size of the data to grow within some polynomial of M, with no restriction on

The Geometry of Differential Privacy: The Sparse and Approximate Cases

by Aleksandar Nikolov, et al. , 2012
"... In this work, we study trade-offs between accuracy and privacy in the context of linear queries over histograms. This is a rich class of queries that includes contingency tables and range queries, and has been a focus of a long line of work [BLR08,RR10,DRV10,HT10,HR10,LHR+10,BDKT12]. For a given set ..."
Abstract - Cited by 16 (5 self) - Add to MetaCart
In this work, we study trade-offs between accuracy and privacy in the context of linear queries over histograms. This is a rich class of queries that includes contingency tables and range queries, and has been a focus of a long line of work [BLR08,RR10,DRV10,HT10,HR10,LHR+10,BDKT12]. For a given set of d linear queries over a database x ∈ RN, we seek to find the differentially private mechanism that has the minimum mean squared error. For pure differential privacy, [HT10, BDKT12] give an O(log2 d) approximation to the optimal mechanism. Our first contribution is to give an O(log2 d) approximation guarantee for the case of (ε, δ)-differential privacy. Our mechanism is simple, efficient and adds carefully chosen correlated Gaussian noise to the answers. We prove its approximation guarantee relative to the hereditary discrepancy lower bound of [MN12], using tools from convex geometry. We next consider this question in the case when the number of queries exceeds the number of individuals in the database, i.e. when d> n, ‖x‖1. The lower bounds used in the previous approximation algorithm no longer apply, and in fact better mechanisms are known in this setting [BLR08,RR10,HR10,GHRU11,GRU12]. Our second main contribution is to give an (ε, δ)-differentially private mechanism that for a given query set A and an upper bound n on ‖x‖1, has mean squared error within polylog(d,N) of the optimal for A and n. This approximation is achieved by coupling the Gaussian noise addition approach with linear regression over the `1 ball. Additionally, we show a similar polylogarithmic approximation guarantee for the best ε-differentially private mechanism in this sparse setting. Our work also shows that for arbitrary counting queries, i.e. A with entries in {0, 1}, there is an ε-differentially private mechanism with expected error Õ(√n) per query, improving on the Õ(n 2 3) bound of [BLR08], and matching the lower bound implied by [DN03] up to logarithmic factors. The connection between hereditary discrepancy and the privacy mechanism enables us to derive the first polylogarithmic approximation to the hereditary discrepancy of a matrix A.

Shape Fitting on Point Sets with Probability Distributions

by Maarten Löffler, Jeff M. Phillips
"... Abstract. We consider problems on data sets where each data point has uncertainty described by an individual probability distribution. We develop several frameworks and algorithms for calculating statistics on these uncertain data sets. Our examples focus on geometric shape fitting problems. We prov ..."
Abstract - Cited by 14 (6 self) - Add to MetaCart
Abstract. We consider problems on data sets where each data point has uncertainty described by an individual probability distribution. We develop several frameworks and algorithms for calculating statistics on these uncertain data sets. Our examples focus on geometric shape fitting problems. We prove approximation guarantees for the algorithms with respect to the full probability distributions. We then empirically demonstrate that our algorithms are simple and practical, solving for a constant hidden by asymptotic analysis so that a user can reliably trade speed and size for accuracy. 1
(Show Context)

Citation Context

... that Ap = {A ∈ A | p ∈ A}. If (P, A) has VC-dimension ν, then (A, P ∗ ) has VC-dimension ≤ 2ν+1 . Thus, if the VC-dimension of (A, P ∗ ) is constant, then the VC-dimension of (P, A) is also constant =-=[20]-=-. When we have a distribution µ : Rd → R +, such that ∫ µ(x) dx = 1, we x∈R can think of this as the set P of all points in Rd , where the weight w of a point p ∈ Rd is µ(p). To simplify notation, we ...

Asymptotically optimal declustering schemes for range queries

by Rakesh K. Sinha, Randeep Bhatia - in 8th International Conference on Database Theory, Lecture Notes In Computer Science , 2001
"... d\Gamma 1 2) for d-dim schemes and to \Omega (log M) for 2-dim schemes, thus proving that the ..."
Abstract - Cited by 12 (2 self) - Add to MetaCart
d\Gamma 1 2) for d-dim schemes and to \Omega (log M) for 2-dim schemes, thus proving that the
(Show Context)

Citation Context

...eorem implies a signicantly better bound for the GRS scheme than what we had proved in an earlier paper [6]. 3 Intuition of Our Results All our schemes are motivated by results in discrepancy theory [=-=23]-=-. We give a very brief description of the relevant results from discrepancy theory. 3 3.1 Discrepancy Theory Given any integer M , the goal is to determine positions of M points in a 2-dimensional uni...

A weighted error metric and optimization method for antialiasing patterns

by Samuli Laine, Timo Aila - COMPUTER GRAPHICS FORUM , 2006
"... Displaying a synthetic image on a computer display requires determining the colors of individual pixels. To avoid aliasing, multiple samples of the image can be taken per pixel, after which the color of a pixel may be computed as a weighted sum of the samples. The positions and weights of the sample ..."
Abstract - Cited by 11 (1 self) - Add to MetaCart
Displaying a synthetic image on a computer display requires determining the colors of individual pixels. To avoid aliasing, multiple samples of the image can be taken per pixel, after which the color of a pixel may be computed as a weighted sum of the samples. The positions and weights of the samples play a major role in the resulting image quality, especially in real-time applications where usually only a handful of samples can be afforded per pixel. This paper presents a new error metric and an optimization method for antialiasing patterns used in image reconstruction. The metric is based on comparing the pattern against a given reference reconstruction filter in spatial domain and it takes into account psychovisually measured angle-specific acuities for sharp features.
(Show Context)

Citation Context

...[Shi91] computes discrepancy by sampling the area with rectangles of different sizes and aspect ratios, and Dobkin et al. [DM93] measure discrepancy based on arbitrary edges through a pixel. Matousek =-=[Mat99]-=- offers a more complete discussion on different types of discrepancy. Quasi Monte Carlo methods [Kel01] typically use deterministic low-discrepancy sequences (e.g. [Hal70]) for generating the sampling...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University