Results 1  10
of
79
Property Testing and its connection to Learning and Approximation
"... We study the question of determining whether an unknown function has a particular property or is fflfar from any function with that property. A property testing algorithm is given a sample of the value of the function on instances drawn according to some distribution, and possibly may query the fun ..."
Abstract

Cited by 498 (68 self)
 Add to MetaCart
We study the question of determining whether an unknown function has a particular property or is fflfar from any function with that property. A property testing algorithm is given a sample of the value of the function on instances drawn according to some distribution, and possibly may query the function on instances of its choice. First, we establish some connections between property testing and problems in learning theory. Next, we focus on testing graph properties, and devise algorithms to test whether a graph has properties such as being kcolorable or having a aeclique (clique of density ae w.r.t the vertex set). Our graph property testing algorithms are probabilistic and make assertions which are correct with high probability, utilizing only poly(1=ffl) edgequeries into the graph, where ffl is the distance parameter. Moreover, the property testing algorithms can be used to efficiently (i.e., in time linear in the number of vertices) construct partitions of the graph which corre...
Correlation Clustering
 MACHINE LEARNING
, 2002
"... We consider the following clustering problem: we have a complete graph on # vertices (items), where each edge ### ## is labeled either # or depending on whether # and # have been deemed to be similar or different. The goal is to produce a partition of the vertices (a clustering) that agrees as mu ..."
Abstract

Cited by 329 (4 self)
 Add to MetaCart
(Show Context)
We consider the following clustering problem: we have a complete graph on # vertices (items), where each edge ### ## is labeled either # or depending on whether # and # have been deemed to be similar or different. The goal is to produce a partition of the vertices (a clustering) that agrees as much as possible with the edge labels. That is, we want a clustering that maximizes the number of # edges within clusters, plus the number of edges between clusters (equivalently, minimizes the number of disagreements: the number of edges inside clusters plus the number of # edges between clusters). This formulation is motivated from a document clustering problem in which one has a pairwise similarity function # learned from past data, and the goal is to partition the current set of documents in a way that correlates with # as much as possible; it can also be viewed as a kind of "agnostic learning" problem. An interesting
Aggregating inconsistent information: ranking and clustering
, 2005
"... We address optimization problems in which we are given contradictory pieces of input information and the goal is to find a globally consistent solution that minimizes the number of disagreements with the respective inputs. Specifically, the problems we address are rank aggregation, the feedback arc ..."
Abstract

Cited by 229 (17 self)
 Add to MetaCart
We address optimization problems in which we are given contradictory pieces of input information and the goal is to find a globally consistent solution that minimizes the number of disagreements with the respective inputs. Specifically, the problems we address are rank aggregation, the feedback arc set problem on tournaments, and correlation and consensus clustering. We show that for all these problems (and various weighted versions of them), we can obtain improved approximation factors using essentially the same remarkably simple algorithm. Additionally, we almost settle a longstanding conjecture of BangJensen and Thomassen and show that unless NP⊆BPP, there is no polynomial time algorithm for the problem of minimum feedback arc set in tournaments.
Polynomial Time Approximation Schemes for Dense Instances of NPHard Problems
, 1995
"... We present a unified framework for designing polynomial time approximation schemes (PTASs) for "dense" instances of many NPhard optimization problems, including maximum cut, graph bisection, graph separation, minimum kway cut with and without specified terminals, and maximum 3satisfiabi ..."
Abstract

Cited by 195 (32 self)
 Add to MetaCart
We present a unified framework for designing polynomial time approximation schemes (PTASs) for "dense" instances of many NPhard optimization problems, including maximum cut, graph bisection, graph separation, minimum kway cut with and without specified terminals, and maximum 3satisfiability. By dense graphs we mean graphs with minimum degree &Omega;(n), although our algorithms solve most of these problems so long as the average degree is &Omega;(n). Denseness for nongraph problems is defined similarly. The unified framework begins with the idea of exhaustive sampling: picking a small random set of vertices, guessing where they go on the optimum solution, and then using their placement to determine the placement of everything else. The approach then develops into a PTAS for approximating certain smooth integer programs where the objective function and the constraints are "dense" polynomials of constant degree.
Quick Approximation to Matrices and Applications
, 1998
"... We give algorithms to find the following simply described approximation to a given matrix. Given an m \Theta n matrix A with entries between say1 and 1, and an error parameter ffl between 0 and 1, we find a matrix D (implicitly) which is the sum of O(1=ffl 2 ) simple rank 1 matrices so that the ..."
Abstract

Cited by 151 (7 self)
 Add to MetaCart
We give algorithms to find the following simply described approximation to a given matrix. Given an m \Theta n matrix A with entries between say1 and 1, and an error parameter ffl between 0 and 1, we find a matrix D (implicitly) which is the sum of O(1=ffl 2 ) simple rank 1 matrices so that the sum of entries of any submatrix (among the 2 m+n ) of (A \Gamma D) is at most fflmn in absolute value. Our algorithm takes time dependent only on ffl and the allowed probability of failure (not on m;n). We draw on two lines of research to develop the algorithms: one is built around the fundamental Regularity Lemma of Szemerédi in Graph Theory and the constructive version of Alon, Duke, Leffman, Rödl and Yuster. The second one is from the papers of Arora, Karger and Karpinski, Fernandez de la Vega and most directly Goldwasser, Goldreich and Ron who develop approximation algorithms for a set of graph problems, typical of which is the maximum cut problem. ?From our matrix approximation, the...
Dependent rounding and its applications to approximation algorithms
 JOURNAL OF THE ACM
, 2006
"... We develop a new randomized rounding approach for fractional vectors defined on the edgesets of bipartite graphs. We show various ways of combining this technique with other ideas, leading to improved (approximation) algorithms for various problems. These include: ffl low congestion multipath rout ..."
Abstract

Cited by 60 (7 self)
 Add to MetaCart
We develop a new randomized rounding approach for fractional vectors defined on the edgesets of bipartite graphs. We show various ways of combining this technique with other ideas, leading to improved (approximation) algorithms for various problems. These include: ffl low congestion multipath routing; ffl richer randomgraph models for graphs with a given degreesequence; ffl improved approximation algorithms for: (i) throughputmaximization in broadcast scheduling, (ii) delayminimization in broadcast scheduling, as well as (iii) capacitated vertex cover; and
An Approximation Algorithm for MaxMin Fair Allocation of Indivisible goods
 In Proc. of the ACM Symposium on Theory of Computing (STOC
"... In this paper, we give the first approximation algorithm for the problem of maxmin fair allocation of indivisible goods. An instance of this problem consists of a set of k people and m indivisible goods. Each person has a known linear utility function over the set of goods which might be different ..."
Abstract

Cited by 58 (2 self)
 Add to MetaCart
(Show Context)
In this paper, we give the first approximation algorithm for the problem of maxmin fair allocation of indivisible goods. An instance of this problem consists of a set of k people and m indivisible goods. Each person has a known linear utility function over the set of goods which might be different from the others’. The goal is to distribute the goods among the people and maximize the minimum utility received by them. 1 The approximation ratio of our algorithm is Ω ( √ k log3). As a crucial part of our k algorithm, we design and analyze an iterative method for rounding a fractional matching on a tree which might be of independent interest. We also provide better bounds when we are allowed to exclude a small fraction of the people from the problem.
A Polynomial Time Approximation Scheme for Inferring Evolutionary Trees from Quartet Topologies and Its Application
 SIAM Journal on Computing
, 2000
"... . Inferring evolutionary trees has long been a challenging problem both for biologists and computer scientists. In recent years research has concentrated on the quartet method paradigm for inferring evolutionary trees. Quartet methods proceed by first inferring the evolutionary history for every set ..."
Abstract

Cited by 39 (3 self)
 Add to MetaCart
(Show Context)
. Inferring evolutionary trees has long been a challenging problem both for biologists and computer scientists. In recent years research has concentrated on the quartet method paradigm for inferring evolutionary trees. Quartet methods proceed by first inferring the evolutionary history for every set of four species (resulting in a set Q of inferred quartet topologies) and then recombining these inferred quartet topologies to form an evolutionary tree. This paper presents two results on the quartet method paradigm. The first is a polynomial time approximation scheme (PTAS) for recombining the inferred quartet topologies optimally. This is an important result since, to date, there have been no polynomial time algorithms with performance guarantees for quartet methods. To achieve this result the natural denseness of the set Q is exploited. The second result is a new technique, called quartet cleaning, that detects and corrects errors in the set Q with performance guarantees. This result h...