Results 1  10
of
239
Exact algorithms for NPhard problems: a survey, in: Combinatorial Optimization—Eureka, you shrink!”,
 Lecture Notes in Computer Science,
, 2003
"... Abstract. We discuss fast exponential time solutions for NPcomplete problems. We survey known results and approaches, we provide pointers to the literature, and we discuss several open problems in this area. The list of discussed NPcomplete problems includes the travelling salesman problem, sched ..."
Abstract

Cited by 151 (4 self)
 Add to MetaCart
(Show Context)
Abstract. We discuss fast exponential time solutions for NPcomplete problems. We survey known results and approaches, we provide pointers to the literature, and we discuss several open problems in this area. The list of discussed NPcomplete problems includes the travelling salesman problem, scheduling under precedence constraints, satisfiability, knapsack, graph coloring, independent sets in graphs, bandwidth of a graph, and many more.
On the Complexity of kSAT
, 2001
"... The kSAT problem is to determine if a given kCNF has a satisfying assignment. It is a celebrated open question as to whether it requires exponential time to solve kSAT for k 3. Here exponential time means 2 $n for some $>0. In this paper, assuming that, for k 3, kSAT requires exponential time ..."
Abstract

Cited by 104 (8 self)
 Add to MetaCart
The kSAT problem is to determine if a given kCNF has a satisfying assignment. It is a celebrated open question as to whether it requires exponential time to solve kSAT for k 3. Here exponential time means 2 $n for some $>0. In this paper, assuming that, for k 3, kSAT requires exponential time complexity, we show that the complexity of kSAT increases as k increases. More precisely, for k 3, define s k=inf[$: there exists 2 $n algorithm for solving kSAT]. Define ETH (ExponentialTime Hypothesis) for kSAT as follows: for k 3, s k>0. In this paper, we show that s k is increasing infinitely often assuming ETH for kSAT. Let s be the limit of s k. We will in fact show that s k (1&d k) s for some constant d>0. We prove this result by bringing together the ideas of critical clauses and the Sparsification Lemma to reduce the satisfiability of a kCNF to the satisfiability of a disjunction of 2 =n k$CNFs in fewer variables for some k $ k and arbitrarily small =>0. We also show that such a disjunction can be computed in time 2 =n for arbitrarily small =>0.
Subexponential parameterized algorithms on graphs of boundedgenus and Hminorfree Graphs
"... ... Building on these results, we develop subexponential fixedparameter algorithms for dominating set, vertex cover, and set cover in any class of graphs excluding a fixed graph H as a minor. Inparticular, this general category of graphs includes planar graphs, boundedgenus graphs, singlecrossing ..."
Abstract

Cited by 63 (22 self)
 Add to MetaCart
... Building on these results, we develop subexponential fixedparameter algorithms for dominating set, vertex cover, and set cover in any class of graphs excluding a fixed graph H as a minor. Inparticular, this general category of graphs includes planar graphs, boundedgenus graphs, singlecrossingminorfree graphs, and anyclass of graphs that is closed under taking minors. Specifically, the running time is 2O(pk)nh, where h is a constant depending onlyon H, which is polynomial for k = O(log² n). We introducea general approach for developing algorithms on Hminorfreegraphs, based on structural results about Hminorfree graphs at the
Tight lower bounds for certain parameterized NPhard problems
 Proc. 19th Annual IEEE Conference on Computational Complexity (CCC’04
, 2004
"... Based on the framework of parameterized complexity theory, we derive tight lower bounds on the computational complexity for a number of wellknown NPhard problems. We start by proving a general result, namely that the parameterized weighted satisfiability problem on deptht circuits cannot be solve ..."
Abstract

Cited by 62 (10 self)
 Add to MetaCart
(Show Context)
Based on the framework of parameterized complexity theory, we derive tight lower bounds on the computational complexity for a number of wellknown NPhard problems. We start by proving a general result, namely that the parameterized weighted satisfiability problem on deptht circuits cannot be solved in time no(k) poly(m), where n is the circuit input length, m is the circuit size, and k is the parameter, unless the (t − 1)st level W [t − 1] of the Whierarchy collapses to FPT. By refining this technique, we prove that a group of parameterized NPhard problems, including weighted sat, dominating set, hitting set, set cover, and feature set, cannot be solved in time no(k) poly(m), where n is the size of the universal set from which the k elements are to be selected and m is the instance size, unless the first level W [1] of the Whierarchy collapses to FPT. We also prove that another group of parameterized problems which includes weighted qsat (for any fixed q ≥ 2), clique, and independent set, cannot be solved in time no(k) unless all search problems in the syntactic class SNP, introduced by Papadimitriou and Yannakakis, are solvable in subexponential time. Note that all these parameterized problems have trivial algorithms of running time either n k poly(m) or O(n k). 1
Satisfiability Allows No Nontrivial Sparsification Unless The PolynomialTime Hierarchy Collapses
 ELECTRONIC COLLOQUIUM ON COMPUTATIONAL COMPLEXITY, REPORT NO. 38 (2010)
, 2010
"... Consider the following twoplayer communication process to decide a language L: The first player holds the entire input x but is polynomially bounded; the second player is computationally unbounded but does not know any part of x; their goal is to cooperatively decide whether x belongs to L at small ..."
Abstract

Cited by 56 (2 self)
 Add to MetaCart
(Show Context)
Consider the following twoplayer communication process to decide a language L: The first player holds the entire input x but is polynomially bounded; the second player is computationally unbounded but does not know any part of x; their goal is to cooperatively decide whether x belongs to L at small cost, where the cost measure is the number of bits of communication from the first player to the second player. For any integer d ≥ 3 and positive real ǫ we show that if satisfiability for nvariable dCNF formulas has a protocol of cost O(n d−ǫ) then coNP is in NP/poly, which implies that the polynomialtime hierarchy collapses to its third level. The result even holds when the first player is conondeterministic, and is tight as there exists a trivial protocol for ǫ = 0. Under the hypothesis that coNP is not in NP/poly, our result implies tight lower bounds for parameters of interest in several areas, namely sparsification, kernelization in parameterized complexity, lossy compression, and probabilistically checkable proofs. By reduction, similar results hold for other NPcomplete problems. For the vertex cover problem on nvertex duniform hypergraphs, the above statement holds for any integer d ≥ 2. The case d = 2 implies that no NPhard vertex deletion problem based on a graph property that is inherited by subgraphs can have kernels consisting of O(k 2−ǫ) edges unless coNP is in NP/poly, where k denotes the size of the deletion set. Kernels consisting of O(k 2) edges are known for several problems in the class, including vertex cover, feedback vertex set, and boundeddegree deletion.
Subexponential Parameterized Algorithms Collapse the Whierarchy (Extended Abstract)
, 2001
"... It is shown that for essentially all MAX SNPhard optimization problems finding exact solutions in subexponential time is not possible unless W [1] = FPT . In particular, we show that O(2 o(k) p(n)) parameterized algorithms do not exist for Vertex Cover, Max Cut, Max cSat, and a number of pr ..."
Abstract

Cited by 54 (3 self)
 Add to MetaCart
It is shown that for essentially all MAX SNPhard optimization problems finding exact solutions in subexponential time is not possible unless W [1] = FPT . In particular, we show that O(2 o(k) p(n)) parameterized algorithms do not exist for Vertex Cover, Max Cut, Max cSat, and a number of problems on bounded degree graphs such as Dominating Set and Independent Set, unless W [1] = FPT . Our results are derived via an approach that uses an extended parameterization of optimization problems and associated techniques to relate the parameterized complexity of problems in FPT to the parameterized complexity of extended versions that are W [1]hard.
Constraint solving via fractional edge covers
 In Proceedings of the of the 17th Annual ACMSIAM Symposium on Discrete Algorithms
, 2006
"... Many important combinatorial problems can be modelled as constraint satisfaction problems, hence identifying polynomialtime solvable classes of constraint satisfaction problems received a lot of attention. In this paper, we are interested in structural properties that can make the problem tractable ..."
Abstract

Cited by 51 (9 self)
 Add to MetaCart
Many important combinatorial problems can be modelled as constraint satisfaction problems, hence identifying polynomialtime solvable classes of constraint satisfaction problems received a lot of attention. In this paper, we are interested in structural properties that can make the problem tractable. So far, the largest structural class that is known to be polynomialtime solvable is the class of bounded hypertree width instances introduced by Gottlob et al. [20]. Here we identify a new class of polynomialtime solvable instances: those having bounded fractional edge cover number. Combining hypertree width and fractional edge cover number, we then introduce the notion of fractional hypertree width. We prove that constraint satisfaction problems with bounded fractional hypertree width can be solved in polynomial time (provided that a the tree decomposition is given in the input). We also prove that certain parameterized constraint satisfaction, homomorphism, and embedding problems are fixedparameter tractable on instances having bounded fractional hypertree width. 1.
A measure & conquer approach for the analysis of exact algorithms
, 2007
"... For more than 40 years Branch & Reduce exponentialtime backtracking algorithms have been among the most common tools used for finding exact solutions of NPhard problems. Despite that, the way to analyze such recursive algorithms is still far from producing tight worstcase running time bounds. ..."
Abstract

Cited by 49 (11 self)
 Add to MetaCart
For more than 40 years Branch & Reduce exponentialtime backtracking algorithms have been among the most common tools used for finding exact solutions of NPhard problems. Despite that, the way to analyze such recursive algorithms is still far from producing tight worstcase running time bounds. Motivated by this we use an approach, that we call “Measure & Conquer”, as an attempt to step beyond such limitations. The approach is based on the careful design of a nonstandard measure of the subproblem size; this measure is then used to lower bound the progress made by the algorithm at each branching step. The idea is that a smarter measure may capture behaviors of the algorithm that a standard measure might not be able to exploit, and hence lead to a significantly better worstcase time analysis. In order to show the potentialities of Measure & Conquer, we consider two wellstudied NPhard problems: minimum dominating set and maximum independent set. For the first problem, we consider the current best algorithm, and prove (thanks to a better measure) a much tighter running time bound for it. For the second problem, we describe a new, simple algorithm, and show that its running time is competitive with the current best time bounds, achieved with far more complicated algorithms (and standard analysis). Our examples
Determinant sums for undirected hamiltonicity
 in Prof. of FOCS’10, 2010
"... We present a Monte Carlo algorithm for Hamiltonicity detection in an nvertex undirected graph running in O ∗ (1.657 n) time. To the best of our knowledge, this is the first superpolynomial improvement on the worst case runtime for the problem since the O ∗ (2 n) bound established for TSP almost fif ..."
Abstract

Cited by 46 (1 self)
 Add to MetaCart
(Show Context)
We present a Monte Carlo algorithm for Hamiltonicity detection in an nvertex undirected graph running in O ∗ (1.657 n) time. To the best of our knowledge, this is the first superpolynomial improvement on the worst case runtime for the problem since the O ∗ (2 n) bound established for TSP almost fifty years ago (Bellman 1962, Held and Karp 1962). It answers in part the first open problem in Woeginger’s 2003 survey on exact algorithms for NPhard problems. For bipartite graphs, we improve the bound to O ∗ (1.414 n) time. Both the bipartite and the general algorithm can be implemented to use space polynomial in n. We combine several recently resurrected ideas to get the results. Our main technical contribution is a new reduction inspired by the algebraic sieving method for kPath (Koutis ICALP 2008, Williams IPL 2009). We introduce the Labeled Cycle Cover Sum in which weareset tocount weightedarclabeled cycle coversoverafinite field ofcharacteristic two. We reduce Hamiltonicity to Labeled Cycle Cover Sum and apply the determinant summation technique for Exact Set Covers (Björklund STACS 2010) to evaluate it. 1
The complexity of constructing pseudorandom generators from hard functions
 COMPUTATIONAL COMPLEXITY
, 2004
"... We study the complexity of constructing pseudorandom generators (PRGs) from hard functions, focussing on constantdepth circuits. We show that, starting from a function f: {0, 1} l → {0, 1} computable in alternating time O(l) with O(1) alternations that is hard on average (i.e. there is a constant ..."
Abstract

Cited by 42 (9 self)
 Add to MetaCart
We study the complexity of constructing pseudorandom generators (PRGs) from hard functions, focussing on constantdepth circuits. We show that, starting from a function f: {0, 1} l → {0, 1} computable in alternating time O(l) with O(1) alternations that is hard on average (i.e. there is a constant ɛ> 0 such that every circuit of size 2 ɛl fails to compute f on at least a 1/poly(l) fraction of inputs) we can construct a PRG: {0, 1} O(log n) → {0, 1} n computable by DLOGTIMEuniform constantdepth circuits of size polynomial in n. Such a PRG implies BP · AC 0 = AC 0 under DLOGTIMEuniformity. On the negative side, we prove that starting from a worstcase hard function f: {0, 1} l → {0, 1} (i.e. there is a constant ɛ> 0 such that every circuit of size 2 ɛl fails to compute f on some input) for every positive constant δ < 1 there is no blackbox construction of a PRG: {0, 1} δn → {0, 1} n computable by constantdepth circuits of size polynomial in n. We also study worstcase hardness amplification, which is the related problem of producing an averagecase hard function starting from a worstcase hard one. In particular, we deduce that there is no blackbox worstcase hardness amplification within the polynomial time hierarchy. These negative results are obtained by showing that polynomialsize constantdepth circuits cannot compute good extractors and listdecodable codes.