Results 1  10
of
435
Infeasibility of instance compression and succinct PCPs for NP
 Electronic Colloquium on Computational Complexity (ECCC
"... The ORSAT problem asks, given Boolean formulae φ1,..., φm each of size at most n, whether at least one of the φi’s is satisfiable. We show that there is no reduction from ORSAT to any set A where the length of the output is bounded by a polynomial in n, unless NP ⊆ coNP/poly, and the PolynomialTi ..."
Abstract

Cited by 66 (1 self)
 Add to MetaCart
(Show Context)
The ORSAT problem asks, given Boolean formulae φ1,..., φm each of size at most n, whether at least one of the φi’s is satisfiable. We show that there is no reduction from ORSAT to any set A where the length of the output is bounded by a polynomial in n, unless NP ⊆ coNP/poly, and the PolynomialTime Hierarchy collapses. This result settles an open problem proposed by Bodlaender et. al. [4] and Harnik and Naor [15] and has a number of implications. • A number of parametric NP problems, including Satisfiability, Clique, Dominating Set and Integer Programming, are not instance compressible or polynomially kernelizable unless NP ⊆ coNP/poly. • Satisfiability does not have PCPs of size polynomial in the number of variables unless NP ⊆ coNP/poly. • An approach of Harnik and Naor to constructing collisionresistant hash functions from oneway functions is unlikely to be viable in its present form. • (BuhrmanHitchcock) There are no subexponentialsize hard sets for NP unless NP is in coNP/poly. We also study probabilistic variants of compression, and show various results about and connections between these variants. To this end, we introduce a new strong derandomization hypothesis, the Oracle Derandomization Hypothesis, and discuss how it relates to traditional derandomization assumptions. Categories and Subject Descriptors
Llull and Copeland voting computationally resist bribery and control
, 2009
"... Control and bribery are settings in which an external agent seeks to influence the outcome of an election. Constructive control of elections refers to attempts by an agent to, via such actions as addition/deletion/partition of candidates or voters, ensure that a given candidate wins. Destructive con ..."
Abstract

Cited by 63 (29 self)
 Add to MetaCart
(Show Context)
Control and bribery are settings in which an external agent seeks to influence the outcome of an election. Constructive control of elections refers to attempts by an agent to, via such actions as addition/deletion/partition of candidates or voters, ensure that a given candidate wins. Destructive control refers to attempts by an agent to, via the same actions, preclude a given candidate’s victory. An election system in which an agent can sometimes affect the result and it can be determined in polynomial time on which inputs the agent can succeed is said to be vulnerable to the given type of control. An election system in which an agent can sometimes affect the result, yet in which it is NPhard to recognize the inputs on which the agent can succeed, is said to be resistant to the given type of control. Aside from election systems with an NPhard winner problem, the only systems previously known to be resistant to all the standard control types were highly artificial election systems created by hybridization. This paper studies a parameterized version of Copeland voting, denoted by Copeland α, where the parameter α is a rational number between 0 and 1 that specifies how ties are valued in the pairwise comparisons of candidates. In every previously studied constructive or destructive
Bidimensionality and Kernels
, 2010
"... Bidimensionality theory appears to be a powerful framework in the development of metaalgorithmic techniques. It was introduced by Demaine et al. [J. ACM 2005] as a tool to obtain subexponential time parameterized algorithms for bidimensional problems on Hminor free graphs. Demaine and Hajiaghayi ..."
Abstract

Cited by 61 (24 self)
 Add to MetaCart
(Show Context)
Bidimensionality theory appears to be a powerful framework in the development of metaalgorithmic techniques. It was introduced by Demaine et al. [J. ACM 2005] as a tool to obtain subexponential time parameterized algorithms for bidimensional problems on Hminor free graphs. Demaine and Hajiaghayi [SODA 2005] extended the theory to obtain polynomial time approximation schemes (PTASs) for bidimensional problems. In this paper, we establish a third metaalgorithmic direction for bidimensionality theory by relating it to the existence of linear kernels for parameterized problems. In parameterized complexity, each problem instance comes with a parameter k and the parameterized problem is said to admit a linear kernel if there is a polynomial time algorithm, called
Satisfiability Allows No Nontrivial Sparsification Unless The PolynomialTime Hierarchy Collapses
 ELECTRONIC COLLOQUIUM ON COMPUTATIONAL COMPLEXITY, REPORT NO. 38 (2010)
, 2010
"... Consider the following twoplayer communication process to decide a language L: The first player holds the entire input x but is polynomially bounded; the second player is computationally unbounded but does not know any part of x; their goal is to cooperatively decide whether x belongs to L at small ..."
Abstract

Cited by 53 (2 self)
 Add to MetaCart
Consider the following twoplayer communication process to decide a language L: The first player holds the entire input x but is polynomially bounded; the second player is computationally unbounded but does not know any part of x; their goal is to cooperatively decide whether x belongs to L at small cost, where the cost measure is the number of bits of communication from the first player to the second player. For any integer d ≥ 3 and positive real ǫ we show that if satisfiability for nvariable dCNF formulas has a protocol of cost O(n d−ǫ) then coNP is in NP/poly, which implies that the polynomialtime hierarchy collapses to its third level. The result even holds when the first player is conondeterministic, and is tight as there exists a trivial protocol for ǫ = 0. Under the hypothesis that coNP is not in NP/poly, our result implies tight lower bounds for parameters of interest in several areas, namely sparsification, kernelization in parameterized complexity, lossy compression, and probabilistically checkable proofs. By reduction, similar results hold for other NPcomplete problems. For the vertex cover problem on nvertex duniform hypergraphs, the above statement holds for any integer d ≥ 2. The case d = 2 implies that no NPhard vertex deletion problem based on a graph property that is inherited by subgraphs can have kernels consisting of O(k 2−ǫ) edges unless coNP is in NP/poly, where k denotes the size of the deletion set. Kernels consisting of O(k 2) edges are known for several problems in the class, including vertex cover, feedback vertex set, and boundeddegree deletion.
Combinatorial Optimization on Graphs of Bounded Treewidth
, 2007
"... There are many graph problems that can be solved in linear or polynomial time with a dynamic programming algorithm when the input graph has bounded treewidth. For combinatorial optimization problems, this is a useful approach for obtaining fixedparameter tractable algorithms. Starting from trees an ..."
Abstract

Cited by 51 (4 self)
 Add to MetaCart
There are many graph problems that can be solved in linear or polynomial time with a dynamic programming algorithm when the input graph has bounded treewidth. For combinatorial optimization problems, this is a useful approach for obtaining fixedparameter tractable algorithms. Starting from trees and seriesparallel graphs, we introduce the concepts of treewidth and tree decompositions, and illustrate the technique with the Weighted Independent Set problem as an example. The paper surveys some of the latest developments, putting an emphasis on applicability, on algorithms that exploit tree decompositions, and on algorithms that determine or approximate treewidth and find tree decompositions with optimal or close to optimal treewidth. Directions for further research and suggestions for further reading are also given.
CompressionBased FixedParameter Algorithms for Feedback Vertex Set and Edge Bipartization
, 2006
"... We show that the NPcomplete Feedback Vertex Set problem, which asks for the smallest set of vertices to remove from a graph to destroy all cycles, is deterministically solvable in O(c k ·m) time. Here, m denotes the number of graph edges, k denotes the size of the feedback vertex set searched for, ..."
Abstract

Cited by 47 (4 self)
 Add to MetaCart
We show that the NPcomplete Feedback Vertex Set problem, which asks for the smallest set of vertices to remove from a graph to destroy all cycles, is deterministically solvable in O(c k ·m) time. Here, m denotes the number of graph edges, k denotes the size of the feedback vertex set searched for, and c is a constant. We extend this to an algorithm enumerating all solutions in O(d k ·m) time for a (larger) constant d. As a further result, we present a fixedparameter algorithm with runtime O(2 k · m 2) for the NPcomplete Edge Bipartization problem, which asks for at most k edges to remove from a graph to make it bipartite.
Incompressibility through Colors and IDs
"... In parameterized complexity each problem instance comes with a parameter k and the parameterized problem is said to admit a polynomial kernel if there are polynomial time preprocessing rules that reduce the input instance down to an instance with size polynomial in k. Many problems have been shown t ..."
Abstract

Cited by 45 (5 self)
 Add to MetaCart
(Show Context)
In parameterized complexity each problem instance comes with a parameter k and the parameterized problem is said to admit a polynomial kernel if there are polynomial time preprocessing rules that reduce the input instance down to an instance with size polynomial in k. Many problems have been shown to admit polynomial kernels, but it is only recently that a framework for showing the nonexistence of polynomial kernels for specific problems has been developed by Bodlaender et al. [6] and Fortnow and Santhanam [15]. With few exceptions, all known kernelization lower bounds result have been obtained by directly applying this framework. In this paper we show how to combine these results with combinatorial reductions which use colors and IDs in order to prove kernelization lower bounds for a variety of basic problems. Below we give a summary of our main results. All our results are under the assumption that the polynomial hierarchy does not collapse to the third level. • We show that the Steiner Tree problem parameterized by the number of terminals and solution size, and the Connected Vertex Cover and Capacitated Vertex Cover problems do not admit a polynomial kernel. The two latter results are surprising because the closely related Vertex Cover problem admits a kernel of size 2k.
Almost 2SAT is fixedparameter tractable
 Journal of Computer and System Sciences
"... Abstract. We consider the following problem. Given a 2CNF formula, is it possible to remove at most k clauses so that the resulting 2CNF formula is satisfiable? This problem is known to different research communities in Theoretical Computer Science under the names ’Almost 2SAT’, ’Allbutk 2SAT’ ..."
Abstract

Cited by 43 (6 self)
 Add to MetaCart
(Show Context)
Abstract. We consider the following problem. Given a 2CNF formula, is it possible to remove at most k clauses so that the resulting 2CNF formula is satisfiable? This problem is known to different research communities in Theoretical Computer Science under the names ’Almost 2SAT’, ’Allbutk 2SAT’, ’2CNF deletion’, ’2SAT deletion’. The status of fixedparameter tractability of this problem is a longstanding open question in the area of Parameterized Complexity. We resolve this open question by proposing an algorithm which solves this problem in O(15 k ∗ k ∗ m 3) and thus we show that this problem is fixedparameter tractable. 1
CrossComposition: A New Technique for Kernelization Lower Bounds
, 2011
"... We introduce a new technique for proving kernelization lower bounds, called crosscomposition. A classical problem L crosscomposes into a parameterized problem Q if an instance of Q with polynomially bounded parameter value can express the logical OR of a sequence of instances of L. Building on wor ..."
Abstract

Cited by 40 (7 self)
 Add to MetaCart
We introduce a new technique for proving kernelization lower bounds, called crosscomposition. A classical problem L crosscomposes into a parameterized problem Q if an instance of Q with polynomially bounded parameter value can express the logical OR of a sequence of instances of L. Building on work by Bodlaender et al. (ICALP 2008) and using a result by Fortnow and Santhanam (STOC 2008) we show that if an NPhard problem crosscomposes into a parameterized problem Q then Q does not admit a polynomial kernel unless the polynomial hierarchy collapses. Our technique generalizes and strengthens the recent techniques of using orcomposition algorithms and of transferring the lower bounds via polynomial parameter transformations. We show its applicability by proving kernelization lower bounds for a number of important graphs problems with structural (nonstandard) parameterizations, e.g., Chromatic Number, Clique, and Weighted Feedback Vertex Set do not admit polynomial kernels with respect to the vertex cover number of the input graphs unless the polynomial hierarchy collapses, contrasting the fact that these problems are trivially fixedparameter tractable for this parameter. We have similar lower bounds for Feedback Vertex Set.
A more effective linear kernelization for Cluster Editing
 Theor. Comput. Sci
, 2009
"... Abstract. In the NPhard Cluster Editing problem, we have as input an undirected graph G andanintegerk ≥ 0. The question is whether we can transform G, by inserting and deleting at most k edges, into a cluster graph, that is, a union of disjoint cliques. We first confirm a conjecture by Michael Fell ..."
Abstract

Cited by 38 (10 self)
 Add to MetaCart
(Show Context)
Abstract. In the NPhard Cluster Editing problem, we have as input an undirected graph G andanintegerk ≥ 0. The question is whether we can transform G, by inserting and deleting at most k edges, into a cluster graph, that is, a union of disjoint cliques. We first confirm a conjecture by Michael Fellows [IWPEC 2006] that there is a polynomialtime kernelization for Cluster Editing that leads to a problem kernel with at most 6k vertices. More precisely, we present a cubictime algorithm that, given a graph G andanintegerk ≥ 0, finds a graph G ′ and an integer k ′ ≤ k such that G can be transformed into a cluster graph by at most k edge modifications iff G ′ can be transformed into a cluster graph by at most k ′ edge modifications, and the problem kernel G ′ has at most 6k vertices. So far, only a problem kernel of 24k vertices was known. Second, we show that this bound for the number of vertices of G ′ can be further improved to 4k. Finally, we consider the variant of Cluster Editing where the number of cliques that the cluster graph can contain is stipulated to be a constant d>0. We present a simple kernelization for this variant leaving a problem kernel of at most (d +2)k + d vertices. 1