Aanderaa–Karp–Rosenberg conjecture

In theoretical computer science, the Aanderaa–Karp–Rosenberg conjecture (also known as the Aanderaa–Rosenberg conjecture or the evasiveness conjecture) is a group of related conjectures about the number of questions of the form "Is there an edge between vertex $$u$$ and vertex $$v$$?" that have to be answered to determine whether or not an undirected graph has a particular property such as planarity or bipartiteness. They are named after Stål Aanderaa, Richard M. Karp, and Arnold L. Rosenberg. According to the conjecture, for a wide class of properties, no algorithm can guarantee that it will be able to skip any questions: any algorithm for determining whether the graph has the property, no matter how clever, might need to examine every pair of vertices before it can give its answer. A property satisfying this conjecture is called evasive.

More precisely, the Aanderaa–Rosenberg conjecture states that any deterministic algorithm must test at least a constant fraction of all possible pairs of vertices, in the worst case, to determine any non-trivial monotone graph property. In this context, a property is monotone if it remains true when edges are added; for example, planarity is not monotone, but non-planarity is monotone. A stronger version of this conjecture, called the evasiveness conjecture or the Aanderaa–Karp–Rosenberg conjecture, states that exactly $$\tbinom{n}{2}=n(n-1)/2$$ tests are needed for. Versions of the problem for randomized algorithms and quantum algorithms have also been formulated and studied.

The deterministic Aanderaa–Rosenberg conjecture was proven by, but the stronger Aanderaa–Karp–Rosenberg conjecture remains unproven. Additionally, there is a large gap between the conjectured lower bound and the best proven lower bound for randomized and quantum query complexity.

Example
The property of being non-empty (that is, having at least one edge) is monotone, because adding another edge to a non-empty graph produces another non-empty graph. There is a simple algorithm for testing whether a graph is non-empty: loop through all of the pairs of vertices, testing whether each pair is connected by an edge. If an edge is ever found in this way, break out of the loop, and report that the graph is non-empty, and if the loop completes without finding any edges, then report that the graph is empty. On some graphs (for instance the complete graphs) this algorithm will terminate quickly, without testing every pair of vertices, but on the empty graph it tests all possible pairs before terminating. Therefore, the query complexity of this algorithm is $$\tbinom{n}{2}=n(n-1)/2$$: in the worst case, the algorithm performs $$n(n-1)/2$$ tests.

The algorithm described above is not the only possible method of testing for non-emptiness, but the Aanderaa–Karp–Rosenberg conjecture implies that every deterministic algorithm for testing non-emptiness has the same worst-case query complexity, $$n(n-1)/2$$. That is, the property of being non-empty is evasive. For this property, the result is easy to prove directly: if an algorithm does not perform $$n(n-1)/2$$ tests, it cannot distinguish the empty graph from a graph that has one edge connecting one of the untested pairs of vertices, and must give an incorrect answer on one of these two graphs.

Definitions
In the context of this article, all graphs will be simple and undirected, unless stated otherwise. This means that the edges of the graph form a set (and not a multiset) and each edge is a pair of distinct vertices. Graphs are assumed to have an implicit representation in which each vertex has a unique identifier or label and in which it is possible to test the adjacency of any two vertices, but for which adjacency testing is the only allowed primitive operation.

Informally, a graph property is a property of a graph that is independent of labeling. More formally, a graph property is a mapping from the class of all graphs to $$\{0,1\}$$ such that isomorphic graphs are mapped to the same value. For example, the property of containing at least one vertex of degree two is a graph property, but the property that the first vertex has degree two is not, because it depends on the labeling of the graph (in particular, it depends on which vertex is the "first" vertex). A graph property is called non-trivial if it does not assign the same value to all graphs. For instance, the property of being a graph is a trivial property, since all graphs possess this property. On the other hand, the property of being empty is non-trivial, because the empty graph possesses this property, but non-empty graphs do not. A graph property is said to be monotone if the addition of edges does not destroy the property. Alternately, if a graph possesses a monotone property, then every supergraph of this graph on the same vertex set also possesses it. For instance, the property of being nonplanar is monotone: a supergraph of a nonplanar graph is itself nonplanar. However, the property of being regular is not monotone.

The big O notation is often used for query complexity. In short, $$f(n)$$ is $$O(g(n))$$ (read as "of the order of $$g(n)$$") if there exist positive constants $$c$$ and $$N$$ such that, for all $$n\ge N$$, $$f(n)\le c\cdot g(n)$$. Similarly, $$f(n)$$ is $$\Omega(g(n))$$ if there exist positive constants $$c$$ and $$N$$ such that, for all $$n\ge N$$, $$f(n)\ge c\cdot g(n)$$. Finally, $$f(n)$$ is $$\Theta(g(n))$$ if it is both $$O(g(n))$$ and $$\Omega(g(n))$$.

Query complexity
The deterministic query complexity of evaluating a function on $$n$$ bits (where the bits may be labeled as $$x_1,x_2,\dots x_n$$) is the number of bits $$x_i$$ that have to be read in the worst case by a deterministic algorithm that computes the function. For instance, if the function takes the value 0 when all bits are 0 and takes value 1 otherwise (this is the OR function), then its deterministic query complexity is exactly $$n$$. In the worst case, regardless of the order it chooses to examine its input, the first $$n-1$$ bits read could all be 0, and the value of the function now depends on the last bit. If an algorithm doesn't read this bit, it might output an incorrect answer. (Such arguments are known as adversary arguments.) The number of bits read are also called the number of queries made to the input. One can imagine that the algorithm asks (or queries) the input for a particular bit and the input responds to this query.

The randomized query complexity of evaluating a function is defined similarly, except the algorithm is allowed to be randomized. In other words, it can flip coins and use the outcome of these coin flips to decide which bits to query in which order. However, the randomized algorithm must still output the correct answer for all inputs: it is not allowed to make errors. Such algorithms are called Las Vegas algorithms. (A different class of algorithms, Monte Carlo algorithms, are allowed to make some error.) Randomized query complexity can be defined for both Las Vegas and Monte Carlo algorithms, but the randomized version of the Aanderaa–Karp–Rosenberg conjecture is about the Las Vegas query complexity of graph properties.

Quantum query complexity is the natural generalization of randomized query complexity, of course allowing quantum queries and responses. Quantum query complexity can also be defined with respect to Monte Carlo algorithms or Las Vegas algorithms, but it is usually taken to mean Monte Carlo quantum algorithms.

In the context of this conjecture, the function to be evaluated is the graph property, and the input can be thought of as a string of size $$n(n-1)/2$$, describing for each pair of vertices whether there is an edge with that pair as its endpoints. The query complexity of any function on this input is at most $$n(n-1)/2$$, because an algorithm that makes $$n(n-1)/2$$ queries can read the whole input and determine the input graph completely.

Deterministic query complexity
For deterministic algorithms, originally conjectured that for all nontrivial graph properties on $$n$$ vertices, deciding whether a graph possesses this property requires $$\Omega(n^2)$$ The non-triviality condition is clearly required because there are trivial properties like "is this a graph?" which can be answered with no queries at all.

The conjecture was disproved by Aanderaa, who exhibited a directed graph property (the property of containing a "sink") which required only $$O(n)$$ queries to test. A sink, in a directed graph, is a vertex of indegree $$n-1$$ and outdegree zero. The existence of a sink can be tested with less than $$3n$$ queries. An undirected graph property which can also be tested with $$O(n)$$ queries is the property of being a scorpion graph, first described in. A scorpion graph is a graph containing a three-vertex path, such that one endpoint of the path is connected to all remaining vertices, while the other two path vertices have no incident edges other than the ones in the path.

Then Aanderaa and Rosenberg formulated a new conjecture (the Aanderaa–Rosenberg conjecture) which says that deciding whether a graph possesses a non-trivial monotone graph property requires $$\Omega(n^2)$$ queries. This conjecture was resolved by by showing that at least $$\tfrac{1}{16}n^2$$ queries are needed to test for any nontrivial monotone graph property. The bound was further improved to $$\tfrac{1}{9}n^2$$ by, to $$\bigl(\tfrac{1}{4}-\varepsilon\bigr)n^2$$ (for any $$\varepsilon>0$$) by , to $$\bigl(\tfrac{8}{25}-\varepsilon\bigr)n^2$$ by , and to $$\bigl(\tfrac{1}{3}-\varepsilon\bigr)n^2$$ by.

Richard Karp conjectured the stronger statement (which is now called the evasiveness conjecture or the Aanderaa–Karp–Rosenberg conjecture) that "every nontrivial monotone graph property for graphs on $$n$$ vertices is evasive." A property is called evasive if determining whether a given graph has this property sometimes requires all $$n(n-1)/2$$ possible queries. This conjecture says that the best algorithm for testing any nontrivial monotone property must (in the worst case) query all possible edges. This conjecture is still open, although several special graph properties have shown to be evasive for all $$n$$. The conjecture has been resolved for the case where $$n$$ is a prime power by using a topological approach. The conjecture has also been resolved for all non-trivial monotone properties on bipartite graphs by. Minor-closed properties have also been shown to be evasive for large $$n$$.

In the conjecture was generalized to properties of other (non-graph) functions too, conjecturing that any non-trivial monotone function that is weakly symmetric is evasive. This case is also solved when $$n$$ is a prime power.

Randomized query complexity
Richard Karp also conjectured that $$\Omega(n^2)$$ queries are required for testing nontrivial monotone properties even if randomized algorithms are permitted. No nontrivial monotone property is known which requires less than $$\tfrac{1}{4}n^2$$ queries to test. A linear lower bound (i.e., $$\Omega(n)$$) on all monotone properties follows from a very general relationship between randomized and deterministic query complexities. The first superlinear lower bound for all monotone properties was given by  who showed that $$\Omega\bigl(n(\log n)^{1/12}\bigr)$$ queries are required. This was further improved by to $$\Omega(n^{5/4})$$, and then by  to $\Omega(n^{4/3})$. This was subsequently improved to the current best known lower bound (among bounds that hold for all monotone properties) of $$\Omega\bigl(n^{4/3}(\log n)^{1/3}\bigr)$$ by.

Some recent results give lower bounds which are determined by the critical probability $$p$$ of the monotone graph property under consideration. The critical probability $$p$$ is defined as the unique number $$p$$ in the range $$[0,1]$$ such that a random graph $$G(n,p)$$ (obtained by choosing randomly whether each edge exists, independently of the other edges, with probability $$p$$ per edge) possesses this property with probability equal to $$\tfrac12$$. showed that any monotone property with critical probability $$p$$ requires $$\Omega\left(\min\left\{\frac{n}{\min(p,1-p)},\frac{n^2}{\log n}\right\}\right)$$ queries. For the same problem, showed a lower bound of $$\Omega(n^{4/3}/p^{1/3})$$.

As in the deterministic case, there are many special properties for which an $$\Omega(n^2)$$ lower bound is known. Moreover, better lower bounds are known for several classes of graph properties. For instance, for testing whether the graph has a subgraph isomorphic to any given graph (the so-called subgraph isomorphism problem), the best known lower bound is $$\Omega(n^{3/2})$$ due to.

Quantum query complexity
For bounded-error quantum query complexity, the best known lower bound is $$\Omega\bigl(n^{2/3}(\log n)^{1/6}\bigr)$$ as observed by Andrew Yao. It is obtained by combining the randomized lower bound with the quantum adversary method. The best possible lower bound one could hope to achieve is $$\Omega(n)$$, unlike the classical case, due to Grover's algorithm which gives an $$O(n)$$-query algorithm for testing the monotone property of non-emptiness. Similar to the deterministic and randomized case, there are some properties which are known to have an $$\Omega(n)$$ lower bound, for example non-emptiness (which follows from the optimality of Grover's algorithm) and the property of containing a triangle. There are some graph properties which are known to have an $$\Omega(n^{3/2})$$ lower bound, and even some properties with an $$\Omega(n^2)$$ lower bound. For example, the monotone property of nonplanarity requires $$\Theta(n^{3/2})$$ queries and the monotone property of containing more than half the possible number of edges (also called the majority function) requires $$\Theta(n^2)$$ queries.