Counterfactual quantum computation

Counterfactual quantum computation is a method of inferring the result of a computation without actually running a quantum computer otherwise capable of actively performing that computation.

Conceptual origin
Physicists Graeme Mitchison and Richard Jozsa introduced the notion of counterfactual computing as an application of quantum computing, founded on the concepts of counterfactual definiteness, on a re-interpretation of the Elitzur–Vaidman bomb tester thought experiment, and making theoretical use of the phenomenon of interaction-free measurement.

After seeing a talk on counterfactual computation by Jozsa at the Isaac Newton Institute, Keith Bowden of the Theoretical Physics Research Unit at Birkbeck College, University of London published a paper in 1997 describing a digital computer that could be counterfactually interrogated to calculate whether a light beam would fail to pass through a maze as an example of this idea.

More recently the idea of counterfactual quantum communication has been proposed and demonstrated.

Outline of the method
The quantum computer may be physically implemented in arbitrary ways but, to date, the common apparatus considered features a Mach–Zehnder interferometer. The quantum computer is set in a superposition of "not running" and "running" states by means such as the quantum Zeno effect. Those state histories are quantum interfered. After many repetitions of very rapid projective measurements, the "not running" state evolves to a final value imprinted into the properties of the quantum computer. Measuring that value allows for learning the result of some types of computations such as Grover's algorithm even though the result was derived from the non-running state of the quantum computer.

Definition
The original formulation of counterfactual quantum computation stated that a set m of measurement outcomes is a counterfactual outcome if there is only one history associated to m and that history contains only "off" (non-running) states, and there is only a single possible computational output associated to m.

A refined definition of counterfactual computation expressed in procedures and conditions is: (i) Identify and label all histories (quantum paths), with as many labels as needed, which lead to the same set m of measurement outcomes, and (ii) coherently superpose all possible histories. (iii) After cancelling the terms (if any) whose complex amplitudes together add to zero, the set m of measurement outcomes is a counterfactual outcome if (iv) there are no terms left with the computer-running label in their history labels, and (v) there is only a single possible computer output associated to m.

Mirror array
In 1997, after discussions with Abner Shimony and Richard Jozsa, and inspired by the idea of the (1993) Elitzur-Vaidman bomb tester, Keith Bowden (Birkbeck College) published a paper describing a digital computer that could be counterfactually interrogated to calculate whether a photon would fail to pass through a maze of mirrors. This so-called mirror array replaces the tentative bomb in Elitzur and Vaidman's device (actually a Mach–Zehnder interferometer). One time in four a photon will exit the device in such a way as to indicate that the maze is not navigable, even though the photon never passed through the mirror array. The mirror array itself is set up in such a way that it is defined by an n by n matrix of bits. The output (fail or otherwise) is itself defined by a single bit. Thus the mirror array itself is an n-squared bit in, 1 bit out digital computer which calculates mazes and can be run counterfactually. Although the overall device is clearly a quantum computer, the part which is counterfactually tested is semi classical.

Experimental demonstration
In 2015, counterfactual quantum computation was demonstrated in the experimental context of "spins of a negatively charged nitrogen-vacancy color center in a diamond". Previously suspected limits of efficiency were exceeded, achieving counterfactual computational efficiency of 85% with the higher efficiency foreseen in principle.