Abstract
Recent advances in quantum technology have led to the development and manufacturing of experimental programmable quantum annealing optimizers that contain hundreds of quantum bits. These optimizers, commonly referred to as ‘DWave’ chips, promise to solve practical optimization problems potentially faster than conventional ‘classical’ computers. Attempts to quantify the quantum nature of these chips have been met with both excitement and skepticism but have also brought up numerous fundamental questions pertaining to the distinguishability of experimental quantum annealers from their classical thermal counterparts. Inspired by recent results in spinglass theory that recognize ‘temperature chaos’ as the underlying mechanism responsible for the computational intractability of hard optimization problems, we devise a general method to quantify the performance of quantum annealers on optimization problems suffering from varying degrees of temperature chaos: A superior performance of quantum annealers over classical algorithms on these may allude to the role that quantum effects play in providing speedup. We utilize our method to experimentally study the DWave Two chip on different temperaturechaotic problems and find, surprisingly, that its performance scales unfavorably as compared to several analogous classical algorithms. We detect, quantify and discuss several purely classical effects that possibly mask the quantum behavior of the chip.
Introduction
Interest in quantum computing originates in the potential of quantum computers to solve certain computational problems much faster than is possible classically, due to the unique properties of Quantum Mechanics^{1,2}. The implications of having at our disposal reliable quantum computing devices are obviously tremendous. The actual implementation of quantum computing devices is however hindered by many challenging difficulties, the most prominent being the control or removal of quantum decoherence^{3}. In the past few years, quantum technology has matured to the point where limited, taskspecific, nonuniversal quantum devices such as quantum communication systems, quantum random number generators and quantum simulators, are being built, possessing capabilities that exceed those of corresponding classical computers.
Recently, a programmable quantum annealing machine, known as the DWave chip^{4}, has been built whose goal is to minimize the cost functions of classicallyhard optimization problems presumably by adiabatically quenching quantum fluctuations. If found useful, the chip could be regarded as a prototype for generalpurpose quantum optimizers, due to the broad range of hard theoretical and practical problems that may be encoded on it.
The capabilities, performance and underlying physical mechanism driving the DWave chip have generated fair amounts of curiosity, interest, debate and controversy within the Quantum Computing community and beyond, as to the true nature of the device and its potential to exhibit clear “quantum signatures”. While some studies have concluded that the behavior of the chip is consistent with quantum opensystem Lindbladian dynamics^{5}, or indirectly observed entanglement^{6,7}, other studies contesting these^{8,9}, pointed to the existence of simple, purely classical, models capable of exhibiting the main characteristics of the chip.
Nonetheless, the debate around the quantum nature of the chip has raised several fundamental questions pertaining to the manner in which quantum devices should be characterized in the absence of clear practical “signatures” such as (quantum) speedups^{10,11,12}. Since quantum annealers are meant to utilize an altogether different mechanism for solving optimization problems than traditional classical devices, methods for quantifying this difference are expected to serve as important theoretical tools while also having vast practical implications.
Here, we propose a method that partly solves the above question by providing a technique to characterize and quantitatively measure how detrimental classical effects are to the performance of quantum annealers. This is done by studying the algorithmic performance of quantum annealers on sets of optimization problems possessing quantifiable, varying degrees of “thermal” or “classical” hardness, which we also define for this purpose. To illustrate the potential of the proposed technique, we apply it to the experimental quantum annealing optimizer, the DWave Two (DW2) chip.
We observe several distinctive phenomena that reveal a strong correlation between the performance of the chip and classical hardness: i) The DWave chip’s typical timetosolution (t_{s}) as a function of classical hardness scales differently, in fact worse, than that of thermal classical algorithms. Specifically, we find that the chip does very poorly on problem instances exhibiting a phenomenon known as “temperature chaos”. ii) Fluctuations in success probability between programming cycles become larger with increasing hardness, pointing to the fact that encoding errors become more pronounced and destructive with increasing hardness. iii) The success probabilities obtained from harder instances are affected more than easy instances by changes in duration of the anneals.
Analyzing the above findings, we identify two major probable causes for the chip’s observed “subclassical” performance, namely i) that its temperature may not be low enough and ii) that encoding errors become more pronounced with increasing hardness. We further offer experiments and simulations designed to detect and subsequently rectify these so as to enhance the capabilities of the chip.
Classical Hardness, temperature chaos and parallel tempering
In order to study the manner in which the performance of quantum annealers correlates with ‘classical hardness’, it is important to first accurately establish the meaning of classical hardness. For that purpose, we refer to spinglass theory^{13}, which deals with spin glasses—disordered, frustrated spin systems that may be viewed as prototypical classicallyhard (also called NPhard) optimization problems, that are so challenging that specialized hardware has been built to simulate them^{14,15,16}.
Currently, the (classical) method of choice to study general spinglass problems is Parallel Tempering (PT, also known as ‘exchange Monte Carlo’)^{17,18}. PT is a refinement of the celebrated yet somewhat outdated Simulated Annealing (SA) algorithm^{19}, that finds optimal assignments (i.e., the groundstate configurations) of given discretevariable cost functions. It is therefore only natural to make use of the performance of PT to characterize and quantify classical hardness.
In PT simulations, one considers N_{T} copies of an Nspin system at temperatures , where each copy undergoes Metropolis spinflip updates independently of other copies. In addition, copies with neighboring temperatures regularly attempt to swap their temperatures with probabilities that satisfy detailed balance^{20}. In this way, each copy performs a temperature randomwalk (see inset of Fig 1). At high temperatures, freeenergy barriers are easily overcome, allowing for a global exploration of configuration space. At lower temperatures on the other hand, the local minima are explored in more detail. A ‘healthy’ PT simulation requires an unimpeded temperature flow: The total length of the simulation should be longer than the temperature ‘mixing time’ τ^{21,22} The time τ may be thought of as the average time it takes each copy to fully traverse the temperature mesh, indicating equilibration of the simulation. Therefore, instances with large τ are harder to equilibrate, which motivates our definition of the mixing time τ as the classical hardness of a given instance.
Despite the popularity of PT, it has also become apparent that not all the spinglass problems can be efficiently solved by the algorithm^{22,23}. The reason is a phenomenon that has become known as Temperature Chaos (TC). TC^{23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38} consists of a sequence of firstorder phase transitions that a given spinglass instance experiences upon lowering its temperature, whereby the dominant configurations minimizing the free energy above the critical temperatures are vastly different than those below them (Fig. 2 depicts such a phase transition that is ‘rounded’ due to the finite size of the system). A given instance may experience zero, one or more transitions at random temperatures, making the study of TC excruciatingly difficult^{23,36,37,38}. Such TC transitions hinder the PT temperature flow, significantly prolonging the mixing time τ. In practice^{23}, it is found that for small systems the large majority of the instances do not suffer any TC transitions and are ‘easy’ (i.e., they are characterized by short mixing times). However, for a minor fraction of them, τ turns out to be inordinately large. Moreover, the larger the system is, the larger the fraction of longτ samples becomes. In the large N limit, these are the shortτ samples that become exponentially rare in N^{23,34}. This provides further motivation for studying TC instances of optimization problems on moderately small experimental devices (even if these problems are rare).
Temperature chaos and quantum annealers
With the advent of quantum annealers, which presumably offer nonthermal mechanisms for finding ground states, it has become only natural to ask whether quantum annealers can be used to solve ‘TCridden’ optimization problems faster than classical techniques such as PT. In this context, the question of how the performance of quantum annealers depends on the ‘classical hardness’ becomes of fundamental interest: If indeed quantum annealers exploit quantum phenomena such as tunneling to traverse energy barriers, one may hope that they will not be as sensitive to the thermal hardness (as defined above) of the optimization problems they solve. As we shall see next, having a practical definition for classical hardness allows us to address the above questions directly.
To illustrate this, in what follows we apply the ideas introduced above to the DW2 quantum annealing optimizer. We accomplish this by first generating an ensemble of instances that are directly embeddable on the DW2 ‘Chimera’ architecture [the reader is referred to the Supplementary Information (SI) for a detailed description of the Chimera lattice and the Dwave chip and its properties]. The chip on which we perform our study is an array of 512 superconducting flux qubits of which only 476 are functional, operating at a temperature of ~15 mK. The DW2 chip is designed to solve a very specific type of problems, namely, Isingtype optimization problems, by adiabatically transitioning the system Hamiltonian from an initial transversefield Hamiltonian to a final classical programmable cost function of a typical spin glass. The latter is given by the Ising Hamiltonian:
The Ising spins, are the variables to be optimized over and the sets and are programmable parameters of the cost function. Here, denotes a sum over all the active edges of the Chimera graph. For simplicity, we conduct our study on randomlygenerated problem instances with h_{i} = 0 and random, equiprobable J_{ij} = ±J couplings (in our energy units J = 1).
Initially, it is not clear whether the task of finding thermallyhard instances on the Chimera is feasible. While on the one hand TC has been observed in spinglasses on the square lattice^{39}, which has the same spatial dimension, D = 2, as the Chimera^{40}, it has also been found that typical Chimeraembeddable instances are easy to solve ^{11,12,40}. As discussed above, system size plays a significant role in this context, as an N~512spin Chimera may simply be too small to have instances exhibiting TC (for instance, on the square lattice one needs to reach spins for TC to be the rule rather than the exception^{39}). Taking a bruteforce approach to resolve this issue, we generated ~80,000 random problem instances (each characterized by a different set of , analyzing each one by running them on a stateoftheart PT algorithm until equilibration is reached (see the SI). This allowed for the calculation of the instances’ classical hardness, namely their temperature mixing times τ (for more details, see Methods, below). The resulting distribution of τ over the instances is shown in Fig. 1. While most instances equilibrate rather quickly (after some 10^{4} Monte Carlo steps), we find that the distribution has a ‘tail’ of hard samples with τ > 10^{6} revealing that hard instances, although rare, do exist (we estimate that 2 samples in 10^{4} have τ > 10^{7}).
To study the DW2 chip, we grouped together instances with similar classical hardness, i.e., similar mixing times, for k = 3, 4, 5, 6 and 7. For each such ‘generation’ of τ, we randomly picked 100 representative instances for running on the chip (only 14 instances with k = 7 were found). As a convergence test of PT on the selected instances, we verified that the groundstate energies reached by PT are the true ones by means of an exact solver.
At the purely classical level, we found, as anticipated^{23}, that classically hard instances differ from easy instances from a thermodynamic point of view as well. Specifically, large τ instances were found to exhibit sharp changes in the average energy at random critical temperatures, consistently with the occurrence of TC (see Fig. 2). For such instances, the true ground states are present during the simulations only below the TC critical temperatures. As the inset shows, the larger τ is, the lower these critical temperatures typically are. Furthermore, classicallyhard instances were found to differ from easier ones in terms of their energy landscape: While for easy instances minimallyexcited states typically reside only a few spin flips away from ground state configurations, for classically hard instances, this is not the case (see the SI).
Effects of temperature chaos on the “DWave Two” chip
Having sorted and analyzed the randomlygenerated instances, we turned to experimentally test the performance of the DWave chip on these (for details see Methods and SI). Our experiments consisted of programming the chip to solve each of the 414 instances over a dense mesh of annealing times in the available range of . The number of attempts, or anneals, that each instance was run for each choice of t_{a} ranged between 10^{5} and 10^{8}. By calculating the success probability of the annealer on each instance and annealing time, a typical timetosolution t_{s} was obtained for each hardnessgroup, or ‘τgeneration’ (see Methods). Interestingly, we found that for easy samples (τ = 10^{3}) the success probability depends only marginally on t_{a}, pointing to the annealer reaching its asymptotic performance on these. As instances become harder, the sensitivity of success probability to t_{a} increases significantly. Nonetheless, for all hardness groups, the typical t_{s} is found to be shortest at the minimallyallowed annealing time of t_{a} = 20 μs (see the SI for a more detailed discussion).
The main results of our investigation are summarized in Fig. 3 depicting the typical time to solution t_{s} of the DW2 chip (averaged over instances of same hardness groups, see Methods) as a function of classical hardness, or ‘τgeneration’. As is clear from the figure, the performance of the chip was found to correlate strongly with the ‘thermal hardness’ parameter, indicating the significant role thermal hardness plays in the annealing process. Interestingly, the response of the chip was found to be affected by thermal hardness even more than PT, i.e., more strongly than the classical thermal response: While for PT the timetosolution scales as t_{s} ~ τ, the scaling of the DWave chip was found to scale as , with . This scaling is rather surprising given that for quantum annealers to perform better than classical ones, one would expect these to be less susceptible to thermal hardness, not more. Nonetheless, it is clear that the notion of classical hardness is very relevant to the Dwave chip.
To complete the picture, we have also tested our instances with two other classical algorithms. The first is Simulated Annealing (SA) which has recently become a popular benchmarking algorithm against which DWave chips are tested^{10,11,12} (see Methods) and the second is the Hamzede FreitasSelby (HFS) algorithm^{41,42} which is the fastest classical algorithm to date for Chimeratype instances. As expected, SA, which is a thermal method, is found to be sensitive to TC, even more than PT, yielding with . Even though the estimate of presumably slightly depends on our choice of success metric and the details of the algorithm, its susceptibility to classical hardness is nonetheless very evident.
The HFS algorithm is a ‘nonthermal’ algorithm (i.e., it does not make use of a temperature parameter). Nonetheless, here too we have found the concept of classical hardness to be very relevant. For the HFS algorithm, we find a scaling of with , implying that the algorithm is significantly less susceptible to thermal hardness than PT (it is worth pointing out that the typical runtime for the HFS algorithm on the hardest, τ = 10^{7}, group problems was found to be ~0.5 s on an Intel Xeon CPU E5462 @ 2.80 GHz, which to our knowledge makes these the hardest known Chimeratype instances to date).
Analysis of findings
The above somewhat lessthanfavorable performance of the experimental Dwave chip on thermallyhard problems is not necessarily a manifestation of the intrinsic limitations of quantum annealing, i.e., it does not necessarily imply that the ‘quantum landscape’ of the tested problems is harder to traverse than the classical one (although this may sometimes be the case^{43,44}). A careful analysis of the results suggests in fact at least two different more probable ‘classical’ causes for the performance of the chip.
First, as already discussed above and is succinctly captured in Fig. 2, temperature is expected to play a key role in DW2 success on instances exhibiting TC. This is because for these, the ground state configurations minimize the free energy only below the lowest critical TC temperature. Even though the working temperature of the DW2 chip is rather low, namely ~15 mK, the crucial figure of merit is the ratio of coupling to temperature T/J [recall Eq. (1)]. Although the nominal value for the chip is (calculated with reference to the classical Hamiltonian at the end of the anneal), any inhomogeneity of the temperature across the chip may render the ratio higher, possibly driving it above typical TC critical temperatures. (We refer to the physical temperature of the chip. However, nonequilibrium systems, e.g., supercooled liquids or glasses, can be characterized by two temperatures^{45}: On the one hand, the physical temperature T which rules fast degrees of freedom that equilibrate. On the other hand, the ‘effective’ temperature refers to slow degrees of freedom that remain out of equilibrium. We are currently investigating whether or not the DW2 chip can analogously be characterized by two such temperatures.)
Another possible cause for the above scaling may be due to the analog nature of the chip. The programming of the coupling parameters and magnetic fields is prone to statistical and systematic errors (also referred to as intrinsic control errors, or ICE). The couplings actually encoded in DW2 are , where is a random error (, according to the manufacturer of the chip). Unfortunately, even tiny changes in coupling values are known to potentially change the groundstate configurations of spin glasses in a dramatic manner^{37,46,47,48}. We refer to this effect as ‘coupling chaos’ (or Jchaos, for short). For an Nbit system, Jchaos seems to become significant for . Empirically^{37,48} , D being the spatial dimension of the system. Note, however, that these estimates refer only to typical instances and small N whereas the assessment of the effects of Jchaos on thermallyhard instances remains an important open problem for classical spin glasses.
Here, we empirically quantify the effects of Jchaos by taking advantage of the many programming cycles and gauge choices each instance has been annealed with (typically between 200 and 2000). Calculating a success probability p for each cycle, we compute the probability distribution of p over different cycles for each instance^{23,38}. We find that while for some instances p is essentially insensitive to programming errors, for other instances (even within the same thermal hardness group), p varies significantly, spanning several orders of magnitude. This is illustrated in Fig. 4 which presents some results based on a straightforward percentile analysis of these distributions. For instance #1 in the figure, the 80th percentile probability is , whereas the probability at the 90th percentile is . Hence the ratio is close to one. Conversely, for instance #35, the values are , and the ratio is , i.e., the success probability drops by an order of magnitude. The inset of Fig. 4 shows the typical ratio as a function of classical hardness, demonstrating the strong correlation between thermal hardness and the devastating effects of Jchaos caused by ICE, namely that the larger τ is, the more probable it is to find instances for which p varies wildly between programming cycles.
Discussion
We have devised a method for quantifying the susceptibility of quantum annealers to classical effects by studying their performance on sets of instances characterized by different degrees of thermal hardness, which we have defined for that purpose as the mixing (or equilibration) time τ of classical thermal algorithms (namely, PT) on these. We find that the 2Dlike Chimera architecture used in the DWave chips does give rise to thermally veryhard, albeit rare, instances. Specifically, we have found samples that exhibit temperature chaos and as a such have very long mixing times, i.e., they are classically exceptionally hard to solve.
We demonstrated that ‘temperature chaos’, or TC, is an inherent property of the (classical) freeenergy landscape of certain (hard) optimization problem instances and is expected to universally hinder the performance of any heuristic optimization algorithm.
Applying our method to an experimental quantum annealing optimizer, the DW2 chip, we have found that its performance is more susceptible to changes in thermal hardness than classical algorithms. This is in contrast with the performance of the bestknown stateoftheart classical solver on Chimera graphs, the ‘nonthermal’ HFS algorithm, which scales (unsurprisingly) better. Our results are not meant to suggest that the DW2 chip is not a quantum annealer, but rather that its performance is greatly impeded by muchundesired classical effects.
We have identified and quantified two probable causes for the observed behavior: A possibly too high temperature, or more probably, Jchaos, the random errors stemming from the digitaltoanalog conversion in the programming of the coupling parameters. One may hope that the scaling of current DW2 chips would significantly improve if one or both of the above issues are resolved. Clearly, lowering the temperature of the chip and/or reducing the error involved in the programming of its parameters are both technologically very ambitious goals, in which case error correcting techniques may prove very useful^{49}. We believe that quantum Monte Carlo simulations of the device will be instrumental to the understanding of the roles that temperature and magnitude of programming errors play in the performance of the chip (and of its classical counterparts). In turn, this will help sharpening the most pressing technological challenges facing the fabrication of these and other future quantum optimizing devices, paving the way to obtaining longawaited insights as to the difference between quantum and classical hardness in the context of optimization. We are currently pursuing these approaches.
Methods
Computation of the mixing time τ
Because we follow ref. 22 we just briefly summarize here the main steps of the procedure. Considering one of the N_{T} system copies in the PT simulation, let us denote the temperature of copy i at Monte Carlo time t by , where (see inset of Fig. 1). At equilibrium, the probability distribution for i_{t} is uniform (namely, 1/N_{T}) hence the exact expectation value of i_{t} is . From the general theory of Markov Chain Monte Carlo^{20}, it follows that the equilibrium timecorrelation function may be written as a sum of exponentially decaying terms:
The mixing time τ is the largest ‘eigentime’ τ_{1}. We compute numerically the correlation function and fit it to the decay of two exponential functions (so we extract the dominant time scale τ and a subleading time scale). The procedure is described in full in ref. 22.
Dwave data acquisition and analysis
Data acquisition
In what follows we briefly summarize the steps of the experimental setup and data acquisition for the anneals performed on the 414 randomlygenerated instances in the various thermalhardness groups.

1
The J_{ij} couplings of each of the 414 instances have been encoded onto the DWave chip using many different choices of annealing times in the allowed range of 20 μs ≤ t_{a} ≤ 20 ms.

2
For each instance and each choice of t_{a} the following process has been repeated hundreds to thousands of times:

a
First, a random ‘gauge’ has been chosen for the instance. A gauge transformation does not change the properties of the optimization problem but has some effect on the performance of the chip which follows from the imperfections of the device that break the gauge symmetry. The different gauges are applied by transforming the original instance , , to the original cost function Eq. (1). The above gauge transformations correspond to the change in configuration spin values. Here, the N gauge parameters were chosen randomly.

b
The chip was then programmed with the gaugetransformed instance (inevitably adding programming bias errors, as mentioned in the main text).

c
The instance was then solved, or annealed, X times within the programming cycle/with the chosen gauge. We chose , the maximally allowed amount.

d
After the X anneals were performed, the number of successes Y, i.e., the number of times a minimizing configuration had been found, was recorded. The probability of success for the instance, for that particular gauge/programming cycle and annealing time t_{a} was then estimated as . Note, that in cases where the probability of success is of the order of 1/X, the probability p will be a rather noisy estimate (see data analysis, below, for a procedure to mitigate this problem).

e
If the prefixed number of cycles for the current instance and t_{a} has not been reached, return to (2a) and choose a new gauge.

a
Data analysis and timetosolution estimates
The analysis of the data acquisition process described above proceeded as follows.

1
For each instance and each annealing time, the total number of hits was calculated, where i sums over all the gauge/programming cycles. Denoting as the total number of annealing attempts, the probability of success for any particular instance and anneal time was then calculated as .

2
The above probability was then converted into an average timetosolution t_{s} for that instance and t_{a} according to t_{s} = t_{a}/P, where the special case of P = 0 designates an estimate of an infinite t_{s}, where in practice the true probability lies below the resolution threshold of 1/X_{tot}.

3
A typical runtime for a hardness group was then obtained by taking the median over all minimal t_{s} values of all the instances in the group.
Simulated Annealing (SA) runs
We employed the following protocol to find ground states for all 414 tested instances. Each annealing schedule was run 1024 times. The temperature range was chosen to be identical to that used in our Parallel Tempering simulations with a starting temperature of , where the system was left to equilibrate for 1000 fulllattice Metropolis sweeps and a final temperature of which is below typical TC critical points (see Fig. 2). The cooling schedule that was chosen is linear in with the temperature decreasing after each fulllattice sweep. In the initial cooling rate, we sent the temperature from T_{max} down to T_{min} in 256 cooling steps. If the number of independent runs reaching a ground state at the end of each anneal was found to be less than one eighth of the total anneals (i.e., if less than 128 runs ended in a ground state), the number of cooling steps was doubled and the runs were restarted. The procedure was iterated until the “oneeighth success rate” criterion has been met. We then took the final number of cooling steps as a figure of merit for the success of the SA algorithm on that particular instance. Finally, we computed the median over the instances of each hardness group in order to obtain a characteristic timetosolution t_{s}. Interestingly, for several of the hardest problems, the “oneeighth success rate” criterion was not met even with 2 × 10^{9} cooling sweeps (40 instances in the τgeneration group 10^{6}, as well as thirteen out of the fourteen instances in the τ ~ 10^{7} group).
Additional Information
How to cite this article: MartinMayor, V. and Hen, I. Unraveling Quantum Annealers using Classical Hardness. Sci. Rep. 5, 15324; doi: 10.1038/srep15324 (2015).
References
Shor, P. W. Polynomial timealgorithms for prime factorization and discrete logarithms on a quantum computer. SIAM J. Comp. 26, 1484–1509 (1997).
Grover, L. K. Quantum mechanics helps in searching for a needle in a haystack. Phys. Rev. Lett. 79, 325–328 (1997).
Schlosshauer, M. Decoherence, the measurement problem and interpretations of quantum mechanics. Rev. Mod. Phys. 76, 1267–1305 (2004).
Johnson, M. W. et al. Quantum annealing with manufactured spins. Nature 473, 194–198 (2011).
Albash, T., Boixo, S., Lidar, D. A. & Zanardi, P. Quantum adiabatic markovian master equations. New Journal of Physics 14, 123016 (2012).
Lanting, T. et al. Entanglement in a quantum annealing processor. Phys. Rev. X 4, 021041 (2014).
Albash, T., Hen, I., Spedalieri, F. M. & Lidar, D. A. Reexamination of the evidence for entanglement in the DWave processor. arXiv:1506.03539 (2015).
Smolin, J. A. & Smith, G. Classical signature of quantum annealing. arXiv:1305.4904 (2013).
Shin, S. W., Smith, G., Smolin, J. A. & Vazirani, U. How “quantum” is the DWave machine? arXiv:1401.7087 (2014).
I. Hen et al. Probing for quantum speedup in spin glass problems with planted solutions. arXiv:1502.01663 (2015).
Ronnow, T. F. et al. Defining and detecting quantum speedup. Science 345, 420–424 (2014).
Boixo, S. et al. Evidence for quantum annealing with more than one hundred qubits. Nat Phys 10, 218–224 (2014).
Young, A. P. Spin Glasses and Random Fields (World Scientific. Singapore, 1998).
Belletti, F. et al. (Janus Collaboration). Simulating spin systems on IANUS, an FPGAbased computer. Comp. Phys. Comm. 178, 208–216 (2008).
Belletti, F. et al. (Janus Collaboration). Janus: An FPGAbased system for highperformance scientific computing. Computing in Science and Engineering 11, 48–58 (2009).
BaityJesi, M. et al. (Janus Collaboration). Janus II: a new generation applicationdriven computer for spinsystem simulations. Comp. Phys. Comm 185, 550–559 (2014).
Hukushima, K. & Nemoto, K. Exchange monte carlo method and application to spin glass simulations. J. Phys. Soc. Japan 65, 1604–1608 (1996).
Marinari, E. In Advances in Computer Simulation (eds. Kertész, J. & Kondor, I. ), 50–81 (SpringerVerlag, 1998).
Kirkpatrick, S., Gelatt Jr., C. D. & Vecchi, M. P. Optimization by simulated annealing. Science 220, 671–680 (1983).
Sokal, A. In Functional Integration: Basics and Applications (eds. DeWittMorette, C., Cartier, P. & Folacci, A. ss), 131–192 (Plenum, 1997).
Fernandez, L. A. et al. Phase transition in the three dimensional Heisenberg spin glass: Finitesize scaling analysis. Phys. Rev. B 80, 024422 (2009).
Alvarez Baños, R. et al. (Janus Collaboration). Nature of the spinglass phase at experimental length scales. J. Stat. Mech. 2010, P06026 (2010).
Fernandez, L. A., MartinMayor, V., Parisi, G. & Seoane, B. Temperature chaos in 3d Ising spin glasses is driven by rare events. EPL 103, 67003 (2013).
McKay, S. R., Berker, A. N. & Kirkpatrick, S. Spinglass behavior in frustrated Ising models with chaotic renormalizationgroup trajectories. Phys. Rev. Lett. 48, 767–770 (1982).
Bray, A. J. & Moore, M. A. Chaotic nature of the spinglass phase. Phys. Rev. Lett. 58, 57–60 (1987).
Banavar, J. R. & Bray, A. J. Chaos in spin glasses: A renormalizationgroup study. Phys. Rev. B 35, 8888–8890 (1987).
Kondor, I. On chaos in spin glasses. J. Phys. A 22, L163–L168 (1989).
Kondor, I. & Végsö, A. Sensitivity of spinglass order to temperature changes. J. Phys. A 26, L641–L646 (1993).
Billoire, A. & Marinari, E. Evidence against temperature chaos in meanfield and realistic spin glasses. J. Phys. A 33, L265–L272 (2000).
Rizzo, T. Against chaos in temperature in meanfield spinglass models. J. Phys. 34, 5531–5549 (2001).
Mulet, R., Pagnani, A. & Parisi, G. Against temperature chaos in naive thoulessandersonpalmer equations. Phys. Rev. B 63, 184438 (2001).
Billoire, A. & Marinari, E. Overlap among states at different temperatures in the sk model. Europhys. Lett. 60, 775–781 (2002).
Krzakala, F. & Martin, O. C. Chaotic temperature dependence in a model of spin glasses. Eur. Phys. J. 28, 199–208 (2002).
Rizzo, T. & Crisanti, A. Chaos in temperature in the SherringtonKirkpatrick model. Phys. Rev. Lett. 90, 137201 (2003).
Parisi, G. & Rizzo, T. Chaos in temperature in diluted meanfield spinglass. J. Phys. A 43, 235003 (2010).
Sasaki, M., Hukushima, K., Yoshino, H. & Takayama, H. Temperature chaos and bond chaos in EdwardsAnderson Ising spin glasses: Domainwall freeenergy measurements. Phys. Rev. Lett. 95, 267203 (2005).
Katzgraber, H. G. & Krzakala, F. Temperature and disorder chaos in threedimensional Ising spin glasses. Phys. Rev. Lett. 98, 017201 (2007).
Billoire, A. Rare events analysis of temperature chaos in the SherringtonKirkpatrick model. J. Stat. Mech. 2014, P04016 (2014).
Thomas, C. K., Huse, D. A. & Middleton, A. A. Zero and low temperature behavior of the twodimensional ±j Ising spin glass. Phys. Rev. Lett. 107, 047203 (2011).
Katzgraber, H. G., Hamze, F. & Andrist, R. S. Glassy chimeras could be blind to quantum speedup: Designing better benchmarks for quantum annealing machines. Phys. Rev. X 4, 021008 (2014).
Hamze, F. & de Freitas, N. From fields to trees. arXiv:1207.4149 (2012).
Selby, A. Efficient subgraphbased sampling of Isingtype models with frustration. arXiv:1409.3934 (2014).
Hen, I. & Young, A. P. Exponential complexity of the quantum adiabatic algorithm for certain satisfiability problems. Phys. Rev. E. 84, 061152 (2011).
Farhi, E. et al. Performance of the quantum adiabatic algorithm on random instances of two optimization problems on regular hypergraphs. Phys. Rev. A 86, 052334 (2012).
Cugliandolo, L. F., Kurchan, J. & Peliti, L. Energy flow, partial equilibration and effective temperatures in systems with slow dynamics. Phys. Rev. E 55, 3898–3914 (1997).
Nifle, M. & Hilhorst, H. J. New criticalpoint exponent and new scaling laws for shortrange Ising spin glasses. Phys. Rev. Lett. 68, 2992–2995 (1992).
NeyNifle, M. Chaos and universality in a fourdimensional spin glass. Phys. Rev. B 57, 492–496 (1998).
Krzakala, F. & Bouchaud, J. P. Disorder chaos in spin glasses. Europhys. Lett. 72, 472–478 (2005).
Pudenz, K. L., Albash, T. & Lidar, D. A. Errorcorrected quantum annealing with hundreds of qubits. Nature Communications 5, 3243 (2014).
Acknowledgements
We thank Luis Antonio Fernández and David Yllanes for providing us with their analysis program for the PT correlation function. We also thank Marco BaityJesi for helping us to prepare the figures. We are indebted to Mohammad Amin, Luis Antonio Fernández, Enzo Marinari, Denis Navarro, Giorgio Parisi, Federico RicciTersenghi and Juan Jesús RuizLorenzo for discussions. We thank Luis Antonio Fernández, Daniel Lidar, Felipe LLanesEstrada, David Yllanes and Peter Young for their reading of a preliminary version of the manuscript. We thank DWave Systems Inc. for granting us access to the chip. We acknowledge the use of algorithms and source code for a classic solver, devised and written by Alex Selby, available for public usage at https://github.com/alex1770/QUBOChimera. Our simulated annealing runs were carried out on the Picasso supercomputer. The authors thankfully acknowledge the computer resources, technical expertise and assistance provided by the staff at the Red Española de Supercomputación. IH acknowledges support by ARO grant number W911NF1210523. VMM was supported by MINECO (Spain) through research contract N^{o} FIS201235719C02.
Author information
Authors and Affiliations
Contributions
V.M.M. and I.H. jointly designed the research project and wrote the manuscript. I.H. conducted the experiments involving the Dwave chip and performed most of the corresponding data analysis, as well as the execution, data collection and subsequent analysis of the HFS algorithm runs. V.M.M. wrote the code for classical PT simulations, performed the classical simulations and analyzed their output. He also performed the “Jchaos” analysis of the Dwave data.
Ethics declarations
Competing interests
The authors declare no competing financial interests.
Electronic supplementary material
Rights and permissions
This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
About this article
Cite this article
MartinMayor, V., Hen, I. Unraveling Quantum Annealers using Classical Hardness. Sci Rep 5, 15324 (2015). https://doi.org/10.1038/srep15324
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/srep15324
This article is cited by

Models in quantum computing: a systematic review
Quantum Information Processing (2021)

Temperature chaos is present in offequilibrium spinglass dynamics
Communications Physics (2021)

Advantages of Unfair Quantum GroundState Sampling
Scientific Reports (2017)

Zerotemperature quantum annealing bottlenecks in the spinglass phase
Nature Communications (2016)

Nested quantum annealing correction
npj Quantum Information (2016)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.