Abstract
It is predicted that quantum computers will dramatically outperform their conventional counterparts. However, largescale universal quantum computers are yet to be built. Boson sampling^{1} is a rudimentary quantum algorithm tailored to the platform of linear optics, which has sparked interest as a rapid way to demonstrate such quantum supremacy^{2,3,4,5,6}. Photon statistics are governed by intractable matrix functions, which suggests that sampling from the distribution obtained by injecting photons into a linear optical network could be solved more quickly by a photonic experiment than by a classical computer. The apparently low resource requirements for large boson sampling experiments have raised expectations of a nearterm demonstration of quantum supremacy by boson sampling^{7,8}. Here we present classical boson sampling algorithms and theoretical analyses of prospects for scaling boson sampling experiments, showing that nearterm quantum supremacy via boson sampling is unlikely. Our classical algorithm, based on Metropolised independence sampling, allowed the boson sampling problem to be solved for 30 photons with standard computing hardware. Compared to current experiments, a demonstration of quantum supremacy over a successful implementation of these classical methods on a supercomputer would require the number of photons and experimental components to increase by orders of magnitude, while tackling exponentially scaling photon loss.
Main
It is believed that new types of computing machines will be constructed to exploit quantum mechanics for an exponential speed advantage in solving certain problems compared with classical computers^{9}. Recent large state and private investments in developing quantum technologies have increased interest in this challenge. However, it is not yet experimentally proven that a large computationally useful quantum system can be assembled, and such a task is highly nontrivial given the challenge of overcoming the effects of errors in these systems.
Boson sampling is a simple task which is native to linear optics and has captured the imagination of quantum scientists because it seems possible that the anticipated supremacy of quantum machines could be demonstrated by a nearterm experiment. The advent of integrated quantum photonics^{10} has enabled large, complex, stable and programmable optical circuitry^{11,12}, while recent advances in photon generation^{13,14,15} and detection^{16,17} have also been impressive. The possibility to generate many photons, evolve them under a large linear optical unitary transformation, then detect them, seems feasible, so the role of a boson sampling machine as a rudimentary but legitimate computing device is particularly appealing. Compared to a universal digital quantum computer, the resources required for experimental boson sampling appear much less demanding. This approach of designing quantum algorithms to demonstrate computational supremacy with nearterm experimental capabilities has inspired a raft of proposals suited to different hardware platforms^{18,19,20}.
Based on a simple architecture, the boson sampling problem is similarly straightforward to state. A number n of indistinguishable noninteracting bosons (for example, photons) should be injected into n input ports of a circuit comprised of a number m of linearly coupled bosonic modes. The circuit should be configured so that the transformation between input and output ports is described by a uniformly (‘Haar’) random unitary matrix. The probability for the n bosons to be detected at given set of n output ports is equal to the square of the absolute value of the permanent of the transfer matrix that describes the transformation.
Although choosing a number of modes m ∼ n^{5} log^{2}n guarantees that the distribution of any n × n submatrix is approximately equal to that of a matrix of elements drawn independently from the complex normal distribution^{1}, the less impractical scaling of m ∼ n^{2} is typically targeted. This polynomial relation between n and m is also important because it ensures a not too large probability that two or more of the bosons arrive at the same output port, that is, bunch; the conjectured hardness applies only to collisionfree events, that is, no bunching. Because approximating the permanent of a random matrix is conjectured to be computationally hard^{1}, calculating any transition probability is intractable; the collection of all of the possible collisionfree transition probabilities (m choose n) constitutes an exponentially large probability distribution, where each element is exponentially hard to calculate. Running an ideal boson sampler would solve the problem of producing samples from this distribution.
Importantly, a strong case for the classical hardness of boson sampling can be made even when the distribution being sampled from is only approximately correct^{1}: assuming certain conjectures from computational complexity theory, there can be no efficient classical algorithm to sample from any distribution within a small distance from the boson sampling distribution.
Current estimates for the regime in which photonic experiments could achieve quantum supremacy have been between 20 and 30 photons^{1,21,22} and, recently, as low as 7 (ref. 7). However, our classical algorithm, based on Metropolised independence sampling (MIS), although necessarily inefficient for large n, was able to output a sample for n = 30 bosons in half an hour on a standard laptop, and would enable an n = 50 sample to be produced in under 10 days on a supercomputer^{23}. MIS^{24} is a specific Markov Chain Monte Carlo (MCMC) method. For an instance of the problem, our aim is to construct a Markov chain where each state in the chain signifies a boson sampling detection event. New states in the chain are proposed from a classical mockup of the boson sampling distribution: the distribution of distinguishable particles, where probabilities of detection events are equal to permanents of real and positive transition matrices, and sampling becomes efficient^{25}. Proposed states are then accepted or rejected by comparison with the actual probability to observe that event for indistinguishable bosons. This means that, at each step, only a relatively small number of permanents must be calculated; a calculation of the full and exponentially large boson sampling distribution is not required.
More precisely, let P_{D} be the distinguishable particle distribution with probability mass function g(x), over the set of tuples of length m with elements in {0,1} which sum to n. And let P_{BS} be the boson sampling distribution over these tuples, with probability mass function f(x). Then, starting at a random (according to P_{D}) tuple x, propose a new random tuple x′. The transition from x to x′ is accepted with probability
Repeating this procedure generates a Markov chain, which will converge such that thereafter, the chain is sampling from P_{BS}.
Not all states in the Markov chain are retained as detection events. The time taken for the Markov chain to converge means that a number of tuples at the beginning of the chain must be discarded, known as the ‘burn in’ period, τ_{burn}. For the size of simulations covered here, empirical tests find that a burn in period of 100 is sufficient for convergence to have occurred. In addition, autocorrelation between states in the chain can occur, for example, because two consecutive states in the chain will be identical whenever a proposed new state x′ is not accepted. We empirically find for the size of problem we tackle that autocorrelation is suppressed with a thinning procedure that retains only every 100th state (see Supplementary Information). Generally, the burn in period and thinning interval are reduced by a greater overlap between target and proposal distributions, as measured by (1), the transition probability. We find that a proposal distribution of the distinguishable particle transition probabilities has a high acceptance rate of ∼40%, a sign that the distributions overlap well. In Fig. 1 we present a schematic of the MISbased approach to boson sampling, alongside schematics of a quantum photonic approach and the boson sampling problem itself.
The classical resources used to produce the thinned Markov chain are far fewer than those required by the brute force approach of calculating all permanents in the full distribution. The asymptotically fastest known algorithm for exactly computing the permanent of an n × n matrix is Ryser’s algorithm^{26}, whose runtime when implemented efficiently is . Generating the first tuple in a sample requires the computation of τ_{burn}n × n realvalued matrix permanents and τ_{burn}n × n complexvalued matrix permanents. Each subsequent sample requires τ_{thin}n × n realvalued and complexvalued matrix permanents to be computed. The relative scaling of the approaches to classical boson sampling using a standard laptop is shown in Fig. 2b, setting both τ_{burn} and τ_{thin} equal to 100. The MIS sampler is over 50 orders of magnitude faster for the n = 30 case than the brute force computation of the entire distribution.
We used this algorithm on a standard laptop to produce samples of size 20,000 for up to 20 bosons, and used a local server, which allowed around 30 times more chains to be run in parallel, to produce 250 samples for 30 bosons in 900 modes in less than five hours. As in the experimental case, a central challenge is to provide evidence for sampling from the correct distribution. Here we addressed this using standard statistical techniques. Testing for distinguishability among photons is one of the main verification tools used in experiments, while here the distribution resulting from distinguishable photons has the desirable property of being somewhat correlated with the ideal boson sampling distribution. It is conceivable (although not necessary) that results from an inaccurate sampling algorithm could sometimes be closer to the distinguishable distribution than the ideal boson sampling distribution. This is especially relevant when the sampling algorithm itself uses samples from the distinguishable distribution. The likelihood ratio tests^{27} in Fig. 2a show a rapid growth in confidence in the hypothesis that these samples are from the indistinguishable boson distribution rather than the distinguishable particle distribution for n = 7,12,20 and 30.
Further results for verification against the distinguishable distribution are shown in Fig. 2c for the case of 7 bosons in 49 modes. For each tuple in a sample size of 20,000 produced by our classical algorithm, we calculate −log(Per(A_{S})^{2}), where A_{S} is the matrix associated to each tuple, and produce a probability mass histogram. The same function and associated histogram is plotted for a sample of 20,000 tuples chosen from the distinguishable particle distribution (note that Per(A_{S}^{2}) gives the probability to observe a transition of distinguishable particles). Figure 2c shows the clear difference between the two distributions, which we analyse with a (bootstrapped) 2sample Kolmogorov–Smirnov (KS) test^{28}. We are able to reject the null hypothesis, that the two samples are chosen from the same distribution, at a significance level of 0.001.
We implement stronger verification tests by comparing (up to computational limits) our MIS algorithm against algorithms that are known to be exact. The KS tests in Fig. 2c between distributions from our MIS algorithm and a rejection sampling algorithm, and between our MIS algorithm and the brute force approach of calculating all permanents in the full distribution, both found large pvalues (see Supplementary Information). This striking level of consistency (including further verification against rejection sampling for 12 photons—see Supplementary Information) is reassuring evidence for the correct operation of our MIS algorithm.
We next compare our classical approach with plausible experimental parameters. It is worth noting that asymptotically, experimental boson sampling will have a slower runtime than our algorithm. This is because photon losses scale exponentially badly with n (refs 1,29). The runtime for an experiment with a transmission probability resulting from fixed loss (generation, coupling, detection) η_{f} and a transmission probability η_{0}, resulting from loss per unit optical circuit depth scales as for an optical circuit depth of d, which is worse than Ryser’s algorithm if d grows with n for any η_{f}, η_{0} < 1. However, the region of interest for quantum supremacy is likely to be restricted to n < 100, where lowloss experiments still have the potential to produce large speedups. Assuming that our MIS sampler continues to perform equally well for larger instance sizes, we can compare its runtime with current and future experiments. The classical and quantum runtimes for an instance of size n bosons in m = n^{2} modes can be estimated as
where a is the time scaling of the classical computer (for computing one real and one complex permanent), the factor of e is an approximation to the probability of obtaining a collisionfree event^{30}, R is the experimental source repetition rate, and η = η_{f}η_{0}^{d} is the experimental transmission probability of a single photon, including the efficiencies of photon generation, coupling, circuit propagation and detection (note that R and η will generally be a function of n). We define the quantum advantage (QA) as the improvement in quantum runtime versus classical runtime measured in orders of magnitude,
We now consider two plausible notions of quantum supremacy. First, we can define supremacy as a speedup so large that it is unlikely to be overcome by algorithmic or hardware improvements to the classical sampler, for which we choose a speedup of ten orders of magnitude. Secondly, we may wish to define supremacy as the point at which a computational task is performed in a practical runtime on a quantum device, for which we choose under a week, but in an impractical runtime on a classical device, for which we choose over a century.
These criteria can be summarized as
To make concrete estimates of future runtimes, we need to fix a and R. Choosing a = 3n × 10^{−15} s as the time scaling for computing one real and one complex matrix permanent recently reported for the supercomputer Tianhe 2 (ref. 23) and R = 10 GHz, which is faster than any experimentally demonstrated photon source to our knowledge, we can plot QA against n and η.
We first note that current approaches using spontaneous parametric down conversion (SPDC) photon pairs are generally inefficient with η < 0.002 (refs 2,3,4,5,11,22,31,32). Recently, improved rates have been demonstrated with quantum dot photon sources^{8,33,34}. The current leading experimental demonstration, however, is still restricted to η ≈ 0.08 for n = 5, where q_{t} ≈ 10^{9}c_{t}. This calculation includes the rate used in the experiment (76n^{−1} MHz) and includes a suppression factor caused by a lower collisionfree event rate using a linear instead of quadratic mode scaling. In Wang et al.^{8} a number of realistic, nearterm experimental improvements are suggested to reach 20photon boson sampling. Using these projections we find that η is increased to approximately 0.35, which would be a major experimental breakthrough. However, as shown in Fig. 2d, in this case we predict that the classical runtime would still be over six orders of magnitude faster. Fig. 2e shows the regions of quantum advantage and quantum supremacy with current and projected experiments.
In ref. 35, the authors showed that the boson sampling problem can be modified to allow for a fixed number of lost photons at the input of the circuit whilst retaining computational hardness. We show that if the overall losses in the experiment are pathindependent then this is equivalent to loss at the input (see Methods). The MIS sampler can be readily adapted to this scenario by adding an initial step which generates a uniformly random input subset, followed by the usual MIS method for this input state. The dashed contours and lines in Fig. 2e and d take into account the adjusted classical and quantum runtimes when up to two lost photons are allowed. Although allowing loss helps the experiments to compete, the complexity of realistic experimental regimes such as losing a constant fraction of photons remains unknown, and it is easy to see that losing too many photons eventually allows the problem to become efficiently solvable classically.
We have shown that the boson sampling problem can probably be solved for 50 photons and any given Haar random unitary matrix, in a feasible time frame using today’s classical technology. Although our classical methods for boson sampling are not optimized, they are roughly as effective as they could possibly be without violating standard complexitytheoretic hardness assumptions. And although likely incremental improvements would push the supremacy threshold further still from current experimental capabilities, our methods serve as the first rigorous benchmark for laboratory experiments. To solve the boson sampling problem at a scale that is beyond the capabilities of the best classical computers, experiments face the initial challenge of efficiently generating upwards of 50 photons in welldefined modes. Lowloss photon propagation is then required in circuitry of several thousands of modes, before arriving at several thousands of highefficiency detectors. Programming different Haar random unitary matrices over those modes requires the precise setting of millions of phase shifters^{36}.
Addressing these challenges could be more feasible through the use of other (nonspatial) encodings such as the time domain^{37,38}, although lowloss programmability in other domains at this scale brings additional challenges.
Although the boson sampling algorithm could be run on a faulttolerant quantum computer, this approach would lose the appealing simplicity of the original proposal. An interesting direction for boson sampling is the development of basic error correction techniques that are far short of what is required for universal digital quantum computing, yet sufficient to enable a demonstration of quantum supremacy^{18}. Such error correction schemes might then be carried over into other photonic quantum technologies. Finally, we note that our classical sampling methods could be more broadly applicable to other sampling algorithms, not least to the many variants of boson sampling that have arisen^{39,40}.
Methods
The boson sampling problem^{1}.
Given U ∈ SU(m), let A be the columnorthonormal, m × n matrix formed by taking the first n columns of U. Also let Φ_{m, n} be the set of all possible tuples of length m of nonnegative integers whose elements sum to n.
For some tuple S = (s_{1}, …, s_{m}) ∈ Φ_{m, n}, let A_{S} be the submatrix of A with s_{i} copies of row i. Boson sampling is the problem of sampling from the probability distribution P_{BS} over Φ_{m, n}, for a given input U, with probabilities defined in terms of permanents of submatrices as:
In this work we restrict ourselves to the collisionfree subspace (CFS), meaning that s_{i} ∈ {0,1} ∀ i. See Supplementary Information for more detail.
Distinguishable particle distribution.
The proposal used in the MIS algorithm is that which describes the same distribution for distinguishable particles (7):
where for a complex matrix A with elements A_{ij}, A^{2} denotes the matrix with elements A_{ij}^{2}. However, there is a classical algorithm which can sample from this distribution in time O(mn) (ref. 25).
Brute force exact sampling.
This method computes all of the probabilities in the CFS and samples from the associated probability mass function, and is computationally demanding given the exponential scaling of the CFS with the size of the problem. This approach requires computing permanents of n × n complexvalued matrices before a single sample can be output. Using the lower bound , for m ≥ n^{2} at least n^{n} permanents must be computed; and even for n = 10 we need to compute more than 17 × 10^{12} permanents. The computation quickly becomes swamped by the number of permanents to compute, rather than the complexity of computing the permanent itself.
Rejection sampling with a uniform proposal.
Rejection sampling is a general approach for exactly sampling from a desired distribution P with probability mass function f(x), given the ability to sample from a distribution Q with probability mass function g(x), where f(x) ≤ λg(x) for some λ and all x. The algorithm proceeds as follows: Step 1: Generate a sample x from Q. Step 2: With probability f(x)/λg(x), output x. Otherwise, go to step 1.
The probability that an accepted sample is generated is
Here we take P to be the boson sampling distribution restricted to the CFS, and Q to be the uniform distribution on the CFS (so ). Note that P is subnormalized, so is not quite a probability distribution. However, the rejection sampling algorithm is blind to this subnormalization (as this is effectively the same as increasing λ), so will generate samples from the renormalized distribution.
Each iteration of rejection sampling requires the computation of one permanent, corresponding to f(x). To use rejection sampling to exactly sample from the target distribution with a uniform proposal most efficiently, it is required to know the maximum value of f(x), which corresponds to the largest permanent of all n × n submatrices A of a given m × m unitary matrix U. Without any bound on this quantity, we would be forced to use the trivial bound μ = 1, corresponding to permanent computations being required to obtain one sample from P.
As we are only attempting to perform approximate boson sampling, we only require a good estimate of μ = max_{x}f(x). It was argued in ref. 1 that sampling from a distribution within total variation distance ε of the real boson sampling distribution P_{BS} should be computationally difficult, for some small constant ε. Imagine that our guess for μ is too small, such that . Then if x is sampled uniformly at random and , step 2 of the rejection sampling algorithm will fail. If we modify the rejection sampling algorithm to simply produce a new uniform sample in this case and repeat, it is easy to see that we can view the modified algorithm as sampling from the truncated distribution P_{low} with probability mass function
Then the total variation distance between P_{low} and P is
So if the probability mass of P above μ is at most ε, we have sampled from a distribution within distance ε of P. We have found that we are able to use a simple random restart hillclimbing algorithm to provide a suitable estimate of μ with computations of n × n matrix permanents.
Random restart hillclimbing algorithm.
We start by randomly sampling a submatrix, represented by the tuple S, from the uniform distribution. For one pass, we greedily try replacing each row in the sampled submatrix by each row from A in turn, accepting only if this increases Pr(S) while also making sure to avoid selecting the same row twice. We perform repeat passes until there is no improvement of the probability over a complete pass. At this point we randomly resample a new starting submatrix and repeat from the beginning. The total number of permanent calculations for one pass is n(m − n).
This method is not guaranteed to find a global maximum. However, in our experiments for n ≤ 7, where we can still compute the full probability mass function exactly, we found the estimates for the maximum probability to be exactly equal to the global maximum in the overwhelming majority of cases. In the range 8 ≤ n ≤ 12, where we no longer are able to compute the exact maximum probability, the bounds from our hillclimbing algorithm also allowed us to sample using rejection sampling efficiently and then compare our results with our MIS sampler. This provided further evidence for both sampling techniques.
Scattershot boson sampling.
We can straightforwardly modify our MISbased method to carry out the scattershot boson sampling problem. For each sample that we wish to output, we can first sample (efficiently) from the uniform distribution on nfold input modes (which fixes the columns of U contributing to submatrices), before running the algorithm in the way described above for a single sample. In this case, τ_{thin} becomes meaningless and we are interested only in τ_{burn}, as we start a new chain for each sample.
Alternatively, our proposal distribution can be changed to include the uniform distribution over nfold input modes, meaning that each state in a given Markov chain can correspond to a different input configuration.
Events with collisions.
Although in this work we do not examine the more general situation of boson sampling where there can be more than one boson in an output mode, we anticipate that relaxing the CFS restriction will not increase the run time of the MIS method. In fact, it is possible that the average run time could be decreased with this relaxation, as there exists an algorithm for computing the permanent which is exponential in matrix rank, rather than matrix size^{41}. However, due to there usually existing very large permanents of subunitary matrices with many repeated rows, relaxing the CFS restraint has an adverse effect on the average run time of our rejection sampling method.
Likelihood ratio test.
We perform a modified version of the likelihood ratio test described by Bentivegna et al. ^{27} (see Supplementary Information for more detail). Defining two hypotheses and , the indistinguishable boson hypothesis and the distinguishable particle hypothesis, respectively, and letting q_{x} be the probability of seeing the sampled event x according to hypothesis , and r_{x} be the corresponding probability under hypothesis , we can write
where N_{events} is the sample size. Normalizing equation (9) such that yields
We note that q_{x} and r_{x} should not simply correspond to Per(A_{x})^{2} and Per(A_{x}^{2}), respectively. The reason for this is that we are restricted to the CFS, and the probability of a sample being collisionfree differs between indistinguishable bosons and distinguishable particles. So q_{x} and r_{x} must be normalized independently, such that they independently sum to 1 over all events x. Doing this exactly would require summing all probabilities in the CFS for the specific instance of the problem being considered. To approximate the normalization of q_{x} efficiently, here we instead average the probability of the output being collisionfree over the Haar measure^{1,30}:
For the hypothesis , we can efficiently sample output tuples, and so the ratio of collisionfree tuples to tuples with collisions in a large sample provides an approximate normalization for r_{x}.
For each problem size, we use this likelihood ratio test to assess the performance of MIS samplers with different τ_{burn} and τ_{thin}. As the proposal distribution for the sampler is the distinguishable particle distribution, we might expect that if the chain has not converged to the target distribution, this will manifest itself as samples looking more like they are from the distinguishable particle distribution than they should. Also, we expect that samplers with larger τ_{burn} and τ_{thin} are more likely to sample from the target distribution. Because of this, we expect to be able to observe a point at which increasing τ_{burn} and τ_{thin} has, on average, no effect on the outcome of a likelihood ratio test between hypotheses and .
Lossy boson sampling.
In this variant^{35}, we assume that n − k photons are lost before entering the circuit enacting the linear optical transfer matrix, so k photons remain. Probabilities in this setting are not determined directly by Per(A_{S})^{2} for some mtuple S, but by the average of this quantity over all possible ways of losing n − k photons from n photons:
where Λ_{n, k} is the set of ksubsets of {1, …, n} and A_{S, T} is the k × k submatrix of A obtained by taking columns of A according to T and rows of A according to S, which remains a subset of {1, …, m}. Note that once again we restrict to the collisionfree subspace, making the assumption that the probability of a collision is low enough that this does not significantly affect the probabilities.
If, at most, a constant number of photons in total are lost before entering the circuit, the lossy boson sampling problem remains hard^{35}. Open problems are whether this is true in the more realistic setting where a constant fraction of photons are lost, and to generalize the loss model to include loss within and after the linear optical circuit. Here we prove a slightly stronger result than that shown in ref. 42; that as long as the overall transfer matrix is proportional to a unitary, loss can always be considered at the input even if the physical loss channels, wherever they occur, are not uniform (that is, are modedependent).
Consider a boson sampling device consisting of an ideal unitary linear optical transformation U on a set of m optical modes which is preceded or succeeded by pathindependent loss. This loss can be modelled by considering a set of m additional virtual ancilla modes such that the optical transfer matrix on all 2m modes remains unitary. A uniform transmission probability of η can then be described by beamsplitters coupling each mode to its corresponding ancilla, resulting in the transfer matrix
and so including the interferometer, the full transfer matrices for input and output losses are
Any mmode optical state can be expressed in a coherent state basis^{43}:
where α〉 = (α_{1}, α_{2}, …, α_{m})^{T}〉 ≡ ⊗ _{i=1}^{m} α_{i}〉 is an mmode coherent state and 〈β = 〈(β_{1}, β_{2}, …β_{m})^{T} = ⊗ _{i=1}^{m}〈β_{i}. A coherent state evolves under a transfer matrix T as
It can then be shown that when the initial state contains vacuum in all ancilla modes, ρ = ρ_{m} ⊗ 0〉〈0_{an}, the same state is produced in the m system modes under the transformations M_{O} and M_{I}
More generally, wherever loss occurs in the experiment, the overall transfer matrix K on the system modes can be efficiently characterized^{44,45}. Since pathdependent loss is usually small in experiments, and can be mitigated by interferometer design^{46}, the matrix K/K_{2} ≈ U. The matrix K can then be embedded into a larger unitary matrix acting on additional modes as before. We note all unitary dilations of K, M_{K} ∈ U(m + p), where p ≥ m, can be parameterized using the Cosine–Sine decomposition as
where K = Acos(Θ)B^{†}, with A, B ∈ U(m) and cos(Θ) = diag(cosθ_{1}, …, cosθ_{m}) with θ_{1} ≤ θ_{2} ≤ … ≤ θ_{m}, is a singular value decomposition of K and X, Y ∈ U(p). In fact, all unitary dilations are related by the choice of X and Y (refs 47,48). Since and the choice of X does not affect ρ_{K} using the cyclic property of the trace as above, setting η = K_{2}^{2}, we see that ρ_{K} = ρ_{I}. Moreover, we have shown that all unitary dilations of a transfer matrix produce the same output state, and therefore any boson sampling experiment with overall pathindependent losses is equivalent to introducing uniform loss channels with transmission probability η at the input, followed by the ideal unitary evolution.
Our MIS method can readily be adapted to deal with loss at the input, by inserting an initial step for each tuple to be output, which generates a uniformly random input subset T. This would be followed by the usual MIS method with permanents of k × k submatrices computed. The core part of the classical sampling procedure for both the lossy and scattershot variants therefore follows precisely that of standard boson sampling. From our analysis of the required burn in period for MIS (see Supplementary Information), we can see that the performance of our sampler will be similar to the standard boson sampling case. That is, it is likely that lossy and scattershot boson sampling is no more difficult classically than standard boson sampling.
Data availability.
The data that support the plots within this paper and other findings of this study are available at the University of Bristol data repository, data.bris, at https://doi.org/10.5523/bris.2ok605tzyel9o20lpmz2kky7wd.
Additional Information
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
 1
Aaronson, S. & Arkhipov, A. Proc. 43rd Annual ACM Symp. Theory Comput. The computational complexity of linear optics. 333–342 (ACM, 2011).
 2
Broome, M. A. et al. Photonic boson sampling in a tunable circuit. Science 339, 794–798 (2013).
 3
Tillmann, M. et al. Experimental boson sampling. Nat. Photon. 7, 540–544 (2013).
 4
Spring, J. B. et al. Boson Sampling on a Photonic Chip. Science 339, 798–801 (2013).
 5
Crespi, A. et al. Integrated multimode interferometers with arbitrary designs for photonic boson sampling. Nat. Photon. 7, 545–549 (2013).
 6
Lund, A. P., Bremner, M. J. & Ralph, T. C. Quantum sampling problems, bosonsampling and quantum supremacy. npj Quant. Inf. 3, 15 (2017).
 7
Latmiral, L., Spagnolo, N. & Sciarrino, F. Towards quantum supremacy with lossy scattershot boson sampling. New J. Phys. 18, 113008 (2016).
 8
Wang, H. et al. Highefficiency multiphoton boson sampling. Nat. Photon. 11, 361–365 (2017).
 9
Shor, P. W. Polynomialtime algorithms for prime factorization and discrete logarithms on a quantum computer. SIAM J. Comput. 26, 1484–1509 (1997).
 10
Politi, A., Cryan, M. J., Rarity, J. G., Yu, S. & O’Brien, J. L. Silicaonsilicon waveguide quantum circuits. Science 320, 646–649 (2008).
 11
Carolan, J. et al. Universal linear optics. Science 349, 711–716 (2015).
 12
Harris, N. C. et al. Quantum transport simulations in a programmable nanophotonic processor. Nat. Photon. 11, 447–452 (2017).
 13
Ding, X. et al. Ondemand single photons with high extraction efficiency and nearunity indistinguishability from a resonantly driven quantum dot in a micropillar. Phys. Rev. Lett. 116, 020401 (2016).
 14
Somaschi, N. et al. Nearoptimal singlephoton sources in the solid state. Nat. Photon. 10, 340–345 (2016).
 15
Spring, J. B. et al. Chipbased array of nearidentical, pure, heralded singlephoton sources. Optica 4, 90–96 (2017).
 16
Lita, A. E., Miller, A. J. & Nam, S. W. Counting nearinfrared singlephotons with 95% efficiency. Opt. Express 16, 3032–3040 (2008).
 17
Hadfield, R. H. Singlephoton detectors for optical quantum information applications. Nat. Photon. 3, 696–705 (2009).
 18
Bremner, M. J., Montanaro, A. & Shepherd, D. J. Achieving quantum supremacy with sparse and noisy commuting quantum computations. Quantum 1, 8 (2017).
 19
Boixo, S. et al. Characterizing quantum supremacy in nearterm devices. Preprint at https://arxiv.org/abs/1608.00263 (2016).
 20
BermejoVega, J., Hangleiter, D., Schwarz, M., Raussendorf, R. & Eisert, J. Architectures for quantum simulation showing quantum supremacy. Preprint at https://arxiv.org/abs/1703.00466 (2017).
 21
Preskill, J. Quantum computing and the entanglement frontier. Preprint at https://arxiv.org/abs/1203.5813 (2012).
 22
Bentivegna, M. et al. Experimental scattershot boson sampling. Sci. Adv. 1, e1400255 (2015).
 23
Wu, J. et al. Computing permanents for boson sampling on Tianhe2 supercomputer. Preprint at https://arxiv.org/abs/1606.05836 (2016).
 24
Liu, J. Metropolized independent sampling with comparisons to rejection sampling and importance sampling. Stat. Comput. 6, 113–119 (1996).
 25
Aaronson, S. & Arkhipov, A. Boson sampling is far from uniform. Quantum. Inf. Comp. 14, 1383–1432 (2014).
 26
Ryser, H. J. Combinatorial Mathematics (Carus Mathematical Monographs, No. 14, 1963).
 27
Bentivegna, M. et al. Bayesian approach to boson sampling validation. Int. J. Quantum Inf. 12, 1560028 (2014).
 28
Praestgaard, J. T. Permutation and bootstrap Kolmogorov–Smirnov tests for the equality of two distributions. Scand. J. Stat. 22, 305–322 (1995).
 29
Rohde, P. P., Motes, K. R., Knott, P. A. & Munro, W. J. Will bosonsampling ever disprove the extended Church–Turing thesis? Preprint at https://arxiv.org/abs/1401.2199 (2014).
 30
Arkhipov, A. & Kuperberg, G. The bosonic birthday paradox. Geom. Topol. Monogr. 18, 1–7 (2012).
 31
Spagnolo, N. et al. Experimental validation of photonic boson sampling. Nat. Photon. 8, 615–620 (2014).
 32
Carolan, J. et al. On the experimental verification of quantum complexity in linear optics. Nat. Photon. 8, 621–626 (2014).
 33
Loredo, J. C. et al. Boson sampling with singlephoton Fock states from a bright solidstate source. Phys. Rev. Lett. 118, 130503 (2017).
 34
He, Y. et al. Timebinencoded boson sampling with a singlephoton device. Phys. Rev. Lett. 118, 190501 (2017).
 35
Aaronson, S. & Brod, D. J. Boson sampling with lost photons. Phys. Rev. A 93, 012335 (2016).
 36
Russell, N. J., Chakhmakhchyan, L., O’Brien, J. L. & Laing, A. Direct dialling of Haar random unitary matrices. New J. Phys. 19, 033007 (2017).
 37
Motes, K. R., Gilchrist, A., Dowling, J. P. & Rohde, P. P. Scalable boson sampling with timebin encoding using a loopbased architecture. Phys. Rev. Lett. 113, 120501 (2014).
 38
Pant, M. & Englund, D. Highdimensional unitary transformations and boson sampling on temporal modes using dispersive optics. Phys. Rev. A 93, 043803 (2016).
 39
Huh, J., Guerreschi, G. G., Peropadre, B., McClean, J. R. & AspuruGuzik, A. Boson sampling for molecular vibronic spectra. Nat. Photon. 9, 615–620 (2015).
 40
Hamilton, C. S. et al. Gaussian boson sampling. Preprint at https://arxiv.org/abs/1612.01199 (2016).
 41
Barvinok, A. I. Two algorithmic results for the traveling salesman problem. Math. Oper. Res. 21, 65–84 (1996).
 42
Berry, D. W. & Lvovsky, A. I. Linearoptical processing cannot increase photon efficiency. Phys. Rev. Lett. 105, 203601 (2010).
 43
Gerry, C. & Knight, P. Introductory Quantum Optics (Cambridge Univ. Press, 2005).
 44
Laing, A. & O’Brien, J. L. Superstable tomography of any linear optical device. Preprint at https://arxiv.org/abs/1208.2868 (2012).
 45
RahimiKeshari, S. et al. Direct characterization of linearoptical networks. Opt. Express 21, 13450–13458 (2013).
 46
Clements, W. R., Humphreys, P. C., Metcalf, B. J., Kolthammer, W. S. & Walmsley, I. A. Optimal design for universal multiport interferometers. Optica 3, 1460–1465 (2016).
 47
Horn, R. A. & Johnson, C. R. Topics in Matrix Analysis (Cambridge Univ. Press, 1991).
 48
Allen, J. & Arceo, D. Matrix Dilations via CosineSine Decomposition Tech. Rep., DTIC Document (SCC, 2006).
Acknowledgements
The authors would like to thank N. Maraviglia and P. Shadbolt for helpful discussions. A.N. is grateful for support from the Wilkinson Foundation. This work was supported by the Engineering and Physical Sciences Research Council (EPSRC) and QUCHIP (H2020FETPROACT32014: Quantum simulation). Fellowship support from EPSRC is acknowledged by R.C. (EP/J019283/1), A.M. (EP/L021005/1) and A.L. (EP/N003470/1).
Author information
Affiliations
Contributions
All authors contributed to discussions and project development. Simulations were carried out by A.N. Data were analysed by A.N., C.S., R.C., A.M. and A.L. The MIS algorithm and the Random Restart HillClimbing routine used in our RS algorithm are by R.C. Additional classical sampling algorithms were designed and implemented by A.N. and R.C., with input from C.S., A.M. and A.L. Code optimization was carried out by E.J. The proof that photon loss can be considered at the input is by P.M.B. The manuscript was written by A.L., A.N. and C.S., with input from R.C. and A.M. The project was conceived and managed by A.L.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing financial interests.
Supplementary information
Supplementary information
Supplementary information (PDF 1056 kb)
Rights and permissions
About this article
Cite this article
Neville, A., Sparrow, C., Clifford, R. et al. Classical boson sampling algorithms with superior performance to nearterm experiments. Nature Phys 13, 1153–1157 (2017). https://doi.org/10.1038/nphys4270
Received:
Accepted:
Published:
Issue Date:
Further reading

Sample caching Markov chain Monte Carlo approach to boson sampling simulation
New Journal of Physics (2020)

Regimes of Classical Simulability for Noisy Gaussian Boson Sampling
Physical Review Letters (2020)

QuantumTeleportationInspired Algorithm for Sampling Large Random Quantum Circuits
Physical Review Letters (2020)

Distinguishability theory for timeresolved photodetection and boson sampling
Physical Review A (2020)

Nearideal spontaneous photon sources in silicon quantum photonics
Nature Communications (2020)