Abstract
Cryptography’s importance in our everyday lives continues to grow in our increasingly digital world. Oblivious transfer has long been a fundamental and important cryptographic primitive, as it is known that general twoparty cryptographic tasks can be built from this basic building block. Here we show the experimental implementation of a 12 random oblivious transfer protocol by performing measurements on polarizationentangled photon pairs in a modified entangled quantum key distribution system, followed by all of the necessary classical postprocessing including oneway error correction. We successfully exchange a 1,366 bit random oblivious transfer string in ~3 min and include a full security analysis under the noisy storage model, accounting for all experimental error rates and finite size effects. This demonstrates the feasibility of using today’s quantum technologies to implement secure twoparty protocols.
Similar content being viewed by others
Introduction
Cryptography, before the modern computing age, was synonymous with the protection of private communications using a variety of encryption techniques, such as a wax seal or substitution cypher. However, with the advent of modern digital computing and an increasingly internetdriven society, important new cryptographic challenges have arisen. People now wish to do business and interact with others they neither know nor trust. In the field of cryptography this is known as secure twoparty computation. Here, we have two users, Alice and Bob, who wish to perform a computation on their private inputs in such a way that they obtain the correct output but without revealing any additional information about their inputs.
A particularly important and familiar example is the task of secure identification, which we perform any time we use a bank’s ATM to withdraw money. Here, honest Alice (the bank) and honest Bob (the legitimate customer) share a password. When authenticating a new session, Alice checks to make sure she is really interacting with Bob by validating his password before dispensing any money. However, we do not want Bob to simply announce his password since a malicious Alice could steal his password and impersonate him in the future. What we require is a method for checking whether Bob’s password is valid without revealing any additional information. While protocols for general twoparty cryptographic tasks such as this one may be very involved, it is known that they can be built from a basic cryptographic building block called oblivious transfer (OT)^{1}.
Many classical cryptography techniques currently in use have their security based on conjectured mathematical assumptions such as the hardness of finding the prime factors of a large number, assumptions that no longer hold once a sufficiently large quantum computer is built^{2}. Alternatively, quantum cryptography offers a means to accomplish cryptographic tasks that are provably secure using fewer assumptions that are ideally much more stringent than those employed classically. However, until now almost all of the experimental work has focused exclusively on quantum key distribution (QKD) , yet there are many other cryptographic primitives^{3,4,5,6} that can make use of quantum mechanics to augment their security. OT is a prime example.
While it has been shown that almost no cryptographic twoparty primitives, save for QKD, are secure if a quantum channel is available and no further restrictions are placed on an adversary^{7,8,9}, twoparty protocols are so pivotal to modern cryptography that we are required to explore scenarios that place realistic restrictions on an adversary, which allow provable security to be restored in important realworld settings. Moreover, if we can efficiently implement lowcost quantum cryptography while making it much harder and more expensive for adversaries to break it as compared with classical schemes, even if not provably secure without added assumptions, then we have still provided a benefit.
One can regain provable security for OT within the noisy storage model^{10,11} under the physical assumption that an adversary does not possess a large reliable quantum memory. This model most accurately captures the difficulty facing a potential adversary since most quantum memories currently suffer from one or more of the following problems: (1) the transfer of the photonic qubit into the physical system used for the memory is noisy; (2) the memory is unable to sufficiently maintain the integrity of the quantum information over time; or (3) the memory suffers from an inability to perform consistently (that is, photons are often lost causing the memory to act as an erasure channel).
With current technology, the quantum information stored in a quantum memory is lost within a few milliseconds^{12,13,14}. While there have been recent demonstrations of systems with coherence times on the order of seconds and even minutes^{15,16}, the highfidelity transfer of quantum information from another physical system into the memory system and its subsequent storage have not yet been demonstrated. Further, a reliable memory also requires faulttolerant error correction built into its architecture. Until these significant experimental challenges are met, security is assured under the noisy storage model. Even with the advent of practical quantum storage, for any storage size with a finite upper bound, security can still be achieved by simply increasing the number of qubits exchanged during the protocol. Moreover, the currently executed protocol holds secure even if a dishonest party obtains a better quantum memory in the future.
Because of its huge potential, securing OT with quantum means has recently received interest from the recent 1outofN OT experiment performed by Chan et al.^{17} However, we note that, in contrast to our implementation, Chan et al. achieve only a weak version of 1outofN OT in which the attacker gains a significant amount of information. In addition, continued theoretical work^{11} has been used to propose a number of different experimentally feasible protocols that could be implemented using today’s technology^{18,19}. Notably, quantum bit commitment secure under the noisy storage model was recently shown by Ng et al.^{20} And while it is indeed known that OT can in principle be built from secure bit commitment and additional quantum communication^{21}, the protocol for such a reduction is inefficient and has not been analysed in a setting where errors are present.
In this work, we show the first experimental implementation of OT secured under the noisy storage model using a modified entangled QKD system and all of the necessary classical postprocessing algorithms including oneway error correction. During a ~3min quantum and classical exchange, we generate 1,366 bits of secure random oblivious transfer (ROT) key accounting for all experimental error rates and finite size effects. Using a new minentropy uncertainty relation to derive much higher OT rates, we examine a number of tradeoffs and analyse the secure parameter regimes. This demonstrates the feasibility of using today’s quantum technologies to implement secure twoparty protocols, most notably the building block necessary to construct a secure identification scheme^{22} to securely authenticate oneself at an ATM one day.
Results
The OT protocol
In a 12 OT protocol, we have two parties, Alice and Bob. Alice holds two secret binary strings Ŝ_{0}, Ŝ_{1}∈{0,1}^{l} of length l. Bob wishes to learn one of these two strings, and the string he decides to learn is given by his choice bit C∈{0,1}. The protocol is called OT because of its security conditions: if Alice is honest, then in learning the string of his choice, Ŝ_{C}, Bob should learn nothing about Alice’s other string, Ŝ_{1−C}; while if Bob is honest, Alice should not be able to discern which string Bob chose to learn, that is, Alice should not learn C. Security is not required in the case where both parties are dishonest. The 12 refers to the fact that Bob learns one and only one of Alice’s two secret strings.
To implement 12 OT we actually first implement 12 ROT that is then converted into 12 OT with an additional step at the end. In the randomized protocol, rather than Alice choosing her two strings she instead receives the two secret binary strings S_{0}, S_{1}∈{0,1}^{l}, chosen uniformly at random, as an output of the protocol. After the ROT protocol is complete, Alice can use these random strings as onetime pads to encrypt her original desired inputs Ŝ_{0} and Ŝ_{1} and send them both to Bob. Bob can then recover Alice’s original intended string by using the S_{C} he learned during the ROT protocol to decrypt his desired Ŝ_{C} thus completing a 12 OT protocol.
For the protocol to hold correct when both parties are honest, it is crucial that Alice and Bob have devices satisfying a minimum set of requirements, for example, the loss and error rates have to be upper bounded. In the correctness proof, Alice and Bob’s devices are assumed not to exceed a maximum loss and error rate. It is these upper bounded values that will be used in the secure ROT string formula (equation 8) to determine the maximum allowable size for the generated ROT string. Whenever their devices fulfill these criteria, the protocol can be executed correctly. On the contrary, in the security proof, a dishonest party is assumed to be all powerful: in particular he/she can use perfect devices to eliminate losses and errors to use them as a cheating advantage instead. The only restriction on a cheating party in this model is that of the quantum memory accessible. We refer the reader to the formal security definition of this protocol in Schaffner^{19}.
Figure 1 outlines our 12 ROT protocol^{18,19}. A source of polarizationentangled photon pairs distributes one photon from each pair to Alice and Bob. In fact, Alice holds the entanglement source as it has been shown that this does not affect either correctness or security. Alice and Bob measure in one of the two usual BB84 bases, H/V (+) or +45°/−45° (×), using passive polarization detectors. For every photon they detect, they record their measurement basis as α_{i} and β_{j}, their bit value as X_{i} and Y_{j}, and their measurement time as t_{Ai} and t_{Bj}, respectively. Bob then sends his timing information to Alice so that she can perform a coincidence search allowing them to sift down to those measurement results where they both detected a photon from a pair. Alice then checks that the number of photons Bob detected falls within the secure interval allowing them to proceed. At this point, Alice obtains basis and bit strings α^{m} and X^{m}, while Bob obtains basis and bit strings β^{m} and Y^{m}, all of which are of length m.
Both parties now wait for a time, Δt, long enough for any stored quantum information of a dishonest party to decohere. Note that the information Bob sends to Alice after this point does not provide her information about his choice bit S_{C}, provided that Bob’s choice of basis is completely uniform. To ensure the uniformity of this choice, an honest Bob can perform symmetrization over all his detectors to make them all equally efficient^{20}, so that Alice cannot make a better guess of his choice bit than a random guess. Hence, it is really only Alice who is waiting here to protect against a dishonest Bob who might have been storing some of his photons. Once the secure wait time is observed, Alice sends Bob her basis information, α^{m}. This allows Bob to divide his bits into two subsets: YI_{C} , where his measurement basis matched Alice’s (α_{i}=β_{i}) leading eventually to the shared bit string S_{C}; and YI_{1−C}, where his measurement basis did not match Alice’s (α_{i}≠β_{i}), which leads to the second bit string S_{1−C} which Bob should know nothing about.
Next, Bob sends his index lists, I_{0} and I_{1}, to Alice so that she can partition her data identically. Note that it is Bob’s choice of labelling the subset where his measurement bases matched Alice’s as I_{C}, which allows him to eventually learn his desired S_{C} from Alice since this is the subset in which error correction will succeed. However, this does not reveal C to Alice since from her point of view, she always receives an I_{0} and I_{1} from Bob and there is never any information in the partitioning that would allow her to deduce C.
Continuing, Alice performs oneway error correction by encoding syndrome information for each of her two subsets, syn(XI_{0}) and syn(XI_{1}), which she then sends to Bob. Bob is able to use the syndrome information for the subset where his measurement bases matched Alice’s, YI_{C}, to correct any errors in his string to obtain his error corrected string y_{cor}. Note that Bob will only be able to correct his string if the devices and quantum channel are operating within their design parameters since this is what the oneway error correction code has been optimized for. If Bob is dishonest, then we do not have to worry about him being able to decode, the goal of the protocol is to ensure that honest Bob can learn his desired string.
The last step is to perform privacy amplification^{23,24}, which is a cryptographic technique that allows a situation of partial ignorance to be turned into a situation of (essentially) complete ignorance. For example, suppose Alice has a long string (here XI_{1−C}) which is difficult, but not impossible, for Bob to guess. That is, Bob’s minentropy about the string XI_{1−C} is large. Applying a randomly chosen twouniversal hash function f_{1−C} to XI_{1−C} allows Alice to obtain a (typically much shorter) string S_{1−C}=f_{1−C}(XI_{1−C}) that is essentially impossible for Bob to guess. That is, if the length of the short string is l, then Bob’s guessing probability is very close to 1/2^{l}, which means that Bob has learned nothing about the short string. The only cost to Alice for performing this is to reduce the size of her output strings somewhat according to equation (8).
To perform privacy amplification, Alice and Bob use the secure ROT string rate formula (equation 8) and the estimates of the error rate and loss due to their devices and the quantum channel to calculate the length of extractable ROT string that they can keep after privacy amplification. Alice applies the twouniversal hash functions f_{0} and f_{1} to her two substrings XI_{0} and XI_{1} and obtains the shorter strings S_{0}=f_{0}(XI_{0}) and S_{1}=f_{1}(XI_{1}) as her private outputs, while Bob applies the appropriate twouniversal hash function to his subset y_{cor} obtaining his desired S_{C}=f_{C}(y_{cor}) as his private output. Privacy amplification is extremely important since Bob has potentially gained a significant amount of information about Alice’s second string, which a secure protocol requires he know nothing about. This extra information has come from the syndrome information that Alice sent for her second subset, syn(XI_{1−C}), where their measurement bases did not match and from Bob’s ability to store partial quantum information and attempt to cheat. In our proof, we show that XI_{1−C} has a high minentropy has a high minentropy conditioned on Bob's system and hence the output S_{1−C} is close to uniform after privacy amplification. Of course, dishonest Bob could use the error correction information to correct errors in his own storage device or he could open his device, remove all its imperfections, and then perform a partial attack to gain extra information without raising the error rate above the allowed limit. Regardless, by subtracting the entire length of the error correction information from the initial extractable ROT string Alice can extract a shorter string S_{1−C} via privacy amplification, which looks uniform to Bob, ensuring the security of the protocol.
As the above protocol has just summarized, even though breaking the security of our protocol requires a large quantum memory with long storage times, neither a quantum memory nor the ability to perform quantum computations are needed to actually run the protocol. Thus, as we shall see below, the technological requirements for honest parties are comparable to QKD and hence well within the reach of current technology.
Experimental parameters
It is important to realize that the techniques used here to prove the security of ROT are fundamentally different than those used in proving the security of QKD. Contrary to QKD, there is no realtime parameter estimation which needs to be performed. Instead, the two parties estimate the parameters before performing the OT protocol. The only requirement is that it should be possible for honest parties to bring certain hardware to execute the protocol. Dishonest parties can have arbitrary devices and still security will be assured based on the following:

1
Alice holds the source and can completely characterize it. Thus, if Alice is honest she can rely on the source following the estimated parameters. If she is dishonest, she is allowed to replace the source with an arbitrary quantum device (it could be a full quantum computer with arbitrary storage since the storage assumption is only needed to deal with a dishonest Bob).

2
Honest Bob can always test his device himself without relying on Alice.

3
The channel parameters (that is, the loss and error rate) can be estimated jointly by Alice and Bob. Depending on their estimate, Alice decides how much error correction information she needs to send. If Bob was dishonest and lied during the estimate then one of the following scenarios happens (both of which are secure): (a) Alice finds there is no n (where n is the number of entangled photon pairs exchanged before losses) such that secure OT can be performed in which case she either demands Bob get a better device or for the two of them to invest in a better channel. (b) Alice tunes n such that security can be obtained for their estimated parameters. Dishonest Bob could have lied and later eliminated his losses and errors for the OT exchange; however, this is already accounted for in the security analysis that assumes a dishonest Bob can have a perfect channel and perfect devices (his only limitation is the memory bound).

4
If the parameters turn out to be different for the honest parties during the protocol (for example, the errors turn out to be much higher) compared to what was estimated earlier, then the protocol may not succeed. Hence, once the parameters are estimated and fixed, the honest parties need to use hardware satisfying these parameters if they want to perform the protocol correctly. However, security is not affected by the actual amount of losses and errors.
To evaluate security, we model our parametric downconversion (PDC) entangled photon source in the standard way^{25} by measuring the mean photon pair number per pulse (μ), which is directly related to the amplitude of the pump laser. Since it is a continuously pumped source, we define our pulse length as the coherence time of our laser. Three other measured parameters are also required: the total transmission efficiency or transmittance, η, the intrinsic detection error rate of the system, e_{det}, and the probability of a dark count in one of Alice’s or Bob’s detectors, p_{dark}. The detection error is the probability that a photon sent by the source causes a click in the wrong detector, which can happen due to deficiencies or misalignments in Alice and Bob’s equipment or due to processes in the quantum channel. For the dark count probability, Alice and Bob take the value of their worst detector. Note that the parameters measured are necessary for allowing correctness of the protocol between two honest parties. In light of this, Alice and Bob perform a device and channel characterization and use these estimates in all subsequent security checks of the system. In the security analysis, a malicious party is assumed to have full power over their devices, such as eliminating all losses and noise in the communication while tricking the honest party to believe otherwise. Their values are summarized in the top of Table 1.
With these definitions we compute a number of probabilities needed for the security statements, which are conditioned on the event that Alice observed a single click, namely: the probability that exactly one entangled photon pair is emitted from the source (all double pair emissions are assumed to give an adversary full information), the probability an honest Bob did not detect a photon and the probability that a dishonest Bob did not detect a photon. Note that due to dark counts in Alice’s detectors. Thus, even if a dishonest Bob removes all loss in the quantum channel and uses perfect detectors, there will still be pairs registered by Alice (she cannot differentiate between dark counts and valid detections of photons from entangled pairs) which a dishonest Bob misses. These probabilities are derived from our PDC model following Wehner et al.^{18} and Schaffner^{19} using the parameters given in Table 1 (more details can be found in the Methods section).
Correctness
Whenever Alice and Bob are both honest, we desire that the protocol runs correctly even in the presence of experimental imperfections except with some failure probability. To quantify this probability, Alice and Bob can beforehand agree on some correctness error, ε >0, which will be used to lower bound the failure probability of this protocol. This parameter can also be seen as a bound on the maximum allowable fluctuations observed during the protocol allowing one to form acceptable intervals around the expected values for , and wherein correctness will hold. For example, the acceptable interval for p_{B, noclick} is given by with
given by invoking Hoeffding’s inequality^{26}. According to Hoeffding’s inequality, the number of detected rounds fall out of this interval with probability <2ε. For our experiment, we chose a correctness error of ε=2.5 × 10^{−7}. Note that in our analysis, there is a correctness error and a security error (discussed in the next section) that we represent by the same variable, ε, since they take the same value, but they could in general be different.
During the protocol, Alice checks whether the number of rounds reported as lost by Bob lies outside the interval that would be expected based on the parameters estimated before starting the protocol. We emphasize that this test is not part of the parameter estimation, but rather ensures that the absolute number of rounds reported lost by Bob is limited. Intuitively, this is a critical step since it prevents a dishonest Bob, who has a perfect channel, from using the fact that he can report rounds as lost to discard some or all of the single photon rounds (for which he only potentially gains partial information) in favour of multiphoton rounds (where he gains full information). Thus, the step prevents a dishonest Bob from performing the equivalent of the wellknown photonnumbersplitting attack from QKD^{27} in our protocol. For more details, please refer to Section III of Wehner et al.^{18}
Error correction is also necessary to correct errors due to the quantum channel, so that Bob can faithfully recover S_{C}. Error correction must be done with a oneway forward error correction protocol to maintain the security of the protocol. The errorcorrecting code is chosen such that Bob can decode faithfully except with a probability at most ε_{EC}. Thus, our ROT protocol will succeed except with probability 2ε+ε_{EC}. For our error correction code, we have found an (empirically tested) upper bound to the error correction failure probability of ε_{EC}≤3.09 × 10^{−3}. Thus, the total correctness error is upper bounded by 2ε+ε_{EC}≈0.0031. Note that while the decoding error has a large effect on the total correctness error of the protocol, it does not affect the security error of the protocol. In particular, the security error can be much lower than the correctness error, which we will see in the next section.
Security
In this section, we show that the security error can be made small, that is, the protocol holds secure against dishonest parties except with error 3ε=7.5 × 10^{−7}. Note that in our analysis, we first fix the security error, ε, and then derive the corresponding rate of OT. Thus, just like in QKD, one can reduce ε at the expense of shortening the final OT string if a particular situation calls for a smaller security error.
To evaluate security for honest Alice, we need to prove that Bob does not know at least one of the extracted strings. This is done by quantifying Bob’s knowledge about the string X^{m} that Alice has. The minentropy serves as a suitable quantification in such cases, and is defined as
where p_{guess}(XBob) is the probability of Bob guessing X correctly, maximized over all possible measurements he can perform on his system. The minentropy can also be understood operationally as the number of bits extractable from X, which appear random to Bob^{28}. To allow for imperfect states, we use the smoothed minentropy that is the minentropy maximized over all joint states of close (in terms of purified distance) to the original state ρ_{X, Bob}, where ρ_{X, Bob} is the joint state of Alice’s variable X and Bob’s entire system.
Note that a dishonest Bob’s strategy, upon receiving information from Alice, can in general be as follows: he performs some arbitrary encoding upon the received state, and stores the state into some quantum memory (possibly retaining some additional classical information). After receiving Alice’s basis information, he utilizes the stored qubits to perform strategic measurements again, gaining more classical information from his measurement outcome. Therefore, the main goal is to show that given all of Bob’s information his guessing probability of X^{m} is still low, that is, Bob’s minentropy is lower bounded.
To do so, we employ the minentropy uncertainty relation recently derived by Ng et al.^{29}, given by
where α^{m} is the vector of measurement bases of an honest Alice, K represents some arbitrary classical side information, while c_{BB84} is a constant that depends solely on the measurements and arises due to bounding a family of Rényi entropies. The quantity is a lower bound (including finite size effects) for the number of single photon pair rounds exchanged^{18}, except with probability ε. We use this to bound the minimum amount of uncertainty Bob has about Alice’s overall string X^{m}. The value of c_{BB84} is given by
The bound on the minentropy in equation (3) also implies that, for security to hold at all we require
to generate a positive m_{1} and by extension, a positive minentropy. This condition provides the regime of parameters, specifically with respect to μ and η, wherein secure ROT can be performed.
Subsequently, we further bound Bob’s minentropy conditioned on his quantum memory. ROT has been proven secure against adversaries whose memories satisfy the strong converse property. This states that the success probability of decoding a randomly chosen nbit string sent through the quantum memory decays exponentially for rates above the classical memory capacity. Recent theoretical results also show that this can be refined by linking security to the entanglement cost^{30} and quantum capacity^{31} of the memory, instead of the classical capacity. To explicitly evaluate security, we follow Wehner et al.^{18} and Schaffner^{19} and model the memory as a ddimensional depolarizing channel, given by
which successfully transmits the state ρ with probability r and otherwise replaces it with the completely mixed state with probability 1−r, as the typical action of a quantum memory specializing to the case of qubits (that is, d=2). A memory has one other pertinent parameter we need; namely, a storage rate, ν, which represents the fraction of qubits that an adversarial Bob can store in memory. Intuitively, the noise (1−r) and storage rate (ν) are related since increasing the amount of quantum error correction in the memory would decrease the noise at the cost of using more memory qubits for error correction thus decreasing the storage rate. However, for this analysis we leave them as two independent parameters. The choices for r and ν in the middle of Table 1 are made for an assumed future quantum storage device subjected to depolarizing noise that retains the state with probability r=0.75. The storage rate ν=0.002 implies we assume that a dishonest party cannot store more than ν n≈1.600 × 10^{5} of all the n=8.00 × 10^{7} qubits. Note that our experiment is quite secure as all current memories decohere after ~1 s and can only store a handful of photons.
There is a second condition necessary for security to hold given by
where C_{N} is the classical capacity of the memory (N). This implies that the classical capacity of the total quantum memory (assumed to satisfy the strong converse property) should be strictly less than the total minentropy of Bob’s knowledge of X^{m}, including some finite size effects.
Finally, we can explore the ROT rate formula^{19,29} that tells Alice and Bob how many secure ROT bits they can keep from the privacy amplification operation. It is given by
where γ^{N} is the strong converse parameter of an adversary’s memory (N) that essentially gives the classical memory capacity, is the rate at which a dishonest Bob would need to store quantum information, is the number of rounds where Alice and Bob both measured a photon from the pair, f is the error correction efficiency relative to the Shannon limit, and p_{err} is the total probability of an error between Alice’s and Bob’s subset where their basis choices matched (including dark counts, their intrinsic detector error (e_{det}), and any errors induced by the quantum channel). It is instructive to think of the three terms in equation 8 in the following way. The first term represents the amount of uncertainty Bob has over at least one of Alice’s strings, regardless of the classical/quantum information he has. This is also the maximum ROT string that could be produced. From this, one has to subtract the second term representing the information potentially leaked to an adversarial party during error correction, just as one would in QKD. This quantity is exactly the length of the error correction syndrome information which (as mentioned earlier) we must subtract from Bob’s initial minentropy to ensure security holds. Finally, the third term represents a safety margin guaranteeing the maximum security failure probability.
We conclude, that security holds for an honest Alice except with probability 3ε=7.5 × 10^{−7}, where the error comes from the finite size effects due to the number of single photon rounds, from smoothing of the minentropy, and the error of privacy amplification. Note that the preceding security discussion used a simplified detector model and neglected vulnerabilities from a number of wellknown attacks in QKD^{32,33,34,35,36}, most of which target the detectors used.
Analysis of secure parameter regimes
The ROT rate is sensitive to four parameters: the depolarizing noise, r; the storage rate, ν; the intrinsic error rate of the system, e_{det}; and the transmittance, η. Further, the secure range of each of these individual parameters depends on the values of the others, although we emphasize that the sensitivity is due, in part, to the fact that we tried to obtain the maximum length of OT string compatible with our n, ε, and storage assumptions. Similarly to QKD, one could make security less sensitive by being less demanding about the output length of the OT string. Nonetheless, we examine here the two most important parameters that determined the final ROT string length; namely, the transmittance and the error rate.
Figure 2a shows the ROT rate versus the transmittance for error rates of 0.93% (red—this experiment), 1.75% (green) and 2.3% (blue) from which we can see why loss is so crucial to the security of ROT. The secure ROT rate quickly drops as the transmittance decreases; indeed, for higher error rate values the ROT rate can quickly become negative even for relatively large transmittances, for example, η~0.15. Already with our table top experiment, we experienced a total transmission efficiency of 1.5%, which just allowed us to get a positive ROT string length. Our transmittance and the corresponding ROT rate are shown by the magenta point on the far left of this graph. The situation only gets worse when one thinks about using the scheme in a distributed quantum network where loss will be even higher.
Just as importantly, Figure 2b shows the ROT rate versus the detection error, e_{det}, for transmittances of η=0.015 (red—this experiment), η=0.1 (green) and η=0.25 (blue). From our system’s transmittance of η=0.015 we can see that to get a positive ROT rate, we needed to observe an error rate of <0.954% over the course of the experiment. Fortunately for this experiment, we managed to decrease the error rate to 0.93%. Our error rate and the corresponding ROT rate are shown by the magenta point on this graph. While the constraint on the maximum allowable error rate does relax somewhat as the transmittance of the system increases, it never gets above a tolerable error rate of 2.38% even at transmittances as high as 25%. This represents an extremely experimentally challenging constraint, one that is much stronger than the typical safe QBER levels for QKD (~11%). This fact alone prevents most of the existing QKD systems in the world from being easily converted into a secure ROT implementation through a classical postprocessing software update alone. Each of these two constraints can be individually relaxed at the cost of more stringent conditions on the other. However, our hope is that future theoretical work on the security of ROT in the noisy storage model can improve these bounds.
Experiment
Before starting the experiment, all the pertinent parameters are estimated/selected and shown in Table 1 in order to allow Alice and Bob to evaluate the length of the ROT strings, given by equation (8), which they could subsequently extract from their data. Over the course of ~50 s, Alice and Bob measured 1,347,412 coincident detection events. With the photon pairs distributed, Alice verified that the number of Bob’s reported measurements fell within the secure interval. After waiting for a minimum time Δt=1 s for any stored quantum information of a dishonest party to decohere, Alice sent her basis measurement information (α) to Bob. Using his choice bit, C=0, Bob partitioned his data into YI_{C=0,} where his measurement basis matched Alice’s (that is, α_{i}=β_{i}) and YI_{1−C=1}, where his measurement basis did not match Alice’s (that is, α_{i}≠β_{i}), truncating both subsets to 600,000 bits (the block size needed for the error correction algorithm). He then sent his partitioning, I_{0} and I_{1}, to Alice.
Alice partitioned her data accordingly, encoded the syndrome information for each subset, and sent it to Bob. With the system operating within the design parameter requirements, Bob was successfully able to errorcorrect his subset of data, YI_{C=0}, where his measurement basis matched Alice’s in 2 min 14 s. Using the estimate of the error rate in equation 8, Alice and Bob calculated the size of their secure final ROT string to be 1,366 bits. To complete the protocol, Alice then chose two 2universal hash functions^{24}, f_{0}, f_{1}∈F, sent her choices to Bob, and calculated her private outputs S_{0}=f_{0}(xI_{0}) and S_{1}=f_{1}(xI_{1}) retaining the last 1,366 bits from the 2universal hash operation in each case. Having chosen C=0, Bob used f_{0} to compute S_{C}=f_{C}(y_{cor}) as his private output, obtaining the ROT string S_{0}, he desired from Alice.
Lastly, if Alice and Bob had wanted to remove the randomness from the protocol to implement OT with specific strings desired by Alice, she could have used S_{0} and S_{1} as onetime pads to encrypt her desired strings and sent the results to Bob. Using his S_{0}, he could then have decrypted Alice’s last communication, recovering his desired string from Alice.
Discussion
It is important to notice that ROT is a fundamentally different cryptographic primitive than QKD and represents an important new tool at our disposal. Using the example of securing one’s banking information at an ATM, it has long been suggested to use a small handheld QKD device to accomplish this^{37}. However, if one were to employ ROT to build a secure identification scheme, as opposed to QKD, one’s bank card number and PIN would never physically be exchanged over any communication line. Rather, the ATM and the user would merely be evaluating the joint equality function, f(x)=f(y), on each of their copies of the login information. This starkly illustrates the difference between ROT and QKD, as well as highlights ROT’s potential in certain situations.
The use of the new minentropy relation^{29} was crucial to making our experimental implementation practical since it could be applied to much smaller block lengths while taking into account finite size effects. To put it in perspective, estimates using the original analysis from Schaffner^{19} suggested we required n on the order of 10^{9}–10^{10} bits requiring Alice and Bob to measure photon pairs for over 4 h. Any realworld secure identification protocol, such as at an ATM, making use of such an ROT implementation would obviously be entirely impractical. However, by employing the new minentropy uncertainty relation we were able to generate positive OT lengths with an n as low as 6.65 × 10^{4}. In fact, the limiting factor in our protocol became the minimum block size necessary for our oneway error correction code to succeed with high probability at the required efficiency.
Though the current work is secure for memories which follow the strong converse property, many important channels exhibit this property; for instance, the depolarizing channel assumed for this work. Moreover, while a few quantum memories have already been shown to work with high fidelity, they are all postselected results. In other words, many photons are lost during their operation causing them to act as erasure channels. An adversary using one of these memories would find its action looking very close to a depolarizing channel since the best they could do is replace any lost photons with a random state. Moreover, recent theoretical results have shown this argument can be refined by linking the success probability of decoding to the entanglement cost^{30} and quantum capacity^{31} of the memory, instead of the classical capacity.
There are a number of shortcomings that should be addressed in future work. As shown earlier, the ROT string length is constrained drastically by the transmittance, η. In order for ROT to be useful over longer distances, for instance, over a future quantum communication network covering a city, the impact of loss on security has to be mitigated. One option could be to analyse the security proof for the case of an entangled photon source more carefully since a similar analysis for QKD has found that entangled systems can tolerate more loss than decoystate prepareandsend systems. Indeed, Wehner et al.^{18} themselves point out that they simplified their security proof for ROT in the case of an entangled source, giving all information in a double pair (and higher) emission to Bob. However, not only could the double pair emission rate be overestimated in the assumed PDC model, but as has been pointed out in connection with QKD, it is far from clear that double pair emissions give much, if anything, to an adversary. In fact, the probability of double pair emission is one of the key quantities limiting the acceptable error rate, transmittance and overall ROT rate; thus, there is likely much to be gained from a more detailed analysis. Another possibility to limit the impact of loss could be to find a secure version of the heralded noiseless amplification scheme of Kocsis et al.^{38} that allowed both Alice and Bob to trust its operation.
Smaller issues to make ROT more practical include relaxing the tolerable error rate to use ROT in a long distance quantum network setting. Even at short distances with the new minentropy relation^{29} requiring us to detect as few as 6.65 × 10^{4} entangled pairs in ~2.5 s, the extremely low error rate required makes it almost impossible to find an efficient error correction code at this block length.
Lastly, parameter estimation in the ROT protocol is fundamentally different than in QKD. Here, parameters are estimated before the protocol begins and then the security statements are evaluated using the agreedupon experimental parameters. This does not present any security problems for the reasons outlined in the Methods section; however, as a practical effect, if the parameters are poorly estimated then honest parties will have a difficult time correctly executing the protocol. Heuristically, it seems one would like to estimate the parameters for a long time such that the error bars on them are smaller than the intervals for Alice’s security checks. This is not a problem since it only needs to be done once in advance and will not affect the number of rounds in the actual ROT protocol. In the case of devices (for example, the source) which are not constant but have relatively slow fluctuations which would increase the error bars on the parameters during a long estimation, Alice and Bob could instead intersperse their ROT experiments with new parameter estimations. It is a very interesting open question for future work whether and how one could incorporate the error bars on the parameter estimates into the security proof.
In this work, we have shown the first experimental implementation of OT using a modified entangled QKD system and secured under the noisy storage model. We performed an updated security analysis using a new minentropy relation including finite statistics to show a drastic improvement in the secure ROT rate. We also examined the pertinent parameters that the ROT rate depends on, the most important being the transmittance and error rate. It will be very interesting to see whether future work on security proofs and experimental implementations can make the protocol even more practical in realworld scenarios.
Methods
Experimental implementation
The security proof for ROT guarantees security against adversaries with quantum memories and a quantum computer; however, the actual implementation of ROT for honest parties does not require these devices and is possible with today’s technology^{18,19}. Thus, we modified our existing entangled QKD system^{39,40} to implement the first ROT protocol secured under the noisy storage model. The key difference to keep in mind is that while loss merely affects the overall key rate in a QKD system, in ROT, loss is integral to the security of the scheme and if not properly bounded can prevent any secure ROT string from being generated.
A schematic of the system is shown in Fig. 3. Entangled photon pairs are produced with a Sagnac interferometric entangled photon source^{40,41,42,43} consisting of a periodically poled KTP (PPKTP) nonlinear optical crystal produced for a collinear configuration placed in the middle of an interferometer loop. Entangled photons are produced by sending 45° polarized light at 404 nm onto a dual wavelength polarizing beamsplitter at the apex of the loop, thus bidirectionally pumping the PPKTP crystal. A dual wavelength halfwaveplate (dual HWP) at 90° on one side of the loop properly polarizes the pump laser so that the crystal produces downconverted polarization correlated photons at 808 nm in both directions around the loop. The dual HWP also rotates the polarization of the downconverted photons travelling clockwise around the loop by 90° thus ensuring that when the photon pairs are split on the dual wavelength polarizing beamsplitter, their path information is erased and true polarizationentangled photon pairs are generated. The first half of the entangled photon pairs are collected directly from one port of the PBS, while the second half of the pairs are collected via a dichroic mirror responsible for removing residual photons from the pump laser.
Since the ROT protocol is useful even over very short distances, such as to securely identify oneself to an ATM machine over ~30 cm, Alice and Bob each measure their half of the entangled photon pairs while located next to the source connected to it with short singlemode optical fibres. The photons are measured with passive polarization detector boxes consisting of: a filter to reject the background light, a 50/50 nonpolarizing beamsplitter (BS) to perform the measurement basis choice, a PBS in the reflected arm of the BS to separate horizontally and vertically polarized photons and a HWP and PBS in the transmitted arm of the BS to separate photons polarized at +45° and −45°. Avalanche photodiode single photon detectors convert the photons into electronic pulses which are recorded with timetagging hardware which saves the measured polarization and detection time with a precision of 156.25 ps. The data is transferred to Alice’s and Bob’s laptops where custom written software then performs the rest of the ROT protocol including entangled photon pair identification (based on the detection times), security checking, sifting, error correction and twouniversal hashing of the final outputs.
While the majority of QKD systems, including ours, have used an interactive algorithm known as Cascade^{44} for error correction, we are not permitted to use it in ROT since the interaction would quickly reveal Bob’s choice bit, C, to Alice. Instead, we chose to implement oneway forward error correction using low density paritycheck (LDPC) codes^{45,46} updated for use in a QKD system^{47,48,49}. LDPC codes are gaining popularity in a number of QKD experiments^{47,48} since they can relieve much of the classical communication bottleneck that interactive error correction algorithms cause. However, there is an important difference between oneway error correction in QKD compared to ROT; namely, in a QKD protocol, error correction is permitted to fail on a block of key. In this case for QKD, Bob would let Alice know decoding failed on that block and they would have three options. They could either publicly reveal the block (in order to maintain an accurate estimate of the QBER) and then throw it away, they could try another oneway code optimized for a higher QBER, or they could revert to a twoway error correction procedure. We are not permitted to do any of these here as the error correction failure probability, ε_{EC}, is the failure probability for running error correction on the entire data set generated during the ROT protocol. A single failure of the error correction protocol requires Alice and Bob to start the protocol over.
In our LDPCbased error correction protocol, Alice and Bob share a large paritycheck matrix, H, which is constructed using a modified version of Hu et al.’s Progressiveedge growth software^{50} with known optimal degree distribution profiles^{51,52}. Alice computes syn(XI_{0}) and syn(XI_{1}) by treating her two bit strings as column vectors and applying H to them. Using the syndromes sent to him from Alice, Bob then employs an iterative message passing (or belief propagation) decoding algorithm known as the sum–product algorithm to correct any errors between his bit string YI_{C} and Alice’s XI_{C} (where their basis choices matched) with high probability. Our sum–product LDPC decoder is written in C# and is based on the Matlab code found in Chan^{49}. With a block size of N=600,000 (matrix dimensions were 68,000 × 600,000) and average row weight of 40, we achieved an error correction efficiency of 1.491. Our particularly low source error rate, required to securely implement the protocol, made it very challenging to find an efficient code and necessitated the large block sizes. This in turn made decoding particularly time intensive with a single block taking on average 2 min 14 s.
For the ROT experiment, the source was pumped with 7 mW of power as a compromise between a high pair rate and minimal error rate. Alice and Bob exchanged timing information in realtime to sort down to coincidence events. These measurements formed their raw strings (X and Y, respectively) combined with their record of measurement basis for each detection (α and β, respectively) and resulted in raw files 1.17 MB in size. It took Alice and Bob ~50 s to measure 1,347,412 coincident detection events which meant that a total of 8.00 × 10^{7} pairs (before losses) were distributed by the source. Due to slightly uneven detection probabilities and the inherent statistics of the source, Alice and Bob had to measure for slightly longer than the minimum time necessary to ensure that they measured 600,000 photon pairs with matching bases and 600,000 pairs with different bases. After error correction, Alice and Bob used an LFSR algorithm^{24}, capable of operating on the entire key at once, to perform privacy amplification. This is a key requirement to maintain security both in ROT and QKD, but one that is rarely, if ever, implemented in experimental and commercial QKD systems. All of the pertinent figures of merit from performing the ROT protocol are summarized in Table 2.
Experimental security parameters
The security proofs for ROT^{18,19} assumed that the states emitted by a PDC source can be written as^{25}
where the probability distribution is given by
and the states Φ_{n}_{AB} are given by
The states are written here in the basis H, V and μ/2 is directly related to the amplitude of the pump laser resulting in a mean photon pair number per pulse of μ. Using this assumption and a few other measured parameters, the authors of Wehner et al.^{18} and Schaffner^{19} derived all of the other necessary quantities used in their proofs of security.
Following this prescription, the first thing we calculate is the parameter μ, which is slightly more involved since we have a continuously pumped source which creates pairs at random times rather than the pulsed source assumed in the model. Using 7 mW of pump power, we measure Alice and Bob’s single photon count rates per second, N_{A} and N_{B}, as well as their coincident detection rate, N_{coin}, measured with a coincidence window of Δt_{coin}=3 ns (see Table 3). We then estimate Alice and Bob’s total transmittances (including their source coupling, polarization analyser and detectors) using the formula η_{A/B}=N_{coin}/N_{B/A} (ref. 53) which yields η_{A}=0.1374 for Alice and η_{B}=0.1092 for Bob. Using these numbers, we back out the loss and estimate the total number of pairs produced at the crystal as 3,145,331 pairs/s with the formula N=N_{A}/η_{A}. For all count rates, Poissonian error bars have been assumed, and Gaussian error propagation has been applied for the error bars on all derived quantities. Finally to calculate μ, we define our pulse length as the coherence time of our laser since the security statements use μ with equations (9, 10, 11) to determine a bound on the possible double pair emission contributions which might expose additional information to a dishonest party. It is well known with PDC sources that pairs separated by more than a coherence time are independent and thus would not give a dishonest party additional information. It is only those pairs generated within the same coherence time which might pose a security risk. Thus, with a coherence length of Δl_{c}=3 mm for our laser (iWave405S laser specifications, Toptica Photonics) and a corresponding coherence time of Δt_{c}=1.0 × 10^{−11} s, we can calculate μ using the formula
There are three other parameters that are necessary for the security statements: the total transmittance, η, which can be calculated from Fig. 3; the intrinsic error rate (QBER) of the system, e_{det} and the probability of obtaining a dark count in one of Alice’s or Bob’s detectors, p_{dark}. The darkcount probability is defined similarly to μ (that is, by multiplying the dark count rate per second by the coherence time), and is taken as the value of their worst detector. Strictly speaking, each detector will have a slightly different detection efficiency which will provide some partial information to the parties about the strings; however, this information can be removed by symmetrizing the losses of each user to their worst detector^{20} so that they become independent of the basis choice. All of these values are summarized in the top half of Table 1.
The security statements (equations 4, 5, 6, 7, 8) require the following important quantities: , which is the probability that only one photon pair is sent; , which is the probability that an honest Bob receives no click from a photon pair in his detector and , which is the minimum probability that a dishonest Bob receives no click from a photon pair in his detector. All of these are derived from the PDC model given by equation (9) with the experimental parameters given in Table 1. For the derivations, we refer the reader to Wehner et al.^{18}
Additional information
How to cite this article: Erven, C. et al. An experimental implementation of oblivious transfer in the noisy storage model. Nat. Commun. 5:3418 doi: 10.1038/ncomms4418 (2014).
References
Kilian, J. inProceedings of the 20th Annual ACM Symposium on Theory of Computing (STOC) 20–31ACM (1988).
Shor, P. inProceedings of the 35th Annual Symposium on the Foundations of Computer Science 124–134IEEE Computer Society (1994).
Wiesner, S. Conjugate coding. Sigact News 15, 78–88 (1983).
Hillery, M., Bužek, V. & Berthiaume, A. Quantum secret sharing. Phys. Rev. A 59, 1829–1834 (1999).
Gutoski, G. & Watrous, J. inProceedings of the 39th Annual ACM Symposium on Theory of Computing 565–574 San Diego, CA, USA, 2007.
Berlin, G. et al. Experimental losstolerant quantum coin flipping. Nat. Commun. 2, 561 (2011).
Mayers, D. Unconditionally secure quantum bit commitment is impossible. Phys. Rev. Lett. 78, 3414–3417 (1997).
Lo, H. & Chau, H. Is quantum bit commitment really possible? Phys. Rev. Lett. 78, 3410–3413 (1997).
Lo, H. Insecurity of quantum secure computations. Phys. Rev. A 56, 1154–1162 (1997).
Wehner, S., Schaffner, C. & Terhal, B. Cryptography from noisy storage. Phys. Rev. Lett. 100, 220502 (2008).
König, R., Wehner, S. & Wullschleger, J. Unconditional security from noisy quantum storage. IEEE Trans. Inf. Theory 58, 1962–1984 (2012).
Bao, X.H. et al. Efficient and longlived quantum memory with cold atoms inside a ring cavity. Nat. Phys. 8, 517–521 (2012).
Zhou, Z.Q., Lin, W.B., Yang, M., Li, C.F. & Guo, G.C. Realization of reliable solidstate quantum memory for photonic polarization qubit. Phys. Rev. Lett. 108, 190505 (2012).
Riedl, S. et al. Boseeinstein condensate as a quantum memory for a photonic polarization qubit. Phys. Rev. A 85, 022318 (2012).
Maurer, P. et al. Roomtemperature quantum bit memory exceeding one second. Science 336, 1283–1286 (2012).
Saeedi, K. et al. Roomtemperature quantum bit storage exceeding 39 minutes using ionized donors in silicon28. Science 342, 830–833 (2013).
Chan, P., LucioMartinez, I., Mo, X., Simon, C. & Tittel, W. Performing private database queries in a realworld environment using a quantum protocol. Preprint at <http://arxiv.org/abs/1303.0865> (2013).
Wehner, S., Curty, M., Schaffner, C. & Lo, H. Implementation of twoparty protocols in the noisystorage model. Phys. Rev. A 81, 052336 (2010).
Schaffner, C. Simple protocols for oblivious transfer and secure identification in the noisyquantumstorage model. Phys. Rev. A 82, 032308 (2010).
Ng, N. H., Joshi, S., Ming, C. C., Kurtsiefer, C. & Wehner, S. Experimental implementation of bit commitment in the noisystorage model. Nat. Commun. 3, 1326 (2012).
Yao, A. C.C. Security of quantum protocols against coherent measurements. inProceedings of the 27th Annual ACM Symposium on Theory of Computing (STOC) 67–75ACM (1995).
Damgård, I., Fehr, S., Salvail, L. & Schaffner, C. Secure identification and QKD in the boundedquantumstorage model. Lecture Notes Comput. Sci. 4622, 342–359 (2007).
Carter, J. & Wegman, M. Universal classes of hash functions. J. Comput. System Sci. 18, 143–154 (1979).
Krawczyk, H. Lfsrbased hashing and authentication. Lecture Notes Comput. Sci. 839, 129–139 (1994).
Kok, P. & Braunstein, S. Postselected versus nonpostselected quantum teleportation using parametric downconversion. Phys. Rev. A 61, 042304 (2000).
Uhlmann, W. Probability inequalities for sums of bounded random variables. J. Am. Stat. Assoc. 58, 13–30 (1963).
Hwang, W.Y. Quantum key distribution with high loss: toward global secure communication. Phys. Rev. Lett. 91, 057901 (2003).
Renner, R. Security of quantum key distribution. Int. J. Quant. Inf. 6, 1–127 (2008).
Ng, N., Berta, M. & Wehner, S. Minentropy uncertainty relation for finitesize cryptography. Phys. Rev. A 86, 042315 (2012).
Berta, M., Fawzi, O. & Wehner, S. Quantum to classical randomness extractors. Lecture Notes Comput. Sci. 7417, 776–793 (2012).
Berta, M., Christandl, M., Brandao, F. & Wehner, S. in2012 IEEE International Symposium on Information Theory Proceedings (ISIT) 900–904 (Cambridge, MA, USA, 2012).
Gerhardt, I. et al. Fullfield implementation of a perfect eavesdropper on a quantum cryptography system. Nat. Commun. 2, 349 (2011).
Lydersen, L., Akhlaghi, M., Majedi, H., Skaar, J. & Makarov, V. Controlling a superconducting nanowire singlephoton detector using tailored bright illumination. New J. Phys. 13, 113042 (2011).
Markarov, V. Controlling passively quenched single photon detectors by bright light. New J. Phys. 11, 065003 (2009).
Qi, B., Fung, C., Lo, H. & Ma, X. Timeshift attack in practical quantum cryptosystems. Quant. Info. Comput. 7, 73–82 (2007).
Makarov, V., Anisimov, A. & Skaar, J. Effects of detector efficiency mismatch on security of quantum cryptosystems. Phys. Rev. A 74, 022313 (2006).
Duligall, J., Godfrey, M., Harrison, K., Munro, W. & Rarity, J. Low cost and compact quantum key distribution. New J. Phys. 8, 249 (2006).
Kocsis, S., Xiang, G., Ralph, T. & Pryde, G. Heralded noiseless amplification of a photon polarization qubit. Nat. Phys. 9, 23–28 (2013).
Erven, C., Couteau, C., Laflamme, R. & Weihs, G. Entangled quantum key distribution over two freespace optical links. Opt. Express 16, 16840–16853 (2008).
Erven, C., Hamel, D., Resch, K., Laflamme, R. & Weihs, G. inQuantum Communication and Quantum Networking, vol. 36 of Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering (eds Akan O.et al. )108–116Springer (2010).
Kim, T., Fiorentino, M. & Wong, F. Phasestable source of polarizationentangled photons using a polarization sagnac interferometer. Phys. Rev. A 73, 012316 (2006).
Fedrizzi, A., Herbst, T., Poppe, A., Jennewein, T. & Zeilinger, A. A wavelengthtunable fibercoupled source of narrowband entangled photons. Opt. Express 15, 15377–15386 (2007).
Predojević, A., Grabher, S. & Weihs, G. Pulsed sagnac source of polarization entangled photon pairs. Opt. Express 20, 25022–25029 (2012).
Brassard, G. & Salvail, L. Secretkey reconciliation by public discussion. Lecture Notes Comput. Sci. 765, 410–423 (1994).
Gallager, R. Low density parity check codes. IRE Trans. Info. Theory IT8, 21–28 (1962).
MacKay, D. & Neal, R. Near shannon limit performance of low density parity check codes. Electron. Lett. 33, 457–458 (1997).
Pearson, D. Highspeed QKD reconciliation using forward error correction. QCMC 299–302 (2004).
Martinez, I., Chan, P., Mo, X., Hosier, S. & Tittel, W. Proofofconcept of realworld quantum key distribution with quantum frames. New J. Phys. 11, 095001 (2009).
Chan, P. LowDensity ParityCheck Codes for Quantum Key Distribution Master’s thesis, Univ. Calgary (2009).
Hu, X., Eleftheriou, E. & Arnold, D. Regular and irregular progressive edgegrowth Tanner graphs. IEEE Trans. Inf. Theory 51, 386–398 (2005).
Elkouss, D., Leverrier, A., Alléaume, R. & Boutros, J. inProceedings of the 2009 IEEE International Conference on Symposium on Information Theory—Volume 3, ISIT'09 1879–1883IEEE (2009).
Mateo, J. Efficient Information Reconciliation for Quantum Key Distribution PhD thesis, Univ. Politecnica de Madrid (2011).
Brida, G., Genovese, M. & Novero, C. An application of twophoton entangled states to quantum metrology. J. Mod. Opt 47, 2099–2104 (2000).
Acknowledgements
Support for this work by NSERC, QuantumWorks, CIFAR, CFI, CIPI, ORF, ORDCF, ERA, ERC (Project "EnSeNa"), and the Bell family fund is gratefully acknowledged. N.N. and S.W. are funded by the Ministry of Education (MOE) and National Research Foundation Singapore, as well as MOE Tier 3 Grant "Random numbers from quantum processes" (MOE2012T31009). The authors would like to thank: Marcos Curty (Universida de Vigo) for many interesting discussions on the practical aspects of the security statements for real implementations of ROT; and Philip Chan (University of Calgary) for numerous explanations and discussions on implementing oneway forward error correction with LDPC codes as well as discussions on the security implications for using oneway codes.
Author information
Authors and Affiliations
Contributions
C.E., S.W. and G.W. conceived the experiment; C.E. constructed the experiment; C.E. and N.G. implemented the efficient oneway error correction; C.E., N.N. and S.W. performed the security analysis; G.W. and R.L. provided materials and support. All authors contributed to the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing financial interests.
Rights and permissions
About this article
Cite this article
Erven, C., Ng, N., Gigov, N. et al. An experimental implementation of oblivious transfer in the noisy storage model. Nat Commun 5, 3418 (2014). https://doi.org/10.1038/ncomms4418
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/ncomms4418
This article is cited by

Experimental cheatsensitive quantum weak coin flipping
Nature Communications (2023)

Stokes metahologram toward optical cryptography
Nature Communications (2022)

Practical quantum tokens without quantum memories and experimental tests
npj Quantum Information (2022)

Probabilistic onetime programs using quantum entanglement
npj Quantum Information (2021)

Continuousvariable protocol for oblivious transfer in the noisystorage model
Nature Communications (2018)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.