Introduction

Uncertainty relations1 are of profound significance in quantum mechanics and quantum information theory. Various important applications of uncertainty relations have been discovered such as entanglement detection2, steering inequalities3 and quantum cryptography4,5,6. The well-known form of the Heisenberg’s uncertainty relations, given by Robertson7, says that the standard deviations of the observables ΔA and ΔB satisfy the following inequality,

As a consequence of the uncertainty relations, it is impossible to determine the exact values of the two incompatible observables simultaneously. However, the lower bound in the above uncertainty inequality may become trivial if the measured state |ψ〉 belongs to the nullspace of the commutator [A, B].

In fact, the uncertainty relations provide a limitation on how much information one can obtain by measuring a physical system, and can be characterized in terms of the probability distributions of the measurement outcomes. In order to overcome the drawback in the product form of variance base uncertainty relations, Deutsch8 introduced the entropic uncertainty relations, which were later improved by Maassen and Uffink9: H(A) + H(B) ≥ −2 log c(A, B), where H is the Shannon entropy, is maximum overlap between the basis elements {|am〉} and {|bn〉} of the eigenbases of A and B, respectively. Recently, the Maassen-Uffink bound has been surprisingly improved by Coles and Piani10, Rudnicki, Puchaa and yczkowski11, for a review on entopic uncertainty relations see refs 12 and 13.

Friedland, Gheorghiu and Gour14 proposed a new concept called “universal uncertainty relations” which are not limited to considering only the well-known entropic functions such as Shannon entropy, Renyi entropy and Tsallis entropy, but also any nonnegative Schur-concave functions. On the other hand, Puchaa, Rudnicki and yczkowski15 independently used majorization technique to establish entropic uncertainty relations similar to “universal uncertainty relations”. Let  and be orthonormal bases of a d-dimensional Hilbert space H. Denote by pm(ρ) = 〈am|ρ|am〉 and qn(ρ) = 〈bn|ρ|bn〉 the probability distributions obtained by measuring the state ρ with respect to these bases, which constitute two probability vectors p(ρ) = (p1, p2, …, pd) and q(ρ) = (q1, q2, …, qd), respectively. It has been shown that the tensor product of the two probability vectors p(ρ) and q(ρ) is majored by a vector ω independent from the state ρ,

where “” stands for “majorization”: in Rd if for all 1 ≤ k ≤ d − 1 and . The down-arrow vector x denotes that the components of x are rearranged in descending order, . The d2-dimensional vector ω is given by

where

with Ik [d] × [d] being a subset of k distinct pairs of indices (m, n) and [d] is the set of the natural numbers from 1 to d. The outer maximum is over all subsets Ik with cardinality k and the inner maximum runs over all density matrices.

Equation (2) is called a universal uncertainty relation, as for any uncertainty measure Φ, a nonnegative Schur-concave function, one has that

The universal uncertainty relation (UUR) (2) generates infinitely many uncertainty relations, for each Φ, in which the right hand side provides a single lower bound.

In relation (2), the state independent vector ω decided by Ωk in Eq. (4) is too hard to evaluate explicitly in general, as it is involved with a highly nontrivial optimization problem. For this reason, only an approximation of Ωk has been presented14,15 to construct a weaker majorization vector . Naturally, how to find a stronger approximation than previous works becomes an interesting open question.

Results

We first introduce a scheme called “joint probability distribution diagram” (JPDD) to consider the optimization problem involved in calculating Ωk. Next, we present a stronger approximation by proposing an analytical formula for Ωk. To facilitate presentation, we denote our stronger approximation as Ωk without ambiguity. All uncertainty relations considered in the paper will be in the absence of quantum side information.

To construct the joint probability distribution diagram, we associate each summand pi(ρ)qj(ρ) in Ωk to a box located at the position (i, j). Then the summation in Ωk corresponds to certain region of boxes (or rather lattice points) in the first quadrant. We configure the region in a combinatorial way. Suppose that p1 ≥ p2 ≥ ··· ≥ pd, q1 ≥ q2 ≥ ··· ≥ qd. Consider the following d × d-matrix

where the entries descend along the rows and columns by assumption. Now, we use a box □ to represent an entry of the matrix. A shadow or grey box in the JPDD means the corresponding entry in the matrix. For example, the top left shadow of the block box specifies the entry p1q1, see Fig. 1. Thus the region corresponding to the summation in will be a special region of the rectangular matrix.

Figure 1
figure 1

The left-top shadow of the block box specifies the entry p1q1.

Our scheme, JPDD, provides a combinatorial method to compute the special region with respect to the Ωk. First, it is easy to see that the top upper left box in JPDD is the maximal element, i.e. Ω1, since p1q1 − piqj ≥ piq1 − piqj = pi(q1 − qj) ≥ 0. The main idea is that each exact solution of Ωk corresponds to a particular region in this matrix.

Suppose that the k-th region is found, i.e. Ωk is obtained, then the next (k + 1)-th region is obtained from the k-th region by adding a special box, which must be “connected” with certain boundary of the k-th region. This iterative procedure enables us to compute all Ωk. Before proving the statement rigorously, we first introduce some terminologies.

[Definition 1] (Different boxes) Two boxes (matrix elements) piqj and pkql are said to be different if they occupy different positions in JPDD, namely, i ≠ k or j ≠ l. The Fig. 2 shows three examples of different boxes. Note that it may happen that even if the numerical values of piqj and pkql are the same, but graphically they are treated as different boxes. “different” and “same” do not imply their quantitative relation. For example, p1q1 may equal to p1q3 in general.

Figure 2
figure 2

Different boxes.

[Definition 2] (Connectedness). Two boxes piqj and pkql in JPDD are connected if there does not exist any box pmqn, different from both piqj and pkql, such that min{piqj, pkql} < pmqn < max{piqj, pkql}. For example, different boxes piqj and pkql are connected if piqj = pkql. For d = 4, if p1 > p2 > p3 > p4, q1 > q2 > q3 > q4, piqj ≠ pkql, pmqn ≠ piqj, pmqn ≠ pkql, for any i, j, k, l {1, 2, 3, 4}, then p2q2 and p3q3 are not connected while p2q3 and p2q2 are connected, see Fig. 3.

Figure 3
figure 3

Connectedness.

[Definition 3] (Connected region). A set of different boxes is called a region, denoted by . A region is connected if , then either or , where is the maximal value of all the elements in . Note that the region of boxes corresponding to a Ωk must contain the top-left element p1q1 in JPDD as its largest element.

For any probability vector p(ρ) on a d-dimensional Hilbert space with pi = 〈ai|ρ|ai〉, let , ,  ··· , . Similarly, are defined similarly for another probability vector q(ρ) on the same Hilbert space. For any sequence k1,  ··· , kn, 1 ≤ n ≤ d, we define In particular, if p1 ≥ ··· ≥ pd, q1 ≥ ··· ≥ qd, then , which can be configured by the Fig. 4. In a JPDD when the first k boxes are chosen, the next (maximal) (k + 1)-th box must appear at the top left corner in the unoccupied region, we give this as follow lemma:

Figure 4
figure 4

related joint probability distribution diagram.

[Lemma]: The maximal k boxes for Ωk in the JPDD can be selected to form a connected region. ■

Lemma gives a way to get Ωk+1 from Ωk in a JPDD. As an example, we show how to get Ω3 from Ω2. Set , , . If c21 ≥ c22, then . If c21 ≤ c22, then . That is, if , then . If , then . Namely, . Thus Ω3 is determined by Ω2.

In general, , subjecting to . By Lemma it follows that

which gives an iterative formula of Ωk in terms of ’s.

We list in Figs 5 and 6 all the possible Ωk for k = 1, 2 ,..., 4. The above example to get Ω3 from Ω2 corresponds to move from the second row to the third. Now we are ready to show the main result.

Figure 5
figure 5

The (k + 1)-th box is fixed by Ωk.

The arrows show the position of the (k + 1)-th box.

Figure 6
figure 6

The tree diagrams to get the JPDDs in the k-th row by adding one shadow box to the JPDDs in the k − 1 row.

[Theorem] The quantities Ωk are given by

The solution given in Eq. (8) can be explained as follows. First, for k = 1, 2, they are solved simply as

where , , the maximum is taken over all indices m ≠ m′, n ≠ n′, and over all n = n′,m ≠ m′. Then, for k = 3 in JPDD, Furthermore, and . Since , we get and We have shown how to calculate Ω1, Ω2 and Ω3. For the cases k ≥ 4, interested readers can calculate Ωk using a similar method and we sketch the details in the Methods.

The above theory enables us to formulate a series of Ωk, based on all quantities we obtain a tighter majorization vector ω. Note that our method is valid when all the maximums are taken over the same quantum state, otherwise our bounds will fail to hold. Even so, our results can outperform BMaj211 to some extent.

Our results enable us to strengthen the bounds on the sum of two Shannon entropies by BJPDD = H(ω), where ω is given by the improved Ωk in Eq. (8) and H is the Shannon entropy. To see this phenomenon, let us first consider a 4-dimensional system with incompatible observables

and the unitary transformation U (Umn = 〈am|bn〉) between them

In Fig. 7, we plot the difference between BJPDD and BMaj211, i.e. BJPDD − BMaj2 (the red line). Clearly, our bound BJPDD is tighter than BMaj2 to some extent. Note that, the entropies are defined with base 2 in general. But in our figures, in order to make them more readable, we take the natural logarithm instead.

Figure 7
figure 7

Difference between BJPDD and BMaj2.

To appreciate the stronger vector ω in the improved UUR, we can consider Shannon entropy in the uncertainty relations to obtain a tighter bound than the previous work16. Namely if we Φ in Eq. (5) as the Shannon entropy H, then we have

where c* ≈ 0.834, the bound H1(c) is the same as the one given in Vincente and Ruiz’s work16, and G(c) = H(ω) with ω given by our formula Eq. (3). The bound G(c) outperforms the Vincente and Ruiz’s bound F(c)16 in the interval [c*, 1]. For further details, see Fig. 8.

Figure 8
figure 8

The vertical coordinate is G(c) − F(c).

The horizontal coordinate is for random runs. It can be seen that our bound outperforms the bound16 100% of the time, while a bound given by Friedland14 outperforms the bound16 around 90% of the time.

Conclusion

In conclusion, we have presented a method called joint probability distribution diagram to strengthen the bounds for the universal uncertainty relations. As an example, we consider the bounds on the sum of the Shannon entropies. As the universal uncertainty relations capture the essence of uncertainty in quantum theory, it is unnecessary to quantify them by particular measures of uncertainty such as Shannon or Renyi entropies. Our results give a way to resolve some important cases in this direction, and is shown to offer a better bound for any uncertainty relations given by the nonnegative Schur-concave functions. Furthermore, how to extend this method to the case of multiple measurements are interesting, which requires further studies.

Methods

Proof of the Lamma

The case of k = 1 is obvious since the maximal element is p1q1. Assume that the statement holds for the case of k − 1: is connected with k1 > … > kn > 0. Suppose on the contrary that the next maximum Ωk = Ωk − 1 + piqj is not connected. Then there are two possibilities: (i) i > n or j > k1, thus we can replace p1qj by piq1 or p1qj and move further to pn+1q1 or to get a possible bigger value for Ωk. (ii) i ≤ n and j > ki, in this case we can also replace the box piqj by a connected box piqj to the region of Ωk − 1 by sliding it leftward or upward. Hence the statement is true by induction. ■

Proof of the Theorem

To calculate , k1 ≥ … ≥ kn and k1 + ··· + kn = k, we note that

where R and S are subsets of distinct indices from [d], |R| is the cardinality of R, and ||·|| is the infinity operator norm which coincides with the maximum eigenvalue of the positive operator. For a given k, there exist sets of k1 ≥ … ≥ kn such that for some n. For any such given k1 ≥ … ≥ kn, the quantity in [.] in Eq. (8) can be calculated. The outer max picks up the largest quantity for all such possible k1, …, kn. Then and . Continuing in this way we have

which gives the improved values of , where max is taken over with k1 ≥ … ≥ kn, and hence ω = (Ω1, Ω2 − Ω1, …, Ωd − Ωd−1, 0, …, 0) for any ρ. ■

Additional Information

How to cite this article: Li, T. et al. Optimal Universal Uncertainty Relations. Sci. Rep. 6, 35735; doi: 10.1038/srep35735 (2016).