Misplaced Pages

Welch bounds

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.

In mathematics, Welch bounds are a family of inequalities pertinent to the problem of evenly spreading a set of unit vectors in a vector space. The bounds are important tools in the design and analysis of certain methods in telecommunication engineering, particularly in coding theory. The bounds were originally published in a 1974 paper by L. R. Welch.

Mathematical statement

If { x 1 , , x m } {\displaystyle \{x_{1},\ldots ,x_{m}\}} are unit vectors in C n {\displaystyle \mathbb {C} ^{n}} , define c max = max i j | x i , x j | {\displaystyle c_{\max }=\max _{i\neq j}|\langle x_{i},x_{j}\rangle |} , where , {\displaystyle \langle \cdot ,\cdot \rangle } is the usual inner product on C n {\displaystyle \mathbb {C} ^{n}} . Then the following inequalities hold for k = 1 , 2 , {\displaystyle k=1,2,\dots } : ( c max ) 2 k 1 m 1 [ m ( n + k 1 k ) 1 ] . {\displaystyle (c_{\max })^{2k}\geq {\frac {1}{m-1}}\left.} Welch bounds are also sometimes stated in terms of the averaged squared overlap between the set of vectors. In this case, one has the inequality 1 m 2 i , j = 1 m | x i , x j | 2 k 1 ( n + k 1 k ) . {\displaystyle {\frac {1}{m^{2}}}\sum _{i,j=1}^{m}|\langle x_{i},x_{j}\rangle |^{2k}\geq {\frac {1}{\binom {n+k-1}{k}}}.}

Applicability

If m n {\displaystyle m\leq n} , then the vectors { x i } {\displaystyle \{x_{i}\}} can form an orthonormal set in C n {\displaystyle \mathbb {C} ^{n}} . In this case, c max = 0 {\displaystyle c_{\max }=0} and the bounds are vacuous. Consequently, interpretation of the bounds is only meaningful if m > n {\displaystyle m>n} . This will be assumed throughout the remainder of this article.

Proof for k = 1

The "first Welch bound," corresponding to k = 1 {\displaystyle k=1} , is by far the most commonly used in applications. Its proof proceeds in two steps, each of which depends on a more basic mathematical inequality. The first step invokes the Cauchy–Schwarz inequality and begins by considering the m × m {\displaystyle m\times m} Gram matrix G {\displaystyle G} of the vectors { x i } {\displaystyle \{x_{i}\}} ; i.e.,

G = [ x 1 , x 1 x 1 , x m x m , x 1 x m , x m ] {\displaystyle G=\left}

The trace of G {\displaystyle G} is equal to the sum of its eigenvalues. Because the rank of G {\displaystyle G} is at most n {\displaystyle n} , and it is a positive semidefinite matrix, G {\displaystyle G} has at most n {\displaystyle n} positive eigenvalues with its remaining eigenvalues all equal to zero. Writing the non-zero eigenvalues of G {\displaystyle G} as λ 1 , , λ r {\displaystyle \lambda _{1},\ldots ,\lambda _{r}} with r n {\displaystyle r\leq n} and applying the Cauchy-Schwarz inequality to the inner product of an r {\displaystyle r} -vector of ones with a vector whose components are these eigenvalues yields

( T r G ) 2 = ( i = 1 r λ i ) 2 r i = 1 r λ i 2 n i = 1 m λ i 2 {\displaystyle (\mathrm {Tr} \;G)^{2}=\left(\sum _{i=1}^{r}\lambda _{i}\right)^{2}\leq r\sum _{i=1}^{r}\lambda _{i}^{2}\leq n\sum _{i=1}^{m}\lambda _{i}^{2}}

The square of the Frobenius norm (Hilbert–Schmidt norm) of G {\displaystyle G} satisfies

| | G | | 2 = i = 1 m j = 1 m | x i , x j | 2 = i = 1 m λ i 2 {\displaystyle ||G||^{2}=\sum _{i=1}^{m}\sum _{j=1}^{m}|\langle x_{i},x_{j}\rangle |^{2}=\sum _{i=1}^{m}\lambda _{i}^{2}}

Taking this together with the preceding inequality gives

i = 1 m j = 1 m | x i , x j | 2 ( T r G ) 2 n {\displaystyle \sum _{i=1}^{m}\sum _{j=1}^{m}|\langle x_{i},x_{j}\rangle |^{2}\geq {\frac {(\mathrm {Tr} \;G)^{2}}{n}}}

Because each x i {\displaystyle x_{i}} has unit length, the elements on the main diagonal of G {\displaystyle G} are ones, and hence its trace is T r G = m {\displaystyle \mathrm {Tr} \;G=m} . So,

i = 1 m j = 1 m | x i , x j | 2 = m + i j | x i , x j | 2 m 2 n {\displaystyle \sum _{i=1}^{m}\sum _{j=1}^{m}|\langle x_{i},x_{j}\rangle |^{2}=m+\sum _{i\neq j}|\langle x_{i},x_{j}\rangle |^{2}\geq {\frac {m^{2}}{n}}}

or

i j | x i , x j | 2 m ( m n ) n {\displaystyle \sum _{i\neq j}|\langle x_{i},x_{j}\rangle |^{2}\geq {\frac {m(m-n)}{n}}}

The second part of the proof uses an inequality encompassing the simple observation that the average of a set of non-negative numbers can be no greater than the largest number in the set. In mathematical notation, if a 0 {\displaystyle a_{\ell }\geq 0} for = 1 , , L {\displaystyle \ell =1,\ldots ,L} , then

1 L = 1 L a max a {\displaystyle {\frac {1}{L}}\sum _{\ell =1}^{L}a_{\ell }\leq \max a_{\ell }}

The previous expression has m ( m 1 ) {\displaystyle m(m-1)} non-negative terms in the sum, the largest of which is c max 2 {\displaystyle c_{\max }^{2}} . So,

( c max ) 2 1 m ( m 1 ) i j | x i , x j | 2 m n n ( m 1 ) {\displaystyle (c_{\max })^{2}\geq {\frac {1}{m(m-1)}}\sum _{i\neq j}|\langle x_{i},x_{j}\rangle |^{2}\geq {\frac {m-n}{n(m-1)}}}

or

( c max ) 2 m n n ( m 1 ) {\displaystyle (c_{\max })^{2}\geq {\frac {m-n}{n(m-1)}}}

which is precisely the inequality given by Welch in the case that k = 1 {\displaystyle k=1} .

Achieving the Welch bounds

In certain telecommunications applications, it is desirable to construct sets of vectors that meet the Welch bounds with equality. Several techniques have been introduced to obtain so-called Welch Bound Equality (WBE) sets of vectors for the k = 1 {\displaystyle k=1} bound.

The proof given above shows that two separate mathematical inequalities are incorporated into the Welch bound when k = 1 {\displaystyle k=1} . The Cauchy–Schwarz inequality is met with equality when the two vectors involved are collinear. In the way it is used in the above proof, this occurs when all the non-zero eigenvalues of the Gram matrix G {\displaystyle G} are equal, which happens precisely when the vectors { x 1 , , x m } {\displaystyle \{x_{1},\ldots ,x_{m}\}} constitute a tight frame for C n {\displaystyle \mathbb {C} ^{n}} .

The other inequality in the proof is satisfied with equality if and only if | x i , x j | {\displaystyle |\langle x_{i},x_{j}\rangle |} is the same for every choice of i j {\displaystyle i\neq j} . In this case, the vectors are equiangular. So this Welch bound is met with equality if and only if the set of vectors { x i } {\displaystyle \{x_{i}\}} is an equiangular tight frame in C n {\displaystyle \mathbb {C} ^{n}} .

Similarly, the Welch bounds stated in terms of average squared overlap, are saturated for all k t {\displaystyle k\leq t} if and only if the set of vectors is a t {\displaystyle t} -design in the complex projective space C P n 1 {\displaystyle \mathbb {CP} ^{n-1}} .


See also

References

  1. Welch, L. (1974-05-01). "Lower bounds on the maximum cross correlation of signals (Corresp.)". IEEE Transactions on Information Theory. 20 (3): 397–399. doi:10.1109/TIT.1974.1055219. ISSN 1557-9654.
  2. Klappenecker, Andreas; Roetteler, Martin (2005-02-11). "Mutually Unbiased Bases are Complex Projective 2-Designs". arXiv:quant-ph/0502031. {{cite journal}}: Cite journal requires |journal= (help)
  3. Belovs, Aleksandrs; Smotrovs, Juris (2008-07-22). "A Criterion for Attaining the Welch Bounds with Applications for Mutually Unbiased Bases". arXiv:0802.0855. {{cite journal}}: Cite journal requires |journal= (help)
  4. ^ Datta, Somantika; Howard, Stephen; Cochran, Douglas (2012-05-29). "Geometry of the Welch Bounds". arXiv:0909.0206. {{cite journal}}: Cite journal requires |journal= (help)
Category: