Misplaced Pages

Schur product theorem: Difference between revisions

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Browse history interactively← Previous editNext edit →Content deleted Content addedVisualWikitext
Revision as of 23:02, 26 November 2013 editLilily (talk | contribs)260 edits link to text at EUDML← Previous edit Revision as of 01:51, 27 November 2013 edit undoChrisGualtieri (talk | contribs)Autopatrolled, Pending changes reviewers, Rollbackers457,369 editsm Checkwiki 61 + General Fixes using AWBNext edit →
Line 1: Line 1:
In ], particularly in ], the '''Schur product theorem''' states that the ] of two ] is also a positive definite matrix. The result is named after ]<ref name="Sch1911">{{Cite doi|10.1515/crll.1911.140.1}}</ref> (Schur 1911, p. 14, Theorem VII) (note that Schur signed as J. Schur in ''Journal für die reine und angewandte Mathematik''<ref>{{Cite doi|10.1007/b105056}}, page 9, Ch. 0.6 ''Publication under J. Schur''</ref><ref>{{Cite doi|10.1112/blms/15.2.97}}</ref>.) In ], particularly in ], the '''Schur product theorem''' states that the ] of two ] is also a positive definite matrix. The result is named after ]<ref name="Sch1911">{{Cite doi|10.1515/crll.1911.140.1}}</ref> (Schur 1911, p.&nbsp;14, Theorem VII) (note that Schur signed as J. Schur in ''Journal für die reine und angewandte Mathematik''.<ref>{{Cite doi|10.1007/b105056}}, page 9, Ch. 0.6 ''Publication under J. Schur''</ref><ref>{{Cite doi|10.1112/blms/15.2.97}}</ref>)


== Proof == == Proof ==
Line 7: Line 7:
It is easy to show that for matrices <math>M</math> and <math>N</math>, the Hadamard product <math>M \circ N</math> considered as a bilinear form acts on vectors <math>a, b</math> as It is easy to show that for matrices <math>M</math> and <math>N</math>, the Hadamard product <math>M \circ N</math> considered as a bilinear form acts on vectors <math>a, b</math> as
: <math>a^T (M \circ N) b = \operatorname{Tr}(M \operatorname{diag}(a) N \operatorname{diag}(b))</math> : <math>a^T (M \circ N) b = \operatorname{Tr}(M \operatorname{diag}(a) N \operatorname{diag}(b))</math>
where <math>\operatorname{Tr}</math> is the matrix ] and <math>\operatorname{diag}(a)</math> is the ] having as diagonal entries the elements of <math>a</math>. where <math>\operatorname{Tr}</math> is the matrix ] and <math>\operatorname{diag}(a)</math> is the ] having as diagonal entries the elements of <math>a</math>.


Since <math>M</math> and <math>N</math> are positive definite, we can consider their square-roots <math>M^{1/2}</math> and <math>N^{1/2}</math> and write Since <math>M</math> and <math>N</math> are positive definite, we can consider their square-roots <math>M^{1/2}</math> and <math>N^{1/2}</math> and write
Line 47: Line 47:
Let <math>M = \sum \mu_i m_i m_i^T</math> and <math>N = \sum \nu_i n_i n_i^T</math>. Then Let <math>M = \sum \mu_i m_i m_i^T</math> and <math>N = \sum \nu_i n_i n_i^T</math>. Then
: <math>M \circ N = \sum_{ij} \mu_i \nu_i (m_i m_i^T) \circ (n_i n_i^T) = \sum_{ij} \mu_i \nu_j (m_i \circ n_j) (m_i \circ n_j)^T</math> : <math>M \circ N = \sum_{ij} \mu_i \nu_i (m_i m_i^T) \circ (n_i n_i^T) = \sum_{ij} \mu_i \nu_j (m_i \circ n_j) (m_i \circ n_j)^T</math>
Each <math>(m_i \circ n_j) (m_i \circ n_j)^T</math> is positive (but, except in the 1-dimensional case, not positive definite, since they are ] 1 matrices) and <math>\mu_i \nu_j > 0</math>, thus the sum giving <math>M \circ N</math> is also positive. Each <math>(m_i \circ n_j) (m_i \circ n_j)^T</math> is positive (but, except in the 1-dimensional case, not positive definite, since they are ] 1 matrices) and <math>\mu_i \nu_j > 0</math>, thus the sum giving <math>M \circ N</math> is also positive.


==== Complete proof ==== ==== Complete proof ====
Line 62: Line 62:


== External links == == External links ==
* at

at


] ]

Revision as of 01:51, 27 November 2013

In mathematics, particularly in linear algebra, the Schur product theorem states that the Hadamard product of two positive definite matrices is also a positive definite matrix. The result is named after Issai Schur (Schur 1911, p. 14, Theorem VII) (note that Schur signed as J. Schur in Journal für die reine und angewandte Mathematik.)

Proof

Proof using the trace formula

It is easy to show that for matrices M {\displaystyle M} and N {\displaystyle N} , the Hadamard product M N {\displaystyle M\circ N} considered as a bilinear form acts on vectors a , b {\displaystyle a,b} as

a T ( M N ) b = Tr ( M diag ( a ) N diag ( b ) ) {\displaystyle a^{T}(M\circ N)b=\operatorname {Tr} (M\operatorname {diag} (a)N\operatorname {diag} (b))}

where Tr {\displaystyle \operatorname {Tr} } is the matrix trace and diag ( a ) {\displaystyle \operatorname {diag} (a)} is the diagonal matrix having as diagonal entries the elements of a {\displaystyle a} .

Since M {\displaystyle M} and N {\displaystyle N} are positive definite, we can consider their square-roots M 1 / 2 {\displaystyle M^{1/2}} and N 1 / 2 {\displaystyle N^{1/2}} and write

Tr ( M diag ( a ) N diag ( b ) ) = Tr ( M 1 / 2 M 1 / 2 diag ( a ) N 1 / 2 N 1 / 2 diag ( b ) ) = Tr ( M 1 / 2 diag ( a ) N 1 / 2 N 1 / 2 diag ( b ) M 1 / 2 ) {\displaystyle \operatorname {Tr} (M\operatorname {diag} (a)N\operatorname {diag} (b))=\operatorname {Tr} (M^{1/2}M^{1/2}\operatorname {diag} (a)N^{1/2}N^{1/2}\operatorname {diag} (b))=\operatorname {Tr} (M^{1/2}\operatorname {diag} (a)N^{1/2}N^{1/2}\operatorname {diag} (b)M^{1/2})}

Then, for a = b {\displaystyle a=b} , this is written as Tr ( A T A ) {\displaystyle \operatorname {Tr} (A^{T}A)} for A = N 1 / 2 diag ( a ) M 1 / 2 {\displaystyle A=N^{1/2}\operatorname {diag} (a)M^{1/2}} and thus is positive. This shows that ( M N ) {\displaystyle (M\circ N)} is a positive definite matrix.

Proof using Gaussian integration

Case of M = N

Let X {\displaystyle X} be an n {\displaystyle n} -dimensional centered Gaussian random variable with covariance X i X j = M i j {\displaystyle \langle X_{i}X_{j}\rangle =M_{ij}} . Then the covariance matrix of X i 2 {\displaystyle X_{i}^{2}} and X j 2 {\displaystyle X_{j}^{2}} is

Cov ( X i 2 , X j 2 ) = X i 2 X j 2 X i 2 X j 2 {\displaystyle \operatorname {Cov} (X_{i}^{2},X_{j}^{2})=\langle X_{i}^{2}X_{j}^{2}\rangle -\langle X_{i}^{2}\rangle \langle X_{j}^{2}\rangle }

Using Wick's theorem to develop X i 2 X j 2 = 2 X i X j 2 + X i 2 X j 2 {\displaystyle \langle X_{i}^{2}X_{j}^{2}\rangle =2\langle X_{i}X_{j}\rangle ^{2}+\langle X_{i}^{2}\rangle \langle X_{j}^{2}\rangle } we have

Cov ( X i 2 , X j 2 ) = 2 X i X j 2 = 2 M i j 2 {\displaystyle \operatorname {Cov} (X_{i}^{2},X_{j}^{2})=2\langle X_{i}X_{j}\rangle ^{2}=2M_{ij}^{2}}

Since a covariance matrix is positive definite, this proves that the matrix with elements M i j 2 {\displaystyle M_{ij}^{2}} is a positive definite matrix.

General case

Let X {\displaystyle X} and Y {\displaystyle Y} be n {\displaystyle n} -dimensional centered Gaussian random variables with covariances X i X j = M i j {\displaystyle \langle X_{i}X_{j}\rangle =M_{ij}} , Y i Y j = N i j {\displaystyle \langle Y_{i}Y_{j}\rangle =N_{ij}} and independt from each other so that we have

X i Y j = 0 {\displaystyle \langle X_{i}Y_{j}\rangle =0} for any i , j {\displaystyle i,j}

Then the covariance matrix of X i Y i {\displaystyle X_{i}Y_{i}} and X j Y j {\displaystyle X_{j}Y_{j}} is

Cov ( X i Y i , X j Y j ) = X i Y i X j Y j X i Y i X j Y j {\displaystyle \operatorname {Cov} (X_{i}Y_{i},X_{j}Y_{j})=\langle X_{i}Y_{i}X_{j}Y_{j}\rangle -\langle X_{i}Y_{i}\rangle \langle X_{j}Y_{j}\rangle }

Using Wick's theorem to develop

X i Y i X j Y j = X i X j Y i Y j + X i Y i X i Y j + X i Y j X j Y i {\displaystyle \langle X_{i}Y_{i}X_{j}Y_{j}\rangle =\langle X_{i}X_{j}\rangle \langle Y_{i}Y_{j}\rangle +\langle X_{i}Y_{i}\rangle \langle X_{i}Y_{j}\rangle +\langle X_{i}Y_{j}\rangle \langle X_{j}Y_{i}\rangle }

and also using the independence of X {\displaystyle X} and Y {\displaystyle Y} , we have

Cov ( X i Y i , X j Y j ) = X i X j Y i Y j = M i j N i j {\displaystyle \operatorname {Cov} (X_{i}Y_{i},X_{j}Y_{j})=\langle X_{i}X_{j}\rangle \langle Y_{i}Y_{j}\rangle =M_{ij}N_{ij}}

Since a covariance matrix is positive definite, this proves that the matrix with elements M i j N i j {\displaystyle M_{ij}N_{ij}} is a positive definite matrix.

Proof using eigendecomposition

Proof of positivity

Let M = μ i m i m i T {\displaystyle M=\sum \mu _{i}m_{i}m_{i}^{T}} and N = ν i n i n i T {\displaystyle N=\sum \nu _{i}n_{i}n_{i}^{T}} . Then

M N = i j μ i ν i ( m i m i T ) ( n i n i T ) = i j μ i ν j ( m i n j ) ( m i n j ) T {\displaystyle M\circ N=\sum _{ij}\mu _{i}\nu _{i}(m_{i}m_{i}^{T})\circ (n_{i}n_{i}^{T})=\sum _{ij}\mu _{i}\nu _{j}(m_{i}\circ n_{j})(m_{i}\circ n_{j})^{T}}

Each ( m i n j ) ( m i n j ) T {\displaystyle (m_{i}\circ n_{j})(m_{i}\circ n_{j})^{T}} is positive (but, except in the 1-dimensional case, not positive definite, since they are rank 1 matrices) and μ i ν j > 0 {\displaystyle \mu _{i}\nu _{j}>0} , thus the sum giving M N {\displaystyle M\circ N} is also positive.

Complete proof

To show that the result is positive definite requires further proof. We shall show that for any vector a 0 {\displaystyle a\neq 0} , we have a T ( M N ) a > 0 {\displaystyle a^{T}(M\circ N)a>0} . Continuing as above, each a T ( m i n j ) ( m i n j ) T a 0 {\displaystyle a^{T}(m_{i}\circ n_{j})(m_{i}\circ n_{j})^{T}a\geq 0} , so it remains to show that there exist i {\displaystyle i} and j {\displaystyle j} for which the inequality is strict. For this we observe that

a T ( m i n j ) ( m i n j ) T a = ( k m i , k n j , k a k ) 2 {\displaystyle a^{T}(m_{i}\circ n_{j})(m_{i}\circ n_{j})^{T}a=\left(\sum _{k}m_{i,k}n_{j,k}a_{k}\right)^{2}}

Since N {\displaystyle N} is positive definite, there is a j {\displaystyle j} for which n j , k a k {\displaystyle n_{j,k}a_{k}} is not 0 for all k {\displaystyle k} , and then, since M {\displaystyle M} is positive definite, there is an i {\displaystyle i} for which m i , k n j , k a k {\displaystyle m_{i,k}n_{j,k}a_{k}} is not 0 for all k {\displaystyle k} . Then for this i {\displaystyle i} and j {\displaystyle j} we have ( k m i , k n j , k a k ) 2 > 0 {\displaystyle \left(\sum _{k}m_{i,k}n_{j,k}a_{k}\right)^{2}>0} . This completes the proof.

References

  1. Attention: This template ({{cite doi}}) is deprecated. To cite the publication identified by doi:10.1515/crll.1911.140.1, please use {{cite journal}} (if it was published in a bona fide academic journal, otherwise {{cite report}} with |doi=10.1515/crll.1911.140.1 instead.
  2. Attention: This template ({{cite doi}}) is deprecated. To cite the publication identified by doi:10.1007/b105056, please use {{cite journal}} (if it was published in a bona fide academic journal, otherwise {{cite report}} with |doi=10.1007/b105056 instead., page 9, Ch. 0.6 Publication under J. Schur
  3. Attention: This template ({{cite doi}}) is deprecated. To cite the publication identified by doi:10.1112/blms/15.2.97, please use {{cite journal}} (if it was published in a bona fide academic journal, otherwise {{cite report}} with |doi=10.1112/blms/15.2.97 instead.

External links

Categories: