Misplaced Pages

Polynomial matrix spectral factorization

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.

Polynomial Matrix Spectral Factorization or Matrix Fejer–Riesz Theorem is a tool used to study the matrix decomposition of polynomial matrices. Polynomial matrices are widely studied in the fields of systems theory and control theory and have seen other uses relating to stable polynomials. In stability theory, Spectral Factorization has been used to find determinantal matrix representations for bivariate stable polynomials and real zero polynomials.

Given a univariate positive polynomial, i.e., p ( t ) > 0 {\displaystyle p(t)>0} for all t R {\displaystyle t\in \mathbb {R} } , the Fejer–Riesz Theorem yields the polynomial spectral factorization p ( t ) = q ( t ) q ¯ ( t ) {\displaystyle p(t)=q(t){\bar {q}}(t)} . Results of this form are generically referred to as Positivstellensatz.

Likewise, the Polynomial Matrix Spectral Factorization provides a factorization for positive definite polynomial matrices. This decomposition also relates to the Cholesky decomposition for scalar matrices A = L L {\displaystyle A=LL^{*}} . This result was originally proven by Norbert Wiener in a more general context which was concerned with integrable matrix-valued functions that also had integrable log determinant. Because applications are often concerned with the polynomial restriction, simpler proofs and individual analysis exist focusing on this case. Weaker positivstellensatz conditions have been studied, specifically considering when the polynomial matrix has positive definite image on semi-algebraic subsets of the reals. Many publications recently have focused on streamlining proofs for these related results. This article roughly follows the recent proof method of Lasha Ephremidze which relies only on elementary linear algebra and complex analysis.

Spectral factorization is used extensively in linear–quadratic–Gaussian control and many algorithms exist to calculate spectral factors. Some modern algorithms focus on the more general setting originally studied by Wiener while others have used Toeplitz matrix advances to speed up factor calculations.

Definition

Consider polynomial matrix P ( t ) = [ p 11 ( t ) p 1 n ( t ) p n 1 ( t ) p n n ( t ) ] , {\displaystyle P(t)={\begin{bmatrix}p_{11}(t)&\ldots &p_{1n}(t)\\\vdots &\ddots &\vdots \\p_{n1}(t)&\cdots &p_{nn}(t)\\\end{bmatrix}},} where each entry p i j ( t ) {\displaystyle p_{ij}(t)} is a complex coefficient polynomial of at most N {\displaystyle N} -degree. If P ( t ) {\displaystyle P(t)} is a positive definite hermitian matrix for all t R {\displaystyle t\in \mathbb {R} } , then there exists a polynomial matrix Q ( t ) = [ q 11 ( t ) q 1 n ( t ) q n 1 ( t ) q n n ( t ) ] , {\displaystyle Q(t)={\begin{bmatrix}q_{11}(t)&\ldots &q_{1n}(t)\\\vdots &\ddots &\vdots \\q_{n1}(t)&\cdots &q_{nn}(t)\\\end{bmatrix}},} such that P ( t ) = Q ( t ) Q ( t ) {\displaystyle P(t)=Q(t)Q(t)^{*}} where Q ( t ) {\displaystyle Q(t)^{*}} is the conjugate transpose. When q i j ( t ) {\displaystyle q_{ij}(t)} is a complex coefficient polynomial or complex coefficient rational function then so are the elements of its conjugate transpose.

We can furthermore find Q ( t ) {\displaystyle Q(t)} which is nonsingular on the lower half plane.

Rational spectral factorization

Let p ( t ) {\displaystyle p(t)} be a rational function where p ( t ) > 0 {\displaystyle p(t)>0} for all t R {\displaystyle t\in \mathbb {R} } . Then there exists a rational function q ( t ) {\displaystyle q(t)} such that p ( t ) = q ( t ) q ¯ ( t ) {\displaystyle p(t)=q(t){\bar {q}}(t)} and q ( t ) {\displaystyle q(t)} has no poles or zeroes in the lower half plane. This decomposition is unique up to multiplication by complex scalars of norm 1 {\displaystyle 1} .

To prove existence write p ( x ) = c i ( x α i ) j ( x β j ) , {\displaystyle p(x)=c{\frac {\prod _{i}(x-\alpha _{i})}{\prod _{j}(x-\beta _{j})}},} where α i β j {\displaystyle \alpha _{i}\neq \beta _{j}} . Letting x {\displaystyle x\to \infty } , we can conclude that c {\displaystyle c} is real and positive. Dividing out by c {\displaystyle {\sqrt {c}}} we reduce to the monic case. The numerator and denominator have distinct sets of roots, so all real roots which show up in either must have even multiplicity (to prevent a sign change locally). We can divide out these real roots to reduce to the case where p ( t ) {\displaystyle p(t)} has only complex roots and poles. By hypothesis we have p ( x ) = i ( x α i ) j ( x β j ) = i ( x α ¯ i ) j ( x β ¯ j ) = p ( x ) ¯ . {\displaystyle p(x)={\frac {\prod _{i}(x-\alpha _{i})}{\prod _{j}(x-\beta _{j})}}={\frac {\prod _{i}(x-{\bar {\alpha }}_{i})}{\prod _{j}(x-{\bar {\beta }}_{j})}}={\overline {p(x)}}.} Since all of the α i , β j {\displaystyle \alpha _{i},\beta _{j}} are complex (and hence not fixed points of conjugation) they both come in conjugate pairs. For each conjugate pair, pick the zero or pole in the upper half plane and accumulate these to obtain q ( t ) {\displaystyle q(t)} . The uniqueness result follows in a standard fashion.

Cholesky decomposition

The inspiration for this result is a factorization which characterizes positive definite matrices.

Decomposition for scalar matrices

Given any positive definite scalar matrix A {\displaystyle A} , the Cholesky decomposition allows us to write A = L L {\displaystyle A=LL^{*}} where L {\displaystyle L} is a lower triangular matrix. If we don't restrict to lower triangular matrices we can consider all factorizations of the form A = V V {\displaystyle A=VV^{*}} . It is not hard to check that all factorizations are achieved by looking at the orbit of L {\displaystyle L} under right multiplication by a unitary matrix, V = L U {\displaystyle V=LU} .

To obtain the lower triangular decomposition we induct by splitting off the first row and first column: [ a 11 a 12 a 12 A 22 ] = [ l 11 0 l 21 L 22 ] [ l 11 l 21 0 L 22 ] = [ l 11 l 11 l 11 l 21 l 11 l 21 l 21 l 21 + L 22 L 22 ] {\displaystyle {\begin{bmatrix}a_{11}&\mathbf {a} _{12}^{*}\\\mathbf {a} _{12}&A_{22}\\\end{bmatrix}}={\begin{bmatrix}l_{11}&0\\\mathbf {l} _{21}&L_{22}\end{bmatrix}}{\begin{bmatrix}l_{11}^{*}&\mathbf {l} _{21}^{*}\\0&L_{22}^{*}\end{bmatrix}}={\begin{bmatrix}l_{11}l_{11}^{*}&l_{11}\mathbf {l} _{21}^{*}\\l_{11}^{*}\mathbf {l} _{21}&\mathbf {l} _{21}\mathbf {l} _{21}^{*}+L_{22}L_{22}^{*}\end{bmatrix}}} Solving these in terms of a i j {\displaystyle a_{ij}} we get l 11 = a 11 {\displaystyle l_{11}={\sqrt {a_{11}}}} l 21 = 1 a 11 a 12 {\displaystyle \mathbf {l} _{21}={\frac {1}{\sqrt {a_{11}}}}\mathbf {a} _{12}} L 22 L 22 = A 22 1 a 11 a 12 a 12 {\displaystyle L_{22}L_{22}^{*}=A_{22}-{\frac {1}{a_{11}}}\mathbf {a} _{12}\mathbf {a} _{12}^{*}}

Since A {\displaystyle A} is positive definite we have a 11 {\displaystyle a_{11}} is a positive real number, so it has a square root. The last condition from induction since the right hand side is the Schur complement of A {\displaystyle A} , which is itself positive definite.

Decomposition for rational polynomial matrices

A rational polynomial matrix is defined as a matrix P ( t ) = [ p 11 ( t ) p 1 n ( t ) p n 1 ( t ) p n n ( t ) ] , {\displaystyle P(t)={\begin{bmatrix}p_{11}(t)&\ldots &p_{1n}(t)\\\vdots &\ddots &\vdots \\p_{n1}(t)&\cdots &p_{nn}(t)\\\end{bmatrix}},} where each entry p i j ( t ) {\displaystyle p_{ij}(t)} is a complex rational function. If P ( t ) {\displaystyle P(t)} is a positive definite Hermitian matrix for all t R {\displaystyle t\in \mathbb {R} } , then by the symmetric Gaussian elimination we performed above, all we need to show is there exists a rational q 11 ( t ) {\displaystyle q_{11}(t)} such that p 11 ( t ) = q 11 ( t ) q 11 ( t ) {\displaystyle p_{11}(t)=q_{11}(t)q_{11}(t)^{*}} , which follows from our rational spectral factorization. Once we have that then we can solve for l 11 ( t ) , l 21 ( t ) {\displaystyle l_{11}(t),\mathbf {l} _{21}(t)} . Since the Schur complement is positive definite for the real t {\displaystyle t} away from the poles and the Schur complement is a rational polynomial matrix we can induct to find L 22 {\displaystyle L_{22}} .

It is not hard to check that we get P ( t ) = L ( t ) L ( t ) {\displaystyle P(t)=L(t)L(t)^{*}} where L ( t ) {\displaystyle L(t)} is a rational polynomial matrix with no poles in the lower half plane.

Extension to polynomial decompositions

One way to prove the existence of polynomial matrix spectral factorization is to apply the Cholesky decomposition to a rational polynomial matrix and modify it to remove lower half plane singularities. That is, given P ( t ) = [ p 11 ( t ) p 1 n ( t ) p n 1 ( t ) p n n ( t ) ] , {\displaystyle P(t)={\begin{bmatrix}p_{11}(t)&\ldots &p_{1n}(t)\\\vdots &\ddots &\vdots \\p_{n1}(t)&\cdots &p_{nn}(t)\\\end{bmatrix}},} where each entry p i j ( t ) {\displaystyle p_{ij}(t)} is a complex coefficient polynomial for all t R {\displaystyle t\in \mathbb {R} } , a rational polynomial matrix L ( t ) {\displaystyle L(t)} with no lower half plane poles exists such that P ( t ) = L ( t ) L ( t ) {\displaystyle P(t)=L(t)L(t)^{*}} . Given a rational polynomial matrix U ( t ) {\displaystyle U(t)} which is unitary valued for real t {\displaystyle t} , there exists another decomposition P ( t ) = L ( t ) L ( t ) = L ( t ) U ( t ) U ( t ) L ( t ) . {\displaystyle P(t)=L(t)L(t)^{*}=L(t)U(t)U(t)^{*}L(t)^{*}.} If det ( L ( a ) ) = 0 {\displaystyle \det(L(a))=0} then there exists a scalar unitary matrix U {\displaystyle U} such that U e 1 = a | a | . {\displaystyle Ue_{1}={\frac {a}{|a|}}.} This implies L ( t ) U {\displaystyle L(t)U} has first column vanish at a {\displaystyle a} . To remove the singularity at a {\displaystyle a} we multiply by U ( t ) = diag ( 1 , , z a ¯ z a , , 1 ) {\displaystyle U(t)=\operatorname {diag} (1,\ldots ,{\frac {z-{\bar {a}}}{z-a}},\ldots ,1)} L ( t ) U U ( t ) {\displaystyle L(t)UU(t)} has determinant with one less zero (by multiplicity) at a, without introducing any poles in the lower half plane of any of the entries.

Example

Consider the following rational matrix decomposition A ( t ) = [ t 2 + 1 2 t 2 t t 2 + 1 ] = [ t i 0 2 t t + i t 2 1 t + i ] [ t + i 2 t t i 0 t 2 1 t i ] . {\displaystyle A(t)={\begin{bmatrix}t^{2}+1&2t\\2t&t^{2}+1\\\end{bmatrix}}={\begin{bmatrix}t-i&0\\{\frac {2t}{t+i}}&{\frac {t^{2}-1}{t+i}}\\\end{bmatrix}}{\begin{bmatrix}t+i&{\frac {2t}{t-i}}\\0&{\frac {t^{2}-1}{t-i}}\\\end{bmatrix}}.} This decomposition has no poles in the upper half plane. However det ( [ t i 0 2 t t + i t 2 1 t + i ] ) = ( t 1 ) ( t i ) ( t + 1 ) t + i , {\displaystyle \det \left({\begin{bmatrix}t-i&0\\{\frac {2t}{t+i}}&{\frac {t^{2}-1}{t+i}}\\\end{bmatrix}}\right)={\frac {(t-1)(t-i)(t+1)}{t+i}},} so we need to modify our decomposition to get rid of the singularity at i {\displaystyle -i} . First we multiply by a scalar unitary matrix U = 1 2 [ 1 1 i i ] , {\displaystyle U={\frac {1}{\sqrt {2}}}{\begin{bmatrix}1&1\\i&-i\\\end{bmatrix}},} such that [ t i 0 2 t t + i t 2 1 t + i ] U = 1 2 [ t i t i i ( t i ) 2 t + i i ( t + i ) ] , {\displaystyle {\begin{bmatrix}t-i&0\\{\frac {2t}{t+i}}&{\frac {t^{2}-1}{t+i}}\\\end{bmatrix}}U={\frac {1}{\sqrt {2}}}{\begin{bmatrix}t-i&t-i\\i{\frac {(t-i)^{2}}{t+i}}&-i(t+i)\\\end{bmatrix}},} becomes a new candidate for our decomposition. Now the first column vanishes at t = i {\displaystyle t=i} , so we multiply through (on the right) by U ( t ) = 1 2 [ t + i t i 0 0 1 ] , {\displaystyle U(t)={\frac {1}{\sqrt {2}}}{\begin{bmatrix}{\frac {t+i}{t-i}}&0\\0&1\end{bmatrix}},} to obtain Q ( t ) = 1 2 [ t i t i i ( t i ) 2 t + i i ( t + i ) ] U ( t ) = 1 2 [ t + i t i i ( t i ) i ( t + i ) ] , {\displaystyle Q(t)={\frac {1}{\sqrt {2}}}{\begin{bmatrix}t-i&t-i\\i{\frac {(t-i)^{2}}{t+i}}&-i(t+i)\\\end{bmatrix}}U(t)={\frac {1}{2}}{\begin{bmatrix}t+i&t-i\\i(t-i)&-i(t+i)\\\end{bmatrix}},} where det ( Q ( t ) ) = i ( 1 t 2 ) . {\displaystyle \det(Q(t))=i(1-t^{2}).} This is our desired decomposition A ( t ) = Q ( t ) Q ( t ) {\displaystyle A(t)=Q(t)Q(t)^{*}} with no singularities in the lower half plane.

Extend analyticity to all of C

After modifications, the decomposition P ( t ) = Q ( t ) Q ( t ) {\displaystyle P(t)=Q(t)Q(t)^{*}} satisfies Q ( t ) {\displaystyle Q(t)} is holomorphic and invertible on the lower half plane. To extend analyticity to the upper half plane we need this key observation: If an invertible rational matrix Q ( t ) {\displaystyle Q(t)} is holomorphic in the lower half plane, Q ( t ) 1 {\displaystyle Q(t)^{-1}} is holomorphic in the lower half plane as well. The analyticity follows from the adjugate matrix formula (since both the entries of Q ( t ) {\displaystyle Q(t)} and det ( Q ( t ) ) 1 {\displaystyle \det(Q(t))^{-1}} are analytic on the lower half plane). The determinant of a rational polynomial matrix can only have poles where its entries have poles, so det ( Q ( t ) ) {\displaystyle \det(Q(t))} has no poles in the lower half plane.

Subsequently, Q ( t ) = ( P ( t ) Q ( t ) 1 ) . {\displaystyle Q(t)=(P(t)Q(t)^{-1})^{*}.} Since Q ( t ) 1 {\displaystyle Q(t)^{-1}} is analytic on the lower half plane, Q ( t ) {\displaystyle Q(t)} is analytic on the upper half plane. Finally if Q ( t ) {\displaystyle Q(t)} has a pole on the real line then Q ( t ) {\displaystyle Q(t)^{*}} has the same pole on the real line which contradicts the hypothesis that P ( t ) {\displaystyle P(t)} has no poles on the real line (i.e. it is analytic everywhere).

The above shows that if Q ( t ) {\displaystyle Q(t)} is analytic and invertible on the lower half plane indeed Q ( t ) {\displaystyle Q(t)} is analytic everywhere and hence a polynomial matrix.

Uniqueness

Given two polynomial matrix decompositions which are invertible on the lower half plane P ( t ) = Q ( t ) Q ( t ) = R ( t ) R ( t ) , {\displaystyle P(t)=Q(t)Q(t)^{*}=R(t)R(t)^{*},} then R ( t ) 1 Q ( t ) Q ( t ) ( R ( t ) ) 1 = I . {\displaystyle R(t)^{-1}Q(t)Q(t)^{*}(R(t)^{*})^{-1}=I.} Since R ( t ) {\displaystyle R(t)} is analytic on the lower half plane and nonsingular, R ( t ) 1 Q ( t ) {\displaystyle R(t)^{-1}Q(t)} is a rational polynomial matrix which is analytic and invertible on the lower half plane. As such, R ( t ) 1 Q ( t ) {\displaystyle R(t)^{-1}Q(t)} is a polynomial matrix which is unitary for all t R {\displaystyle t\in \mathbb {R} } . This means that if q i ( t ) {\displaystyle \mathbf {q} _{i}(t)} is the i t h {\displaystyle i^{th}} row of Q ( t ) {\displaystyle Q(t)} then q i ( t ) q i ( t ) = 1 {\displaystyle \mathbf {q} _{i}(t)\mathbf {q} _{i}(t)^{*}=1} . For real t {\displaystyle t} this is a sum of non-negative polynomials which sums to a constant, implying each of the summands are in fact constant polynomials. Then Q ( t ) = R ( t ) U {\displaystyle Q(t)=R(t)U} where U {\displaystyle U} is a scalar unitary matrix.

See also

Remarks

  1. Subsection needs work. First a better distinction must be made between the "ordinary" and "complex" Fejer-Riesz theorm. See Dritschel 2010 or https://encyclopediaofmath.org/Fej%C3%A9r-Riesz_theorem

Notes

  1. Grinshpan et al. 2016, pp. 1–26.
  2. Wiener & Masani 1957, pp. 111–150.
  3. Tim N.T. Goodman Charles A. Micchelli Giuseppe Rodriguez Sebastiano Seatzu (1997). "Spectral factorization of Laurent polynomials". Advances in Computational Mathematics. 7 (4): 429–454. doi:10.1023/A:1018915407202. S2CID 7880541.
  4. Aljaž Zalar (2016). "Matrix Fejér–Riesz theorem with gaps". Journal of Pure and Applied Algebra. 220 (7): 2533–2548. arXiv:1503.06034. doi:10.1016/j.jpaa.2015.11.018. S2CID 119303900.
  5. Zalar, Aljaž (2016-07-01). "Matrix Fejér–Riesz theorem with gaps". Journal of Pure and Applied Algebra. 220 (7): 2533–2548. arXiv:1503.06034. doi:10.1016/j.jpaa.2015.11.018. S2CID 119303900.
  6. Lasha Ephremidze (2009). "A Simple Proof of the Matrix-Valued Fejer–Riesz Theorem". Journal of Fourier Analysis and Applications. 15 (1): 124–127. arXiv:0708.2179. Bibcode:2009JFAA...15..124E. CiteSeerX 10.1.1.247.3400. doi:10.1007/s00041-008-9051-z. S2CID 115163568. Retrieved 2017-05-23.
  7. Ephremidze, Lasha (2014). "An Elementary Proof of the Polynomial Matrix Spectral Factorization Theorem". Proceedings of the Royal Society of Edinburgh, Section A. 144 (4): 747–751. CiteSeerX 10.1.1.755.9575. doi:10.1017/S0308210512001552. S2CID 119125206.
  8. Sayed & Kailath 2001, pp. 467–496.
  9. Janashia, Lagvilava & Ephremidze 2011, pp. 2318–2326.
  10. Bini et al. 2003, pp. 217–227.

References

Categories: