Misplaced Pages

Derivation of the Routh array

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Mathematical proof
This article includes a list of references, related reading, or external links, but its sources remain unclear because it lacks inline citations. Please help improve this article by introducing more precise citations. (February 2011) (Learn how and when to remove this message)

The Routh array is a tabular method permitting one to establish the stability of a system using only the coefficients of the characteristic polynomial. Central to the field of control systems design, the Routh–Hurwitz theorem and Routh array emerge by using the Euclidean algorithm and Sturm's theorem in evaluating Cauchy indices.

The Cauchy index

Given the system:

f ( x ) = a 0 x n + a 1 x n 1 + + a n ( 1 ) = ( x r 1 ) ( x r 2 ) ( x r n ) ( 2 ) {\displaystyle {\begin{aligned}f(x)&{}=a_{0}x^{n}+a_{1}x^{n-1}+\cdots +a_{n}&{}\quad (1)\\&{}=(x-r_{1})(x-r_{2})\cdots (x-r_{n})&{}\quad (2)\\\end{aligned}}}

Assuming no roots of f ( x ) = 0 {\displaystyle f(x)=0} lie on the imaginary axis, and letting

N {\displaystyle N} = The number of roots of f ( x ) = 0 {\displaystyle f(x)=0} with negative real parts, and
P {\displaystyle P} = The number of roots of f ( x ) = 0 {\displaystyle f(x)=0} with positive real parts

then we have

N + P = n ( 3 ) {\displaystyle N+P=n\quad (3)}

Expressing f ( x ) {\displaystyle f(x)} in polar form, we have

f ( x ) = ρ ( x ) e j θ ( x ) ( 4 ) {\displaystyle f(x)=\rho (x)e^{j\theta (x)}\quad (4)}

where

ρ ( x ) = R e 2 [ f ( x ) ] + I m 2 [ f ( x ) ] ( 5 ) {\displaystyle \rho (x)={\sqrt {{\mathfrak {Re}}^{2}+{\mathfrak {Im}}^{2}}}\quad (5)}

and

θ ( x ) = tan 1 ( I m [ f ( x ) ] / R e [ f ( x ) ] ) ( 6 ) {\displaystyle \theta (x)=\tan ^{-1}{\big (}{\mathfrak {Im}}/{\mathfrak {Re}}{\big )}\quad (6)}

from (2) note that

θ ( x ) = θ r 1 ( x ) + θ r 2 ( x ) + + θ r n ( x ) ( 7 ) {\displaystyle \theta (x)=\theta _{r_{1}}(x)+\theta _{r_{2}}(x)+\cdots +\theta _{r_{n}}(x)\quad (7)}

where

θ r i ( x ) = ( x r i ) ( 8 ) {\displaystyle \theta _{r_{i}}(x)=\angle (x-r_{i})\quad (8)}

Now if the i root of f ( x ) = 0 {\displaystyle f(x)=0} has a positive real part, then (using the notation y=(RE,IM))

θ r i ( x ) | x = j = ( x r i ) | x = j = ( 0 R e [ r i ] , I m [ r i ] ) = ( | R e [ r i ] | , ) = π + lim ϕ tan 1 ϕ = 3 π 2 ( 9 ) {\displaystyle {\begin{aligned}\theta _{r_{i}}(x){\big |}_{x=-j\infty }&=\angle (x-r_{i}){\big |}_{x=-j\infty }\\&=\angle (0-{\mathfrak {Re}},-\infty -{\mathfrak {Im}})\\&=\angle (-|{\mathfrak {Re}}|,-\infty )\\&=\pi +\lim _{\phi \to \infty }\tan ^{-1}\phi ={\frac {3\pi }{2}}\quad (9)\\\end{aligned}}}

and

θ r i ( x ) | x = j 0 = ( | R e [ r i ] | , 0 ) = π tan 1 0 = π ( 10 ) {\displaystyle \theta _{r_{i}}(x){\big |}_{x=j0}=\angle (-|{\mathfrak {Re}}|,0)=\pi -\tan ^{-1}0=\pi \quad (10)}

and

θ r i ( x ) | x = j = ( | R e [ r i ] | , ) = π lim ϕ tan 1 ϕ = π 2 ( 11 ) {\displaystyle \theta _{r_{i}}(x){\big |}_{x=j\infty }=\angle (-|{\mathfrak {Re}}|,\infty )=\pi -\lim _{\phi \to \infty }\tan ^{-1}\phi ={\frac {\pi }{2}}\quad (11)}

Similarly, if the i root of f ( x ) = 0 {\displaystyle f(x)=0} has a negative real part,

θ r i ( x ) | x = j = ( x r i ) | x = j = ( 0 R e [ r i ] , I m [ r i ] ) = ( | R e [ r i ] | , ) = 0 lim ϕ tan 1 ϕ = π 2 ( 12 ) {\displaystyle {\begin{aligned}\theta _{r_{i}}(x){\big |}_{x=-j\infty }&=\angle (x-r_{i}){\big |}_{x=-j\infty }\\&=\angle (0-{\mathfrak {Re}},-\infty -{\mathfrak {Im}})\\&=\angle (|{\mathfrak {Re}}|,-\infty )\\&=0-\lim _{\phi \to \infty }\tan ^{1}\phi =-{\frac {\pi }{2}}\quad (12)\\\end{aligned}}}

and

θ r i ( x ) | x = j 0 = ( | R e [ r i ] | , 0 ) = tan 1 0 = 0 ( 13 ) {\displaystyle \theta _{r_{i}}(x){\big |}_{x=j0}=\angle (|{\mathfrak {Re}}|,0)=\tan ^{-1}0=0\,\quad (13)}

and

θ r i ( x ) | x = j = ( | R e [ r i ] | , ) = lim ϕ tan 1 ϕ = π 2 ( 14 ) {\displaystyle \theta _{r_{i}}(x){\big |}_{x=j\infty }=\angle (|{\mathfrak {Re}}|,\infty )=\lim _{\phi \to \infty }\tan ^{-1}\phi ={\frac {\pi }{2}}\,\quad (14)}

From (9) to (11) we find that θ r i ( x ) | x = j x = j = π {\displaystyle \theta _{r_{i}}(x){\Big |}_{x=-j\infty }^{x=j\infty }=-\pi } when the i root of f ( x ) {\displaystyle f(x)} has a positive real part, and from (12) to (14) we find that θ r i ( x ) | x = j x = j = π {\displaystyle \theta _{r_{i}}(x){\Big |}_{x=-j\infty }^{x=j\infty }=\pi } when the i root of f ( x ) {\displaystyle f(x)} has a negative real part. Thus,

θ ( x ) | x = j x = j = ( x r 1 ) | x = j x = j + ( x r 2 ) | x = j x = j + + ( x r n ) | x = j x = j = π N π P ( 15 ) {\displaystyle \theta (x){\Big |}_{x=-j\infty }^{x=j\infty }=\angle (x-r_{1}){\Big |}_{x=-j\infty }^{x=j\infty }+\angle (x-r_{2}){\Big |}_{x=-j\infty }^{x=j\infty }+\cdots +\angle (x-r_{n}){\Big |}_{x=-j\infty }^{x=j\infty }=\pi N-\pi P\quad (15)}

So, if we define

Δ = 1 π θ ( x ) | j j ( 16 ) {\displaystyle \Delta ={\frac {1}{\pi }}\theta (x){\Big |}_{-j\infty }^{j\infty }\quad (16)}

then we have the relationship

N P = Δ ( 17 ) {\displaystyle N-P=\Delta \quad (17)}

and combining (3) and (17) gives us

N = n + Δ 2 {\displaystyle N={\frac {n+\Delta }{2}}} and P = n Δ 2 ( 18 ) {\displaystyle P={\frac {n-\Delta }{2}}\quad (18)}

Therefore, given an equation of f ( x ) {\displaystyle f(x)} of degree n {\displaystyle n} we need only evaluate this function Δ {\displaystyle \Delta } to determine N {\displaystyle N} , the number of roots with negative real parts and P {\displaystyle P} , the number of roots with positive real parts.

Graph of θ versus tan(θ)
Figure 1
tan ( θ ) {\displaystyle \tan(\theta )} versus θ {\displaystyle \theta }

In accordance with (6) and Figure 1, the graph of tan ( θ ) {\displaystyle \tan(\theta )} vs θ {\displaystyle \theta } , varying x {\displaystyle x} over an interval (a,b) where θ a = θ ( x ) | x = j a {\displaystyle \theta _{a}=\theta (x)|_{x=ja}} and θ b = θ ( x ) | x = j b {\displaystyle \theta _{b}=\theta (x)|_{x=jb}} are integer multiples of π {\displaystyle \pi } , this variation causing the function θ ( x ) {\displaystyle \theta (x)} to have increased by π {\displaystyle \pi } , indicates that in the course of travelling from point a to point b, θ {\displaystyle \theta } has "jumped" from + {\displaystyle +\infty } to {\displaystyle -\infty } one more time than it has jumped from {\displaystyle -\infty } to + {\displaystyle +\infty } . Similarly, if we vary x {\displaystyle x} over an interval (a,b) this variation causing θ ( x ) {\displaystyle \theta (x)} to have decreased by π {\displaystyle \pi } , where again θ {\displaystyle \theta } is a multiple of π {\displaystyle \pi } at both x = j a {\displaystyle x=ja} and x = j b {\displaystyle x=jb} , implies that tan θ ( x ) = I m [ f ( x ) ] / R e [ f ( x ) ] {\displaystyle \tan \theta (x)={\mathfrak {Im}}/{\mathfrak {Re}}} has jumped from {\displaystyle -\infty } to + {\displaystyle +\infty } one more time than it has jumped from + {\displaystyle +\infty } to {\displaystyle -\infty } as x {\displaystyle x} was varied over the said interval.

Thus, θ ( x ) | j j {\displaystyle \theta (x){\Big |}_{-j\infty }^{j\infty }} is π {\displaystyle \pi } times the difference between the number of points at which I m [ f ( x ) ] / R e [ f ( x ) ] {\displaystyle {\mathfrak {Im}}/{\mathfrak {Re}}} jumps from {\displaystyle -\infty } to + {\displaystyle +\infty } and the number of points at which I m [ f ( x ) ] / R e [ f ( x ) ] {\displaystyle {\mathfrak {Im}}/{\mathfrak {Re}}} jumps from + {\displaystyle +\infty } to {\displaystyle -\infty } as x {\displaystyle x} ranges over the interval ( j , + j ) {\displaystyle (-j\infty ,+j\infty \,)} provided that at x = ± j {\displaystyle x=\pm j\infty } , tan [ θ ( x ) ] {\displaystyle \tan} is defined.

Graph of θ versus −cotan(θ)
Figure 2
cot ( θ ) {\displaystyle -\cot(\theta )} versus θ {\displaystyle \theta }

In the case where the starting point is on an incongruity (i.e. θ a = π / 2 ± i π {\displaystyle \theta _{a}=\pi /2\pm i\pi } , i = 0, 1, 2, ...) the ending point will be on an incongruity as well, by equation (17) (since N {\displaystyle N} is an integer and P {\displaystyle P} is an integer, Δ {\displaystyle \Delta } will be an integer). In this case, we can achieve this same index (difference in positive and negative jumps) by shifting the axes of the tangent function by π / 2 {\displaystyle \pi /2} , through adding π / 2 {\displaystyle \pi /2} to θ {\displaystyle \theta } . Thus, our index is now fully defined for any combination of coefficients in f ( x ) {\displaystyle f(x)} by evaluating tan [ θ ] = I m [ f ( x ) ] / R e [ f ( x ) ] {\displaystyle \tan={\mathfrak {Im}}/{\mathfrak {Re}}} over the interval (a,b) = ( + j , j ) {\displaystyle (+j\infty ,-j\infty )} when our starting (and thus ending) point is not an incongruity, and by evaluating

tan [ θ ( x ) ] = tan [ θ + π / 2 ] = cot [ θ ( x ) ] = R e [ f ( x ) ] / I m [ f ( x ) ] ( 19 ) {\displaystyle \tan=\tan=-\cot=-{\mathfrak {Re}}/{\mathfrak {Im}}\quad (19)}

over said interval when our starting point is at an incongruity. This difference, Δ {\displaystyle \Delta } , of negative and positive jumping incongruities encountered while traversing x {\displaystyle x} from j {\displaystyle -j\infty } to + j {\displaystyle +j\infty } is called the Cauchy Index of the tangent of the phase angle, the phase angle being θ ( x ) {\displaystyle \theta (x)} or θ ( x ) {\displaystyle \theta '(x)} , depending as θ a {\displaystyle \theta _{a}} is an integer multiple of π {\displaystyle \pi } or not.

The Routh criterion

To derive Routh's criterion, first we'll use a different notation to differentiate between the even and odd terms of f ( x ) {\displaystyle f(x)} :

f ( x ) = a 0 x n + b 0 x n 1 + a 1 x n 2 + b 1 x n 3 + ( 20 ) {\displaystyle f(x)=a_{0}x^{n}+b_{0}x^{n-1}+a_{1}x^{n-2}+b_{1}x^{n-3}+\cdots \quad (20)}

Now we have:

f ( j ω ) = a 0 ( j ω ) n + b 0 ( j ω ) n 1 + a 1 ( j ω ) n 2 + b 1 ( j ω ) n 3 + ( 21 ) = a 0 ( j ω ) n + a 1 ( j ω ) n 2 + a 2 ( j ω ) n 4 + ( 22 ) + b 0 ( j ω ) n 1 + b 1 ( j ω ) n 3 + b 2 ( j ω ) n 5 + {\displaystyle {\begin{aligned}f(j\omega )&=a_{0}(j\omega )^{n}+b_{0}(j\omega )^{n-1}+a_{1}(j\omega )^{n-2}+b_{1}(j\omega )^{n-3}+\cdots &{}\quad (21)\\&=a_{0}(j\omega )^{n}+a_{1}(j\omega )^{n-2}+a_{2}(j\omega )^{n-4}+\cdots &{}\quad (22)\\&+b_{0}(j\omega )^{n-1}+b_{1}(j\omega )^{n-3}+b_{2}(j\omega )^{n-5}+\cdots \\\end{aligned}}}

Therefore, if n {\displaystyle n} is even,

f ( j ω ) = ( 1 ) n / 2 [ a 0 ω n a 1 ω n 2 + a 2 ω n 4 ] ( 23 ) + j ( 1 ) ( n / 2 ) 1 [ b 0 ω n 1 b 1 ω n 3 + b 2 ω n 5 ] {\displaystyle {\begin{aligned}f(j\omega )&=(-1)^{n/2}{\big }&{}\quad (23)\\&+j(-1)^{(n/2)-1}{\big }&{}\\\end{aligned}}}

and if n {\displaystyle n} is odd:

f ( j ω ) = j ( 1 ) ( n 1 ) / 2 [ a 0 ω n a 1 ω n 2 + a 2 ω n 4 ] ( 24 ) + ( 1 ) ( n 1 ) / 2 [ b 0 ω n 1 b 1 ω n 3 + b 2 ω n 5 ] {\displaystyle {\begin{aligned}f(j\omega )&=j(-1)^{(n-1)/2}{\big }&{}\quad (24)\\&+(-1)^{(n-1)/2}{\big }&{}\\\end{aligned}}}

Now observe that if n {\displaystyle n} is an odd integer, then by (3) N + P {\displaystyle N+P} is odd. If N + P {\displaystyle N+P} is an odd integer, then N P {\displaystyle N-P} is odd as well. Similarly, this same argument shows that when n {\displaystyle n} is even, N P {\displaystyle N-P} will be even. Equation (15) shows that if N P {\displaystyle N-P} is even, θ {\displaystyle \theta } is an integer multiple of π {\displaystyle \pi } . Therefore, tan ( θ ) {\displaystyle \tan(\theta )} is defined for n {\displaystyle n} even, and is thus the proper index to use when n is even, and similarly tan ( θ ) = tan ( θ + π ) = cot ( θ ) {\displaystyle \tan(\theta ')=\tan(\theta +\pi )=-\cot(\theta )} is defined for n {\displaystyle n} odd, making it the proper index in this latter case.

Thus, from (6) and (23), for n {\displaystyle n} even:

Δ = I + I m [ f ( x ) ] R e [ f ( x ) ] = I + b 0 ω n 1 b 1 ω n 3 + a 0 ω n a 1 ω n 2 + ( 25 ) {\displaystyle \Delta =I_{-\infty }^{+\infty }{\frac {-{\mathfrak {Im}}}{{\mathfrak {Re}}}}=I_{-\infty }^{+\infty }{\frac {b_{0}\omega ^{n-1}-b_{1}\omega ^{n-3}+\cdots }{a_{0}\omega ^{n}-a_{1}\omega ^{n-2}+\ldots }}\quad (25)}

and from (19) and (24), for n {\displaystyle n} odd:

Δ = I + R e [ f ( x ) ] I m [ f ( x ) ] = I + b 0 ω n 1 b 1 ω n 3 + a 0 ω n a 1 ω n 2 + ( 26 ) {\displaystyle \Delta =I_{-\infty }^{+\infty }{\frac {{\mathfrak {Re}}}{{\mathfrak {Im}}}}=I_{-\infty }^{+\infty }{\frac {b_{0}\omega ^{n-1}-b_{1}\omega ^{n-3}+\ldots }{a_{0}\omega ^{n}-a_{1}\omega ^{n-2}+\ldots }}\quad (26)}

Lo and behold we are evaluating the same Cauchy index for both: Δ = I + b 0 ω n 1 b 1 ω n 3 + a 0 ω n a 1 ω n 2 + ( 27 ) {\displaystyle \Delta =I_{-\infty }^{+\infty }{\frac {b_{0}\omega ^{n-1}-b_{1}\omega ^{n-3}+\ldots }{a_{0}\omega ^{n}-a_{1}\omega ^{n-2}+\ldots }}\quad (27)}

Sturm's theorem

Sturm gives us a method for evaluating Δ = I + f 2 ( x ) f 1 ( x ) {\displaystyle \Delta =I_{-\infty }^{+\infty }{\frac {f_{2}(x)}{f_{1}(x)}}} . His theorem states as follows:

Given a sequence of polynomials f 1 ( x ) , f 2 ( x ) , , f m ( x ) {\displaystyle f_{1}(x),f_{2}(x),\dots ,f_{m}(x)} where:

1) If f k ( x ) = 0 {\displaystyle f_{k}(x)=0} then f k 1 ( x ) 0 {\displaystyle f_{k-1}(x)\neq 0} , f k + 1 ( x ) 0 {\displaystyle f_{k+1}(x)\neq 0} , and sign [ f k 1 ( x ) ] = sign [ f k + 1 ( x ) ] {\displaystyle \operatorname {sign} =-\operatorname {sign} }

2) f m ( x ) 0 {\displaystyle f_{m}(x)\neq 0} for < x < {\displaystyle -\infty <x<\infty }

and we define V ( x ) {\displaystyle V(x)} as the number of changes of sign in the sequence f 1 ( x ) , f 2 ( x ) , , f m ( x ) {\displaystyle f_{1}(x),f_{2}(x),\dots ,f_{m}(x)} for a fixed value of x {\displaystyle x} , then:

Δ = I + f 2 ( x ) f 1 ( x ) = V ( ) V ( + ) ( 28 ) {\displaystyle \Delta =I_{-\infty }^{+\infty }{\frac {f_{2}(x)}{f_{1}(x)}}=V(-\infty )-V(+\infty )\quad (28)}

A sequence satisfying these requirements is obtained using the Euclidean algorithm, which is as follows:

Starting with f 1 ( x ) {\displaystyle f_{1}(x)} and f 2 ( x ) {\displaystyle f_{2}(x)} , and denoting the remainder of f 1 ( x ) / f 2 ( x ) {\displaystyle f_{1}(x)/f_{2}(x)} by f 3 ( x ) {\displaystyle f_{3}(x)} and similarly denoting the remainder of f 2 ( x ) / f 3 ( x ) {\displaystyle f_{2}(x)/f_{3}(x)} by f 4 ( x ) {\displaystyle f_{4}(x)} , and so on, we obtain the relationships:

f 1 ( x ) = q 1 ( x ) f 2 ( x ) f 3 ( x ) ( 29 ) f 2 ( x ) = q 2 ( x ) f 3 ( x ) f 4 ( x ) f m 1 ( x ) = q m 1 ( x ) f m ( x ) {\displaystyle {\begin{aligned}&f_{1}(x)=q_{1}(x)f_{2}(x)-f_{3}(x)\quad (29)\\&f_{2}(x)=q_{2}(x)f_{3}(x)-f_{4}(x)\\&\ldots \\&f_{m-1}(x)=q_{m-1}(x)f_{m}(x)\\\end{aligned}}}

or in general

f k 1 ( x ) = q k 1 ( x ) f k ( x ) f k + 1 ( x ) {\displaystyle f_{k-1}(x)=q_{k-1}(x)f_{k}(x)-f_{k+1}(x)}

where the last non-zero remainder, f m ( x ) {\displaystyle f_{m}(x)} will therefore be the highest common factor of f 1 ( x ) , f 2 ( x ) , , f m 1 ( x ) {\displaystyle f_{1}(x),f_{2}(x),\dots ,f_{m-1}(x)} . It can be observed that the sequence so constructed will satisfy the conditions of Sturm's theorem, and thus an algorithm for determining the stated index has been developed.

It is in applying Sturm's theorem (28) to (29), through the use of the Euclidean algorithm above that the Routh matrix is formed.

We get

f 3 ( ω ) = a 0 b 0 f 2 ( ω ) f 1 ( ω ) ( 30 ) {\displaystyle f_{3}(\omega )={\frac {a_{0}}{b_{0}}}f_{2}(\omega )-f_{1}(\omega )\quad (30)}

and identifying the coefficients of this remainder by c 0 {\displaystyle c_{0}} , c 1 {\displaystyle -c_{1}} , c 2 {\displaystyle c_{2}} , c 3 {\displaystyle -c_{3}} , and so forth, makes our formed remainder

f 3 ( ω ) = c 0 ω n 2 c 1 ω n 4 + c 2 ω n 6 ( 31 ) {\displaystyle f_{3}(\omega )=c_{0}\omega ^{n-2}-c_{1}\omega ^{n-4}+c_{2}\omega ^{n-6}-\cdots \quad (31)}

where

c 0 = a 1 a 0 b 0 b 1 = b 0 a 1 a 0 b 1 b 0 ; c 1 = a 2 a 0 b 0 b 2 = b 0 a 2 a 0 b 2 b 0 ; ( 32 ) {\displaystyle c_{0}=a_{1}-{\frac {a_{0}}{b_{0}}}b_{1}={\frac {b_{0}a_{1}-a_{0}b_{1}}{b_{0}}};c_{1}=a_{2}-{\frac {a_{0}}{b_{0}}}b_{2}={\frac {b_{0}a_{2}-a_{0}b_{2}}{b_{0}}};\ldots \quad (32)}

Continuing with the Euclidean algorithm on these new coefficients gives us

f 4 ( ω ) = b 0 c 0 f 3 ( ω ) f 2 ( ω ) ( 33 ) {\displaystyle f_{4}(\omega )={\frac {b_{0}}{c_{0}}}f_{3}(\omega )-f_{2}(\omega )\quad (33)}

where we again denote the coefficients of the remainder f 4 ( ω ) {\displaystyle f_{4}(\omega )} by d 0 {\displaystyle d_{0}} , d 1 {\displaystyle -d_{1}} , d 2 {\displaystyle d_{2}} , d 3 {\displaystyle -d_{3}} , making our formed remainder

f 4 ( ω ) = d 0 ω n 3 d 1 ω n 5 + d 2 ω n 7 ( 34 ) {\displaystyle f_{4}(\omega )=d_{0}\omega ^{n-3}-d_{1}\omega ^{n-5}+d_{2}\omega ^{n-7}-\cdots \quad (34)}

and giving us

d 0 = b 1 b 0 c 0 c 1 = c 0 b 1 b 0 c 1 c 0 ; d 1 = b 2 b 0 c 0 c 2 = c 0 b 2 b 0 c 2 c 0 ; ( 35 ) {\displaystyle d_{0}=b_{1}-{\frac {b_{0}}{c_{0}}}c_{1}={\frac {c_{0}b_{1}-b_{0}c_{1}}{c_{0}}};d_{1}=b_{2}-{\frac {b_{0}}{c_{0}}}c_{2}={\frac {c_{0}b_{2}-b_{0}c_{2}}{c_{0}}};\ldots \quad (35)}

The rows of the Routh array are determined exactly by this algorithm when applied to the coefficients of (20). An observation worthy of note is that in the regular case the polynomials f 1 ( ω ) {\displaystyle f_{1}(\omega )} and f 2 ( ω ) {\displaystyle f_{2}(\omega )} have as the highest common factor f n + 1 ( ω ) {\displaystyle f_{n+1}(\omega )} and thus there will be n {\displaystyle n} polynomials in the chain f 1 ( x ) , f 2 ( x ) , , f m ( x ) {\displaystyle f_{1}(x),f_{2}(x),\dots ,f_{m}(x)} .

Note now, that in determining the signs of the members of the sequence of polynomials f 1 ( x ) , f 2 ( x ) , , f m ( x ) {\displaystyle f_{1}(x),f_{2}(x),\dots ,f_{m}(x)} that at ω = ± {\displaystyle \omega =\pm \infty } the dominating power of ω {\displaystyle \omega } will be the first term of each of these polynomials, and thus only these coefficients corresponding to the highest powers of ω {\displaystyle \omega } in f 1 ( x ) , f 2 ( x ) , {\displaystyle f_{1}(x),f_{2}(x),\dots } , and f m ( x ) {\displaystyle f_{m}(x)} , which are a 0 {\displaystyle a_{0}} , b 0 {\displaystyle b_{0}} , c 0 {\displaystyle c_{0}} , d 0 {\displaystyle d_{0}} , ... determine the signs of f 1 ( x ) {\displaystyle f_{1}(x)} , f 2 ( x ) {\displaystyle f_{2}(x)} , ..., f m ( x ) {\displaystyle f_{m}(x)} at ω = ± {\displaystyle \omega =\pm \infty } .

So we get V ( + ) = V ( a 0 , b 0 , c 0 , d 0 , ) {\displaystyle V(+\infty )=V(a_{0},b_{0},c_{0},d_{0},\dots )} that is, V ( + ) {\displaystyle V(+\infty )} is the number of changes of sign in the sequence a 0 n {\displaystyle a_{0}\infty ^{n}} , b 0 n 1 {\displaystyle b_{0}\infty ^{n-1}} , c 0 n 2 {\displaystyle c_{0}\infty ^{n-2}} , ... which is the number of sign changes in the sequence a 0 {\displaystyle a_{0}} , b 0 {\displaystyle b_{0}} , c 0 {\displaystyle c_{0}} , d 0 {\displaystyle d_{0}} , ... and V ( ) = V ( a 0 , b 0 , c 0 , d 0 , . . . ) {\displaystyle V(-\infty )=V(a_{0},-b_{0},c_{0},-d_{0},...)} ; that is V ( ) {\displaystyle V(-\infty )} is the number of changes of sign in the sequence a 0 ( ) n {\displaystyle a_{0}(-\infty )^{n}} , b 0 ( ) n 1 {\displaystyle b_{0}(-\infty )^{n-1}} , c 0 ( ) n 2 {\displaystyle c_{0}(-\infty )^{n-2}} , ... which is the number of sign changes in the sequence a 0 {\displaystyle a_{0}} , b 0 {\displaystyle -b_{0}} , c 0 {\displaystyle c_{0}} , d 0 {\displaystyle -d_{0}} , ...

Since our chain a 0 {\displaystyle a_{0}} , b 0 {\displaystyle b_{0}} , c 0 {\displaystyle c_{0}} , d 0 {\displaystyle d_{0}} , ... will have n {\displaystyle n} members it is clear that V ( + ) + V ( ) = n {\displaystyle V(+\infty )+V(-\infty )=n} since within V ( a 0 , b 0 , c 0 , d 0 , ) {\displaystyle V(a_{0},b_{0},c_{0},d_{0},\dots )} if going from a 0 {\displaystyle a_{0}} to b 0 {\displaystyle b_{0}} a sign change has not occurred, within V ( a 0 , b 0 , c 0 , d 0 , ) {\displaystyle V(a_{0},-b_{0},c_{0},-d_{0},\dots )} going from a 0 {\displaystyle a_{0}} to b 0 {\displaystyle -b_{0}} one has, and likewise for all n {\displaystyle n} transitions (there will be no terms equal to zero) giving us n {\displaystyle n} total sign changes.

As Δ = V ( ) V ( + ) {\displaystyle \Delta =V(-\infty )-V(+\infty )} and n = V ( + ) + V ( ) {\displaystyle n=V(+\infty )+V(-\infty )} , and from (18) P = ( n Δ / 2 ) {\displaystyle P=(n-\Delta /2)} , we have that P = V ( + ) = V ( a 0 , b 0 , c 0 , d 0 , ) {\displaystyle P=V(+\infty )=V(a_{0},b_{0},c_{0},d_{0},\dots )} and have derived Routh's theorem -

The number of roots of a real polynomial f ( z ) {\displaystyle f(z)} which lie in the right half plane R e ( r i ) > 0 {\displaystyle {\mathfrak {Re}}(r_{i})>0} is equal to the number of changes of sign in the first column of the Routh scheme.

And for the stable case where P = 0 {\displaystyle P=0} then V ( a 0 , b 0 , c 0 , d 0 , ) = 0 {\displaystyle V(a_{0},b_{0},c_{0},d_{0},\dots )=0} by which we have Routh's famous criterion:

In order for all the roots of the polynomial f ( z ) {\displaystyle f(z)} to have negative real parts, it is necessary and sufficient that all of the elements in the first column of the Routh scheme be different from zero and of the same sign.

References

  • Hurwitz, A., "On the Conditions under which an Equation has only Roots with Negative Real Parts", Rpt. in Selected Papers on Mathematical Trends in Control Theory, Ed. R. T. Ballman et al. New York: Dover 1964
  • Routh, E. J., A Treatise on the Stability of a Given State of Motion. London: Macmillan, 1877. Rpt. in Stability of Motion, Ed. A. T. Fuller. London: Taylor & Francis, 1975
  • Felix Gantmacher (J.L. Brenner translator) (1959) Applications of the Theory of Matrices, pp 177–80, New York: Interscience.
Categories: