Misplaced Pages

Bernstein polynomial

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Type of polynomial used in Numerical Analysis For the Bernstein polynomial in D-module theory, see Bernstein–Sato polynomial.
This article includes a list of general references, but it lacks sufficient corresponding inline citations. Please help to improve this article by introducing more precise citations. (June 2016) (Learn how and when to remove this message)
Bernstein polynomials approximating a curve

In the mathematical field of numerical analysis, a Bernstein polynomial is a polynomial expressed as a linear combination of Bernstein basis polynomials. The idea is named after mathematician Sergei Natanovich Bernstein.

Polynomials in Bernstein form were first used by Bernstein in a constructive proof for the Weierstrass approximation theorem. With the advent of computer graphics, Bernstein polynomials, restricted to the interval , became important in the form of Bézier curves.

A numerically stable way to evaluate polynomials in Bernstein form is de Casteljau's algorithm.

Bernstein basis polynomials for 4th degree curve blending

Definition

Bernstein basis polynomials

The n +1 Bernstein basis polynomials of degree n are defined as

b ν , n ( x ) : = ( n ν ) x ν ( 1 x ) n ν , ν = 0 , , n , {\displaystyle b_{\nu ,n}(x)\mathrel {:} \mathrel {=} {\binom {n}{\nu }}x^{\nu }\left(1-x\right)^{n-\nu },\quad \nu =0,\ldots ,n,}

where ( n ν ) {\displaystyle {\tbinom {n}{\nu }}} is a binomial coefficient.

So, for example, b 2 , 5 ( x ) = ( 5 2 ) x 2 ( 1 x ) 3 = 10 x 2 ( 1 x ) 3 . {\displaystyle b_{2,5}(x)={\tbinom {5}{2}}x^{2}(1-x)^{3}=10x^{2}(1-x)^{3}.}

The first few Bernstein basis polynomials for blending 1, 2, 3 or 4 values together are:

b 0 , 0 ( x ) = 1 , b 0 , 1 ( x ) = 1 x , b 1 , 1 ( x ) = x b 0 , 2 ( x ) = ( 1 x ) 2 , b 1 , 2 ( x ) = 2 x ( 1 x ) , b 2 , 2 ( x ) = x 2 b 0 , 3 ( x ) = ( 1 x ) 3 , b 1 , 3 ( x ) = 3 x ( 1 x ) 2 , b 2 , 3 ( x ) = 3 x 2 ( 1 x ) , b 3 , 3 ( x ) = x 3 {\displaystyle {\begin{aligned}b_{0,0}(x)&=1,\\b_{0,1}(x)&=1-x,&b_{1,1}(x)&=x\\b_{0,2}(x)&=(1-x)^{2},&b_{1,2}(x)&=2x(1-x),&b_{2,2}(x)&=x^{2}\\b_{0,3}(x)&=(1-x)^{3},&b_{1,3}(x)&=3x(1-x)^{2},&b_{2,3}(x)&=3x^{2}(1-x),&b_{3,3}(x)&=x^{3}\end{aligned}}}

The Bernstein basis polynomials of degree n form a basis for the vector space Π n {\displaystyle \Pi _{n}} of polynomials of degree at most n with real coefficients.

Bernstein polynomials

A linear combination of Bernstein basis polynomials

B n ( x ) : = ν = 0 n β ν b ν , n ( x ) {\displaystyle B_{n}(x)\mathrel {:} \mathrel {=} \sum _{\nu =0}^{n}\beta _{\nu }b_{\nu ,n}(x)}

is called a Bernstein polynomial or polynomial in Bernstein form of degree n. The coefficients β ν {\displaystyle \beta _{\nu }} are called Bernstein coefficients or Bézier coefficients.

The first few Bernstein basis polynomials from above in monomial form are:

b 0 , 0 ( x ) = 1 , b 0 , 1 ( x ) = 1 1 x , b 1 , 1 ( x ) = 0 + 1 x b 0 , 2 ( x ) = 1 2 x + 1 x 2 , b 1 , 2 ( x ) = 0 + 2 x 2 x 2 , b 2 , 2 ( x ) = 0 + 0 x + 1 x 2 b 0 , 3 ( x ) = 1 3 x + 3 x 2 1 x 3 , b 1 , 3 ( x ) = 0 + 3 x 6 x 2 + 3 x 3 , b 2 , 3 ( x ) = 0 + 0 x + 3 x 2 3 x 3 , b 3 , 3 ( x ) = 0 + 0 x + 0 x 2 + 1 x 3 {\displaystyle {\begin{aligned}b_{0,0}(x)&=1,\\b_{0,1}(x)&=1-1x,&b_{1,1}(x)&=0+1x\\b_{0,2}(x)&=1-2x+1x^{2},&b_{1,2}(x)&=0+2x-2x^{2},&b_{2,2}(x)&=0+0x+1x^{2}\\b_{0,3}(x)&=1-3x+3x^{2}-1x^{3},&b_{1,3}(x)&=0+3x-6x^{2}+3x^{3},&b_{2,3}(x)&=0+0x+3x^{2}-3x^{3},&b_{3,3}(x)&=0+0x+0x^{2}+1x^{3}\end{aligned}}}

Properties

The Bernstein basis polynomials have the following properties:

  • b ν , n ( x ) = 0 {\displaystyle b_{\nu ,n}(x)=0} , if ν < 0 {\displaystyle \nu <0} or ν > n . {\displaystyle \nu >n.}
  • b ν , n ( x ) 0 {\displaystyle b_{\nu ,n}(x)\geq 0} for x [ 0 ,   1 ] . {\displaystyle x\in .}
  • b ν , n ( 1 x ) = b n ν , n ( x ) . {\displaystyle b_{\nu ,n}\left(1-x\right)=b_{n-\nu ,n}(x).}
  • b ν , n ( 0 ) = δ ν , 0 {\displaystyle b_{\nu ,n}(0)=\delta _{\nu ,0}} and b ν , n ( 1 ) = δ ν , n {\displaystyle b_{\nu ,n}(1)=\delta _{\nu ,n}} where δ {\displaystyle \delta } is the Kronecker delta function: δ i j = { 0 if  i j , 1 if  i = j . {\displaystyle \delta _{ij}={\begin{cases}0&{\text{if }}i\neq j,\\1&{\text{if }}i=j.\end{cases}}}
  • b ν , n ( x ) {\displaystyle b_{\nu ,n}(x)} has a root with multiplicity ν {\displaystyle \nu } at point x = 0 {\displaystyle x=0} (note: if ν = 0 {\displaystyle \nu =0} , there is no root at 0).
  • b ν , n ( x ) {\displaystyle b_{\nu ,n}(x)} has a root with multiplicity ( n ν ) {\displaystyle \left(n-\nu \right)} at point x = 1 {\displaystyle x=1} (note: if ν = n {\displaystyle \nu =n} , there is no root at 1).
  • The derivative can be written as a combination of two polynomials of lower degree: b ν , n ( x ) = n ( b ν 1 , n 1 ( x ) b ν , n 1 ( x ) ) . {\displaystyle b'_{\nu ,n}(x)=n\left(b_{\nu -1,n-1}(x)-b_{\nu ,n-1}(x)\right).}
  • The k-th derivative at 0: b ν , n ( k ) ( 0 ) = n ! ( n k ) ! ( k ν ) ( 1 ) ν + k . {\displaystyle b_{\nu ,n}^{(k)}(0)={\frac {n!}{(n-k)!}}{\binom {k}{\nu }}(-1)^{\nu +k}.}
  • The k-th derivative at 1: b ν , n ( k ) ( 1 ) = ( 1 ) k b n ν , n ( k ) ( 0 ) . {\displaystyle b_{\nu ,n}^{(k)}(1)=(-1)^{k}b_{n-\nu ,n}^{(k)}(0).}
  • The transformation of the Bernstein polynomial to monomials is b ν , n ( x ) = ( n ν ) k = 0 n ν ( n ν k ) ( 1 ) n ν k x ν + k = = ν n ( n ) ( ν ) ( 1 ) ν x , {\displaystyle b_{\nu ,n}(x)={\binom {n}{\nu }}\sum _{k=0}^{n-\nu }{\binom {n-\nu }{k}}(-1)^{n-\nu -k}x^{\nu +k}=\sum _{\ell =\nu }^{n}{\binom {n}{\ell }}{\binom {\ell }{\nu }}(-1)^{\ell -\nu }x^{\ell },} and by the inverse binomial transformation, the reverse transformation is x k = i = 0 n k ( n k i ) 1 ( n i ) b n i , n ( x ) = 1 ( n k ) j = k n ( j k ) b j , n ( x ) . {\displaystyle x^{k}=\sum _{i=0}^{n-k}{\binom {n-k}{i}}{\frac {1}{\binom {n}{i}}}b_{n-i,n}(x)={\frac {1}{\binom {n}{k}}}\sum _{j=k}^{n}{\binom {j}{k}}b_{j,n}(x).}
  • The indefinite integral is given by b ν , n ( x ) d x = 1 n + 1 j = ν + 1 n + 1 b j , n + 1 ( x ) . {\displaystyle \int b_{\nu ,n}(x)\,dx={\frac {1}{n+1}}\sum _{j=\nu +1}^{n+1}b_{j,n+1}(x).}
  • The definite integral is constant for a given n: 0 1 b ν , n ( x ) d x = 1 n + 1   for all  ν = 0 , 1 , , n . {\displaystyle \int _{0}^{1}b_{\nu ,n}(x)\,dx={\frac {1}{n+1}}\quad \ \,{\text{for all }}\nu =0,1,\dots ,n.}
  • If n 0 {\displaystyle n\neq 0} , then b ν , n ( x ) {\displaystyle b_{\nu ,n}(x)} has a unique local maximum on the interval [ 0 , 1 ] {\displaystyle } at x = ν n {\displaystyle x={\frac {\nu }{n}}} . This maximum takes the value ν ν n n ( n ν ) n ν ( n ν ) . {\displaystyle \nu ^{\nu }n^{-n}\left(n-\nu \right)^{n-\nu }{n \choose \nu }.}
  • The Bernstein basis polynomials of degree n {\displaystyle n} form a partition of unity: ν = 0 n b ν , n ( x ) = ν = 0 n ( n ν ) x ν ( 1 x ) n ν = ( x + ( 1 x ) ) n = 1. {\displaystyle \sum _{\nu =0}^{n}b_{\nu ,n}(x)=\sum _{\nu =0}^{n}{n \choose \nu }x^{\nu }\left(1-x\right)^{n-\nu }=\left(x+\left(1-x\right)\right)^{n}=1.}
  • By taking the first x {\displaystyle x} -derivative of ( x + y ) n {\displaystyle (x+y)^{n}} , treating y {\displaystyle y} as constant, then substituting the value y = 1 x {\displaystyle y=1-x} , it can be shown that ν = 0 n ν b ν , n ( x ) = n x . {\displaystyle \sum _{\nu =0}^{n}\nu b_{\nu ,n}(x)=nx.}
  • Similarly the second x {\displaystyle x} -derivative of ( x + y ) n {\displaystyle (x+y)^{n}} , with y {\displaystyle y} again then substituted y = 1 x {\displaystyle y=1-x} , shows that ν = 1 n ν ( ν 1 ) b ν , n ( x ) = n ( n 1 ) x 2 . {\displaystyle \sum _{\nu =1}^{n}\nu (\nu -1)b_{\nu ,n}(x)=n(n-1)x^{2}.}
  • A Bernstein polynomial can always be written as a linear combination of polynomials of higher degree: b ν , n 1 ( x ) = n ν n b ν , n ( x ) + ν + 1 n b ν + 1 , n ( x ) . {\displaystyle b_{\nu ,n-1}(x)={\frac {n-\nu }{n}}b_{\nu ,n}(x)+{\frac {\nu +1}{n}}b_{\nu +1,n}(x).}
  • The expansion of the Chebyshev Polynomials of the First Kind into the Bernstein basis is T n ( u ) = ( 2 n 1 ) ! ! k = 0 n ( 1 ) n k ( 2 k 1 ) ! ! ( 2 n 2 k 1 ) ! ! b k , n ( u ) . {\displaystyle T_{n}(u)=(2n-1)!!\sum _{k=0}^{n}{\frac {(-1)^{n-k}}{(2k-1)!!(2n-2k-1)!!}}b_{k,n}(u).}

Approximating continuous functions

Let ƒ be a continuous function on the interval . Consider the Bernstein polynomial

B n ( f ) ( x ) = ν = 0 n f ( ν n ) b ν , n ( x ) . {\displaystyle B_{n}(f)(x)=\sum _{\nu =0}^{n}f\left({\frac {\nu }{n}}\right)b_{\nu ,n}(x).}

It can be shown that

lim n B n ( f ) = f {\displaystyle \lim _{n\to \infty }{B_{n}(f)}=f}

uniformly on the interval .

Bernstein polynomials thus provide one way to prove the Weierstrass approximation theorem that every real-valued continuous function on a real interval can be uniformly approximated by polynomial functions over  R {\displaystyle \mathbb {R} } .

A more general statement for a function with continuous k derivative is

B n ( f ) ( k ) ( n ) k n k f ( k )   and   f ( k ) B n ( f ) ( k ) 0 , {\displaystyle {\left\|B_{n}(f)^{(k)}\right\|}_{\infty }\leq {\frac {(n)_{k}}{n^{k}}}\left\|f^{(k)}\right\|_{\infty }\quad \ {\text{and}}\quad \ \left\|f^{(k)}-B_{n}(f)^{(k)}\right\|_{\infty }\to 0,}

where additionally

( n ) k n k = ( 1 0 n ) ( 1 1 n ) ( 1 k 1 n ) {\displaystyle {\frac {(n)_{k}}{n^{k}}}=\left(1-{\frac {0}{n}}\right)\left(1-{\frac {1}{n}}\right)\cdots \left(1-{\frac {k-1}{n}}\right)}

is an eigenvalue of Bn; the corresponding eigenfunction is a polynomial of degree k.

Probabilistic proof

This proof follows Bernstein's original proof of 1912. See also Feller (1966) or Koralov & Sinai (2007).

Motivation

We will first give intuition for Bernstein's original proof. A continuous function on a compact interval must be uniformly continuous. Thus, the value of any continuous function can be uniformly approximated by its value on some finite net of points in the interval. This consideration renders the approximation theorem intuitive, given that polynomials should be flexible enough to match (or nearly match) a finite number of pairs ( x , f ( x ) ) {\displaystyle (x,f(x))} . To do so, we might (1) construct a function close to f {\displaystyle f} on a lattice, and then (2) smooth out the function outside the lattice to make a polynomial.

The probabilistic proof below simply provides a constructive method to create a polynomial which is approximately equal to f {\displaystyle f} on such a point lattice, given that "smoothing out" a function is not always trivial. Taking the expectation of a random variable with a simple distribution is a common way to smooth. Here, we take advantage of the fact that Bernstein polynomials look like Binomial expectations. We split the interval into a lattice of n discrete values. Then, to evaluate any f(x), we evaluate f at one of the n lattice points close to x, randomly chosen by the Binomial distribution. The expectation of this approximation technique is polynomial, as it is the expectation of a function of a binomial RV. The proof below illustrates that this achieves a uniform approximation of f. The crux of the proof is to (1) justify replacing an arbitrary point with a binomially chosen lattice point by concentration properties of a Binomial distribution, and (2) justify the inference from x X {\displaystyle x\approx X} to f ( x ) f ( X ) {\displaystyle f(x)\approx f(X)} by uniform continuity.

Bernstein's proof

Suppose K is a random variable distributed as the number of successes in n independent Bernoulli trials with probability x of success on each trial; in other words, K has a binomial distribution with parameters n and x. Then we have the expected value E [ K n ] = x   {\displaystyle \operatorname {\mathcal {E}} \left=x\ } and

p ( K ) = ( n K ) x K ( 1 x ) n K = b K , n ( x ) {\displaystyle p(K)={n \choose K}x^{K}\left(1-x\right)^{n-K}=b_{K,n}(x)}

By the weak law of large numbers of probability theory,

lim n P ( | K n x | > δ ) = 0 {\displaystyle \lim _{n\to \infty }{P\left(\left|{\frac {K}{n}}-x\right|>\delta \right)}=0}

for every δ > 0. Moreover, this relation holds uniformly in x, which can be seen from its proof via Chebyshev's inequality, taking into account that the variance of 1⁄n K, equal to 1⁄n x(1−x), is bounded from above by 1⁄(4n) irrespective of x.

Because ƒ, being continuous on a closed bounded interval, must be uniformly continuous on that interval, one infers a statement of the form

lim n P ( | f ( K n ) f ( x ) | > ε ) = 0 {\displaystyle \lim _{n\to \infty }{P\left(\left|f\left({\frac {K}{n}}\right)-f\left(x\right)\right|>\varepsilon \right)}=0}

uniformly in x for each ϵ > 0 {\displaystyle \epsilon >0} . Taking into account that ƒ is bounded (on the given interval) one finds that

lim n E ( | f ( K n ) f ( x ) | ) = 0 {\displaystyle \lim _{n\to \infty }{\operatorname {\mathcal {E}} \left(\left|f\left({\frac {K}{n}}\right)-f\left(x\right)\right|\right)}=0}

uniformly in x. To justify this statement, we use a common method in probability theory to convert from closeness in probability to closeness in expectation. One splits the expectation of | f ( K n ) f ( x ) | {\displaystyle \left|f\left({\frac {K}{n}}\right)-f\left(x\right)\right|} into two parts split based on whether or not | f ( K n ) f ( x ) | < ϵ {\displaystyle \left|f\left({\frac {K}{n}}\right)-f\left(x\right)\right|<\epsilon } . In the interval where the difference does not exceed ε, the expectation clearly cannot exceed ε. In the other interval, the difference still cannot exceed 2M, where M is an upper bound for |ƒ(x)| (since uniformly continuous functions are bounded). However, by our 'closeness in probability' statement, this interval cannot have probability greater than ε. Thus, this part of the expectation contributes no more than 2M times ε. Then the total expectation is no more than ϵ + 2 M ϵ {\displaystyle \epsilon +2M\epsilon } , which can be made arbitrarily small by choosing small ε.

Finally, one observes that the absolute value of the difference between expectations never exceeds the expectation of the absolute value of the difference, a consequence of Holder's Inequality. Thus, using the above expectation, we see that (uniformly in x)

lim n | E f ( K n ) E f ( x ) | lim n E ( | f ( K n ) f ( x ) | ) = 0 {\displaystyle \lim _{n\to \infty }{\left|\operatorname {\mathcal {E}} f\left({\frac {K}{n}}\right)-\operatorname {\mathcal {E}} f\left(x\right)\right|}\leq \lim _{n\to \infty }{\operatorname {\mathcal {E}} \left(\left|f\left({\frac {K}{n}}\right)-f\left(x\right)\right|\right)}=0}

Noting that our randomness was over K while x is constant, the expectation of f(x) is just equal to f(x). But then we have shown that E x f ( K n ) {\displaystyle \operatorname {{\mathcal {E}}_{x}} f\left({\frac {K}{n}}\right)} converges to f(x). Then we will be done if E x f ( K n ) {\displaystyle \operatorname {{\mathcal {E}}_{x}} f\left({\frac {K}{n}}\right)} is a polynomial in x (the subscript reminding us that x controls the distribution of K). Indeed it is:

E x [ f ( K n ) ] = K = 0 n f ( K n ) p ( K ) = K = 0 n f ( K n ) b K , n ( x ) = B n ( f ) ( x ) {\displaystyle \operatorname {{\mathcal {E}}_{x}} \left=\sum _{K=0}^{n}f\left({\frac {K}{n}}\right)p(K)=\sum _{K=0}^{n}f\left({\frac {K}{n}}\right)b_{K,n}(x)=B_{n}(f)(x)}

Uniform convergence rates between functions

In the above proof, recall that convergence in each limit involving f depends on the uniform continuity of f, which implies a rate of convergence dependent on f 's modulus of continuity ω . {\displaystyle \omega .} It also depends on 'M', the absolute bound of the function, although this can be bypassed if one bounds ω {\displaystyle \omega } and the interval size. Thus, the approximation only holds uniformly across x for a fixed f, but one can readily extend the proof to uniformly approximate a set of functions with a set of Bernstein polynomials in the context of equicontinuity.

Elementary proof

The probabilistic proof can also be rephrased in an elementary way, using the underlying probabilistic ideas but proceeding by direct verification:

The following identities can be verified:

  1. k ( n k ) x k ( 1 x ) n k = 1 {\displaystyle \sum _{k}{n \choose k}x^{k}(1-x)^{n-k}=1} ("probability")
  2. k k n ( n k ) x k ( 1 x ) n k = x {\displaystyle \sum _{k}{k \over n}{n \choose k}x^{k}(1-x)^{n-k}=x} ("mean")
  3. k ( x k n ) 2 ( n k ) x k ( 1 x ) n k = x ( 1 x ) n . {\displaystyle \sum _{k}\left(x-{k \over n}\right)^{2}{n \choose k}x^{k}(1-x)^{n-k}={x(1-x) \over n}.} ("variance")

In fact, by the binomial theorem

( 1 + t ) n = k ( n k ) t k , {\displaystyle (1+t)^{n}=\sum _{k}{n \choose k}t^{k},}

and this equation can be applied twice to t d d t {\displaystyle t{\frac {d}{dt}}} . The identities (1), (2), and (3) follow easily using the substitution t = x / ( 1 x ) {\displaystyle t=x/(1-x)} .

Within these three identities, use the above basis polynomial notation

b k , n ( x ) = ( n k ) x k ( 1 x ) n k , {\displaystyle b_{k,n}(x)={n \choose k}x^{k}(1-x)^{n-k},}

and let

f n ( x ) = k f ( k / n ) b k , n ( x ) . {\displaystyle f_{n}(x)=\sum _{k}f(k/n)\,b_{k,n}(x).}

Thus, by identity (1)

f n ( x ) f ( x ) = k [ f ( k / n ) f ( x ) ] b k , n ( x ) , {\displaystyle f_{n}(x)-f(x)=\sum _{k}\,b_{k,n}(x),}

so that

| f n ( x ) f ( x ) | k | f ( k / n ) f ( x ) | b k , n ( x ) . {\displaystyle |f_{n}(x)-f(x)|\leq \sum _{k}|f(k/n)-f(x)|\,b_{k,n}(x).}

Since f is uniformly continuous, given ε > 0 {\displaystyle \varepsilon >0} , there is a δ > 0 {\displaystyle \delta >0} such that | f ( a ) f ( b ) | < ε {\displaystyle |f(a)-f(b)|<\varepsilon } whenever | a b | < δ {\displaystyle |a-b|<\delta } . Moreover, by continuity, M = sup | f | < {\displaystyle M=\sup |f|<\infty } . But then

| f n ( x ) f ( x ) | | x k n | < δ | f ( k / n ) f ( x ) | b k , n ( x ) + | x k n | δ | f ( k / n ) f ( x ) | b k , n ( x ) . {\displaystyle |f_{n}(x)-f(x)|\leq \sum _{|x-{k \over n}|<\delta }|f(k/n)-f(x)|\,b_{k,n}(x)+\sum _{|x-{k \over n}|\geq \delta }|f(k/n)-f(x)|\,b_{k,n}(x).}

The first sum is less than ε. On the other hand, by identity (3) above, and since | x k / n | δ {\displaystyle |x-k/n|\geq \delta } , the second sum is bounded by 2 M {\displaystyle 2M} times

| x k / n | δ b k , n ( x ) k δ 2 ( x k n ) 2 b k , n ( x ) = δ 2 x ( 1 x ) n < 1 4 δ 2 n 1 . {\displaystyle \sum _{|x-k/n|\geq \delta }b_{k,n}(x)\leq \sum _{k}\delta ^{-2}\left(x-{k \over n}\right)^{2}b_{k,n}(x)=\delta ^{-2}{x(1-x) \over n}<{1 \over 4}\delta ^{-2}n^{-1}.}
(Chebyshev's inequality)

It follows that the polynomials fn tend to f uniformly.

Generalizations to higher dimension

Bernstein polynomials can be generalized to k dimensions – the resulting polynomials have the form Bi1(x1) Bi2(x2) ... Bik(xk). In the simplest case only products of the unit interval are considered; but, using affine transformations of the line, Bernstein polynomials can also be defined for products × × ... × . For a continuous function f on the k-fold product of the unit interval, the proof that f(x1, x2, ... , xk) can be uniformly approximated by

i 1 i 2 i k ( n 1 i 1 ) ( n 2 i 2 ) ( n k i k ) f ( i 1 n 1 , i 2 n 2 , , i k n k ) x 1 i 1 ( 1 x 1 ) n 1 i 1 x 2 i 2 ( 1 x 2 ) n 2 i 2 x k i k ( 1 x k ) n k i k {\displaystyle \sum _{i_{1}}\sum _{i_{2}}\cdots \sum _{i_{k}}{n_{1} \choose i_{1}}{n_{2} \choose i_{2}}\cdots {n_{k} \choose i_{k}}f\left({i_{1} \over n_{1}},{i_{2} \over n_{2}},\dots ,{i_{k} \over n_{k}}\right)x_{1}^{i_{1}}(1-x_{1})^{n_{1}-i_{1}}x_{2}^{i_{2}}(1-x_{2})^{n_{2}-i_{2}}\cdots x_{k}^{i_{k}}(1-x_{k})^{n_{k}-i_{k}}}

is a straightforward extension of Bernstein's proof in one dimension.

See also

Notes

  1. ^ Lorentz 1953
  2. Mathar, R. J. (2018). "Orthogonal basis function over the unit circle with the minimax property". Appendix B. arXiv:1802.09518 .
  3. Rababah, Abedallah (2003). "Transformation of Chebyshev-Bernstein Polynomial Basis". Comp. Meth. Appl. Math. 3 (4): 608–622. doi:10.2478/cmam-2003-0038. S2CID 120938358.
  4. Natanson (1964) p. 6
  5. ^ Feller 1966
  6. ^ Beals 2004
  7. Natanson (1964) p. 3
  8. Bernstein 1912
  9. Koralov, L.; Sinai, Y. (2007). ""Probabilistic proof of the Weierstrass theorem"". Theory of probability and random processes (2nd ed.). Springer. p. 29.
  10. Lorentz 1953, pp. 5–6
  11. Goldberg 1964
  12. Akhiezer 1956
  13. Burkill 1959
  14. Hildebrandt, T. H.; Schoenberg, I. J. (1933), "On linear functional operations and the moment problem for a finite interval in one or several dimensions", Annals of Mathematics, 34 (2): 327, doi:10.2307/1968205, JSTOR 1968205

References

External links

Categories: