1. Trang chủ
  2. » Khoa Học Tự Nhiên

Handbook of mathematics for engineers and scienteists part 150 ppt

7 86 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 7
Dung lượng 413,71 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

The characteristic function corresponding to the m-dimensional marginal distribution of m out of n variables X1,.. , X n can be obtained from the characteristic function 20.2.5.29 if the

Trang 1

For continuous random variables,

f(t) =

 +∞ .

 + exp



in j=1t j x j



p (x) dx.

The inversion formula for a continuous distribution has the form

p(x) = 1

(2π)n

 +∞ .

 + exp



–in

j=1t j x j



f (t) dt,

where dt = dt1 dt n.

If the initial moments of a random variable X exist, then

E{X r1

1 X n r n}= i

n j=1r j ∂ r1 +···+r n f(t)

∂t r1

1 ∂t r n n





t1 =···=t n 0.

The characteristic function corresponding to the m-dimensional marginal distribution

of m out of n variables X1, , X n can be obtained from the characteristic function

(20.2.5.29) if the variables t j corresponding to the random variables that are not contained

in the m-dimensional marginal distribution are replaced by zeros.

CONTINUITY THEOREM FOR CHARACTERISTIC FUNCTIONS The weak convergence

F X n → F of a sequence of distribution functions F1 (x), F2(x), is equivalent to the uniform convergence f n (t) → f(t) of characteristic functions on each finite interval.

20.2.5-12 Independency of random variables

Random variables X1, , X n are said to be independent if the events{X1 S1}, ,

{X n S n}are independent for any measurable sets S1, , S n For this, it is necessary and sufficient that

P (X1 S1, , X nS n) =

n



k=1

P (X kS k). (20.2.5.30)

Relation (20.2.5.30) is equivalent to one of the following three:

1 In the general case: for any xRn,

FX (x) =

n



k=1

F X k (x k)

2 For absolutely continuous distributions: for any x Rn (except possibly for a set of

measure zero),

pX (x) =

n



k=1

p X k (x k)

3 For discrete distributions: for any xRn,

P (X1 = x1, , X n = x n) =

n



k=1

P (X k = x k)

The joint distribution of independent random variables is uniquely determined by their individual distributions Independent random variables are uncorrelated, but the converse

is not true in general

Random variables X1, , X nare independent if and only if the characteristic function

of the multivariate random variable X is equal to the product of the characteristic functions

of the random variables X1, , X n

Trang 2

1068 PROBABILITYTHEORY

20.3 Limit Theorems

20.3.1 Convergence of Random Variables

20.3.1-1 Convergence in probability

A sequence of random variables X1, X2, is said to converge in probability to a random variable X(X n −→ X) if P

lim

n→∞ P(|X n – X| ≥ε) =0 (20.3.1.1)

for each ε >0, i.e., if for any ε >0and δ >0there exists a number N , depending on ε and

δ, such that the inequality

P(|X n – X|> ε) < δ

holds for n > N A sequence of k-dimensional random variables X nis said to converge in

probability to a random variable X if each coordinate of the random variable Xnconverges

in probability to the respective coordinate of the random variable X.

20.3.1-2 Almost sure convergence (convergence with probability 1)

A sequence of random variables X1, X2, is said to converge almost surely (or with

probability1) to a random variable X (X n −→ X) ifa.s.

P [ωΩ : lim

n→∞ X n (ω) = X(ω)] =1 (20.3.1.2)

A sequence X n → X converges almost surely if and only if

P

;

m=1

{|X n+m – X| ≥ε} −→

n→∞0

for each ε >0

Convergence almost surely implies convergence in probability The converse statement

is not true in general

20.3.1-3 Convergence in mean

A sequence of random variables X1, X2, with finite pth initial moments (p =1,2, )

is said to converge in pth mean to a random variable X (E{X p}<∞) if

lim

n→∞ E{|X n – X|p}=0 (20.3.1.3)

Convergence in pth mean, for p =2is called convergence in mean square If X n → X

in pth mean then X n → X in p1 th mean for all p1≤p

Convergence in pth mean implies convergence in probability The converse statement

is not true in general

Trang 3

20.3.1-4 Convergence in distribution.

Suppose that a sequence F1(x), F2(x), of cumulative distribution functions converges

to a distribution function F (x),

lim

n→∞ F n (x) = F (x), (20.3.1.4)

for every point x at which F (x) In this case, we say that the sequence X1, X2, of the corresponding random variables converges to the random variable X in distribution The random variables X1, X2, can be defined on different probability spaces.

A sequence F1(x), F2(x), of distribution functions weakly converges to a distribution function F (x) (F n → F ) if

lim

n→∞ E{h (X n)}= E{h (X)} (20.3.1.5)

for any bounded continuous function h as n → ∞.

Convergence in distribution and weak convergence of distribution functions are equiv-alent

The weak convergence F X n → F for random variables having a probability density

function means the convergence

 +∞ g (x)p X n (x) dx →

 +∞ g (x)p(x) dx (20.3.1.6)

for any bounded continuous function g(x).

20.3.2 Limit Theorems

20.3.2-1 Law of large numbers

The law of large numbers consists of several theorems establishing the stability of average results and revealing conditions for this stability to occur

The notion of convergence in probability is most often used for the case in which the limit

random variable X has the degenerate distribution concentrated at a point a (P (ξ = a) =1) and

X n= 1

n

n



k=1

Y k, (20.3.2.1)

where Y1, Y2, are arbitrary random variables.

A sequence Y1, Y2, satisfies the weak law of large numbers if the limit relation

lim

n→∞ P

1

n

n



k=1

Y k – a≥ε

≡ lim

n→∞ P(|X n – a| ≥ε) =0 (20.3.2.2)

holds for any ε >0

If the relation

P



ω Ω : lim

n→∞

1

n

n



k=1

Y k = a



P ωΩ : lim

n→∞ X n = a

=1 (20.3.2.3)

Trang 4

1070 PROBABILITYTHEORY

is satisfied instead of (20.3.2.2), i.e., the sequence X n converges to the number a with

probability1, then the sequence Y1, Y2, satisfies the strong law of large numbers.

Markov inequality For any nonnegative random variable X that has an expectation

E{X}, the inequality

P (Xε)≤ E{X}

holds for each ε >0

Chebyshev inequality For any random variable X with finite variance, the inequality

P(|X – E{X}| ≥ε)≤ Var{X}

ε2 (20.3.2.5)

holds for each ε >0

CHEBYSHEV THEOREM If X1, X2, is a sequence of independent random variables

with uniformly bounded finite variances, Var{X1} ≤C, Var{X2} ≤C , , then the limit

relation

lim

n→∞ P

1

n

n



k=1

X k– 1

n

n



k=1

E{X k} < ε

=1 (20.3.2.6)

holds for each ε >0

BERNOULLI THEOREM Let μ n be the number of occurrences of an event A (the number

of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of the trials Then the sequence of relative frequencies μ n /n of the occurrence of the event A in n independent trials converges in probability to p = P (A) as n → ∞; i.e., the limit relation

lim

n→∞ P

μ n

n – p < ε

=1 (20.3.2.7)

holds for each ε >0

POISSON THEOREM If in a sequence of independent trials the probability that an event

A occurs in the kth trial is equal to p k, then

lim

n→∞ P

μ n

np1+· · · + p n

n



 < ε=1 (20.3.2.8)

KOLMOGOROV THEOREM If a sequence of independent random variables X1, X2,

satisfies the condition



k=1

Var{X k}

k2 < +∞, (20.3.2.9) then it obeys the strong law of large numbers

The existence of the expectation is a necessary and sufficient condition for the strong law of large numbers to apply to a sequence of independent identically distributed random variables

Trang 5

20.3.2-2 Central limit theorem.

A random variable X n with distribution function F X n is asymptotically normally distributed

if there exists a sequence of pairs of real numbers m n , σ n2 such that the random variables

(X n – m n /σ n) converge in probability to a standard normal variable This occurs if and

only if the limit relation

lim

n→∞ P (X n + aσ n < X n < X n + bσ n) =Φ(b) – Φ(a), (20.3.2.10) whereΦ(x) is the distribution function of the standard normal law, holds for any a and b (b > a).

LYAPUNOV CENTRAL LIMIT THEOREM If X1, , X n is a sequence of independent random variables satisfying the Lyapunov condition

lim

n→∞

n

k=1α3(X k)

n

k=1Var{X k} =0,

where α3(X k)is the third initial moment of the random variable X k, then the sequence of random variables

Y n=

n

k=1(X k – E{X k}) n

k=1Var{X k}

converges in distribution to the normal law, i.e., the following limit exists:

lim

n→∞ P

 n

k=1(X k – E{X k}) n

k=1Var{X k} < t



= 1

2π

 t∞ e

t2/ dt=Φ(t). (20.3.2.11)

LINDEBERG CENTRAL LIMIT THEOREM Let X1, X2, be a sequence of independent identically distributed random variables with finite expectation E{X k} = m and finite

variance σ2 Then, as n → ∞, the random variable 1n n

k=1X khas an asymptotically normal

probability distribution with parameters (m, σ2/n)

Let μ n be the number of occurrences of an event A (the number of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of the trials Then the sequence of relative frequencies μ n /n

has an asymptotically normal probability distribution with parameters (p, p(1– p)/n).

20.4 Stochastic Processes

20.4.1 Theory of Stochastic Processes

20.4.1-1 Notion of stochastic process

Let a family

ξ (t) = ξ(ω, t), ωΩ, (20.4.1.1)

of random variables depending on a parameter t  T be given on a probability space (Ω, F, P ) The variable ξ(t), t T, can be treated as a random function of the variable

tT The values of this function are the values of the random variable (20.4.1.1) The

random function ξ(t) of the independent variable t is called a stochastic process If a random

Trang 6

1072 PROBABILITYTHEORY

outcome ωΩ occurs, then the actual process is described by the corresponding trajectory,

which is called a realization of the process, or a sample function.

A stochastic process can be simply a numerical function ξ(t) of time admitting different

realizations ξ(ω, t) (one-dimensional stochastic process) or a vector function Ξ(t)

(multi-dimensional, or vector, stochastic process) The study of a multidimensional stochastic

process can be reduced to the study of one-dimensional stochastic processes by a

transfor-mation taking Ξ(t) = (ξ1(t), , ξ n (t)) to the auxiliary process

ξa(t) = Ξ(t)a =

n



i=1

a i ξ i (t), (20.4.1.2)

where a = (a1, , a n ) is an arbitrary k-dimensional vector Therefore, the study of one-dimensional stochastic processes ξ(t) is the main point in the theory of stochastic processes.

If the parameter t ranges in some interval of the real line R, then the stochastic process is called a stochastic process with continuous time, and if the parameter t takes integer values, then the process is called a stochastic process with discrete time (a random sequence).

To describe a stochastic process, one should specify an infinite set of compatible

finite-dimensional probability distributions of the random vectors Ξ(t) corresponding to all pos-sible finite subsets t = (t1, , t n) of values of the argument

Remark Specifying compatible finite-dimensional probability distributions of random vectors may be

insufficient for specifying the probabilities of events depending on the values of ξ(t) on an infinite set of values

of the parameter t; i.e., this does not uniquely determine the stochastic process ξ(t).

Example Suppose that ξ(t) = cos(ωt +Φ), 0 ≤t≤ 1, is a harmonic oscillation with random phaseΦ, Z

is a random variable uniformly distributed on the interval [0, 1], and the stochastic process ζ(t), 0 ≤t≤ 1, is given by the relation

ζ (t) =



ξ (t) for tZ,

ξ (t) +3 for t = Z.

Since P [(Z = t1 )∪ · · · ∪ (Z = t n) 

= 0for any finite set t = (t1, , t n), we see that all finite-dimensional

distributions of the stochastic processes ξ(t) and ζ(t) are the same At the same time, these processes differ

from each other.

Specifying the set of finite-dimensional probability distributions often permits one to

clarify whether there exists at least one stochastic process ξ(t) with finite-dimensional

distributions whose realizations satisfy a certain property (for example, are continuous or differentiable)

20.4.1-2 Correlation function

Let ξ(t) and ζ(t) be real stochastic processes.

The autocorrelation function of a stochastic process is defined as the function

B ξξ (t, s) = E5

ξ (t) – E{ξ (t)}ξ (s) – E{ξ (s)}6, (20.4.1.3) which is the second central moment function

Remark. The autocorrelation function of a stochastic process is also called the covariance function.

The mixed second moment, i.e., the function

B ξζ (t, s) = E{ξ (t)ζ(s)}, (20.4.1.4)

of values of ξ(t) and ζ(t) at two points is called the cross-correlation function

(cross-covariance function).

Trang 7

The mixed central moment (covariance), i.e., the function

b ξζ (t, s) = E5

ξ (t) – E{ξ (t)}ζ (s) – E{ζ (s)}6, (20.4.1.5)

is called the central cross-correlation function.

The correlation coefficient, i.e., the function

ρ ξζ (t, s) = √ b ξζ (t, s)

Var{ξ (t)}Var{ζ (s)}, (20.4.1.6)

is called the normalized cross-correlation function.

The following relations hold:

E{ξ2(t)}= B ξξ (t, t)≥ 0,

|B ξξ (t, s)|2 ≤B ξξ (t, t)B ξξ (s, s),

Var{ξ (t)}= B ξξ (t, t) – [E{ξ (t)}]2,

Cov[ξ(t), ζ(s)] = B ξζ (t, s) – ξ(t)ζ(s).

Suppose that ξ(t) and ζ(t) are complex stochastic processes (essentially, two-dimensional

stochastic processes) The autocorrelation function and the cross-correlation function are determined by the relations

B ξξ (t, s) = E5

ξ (t)ξ(s)6

= B ξξ (s, t), B ξζ (t, s) = E5

ξ (t)ζ(s)6

= B ζξ (s, t), (20.4.1.7)

where ξ(t), B ξξ (s, t), and B ζξ (s, t) are the function conjugate to a function ξ(t), B ξξ (s, t), and B ζξ (s, t), respectively.

20.4.1-3 Differentiation and integration of stochastic process

To differentiate a stochastic process is to calculate the limit

lim

h→0

ξ (t + h) – ξ(t)

h = ξ t  (t). (20.4.1.8)

If the limit is understood in the sense of convergence in mean square (resp., with probabil-ity 1), then the differentiation is also understood in mean square (resp., with probabilprobabil-ity 1) The following formulas hold:

E{ξ 

t (t)}=dE{ξ (t)}

dt , B ξ 

t ξ 

t (t, s) = ∂

2B

ξξ (t, s)

∂t∂s , B ξ 

t ξ (t, s) = ∂B ξξ ∂t (t, s) (20.4.1.9)

a ξ (t) dt (20.4.1.10)

of a stochastic process ξ(t) defined on an interval [a, b] with autocorrelation function

B ξξ (t, s) is the limit in mean square of the integral sums

n



k=1

ξ (s k )(t k – t k–1)

as the diameter of the partition a = t0< t1<· · · < t n = b, where s k [t k , t k–1] tends to zero

... number of occurrences of an event A (the number of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of. .. as a random function of the variable

tT The values of this function are the values of the random variable (20.4.1.1) The

random function ξ(t) of the... n) of values of the argument

Remark Specifying compatible finite-dimensional probability distributions of random vectors may be

insufficient for specifying

Ngày đăng: 02/07/2014, 13:20

🧩 Sản phẩm bạn có thể quan tâm