The characteristic function corresponding to the m-dimensional marginal distribution of m out of n variables X1,.. , X n can be obtained from the characteristic function 20.2.5.29 if the
Trang 1For continuous random variables,
f(t) =
+∞ –∞ .
+∞ –∞ exp
in j=1t j x j
p (x) dx.
The inversion formula for a continuous distribution has the form
p(x) = 1
(2π)n
+∞ –∞ .
+∞ –∞ exp
–in
j=1t j x j
f (t) dt,
where dt = dt1 dt n.
If the initial moments of a random variable X exist, then
E{X r1
1 X n r n}= i–
n j=1r j ∂ r1 +···+r n f(t)
∂t r1
1 ∂t r n n
t1 =···=t n 0.
The characteristic function corresponding to the m-dimensional marginal distribution
of m out of n variables X1, , X n can be obtained from the characteristic function
(20.2.5.29) if the variables t j corresponding to the random variables that are not contained
in the m-dimensional marginal distribution are replaced by zeros.
CONTINUITY THEOREM FOR CHARACTERISTIC FUNCTIONS The weak convergence
F X n → F of a sequence of distribution functions F1 (x), F2(x), is equivalent to the uniform convergence f n (t) → f(t) of characteristic functions on each finite interval.
20.2.5-12 Independency of random variables
Random variables X1, , X n are said to be independent if the events{X1 S1}, ,
{X n S n}are independent for any measurable sets S1, , S n For this, it is necessary and sufficient that
P (X1 S1, , X nS n) =
n
k=1
P (X kS k). (20.2.5.30)
Relation (20.2.5.30) is equivalent to one of the following three:
1 In the general case: for any xRn,
FX (x) =
n
k=1
F X k (x k)
2 For absolutely continuous distributions: for any x Rn (except possibly for a set of
measure zero),
pX (x) =
n
k=1
p X k (x k)
3 For discrete distributions: for any xRn,
P (X1 = x1, , X n = x n) =
n
k=1
P (X k = x k)
The joint distribution of independent random variables is uniquely determined by their individual distributions Independent random variables are uncorrelated, but the converse
is not true in general
Random variables X1, , X nare independent if and only if the characteristic function
of the multivariate random variable X is equal to the product of the characteristic functions
of the random variables X1, , X n
Trang 21068 PROBABILITYTHEORY
20.3 Limit Theorems
20.3.1 Convergence of Random Variables
20.3.1-1 Convergence in probability
A sequence of random variables X1, X2, is said to converge in probability to a random variable X(X n −→ X) if P
lim
n→∞ P(|X n – X| ≥ε) =0 (20.3.1.1)
for each ε >0, i.e., if for any ε >0and δ >0there exists a number N , depending on ε and
δ, such that the inequality
P(|X n – X|> ε) < δ
holds for n > N A sequence of k-dimensional random variables X nis said to converge in
probability to a random variable X if each coordinate of the random variable Xnconverges
in probability to the respective coordinate of the random variable X.
20.3.1-2 Almost sure convergence (convergence with probability 1)
A sequence of random variables X1, X2, is said to converge almost surely (or with
probability1) to a random variable X (X n −→ X) ifa.s.
P [ωΩ : lim
n→∞ X n (ω) = X(ω)] =1 (20.3.1.2)
A sequence X n → X converges almost surely if and only if
P
∞
;
m=1
{|X n+m – X| ≥ε} −→
n→∞0
for each ε >0
Convergence almost surely implies convergence in probability The converse statement
is not true in general
20.3.1-3 Convergence in mean
A sequence of random variables X1, X2, with finite pth initial moments (p =1,2, )
is said to converge in pth mean to a random variable X (E{X p}<∞) if
lim
n→∞ E{|X n – X|p}=0 (20.3.1.3)
Convergence in pth mean, for p =2is called convergence in mean square If X n → X
in pth mean then X n → X in p1 th mean for all p1≤p
Convergence in pth mean implies convergence in probability The converse statement
is not true in general
Trang 320.3.1-4 Convergence in distribution.
Suppose that a sequence F1(x), F2(x), of cumulative distribution functions converges
to a distribution function F (x),
lim
n→∞ F n (x) = F (x), (20.3.1.4)
for every point x at which F (x) In this case, we say that the sequence X1, X2, of the corresponding random variables converges to the random variable X in distribution The random variables X1, X2, can be defined on different probability spaces.
A sequence F1(x), F2(x), of distribution functions weakly converges to a distribution function F (x) (F n → F ) if
lim
n→∞ E{h (X n)}= E{h (X)} (20.3.1.5)
for any bounded continuous function h as n → ∞.
Convergence in distribution and weak convergence of distribution functions are equiv-alent
The weak convergence F X n → F for random variables having a probability density
function means the convergence
+∞ –∞ g (x)p X n (x) dx →
+∞ –∞ g (x)p(x) dx (20.3.1.6)
for any bounded continuous function g(x).
20.3.2 Limit Theorems
20.3.2-1 Law of large numbers
The law of large numbers consists of several theorems establishing the stability of average results and revealing conditions for this stability to occur
The notion of convergence in probability is most often used for the case in which the limit
random variable X has the degenerate distribution concentrated at a point a (P (ξ = a) =1) and
X n= 1
n
n
k=1
Y k, (20.3.2.1)
where Y1, Y2, are arbitrary random variables.
A sequence Y1, Y2, satisfies the weak law of large numbers if the limit relation
lim
n→∞ P
1
n
n
k=1
Y k – a≥ε
≡ lim
n→∞ P(|X n – a| ≥ε) =0 (20.3.2.2)
holds for any ε >0
If the relation
P
ω Ω : lim
n→∞
1
n
n
k=1
Y k = a
≡P ωΩ : lim
n→∞ X n = a
=1 (20.3.2.3)
Trang 41070 PROBABILITYTHEORY
is satisfied instead of (20.3.2.2), i.e., the sequence X n converges to the number a with
probability1, then the sequence Y1, Y2, satisfies the strong law of large numbers.
Markov inequality For any nonnegative random variable X that has an expectation
E{X}, the inequality
P (X ≥ε)≤ E{X}
holds for each ε >0
Chebyshev inequality For any random variable X with finite variance, the inequality
P(|X – E{X}| ≥ε)≤ Var{X}
ε2 (20.3.2.5)
holds for each ε >0
CHEBYSHEV THEOREM If X1, X2, is a sequence of independent random variables
with uniformly bounded finite variances, Var{X1} ≤C, Var{X2} ≤C , , then the limit
relation
lim
n→∞ P
1
n
n
k=1
X k– 1
n
n
k=1
E{X k} < ε
=1 (20.3.2.6)
holds for each ε >0
BERNOULLI THEOREM Let μ n be the number of occurrences of an event A (the number
of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of the trials Then the sequence of relative frequencies μ n /n of the occurrence of the event A in n independent trials converges in probability to p = P (A) as n → ∞; i.e., the limit relation
lim
n→∞ P
μ n
n – p < ε
=1 (20.3.2.7)
holds for each ε >0
POISSON THEOREM If in a sequence of independent trials the probability that an event
A occurs in the kth trial is equal to p k, then
lim
n→∞ P
μ n
n – p1+· · · + p n
n
< ε=1 (20.3.2.8)
KOLMOGOROV THEOREM If a sequence of independent random variables X1, X2,
satisfies the condition
∞
k=1
Var{X k}
k2 < +∞, (20.3.2.9) then it obeys the strong law of large numbers
The existence of the expectation is a necessary and sufficient condition for the strong law of large numbers to apply to a sequence of independent identically distributed random variables
Trang 520.3.2-2 Central limit theorem.
A random variable X n with distribution function F X n is asymptotically normally distributed
if there exists a sequence of pairs of real numbers m n , σ n2 such that the random variables
(X n – m n /σ n) converge in probability to a standard normal variable This occurs if and
only if the limit relation
lim
n→∞ P (X n + aσ n < X n < X n + bσ n) =Φ(b) – Φ(a), (20.3.2.10) whereΦ(x) is the distribution function of the standard normal law, holds for any a and b (b > a).
LYAPUNOV CENTRAL LIMIT THEOREM If X1, , X n is a sequence of independent random variables satisfying the Lyapunov condition
lim
n→∞
n
k=1α3(X k)
n
k=1Var{X k} =0,
where α3(X k)is the third initial moment of the random variable X k, then the sequence of random variables
Y n=
n
k=1(X k – E{X k}) n
k=1Var{X k}
converges in distribution to the normal law, i.e., the following limit exists:
lim
n→∞ P
n
k=1(X k – E{X k}) n
k=1Var{X k} < t
= √1
2π
t –∞ e
–t2/ dt=Φ(t). (20.3.2.11)
LINDEBERG CENTRAL LIMIT THEOREM Let X1, X2, be a sequence of independent identically distributed random variables with finite expectation E{X k} = m and finite
variance σ2 Then, as n → ∞, the random variable 1n n
k=1X khas an asymptotically normal
probability distribution with parameters (m, σ2/n)
Let μ n be the number of occurrences of an event A (the number of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of the trials Then the sequence of relative frequencies μ n /n
has an asymptotically normal probability distribution with parameters (p, p(1– p)/n).
20.4 Stochastic Processes
20.4.1 Theory of Stochastic Processes
20.4.1-1 Notion of stochastic process
Let a family
ξ (t) = ξ(ω, t), ωΩ, (20.4.1.1)
of random variables depending on a parameter t T be given on a probability space (Ω, F, P ) The variable ξ(t), t T, can be treated as a random function of the variable
tT The values of this function are the values of the random variable (20.4.1.1) The
random function ξ(t) of the independent variable t is called a stochastic process If a random
Trang 61072 PROBABILITYTHEORY
outcome ωΩ occurs, then the actual process is described by the corresponding trajectory,
which is called a realization of the process, or a sample function.
A stochastic process can be simply a numerical function ξ(t) of time admitting different
realizations ξ(ω, t) (one-dimensional stochastic process) or a vector function Ξ(t)
(multi-dimensional, or vector, stochastic process) The study of a multidimensional stochastic
process can be reduced to the study of one-dimensional stochastic processes by a
transfor-mation taking Ξ(t) = (ξ1(t), , ξ n (t)) to the auxiliary process
ξa(t) = Ξ(t)⋅a =
n
i=1
a i ξ i (t), (20.4.1.2)
where a = (a1, , a n ) is an arbitrary k-dimensional vector Therefore, the study of one-dimensional stochastic processes ξ(t) is the main point in the theory of stochastic processes.
If the parameter t ranges in some interval of the real line R, then the stochastic process is called a stochastic process with continuous time, and if the parameter t takes integer values, then the process is called a stochastic process with discrete time (a random sequence).
To describe a stochastic process, one should specify an infinite set of compatible
finite-dimensional probability distributions of the random vectors Ξ(t) corresponding to all pos-sible finite subsets t = (t1, , t n) of values of the argument
Remark Specifying compatible finite-dimensional probability distributions of random vectors may be
insufficient for specifying the probabilities of events depending on the values of ξ(t) on an infinite set of values
of the parameter t; i.e., this does not uniquely determine the stochastic process ξ(t).
Example Suppose that ξ(t) = cos(ωt +Φ), 0 ≤t≤ 1, is a harmonic oscillation with random phaseΦ, Z
is a random variable uniformly distributed on the interval [0, 1], and the stochastic process ζ(t), 0 ≤t≤ 1, is given by the relation
ζ (t) =
ξ (t) for t≠Z,
ξ (t) +3 for t = Z.
Since P [(Z = t1 )∪ · · · ∪ (Z = t n)
= 0for any finite set t = (t1, , t n), we see that all finite-dimensional
distributions of the stochastic processes ξ(t) and ζ(t) are the same At the same time, these processes differ
from each other.
Specifying the set of finite-dimensional probability distributions often permits one to
clarify whether there exists at least one stochastic process ξ(t) with finite-dimensional
distributions whose realizations satisfy a certain property (for example, are continuous or differentiable)
20.4.1-2 Correlation function
Let ξ(t) and ζ(t) be real stochastic processes.
The autocorrelation function of a stochastic process is defined as the function
B ξξ (t, s) = E5
ξ (t) – E{ξ (t)}ξ (s) – E{ξ (s)}6, (20.4.1.3) which is the second central moment function
Remark. The autocorrelation function of a stochastic process is also called the covariance function.
The mixed second moment, i.e., the function
B ξζ (t, s) = E{ξ (t)ζ(s)}, (20.4.1.4)
of values of ξ(t) and ζ(t) at two points is called the cross-correlation function
(cross-covariance function).
Trang 7The mixed central moment (covariance), i.e., the function
b ξζ (t, s) = E5
ξ (t) – E{ξ (t)}ζ (s) – E{ζ (s)}6, (20.4.1.5)
is called the central cross-correlation function.
The correlation coefficient, i.e., the function
ρ ξζ (t, s) = √ b ξζ (t, s)
Var{ξ (t)}Var{ζ (s)}, (20.4.1.6)
is called the normalized cross-correlation function.
The following relations hold:
E{ξ2(t)}= B ξξ (t, t)≥ 0,
|B ξξ (t, s)|2 ≤B ξξ (t, t)B ξξ (s, s),
Var{ξ (t)}= B ξξ (t, t) – [E{ξ (t)}]2,
Cov[ξ(t), ζ(s)] = B ξζ (t, s) – ξ(t)ζ(s).
Suppose that ξ(t) and ζ(t) are complex stochastic processes (essentially, two-dimensional
stochastic processes) The autocorrelation function and the cross-correlation function are determined by the relations
B ξξ (t, s) = E5
ξ (t)ξ(s)6
= B ξξ (s, t), B ξζ (t, s) = E5
ξ (t)ζ(s)6
= B ζξ (s, t), (20.4.1.7)
where ξ(t), B ξξ (s, t), and B ζξ (s, t) are the function conjugate to a function ξ(t), B ξξ (s, t), and B ζξ (s, t), respectively.
20.4.1-3 Differentiation and integration of stochastic process
To differentiate a stochastic process is to calculate the limit
lim
h→0
ξ (t + h) – ξ(t)
h = ξ t (t). (20.4.1.8)
If the limit is understood in the sense of convergence in mean square (resp., with probabil-ity 1), then the differentiation is also understood in mean square (resp., with probabilprobabil-ity 1) The following formulas hold:
E{ξ
t (t)}=dE{ξ (t)}
dt , B ξ
t ξ
t (t, s) = ∂
2B
ξξ (t, s)
∂t∂s , B ξ
t ξ (t, s) = ∂B ξξ ∂t (t, s) (20.4.1.9)
a ξ (t) dt (20.4.1.10)
of a stochastic process ξ(t) defined on an interval [a, b] with autocorrelation function
B ξξ (t, s) is the limit in mean square of the integral sums
n
k=1
ξ (s k )(t k – t k–1)
as the diameter of the partition a = t0< t1<· · · < t n = b, where s k [t k , t k–1] tends to zero
... number of occurrences of an event A (the number of successes) in n independent trials, and let p = P (A) be the probability of the occurrence of the event A (the probability of success) in each of. .. as a random function of the variabletT The values of this function are the values of the random variable (20.4.1.1) The
random function ξ(t) of the... n) of values of the argument
Remark Specifying compatible finite-dimensional probability distributions of random vectors may be
insufficient for specifying