1. Trang chủ
  2. » Khoa Học Tự Nhiên

Handbook of mathematics for engineers and scienteists part 154 doc

7 54 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 7
Dung lượng 432,45 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

Any rule table, function, graph, or otherwise that permits one to find the probabilities of events A ⊆ F is usually called the distribution law of a random variable.. In general, random

Trang 1

20.2 Random Variables and Their Characteristics

20.2.1 One-Dimensional Random Variables

20.2.1-1 Notion of random variable Distribution function of random variable

A random variable X is a real function X = X(ω), ω Ω, on a probability space Ω such that the set{ω : X(ω)x}belongs to the σ-algebra F of events for each real x.

Any rule (table, function, graph, or otherwise) that permits one to find the probabilities

of events A ⊆ F is usually called the distribution law of a random variable In general,

random variables can be discrete or continuous.

The cumulative distribution function of a random variable X is the function F X (x) whose value at each point x is equal to the probability of the event{X < x}:

F X (x) = F (x) = P (X < x). (20.2.1.1) Properties of the cumulative distribution function:

1 F (x) is bounded, i.e.,0 ≤F (x)≤ 1

2 F (x) is a nondecreasing function for x(–∞, ∞); i.e., if x2> x1, then F (x2)≥F (x1)

3 lim

x→–∞ F (x) = F (– ∞) =0

4 lim

x→+∞ F (x) = F (+ ∞) =1

5 The probability that a random variable X lies in the interval [x1, x2) is equal to the increment of its cumulative distribution function on this interval; i.e.,

P (x1≤X < x2) = F (x2) – F (x1)

6 F (x) is left continuous; i.e., lim

x→x0– 0F (x) = F (x0).

20.2.1-2 Discrete random variables

A random variable X is said to be discrete if the set of its possible values (the spectrum of

the discrete random variable) is at most countable A discrete distribution is determined by

a finite or countable set of probabilities P (X = x i) such that



i

P (X = x i) =1

To define a discrete random variable, it is necessary to specify the values x1, x2, and the corresponding probabilities p1, p2, , where p i = P (X = x i)

Remark. In what follows, we assume that the values of a discrete random variable X are arranged in

ascending order.

In this case, the cumulative distribution function of a discrete random variable X is the

step function defined as the sum

F (x) = 

x i<x

P (X = x i) (20.2.1.2)

It is often convenient to write out the cumulative distribution function using the function

θ (x) such that θ(x) =1for x >0and θ(x) =0for x≤ 0:

F (x) =

i

P (X = x i )θ(x – x i)

Trang 2

For discrete random variables, one can introduce the notion of probability density

function p(x) by setting

p (x) =

i

P (X = x i )δ(x – x i),

where δ(x) is the delta function.

20.2.1-3 Continuous random variables Probability density function

A random variable X is said to be continuous if its cumulative distribution function F X (x)

can be represented in the form

F (x) =

 x

∞ p (y) dy. (20.2.1.3)

The function p(x) is called the probability density function of the random variable X.

Obviously, relation (20.2.1.3) is equivalent to the relation

p (x) = lim Δx→0

P (xXx+Δx)

dF (x)

dx (20.2.1.4)

The differential dF (x) = p(x) dxP (xX < x + dx) is called a probability element.

Properties of the probability density function:

1 p(x)≥ 0

2 P (x1 ≤X < x2) =7x2

x1 p (y) dy.

3 7+

∞ p (x) dx =1

4 P (xξ < x + Δx)p (x) Δx.

5 For continuous random variables, one always has P (X = x) =0, but the event{X = x}

is not necessarily impossible

6 For continuous random variables,

P (x1 ≤X < x2) = P (x1< X < x2) = P (x1< Xx2) = P (x1≤Xx2).

20.2.1-4 Unified description of probability distribution

Discrete and continuous probability distributions can be studied simultaneously if the prob-ability of each event{aX < b}is represented in terms of the integral

P (aX < b) =

 b

a dF (x), (20.2.1.5)

where F (x) = P (X < x) is the cumulative distribution function of the random variable X.

For a continuous distribution, the integral (20.2.1.5) becomes the Riemann integral For a discrete distribution, the integral can be reduced to the form

P (aX < b) = 

a x i<b

P (X = x i)

In particular, the integral can also be used in the case of mixed distributions, i.e., distributions that are partially continuous and partially discrete

Trang 3

20.2.1-5 Symmetric random variables.

A random variable X is said to be symmetric if the condition

P (X < –x) = P (X > x) (20.2.1.6)

holds for all x.

Properties of symmetric random variables:

1 P (|X|< x) = F (x) – F (–x) =2F (x) –1

2 F (0) =0.5

3 If a moment α2k+1of odd order about the origin (see Paragraph 20.2.2-3) exists, then it

is zero

4 If t γ is the quantile of level γ (see Paragraph 20.2.2-5), then t γ = –t1–γ

A random variable Y is said to be symmetric about its expected value if the random variable X = Y – E{Y}is symmetric, where E{Y} is the expected value of a random

variable Y (see Paragraph 20.2.2-1).

Properties of random variables symmetric about the expected value:

1 P (|Y – E{Y}|< x) =2F Y (x + E{Y}) –1

2 F Y (E{Y}) =0.5

3 If a central moment μ2k+1of odd order (see Paragraph 20.2.2-3) exists, then it is equal

to zero

20.2.1-6 Functions of random variables

Suppose that a random variable Y is related to a random variable X by a functional dependence Y = f (X) If X is discrete, then, obviously, Y is also discrete To find the distribution law of the random variable Y , it suffices to calculate the values f (x i) If there

are repeated values among y i = f (x i), then these repeated values are taken into account only once, the corresponding probabilities being added

If X is a continuous random variable with probability density function p X (x), then, in general, the random variable Y is also continuous The cumulative distribution function

of Y is given by the formula

F Y (y) = P (η < y) = P [f (x) < y] =



f(x)<y p X (x) dx. (20.2.1.7)

If the function y = f (x) is differentiable and monotone on the entire range of the argument x, then the probability density function p Y (y) of the random variable Y is given

by the formula

p Y (y) = p X [ψ(y)]|ψ 

y (y)|, (20.2.1.8)

where ψ is the inverse function of f (x).

If f (x) is a nonmonotonic function, then the inverse function is nonunique and the probability density function of the random variable y is the sum of as many terms as there are values (for a given y) of the inverse function:

p Y (y) =

k



i=1

p X [ψ i (y)][ψ i (y)] 

y, (20.2.1.9)

where ψ1(y), , ψ k (y) are the values of the inverse function for a given y.

Trang 4

Example 1 Suppose that a random variable X has the probability density

p X (x) = √1

2π e

x2/2

Find the distribution of the random variable Y = X2.

In this case, y = f (x) = x2 According to (20.2.1.7), we obtain

F Y (y) =



x2<y

1

2π e

x2/2dx= 1

2π

 √y

√y e

x2/2dx= 2

2π

 √y

0 e

x2/2dx= 1

2π

 y

0

et/2

t dt.

Example 2 Suppose that a random variable X has the probability density

p X (x) = √1

*

(x – a)

2

2σ2

+

.

Find the probability density of the random variable Y = e X.

For y >0, the cumulative distribution function of the random variable Y = e Xis determined by the relations

F Y (y) = P (Y < y) = P (e X < y) = P (X < ln y) = F X (ln y).

We differentiate this relation and obtain

p (y) = dF Y (y)

dy = dF X (ln y)

dy = p X (ln y)1

y = 1

*

(ln y – a)

2

2σ2

+

for y >0

The distribution of Y is called the log-normal distribution.

Example 3 Suppose that a random variable X has the probability density p X (x) for x (–∞, ∞) Then the probability density of the random variable Y =|X|is given by the formula p Y (y) = p X (x) + p X (–x) (y≥ 0 ).

In particular, if X is symmetric, then p Y (y) =2p X (y) (y≥ 0 ).

20.2.2 Characteristics of One-Dimensional Random Variables

20.2.2-1 Expectation

The expectation (expected value) E{X}of a discrete or continuous random variable X is

the expression given by the formula

E{X}=

 +

∞ x dF (x) =



i

x i p i in the discrete case,

 +

∞ xp (x) dx in the continuous case.

(20.2.2.1)

For the existence of the expectation (20.2.2.1), it is necessary that the corresponding series or integral converge absolutely

The expectation is the main characteristic defining the “position” of a random variable, i.e., the number near which its possible values are concentrated

We note that the expectation is not a function of the variable x but a functional describing the properties of the distribution of the random variable X There are distributions for which

the expectation does not exist

Example 1 For the Cauchy distribution given by the probability density function

p(x) = 1

π(1+ x2), x (–∞, +∞), the expectation does not exist because the integral 7+

|x|/π(1+ x2) dx diverges.

Trang 5

20.2.2-2 Expectation of function of random variable.

If a random variable Y is related to a random variable X by a functional dependence

Y = f (X), then the expectation of the random variable Y = f (X) can be determined by two methods The first method is to construct the distribution of the random variable Y and then use already known formulas to find E{Y} The second method is to use the formulas

E{Y}= E{f (X)}=



i

f (x i )p i in the discrete case,

 +

∞ f (x)p(x) dx in the continuous case

(20.2.2.2)

if these expressions exist in the sense of absolute convergence

Example 2 Suppose that a random variable X is uniformly distributed in the interval (–π/2, π/2 ), i.e.,

p(x) =1/π for x(–π/2, π/2) Then the expectation of the random variable Y = sin(X) is equal to

E{Y} =

 +

f (x)p(x) dx =

 π/2

π/2

1

π sin x dx =0 Properties of the expectation:

1 E{C}= C for any real C.

2 E{αX + βY}= αE{X}+ βE{Y}for any real α and β.

3 E{X} ≤E{Y}if X(ω)Y (ω), ω

4 E



k=1X k

4

= 

k=1E{X k}if the series 

k=1E{|X k|}converges

5 g(E{X})≤E{g (X)}for convex functions g(X).

6 Any bounded random variable has a finite expectation

7 |E{X}| ≤E{|X|}

8 The Cauchy–Schwarz inequality (E{|XY|})2≤(E{X})2(E{Y})2holds

9 E5n

k=1X k6

=n

k=1E{X k}for mutually independent random variables X1, , X n 20.2.2-3 Moments

The expectation E{(X – a) k}is called the kth moment of a random variable X about a The

moments about zero are usually referred to simply as the moments of a random variable

(Sometimes they are called initial moments.) The kth moment satisfies the relation

α k = E{X k}=

 +

∞ x

k dF (x) =



i

x k

i p i in the discrete case,

 +

∞ x

k p (x) dx in the continuous case. (20.2.2.3)

If a = E{X}, then the kth moment of the random variable X about a is called the kth

central moment The kth central moment satisfies the relation

μ k = E{(X –E{X})k}=



i

(x i – E{X})k p i in the discrete case,

 +

∞ (x – E{X})k p (x) dx in the continuous case.

(20.2.2.4)

Trang 6

In particular, μ0=1for any random variable.

The number m k = E{|X – a|k}is called the kth absolute moment of X about a The existence of a kth moment α k or μ k implies the existence of the moments α mand

μ m of orders mk ; if the integral (or series) for α k or μ kdiverges, then all integrals (series)

for α m and μ m of orders mkalso diverge

There is a simple relationship between the central and initial moments:

μ k =

k



m=0

C m

k α m (α1)k–m, α0=1; α k =

k



m=0

C m

k μ m (α1)k–m. (20.2.2.5)

Relations (20.2.2.5) can be represented in the following easy-to-memorize symbolic form:

μ k = (α – α1)k , α k = (μ + α1)k, where it is assumed that after the right-hand sides have been

multiplied out according to the binomial formula, the expressions α m and μ mare replaced

by α m and μ m, respectively

If the probability distribution is symmetric about its expectation, then all existing central

moments μ k of even order k are zero.

The probability distribution is uniquely determined by the moments α0, α1, provided

that they all exist and the series 

m=0|α m|t m /m ! converges for some t >0

20.2.2-4 Variance

The variance of a random variable is the measure Var{X}of the deviation of a random

variable X from its expectation E{X}, determined by the relation

Var{X}= E{(X – E{X})2} (20.2.2.6) The variance Var{X} is the second central moment of the random variable X The

variance can be determined by the formulas

Var{X}=

 +

∞ (x – E{X})2dF (x)

=



i

(x i – E{X})2p i in the discrete case,

 +

∞ (x – E{X})2p (x) dx in the continuous case

(20.2.2.7)

The variance characterizes the spread in values of the random variable X about its

expectation

Properties of the variance:

1 Var{C}=0for any real C.

2 The variance is nonnegative: Var{X} ≥ 0

3 Var{αX + β}= α2Var{X}for any real numbers α and β.

4 Var{X}= E{X2}– (E{X})2

5 min

m E{(X – m)2}= Var{X}and is attained for m = E{X}

6 Var{X1 +· · · + X n} = Var{X1}+· · · + Var{X n} for pairwise independent random

variables X1, , X n

7 If X and Y are independent random variables, then

Var{XY}= Var{X}Var{Y}+ Var{X}(E{Y})2+ Var{Y}(E{X})2

Trang 7

20.2.2-5 Numerical characteristics of random variables.

A quantile of level γ of a one-dimensional distribution is a number t γ for which the value

of the corresponding distribution function is equal to γ; i.e.,

P (X < t γ ) = F (t γ ) = γ (0< γ <1) (20.2.2.8) Quantiles exist for each probability distribution, but they are not necessarily uniquely

determined Quantiles are widely used in statistics The quantile t1 2 is called the

me-dian Med{ X} For n =4, the quantiles t m/n are called quartiles, for n =10, they are called

deciles, and for n =100, they are called percentiles.

A mode Mode{X}of a continuous probability distribution is a point of maximum of

the probability density function p(x) A mode of a discrete probability distribution is a

value Mode{X} preceded and followed by values associated with probabilities smaller

than p(Mode{X})

Distributions with one, two, or more modes are said to be unimodal, bimodal, or

multimodal, respectively.

The standard deviation (root-mean-square deviation) of a random variable X is the

square root of its variance,

σ= Var{X} The standard deviation has the same dimension as the random variable itself

The coefficient of variation is the ratio of the standard deviation to the expected value,

E{X}.

The asymmetry coefficient, or skewness, is defined by the formula

γ1 = μ3

2)3 2. (20.2.2.9)

If γ1>0, then the distribution curve is more flattened to the right of the mode Mode{X};

if γ1 < 0, then the distribution curve is more flattened to the left of the mode Mode{X}

(see Fig 20.1) (As a rule, this applies to continuous random variables.)

Mode{ }X x

Mode{ }X

Figure 20.1 Relationship of the distribution curve and the asymmetry coefficient.

The excess coefficient, or excess, or kurtosis, is defined by the formula

γ2= μ4

μ2 2

One says that for γ2=0the distribution has a normal excess, for γ2>0the distribution has

a positive excess, and for γ2 <0the distribution has a negative excess

Remark. The coefficients γ2and γ2+ 3or (γ2+ 3)/2are often used instead of γ1and γ2.

Pearson’s first skewness coefficient for a unimodal distribution is defined by the formula

s= E{X}– Mode{X}

... Var{X} The standard deviation has the same dimension as the random variable itself

The coefficient of variation is the ratio of the standard deviation to the expected... this applies to continuous random variables.)

Mode{ }X x

Mode{ }X

Figure 20.1 Relationship of the distribution curve and the asymmetry coefficient.... γ2=0the distribution has a normal excess, for γ2>0the distribution has

a positive excess, and for γ2 <0the distribution has a negative

Ngày đăng: 02/07/2014, 13:20

🧩 Sản phẩm bạn có thể quan tâm