1. Trang chủ
  2. » Khoa Học Tự Nhiên

Báo cáo hóa học: " Research Article Nonexpansive Matrices with Applications to Solutions of Linear Systems by Fixed Point Iterations Teck-Cheong Lim" docx

13 292 0
Tài liệu đã được kiểm tra trùng lặp

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 13
Dung lượng 510,02 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

We characterizei matrices which are nonexpansive with respect to some matrix norms, and ii matrices whose average iterates approach zero or are bounded.. A matrix norm on M n is called a

Trang 1

Volume 2010, Article ID 821928, 13 pages

doi:10.1155/2010/821928

Research Article

Nonexpansive Matrices with Applications

to Solutions of Linear Systems by Fixed

Point Iterations

Teck-Cheong Lim

Department of Mathematical Sciences, George Mason University, 4400, University Drive,

Fairfax, VA 22030, USA

Correspondence should be addressed to Teck-Cheong Lim,tlim@gmu.edu

Received 28 August 2009; Accepted 19 October 2009

Academic Editor: Mohamed A Khamsi

Copyrightq 2010 Teck-Cheong Lim This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited

We characterizei matrices which are nonexpansive with respect to some matrix norms, and ii matrices whose average iterates approach zero or are bounded Then we apply these results to iterative solutions of a system of linear equations

Throughout this paper,R will denote the set of real numbers, C the set of complex numbers, andM nthe complex vector space of complexn × n matrices A function  ·  : M n → R is a

matrix norm if for all A, B ∈ M n, it satisfies the following five axioms:

1 A ≥ 0;

2 A  0 if and only if A  0;

3 cA  |c|A for all complex scalars c;

4 A  B ≤ A  B;

5 AB ≤ A B.

Let| · | be a norm on Cn Define ·  on M nby

A  max

This norm onM n is a matrix norm, called the matrix norm induced by| · | A matrix norm on

M n is called an induced matrix norm if it is induced by some norm onCn If · 1is a matrix norm onM n, there exists an induced matrix norm · 2onM nsuch thatA2 ≤ A1for all

Trang 2

A ∈ M ncf 1, page 297 Indeed one can take  · 2to be the matrix norm induced by the norm| · | on Cndefined by

whereCx is the matrix in M nwhose columns are all equal tox For A ∈ M n,ρA denotes

the spectral radius ofA.

Let | · | be a norm in Cn A matrix A ∈ M n is a contraction relative to | · | if it is a contraction as a transformation fromCnintoCn; that is, there exists 0≤ λ < 1 such that

Evidently this means that for the matrix norm ·  induced by | · |, A < 1 The following

theorem is well knowncf 1, Sections 5.6.9–5.6.12

Theorem 1 For a matrix A ∈ M n , the following are equivalent:

a A is a contraction relative to a norm in C n ;

b A < 1 for some induced matrix norm  · ;

c A < 1 for some matrix norm  · ;

d limk → ∞ A k  0;

e ρA < 1.

Thatb follows from c is a consequence of the previous remark about an induced matrix norm being less than a matrix norm Since all norms onM n are equivalent, the limit ind can be relative to any norm onM n, so thatd is equivalent to all the entries of A kconverge

to zero ask → ∞, which in turn is equivalent to lim k → ∞ A k x  0 for all x ∈ C n

In this paper, we first characterize matrices inM n that are nonexpansive relative to some

norm| · | on Cn, that is,

Then we characterize thoseA ∈ M nsuch that

A k 1

k



I  A  A2 · · ·  A k−1

5

converges to zero ask → ∞, and those that {A k:k  0, 1, 2, } is bounded.

Finally we apply our theory to approximation of solution ofAx  b using iterative

methodsfixed point iteration methods

Trang 3

Theorem 2 For a matrix A ∈ M n , the following are equivalent:

a A is nonexpansive relative to some norm on C n ;

b A ≤ 1 for some induced matrix norm  · ;

c A ≤ 1 for some matrix norm  · ;

d {A k:k  0, 1, 2, } is bounded;

e ρA ≤ 1, and for any eigenvalue λ of A with |λ|  1, the geometric multiplicity is equal to

the algebraic multiplicity.

Proof As in the previous theorem,a, b, and c are equivalent Assume that b holds Let the norm ·  be induced by a vector norm | · | of Cn Then



A k x ≤A k |x| ≤ A k |x| ≤ |x|, k  0, 1, 2, , 6

proving thatA k x is bounded in norm | · | for every x ∈ C n Takingx  e i, we see that the set

of all columns ofA k , k  0, 1, 2, , is bounded This proves that A k , k  0, 1, 2, , is bounded

in maximum column sum matrix norm1, page 294, and hence in any norm in Mn Note that the last part of the proof also follows from the Uniform Boundedness Principlesee, e.g.,

2, Corollary 21, page 66

Now we prove thatd implies e Suppose that A has an eigenvalue λ with λ > 1.

Letx be an eigenvector corresponding to λ Then



ask → ∞, where  ·  is any vector norm of C n This contradictsd Hence |λ| ≤ 1 Now

suppose thatλ is an eigenvalue with |λ|  1 and the Jordan block corresponding to λ is not

diagonal Then there exist nonzero vectorsv1, v2such thatAv1  λv1, Av2  v1 λv2 Let

u  v1  v2 Then

A k u  λ k−1 λ  kv1 λ k v2, 8

and A k u ≥ kv1 − v1 − v2 It follows that A k u, k  0, 1, 2, , is unbounded,

contradictingd Hence d implies e

Lastly we prove thate implies c Assume that e holds A is similar to its Jordan

canonical form J whose nonzero off-diagonal entries can be made arbitrarily small by

similarity 1, page 128 Since the Jordan block for each eigenvalue with modulus 1 is diagonal, we see that there is an invertible matrix S such that the l1-sum of each row of

SAS−1is less than or equal to 1, that is,SAS−1∞≤ 1, where  · ∞is the maximum row sum matrix norm1, page 295 Define a matrix norm  ·  by M  SMS−1∞ Then we have

A ≤ 1.

Letλ be an eigenvalue of a matrix A ∈ M n The index ofλ, denoted by indexλ is the

smallest value ofk for which rankA − λI k  rankA − λI k11, pages 148 and 131 Thus conditione above can be restated as ρA ≤ 1, and for any eigenvalue λ of A with |λ|  1,

indexλ  1

Trang 4

LetA ∈ M n Consider

A k 1kI  A  · · ·  A k−1

We callA k thek-average of A As with A k, we have A k x → 0 for every x if and only if

A k → 0 in M n, and thatA k x is bounded for every x if and only if A kis bounded inM n We have the following theorem

Theorem 3 Let A ∈ M n Then

a A k , k  1, 2, , converges to 0 if and only if A ≤ 1 for some matrix norm  ·  and that

1 is not an eigenvalue of A,

b A k , k  1, 2, , is bounded if and only if ρA ≤ 1, indexλ ≤ 2 for every eigenvalue λ with |λ|  1 and that index1  1 if 1 is an eigenvalue of A.

Proof First we prove the su fficiency part of a Let x be a vector in C n Let

y k 1

k



I  A  · · ·  A k−1

ByTheorem 2for any eigenvaluesλ of A either |λ| < 1 or |λ|  1 and indexλ  1.

If A is written in its Jordan canonical form A  SJS−1, then the k-average of A is

SJS−1, whereJis thek-average of J Jis in turn composed of thek-average of each of its

Jordan blocks For a Jordan block ofJ of the form

λ 1

λ 1

· ·

· 1

λ

|λ| must be less than 1 Its k-average has constant diagonal and upper diagonals Let D jbe the constat value of itsjth upper diagonal D0being the diagonal and let Sj  kD j Then

Cm, n  0 for n > m

S0 1− λ k

1− λ ,

S j  C j, j  C j  1, j λ  · · ·  C k − 1, j λ k−1−j , j  1, 2, , n − 1.

12

Using the relationCm  1, j − Cm, j  Cm, j − 1, we obtain

S j − λS j  S j−1 − λ k−j C k, j 13

Thus, we haveS0 → 1/1 − λ as k → ∞ By induction, using 13 above and the fact that

λ k−j Ck, j → 0 as k → ∞, we obtain S j → 1/1 − λ j1ask → ∞ Therefore D j  S j /k  O1/k as k → ∞.

Trang 5

If the Jordan block is diagonal of constant valueλ, then λ / 1, |λ| ≤ 1 and the k-average

of the block is diagonal of constant value1 − λ k /k1 − λ  O1/k.

We conclude thatA k   O1/k and hence y k  ≤ A k x  O1/k as k → ∞.

Now we prove the necessity part of a If 1 is an eigenvalue of A and x is a

corresponding eigenvector, thenA k x  x / 0 for every k and of course B k x fails to converge

to 0 Ifλ is an eigenvalue of A with |λ| > 1 and x is a corresponding eigenvector, then

A k x 



 λ

k− 1

kλ − 1





 x ≥ |λ|

k− 1

k|λ − 1| x. 14

which approaches to ∞ as k → ∞ If λ is an eigenvalue of A with |λ|  1, λ / 1, and

indexλ ≥ 2, then there exist nonzero vectors v1, v2such thatAv1   λv1, Av2  v1 λv2 Then by using the identity

1 2λ  3λ2 · · ·  k − 1λ k−2 1− λ k−1

1 − λ2 − k − 1 λ k−1

we get

A k v2 



1− λ k−1

k1 − λ2 −



1−k1

λ k−1

1− λ



v1 1− λ k

k1 − λ v2. 16

It follows that limk → ∞ A k v2 does not exist This completes the proof of part a

Suppose that A satisfies the conditions in b and that A  SJS−1 is the Jordan canonical form of A Let λ be an eigenvalue of A and let v be a column vector of S

corresponding to λ If |λ| < 1, then the restriction B of A to the subspace spanned by

v, Av, A2v, is a contraction, and we have A k v  B k v ≤ v If |λ|  1, and λ / 1,

then by conditions in b either Av  λv, or there exist v1, v2 with v  v2 such that

Av1   λv1, Av2  v1 λv2 In the former case, we haveA k  ≤ v and in the latter case,

we see from16 that A k v  A k v2 is bounded Finally if λ  1 then since index1  1, we

haveAv  v and hence A k v  v In all cases, we proved that A k v, k  0, 1, 2, , is bounded.

Since column vectors ofS form a basis for C n, the sufficiency part of b follows

Now we prove the necessity part ofb If A has an eigenvalue λ with |λ| > 1 and

eigenvectorv, then as shown above A k v → ∞ as k → ∞ If A has 1 as an eigenvalue and

index1 ≥ 2, then there exist nonzero vectors v1, v2 such thatAv1  v1 andAv2  v1 v2 ThenA k v2  k−1/2v2which is unbounded Ifλ is an eigenvalue of A with |λ|  1, λ / 1

and indexλ ≥ 3, then there exist nonzero vectors v1, v2andv3such thatAv1  λv1, Av2 

v1  λv2andAv3   v2 λv3 By expandingA j v3, j  0, 1, 2, , k − 1 and using the identity

k−1



j2

C j, 2 λ j−2 1

1 − λ2



1− λ k−2

1− λ 

1

2k − 2λ k−2 k − 1λ − k  1



, 17

Trang 6

we obtain

A k v3  1

1 − λ2



1− λ k−2

k1 − λ

1

2k − 2λ k−2k − 1

k λ −

k  1 k



v1





1− λ k−1

k1 − λ2 −



1− 1

k

λ k−1

1− λ



v2 1− λ k

k1 − λ v3

18

which approaches to∞ as k → ∞ This completes the proof.

We now consider applications of preceding theorems to approximation of solution of

a linear systemAx  b, where A ∈ M nandb a given vector in C n LetQ be a given invertible

matrix inM n.x is a solution of Ax  b if and only if x is a fixed point of the mapping T

defined by

Tx I − Q−1Ax  Q−1b. 19

T is a contraction if and only if I − Q−1A is In this case, by the well known Contraction

Mapping Theorem, given any initial vector x0, the sequence of iterates x k  T k x0, k 

0, 1, 2, , converges to the unique solution of Ax  b In practice, given x0, each successive

x kis obtained fromx k−1by solving the equation

The classical methods of Richardson, Jacobi, and Gauss-Seidelsee, e.g., 3 have Q  I, D,

and L respectively, where I is the identity matrix, D the diagonal matrix containing the

diagonal of A, and L the lower triangular matrix containing the lower triangular portion

ofA Thus byTheorem 1we have the following known theorem

Theorem 4 Let A, Q ∈ M n , with Q invertible Let b, x0 ∈ Cn If ρI − Q−1A < 1, then A is invertible and the sequence x k , k  1, 2, , defined recursively by

converges to the unique solution of Ax  b.

Theorem 4fails ifρI − Q−1A  1, For a simple 2 × 2 example, let Q  I, b  0, A  2I

andx0any nonzero vector

We need the following lemma in the proof of the next two theorems For a matrix

A ∈ M n, we will denoteRA and NA the range and the null space of A respectively.

Lemma 5 Let A be a singular matrix in M n such that the geometric multiplicity and the algebraic multiplicity of the eigenvalue 0 are equal, that is, index 0  1 Then there is a unique projection

P A whose range is the range of A and whose null space is the null space of A, or equivalently, C n 

RA ⊕ NA Moreover, A restricted to RA is an invertible transformation from RA onto RA.

Trang 7

Proof If A  SJS−1is a Jordan canonical form ofA where the eigenvalues 0 appear at the end

portion of the diagonal ofJ, then the matrix

P A  S

1

·

· 1 0

·

· 0

is the required projection ObviouslyA maps RA into RA If z ∈ RA and Az  0, then

z ∈ NA ∩ RA  {0} and so z  0 This proves that A is invertible on RA.

Remark 6 Under the assumptions ofLemma 5, we will call the component of a vectorc in NA the projection of c on NA along RA Note that by definition of index, the condition

in the lemma is equivalent toNA2  NA.

Theorem 7 Let A be a matrix in M n and b a vector in C n Let Q be an invertible matrix in M n and let B  I − Q−1A Assume that ρB ≤ 1 and that indexλ  1 for every eigenvalue λ of B with modulus 1, that is, B is nonexpansive relative to a matrix norm Starting with an initial vector x0 in

Cn define x k recursively by

for k  1, 2, Let

y k x0  x1 · · ·  x k−1

If Ax  b is consistent, that is, has a solution, then y k , k  1, 2, , converge to a solution vector z with rate of convergence y k − z  O1/k If Ax  b is inconsistent, then lim k x k  limk y k 

∞ More precisely, lim k x k /k  cand lim k y k /k  c/2, where c  Q−1b and cis the projection of c

on NA  NQ−1A along RQ−1A.

Proof First we assume that A is invertible so that I − B  Q−1A is also invertible Let T be the

mapping defined byTx  Bxc Then T k x  B k xcBc· · ·B k−1 c Let s  cBc· · ·B k−1 c.

Thens − Bs  c − B k c and hence s  I − B−1c − I − B−1B k c  I − B−1c − B k I − B−1c Let

z  I − B−1c  A−1b z is the unique solution of Ax  b and

T k x  B k x  z − B k z  B k x − z  z. 25

Trang 8

Since the sequencex kin the theorem isT k x0, we have

y k 1kI  B  · · ·  B k−1

x0− z  z  B k x0− z  z. 26

SinceI − B is invertible, 1 is not an eigenvalue of B, and byTheorem 3 parta y k − z 

B k x0−z → 0 as k → ∞ Moreover, from the proof of the same theorem, y k −z  O1/k.

Next we consider the case when A is not invertible Since Q is invertible, we have RQ−1A  Q−1RA and NQ−1A  NA The index of the eigenvalue 0 of Q−1A is the

index of eigenvalue 1 ofB  I − Q−1A Thus byLemma 5,Cn  Q−1RA ⊕ NA For every

vectorv ∈ C n, let v r andv ndenote the component ofv in the subspace Q−1RA and

NA, respectively.

Assume thatAx  b is consistent, that is, b ∈ RA Then c ∈ RQ−1A ByLemma 5, the restriction ofQ−1A on its range is invertible, so there exists a unique zinRQ−1A such

thatQ−1Az c, or equivalently, I − Bz c For any vector x, we have

T k x  B k x  c  Bc  · · ·  B k−1 c

 B K

x r  x n

I  B  · · ·  B k−1

I − Bz

 B k

x r

 x n  z− B k

z

 B k

x r − z

 x n  z.

27

SinceB maps RQ−1A into RQ−1A and I −B  Q−1A restricted to RQ−1A is invertible, we

can apply the preceding proof and conclude that the sequencey kas defined before converges

toz  x0n  zandy k − z  O1/k Now Az  Ax n0   Az  Az  Qc  b, showing

thatz is a solution of Ax  b.

Assume now thatb /∈ RA, that is, Ax  b is inconsistent Then c /∈ RQ−1A and c 

c r  c nwithc n / 0 As in the preceding case there exists a unique z∈ RQ−1A such that

I − Bz  c r Note that for ally ∈ NA, By  I − Q−1Ay  y Thus for any vector x

and any positive integerj

x j  T j x

 B j x  c  Bc  · · ·  B j−1 c

 B j

x r  x n

I  B  · · ·  B j−1

I − Bz jc n

 B j

x r

 x n  z− B j

z  jc n

 B j

xr − z

 x n  z jc n ,

y k 1kx  Tx  · · ·  T k−1 x

 B kx r − z

 x n  zk − 1

2 c n ,

28

Trang 9

whereB k  I  B  · · ·  B k−1  As in the preceding case, B k x r − z, k  0, 1, 2, is bounded

andB k x r −z, k  1, 2, , converges to 0 Thus lim k → ∞ x k /k  cn and lim k → ∞ y k /k 

c n /2, and hence lim k → ∞ x k  limk → ∞ y k  ∞ This completes the proof

Next we consider another kind of iteration in which the nonlinear case was considered

in Ishikawa 4 Note that the type of mappings in this case is slightly weaker than nonexpansivitysee condition c in the next lemma

Lemma 8 Let B be an n × n matrix The following are equivalent:

a for every 0 < μ < 1, there exists a matrix norm  ·  μ such that μI  1 − μB μ ≤ 1,

b for every 0 < μ < 1, there exists an induced matrix norm · μ such that μI1−μB μ ≤ 1,

c ρB ≤ 1 and index1  1 if 1 is an eigenvalue of B.

Proof As in the proof ofTheorem 2,a and b are equivalent For 0 < μ < 1, denote μI 

1 − μB by Bμ Suppose now that a holds Let λ be an eigenvalue of B Then μ  1 −

μλ is an eigenvalue of Bμ By Theorem 2 |μ  1 − μλ| ≤ 1 for every 0 < μ < 1 and

hence|λ| ≤ 1 If 1 is an eigenvalue of B, then it is also an eigenvalue of Bμ ByTheorem 2, the index of 1, as an eigenvalue of Bμ, is 1 Since obviously B and Bμ have the same

eigenvectors corresponding to the eigenvalue 1, the index of 1, as an eigenvalue ofB, is also

1 This provesc

Now assumec holds Since |μ  1 − μλ| < 1 for |λ|  1, λ / 1, every eigenvalue of

Bμ, except possibly for 1, has modulus less than 1 Reasoning as above, if 1 is an eigenvalue

ofBμ, then its index is 1 Therefore byTheorem 2,a holds This completes the proof

Theorem 9 Let A ∈ M n and b ∈ C n Let Q be an invertible matrix in M n , and B  I − Q−1A Suppose ρB ≤ 1 and that index1  1 if 1 is an eigenvalue of B Let 0 < μ < 1 be fixed Starting with an initial vector x0 , define x k , y k , k  0, 1, 2, , recursively by

y0  x0,

Qx k   Q − A y k−1  b,

y k  μy k−1 1− μ x k

29

If Ax  b is consistent, then y k , k  0, 1, 2, , converges to a solution vector z of Ax  b with rate

of convergence given by

y k − z  oζ k

where ζ is any number satisfying

Trang 10

If Ax  b is inconsistent, then lim k → ∞ y k   ∞; more precisely,

lim

k → ∞

y k

k 

where c n is the projection of c on NA along RQ−1A.

Proof Let c  Q−1b, B1  μI  1 − μB  I − 1 − μQ−1A, and Tx  B1x  1 − μc Then

y k  T k x0

First we assume thatA is invertible Then I − B1  1 − μQ−1A is invertible and 1 is

not an eigenvalue ofB1; thusρB1  < 1 Let z  1 − μI − B1−1c  A−1b We have

y k  T k x0

 B k

1x0 1− μ c  B1c  · · ·  B k−1

1 c

 B k

1x0 1− μ

1− μ



I  B1  · · ·  B k−1

1



I − B1z

 B k

1x0− z  z.

33

By a well known theoremsee, e.g 1, y k − z  oζ k  for every ζ > ρB1

Assume now thatA is not invertible and b ∈ RA Then c is in the range of Q−1A.

Since B  I − Q−1A satisfies the condition in Lemma 8, Q−1A satisfies the condition in

Lemma 5 Thus the restriction ofQ−1A on its range is invertible and there exists zinRQ−1A

such that Q−1Az  c, or equivalently, I − B1z  1 − μc For any vector x  x0, we have

y k  T k x

 B k

1x  1− μ c  B1c  · · ·  B k−1

1 c

 B k

1



x r  x n

I  B1  · · ·  B k−1

1



I − B1z

 B k

1



x r

 x n  z− B k

1

z

 B k

1



x r − z

 x n  z.

34

SinceB1mapsRQ−1A into RQ−1A and I − B  Q−1A restricted to RQ−1A is invertible,

we can apply the preceding proof and conclude that the sequencey k , k  0, 1, 2, converges

toz  x n  zandy k − z  oζ k  z solves Ax  b since Az  Ax n   Az  Az 

Qc  b.

... restricted to RA is an invertible transformation from RA onto RA.

Trang 7

Proof If A ...

which approaches to< i>∞ as k → ∞ This completes the proof.

We now consider applications of preceding theorems to approximation of solution of

a linear systemAx  b,... class="text_page_counter">Trang 5

If the Jordan block is diagonal of constant valueλ, then λ / 1, |λ| ≤ and the k-average

of the

Ngày đăng: 21/06/2014, 20:20

TỪ KHÓA LIÊN QUAN

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN

🧩 Sản phẩm bạn có thể quan tâm