1. Trang chủ
  2. » Khoa Học Tự Nhiên

Báo cáo hóa học: " Review Article Super-Relaxed (η)-Proximal Point Algorithms, Relaxed (η)-Proximal Point Algorithms, Linear Convergence Analysis, and Nonlinear Variational Inclusions" doc

47 193 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Tiêu đề Review Article Super-Relaxed (η)-Proximal Point Algorithms, Relaxed (η)-Proximal Point Algorithms, Linear Convergence Analysis, and Nonlinear Variational Inclusions
Tác giả Ravi P. Agarwal, Ram U. Verma
Trường học Florida Institute of Technology
Chuyên ngành Mathematics
Thể loại Review Article
Năm xuất bản 2009
Thành phố Melbourne
Định dạng
Số trang 47
Dung lượng 678,66 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

We focus mostly on applications of the super-relaxed η- proximal point algorithm to the context of solving a class of nonlinear variational inclusionproblems, based on the notion of maxi

Trang 1

Volume 2009, Article ID 957407, 47 pages

doi:10.1155/2009/957407

Review Article

Convergence Analysis, and Nonlinear

Variational Inclusions

Ravi P Agarwal1, 2 and Ram U Verma1, 3

1 Department of Mathematical Sciences, Florida Institute of Technology, Melbourne, FL 32901, USA

2 Department of Mathematics and Statistics, King Fahd University of Petroleum and Minerals,

Dhahran 31261, Saudi Arabia

3 International Publications (USA), 12085 Lake Cypress Circle, Suite I109, Orlando, FL 32828, USA

Correspondence should be addressed to Ravi P Agarwal,agarwal@fit.edu

Received 26 June 2009; Accepted 30 August 2009

Recommended by Lai Jiu Lin

We glance at recent advances to the general theory of maximalset-valued monotone mappingsand their role demonstrated to examine the convex programming and closely related field ofnonlinear variational inequalities We focus mostly on applications of the super-relaxed η-

proximal point algorithm to the context of solving a class of nonlinear variational inclusionproblems, based on the notion of maximalη-monotonicity Investigations highlighted in this

others have played a significant part as well in generalizing the proximal point algorithm

or super-relaxed η-proximal point algorithm, the fundamental model for Rockafellar’s case

does the job Furthermore, we attempt to explore possibilities of generalizing the Yosidaregularization/approximation in light of maximalη-monotonicity, and then applying to first-

order evolution equations/inclusions

Copyrightq 2009 R P Agarwal and R U Verma This is an open access article distributed underthe Creative Commons Attribution License, which permits unrestricted use, distribution, andreproduction in any medium, provided the original work is properly cited

1 Introduction and Preliminaries

We begin with a real Hilbert space X with the norm  ·  and the inner product ·, · We

consider the general variational inclusion problem of the following form Find a solution to

where M : X → 2 X is a set-valued mapping on X.

Trang 2

In the first part, Rockafellar 1 introduced the proximal point algorithm, andexamined the general convergence and rate of convergence analysis, while solving1.1 by

showing when M is maximal monotone, that the sequence {x k} generated for an initial point

converges weakly to a solution of1.1, provided that the approximation is made sufficiently

accurate as the iteration proceeds, where P k  I  c k M−1for a sequence{c k} of positive realnumbers that is bounded away from zero, and in second part using the first part and furtheramending the proximal point algorithm succeeded in achieving the linear convergence Itfollows from1.2 that x k1is an approximate solution to inclusion problem

of the Lipschitz continuity of M−1at 0 plays the crucial part Let us recall these results

Theorem 1.1 see 1 Let X be a real Hilbert space Let M : X → 2 X be maximal monotone, and let xbe a zero of M Let the sequence {x k } be generated by the iterative procedure

k0  k < ∞, and {c k } is bounded away from zero Suppose that the

sequence {x k } is bounded in the sense that there exists at least one solution to 0 ∈ Mx.

Then the sequence {x k } converges weakly to xfor 0 ∈ Mx with

is crucial; otherwise we may end up getting a nonconvergent sequence even with having just

 k → 0 and X one dimensional Consider any maximal monotone mapping M such that the

Trang 3

set T−10  {x : 0 ∈ Mx}, that is known always to be convex and contains more than one element Then it turns out that T−1x contains a nonconvergent sequence {x k} such that

Next we look, unlikeTheorem 1.1, at 1, Theorem 2 in which Rockafellar achieved

a linear convergence of the sequence by considering the Lipschitz continuity of M−1 at 0instead

Theorem 1.3 see 1 Let X be a real Hilbert space Let M : X → 2 X be maximal monotone, and let xbe a zero of M Let the sequence {x k } be generated by the iterative procedure

k0 δ k < ∞, and {c k } is bounded away from zero Suppose that the

sequence {x k } is bounded in the sense that there exists at least one solution to 0 ∈ Mx In addition,

let M−1be Lipschitz continuous at 0 with modulus a, and

Trang 4

Later on Rockafellar1 appliedTheorem 1.1 to a minimization problem regarding

function f : X → −∞, ∞, where f is lower semicontinuous convex and proper by taking

M  ∂f It is well known that in this situation ∂f is maximal monotone, and further

That means, the proximal point algorithm for M  ∂f is a minimizing method for f.

There is an abundance of literature on proximal point algorithms with applicationsmostly followed by the work of Rockafellar1, but we focus greatly on the work of Ecksteinand Bertsekas 2, where they have relaxed the proximal point algorithm in the followingform and applied to the Douglas-Rachford splitting method Now let us have a look at therelaxed proximal point algorithm introduced and studied in2

Algorithm 1.4 Let M : X → 2 X be a set-valued maximal monotone mapping on X with

0∈ rangeM, and let the sequence {x k} be generated by the iterative procedure

1.19

are scalar sequences

As a matter of fact, Eckstein and Bertsekas2 appliedAlgorithm 1.4to approximate

a weak solution to 1.1 In other words, they established Theorem 1.1 using the relaxedproximal point algorithm instead

Theorem 1.5 see 2, Theorem 3 Let M : X → 2X be a set-valued maximal monotone mapping

on X with 0 ∈ rangeM, and let the sequence {x k } be generated by Algorithm 1.4 If the scalar sequences { k }, {α k }, and {c k } satisfy

E1 Σ∞

k0  k < ∞, Δ1 inf α k > 0, Δ2 sup α k < 2, c  inf c k > 0, 1.20

then the sequence {x k } converges weakly to a zero of M.

Trang 5

Convergence analysis for Algorithm 1.4 is achieved using the notion of the firmnonexpansiveness of the resolvent operator I  c k M−1 Somehow, they have notconsidered applying Algorithm 1.4 to Theorem 1.3 to the case of the linear convergence.The nonexpansiveness of the resolvent operatorI  c k M−1 poses the prime difficulty toalgorithmic convergence, and may be, this could have been the real steering for Rockafellar

to the Lipschitz continuity of M−1instead That is why the Yosida approximation turned out

to be more effective in this scenario, because the Yosida approximation

M c k  c−1

k

takes care of the Lipschitz continuity issue

As we look back into the literature, general maximal monotonicity has played a greaterrole to studying convex programming as well as variational inequalities/inclusions Later itturned out that one of the most fundamental algorithms applied to solve these problems wasthe proximal point algorithm In2, Eckstein and Bertsekas have shown that much of thetheory of the relaxed proximal point algorithm and related algorithms can be passed along

to the Douglas-Rachford splitting method and its specializations, for instance, the alternatingdirection method of multipliers

Just recently, Verma3 generalized the relaxed proximal point algorithm and applied

to the approximation solvability of variational inclusion problems of the form1.1 Recently,

a great deal of research on the solvability of inclusion problems is carried out using resolventoperator techniques, that have applications to other problems such as equilibria problems

in economics, optimization and control theory, operations research, and mathematicalprogramming

In this survey, we first discuss in detail the history of proximal point algorithmswith their applications to general nonlinear variational inclusion problems, and then werecall some significant developments, especially the relaxation of proximal point algorithmswith applications to the Douglas-Rachford splitting method At the second stage, we turnour attention to over-relaxed proximal point algorithms and their contribution to thelinear convergence We start with some introductory materials to the over-relaxed η-

proximal point algorithm based on the notion of maximal η-monotonicity, and recall

some investigations on approximation solvability of a general class of nonlinear inclusionproblems involving maximalη-monotone mappings in a Hilbert space setting As a matter

fact, we examine the convergence analysis of the over-relaxedη-proximal point algorithm

for solving a class of nonlinear inclusions Also, several results on the generalized firmnonexpansiveness and generalized resolvent mapping are given Furthermore, we explorethe real impact of recently obtained results on the celebrated work of Rockafellar, mostimportantly in the case of over-relaxedor super-relaxed proximal point algorithms Formore details, we refer the reader1 55

We note that the solution set for1.1 turns out to be the same as of the Yosida inclusion

where M ρ  MI  ρM−1 is the Yosida regularization of M, while there is an equivalent form ρ−1I − I  ρM−1, that is characterized as the Yosida approximation of M with

Trang 6

parameter ρ > 0 It seems in certain ways that it is easier to solve the Yosida inclusion than

1.1 In other words, M ρ provides better solvability conditions under right choice for ρ than

M itself To prove this assertion, let us recall the following existence theorem.

Theorem 1.6 Let M : X → 2 X be a set-valued maximal monotone mapping on X Then the following statements are equivalent.

i An element u ∈ X is a solution to 0 ∈ M ρ u.

equa-Yosida approximation can be generalized in the context of solving first-order evolutionequations/inclusions In Zeidler52, Lemma 31.7, it is shown that the Yosida approximation

M ρis2ρ−1-Lipschitz continuous, that is,

ρM−1, though the result does not seem to be much application oriented, while if we apply

the firm nonexpansiveness of the resolvent operator R M

ρ  I  ρM−1, we can achieve, asapplied in5, more application-oriented results as follows:

Trang 7

Proof For any x, y ∈ DM, we have

This completes the proof

We note that from applications’ point of view, it seems that the result

Indeed, the Yosida approximation M ρ  ρ−1I − I  ρM−1 and its equivalent form MI 

ρM−1are related to this identity Let us consider

Trang 8

Suppose that u ∈ I − I  ρM−1w, then we have

Trang 9

Note that when M : X → 2 Xis maximal monotone, mappings

are single valued, in fact maximal monotone and nonexpansive

The contents for the paper are organized as follows.Section 1 deals with a generalhistorical development of the relaxed proximal point algorithm and its variants inconjunction with maximal η-monotonicity, and with the approximation solvability of a

class of nonlinear inclusion problems using the convergence analysis for the proximalpoint algorithm as well as for the relaxed proximal point algorithm Section 2 introducesand derives some results on unifying maximal η-monotonicity and generalized firm

nonexpansiveness of the generalized resolvent operator In Section 3, the role of the relaxedη-proximal point algorithm is examined in detail in terms of its applications to

over-approximating the solution of the inclusion problem 1.1 Finally, Section 4 deals withsome important specializations that connect the results on general maximal monotonicity,especially to several aspects of the linear convergence

In this section we discus some results based on basic properties of maximal η-monotonicity, and then we derive some results involving η-monotonicity and the generalized firm nonexpansiveness Let X denote a real Hilbert space with the norm  ·  and inner product

·, · Let M : X → 2 X be a multivalued mapping on X We will denote both the map M and its graph by M, that is, the set {x, y : y ∈ Mx} This is equivalent

to stating that a mapping is any subset M of X × X, and Mx  {y : x, y ∈

M} If M is single valued, we will still use Mx to represent the unique y such that

x, y ∈ M rather than the singleton set {y} This interpretation will much depend on the context The domain of a map M is defined as its projection onto the first argument

by

domM x ∈ X : ∃y ∈ X :

x, y

∈ M {x ∈ X : Mx / ∅}. 2.1domT  X will denote the full domain of M, and the range of M is defined by

rangeM y ∈ X : ∃x ∈ X :

x, y

The inverse M−1of M is {y, x : x, y ∈ M} For a real number ρ and a mapping M, let

ρM  {x, ρy : x, y ∈ M} If L and M are any mappings, we define

Trang 10

Definition 2.1 Let M : X → 2 X be a multivalued mapping on X The map M is said to be

i monotone if

u− v, u − v  ≥ 0 ∀u, u, v, v ∈ graph M, 2.4

ii r-strongly monotone if there exists a positive constant r such that

u− v, u − v  ≥ ru − v2 ∀u, u, v, v ∈ graph M, 2.5

iii strongly monotone if

vi m-relaxed monotone if there exists a positive constant m such that

u− v, u − v  ≥ −mu − v2 ∀u, u, v, v ∈ graph M, 2.11

vii cocoercive if

u− v, u − v ≥ u− v∗2 ∀u, u, v, v ∈ graph M, 2.12

viii c-cocoercive if there is a positive constant c such that

u− v, u − v ≥ c u− v∗2 ∀u, u, v, v ∈ graph M. 2.13

Trang 11

Definition 2.2 Let M : X → 2 X be a mapping on X The map M is said to be

i nonexpansive if

u− v ≤ u − v ∀u, u, v, v ∈ graph M, 2.14

ii firmly nonexpansive if

u− v∗2≤ u− v, u − v  ∀u, u, v, v ∈ graph M, 2.15

iii c-firmly nonexpansive if there exists a constant c > 0 such that

u− v∗2≤ cu− v, u − v ∀ u, u, v, v ∈ graph M. 2.16

In light of Definitions2.1vii and2.2ii, notions of cocoerciveness and firm pansiveness coincide, but differ in applications much depending on the context

nonex-Definition 2.3 A map η : X × X → X is said to be

Trang 12

Definition 2.4 Let M : X → 2 X be a multivalued mapping on X, and let η : X × X → X be another mapping The map M is said to be

i η-monotone if

u− v, η u, v ≥ 0 ∀u, u, v, v ∈ graph M, 2.21

ii r, η-strongly monotone if there exists a positive constant r such that

u− v, η u, v ≥ ru − v2 ∀u, u, v, v ∈ graph M, 2.22

iii η-strongly monotone if

u− v, η u, v ≥ u − v2 ∀u, u, v, v ∈ graph M, 2.23

iv r, η-strongly pseudomonotone if

u, η u, v ≥ 0 ∀u, u, v, v ∈ graph M, 2.27

vi m, η-relaxed monotone if there exists a positive constant m such that

u− v, η u, v ≥ −mu − v2 ∀u, u, v, v ∈ graph M, 2.28

vii c, η-cocoercive if there is a positive constant c such that

u− v, η u, v ≥ cu− v∗2 ∀u, u, v, v ∈ graph M. 2.29

Trang 13

Definition 2.5 A map M : X → 2 Xis said to be maximalη-monotone if

1 M is η-monotone,

2 RI  cM  X for c > 0.

Proposition 2.6 Let η : X × X → X be a t-strongly monotone mapping, and let M : X → 2 X be

a maximal η-monotone mapping Then I  cM is maximal η-monotone for c > 0, where I is the

identity mapping.

Proof The proof follows on applyingDefinition 2.5

Proposition 2.7 see 4 Let η : X × X → X be t-strongly monotone, and let M : X → 2 X be maximal η-monotone Then generalized resolvent operator I  cM−1is single valued, where I is the identity mapping.

Proof For a given u ∈ X, consider x, y ∈ I  cM−1u for c > 0 Since M is maximal

Since η is t-strongly monotone, it implies x  y Thus, I  cM−1is single valued

Definition 2.8 Let η : X × X → X be t-strongly monotone, and let M : X → 2 Xbe maximal

η-monotone Then the generalized resolvent operator J M,η

c : X → X is defined by

Proposition 2.9 see 4 Let X be a real Hilbert space, let M : X → 2 X be maximal η-monotone,

and let η : X × X → X be t-strongly monotone Then the resolvent operator associated with M and defined by

Trang 14

Proof For any u, v ∈ X, it follows from the definition of the resolvent operator J ρ M,ηthat

Proposition 2.10 see 4 Let X be a real Hilbert space, let M : X → 2 X be maximal

η-monotone, and let η : X × X → X be t-strongly monotone.

If, in addition, (for γ > 0)

Trang 15

Proof We include the proof for the sake of the completeness To prove2.39, we apply 2.38

toProposition 2.9, and we get

When γ  1 and t > 1 inProposition 2.10, we have the following

Proposition 2.11 Let X be a real Hilbert space, let M : X → 2 X be maximal η-monotone, and let

For t  1 and γ > 1 inProposition 2.10, we find a result of interest as follows

Proposition 2.12 Let X be a real Hilbert space, let M : X → 2 X be maximal η-monotone, and let

Trang 16

Proposition 2.13 Let X be a real Hilbert space, let M : X → 2 X be maximal η-monotone, and let

This section deals with the over-relaxed η-proximal point algorithm and its application

to approximation solvability of the inclusion problem 1.1 based on the maximal

η-monotonicity Furthermore, some results connecting theη-monotonicity and corresponding

resolvent operator are established, that generalize the results on the firm nonexpansiveness

2, while the auxiliary results on maximal η-monotonicity and general maximal

mono-tonicity are obtained

Theorem 3.1 Let X be a real Hilbert space, and let M : X → 2 X be maximal η-monotone Then

the following statements are mutually equivalent.

i An element u ∈ X is a solution to 1.1.

ii For an u ∈ X, one has

η-mono-Next, we present a generalization to the relaxed proximal point algorithm3 based

on the maximalη-monotonicity.

Trang 17

Algorithm 3.2see 4 Let M : X → 2 Xbe a set-valued maximalη-monotone mapping on

X with 0 ∈ rangeM, and let the sequence {x k} be generated by the iterative procedure

Algorithm 3.3 Let M : X → 2 X be a set-valued maximalη-monotone mapping on X with

0∈ rangeM, and let the sequence {x k} be generated by the iterative procedure

Trang 18

Algorithm 3.4 Let M : X → 2 X be a set-valued maximalη-monotone mapping on X with

0∈ rangeM, and let the sequence {x k} be generated by the iterative procedure

are scalar sequences

In the following result4, we observe that Theorems1.1and1.3are unified and are

generalized to the case of the η-maximal monotonicity and super-relaxed proximal point

algorithm Also, we notice that this result in certain respects demonstrates the importance

of the firm nonexpansiveness rather than of the nonexpansiveness

Theorem 3.5 see 4 Let X be a real Hilbert space Let M : X → 2 X be maximal η-monotone,

and let xbe a zero of M Let η : X × X → X be t-strongly monotone Furthermore, assume (for

Trang 19

Then one has (for t ≥ 1)

In addition, suppose that the sequence {x k } is generated by Algorithm 3.2 as well, and that

M−1 is a-Lipschitz continuous at 0, that is, there exists a unique solution zto 0 ∈ Mz

(equivalently, M−10  {z}) and for constants a ≥ 0 and b > 0, one has

a2/c∗2 2γt − 1a2, α∗  lim supk → ∞ α k , and sequences {α k } and {c k } satisfy

α k ≥ 1, c k  c≤ ∞, inf k≥0α k > 0, and sup k≥0 α k < 2γt/2γt − 1.

Proof Suppose that xis a zero of M For all k ≥ 0, we set

J k I − J M,η

Therefore, J kx∗  0 Then, in light ofTheorem 3.1, any solution to1.1 is a fixed point of

J c M,η k , and hence a zero of J k

Next, the proof of 3.17 follows from a regular manipulation, and the followingequality:

Trang 20

Before we start establishing linear convergence of the sequence{x k }, we express {x k} in light

Trang 21

Now we find the estimate leading to the boundedness of the sequence{x k},

Thus, the sequence{x k} is bounded

We further examine the estimate

Trang 22

Now we turn our attentionusing the previous argument to linear convergence of thesequence{x k} Since limk → ∞ J kx k   0, it implies for k large that c−1

k J kx k  ∈ MJ M,η

c k x k.Moreover,c−1

k J kx k  ≤ b for k ≥ k and b > 0 Therefore, in light of 3.19, by taking w 

... manipulation, and the followingequality:

Trang 20

Before we start establishing linear convergence of... the relaxed proximal point algorithm3 based

on the maximalη-monotonicity.

Trang 17

Algorithm...

Trang 22

Now we turn our attentionusing the previous argument to linear convergence of thesequence{x

Ngày đăng: 21/06/2014, 20:20

Nguồn tham khảo

Tài liệu tham khảo Loại Chi tiết
9 J. Douglas Jr. and H. H. Rachford Jr., “On the numerical solution of heat conduction problems in two and three space variables,” Transactions of the American Mathematical Society, vol. 82, pp. 421–439, 1956 Sách, tạp chí
Tiêu đề: On the numerical solution of heat conduction problems in twoand three space variables,” "Transactions of the American Mathematical Society
Năm: 1956
10 J. Eckstein, Splitting methods for monotone operators with applications to parallel optimization, Doctoral dissertation, Department of Civil Engineering, Massachusetts Institute of Technology, Cambridge, Mass, USA, 1989 Sách, tạp chí
Tiêu đề: Splitting methods for monotone operators with applications to parallel optimization
Năm: 1989
11 J. Eckstein, “Nonlinear proximal point algorithms using Bregman functions, with applications to convex programming,” Mathematics of Operations Research, vol. 18, no. 1, pp. 202–226, 1993 Sách, tạp chí
Tiêu đề: Nonlinear proximal point algorithms using Bregman functions, with applications toconvex programming,” "Mathematics of Operations Research
Năm: 1993
14 M. C. Ferris, “Finite termination of the proximal point algorithm,” Mathematical Programming, vol. 50, no. 3, pp. 359–366, 1991 Sách, tạp chí
Tiêu đề: Finite termination of the proximal point algorithm,” "Mathematical Programming
Năm: 1991
15 O. G ¨uler, “On the convergence of the proximal point algorithm for convex minimization,” SIAM Journal on Control and Optimization, vol. 29, no. 2, pp. 403–419, 1991 Sách, tạp chí
Tiêu đề: On the convergence of the proximal point algorithm for convex minimization,” "SIAM"Journal on Control and Optimization
Năm: 1991
16 B. Martinet, “R´egularisation d’in´equations variationnelles par approximations successives,” Revue Francáaise d’Informatique et de Recherche Op´erationnelle, S´erie Rouge, vol. 4, no. 3, pp. 154–158, 1970.17 G. J. Minty, “Monotone nonlinear operators in Hilbert space,” Duke Mathematical Journal, vol. 29, pp.341–346, 1962 Sách, tạp chí
Tiêu đề: R´egularisation d’in´equations variationnelles par approximations successives,” "Revue"Francáaise d’Informatique et de Recherche Op´erationnelle, S´erie Rouge", vol. 4, no. 3, pp. 154–158, 1970.17 G. J. Minty, “Monotonenonlinearoperators in Hilbert space,” "Duke Mathematical Journal
Năm: 1962
20 A. Moudafi and M. Th´era, “Finding a zero of the sum of two maximal monotone operators,” Journal of Optimization Theory and Applications, vol. 94, no. 2, pp. 425–448, 1997 Sách, tạp chí
Tiêu đề: Finding a zero of the sum of two maximal monotone operators,” "Journal"of Optimization Theory and Applications
Năm: 1997
21 J.-S. Pang, “Complementarity problems,” in Handbook of Global Optimization, R. Horst and P. Pardalos, Eds., vol. 2 of Nonconvex Optimization and Its Applications, pp. 271–338, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1995.22 S. M. Robinson, “Composition duality and maximal monotonicity,” Mathematical Programming, vol.85, no. 1, pp. 1–13, 1999 Sách, tạp chí
Tiêu đề: Handbook of Global Optimization
Tác giả: J.-S. Pang
Nhà XB: Kluwer Academic Publishers
Năm: 1995
23 S. M. Robinson, “Linear convergence of epsilon-subgradient descent methods for a class of convex functions,” Mathematical Programming, vol. 86, pp. 41–50, 1999 Sách, tạp chí
Tiêu đề: Linear convergence of epsilon-subgradient descent methods for a class of convexfunctions,” "Mathematical Programming
Năm: 1999
24 R. T. Rockafellar, “On the maximal monotonicity of subdifferential mappings,” Pacific Journal of Mathematics, vol. 33, pp. 209–216, 1970 Sách, tạp chí
Tiêu đề: On the maximal monotonicity of subdifferential mappings,” "Pacific Journal of"Mathematics
Năm: 1970
25 R. T. Rockafellar, “Augmented Lagrangians and applications of the proximal point algorithm in convex programming,” Mathematics of Operations Research, vol. 1, no. 2, pp. 97–116, 1976 Sách, tạp chí
Tiêu đề: Augmented Lagrangians and applications of the proximal point algorithm inconvex programming,” "Mathematics of Operations Research
Năm: 1976
26 R. T. Rockafellar and R. J.-B. Wets, Variational Analysis, Springer, Berlin, Germany, 2004.27 M. V. Solodov and B. F. Svaiter, “An inexact hybrid generalized proximal point algorithm and some new results on the theory of Bregman functions,” Mathematics of Operations Research, vol. 25, no. 2, pp.214–230, 2000 Sách, tạp chí
Tiêu đề: Variational Analysis
Tác giả: R. T. Rockafellar, R. J.-B. Wets
Nhà XB: Springer
Năm: 2004
28 M. V. Solodov and B. F. Svaiter, “Forcing strong convergence of proximal point iterations in a Hilbert space,” Mathematical Programming, vol. 87, no. 1, pp. 189–202, 2000 Sách, tạp chí
Tiêu đề: Forcing strong convergence of proximal point iterations in a Hilbertspace,” "Mathematical Programming
Năm: 2000
29 W. Takahashi, “Approximating solutions of accretive operators by viscosity approximation methods in Banach spaces,” in Applied Functional Analysis, pp. 225–243, Yokohama, Yokohama, Japan, 2007 Sách, tạp chí
Tiêu đề: Approximating solutions of accretive operators by viscosity approximation methodsin Banach spaces,” in "Applied Functional Analysis
Năm: 2007
30 P. Tossings, “The perturbed proximal point algorithm and some of its applications,” Applied Mathematics and Optimization, vol. 29, no. 2, pp. 125–159, 1994 Sách, tạp chí
Tiêu đề: The perturbed proximal point algorithm and some of its applications,” "Applied"Mathematics and Optimization
Năm: 1994
31 P. Tseng, “Applications of a splitting algorithm to decomposition in convex programming and variational inequalities,” SIAM Journal on Control and Optimization, vol. 29, no. 1, pp. 119–138, 1991 Sách, tạp chí
Tiêu đề: Applications of a splitting algorithm to decomposition in convex programming andvariational inequalities,” "SIAM Journal on Control and Optimization
Năm: 1991
32 P. Tseng, “Alternating projection-proximal methods for convex programming and variational inequalities,” SIAM Journal on Optimization, vol. 7, no. 4, pp. 951–965, 1997 Sách, tạp chí
Tiêu đề: Alternating projection-proximal methods for convex programming and variationalinequalities,” "SIAM Journal on Optimization
Năm: 1997
Proceedings of the Royal Irish Academy. Section A, vol. 97, no. 1, pp. 83–86, 1997.35 R. U. Verma, “New class of nonlinear A-monotone mixed variational inclusion problems and resolvent operator technique,” Journal of Computational Analysis and Applications, vol. 8, no. 3, pp. 275–285, 2006 Sách, tạp chí
Tiêu đề: New class of nonlinear A-monotone mixed variational inclusion problems and resolvent operator technique
Tác giả: R. U. Verma
Nhà XB: Journal of Computational Analysis and Applications
Năm: 2006
36 R. U. Verma, “Nonlinear A-monotone variational inclusions systems and the resolvent operator technique,” Journal of Applied Functional Analysis, vol. 1, no. 2, pp. 183–189, 2006 Sách, tạp chí
Tiêu đề: Nonlinear "A"-monotone variational inclusions systems and the resolvent operatortechnique,” "Journal of Applied Functional Analysis
Năm: 2006
37 R. U. Verma, “A-monotonicity and its role in nonlinear variational inclusions,” Journal of Optimization Theory and Applications, vol. 129, no. 3, pp. 457–467, 2006 Sách, tạp chí
Tiêu đề: A"-monotonicity and its role in nonlinear variational inclusions,” "Journal of Optimization"Theory and Applications
Năm: 2006

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN

🧩 Sản phẩm bạn có thể quan tâm