This paper proposes an algorithmic framework for solving parametric optimization problems which we call adjoint-based predictor-corrector sequential convex programming.. predictor-correc
Trang 1ADJOINT-BASED PREDICTOR-CORRECTOR SEQUENTIAL CONVEX PROGRAMMING FOR PARAMETRIC NONLINEAR
QUOC TRAN DINH†, CARLO SAVORGNAN‡, AND MORITZ DIEHL‡
Abstract This paper proposes an algorithmic framework for solving parametric optimization
problems which we call adjoint-based predictor-corrector sequential convex programming After senting the algorithm, we prove a contraction estimate that guarantees the tracking performance of the algorithm Two variants of this algorithm are investigated The first can be used to treat on- line parametric nonlinear programming problems when the exact Jacobian matrix is available, while the second variant is used to solve nonlinear programming problems The local convergence of these variants is proved An application to a large-scale benchmark problem that originates from nonlinear model predictive control of a hydro power plant is implemented to examine the performance of the algorithms.
pre-Key words predictor-corrector path-following, sequential convex programming, adjoint-based
optimization, parametric nonlinear programming, online optimization
AMS subject classifications 49M37, 65K05, 90C31 DOI 10.1137/110844349
1 Introduction In this paper, we consider a parametric nonconvex
optimiza-tion problem of the form
where f :Rn → R is convex, g : R n → R mis nonlinear, Ω⊆ R n is a nonempty, closed
convex set, and the parameter ξ belongs to a given subset P ⊆ R p Matrix M ∈ R m ×p
plays the role of embedding the parameter ξ into the equality constraints in a linear way Throughout this paper, f and g are assumed to be differentiable on their domain Problem P(ξ) includes many (parametric) nonlinear programming problems such as
standard nonlinear programs, nonlinear second order cone programs, and nonlinearsemidefinite programs [32, 39, 47] The theory of parametric optimization has beenextensively studied in many research papers and monographs; see, e.g., [7, 25, 42].This paper deals with the efficient calculation of approximate solutions to a se-
quence of problems of the form P(ξ), where the parameter ξ is slowly varying In
∗Received by the editors August 15, 2011; accepted for publication (in revised form) July 6, 2012;
published electronically October 9, 2012 This research was supported by Research Council KUL CoE EF/05/006; Optimization in Engineering (OPTEC); IOF-SCORES4CHEM, GOA/10/009(MaNet), GOA/10/11; the Flemish Government through projects G.0452.04, G.0499.04, G.0211.05, G.0226.06, G.0321.06, G.0302.07, G.0320.08, G.0558.08, G.0557.08, G.0588.09, and G.0377.09; ICCoS, AN- MMM, and MLDM; IWT PhD grants, Belgian Federal Science Policy Office, IUAP P6/04; EU, ERNSI; FP7-HDMPC, FP7-EMBOCON; AMINAL; Helmholtz-viCERP, COMET-ACCM, ERC- HIGHWIND, and ITN-SADCO.
http://www.siam.org/journals/siopt/22-4/84434.html
†Department of Electrical Engineering (ESAT-SCD) and Optimization in Engineering Center
(OPTEC), K.U Leuven, B-3001 Leuven, Belgium, and Department of Informatics, Vietnam National University, Hanoi, Vietnam (quoc.trandinh@esat.kuleuven.be).
Mathematics-Mechanics-‡Department of Electrical Engineering (ESAT-SCD) and Optimization in Engineering
Center (OPTEC), K.U Leuven, B-3001 Leuven, Belgium (carlo.savorgnan@esat.kuleuven.be, moritz.diehl@esat.kuleuven.be).
1258
Trang 2other words, for a sequence{ξ k } k ≥0 such thatξ k+1− ξ k is small, we want to solve
in the result
In practice, sequences of problems of the form P(ξ) arise in the framework of
real-time optimization, moving horizon estimation, and online data assimilation as well as
in nonlinear model predictive control (NMPC) A practical obstacle in these tions is the time limitation imposed on solving the underlying optimization problemfor each value of the parameter Instead of solving completely a nonlinear program
applica-at each sample time [3, 4, 5, 29], several online algorithms approximapplica-ately solve theunderlining nonlinear optimization problem by performing the first iteration of exactNewton, sequential quadratic programming (SQP), Gauss–Newton, or interior pointmethods [17, 40, 54] In [17, 40] the authors only considered the algorithms in theframework of SQP methods This approach has been proved to be efficient in prac-tice and is widely used in many applications [14] Recently, Zavala and Anitescu [54]proposed an inexact Newton-type method for solving online optimization problemsbased on the framework of generalized equations [7, 42]
Other related work considers practical problems which possess general convexitystructure such as second order cone and semidefinite cone constraints and nonsmoothconvexity [21, 47] In these applications, standard optimization methods may notperform satisfactorily Many algorithms for nonlinear second order cone and nonlinearsemidefinite programming have recently been proposed and found many applications
in robust optimal control, experimental design, and topology optimization; see, e.g.[2, 21, 23, 33, 47] These approaches can be considered as generalizations of theSQP method Although solving semidefinite programming problems is in generaltime consuming due to matrix operations, in some practical applications, problemsmay possess few expensive constraints such as second order cone or semidefinite coneconstraints In this case handling these constraints directly in the algorithm may bemore efficient than transforming them into scalar constraints
Contribution The contribution of this paper is as follows:
(a) We start our paper by proposing a generic framework called the
adjoint-based predictor-corrector sequential convex programming (APCSCP) method
for solving parametric optimization problems of the form P(ξ) The algorithm
is specially suited for solving nonlinear MPC problems where the evaluations
of the derivatives are time-consuming For example, it can show advantageswith respect to standard techniques when applied to problems in which thenumber of state variables in the dynamic system is much larger than thenumber of control variables
(b) We prove the stability of the tracking error for this algorithm (Theorem 3.5).(c) In the second part of the paper the theory is specialized to the nonparametriccase where a single optimization problem is solved The local convergence ofthis variant is also obtained
(d) Finally, we present a numerical application to large-scale nonlinear modelpredictive control of a hydro power plant with 259 state variables and 10controls The performance of our algorithms is compared with a standardreal-time Gauss–Newton method and a conventional MPC approach.APCSCP is based on three main ideas: sequential convex programming, predictor-corrector path-following, and adjoint-based optimization We briefly explain thesemethods in the following
Trang 31.1 Sequential convex programming The sequential convex programming
(SCP) method is a local nonconvex optimization technique SCP solves a sequence
of convex approximations of the original problem by convexifying only the vex parts and preserving the structures that can efficiently be exploited by convexoptimization techniques [9, 36, 38] Note that this method is different from SQPmethods where quadratic programs are used as approximations of the problem Thisapproach is useful when the problem possesses general convex structures such as conicconstraints, a cost function depending on matrix variables, or convex constraintsresulting from a low-level problem in multilevel settings [2, 15, 47] Due to the com-plexity of these structures, standard optimization techniques such as SQP and Gauss–Newton-type methods may not be convenient to apply In the context of nonlinearconic programming, SCP approaches have been proposed under the names sequentialsemidefinite programming (SSDP) or SQP-type methods [12, 21, 23, 32, 33, 47] Ithas been shown in [18] that the superlinear convergence is lost if the linear semidef-inite programming subproblems in the SSDP algorithm are convexified In [35] theauthors considered a nonlinear program in the framework of a composite minimizationproblem, where the inner function is linearized to obtain a convex subproblem which
noncon-is made strongly convex by adding a quadratic proximal term
In this paper, following the work in [21, 24, 50, 52], we apply the SCP approach
to solve problem P(ξ) The nonconvex constraint g(x) + M ξ = 0 is linearized at each
iteration to obtain a convex approximation The resulting subproblems can be solved
by exploiting convex optimization techniques
We would like to point out that the term “sequential convex programming” wasalso used in structural optimization; see, e.g [22, 55] The cited papers are related
to methods of moving asymptotes introduced by Svanberg [49]
1.2 Predictor-corrector path-following methods In order to illustrate the
idea of the predictor-corrector path-following method [13, 54] and to distinguish itfrom the other “predictor-corrector” concepts, e.g., the well-known predictor-correctorinterior point method proposed by Mehrotra in [37], we summarize the concept ofpredictor-corrector path-following methods in the case Ω≡ R n as follows
The KKT system of problem P(ξ) can be written as F (z; ξ) = 0, where z = (x, y)
is its primal-dual variable The solution z ∗ (ξ) that satisfies the KKT condition for a
given ξ is in general a smooth map By applying the implicit function theorem, the derivative of z ∗(·) is expressed as
this case, linearization at (¯z, ¯ ξ) yields a formula that one step of a predictor-corrector
path-following method needs to satisfy:
Trang 4Written explicitly, it delivers the solution guess ˆz for the next parameter ˆ ξ as
guess ¯z if we employ the parameter embedding in the problem formulation [14].
Based on the above analysis, the predictor-corrector path-following method onlyperforms the first iteration of the exact Newton method for each new problem Inthis paper, by applying the generalized equation framework [42, 43], we generalizethis idea to the case where more general convex constraints are considered Whenthe parameter does not enter linearly into the problem, we can always reformulate
this problem as P(ξ) by using intermediate variables In this case, the derivatives
with respect to these intermediate variables contain the information of the predictorterm Finally, we notice that the real-time iteration scheme proposed in [17] can beconsidered as a variant of the above predictor-corrector method in the SQP context
1.3 Adjoint-based method From a practical point of view, most of the time
spent on solving optimization problems resulting from simulation-based methods isneeded to evaluate the functions and their derivatives [6] Adjoint-based methodsrely on the observation that it is not necessary to use exact Jacobian matrices ofthe constraints Moreover, in some applications, the time needed to evaluate all thederivatives of the functions exceeds the time available to compute the solution of theoptimization problem The adjoint-based Newton-type methods in [19, 28, 45] canwork with an inexact Jacobian matrix and only require an exact evaluation of theLagrange gradient using adjoint derivatives to form the approximate optimizationsubproblems in the algorithm This technique still allows the algorithm to converge
to the exact solutions but can save valuable time in the online performance of thealgorithm
1.4 A tutorial example The idea of the APCSCP method is illustrated in
the following simple example
Example 1.1 (tutorial example) Let us consider a simple nonconvex parametric
ξ − 1) T is a stationary point of problem (1.4) which
is also the unique global optimum It is clear that problem (1.4) satisfies the strong
second order sufficient condition (SSOSC) at x ∗
Trang 5Note that the constraint x2− x2+ 1 ≤ 0 can be rewritten as a second order
(1.4) can be cast into the form of P(ξ).
parameter ξ Instead of solving the nonlinear optimization problem at each ξ k untilcomplete convergence, APCSCP only performs the first step of the SCP algorithm to
obtain an approximate solution x k at ξ k Notice that the convex subproblem needed
We compare this method with other known real-time iteration algorithms The first
is the real-time iteration with an exact SQP method, and the second is the real-timeiteration with an SQP method using a projected Hessian [17, 31] In the secondalgorithm, the Hessian matrix of the Lagrange function is projected onto the cone ofsymmetric positive semidefinite matrices to obtain a convex quadratic programmingsubproblem
Figures 1.1 and 1.2 illustrate the performance of three methods when ξ k = 1.2 +
kΔξ k for k = 0, , 9 and Δξ k = 0.25 Figure 1.1 presents the approximate solution
trajectories given by three methods, while Figure 1.2 shows the tracking errors and
the cone constraint violations of those methods The initial point x0of three methods
is chosen at the true solution of P(ξ0) We can see that the performance of the exactSQP and the SQP using projected Hessian is quite similar However, the secondorder cone constraint(x1, 1) T
preserves the feasibility and better follows the exact solution trajectory Note thatthe subproblem in the exact SQP method is a nonconvex quadratic program, a convex
QP in the projected SQP case, and a second order cone constrained program (1.5) inthe SCP method
1 1.5 2 2.5 3
Exact−SQP
0 0.02 0.04 0.06 0.08
Projected−SQP
0 0.02 0.04 0.06
SCP
||x k −x*(ξk)||
SOC const viol.
Fig 1.2 The tracking error and the cone constraint violation of three methods ( k = 0, , 9).
Trang 61.5 Notation Throughout this paper, we use the notation ∇f for the
val-ued function g, and S n (resp., S n
+ and S n
(resp., positive semidefinite and positive definite) matrices The notation· stands
B(x, r) := {y ∈ R n | y − x < r} and ¯ B(x, r) is its closure.
The rest of this paper is organized as follows Section 2 presents a generic work of the APCSCP algorithm Section 3 proves the local contraction estimate for
frame-APCSCP and the stability of the approximation error Section 4 considers an
adjoint-based SCP algorithm for solving nonlinear programming problems as a special case.
The last section presents computational results for an application of the proposedalgorithms in NMPC of a hydro power plant
2 An APCSCP algorithm In this section, we present a generic algorithmic
framework for solving the parametric optimization problem P(ξ) Traditionally, at
com-pletely converged solution ¯z(ξ k) Exploiting the real-time iteration idea [14, 17] inour algorithm below, only one convex subproblem is solved to get an approximated
solution z k at ξ k to ¯z(ξ k)
Suppose that z k := (x k , y k)∈ Ω×R m is a given KKT point of P(ξ k) (more details
x L(¯z k), where L is the
correction term of the inconsistency between A k and g (x k ) Vector y k is referred to
semidefinite, the subproblem P(z k , A k , H k ; ξ) is convex Here, z k , A k and H k areconsidered as parameters
Remark 1 Note that computing the term g (x k)T y k of the correction vector s k does not require the whole Jacobian matrix g (x k), which is usually time-consuming
to evaluate
When implementing the algorithm, the evaluation of the directional derivatives
η k := g (x k)T y k can be done by the reverse mode (or adjoint mode) of automatic
dif-ferentiation By using this technique, we can evaluate an adjoint directional derivative
vector of the form g (x k)T y k without evaluating the whole Jacobian matrix g (x k) of
the vector function g More details about automatic differentiation can be found
in a monograph [26] or at http://www.autodiff.org In the NMPC framework, the
constraint function g is usually obtained from a dynamic system of the form
(2.1)
˙
η(t) = G(η(t), x, t), t0≤ t ≤ t f , η(t0) = η0(x),
by applying a direct transcription, where η is referred to as a state vector and x
is a parameter vector The adjoint directional derivative vector g (x) T y is nothing
Trang 7more than the gradient vector ∂V ∂x of the function V (x) := g(x) T y In the dynamic
system context, this function V is a special case of the general functional V (x) :=
e(η(t f)) +t f
t0 v(η, x, t)dt By simultaneously integrating the dynamic system and its
adjoint sensitivity system ˙λ = −G T
η and λ(t f) =∇ η e(η(t f)), we can evaluate the
gradient vector of V with respect to x as dV
t0 (v x + λ T G x )dt, where
λ(t0) is the solution of the adjoint system at t0 Note that the cost of integrating theadjoint system is of the same order as integrating the forward dynamics and, crucially,
independent of the dimension of x Adjoint differentiation of dynamic systems is
performed, e.g., in an open-source software package, Sundials [46] For more details
of adjoint sensitivity analysis of dynamic systems, see [10, 46]
The APCSCP algorithmic framework is described as follows
Initialization For a given parameter ξ0 ∈ P, solve approximately (offline)
P(ξ0) to get an approximate KKT point z0 := (x0, y0) Compute g(x0), find a
ma-trix A0 which approximates g (x0), and H0 ∈ S n
Step 1 Get a new parameter value ξ k+1 ∈ P.
Step 2 Solve the convex subproblem P(z k , A k , H k ; ξ k+1) to obtain a solution x k+1
and the corresponding multiplier y k+1.
Step 3 Evaluate g(x k+1), update (or recompute) matrices A k+1and H k+1∈ S n
The core step of Algorithm 1 is to solve the convex subproblem P(z k , A k , H k ; ξ)
at each iteration In Algorithm 1 we do not mention explicitly the method to solve
P(z k , A k , H k ; ξ) In practice, to reduce the computational time, we can either
imple-ment an optimization method which exploits the structure of the problem, e.g., blockstructure or separable structure [22, 51, 55], or rely on several efficient methods andsoftware tools that are available for convex optimization [9, 38, 39, 48, 53] In this
paper, we are most interested in the case where one evaluation of g is very expensive.
A possible simple choice of H k is H k = 0 for all k ≥ 0.
The initial point z0is obtained by solving offline P(ξ
0) However, as we will show
later (Corollary 3.6), if we choose z0 close to the set of KKT points Z ∗ (ξ
0) of P(ξ0)
(not necessarily an exact solution), then the new KKT point z1 of P(z0, A
0, H0; ξ1)
is still close to Z ∗ (ξ
1) of P(ξ1) provided thatξ1− ξ0 is sufficiently small Hence, in
practice, we only need to solve approximately problem P(ξ0) to get a starting point z0.
In the NMPC framework, the parameter ξ usually coincides with the initial state
g (x k ), the exact Jacobian matrix of g at x k and H k ≡ 0, then this algorithm collapses
to the real-time SCP method considered in [52].
3 Contraction estimate In this section, we will show that under certain
as-sumptions, the sequence{z k } k ≥0 generated by Algorithm 1 remains close to the quence of the true KKT points{¯z k } k ≥0 of problem P(ξ k) Without loss of generality,
se-we assume that the objective function f is linear, i.e., f (x) = c T x, where c ∈ R n is
given Indeed, since f is convex, by using a slack variable s, we can reformulate P(ξ)
as a nonlinear program min(x,s)
s | g(x) + Mξ = 0, x ∈ Ω, f(x) ≤ s
Trang 83.1 KKT condition as a generalized equation Let us first define the
Lagrange function of problem P(ξ) as
Note that the first line of (3.1) implicitly includes the constraint x ∈ Ω.
A pair (¯x(ξ), ¯ y(ξ)) satisfying (3.1) is called a KKT point of P(ξ) and ¯ x(ξ) is
denote by Z ∗ (ξ) and X ∗ (ξ) the set of KKT points and the set of stationary points
of P(ξ), respectively In what follows, we use the letter z for the pair of (x, y), i.e.,
z := (x T , y T)T
Throughout this paper, we require the following assumptions, which are standard
in optimization
Assumption 1 The function g is twice differentiable on their domain.
Assumption 2 For a given ξ0 ∈ P, problem P(ξ0) has at least one KKT point
parametric generalized equation as follows:
M
Generalized equations are an essential tool to study many problems
in nonlinear analysis, perturbation analysis, variational calculations, and optimization[8, 34, 43]
Suppose that for some ξ k ∈ P, the set of KKT points Z ∗ (ξ k ) of P(ξ k) is nonempty.
For any fixed ¯z k ∈ Z ∗ (ξ k), we define the following set-valued mapping:
(3.5) L(z; ¯ z k , ξ k ) := F (¯ z k ) + F (¯z k )(z − ¯z k ) + Cξ k+N K (z).
We also define the inverse mapping L −1:Rn +m → R n +m of L( ·; ¯z k , ξ k) as follows:
z ∈ R n +m : δ ∈ L(z; ¯z k , ξ k) .
Now, we consider the KKT condition of the subproblem P(z k , A k , H k ; ξ) For
given neighborhoods B(¯z k , r z) of ¯z k and B(ξ k , r ξ ) of ξ k , and z k ∈ B(¯z k , r z ), ξ k+1 ∈
Trang 9B(ξ k , r ξ ), and given matrices A k and H k ∈ S n
+, let us consider the convex
subprob-lem P(z k , A k , H k ; ξ k+1) with respect to the parameter (z k , A k , H k , ξ k+1) The KKT
condition of this problem is expressed as follows
(3.7)
0∈ c + m(z k , A k ) + H k (x − x k ) + A T k y + NΩ(x),
0 = g(x k ) + A k (x − x k ) + M ξ k+1,
holds for the subproblem P(z k , A k , H k ; ξ k+1), i.e.,
ri(Ω)∩ x ∈ R n | g(x k ) + A k (x − x k ) + M ξ k+1= 0 = ∅,
where ri(Ω) is the relative interior of Ω Then by convexity of Ω, a point z k+1 :=
(x k+1, y k+1) is a KKT point of P(z k , A k , H k ; ξ k+1) if and only if x k+1is a solution to
P(z k , A k , H k ; ξ k+1) associated with the multiplier y k+1.
Since g is twice differentiable by Assumption 1 and f is linear, for a given z = (x, y), we have
the Hessian matrix of the Lagrange function L, where ∇2g i(·) is the Hessian matrix
of g i (i = 1, , m) Let us define the following matrix:
k , and ξ k+1 are considered as parameters Note that if A k = g (x k) and
3.2 The strong regularity concept We recall the following definition of the
strong regularity concept This definition can be considered as the strong regularity
of the generalized equation (3.4) in the context of nonlinear optimization; see [42].Definition 3.1 Let ξ k ∈ P such that the set of KKT points Z ∗ (ξ k ) of P(ξ k ) is
nonempty Let ¯ z k ∈ Z ∗ (ξ
k ) be a given KKT point of P(ξ k ) Problem P(ξ k ) is said to
be strongly regular at ¯ z k if there exist neighborhoods B(0, ¯r δ ) of the origin and B(¯z k , ¯ r z)
of ¯ z k such that the mapping z ∗
Trang 10From the definition of L −1 where strong regularity holds, there exists a unique
is equivalent to SSOSC [20] In order to interpret the strong regularity condition of
Here, δ = (δ c , δ g)∈ B(0, ¯r δ ) is a perturbation Problem P(ξ k ) is strongly regular at ¯ z k
if and only if (3.12) has a unique KKT point z ∗
k (δ) in B(¯z k , ¯ r z ) and z ∗
k(·) is Lipschitz
continuous inB(0, ¯r δ ) with a Lipschitz constant γ.
Example 3.2 Let us recall example 1.1 in section 1 The optimal multipliers
2+ 2− 4ξ = 0 and x2 − x2 + 1 ≤ 0 are
y ∗ = (2√
ξ − 1)[8ξ2− ξ √ ξ] −\1 > 0 and y ∗ = [8
ξ2− ξ √ ξ] −1 > 0, respectively.
Since the last inequality constraint is active, while x ≥ 0 is inactive, we can easily
compute the critical cone as C(x ∗
2 > 0 which says that the strict complementarity condition holds.
Therefore, problem (1.1) satisfies the the strong second order sufficient condition On
the other hand, it is easy to check that the LICQ condition holds for (1.1) at x ∗
applying [42, Theorem 4.1], we can conclude that (1.4) is strongly regular at (x ∗
The following lemma shows the nonemptiness of Z ∗ (ξ) in the neighborhood of ξ k.
a given ξ k ∈ P Suppose further that problem P(ξ k ) is strongly regular at ¯ z k for a given ¯ z k ∈ Z ∗ (ξ
k ) Then there exist neighborhoods B(ξ k , r ξ ) of ξ k and B(¯z k , r z ) of
Proof Since the KKT condition of P(ξ k) is equivalent to the generalized equation
(3.4) with ξ = ξ k, by applying [42, Theorem 2.1] we conclude that there exist borhoods B(ξ k , r ξ ) of ξ k and B(¯z k , r z) of ¯z k such that Z ∗ (ξ
neigh-k+1) is nonempty for all
ξ k+1∈ B(ξ k , r ξ ) and Z ∗ (ξ
k+1)∩ B(¯z k , r z) contains only one point ¯z k+1 On the other
hand, sinceF (¯ z k ) + Cξ k − F (¯z k)− Cξ k+1=M(ξ k − ξ k+1) ≤ M ξ k+1− ξ k ,
by using the formula [42, 2.4], we obtain the estimate (3.13)
Trang 113.3 A contraction estimate for APCSCP using an inexact Jacobian matrix In order to prove a contraction estimate for APCSCP, throughout this sec-
tion we make the following assumption
Assumption 3 For a given ¯ z k ∈ Z ∗ (ξ k ), k ≥ 0, the following conditions are
k is defined by (3.9) and γ is the constant in Definition 3.1.
(b) The Jacobian mapping F (·) is Lipschitz continuous on B(¯z k , r z) around ¯z k,i.e., there exists a constant 0≤ ω < +∞ such that
(3.15) F (z) − F (¯z k) ≤ ωz − ¯z k ∀ z ∈ B(¯z k , r z ).
Note that Assumption 3 is commonly used in the theory of Newton-type andGauss–Newton methods [13, 16], where the residual term is required to be sufficientlysmall in a neighborhood of the local solution From the definition of ˜F
k depends on the norms of ∇2
x L(¯z k)−H k and g (¯x k)−A k These
∇2
x L(¯z k ) and the Jacobian matrix g (¯x k), respectively On the one hand, Assumption
3(a) requires the positive definiteness of H k to be an approximation of∇2
x L (which is
not necessarily positive definite) On the other hand, it requires that matrix A k is a
stationary point ¯x k Note that the matrix H k in the Newton-type method proposed
in [7] is not necessarily positive definite
Now, let us define the following mapping:
where ˜F
Lipschitz continuous in a neighborhood of ¯v k := ˜F
k¯k − F (¯z k)− Cξ k
exist neighborhoods B(ξ k , r ξ ) and B(¯z k , r z ) such that if we take any z k ∈ B(¯z k , r z ) and
ξ k+1∈ B(ξ k , r ξ ), the mapping J k defined by (3.16) is single-valued in a neighborhood B(¯v k , r v ), where ¯ v k := ˜F
k¯k − F (¯z k)− Cξ k Moreover, the following inequality holds:
(3.17) J k (v) − J k (v ) ≤ β v − v ∀v, v ∈ B(¯v k , r v ),
where β := 1−γκ γ > 0 is a Lipschitz constant.
Proof Let us fix a neighborhood B(¯v k , r v) of ¯v k Suppose for contradiction that
J k is not single-valued inB(¯v k , r v ); then for a given v the set J k (v) contains at least two points z and z such thatz − z = 0 We have
Trang 12Now, using the strong regularity assumption of P(ξ k) at ¯z k, it follows from (3.20)that
Finally, we prove the Lipschitz continuity of J k Let z = J k (v) and z = J k (v ),
where v, v ∈ B(¯v k , r v) Similar to (3.20), these expressions can be written equivalentlyto
δ ∈ F (¯z k ) + F (¯z k )(z − ¯z k ) + Cξ k+N K (z)
and(3.22)
Trang 13By using again the strong regularity assumption, it follows from (3.22) and (3.23)that
which shows that J k satisfies (3.17) with a constant β := 1−γκ γ > 0.
Let us recall that if z k+1 is a KKT point of the convex subproblem P(z k , A k , H k;
ξ k+1), then
0∈ ˜ F
k (z k+1− z k ) + F (z k ) + Cξ k+1+N K (z k+1).
According to Lemma 3.4, if z k ∈ B(¯z k , r z ), then problem P(z k , A k , H k ; ξ) is uniquely
solvable We can write its KKT condition equivalently as
Since ¯z k+1 is the solution of (3.4) at ξ k+1, we have 0 = F (¯ z k+1) + Cξ k+1 + ¯u k+1,
where ¯u k+1∈ N K(¯z k+1) Moreover, since ¯z k+1= J k( ˜F
The main result of this section is stated in the following theorem
Then, for k ≥ 0 and ¯z k ∈ Z ∗ (ξ k ), if P(ξ k ) is strongly regular at ¯ z k , then there exist neighborhoods B(¯z k , r z ) and B(ξ k , r ξ ) such that the following hold:
(a) The set of KKT points Z ∗ (ξ k+1) of P(ξ k+1) is nonempty for any ξ k+1 ∈ B(ξ k , r ξ ).
(b) If, in addition, Assumption 3(a) is satisfied, then subproblem P(z k , A k , H k;
ξ k+1) is uniquely solvable in the neighborhood B(¯z k , r z ).
(c) Moreover, if, in addition, Assumption 3(b) is satisfied, then the sequence
{z k } k ≥0 generated by Algorithm 1, where ξ k+1 ∈ B(ξ k , r ξ ), guarantees
Proof We prove the theorem by induction For k = 0, we have Z ∗ (ξ0) is nonempty
by Assumption 2 Now, we assume Z ∗ (ξ k ) is nonempty for some k ≥ 0 We will prove
k ) such that P(ξ k) is strong regular at ¯z k Now, by applying Lemma 3.3 to
prob-lem P(ξ k), we conclude that there exist neighborhoodsB(¯z k , r z) of ¯z k and B(ξ k , r ξ)
of ξ k such that Z ∗ (ξ k
+1) is nonempty for any ξ k+1∈ B(ξ k , r ξ)
... tool to study many problemsin nonlinear analysis, perturbation analysis, variational calculations, and optimization[ 8, 34, 43]
Suppose that for some ξ k ∈ P, the... regularity
of the generalized equation (3.4) in the context of nonlinear optimization; see [42].Definition 3.1 Let ξ k ∈ P such that the set of KKT points Z ∗... it is easy to check that the LICQ condition holds for (1.1) at x ∗
applying [42, Theorem 4.1], we can conclude that (1.4) is strongly regular at (x ∗