Вычислительные технологии Том 4, № 3, 1999
GENERALIZED SEMI-INFINITE OPTIMIZATION: ON SOME FOUNDATIONS
G.-W. Weber Darmstadt University of Technology Department of Mathematics, Darmstadt, Germany e-mail: [email protected]
Рассматривается задача полубесконечной оптимизации в общем виде
( Min f (x) на Msi[h, g], где
Vsi (f,h,g,u,v) <j Msi [h, g] := { x € IRn | hi (x) = 0 (i € I),
( g(x,y) > 0 (y € Y(x)) },
в которой Y (x) = Mf [u(x, -),v(x, ■)] (x € IRn) — допустимые множества в смысле оптимизации с конечными ограничениями. Предполагается, что некоторые ограничения выполнены для Y (x), (LICQ или MFCQ), локально (или глобально) по x. Задача Psi(f,h,g,u,v) может быть локально (или глобально) представлена как обычная полубесконечная оптимизационная задача. Таким образом, используются два различных подхода, каждый из которых с или без предположения компактности на Y (x). Более того, для Psi (f, h, g, u, v) мы предлагаем необходимые и, при некоторых дополнительных предположениях, достаточные условия оптимальности первого порядка, которые в специальном случае были впервые предложены Кайзером и Крабсом.
1. Introduction, a first equivalent model
In semi-infinite optimization very often the index set Y of maybe infinitely many inequality constraints does not depend on the state x £ Mn. The set of equality constraints is finite. For problems of this kind which we call ordinary semi-infinite optimization problems, a lot of research has been done; we mention, e.g., [4, 13, 26] and, from special viewpoints, [21, 23, 24]. In this research, however, we are concerned with generalized semi-infinite optimization problems. These problems have the form
Minimize f (x) on MSI[h,g], where
Psi (f,h,g,u,v)
Msi[h,g] := { x £ Rn | hi(x) = 0 (г £ I),
g(x,y) > 0 (y £ Y(x)) }
and
Y(x) := Mf[u(x, ■), v(x, ■)] (x £ IRn),
© G.-W. Weber, 1999.
i.e., now the index set Y, with its perhaps infinitely many elements, depends on the state x £ Rn. Moreover, Y(x) is a feasible set in the sense of finitely constrained optimization (F), i.e., the set of inequality constraints has only finitely many elements. For our problem, SI abbreviates its semi-infiniteness.
Let h, u, v comprise the component functions hi : Rn ^ R, i £ I := {1,... ,m}, uk : Rn xRq ^ R, k £ K := {1,...,r}, and vg : Rn xRq ^ R, t £ L := {1,... ,s}, respectively. Moreover, we assume that f : Rn ^ R, g : Rn x Rq ^ R, hi (i £ I), uk (k £ K) and vg (t £ L) are C^functions (continuously differentiable). For each C^function, e.g. for f, Df (x) and DT f (x), denote the row- and the column-vector of the first order partial derivatives d
——f(x) (k £ {1,...,n}; x £ Rn), respectively. Let, e.g., Dxg(x,y), Dyg(x,y), analogously
dxK
comprise the partial derivatives of g due to the coordinates xK and ya, respectively (and so on for the other defining functions).
For problems of that generalized form we refer, from the topological (generic) point of view, to [12, 20]. Hereby, a basic aspect is the question whether the so-called Reduction-Ansatz applies or, to what an extent it is violated. That Reduction-Ansatz means the opportunity locally (e.g., around certain critical points) to express the semi-infinite optimization problem as a finite optimization problem (cf. [11, 33]). This paper, however, is based on the question of local (or global) representability of our generalized semi-infinite optimization problem as an ordinary semi-infinite optimization problem. Hereby, we walk along two different ways, the first of which is presented in Section 1, the other one in Section 2. The two ways (approaches, models) are (critically) commented and compared.
In both of these ways, we assume for Y(x) C Rq, locally in x, the linear independence constraint qualification, the Mangasarian — Fromovitz constraint qualification, or a connectedness property. Finally, we shall refer to a local minimum x and to a small neighborhood of this point. Then, approaching our models with the help of appropriate other assumptions, we are in a position to derive first order necessary or sufficient optimality conditions, and, in the further paper [32], the iteration procedures which were presented by Kaiser and Krabs [28]. In [28] the authors in particular refer to the special case of an interval Y(x) := [a(x), b(x)] (q = 1). Kaiser's and Krabs' investigation was motivated by a concrete problem from mechanical engineering (see [2]). That problem consists in
— optimizing the layout of a special assembly line:
under certain constraints, the objects to be transported in periodic time intervals, shall be moved as far as possible in each period. (For the modelling see also [27].)
As some further examples which can, under appropriate assumptions, be modelled as a generalized semi-infinite optimization problem we mention the following problems:
— reverse Chebychev approximation:
motivated by the approximation of a thermocouple characteristic in chemical engineering [34] or by digital filtering (cf. [15, 19, 25]),
— maneuverability of a robot (cf. [7, 19, 25]),
— time optimal control (e. g., time minimal heating or cooling of a ball of some homogeneous material; cf. [25]),
— structure and stability in optimal control of an ordinary differential equation (cf. [31]).
We remark that in [28] there are also considerations on sets Y(x) which do not have the
structure of a feasible set.
This present research has also some relations with the recent article [19] of Jongen, Ruckmann and Stein. While in [19] as the main result a general Fritz — John type necessary optimality
theorem is proved, here our necessary optimality conditions are closer to the Kuhn —Tucker type. Mainly, however, we shall for a local minimum x describe the necessary nonnegativity of the function £ ^ Df (x)£ (£ £ Rn) over some linearized tangent cone of the feasible set at x. Finally, in [19] a Kuhn — Tucker theorem is concluded, too. Moreover, in [19] assumptions on compactness of Y(x) (x £ Rn) and on upper semi-continuity of Y(•) in the sense of Berge are made, while no constraint qualification is assumed. In our paper such a compactness (or boundedness) assumption is discussed, but not always necessary, and a (pointwise) constraint qualification or some connectedness property with respect to Y(•) is assumed. (The article [19] and the present research with their results, proofs and further considerations were started independently, under different aspects and motivating examples.)
These problem representations and optimality conditions play an important part for creating methods to find local or global minima. In this context of iteration procedures, topological questions of the behavior of the feasible set MSI[h, g] under data perturbations naturally arise. In [32] these questions are treated by means of generalizing some of the results on ordinary semi-infinite optimization made by Jongen, Twilt and Weber ([21]).
We introduce the following two boundedness assumptions:
Assumption A (Boundedness). There is a bounded open neighborhood U C Rn of Msi[h,g](c U), such that ]Jxey Y(x) is bounded, and, as a weaker condition,
Assumption Auo (Boundedness, locally). There is a bounded set U0 C Rn with Msi[h,g] HU0 = 0 such that Uxeyo Y(x) is bounded.
Hereby, e.g., U denotes the closure of U. In the sequel we shall precisely discuss, to what an extent global or local assumptions on compactness, or constraint qualifications, should be satisfied.
Under these differentiability and, hence, continuity assumptions all the sets Y(x) (x £ Rn) are closed. By means of a small argumentation on continuity and compactness, we see that the bounded sets (x), UxeyoY(x) from Assumptions A and Ayo, respectively, are even
compact. Hence, from now on we may call the Assumptions A and Ayo, compactness assumptions.
Remark1. In the case of any constraint qualification on Y(x) (for all x £U or for all x £ U0) which we shall make, and if due to no converging sequence (xK)K&N (in U or U0, respectively) and for no sequence (yK)K£N lying "at infinity", the sequences (uk(xK,yK))K&N (k £ K), and (mmi^L vg(xK ,yK))K&N, approach 0 or their members are non negative, respectively, then that compactness (of the unions) is not only sufficient but also necessary for the boundedness (compactness) of each of these sets Y(x). This fact follows from a topological consideration based on [21].
For the set of active inequality constraints at a elements x1 £ MSI[h,g], x2 £ Rn and at an element y £ Mf[u(x2, •),v(x2, •)] (or, lateron, y £ Rq), respectively, we write
Definition 1.1. Let points x £ Rn and y £ Y(x) be given. We say that the linear independence constraint qualification, in short: LICQ, holds at y as an element of the feasible set Mf[u(x, •),v(x, •)] if the vectors
Dyuk(x,y) (k £ K), Dyvg(x,y) (t £ L0(x,y)) xIf the reader is not so much interested in additional theoretical details, then he may skip the remark.
Yo(x1):= {y e Y(x1)\g(x1,y) = 0], Lo(x2,y) := {£ e L\v£(x2,y) = 0}.
(1.1) (1.2)
are linearly independent.
The linear independence constraint qualification (LICQ) is said to hold for Mf[u(x, ■), v(x, ■)] if LICQ is fulfilled for all y £ Y(x).
Assumption B (LICQ). LICQ holds for all sets MF[u(x,-),v(x, ■)] (x £ U) where U C _Rn is an open neighborhood of MSI[h,g] (cf., e.g., Assumption A), or
Assumption By 0 (LICQ, locally). Referring to some given open set U0 C (cf., e.g., Assumption Au0) LICQ holds for all sets Mf[u(x, ■), v(x, -)](x £ U0).
The topological investigation [18] on parametric finitely constrained optimization indicates that the Assumption B on (with the parameter x) global validity of LICQ is very strong (cf. also [10, 22]). As for our purposes of optimality conditions we do not globally need to represent our generalized semi-infinite problem by an ordinary one, but locally, Assumption Byo will finally be sufficient. Hereby, on the one hand the special case U0 = U will be included in our general considerations. On the other hand, due to a small neighborhood U1 of a given point x £ iRn, the validity for Y(x) of both LICQ and Assumption Ayi is sufficient for Assumption Byo to hold, where U0 C U1 is a sufficiently small other neighborhood of x.
Only for both the ease of the following (re)presentation of our problem with the help of an ordinary semi-infinite optimization problem and in order to work out the case of locally holding compactness being important for iteration procedures, we may without loss of generality make the Assumption Ayo.
Nevertheless, until we arrive at that ordinary semi-infinite problem pOi(f, h, g0, u0, v0), our explanations need some technical effort.2
Then, Assumption Byo gives us the opportunity, for a fixed x £ U0 locally around a given point y £ Y(x) to linearize Y(x) for x close to x (cf. [16, 22]). Therefore we define z = F(x, y) as follows:
zi := ui(x,y)
zr := ur(x,y) Zr+i := v^i (x,y)
Zr+p := vip(x,y)
zr+p+i := Cl(y - y)
zq : Cq-r-p(y — y) „
((x,y) £ Rn x Rq),
(1.3a)
where p is the cardinality |L0(x, y)| of L0(x, y) = j^1,... and where the vectors £ (n £ {1,... , q — r — p}) complete the set { Dyuk(x, y) | k £ K } U { Dyv^(x,y) | € £ L0(x,y) } to a basis of . Let us put
:F(x, y) := (x, fP(x, y)) ((x, y) £ IT x lRq).
(1.3b)
2If the reader is not very interested in the topological details, he might skip those details and immediately turn to Theorem 1.2.
Then, we realize that the partitioned matrix of derivatives of the component functions
(
DFF(x,y)
O
\
nxq
V
DxF(x,y)
Dy F(x, y)
(1.4a)
/
where In = unit-matrix of lRn and Onxq = zero-matrix of type n x q, is nonsingular if and only if the matrix
/ : \
Dy F(x,y) =
DyUk (x,y) (k G K)
DyVf(x, y) (i G Lo(x,y))
Ci (n G {1,
, q — r — p})
(1.4b)
V : /
is nonsingular. The latter condition, however, is guaranteed by Assumption B^o and by the choice of the vectors . Now, by means of applying the inverse function theorem at (x, y) on F we conclude that there exist open and bounded neighborhoods U C lRn, U2 C lRq around (x, y) respectively, such that F := F"|(Ui x U2) : U x U2 ^ W := F7(Ui x U2) is a C 1-diffeomorphism. Now, let || • denote the maximum norm of the underlying Euclidean space. Shrinking U1, U2, if necessary, we can guarantee that on the one hand W is an axis-parallel open cube around (x, 0q) G lRn x lRq: This means W = C1 xC2, where C1 = C (x, p1) := {x G lRn | ||x — X||^ < p1 }, C2 = C(0p,p2) stand for the open cubes C1, C2 of the || • ||^-radii (= half the length) p1, p2, around x, 0p, respectively, and being parallel with respect to the axis. On the other hand, we have L0(x,y) C L0(x,y) for all (x,y) C U1 x U2. Then, for each x G U1 the mapping := (F(x, 0)|U2 : U2 —► C2 is a C 1-diffeomorphism which also transforms the (relative) neighborhood Y(x) nU2 of y onto the (relative) neighborhood
({0r} x Hp x lRq-r-p) nC2 C Rq
of 0q: Hereby, Hp denotes the nonnegative orthant { z G Rp | zg > 0 (£ G {1,...,p}) } of Rp. We call a canonical local change of the coordinates (of) y.
For all points x G U1, z G C2 we have the pre-image for F(x, •), F(x, •) — z of the corner point 0q pointwise being given by means of implicit C1 -functions y(-), y(-, •) of x and (x,z), respectively, i. e.
0-1(°q)= y(x), |L0(x,y(x))| = P (x GU1),
= y(x,z)-
Performing the construction of a C 1-family (4>X)x&uf of diffeomorphisms 0X : Uf
(1.5a) (1.5b)
^ Cf for
each a = (x,y) where x G U0, y G Y(x). In particular, we may choose an open covering (Uf X UDaeAof A := {(x, y) | x G U0, y G Y (x)} where X G Uf C Uf C Uf, y G ¿/f C ¿if C Uf, and where Wf := JF(U/f X Uf) is an (axis-parallel) open subcube Wf = ¿^f X Cf of W = Wf.
n
From Assumption Ay0, by means of an argumentation with (sub)sequences we conclude that A is compact. Hence, there exist finitely many points aj = (xj, yj) 6 A, j £ J := {1,... , w}, such that (UfJ x Uf )jeJ is an open covering of A.
We note that in the general case without a compactness assumption, this subcovering can be chosen with JC IN and locally finite. This means: for each a:= (x, y) £ {(x, y) £U0x!Rq|y £ Y(x)} there is a neighbourhood Vsuch that VH (Uf x Uf) = 0 for only finitely many j £ J. (Hereby, the open set U0 needs not to be bounded.)
Let us give the idea how to achieve such a locally finite structure. Therefore, we can decompose the set A by means of intersecting it with the countably many compact cubes C = C1 x C2 := [2v, 2v + 1]n+q C 1Rn x 1Rq (v £ Z, i. e., v is an integer). These cubes a]together cover 1Rn x 1Rq. Let some v0 £ Z be given. The intersection An CV0 = {(x,y) | x £ U0 H C^0, y £ Y(x) H C20} is compact. Hence, in CV0 we are in a similar situation as under Assumption Ay0, such that we may choose a finite open covering OV0 of An CV0. Taking into account all these open coverings Ov (v £ Z) and enumerating all their members (open sets) by means of j £ J C IN, we finally arrive at a suitable locally finite open covering of A. If A is actually known not to be compact (the case of unboundedness), then we may even choose J = IN.
Now, with our (local) linearizations of Y(x) (x £Uf, a £ A) we are able equivalently to represent our inequality constraints on x (on the "upper stage") without x-dependence of the
index set. In fact, writing p> := |L0(xj, yj) |, := ({0r} x x Mq-r-pj) H Cf (j £ {1,..., w}) and j C2, C2 for , , Cf, we have for each x £ U0:
g(x,y) > 0 for all y £ Y(x) ^ g(x, (0X)-1 (z)) > 0 for all z £ ({0r} x HpJ x Mq-r-p3) n Cj,
if x £ Ufj, j £ J;
^ g(x, (0X)-1(z)) > 0 for all z £ Zj, if x £ Uf, j £ J.
Finally, let us for each k £ {1,...,k0} by means of our set inclusions from above and of a C^- partition of unity (cf. [14, 17]) glue together (x,z) ^ g(x, (0X)-1(z)) with 0 in Vj := Uf x C2. Namely, we let the resulting function coincide with g(x, (0X)-1(z)) for all
(x, z) £ Vj and with 0 on Vj \ Vj. Hereby, l^j, and Vj are open subsets of Vj being chosen such that with Vj := Ullf x ¿j it holds Vj C V, Vj C Vj, V C Vj. So, we have immediately extended those functions gj0 by means of 0 outside of Vj (j £ J; note that y in (1.5b) is C1). As each of our gluing partitions of unity is an ((x, z)-dependent) convex combinations of values which are lower bounded by 0, it has the same property, too. Hence, we may for each x £ U0 conclude from (1.6):
g(x, y) > 0 for all y £ Y(x) ^^ g°(x,z) > 0 for all z £ Zj, j £ J. (1.7)
We note that, by definition, each of the new index sets Zj is a (q — r)-dimensional closed cube with 0q being one of its corner points. In particular, the sets Zj are feasible sets in the sense of finite optimization,
Zj = mf[m0,v°] = {0r} x [0,bj] x ■ ■ ■ x [0,bq-r] (1.8)
(1.6)
where «0 = (M01,... ,M0r), u0k(z) := zk (z £ IRq, k £ K = {1,... , r}, j £ J) and where the
components j of = (vj01,..., vjl2(q-r)) reflect the boundary points of q — r coordinate-wise intervals: v0 (z) = zi > 0, v0 (z) = —zi > —&i (£ £ {1,...,q — r}). Moreover, because of their forms these feasible sets Zj are compact and fulfill the condition LICQ. Let us shortly write g0 := (g°,... , g°°). Now, we finally arrived at the problem
Minimize f(x) on MSi[h,g0], where
[h, g0] := { x £ Rn | hi(x) = 0 (i £ I), g0(x,z) > 0 (z £ Zj, j £ J) }.
We remember that in our general case where only the Assumption By0 is made, J C IN needs not to be finite. Hence, g0 may be a sequence, e. g., g0 = (g^jgjv.
Theorem 1.2. Let the Assumption Bu0 on LICQ hold, due to a given open set U0 C IRn, MSi[h,g] H U0 = 0, and for the given generalized semi-infinite optimization problem P si(f, h,g,u, v). We assume that the defining functions of this problem are of class C1.
Then, in U0, Psi(f, h,g,u,v) can always equivalently be expressed as an ordinary semiinfinite optimization problem PSi (f, h,g0,«0,v0) with defining functions of class C1, too.
Moreover, for its feasible set MJ1 [h,g0] we have
M|1 [h, g0] hU0 = Msi[h, g] nU0 , (1.9a)
where the sets Zj = Mf[u^^] (j £ J) of inequality constraints are compact and fulfill LICQ.
If also Assumption Ay0 holds, then J is finite and in the special case U0 = U (Assumption A) (1.9a) means
MOi [h, g0] = Msi [h, g]. (1.9b)
Note that in the special case U0 = U the equation (1.9b) follows from both (1.9a) and the inclusions MSi[h,g] C U, MJi[h,g0] C U. Hereby the last inclusion is guaranteed by the construction of g0. We emphasize that in this case of Assumption A our Theorem 1.2 gives us a locally equivalent formulation of PSi(f, h,g,«,v) as an ordinary semi-infinite problem.
In analogy with (1.1), (1.2) we introduce for each x £ MJi[h, g0] the following "active sets":
Z0(x) := { z £ Zj | g0(x, z) = 0 } (j £ J), (1.10)
Z0O(x) := { (j,z) £ J x Rq | z £ Z0(x) }. (1.11)
Remark. We note that there is some ambiguity in the activity behaviour between our generalized semi-infinite problem and the ordinary semi-infinite problem. Namely, because of Psi(f, h, g0, u0, v0) being introduced by means of open coverings there may be a point x £ Msi[h, g] with an active index y £ Y0(x) corresponding to (more than one) different indices (jK,zjK) £ Z0O(x) (k £{1,...,k'}, K £ IV, K > 1, |{j 1,...,jK}| = k').
Finally, in the context of Kuhn — Tucker conditions we have to face a (special) disadvantage which comes from the definition of Zj (j £ J). Namely, for the definition of theses sets further inequalities are involved which do not represent one of the inequalities vi (£ £ L). This should give rise to care for some fineness of our open coverings.
In our second approach on expressing P i(f, h, g, v) (K = 0) as an ordinary semi-infinite problem in Section 2, that ambiguity and this disadvantage do not exist. Moreover, the second approach does not need the formalism of changing the coordinates (diffeomorphisms, inversions).
Let us make a last technical preparation for the next section. Therefore, we assume x G U0 to be a fixed feasible, maybe a local minimal point for (f,h,g,u,v). Then, aj = (xj,yj) (j G J) can always be chosen such that x = xj whenever x G Ua for some j G J.
Before we turn to optimality conditions for (f,h,g,u,v), i.e. for (f, h, g0, u0, v0) in U0, we make two remarks on the new formulation of our problem.
Remarks. (a) (An analytical remark.) Besides the critical comment from the remark above and in comparison with the (maybe infinitely many) index sets Y(x) (x G U0) the (locally in x, finitely many) index sets Zj (j G J) have the further advantage of being linearized. These sets were introduced more implicitly (inverse or implicit function theorem); however, there is some information on their sizes.
Indeed, for each of the sets Zj (j G J) we have a "controlling" parameter ftj > bg (£ G {1,... , q — r}) in order to estimate in the maximum norm || • the (coordinate-wise defined) size max{ bg | £ G {1,... , q — r}} of Zj. Therefore, we consider the proof of the inverse function theorem which is based on a suitable application of Banach's fixed point theorem (see, e.g., [1]). Then, we see in view of (1.3, a, b) that should for each (x,y) G lRn x lRq with
||(x,y) — (xj,yj)||^ < satisfy ||ln+q — DF"(x,y)||^ < ^. In detail, the last inequality means
O(q-r— pj)xn
C
- Iq
1
< -
" 2
where A =(..., DT uk (k G K),..., DTX v (£ G Lo(xj ,yj)), ... )T, B = (
DT ufc (k G K),...,
DTTvg(£G L0(xj, yj)), ... )T are evaluated at (x, y), and C =(..., ^T (n G {1,... , q—r — p}) ,...).
(b) Moreover, one can perform translations which transform the z-space Rq such that the finitely or countably many sets Zj (j G J) become pairwise disjoint, with maybe noncompact union Z. Then, one can glue together the transformed functions (j G J) to real-valued C1-
_R. In this way one could in U0 equivalently express (f, h,g,u,v),
0 „,0 „,0\
functions g° : !RnxJRq
po ( f h g0 u0 v0
inequality constraint function on the "upper stage inequality constraints.
(f , h, g0 , u0 , v0) as an ordinary semi-infinite problem (f , h, , u0 , v0) which has only one
but a maybe noncompact index set Z of
2. Optimality conditions, a second equivalent model
With the preparations of Section 1, we are able to generalize the results on necessary or sufficient optimality conditions from Kaiser and Krabs ([28]). In fact, due to the case where Y(x) is an interval [a(x),b(x)] C lR (q =1), in [28] the optimality conditions for a generalized semiinfinite problem could be traced back to optimality conditions for an ordinary semi-infinite problem. Now, we can extend this tracing back for cases of higher dimensional manifolds Y(x) with generalized boundary (cf. [16]). For ordinary semi-infinite optimization problems, optimality conditions have been worked out; cf. [13, 26]. While hereby in [26] a compactness assumption is made corresponding to our Assumption A, we may even use [26] for a noncompact fixed index set Y of inequality constraints. For this generalization we can replace the topology of uniform convergence on Y by the topology of uniform convergence on all the compact subsets of Y. For more information on Whitney's weak topologies C^ (k G lV0 := lV U {0}) we refer to [14, 17].
We need some more notation. Whenever we disregard the inequality constraints g(x,y) > 0 (y G Y(x)) then we denote the feasible set by M [h]. In [28], instead of M[h] arbitrary
nonempty sets X C IRn with convex tangent cone TXX ([26]) at a minimum x £ X are considered. However, a theorem of Whitney (cf. [3], Theorem 3.3) tells us that each closed set X C IRn can be represented by means of a C^-function h as X = M[h].
Now, for each x £ M[h] at which LICQ holds,
Tx M[h] := { £ £ IRn | Dhi(x) £ = 0 (i £ I) } (2.1)
stands for the tangent space at x on M[h]. If x £ MSi[h, g] this space contains the (linearized) tangent cone at x on MSi[h,g]:
Cx Msi[h,g] := { £ £ Tx M[h] | D^gj^z) £ > 0 ((j,z) £ Z0O(x)) }. (2.2)
Let x := x £ MSi[h, g] be a local minimum for PSi(f, h, g, u, v) where LICQ is fulfilled at x as an element of M[h] (in short: fulfilled at x £ M[h]). Hence, we refer to all x £ MSi[h, g]HU0 being in competition, where U0 is a suitable neighborhood of x. In the case ZS(x) = 0 then we learn from [26], Section 3.1, that it holds
Df (x) £ > 0 for all £ £ T£ M[h]. (2.3)
In the general case where ZO(x) = 0 is admitted, but where moreover the (relatively) open tangent cone
C| Msi[h, g] := { £ £ T* M[h] | Dj^) £ > 0 ((j,z) £ Z0s(x)) } (2.4a) is also = 0, then we conclude with [26], Theorem III.3.5 and Lemma III.3.15:
Cxi Msi[h,g] = C Msi[h,g], (2.4b)
Df (x) £ > 0 for all £ £ C Msi[h,g]. (2.5)
The notations in (2.2), (2.4a) are justified by the local representation (1.9a).
Before we evaluate the necessary optimality condition (2.5) in the following result, let us recall the by x parametrized new local coordinates around yj (j £ J). They are of the form 0X(y) = F^(x,y), with F := Fj given by (1.3a) for (x,y) := (xj,yj) (j £ J). By means of calculus an explicit representation of the Dx-derivative of the implicit function yj (x, z) = (0j)-1(z) (cf. (1.5b)) can be found:
Gj (x, z) := Dxyj(x,z) =
= — (DyFj(x, (0X)-1(z)))-1DxFFj(x, (0X)-1(z)) ((x,z) £ Uf x Cj). (2.6)
The derivatives on the right hand side are well known (cf. (1.3a), (1.4b)). In particular, the last q — r — pj components of DXFFj vanish. With these explanations for a further evaluation and with the definition of the problem Psi(f, h, g, u0, v0), now we state:
Theorem 2.1 (Theorem on a necessary optimality condition (N1)). Let x £
Msi[h, g] be a local minimum for the generalized semi-infinite optimization problem Psi(f, h, g, u,v), say: minimal on Msi[h, g] HU0 where U0 is some open neighborhood of x, and let Assumption BU0 hold. Moreover, let LICQ be fulfilled at x £ M[h] and the (relatively open linearized) tangent cone C| Msi[h,g] be nonempty.
Then, referring to canonical C1-smooth local changes of the coordinates of y, we have
Df (x) £ > 0 (2.7)
for all £ G Rn with
Dhi(x) £ = 0 for a// i G /,
(2.8)
(Dxg(x, (0X )-1(z )) + (0X)-1(z))Gj (x,z)) £ > 0 fora// (j,z) G Z0o(x). (2.9)
We note that, on the one hand, (2.9) can also be expressed in the original variable y, namely referring to active indices y = (0X)-1(z) G Y0(x).
On the other hand, (2.9) precisely means Dxg0(x,z) £ > 0 for all (j, z) G ZO(x). Then, we learn from [26], Theorem III.3.16 and what follows, in the presence of equality constraints, under the assumptions of Theorem 2.1 and Assumption Ayo, that the implication of the previous theorem can equivalently be expressed as the following Kuhn — Tucker condition. Namely, with R+ denoting the set of nonnegative real numbers and referring to the gradients we have
(KT)C
There is a finite subset Zfj(x) := { (jK, zK) |kG{1,...,k}} of ZO(x) and numbers Aj G R (i G I), G R+ ( k G {1,... , k}), where k G iV0, such that
Df (x) = Em=1 Aj Dhj(x) + EL1 ^ Dag°K(x,zK).
For our local minimum X this conclusion can also be attained by means of [13], Satz 3.1.14b) (see also [30]). Indeed, Assumption BMo, LICQ for X G M[h] together with C| Msi[h,g] = 0 precisely means the extended Mangasarian — Fromovitz constraint qualification EMFCQ ([13, 21, 32]) at X as an element of MS1[h,g0]. The finite version of EMFCQ, called MFcQ, will be introduced below.
In view of (1.3a), (2.6), (2.9), of the chain rule and of the choice of aj (j G J), (KT)0 can further be evaluated. In this way we get a Kuhn — Tucker theorem on our local minimum X of PS1(f, h, g, u, v), which can also be proved by means of the Kuhn — Tucker theorem of Jongen, Riickmann and Stein ([19]). Actually, provided that
for each (j, z) G Z0(X) the point z does not belong to a (boundary) face { z G Zj | Zk = bK—r }, { z G Zj | zCT = 0 } (k G{r + 1,...,q}, a G {r + + 1,...,q}) of Zj,
and, implicitly referring to the set Zfi(X) from (KT)0, then we have:
(KT)
There is a finite subset Y0fj(x) := { yK | yK = (0X ) (zK), k G {1,... , k} } of Yo(x) and numbers Aj G R, G R+, G R, G R+ (i G I, k G K, £ G L0(x,yK), k G {1,... ,k}), where k G IV0, such that
Df (k) = EI=1 Aj Dhj(x) + EK=1 ^ Dxg(x,yK) —
— E aK,fc DTufc(x,yK) — E ftK,gDJvg(x,yK).
fce{i,...,r}
eeL0(x,yK)
K6{1,...,ft}
Indeed, in virtue of our boundary condition, each active inequality constraint on z (in new variables) always represents an active (original) inequality constraint on y. Within the context of both our local linearizations (LICQ) and under the boundary condition from above, we also refer to [15] for Kuhn — Tucker conditions from finitely constrained optimization. In particular,
d
there we learn the nonnegativity of the "Lagrange multipliers" being —-g0K(x,zK) up
to the factor (w = € = £ L0(x,yK), k £ {1,...,K}; cf. (1.3a)). Moreover, that boundary condition on the geometry of ZO (x) is the content of Assumption F which will lateron (in the case K = 0) together with that nonnegativity be introduced. Let us already note that, by definition of Zj (j £ J), our condition (that assumption) can also easily (but nonlinearly) be expressed in the original coordinates of y.
Example 2.2 (cf. [28]). Let us turn to the problem formulation from [28] being motivated from a mechanical engineering model (see Section 1), where we still assume all defining data to be of class C1. Then we are in the special case where K = 0, Y(x) = [a(x),b(x)], say, in our formulation with (maybe) I = 0, for all x £ IRn and where a(x) < b(x) for all x £ U, U C IRn being a (possibly bounded) neighborhood of Msi[h,g]. Consequently, we can easily choose a new coordinate z by means of parametrizing the interval [a(x),b(x)]. Hence, we get z = 0X(y), with 0-1(z) = a(x) + z ■ (b(x) — a(x)) (w = 1) and the new index set Z = [0,1]. In this easy example, the special form of Y(x) guarantees LICQ and, hence, Assumption B to be fulfilled. However, the diffeomorphic representation of Y(x) = Mf[v(x, ■)], with v1(x,y) = —a(x)+y, v2(x, ■) = b(x)—y, performed in the general way of Section 1, would lead to a bit more notation. Hereby, we would refer to mappings (0X)-1(z) = z + a(x), (0X)-1(z) = — z + b(x).
Now, with our preferred new coordinate, (2.9) can be subdivided by our case study as follows:
(DXg(x, a(x) + z ■ (b(x) — a(x))) £ > 0 for all z £ Z0O(x) n (0,1)
(2.10a)
where the additional shift term from (2.9) vanishes, and
(DXg(x, a(x)) + Dyg(x, a(x))Da(x)) £ > 0 if 0 £ Z°(x) (DXg(x, b(x)) + Dyg(x, b(x))Db(x)) £ > 0 if 1 £ Z0O(x)
(2.10b) (2.10c)
In order to realize the sufficiency for (local or even global) optimality of the condition (2.7) for all £ £ IRn with (2.8), (2.9), we may suitably take over from [28] three more assumptions into our model. Hereby, we refer to a given point x £ Msi[h,g]. For the conditions involved in the Assumptions D and E we refer to [26].
Assumption C. The set M[h] is star-shaped with x being a star point, i. e.
x + A ■ (x — x) £ M[h] for all x £ M[h], A £ [0,1].
Assumption D. For all j £ J, z £ Zj, the functions g0(-,z) are quasi-concave on M[h] with respect to x; i. e., for each x £ M[h] the following implication holds:
g0 (x,z) > g0(x, z) Dxg0(x,z)(x — x) > 0. (2.11)
The implication (2.11) can be rewritten in the original data by writing the right hand side as the inequality from (2.9), whereby £ = x — x.
Assumption E. The function f is pseudo-convex on M[h] with respect to x; i. e., for each x £ M[h] the following implication holds:
Df (x) (x — x) > 0 f (x) > f (x). (2.12)
We also introduce the corresponding local versions Assumptions Cyo, Dyo, Eyo. Therefore, M[h] becomes replaced by M[h] RU0, where U0 C Rn is open.
For more information on quasi-concavity, pseudo-convexity and related conditions in the contexts of optimization (inf-compactness and solutions) and of differential geometry (implicit or parametrized surfaces), we refer to [5] and [6], respectively. The development of iteration procedures underlines the practical importance of our different assumptions ([32]).
Example 2.3 (cf. [28]). Assumption C is fulfilled if, e.g., M[h] is convex, and it implies arc-wise connectedness of M [h]. Without severe restrictions we may think M[h] C U.
The following example (due to [28]) for the Assumptions D, E continues Example 2.2 for the convex set M[h]: g(x,y) := g1(y) + g2(x) with g1, g2 being concave, f is convex, and a, b are affinely linear. These properties may hold globally, or, referring to a further (possibly bounded) open set V C IR for all x G U, y G V, where Y(x) C V (x G U). Note, that then g0(x,z) = g(x,0-1(z)) = g1(a(x) + z ■ (b(x) — a(x))) + g2(x) is concave in x such that g0(x,z) — g0(x,z) < Dxg0(x,z) (x — x), and f (x) — f (x) > Df (x) (x — x) for all x, x. We choose x := x.
In this example, the local versions of our assumptions can easily be given, too.
Theorem 2.4 on a sufficient optimality condition (S1)3. (a) Let the Assumption B hold for PSI(f, h,g,u,v). Moreover, for some given x G MSI[h,g] we assume LICQ for x G M[h] and the Assumptions C, D, E to be fulfilled.
If, moreover, the condition (2.7) holds for each £ G with (2.8), (2.9), then x is a global minimum for PSI(f, h, g, u, v).
(b) In (a) we replace the Assumptions B-E by the Assumptions Byo -Ejo where U0 is some open neighborhood of the point x. Then, under the further assumption of LICQ at x G M[h] and of (2.7) for all £ G with (2.8), (2.9), x turns out to be a local minimum for Psi(f, h,g,u,v).
Proof. It is enough to demonstrate the first part of the theorem because then the second part immediately follows. Indeed, a global minimum for f on MSI[h, g]RU0 is a local minimum for PSI(f, h,g,u,v). Now, for each given x G MSI[h,g] we have to show f (x) < f (x).
From our Assumption C we conclude x — x G Tx M[h]. This means the validity of (2.8) for £ := x — x. For each given (j, z) G ZO(x) we have g°(x, z) — g°(x, z) = g°(x, z) > 0. Then, for £ the inequality from (2.9) holds because of Assumption D.
Now, as in the case of (2.8), (2.9), the inequality (2.7) holds by assumption, it follows Df (x) (x — x) > 0. Finally, Assumption E allows us to state f (x) < f (x).
We introduce x-dependent subsets M(x) (x G IRn) of the feasible set MSI[h,g] by restrictively implying the feasible sets Y(x) of the "lower stage? (for a general introduction see [28]):
M(x) = {x G Rn | Y(x) C Y(x), 1 (213)
hj(x) = 0 (i G I), g(x,y) > 0 (y G Y(x)) }. J (2'13)
Lemma 2.5 on a necessary optimality condition (N2)4. Let a local or global minimum x of PSI(f, h,g,u,v) be given. Then, it holds x G M(x), and x is a local or global minimum for f on M(x) respectively.
Proof. We have the following representation of MSI[h,g]:
Msi[h, g] = Ux€M[h] M(x). (2.14)
3Cf. also [28], Satz 2.
4Cf. [28]
Namely, from (2.13) we conclude the implications
x £ MSI[h,g] ^^ x £ M(x) (x £ M[h]), (2.15a)
x £ M(x) for some x £ M[h] x £ MSI[h,g], (2.15b)
from which the inclusions "C, D" for (2.14) follow, respectively. In view of (2.15a), (2.14), a global minimum x := x belongs to M(x) and it minimizes f on M(x). For a local minimum x the corresponding assertions follow analogously, referring to some suitable neighborhood U0 of x.
In view of (2.14) and as far as the two conditions of being a local or global minimum are concerned, respectively, the reverse implication of Lemma 2.5 cannot hold in general.
For a given feasible (e.g., locally minimal) point x we can, after some preparations and assumptions, express the set M(x) as a feasible set:
M (x) = Msi ,x[h, gv], (2.16a)
namely in the way of defining which is subsequently described and proved5.
Now, firstly we replace each of our possible compactness assumptions by the following condition on global arc-wise connectedness (which can be interpreted as some "stiffness"):
Assumption Av (Connectedness). There is a neighborhood U of MSI[h,g] such that for all x1, x2 £ U it holds (Y(x1))0 H (Y(x2))° = 0, and each of the sets Y(x) (x £ U) is arc-wise connected.
Hereby, (Y(x))° denotes the interior of Y(x), relatively in M[u(x, ■)] (x £ U). Then, under our further Assumption B, referring to the open set U from Assumption Av, we have the representation (cf. [15])
(Y(x))° = { y £ M[u(x, ■)] | min^eL v*(x,y) > 0 } C Y(x) (x £ U). (2.17a)
As a local version we introduce, referring to those relative topologies again,
Assumption A^ (Connectedness, locally). Referring to some open set U0 C IRn, for each x1, x2 £ U0 the arc components K^, (Yj £ rj, j £ {1, 2}) of Y(x1), Y(x2) pairwise correspond to each other in such a way that r := r1 = r2 and, pairwise, y := 71 = y2 (Yj £ r, j £ {1, 2}), and, moreover, (K^)° H (K^)° = 0 for all corresponding arc components K1, K^ of Y(x1) and Y(x2), respectively (y £ r).
As for the purpose of our problem representation we may weaken the Assumptions B and Bu0, we introduce the following constraint qualification which is implied by LICQ (cf. [17, 29]): Definition 2.6. Let points x £ IRn and y £ Y(x) be given. We say that the Mangasarian — Fromovitz constraint qualification, in short: MFCQ, holds at y as an element of the feasible set Mf[u(x, ■), v(x, ■)] if the following conditions are fulfilled:
MF1. The vectors Dy(x,y) (k £ K) are linearly independent. MF2. There exists a vector £ £ IRn satisfying
Dyufc(x,y) £ = 0 (k £ K),
Dyv^(x,y) £ > 0 (€ £ L0(x,y)). Such a vector £ is called an MF-vector.
5If the reader is not so interested in the technical details, then he might after a short study of the assumptions which follow, of (2.22) and Definition 2.6, more directly turn to Theorem 2.7.
The Mangasarian — Fromovitz constraint qualification (MFCQ) is said to hold for the feasible set Mf[u(x, -),v(x, ■)] if MFCQ is fulfilled for all its elements y £ Y(x).
Assumption B , Buo (MFCQ, globally or locally). In the Assumptions B, Buo we replace LICQ by MFCQ.
Whenever we have K = 0, Assumption Auo being fulfilled and U0 being a sufficiently small neighborhood of x, then Assumption Buo, or Buo, already implies Assumption A^o. From [8,16] we can learn that under the Assumptions Auo, Buo, moreover, r is of finite cardinality.
Therefore, maybe we have to turn to a smaller neighborhood U1 C U0. If however, U0 (or, globally, U), considered as a bounded parameter set, is arc-wise connected, too, then such a shrinkening is not necessary.
Indeed, then our correspondences can be expressed by means of global homeomorphisms. For this implication which in fact does not need LICQ but MFCQ, we refer to [9, 21]. Moreover, for some related situation on the upper level, our implication of homeomorphical correspondence is stated in [32].
Finally, that special parametrical aspect on the presence of arc-wise connectedness is (for one parameter) given in [22].
In order firstly to give a global problem discussion and for the ease of exposition, we begin with making the Assumptions Av, B . Lateron, however, we shall see and discuss that locally the Assumptions A^o and Buo (or Buo) are appropriate for the goals of representation, optimality conditions and, in [32], convergence of iteration procedures.
Now, we introduce the following defining inequality constraint functions:
gv = (gv,g2v ,g3v), gv := g, gV = (02V-->02V.r)
where
gV
gVk(x,y):= (x,y) (k e K),
(g3,i>--->g3,sg^(x,y):= (x,y) e L)
(2.18)
and the corresponding index sets
Yv 1 := Y(x), Yv2'fc(x) := Y(x) (x £ IRn, k £ K), Yv:= Y(x) H Y/(x) (€ £ L). (2.19a) Hereby, for each € £ L the definition
Y/(x) := { y £ IRq | v/(x, y)
0 }
(2.19b)
means that Yv3,/ comes from Y(x) = Mf[u(x,-),v(x, ■)] by deleting v/(x, ■) as an inequality constraint, but by treating v/(x, ■) as an equality constraint. In this sense we may with the help of suitable defining functions also write
Y
v 3/
Mf [u
VI vv I
'] (* e L).
(2.19c)
In the special case where Assumption B is fulfilled we may state that all the feasible sets given in (2.19a) fulfill LICQ. Hence, they are manifolds with generalized boundaries (cf. [16]). However, if only Assumption B holds, then these sets need not all to fulfill MFCQ (namely, consider Yv3,/ (€ £ L)). But if they fulfill MFCQ, then they are manifolds with Lipschitzian boundaries (cf. [9, 21, 30]). The same statements can also locally be made referring to Buo and Buo, respectively.
Moreover, let us for each x £ U denote the corresponding "active" subsets by
YoV 1(x), YoV2,k(x)(k G K), YoV3/(x)(£ G L),
(2.20a)
where
YV2,k(x) := { y G Y(x) | ufc(x,y) = 0 } (k G K),
YV3,'(x) := { y G Y(x) | v,(x,y)= v,(x,y) = 0 } (£ G L),
(2.20b) (2.20c)
and, without misunderstandings using the index j also in the second approach, namely for the following enumerated union:
YV1 (x),
YV(x)
0
{ (j,y) I y G < YoV2,k(x), YV V(x),
if j = 1 if j = 2 if j = 3
for some k G K, £ G L }. (2.21)
Let the functions defining the index sets of (2.19a) systematically be ordered in a way being compatible with the indices from (2.21). Then, we comprise them by uv and vv, respectively; cf. (2.19c) for some finite subfamilies of defining component functions.
Now, we can give the following proof of (2.16a) and, hence, we may represent the minimization problem for f on M(x) by
Minimize f (x) on MSI,x [h,gv], where
Psx,x(f,h,gV,uV,vV)
[h,gV] := {x G IRn | h*(x) = 0 (i G I), gV(x,y) > 0 (y G YV1), glk(x,y) > 0 (y G YV2'fc(x), k G K),
(2.22)
glk(x,y) > 0 (y G Yv3'k, £ G L) }. J
Proof of (2.16a). The last r+s inequality constraints in (2.22) precisely reflect the inclusion Y(x) C Y(x). Hereby, the implication "of this equivalence is not hard to realize; let us turn to " Therefore, we note in the following indirect way, that the Assumptions Av,
B do not allow some y G Y(x) fulfilling these inequalities, say y G (Y(x))°, to lie outside of Y(x). This (relative) interior position can be guaranteed by means of a small inward shift of y. Hereby, we note that because of Assumption B on MFCQ, Y(x) is a manifold with Lipschitzian boundary (cf. [9]).
Otherwise, as in view of the definition of gfk (k G K) the constraints uk = 0 do not cause difficulties and as by Assumption Av there is also a point y0 G (Y(x))° R (Y(x))°, there exists an arc C in Y(x) connecting y with y0. Using the topological structure of Y(x) again, we may say: C C (Y(x))°. This arc has to meet the (relative) boundary SY(x) of Y(x) in M[u(x, ■)] at a point y*. Because of Assumption B we have the representation (cf. [9], Theorem A)
ÔY(x) = { y G M[u(x, ■)] | min^eL v^(x,y) = 0 } Ç Y(x).
(2.17b)
Hence, there is an index £* G L with v^» (X,y*) = 0, such that we conclude gj^* (X,y*) > 0,
i. e.
v^* (x,y*) < 0.
(2.23a)
From y* £ C C (Y(x))°, however, it follows
v* (x, y*) > 0, (2.23b)
in contradiction with (2.23a).
Our problem PSI,x(f, h,gv,uv, vv) is of generalized semi-infinite type such that under suitable assumptions (especially Assumption Buo) the results given in the Theorems 2.1 and 2.4 with their necessary and sufficient optimality conditions, namely N1. and S1., could easily be formulated. However, if there are no equality constraints on the lower stage of the original problem, i. e. if K = 0, then we have turned to an ordinary semi-infinite optimization problem, called
PS I, x(f,h,gv,vv).
Furthermore, then, its feasible set and its active index sets are denoted by
MSi,x[h,gv] := Msi,x[h,gv], Y?(x) := Y0v(x) (x £ MSx>4[h,gv]).
Making the Assumptions A^o (or K = 0, Auo and shrinkening U0, if necessary) and Buo (or Buo), our considerations remain valid. Hereby, our indirect argumentation refers to some pair (K}o, K^o) of corresponding components with y £ K^, y0 £ K^ H K^o. In particular, we
have
M(x) nU0 = Msi[h,gv] nU0, and = MSX,x[h,gv] nU0 if K = 0. (2.16b)
Now, let us dispense with our goal of global representation, but turn to the local model in the case K = 0.
With the help of the considerations at the beginning of this section and by means of Lemma 2.5, we may state now:
Theorem 2.7 on a necessary optimality condition (N3). Let x £ MSI[h,g] be a local minimum for the generalized semi-infinite optimization problem PSI(f, h,g,v) (K = 0), say: minimal on MSI[h, g] HU0 where U0 is some open neighborhood of x, and let the Assumptions A^o, or Auo, and B uo hold.
Moreover, we assume LICQ to be fulfilled at x £ M[h] and the (relatively open linearized) tangent cone C,* MSIx [h,gv] to be nonempty.
Then, referring to the ordinary semi-infinite optimization problem PSi x(f, h, gv, vv) we have
Df (x) £ > 0 (2.24)
for all £ £ IRn with
Dxh,(x) £ = 0 for all i £ I, (2.25)
Dxg/(x,y) £ > 0 for all (j, y) £ Y00(x). (2.26)
Hereby, (2.26) can equivalently be formulated as
Dxg(x,y) £ > 0 for all y £ Yj(x), (2.27a)
Dxv/(x, y) £ < 0 for all y £ Y(x), € £ L0(x,y). (2.27b)
Because of the absence of an intrinsic diffeomorphism, (2.26) does not reveal an (additional) shift-term as it is given in (2.9).
As we did for Theorem 2.1, under the assumptions of Theorem 2.7 and Assumption Auo we can again express the implication of the previous theorem as a Kuhn — Tucker condition.
Namely, this time on the one hand we have a condition (KT)V being analogous with (KT)° for K = 0. Hereby, gV takes over the part of g0. On the other hand we realize the following specification of (KT) for the case K = 0, where for simplicity we use some old notations again:
There are finite subsets Yfi(X) = {yK | k6{1,...,k}} C Y0(X), Yfi(X) = { y'« | k' G {1,..., k'} } C Y(X), and numbers A, G R, (KT) G R+, /«'/ G R+ (i G I,£ G Lo(x, y'K'), kG{1,...,k}, k' g{1,...,k'})
( ) | where K, k' G IN0, such that
Df (X) = E7=1 Ai Dhi(X) + E^¡=1 ^K Dxg(X,yK) - E 3«'/ Ax^(X,y'K).
teL0(x,y'rJ)
Provided that the extended Mangasarian — Fromovitz constraint qualification of Jongen, Ruckmann and Stein ([19]) holds, this Kuhn — Tucker result on a local minimum X of PSI(f, h,g,v) also follows from the Kuhn — Tucker theorem of [19]. Now, we may realize that for PSj x(f, h,gV,vV) neither the ambiguity nor the disadvantage exists, which were remarked for the problem Pjj(f,h, g0,u0,v0) of the first approach (Section 1).
However, now, of course K = 0 means a restriction of the generality. Moreover, in Theorem 2.7 there are more inequality constraints involved into the nonemptiness condition on the relatively open linearized tangent cone than in Theorem 2.1.
Example 2.8 (cf. [28]). Continuing Example 2.2 (and, hence, Example 2.3) where we are in the case K = 0, (2.27b) can be written as follows:
Da(X) £ > 0 and Db(X) £ < 0. (2.28)
Here, we are in the special case q =1 which will become important below.
Let us for Psj(f,h,g,v) compare our necessary optimality conditions, namely N1. with N3., in particular, (2.9) with (2.27a,b). Therefore, we make the Assumptions A^o, or Ayo, and Byo, where U0 is some open neighborhood of X. Using (2.6), here we may express (2.9) as
(
Dxg(x, (0X)-1(z)) £ + Dzg°(X,z)
\
-DxVeXx, (0X)-1(z)) (£ G Lo(X, (0X)-l(z)))
O
£ > 0
(q-pj ) Xn
((j,z) G Z0(X)).
(2.29)
For each given index (j,z) G Z0(X) we know that z is a local minimum for g0(X, •) on Zj Ç Rq. Recall that Zj is an axis-parallel cube of the form (see (1.8), where r may also be positive)
Zj = [0,bj] x • • • x [0,bq] Ç Hpj x Rq-pj, p = |Lo(X,yj)|.
If z is not lying on the relative boundary
d+Zj := { z G Zj | Z(j = j for some a G {1, or za = 0 for some a G {p + 1,... , q} }
,q},
of Zj in IHpj x IRq pJ then for the local minimum z it holds
d
— g0(x,z) > 0 (i £{1,...,q}), (2.30)
where for i £ { p + 1,... , q} we know that 0 is attained (= 0). Let us put the set
ZO+ (£) := { (j, z) £ Z0O(x) | z £ d+Zj }. (2.31)
Let us remark that this notation of a set can indeed, with r leading components 0 of the vectors z, immediately be generalized for the case r > 0 (K = 0).
Now, we make an assumption on a special fineness of the open coverings from Section 1: Assumption F (Technical fineness). It holds ZO+ (x) = ZO(x).
In order not to go too much into the technicalities, we only note that this assumption rules out both any kind of activity of originally free coordinates, and any negativity in (2.30) for indices z £ d+Zj, i £ L0(x, (0xi)-1(z)). Such a nonnegativity, coming say, from the orientation
within our local linearization, would mean that a Lagrange multiplier ;i = —— (x, z) (> 0)
_ _
on the lower stage (i.e., ;i = • ;i, j = jK, i = , in the sense of (1.3a), (KT); r = 0) becomes negative.
Now, if the inequalities (2.27a,b) hold then, in view of (2.30), the sum on the left hand side of (2.29) turns out to be sum of nonnegative numbers. Hence, (2.9) is satisfied. Hereby, it is even enough in (2.27b) to refer to Y(x) instead of Y(x).
Let us think about the reverse direction of this implication. If we are in the special case q =1, if, moreover, (2.9) holds and the following technical condition
(for each (j, z) £ ZO(x) and each active i £ L, i. e. (by LICQ){i} = L0 (x,y)
where y = (0X)-1(z), there is a multiplier x = Xz i < 0 solving the equation , .
(2.32)
Ar0(x,y) = xA^iO^y^
is fulfilled, then the reverse implication holds due to each given index (j, z) £ ZO(x). Indeed, whenever there is a corresponding active index i, from (2.32) we conclude the inequality
d
(xZ,i - dz^0(x,z^ Dxvi(x,y) £ > 0
from which the inequality Dxvi(;r,y) £ < 0 follows by means of (2.30) and of x^ i < 0. With the help of the previous inequality, of x^ i < 0 and (2.32), we realize the validity of Dxg(;r,y) £ > 0. Hence, we have concluded (2.27a), and (2.27b) with Y(x) being substituted by its subset Y (x). Of course, the same implication can be stated whenever there is no active index i.
We consider the last reflections in the context of our necessary optimality conditions. Lemma 2.9 (cf. also [28], Satz 4). Let a point x £ MSI[h,g] be given for the problem Psi(f, h,g,v) and the Assumptions BUo, F hold where U0 is some open neighborhood of x. Then, the following relations hold between the necessary optimality conditions (N1.,3.):
(a) (2.7) for all £ £ 1Rn with (2.8), (2.9) (2.7) for all £ £ 1Rn with (2.8), (2.27a,b). Here, in (2.27b) the set Y(x) may be replaced by its subset Y0(x).
(b) If, moreover, q = 1 and the condition (TC) holds, then we have:
(2.7) for all £ £ 1Rn with (2.8), (2.27a,b), where Y0(x) is replaced by Y(x) in (2.27b)
(2.7) for all £ £ 1Rn with (2.8), (2.9).
Looking once again at our (standard) Example 2.8, then we see for those (special) situations that the above substitution of Y(X) by Y°(X) and, hence, the subsequent problem modification, need not to be performed.
In order to formulate a further (general) sufficient optimality condition, let us modify the definition of PSI(f, h,gv,vv) a bit. Therefore, we introduce the following auxiliary feasible set and the corresponding auxiliary ordinary semi-infinite optimization problem:
- : the feasible set [h,gv] (K = 0; cf. (2.22)) up to replacing
Y(X) by Y°(X) in the definition of Yv3 ^ in (2.19a) P5I° x : Minimize f (x) on MJI° -
Based on the considerations at the beginning of this section we learn from Lemma 2.9(b), that in the case of q =1 and under (TC) the condition
(2.7) for all £ G lRn with (2.8), (2.9)
is a necessary optimality condition at X with respect to P^iV Then, however, we may by means of Theorem 2.4 state the following reversion of Lemma 2.5. This optimality criterion generalizes [28], Satz 4, Zusatz.
Corollary 2.10 (Theorem on a sufficient optimality condition (S2)). Let for the
generalized semi-infinite optimization problem PSI(f,h,g,v) (K = 0) a point X G lRn with X G M(X) be given, and the Assumptions B, C, D, E, F hold (C-F referring to X) or, locally on a neighborhood U° of X, the Assumptions BUo-Eyo, F, be made. Moreover, let q =1, the condition (TC) hold, X be a global or, say on U°, a local minimizer for f on Mjl°x, LICQ for X G M[h] and CJMsI?x = 0 be fulfilled.
Then, X G MSI[h,g] and X is a global or, with respect to the neighborhood U°, a local minimum for PSI(f, h, g, v), respectively.
3. Concluding remark
In this paper we were concerned with some foundations of generalized semi-infinite optimization. Hereby, the relations with other investigations in literature were taken into consideration. In the modelling and in the results the local-global aspect was worked out. Based on two different approaches we arrived at representations of our generalized semi-infinite optimization problem by means of ordinary ones, and at both necessary and sufficient optimality conditions of first order. The two approaches were discussed, and a continued example was given.
This present investigation also serves as a preparation of numerical concepts for the purpose of solving our generalized semi-infinite optimization problem. Hereby, we have to realize and to use some properties on the topological behavior of the feasible sets which are involved (see [32]).
Acknowledgment. The author expresses his gratitude to the Professors Dr. W. Krabs, Dr. K. G. Roesner, Dr. Yu Shokin and to Dipl.-Math. A. Reibold for their encouragement and support.
References
[1] BARNER M., FlüHR F. Analysis II. Walter de Gruyter, Berlin, New York, 1983.
[2] BÖHM E., MOST M. Konstruktiver Entwurf eines Schwingförderers. Students' working paper, Darmstadt University of Technology, Darmstadt, 1981.
[3] BRÖCKER Th., Lander L. Differential Germs and Catastrophes. London Math. Society Lecture Note Series, 17. Cambridge University Press, 1975.
[4] BROSOWSKI B. Parametric Semi-Infinite Optimization. Peter Lang Verlag, Frankfurt a.M., Bern, New York, 1982.
[5] CROUZEIX J. P., Ferland J., ZAlinescu C. a-convex sets and strong quasiconvexity. Preprint Université Blaise Pascal. Aubiere Cedex, 1996.
[6] GöRG A. Die Flachenschnittmethode zur Erzeugung GCn-stetiger Flächenübergänge. Eine Blending-Methode für implizite und parametrisierte Flachen: Thesis, Darmstadt University of Technology, Darmstadt, 1996.
[7] Graettinger T. J. Krogh B. H. The acceleration radius: a global performance measure for robotic manipulators, IEEE J. of Robotics and Automation, 4, 1988, 60-69.
[8] Guddat J., Jongen H. Th. Structural stability in nonlinear optimization. Optimization, 18, 1987, 617-631.
[9] Guddat J., Jongen H.Th., RUckmann J.-J. On stability and stationary points in nonlinear optimization. J. Austral. Math. Soc., Ser.B, 28, 1986, 36-56.
[10] Guerra VASQUEZ F., Guddat J., Jongen H.Th. Parametric Optimization: Singularities, Pathfollowing and Jumps, John Wiley, 1990.
[11] Hettich R., Jongen H.Th. Semi-infinite programming: conditions of optimality and applications. In: "Optimization Techniques", Part 2. Ed. J. Stoer. Lecture Notes in Control and Inform. Sci., 7, Springer Verlag, 1978, 1-11.
[12] Hettich R., Jongen H.Th., Stein O. On continuous deformations of semi-infinite optimization problems (1995). To appear in: "Approximation and Optimization in the Caribbean II". Eds. M. Florenzano, J. Guddat, M. Jimenez, H.Th. Jongen, Lopez Lagomasino G., F Marcellan. Peter Lang Verlag, Frankfurt a.M.
[13] Hettich R., Zencke P. Numerische Methoden der Approximation und semi-infiniten Optimierung. Teubner Studienbucher, Stuttgart, 1982.
[14] Hirsch M.W. Differential Topology. Springer Verlag, 1976.
[15] HOFFMANN A., Reinhard R. On reverse Chebychev approximation problems. Preprint Ilmenau University of Technology. Ilmenau, 1994, submitted for publication.
[16] Jongen H.Th., Jonker P., Twilt F. Nonlinear Optimization in lRn, I. Morse Theory, Chebychev Approximation, Peter Lang Verlag, Frankfurt a.M., Bern, New York, 1983.
[17] Jongen H.Th., Jonker P., Twilt F. Nonlinear Optimization in Rn, II. Transversality, Flows, Parametric Aspects. Peter Lang Verlag, Frankfurt a.M., Bern, New York, 1986.
[18] Jongen H. Th., Jonker P., Twilt F. Critical sets in parametric optimization. Mathem. Programming, 34, 1986, 333-353.
[19] Jongen H.Th., RUckmann J.-J., Stein O. Generalized semi-infinite optimization: a first order optimality condition and examples. Mathem. Programming, 83, 1998, 145-158.
[20] JONGEN H.Th., Stein O. On generic one-parametric semi-infinite optimization. SIAM J. Optim., 7, 1997, 1103-1137.
[21] JONGEN H.Th., Twilt F., Weber G.-W. Semi-infinite optimization: structure and stability of the feasible set. J. Optim. Theory Appl., 72, 1992, 529-552.
[22] JONGEN H.Th., Weber G.-W. On parametric nonlinear programming. Annals of Operations Research, 27, 1990, 253-284.
[23] JONGEN H.Th., Zwier G. On the local structure of the feasible set in semi-infinite optimization. In: "Parametric Optimization and Approximation". Eds. B. Brosowski and F. Deutsch. ISNM, 72, Birkhauser Verlag, Basel, Boston, Stuttgart, 1985, 185-202.
[24] JONGEN H.Th., Zwier G. On regular semi-infinite optimization. In: "Infinite Programming". Eds. E. J. Anderson and A. B. Philpott Lecture Notes in Econom. and Math. Systems, Springer Verlag, 259, 1985, 53-64.
[25] KAPLAN A., Tichatschke R. On a class of terminal variational problems, to appear in: "Parametric Optimization and Related Topics IV". Eds. J. Guddat et al. Enschede, the Netherlands, 1995.
[26] Krabs W. Optimization and Approximation. John Wiley, 1979.
[27] Krabs W. Einfuhrung in die lineare und nichtlineare Optimierung fur Ingenieure, Teubner, Leipzig, Stuttgart, 1983.
[28] KAISER C., Krabs W. Ein Problem der semi-infiniten Optimierung im Maschinenbau und seine Verallgemeinerung. Working paper, Darmstadt University of Technology, Darmstadt, 1986.
[29] Mangasarian O. L., FröMöVITZ S. The Fritz —John necessary optimality condition in the presence of equality and inequality constraints. J. Math. Anal. Appl., 17, 1967, 37-47.
[30] Weber G.-W. Charakterisierung struktureller Stabilität in der nichtlinearen Optimierung. Thesis. Aachen University of Technology. Appeared in: "Aachener Beitröge zur Mathematik", 5, Augustinus-Buchhandlung, Aachen, 1992.
[31] Weber G.-W. Optimal control theory: on the global structure and connections with optimization. Part 2. Preprint Darmstadt University of Technology. Darmstadt, 1998, submitted for publication.
[32] Weber G.-W. Generalized semi-infinite optimization: on iteration procedures and topological aspects. In: "Similarity Methods, International Workshop". Eds. B. Kroplin, St. Rudolph, St. Bruckner Institute of Statics and Dynamics of Aviation and Space Travel Constructions, Stuttgart, 1998, 281-309.
[33] WETTERLING W. W. E. Definitheitsbedingungen fur relative Extrema bei Optimierungsund Approximationsaufgaben. Numer. Math., 15, 1970, 122-136.
[34] ZlöKARNIK M. Dimensional Analysis and Scale-up in Chemical Engineering. Springer Verlag, Berlin, Heidelberg, New York, 1991.
Received for publication March 2, 1999