Document not found! Please try again

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS ...

14 downloads 672 Views 488KB Size Report
We seek “robust” solutions to such programs, that is, solutions which minimize the (worst-case) objective while satisfying the constraints for every possible value ...
c 1998 Society for Industrial and Applied Mathematics °

SIAM J. OPTIM. Vol. 9, No. 1, pp. 33–52

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS∗ ´ LEBRET† LAURENT EL GHAOUI† , FRANCOIS OUSTRY† , AND HERVE Abstract. In this paper we consider semidefinite programs (SDPs) whose data depend on some unknown but bounded perturbation parameters. We seek “robust” solutions to such programs, that is, solutions which minimize the (worst-case) objective while satisfying the constraints for every possible value of parameters within the given bounds. Assuming the data matrices are rational functions of the perturbation parameters, we show how to formulate sufficient conditions for a robust solution to exist as SDPs. When the perturbation is “full,” our conditions are necessary and sufficient. In this case, we provide sufficient conditions which guarantee that the robust solution is unique and continuous (H¨ older-stable) with respect to the unperturbed problem’s data. The approach can thus be used to regularize ill-conditioned SDPs. We illustrate our results with examples taken from linear programming, maximum norm minimization, polynomial interpolation, and integer programming. Key words. convex optimization, semidefinite programming, uncertainty, robustness, regularization AMS subject classifications. 93B35, 49M45, 90C31, 93B60 PII. S1052623496305717

Notation. For a matrix X, kXk denotes the largest singular value. If X is square, X  0 (resp., X  0) means X is symmetric and positive semidefinite (resp., definite). For X  0, X 1/2 denotes the symmetric square root of X. The notation Ip denotes the p × p identity matrix; the subscript is omitted when it can be inferred from context. 1. Introduction. A semidefinite program (SDP) consists of minimizing a linear objective under a linear matrix inequality (LMI) constraint; precisely, (1)

P0 :

minimize cT x subject to F (x) = F0 +

m X

xi Fi  0,

i=1

where c ∈ Rm − {0} and the symmetric matrices Fi = FiT ∈ Rn×n , i = 0, . . . , m, are given. SDPs are convex optimization problems and can be solved in polynomial time with, e.g., primal-dual interior-point methods [24, 35, 26, 19, 2]. SDPs include linear programs and convex quadratically constrained quadratic programs, and arise in a wide range of engineering applications; see, e.g., [12, 1, 35, 22]. In the SDP (1), the “data” consist of the objective vector c and the matrices F0 , . . . , Fm . In practice, these data are subject to uncertainty. An extensive body of work has concentrated on the sensitivity issue, in which the perturbations are assumed to be infinitesimal, and regularity of optimal values and solution(s), as functions of the data matrices, is studied. Recent references on sensitivity analysis include [30, 31, 10] for general nonlinear programs, [33] for semi-infinite programs, and [32] for semidefinite programs. When the perturbation affecting the data of the problem is not necessarily small, a sensitivity analysis is not sufficient. For general optimization problems, a whole field ∗ Received by the editors June 21, 1996; accepted for publication (in revised form) September 22, 1997; published electronically October 30, 1998. http://www.siam.org/journals/siopt/9-1/30571.html † Ecole Nationale Sup´ erieure de Techniques Avanc´ees, 32, Bd. Victor, 75739 Paris, France ([email protected], [email protected], [email protected]).

33

34

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

of study (stochastic programming) concentrates on the case where the perturbation is stochastic with known statistics. One object of this field is to study the impact of, say, a random objective on the distribution of optimal values (this problem is called the “distribution problem”). References relevant to this approach to the perturbation problem include [15, 9, 29]. We are not aware of special references for general SDPs with randomly perturbed data except for the last section of [30], some exercises in the course notes of [13], and section 2.6 in [23]. The main objective of this paper is to quantify the effect of unknown but bounded deterministic perturbation of problem data on solutions. In our framework, the perturbation is not necessarily small, and we seek a solution that is “robust,” that is, remains feasible despite the allowable, not necessarily small, perturbation. Our aim is to obtain (approximate) robust solutions via SDP. Links between regularity of solutions and robustness are, of course, expected. One of our side objectives is to clarify these links to some extent. This paper extends results given in [16] for the least-squares problem. The approach developed here can be viewed as a special case of stochastic programming in which the distribution of the perturbation is uniform. The ideas developed in this paper draw mainly from two sources: control theory, in which we have found the tools for robustness analysis [36, 17, 12] and some recent work on sensitivity analysis of optimization problems by Shapiro [31] and Bonnans, Cominetti, and Shapiro [10]. Shortly after completion of our manuscript, we became aware of the ongoing work of Ben-Tal and Nemirovski on the same subject. In [7], they apply similar ideas to a truss topology design problem and derive very efficient algorithms for solving the corresponding robustness problem. In [8], the general problem of tractability of obtaining a robust solution is studied, and “tractable counterparts” of a large class of uncertain SDPs are given. The case of robust linear programming, under quite general assumptions on the perturbation bounds, is studied in detail in [6]. Our paper can be seen as a complement of [8], giving ways to cope with (not necessarily) tractable robust SDPs by means of upper bounds. (In particular, our paper handles the case of nonlinear dependence of the data on the uncertainties.) A unified treatment, and more results, will appear in [4]. The paper is divided as follows. Our problem is defined in section 2. In section 3, we show how to compute upper bounds on our problem via SDP. We give special attention to the so-called full perturbations case, for which our results are nonconservative. In section 4, we examine sensitivity of the robust solutions in the full perturbations case. We provide conditions which guarantee that the robust solution is unique and a regular function of the data matrices. We then consider several interesting examples in section 5, such as robust linear programming, robust norm minimization, and error-in-variables SDPs. 2. Problem definition. 2.1. Robust SDPs. Let F(x, ∆) be a symmetric matrix-valued function of two variables x ∈ Rm , ∆ ∈ Rp×q . In the following, we consider x to be the decision variable, and we think of ∆ as a perturbation. We assume that ∆ is unknown but bounded. Precisely, we assume that ∆ is known to belong to a given linear subspace D of Rp×q , and in addition, k∆k ≤ ρ, where ρ ≥ 0 is given. In section 2.2, we will be more precise about the dependence of F on ∆.

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

35

We define the robust feasible set by   for every ∆ ∈ D, k∆k ≤ ρ, m (2) . Xρ = x ∈ R F(x, ∆) is well defined and F(x, ∆)  0 Now let c(∆) be a vector-valued rational function of the perturbation ∆, such that c(0) = c. We consider the following min-max problem: (3)

minimize

max

∆∈D, k∆k≤ρ

c(∆)T x subject to x ∈ Xρ .

From now on, we assume that the function c(∆) is independent of ∆ (in other words, the objective vector c is not subject to perturbation). This is done with no loss of generality: introduce a slack variable λ and define     x 0 ˜ x, ∆) = diag(F(x, ∆), λ − c(∆)T x). x ˜= , c˜ = , F(˜ λ 1 Problem (3) can be formulated as ˜ subject to x ˜ ∈ X˜ρ , minimize c˜T x ˜ where X˜ρ is the robust feasible set corresponding to the function F. In the following, we thus consider a problem of the form (4)

Pρ :

minimize cT x subject to x ∈ Xρ

and refer to it as a robust semidefinite problem (RSDP). In general, although Xρ is convex, Pρ is not a tractable problem—in particular, it is not an SDP. Our aim is to find a convex, inner approximation of Xρ that is described by a linear matrix inequality constraint. This inner approximation is then used to find an upper bound on the optimal value of Pρ by solving an SDP. In some cases, we can prove our results are nonconservative, that is, as in the so-called “full perturbation” case. We refer to the set X0 (resp., problem P0 , i.e., (1)) as the nominal feasible set (resp., nominal SDP). We shall assume that the nominal SDP is feasible, that is, X0 6= ∅. Of course, the robust feasible set Xρ may become empty for some ρ > 0; we return to this question later. 2.2. Linear-fractional representation. In this paper, we restrict our attention to functions F that are given by a “linear-fractional representation” (LFR): (5)

F(x, ∆) = F (x) + L∆(I − D∆)−1 R(x) + R(x)T (I − ∆T DT )−1 ∆T LT ,

where F (x) is defined in (1), R(·) is an affine mapping taking values in Rq×n , and L ∈ Rn×p and D ∈ Rq×p are given matrices. Together, the mappings F (·), R(·), the matrices L, D, the subspace D, and the scalar ρ constitute our perturbation model for the nominal SDP (1). The above class of models seems quite specialized. In fact, these models can be used in a wide variety of situations, for example, in the case where the (matrix) coefficients Fi in P0 are rational functions of the perturbation. The representation lemma, given below, and the examples of section 5 illustrate this point. A constructive proof of the following result can be found in [37]. Lemma 2.1. For any rational matrix function M : Rk → Rn×c , with no singularities at the origin, there exist nonnegative integers r1 , . . . , rk , and matrices M ∈ Rn×c ,

36

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

L ∈ Rn×N , R ∈ RN ×c , D ∈ RN ×N , with N = r1 + · · · + rk , such that M has the following linear-fractional representation (LFR): For all δ where M is defined, (6)

M(δ) = M + L∆ (I − D∆)

−1

R, where ∆ = diag (δ1 Ir1 , . . . , δk Irk ) .

Using the LFR lemma, we may devise LFR models for SDPs, where a perturbation vector δ ∈ Rk enters rationally in the coefficient matrices. The resulting set D of perturbation matrices ∆ is then a set of diagonal matrices of repeated elements, as in (6). Componentwise bounds on the vector δ, such as |δ|i ≤ ρ, i = 1, . . . , k, translate into a norm-bound k∆k ≤ ρ on the corresponding matrix ∆. 2.3. A special case. We distinguish a special case for which exact (nonconservative) results can be obtained via SDP. This is when F(x, ∆) is block diagonal, each block being independently perturbed—precisely, when  ¡ (7) F(x, ∆) = diag F1 (x, ∆1 ), . . . , FL (x, ∆L ) , where each Fi (x, ∆i ) assumes the form shown in section 2.2 for appropriate Li , Ri , Di , with ∆i ∈ Rpi ×pi , i = 1, . . . , L, and D consists of block-diagonal matrices of the form ª  D = ∆ = diag(∆1 , . . . , ∆L ), ∆i ∈ Rpi ×qi . We refer to this situation as the block-full perturbation case. When L = 1, we speak of the full perturbation case. As will be seen later, all results given for L = 1 can be generalized to the case L > 1. 3. Robust solutions for SDPs. Unless otherwise specified, we fix ρ > 0. 3.1. Full perturbations case. In this section, we consider the full perturbations case, that is, D = Rp×q . We assume kDk < ρ−1 , which is a necessary and sufficient condition for F(x, ∆) to be well defined for every x ∈ Rm and ∆ ∈ Rp×q , k∆k ≤ ρ. The following lemma is a simple corollary of a classic result on quadratic inequalities, referred to as the S-procedure [12]. Its proof is detailed in [16]. Lemma 3.1. Let F = F T , L, R, D be real matrices of appropriate size. We have det(I − D∆) 6= 0 and (8)

F + L∆(I − D∆)−1 R + RT (I − D∆)−T ∆T LT  0

for every ∆, k∆k ≤ 1, if and only if kDk < 1 and there exists a scalar τ such that   F − τ LLT RT − τ LDT (9)  0. R − τ DLT τ (I − DDT ) A direct application of the above lemma shows that, in the full perturbations case, the RSDP (4) is an SDP. Theorem 3.1. When D = Rp×q , the RSDP (4) and a corresponding solution x can be computed by solving the SDP in variables x, τ :   F (x) − τ LLT R(x)T − τ LDT (10)  0. minimize cT x subject to R(x) − τ DLT τ (ρ−2 I − DDT )

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

37

Special barrier functions adapted to a conic formulation of the problem can be devised and yield an interior-point algorithm that has the same complexity as the nominal problem; see [24]. We may define the maximum allowable perturbation level, which is the largest number ρmax such that Xρ 6= ∅ for every ρ, 0 ≤ ρ ≤ ρmax (note ρmax > 0 since X0 6= ∅). Computing ρmax is a (quasi-convex) generalized eigenvalue minimization problem [24, 11]:   F (x) − τ LLT R(x)T − τ LDT  0. minimize λ subject to R(x) − τ DLT τ (λI − DDT ) Remark. The above exact results are readily generalized to the block-full perturbation case (L > 1) as defined in section 2.2. 3.2. Structured case. We now turn to the general case (D is now an arbitrary linear subspace). In this section, we associate with D the following linear subspace: ∆  B = (S, T, G) ∈ Rp×p × Rq×q × Rq×p ª (11) S∆ = ∆T, G∆ = −∆T GT for every ∆ ∈ D . As shown in [16], a general instance of problem (4) is NP-hard. Therefore, we look for upper bounds on its optimal value. The following lemma is a generalization of Lemma 3.1 that traces back to [17]. Its proof is detailed in [16]. Lemma 3.2. Let F = F T , L, R, D be real matrices of appropriate size. Let D be a subspace of Rp×q , and denote by B the set of matrices associated with D as in (11). We have det(I − D∆) 6= 0 and F + L∆(I − D∆)−1 R + RT (I − D∆)−T ∆T LT  0

(12)

for every ∆ ∈ D, k∆k ≤ 1 if there exist a triple (S, T, G) ∈ B such that S  0, T  0, and   RT − LSDT + LG F − LSLT (13)  0. R − DSL − GLT T − GDT + DG − DSDT Using Lemma 3.2, we obtain the following result. Theorem 3.2. An upper bound on the RSDP (4) and a corresponding solution x can be computed by solving the SDP in variables x, S, T, G : 

inf cT x subject to (S, T, G) ∈ B, S  0, T  0,  R(x)T − LSDT + LG F (x) − LSLT  0. T −2 R(x) − DSL − GL ρ T − GDT + DG − DSDT

Note that when the perturbation is full, the variable G is zero and S, T are of the form τ Ip , τ Iq , resp., for some τ ≥ 0. We then recover the exact results of section 3.1. As before, we may define the maximum allowable perturbation level, which is the largest number ρmax such that Xρ 6= ∅ for every ρ, 0 ≤ ρ ≤ ρmax . Computing a lower bound on this number is a (quasi-convex) generalized eigenvalue minimization problem: (14)



inf λ subject to (S, T, G) ∈ B, S  0, T  0,  R(x)T − LSDT + LG F (x) − LSLT  0. T λT − GDT + DG − DSDT R(x) − DSL − GL

38

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

4. Uniqueness and regularity of robust solutions. In this section, we derive uniqueness and regularity results for the RSDP in the case of full perturbations. As before, we first take L = 1 (one block), that is, D = Rp×q . The results of this section remain valid in the general case L > 1 (several blocks). We fix ρ, 0 < ρ < ρmax . For simplicity of notation (and without loss of generality) we take ρ = 1 (and thus ρmax > 1). For well-posedness reasons, we must assume kDk < 1. We make the further assumption that D = 0 (in other words, F(x, ∆) is affine in ∆). In section 4.5, we show how the case D 6= 0 can be treated. For full perturbations and D = 0, the RSDP is the SDP   F (x) − τ LLT R(x)T (15)  0. minimize cT x subject to R(x) τI 4.1. Hypotheses. We assume that the SDP (15) (with D = 0) satisfies the following hypotheses: H1. The Slater condition holds, that is, the problem is strictly feasible. H2. The problem is inf-compact, meaning that any unbounded sequence (xk ) of feasible points (if any) produces an unbounded sequence of objectives. An equivalent condition is that the Slater condition holds for the dual problem [28, p. 317, Thm. 30.4]. H3. (a) The nullspace of the matrix λR0 +

m X

xi Ri

i=1

is independent of (λ, x) 6= (0, 0) and not equal to the whole space. (b) For every x,   LT is full column-rank. R(x) Hypotheses H1 and H2 ensure, in particular, the existence of optimal points for problem (15) and its dual. Hypotheses H3(a) and (b) are difficult to check in general, but sometimes can be easily tested in practical examples, as seen in section 5. We note that H3(a) implies that R(x) 6= 0 for every x. Hypothesis H1 is equivalent to Robinson’s condition [27], which can be expressed in terms of   F (x) − τ LLT R(x)T . F(x, τ ) = R(x) τI Robinson’s condition is stated in [27] as the existence of x0 ∈ Rm , τ0 ∈ R such that  ¡ + , 0 ∈ int F(x0 , τ0 ) + dF(x0 , τ0 )Rm+1 − Sn+q + is the set of positive semidefinite matrices where dF is the differential of F, and Sn+q of order n + q. The equivalence between H1 and Robinson’s assumption is not true, in general. Here, this equivalence stems from the fact that the problem is convex and + has a nonempty interior. that the cone Sn+q Remark. Hypothesis H1 holds if and only if it holds for the nominal problem (1) (recall our assumption ρmax > 1). Also, hypothesis H2 implies L 6= 0 (otherwise, we can let τ → ∞ without affecting the objective value). If H2 holds for the nominal problem and L 6= 0, then H2 holds for the RSDP (15).

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

39

4.2. An equivalent nonlinear program. Let xopt , τopt be optimal for (15). Hypothesis H3(a) ensures that any τ that is feasible for (15) is nonzero (otherwise, R(x) would be zero for some x). We thus have τopt > 0. We introduce some notation. For x ∈ Rm , Z ∈ Rn×n , τ > 0 and µ ∈ R, define     c x d= , y= , Y = diag(Z, µ), 0 τ 1 R(x)T R(x), G(y) = diag(G(y), τ − .99τopt ), τ L(y, Y ) = dT y − TrY G(y).

G(y) = F (x) − τ LLT −

Using Schur complements and τopt > 0, we obtain that problem (15) can be rewritten as (16)

minimize dT y subject to G(y)  0

and that yopt = [xTopt τopt ]T is optimal for (16). Our aim is first to prove that the so-called quadratic growth condition [10] holds at yopt for problem (16). Then, we will apply the results of [10] to obtain uniqueness and regularity theorems. 4.3. Checking the quadratic growth condition. Following [10], we say that the quadratic growth condition (QGC) holds at yopt if there exists a scalar α > 0 such that, for every feasible y, dT y ≥ dT yopt + αky − yopt k2 + o(ky − yopt k2 ). Roughly speaking, this condition guarantees that yopt is not on a facet on the boundary of the feasible set. Define the set of dual variables associated with yopt by   ∂G (yopt ) = di , i = 1, . . . , m + 1 . Y(yopt ) = Y = diag(Z, µ) Y  0, TrY ∂yi The following result is a direct consequence of a general result by Bonnans, Cominetti, and Shapiro [10]. Roughly speaking, this result states that, if an optimization problem satisfies Robinson’s condition and has an optimal point, and if a certain “curvature” condition is satisfied, then the QGC holds at that point. Theorem 4.1. With the notation above, if H1 and H2 hold, and if (17)

∃ Y ∈ Y(yopt ) such that ∇2yy L(yopt , Y ) > 0,

then problem (16) satisfies the QGC. The following theorem is proven in appendix A. Theorem 4.2. If H1–H3 hold, problem (15) satisfies the quadratic growth condition at every optimal point yopt . Consequently, there exists a unique solution to the SDP (15). Remark. Note that the QGC is satisfied independent of the objective vector. This means that the boundary of the feasible set is strictly convex (it contains no facets). 4.4. Regularity results. In problem (15), the data consist of the matrices L, and Fi , Ri , i = 0, . . . , m. We seek to examine the sensitivity of the problem with respect to small variations in Fi , Li , and Ri .

40

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

In this section, we consider matrices L(u), Ri (u), and Fi (u), i = 0, . . . , m that are functions of class C 1 of a (small) parameter vector u. Define F (x, u) = F0 (u) +

m X

xi Fi (u), R(x, u) = R0 (u) +

i=1

m X

xi Ri (u).

i=1

We denote by P(u) the corresponding problem (15), where F (·), R(·), and L are replaced by F (·, u), R(·, u), and L(u). We assume that F (·, 0) = F (·), R(·, 0) = R(·), and L(0) = L, so that P(0) is (15). We first note that, in the vicinity of u = 0, problem P(u) satisfies the hypotheses H1 and H2 if P(0) does. In this case, for every  > 0 we may define the set S (u) of -suboptimal points of P(u):  ª S (u) = x x is feasible for P(u) and cT x ≤ v(u) +  , where v(u) is the optimal value of P(u). Recall that, if P0 satisfies hypotheses H1 and H2, the optimal value v(u) is continuous, and even directionally differentiable, at u = 0 [32, Thm. 5.1]. With the QGC in force, and using [31, Thm. 4.1], we can give quite complete regularity results for the robust solutions. Theorem 4.3. If hypotheses H1–H3 hold for P(0), then for every  = O(u), there exists a γ > 0 and a neighborhood V of u = 0 such that for every u ∈ V and x ∈ S (u), we have (18)

kx − x(0)k ≤ γkuk1/2 .

When H1–H3 hold for P(0), the above theorem states that every (sufficiently) suboptimal solution to P(0) is H¨ older-stable (with coefficient 1/2). This is true, in particular, for any optimal solution of P(u) (that is, for  = 0). The fact that the theorem remains true for  > 0 guarantees regularity of numerical solutions to the RSDP. The main consequence is that even if the nominal SDP is ill conditioned (with respect to variations in the Fi ’s), the RSDP becomes well conditioned for every ρ > 0. Now assume ρ 6= 1. We seek to examine the behavior of problem (10) (with D = 0) when the uncertainty level ρ for 0 < ρ < ρmax varies. This is a special case of the problem examined above, with u = ρ, F (·, u) = F (·), R(·, u) = R(·), L(u) = ρL. Corollary 4.1. For every ρ, 0 < ρ < ρmax , the solution to (10) (with D = 0) is unique and satisfies the regularity results (written with u = ρ) of Theorem 4.3. Remark. The results of this section are all valid in the block-full perturbation case (L > 1), as defined in section 2.2. Of course, the conditions given in H3 should be understood blockwise. 4.5. Case D 6= 0. When D 6= 0, we can get back to the case D = 0 as follows. Recall that we have kDk < 1 in order to ensure that F(x, ∆) is defined everywhere on D. With this assumption, we can define, for x ∈ Rm and τ > 0, ˜ = L(I − DT D)−1/2 , L ˜ R(x) = (I − DDT )−1/2 R(x), F˜ (x) = F (x) − LDT (I − DDT )−1 R(x) − R(x)T (I − DDT )−1 LT , ˜ T R(x). ˜ ˜ ˜L ˜ T − 1 R(x) G(y) = F˜ (x) − τ L τ

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

41

Using Schur complements, we have, for every x and τ > 0,   F (x) − τ LLT R(x)T − τ LDT ˜  0. G(y)  0 if and only if R(x) − τ DLT τ (I − DDT ) ˜ R(·) ˜ if and only if it holds for L, R(·). We can then follow Hypothesis H3 holds for L, the steps detailed previously. Corollary 4.2. If the SDP (10) (with ρ = 1) satisfies H1–H3 and if kDk < 1, then the results of Theorem 4.3 hold. 5. Examples. 5.1. Unstructured perturbations. Assume F(x, ∆) = F (x) + ∆0 + ∆T0 +

m X

xi (∆i + ∆Ti ),

i=1

where ∆ = [∆0 . . . ∆m ]. This case corresponds to the representation in section 5, with   1 (19) L = I, R(x) = ⊗ I, D = 0, D = Rn×nm . x Using Lemma 3.2, we obtain that problem (4) is equivalent to the SDP     1 xT ⊗ ρI F (x) − τ I T T  (20) minimize c x subject to  0. 1 xT ⊗ ρI τI It turns out that we may get rid of the variable τ and get back to a convex problem of the same size as that of the unperturbed problem (1). To see this, first note that every feasible variable τ in problem (20) is strictly positive. Use Schur complements to rewrite the matrix inequality in (20) as   1 + kxk2 I, τ > 0. F (x)  τ + ρ2 τ Minimizing (over variable τ ) the scalar in the left-hand side of the above inequality shows that the RSDP (1) is equivalent to p (21) minimize cT x subject to F (x)  2ρ kxk2 + 1 · I. Formulation (21) is more advantageous than (20), since (21) involves a (convex) matrix inequality constraint of the same size as the original problem. As noted before, special barrier functions can be devised for this problem and yield an interior-point algorithm that has the same complexity as the original problem; see [24]. We note that, with the above choice for L, R, hypothesis H3 holds, which yields the following result. Theorem 5.1. The optimal value of the RSDP (20) can be computed by solving the convex problem (21). If (21) satisfies hypotheses H1 and H2, then for every ρ > 0, the solution is unique and satisfies the regularity conditions of Theorem 4.3. Remark. A sufficient condition for hypotheses H1 and H2 to hold for (21) is that they hold for the nominal problem. A more restrictive sufficient condition is that the nominal feasible set X0 is nonempty and bounded, and ρ < ρmax .

42

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

10 8 6 4 2 x

0

ρ=3

−2

ρ=1

−4

ρ = .4 ρ = .3 ρ = .2 ρ = .1 ρ=0

−6 −8 −10 −10

−8

−6

−4

−2

0

2

4

6

8

10

Fig. 1. Nominal and robust solutions of an SDP, with a 5 × 5 matrix F (x). Here ρmax = 5.

5.2. Robust center of a linear matrix inequality. In this section, for ρ > 0, we consider the SDP (21) and corresponding feasible (convex) set Xρ . We assume that X0 is nonempty and bounded, and that P0 is strictly feasible. Then, for every ρ, 0 < ρ < ρmax , Xρ is nonempty and bounded, and we can define a (unique) solution x(ρ) to the strictly convex problem (21). In view of Corollary 4.1, x(ρ) is a continuous function of ρ in ]0 ρmax [. Since (Xρ ) is a decreasing family of bounded sets, we may define (22)

x∗ =

lim x(ρ).

ρ→ρmax

Note that x∗ is independent on the objective vector c. Thus, to the matrix inequality F (x)  0, we may associate the robust center, defined by (22). The robust center has the property of being the most tolerant (with respect to unstructured perturbation) among the feasible points. An example is depicted in Fig. 1. The nominal feasible set X0 is described by a linear matrix inequality F (x)  0, where F is a 5 × 5 matrix. For various values of ρ, we seek to minimize x2 . The dashed lines correspond to the optimal objectives. As ρ increases, we observe that the robust feasible sets shrink. A crucial property of these robust sets is that they do not possess any straight faces, as observed in the figure. For ρ = ρmax ' 5, the robust feasible set is a singleton (in this example, x? = 0). When ρ = 0, the optimal solution is not unique and not continuous with respect to changes in the coefficient matrices Fi , i = 0, 1, 2 (although the optimal value is continuous). Since the sets Xρ become strictly convex as soon as ρ > 0, the resulting robust solutions are continuous. 5.3. Robust linear programs. An interesting special case arises with linear programming (LP). Consider the LP minimize cT x subject to aTi x ≥ bi , i = 1, . . . , L. Assume that the ai ’s and bi ’s are subject to unstructured perturbations. The perturbed value of [aTi bi ]T is [aTi bi ]T + δi , where kδi k2 ≤ ρ, i = 1, . . . , L. We seek a

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

43

robust solution to our problem, which is a special case of the block-full perturbation case referred to in section 2.2, with F given by (7), and Fi (x, ∆i ) = aTi x − bi + 2[xT − 1]∆i , i = 1, . . . , L, where ∆i = δi /2, and D is the set of diagonal, L × L matrices. The robust LP is q minimize cT x subject to aTi x − ρ kxk22 + 1 ≥ bi , i = 1, . . . , L. (23) The above program is readily written as an SDP by introducing slack variables. In fact, the robust LP is a second-order cone program (SOCP) for which efficient specialpurpose interior-point methods are available [24, 20, 23]. We note that hypothesis H3 holds blockwise. This yields the following result. Theorem 5.2. The optimal value of the robust LP can be computed by solving the convex problem (23). If the latter satisfies hypotheses H1 and H2, then for every ρ, 0 < ρ < ρmax , the solution is unique and satisfies the regularity conditions of Theorem 4.3. In [6], robust linear programming is studied in detail. For a wide class of perturbation models, where the data of every linear constraint vary in an ellipsoid, explicit robust solutions are constructed using convex SOCPs. Reference [23] also provides examples of robust linear programs solved via SOCP. 5.4. Robust eigenvalue minimization. Consider the case where the nominal problem consists of minimizing the largest eigenvalue of a matrix-valued function F (x): (24)

minimize λmax (F (x)).

When F (·) is subject to unstructured perturbations (as defined in section 5.1), the robust version of the problem is p minimize λ + 2ρ kxk2 + 1 subject to λI  F (x), or equivalently (25)

p minimize λmax (F (x)) + 2ρ kxk2 + 1.

Let ρ > 0. When written in an SDP form, the above problem satisfies the hypotheses H1–H3. From Theorem 4.3 we obtain that the solution is unique. If we consider that the data of the above problem consist of the matrices Fi , i = 0, . . . , m, then we know that the corresponding solution is H¨ older-stable (with coefficient 1/2). Since the problem is unconstrained, we can use a result of Shapiro [31, Thm. 3.1], by which we conclude that the solution is actually Lipschitz stable (inequality (18) holds with the exponent 1/2 replaced by 1). Finally, using the results from Attouch [3], we can show that computing the solution for ρ → 0 amounts to selecting the minimum norm solution among the solutions of the nominal problem. Theorem 5.3. The optimal value of the min-max problem (24) can be computed by solving the convex problem (25). For every ρ > 0, the solution is unique and is Lipschitz stable with respect to perturbations in Fi , i = 0, . . . , m. When ρ → 0, the solution converges to the minimum norm solution of the nominal problem (24). Remark. In this case, the RSDP is a regularized version of the nominal SDP, which belongs to the class of Tikhonov regularizations [34]. The regularization parameter

44

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

is 2ρ and is chosen according to some a priori information on uncertainty associated with the nominal problem’s data. Taking ρ close to zero can be used as a selection procedure for choosing a particular (minimum norm, regular) solution among the (not necessarily unique and/or regular) solutions of the nominal problem. Problem (25) is particularly suitable to the recent so-called U-Newton algorithms for solving problem (24). These methods, described in [21, 25], require that the Hessian of the “smooth part” (the so-called U-Hessian) of the objective of (24) be positive definite. For general mappings F (·), this property is not guaranteed. However, when looking at the robust problem (25), we see that the modified U-Hessian is guaranteed to be positive definite for every x and ρ > 0. This indicates that the RSDP approach may be used to devise robust algorithms for solving SDPs. 5.5. Robust SOCPs. An SOCP is a problem of the form (26)

minimize cT x subject to kCi x + di k ≤ eTi x + fi , i = 1, . . . , L,

where Ci ∈ Rni ×m , di ∈ Rni , ei ∈ Rm , fi ∈ R, i = 1, . . . , L. SOCPs can be formulated as SDPs, but special-purpose, more efficient algorithms can be devised for them; see [24, 5, 23]. Assuming that Ci , di , ei , fi are subject to linear—or even rational—uncertainty, we may formulate the corresponding RSDP as an SDP. This SDP can be written as an SOCP if the uncertainty is unstructured and affects each constraint independently. The subject of robust SOCPs is explored in [5] in detail. Explicit SDPs that yield robust counterparts to SOCPs nonconservatively are given for a wide class of uncertainty structures. In some cases, albeit not all, the robust counterpart is itself an SOCP. In [16, 14], the special case of least-squares problems with uncertainty in the data is studied at length. 5.6. Robust maximum norm minimization. Several engineering problems take the form (27)

minimize kH(x)k,

where H(x) = H0 +

m X

xi Hi ,

i=1

and Hi , i = 1, . . . , m are given p×q matrices. A well-known instance of this problem is the linear least-squares problem, with H(x) = Ax − b. Another example is a minimal norm extension problem for a Hankel operator studied in [18], in which H0 is a given (arbitrary) n × n Hankel matrix and Hi , i = 1, . . . , m is the n × n Hankel matrix associated with the polynomial 1/z i . In practice, the matrices Hi , i = 0, . . . , m are subject to perturbation, which motivates a study of the robust version of problem (27). Note that the least-squares case is extensively studied in [16]. Consider the full perturbation case, which occurs when each Hi is perturbed independently in a linear manner. Precisely, consider the matrix-valued function H(x, ∆) = H0 + ∆0 +

m X i=1

xi (Hi + ∆i ),

45

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

where ∆ = [∆0 . . . ∆m ]. For a given ρ > 0, we address the min-max problem (28)

min max kH(x, ∆)k. x

k∆k≤ρ

This problem is an RSDP for which we can get exact results using SDP. Indeed, for every x ∈ Rm and λ ≥ 0, the property max kH(x, ∆)k ≤ λ

k∆k≤ρ

is equivalent to F(x, λ, ∆)  0 for every ∆, k∆k ≤ ρ, where F(x, λ, ∆) = F (x, λ) + L∆R(x) + R(x)T ∆T LT , where

 F (x, λ) =

λI H(x)T

H(x) λI



 , L=

I 0



 , R(x) =

 0

1 x



 ⊗I

.

We thus write problem (28) as (4), where the perturbation set D is Rp×q . Applying Theorem 3.2, we obtain that the RSDP above is equivalent to the SDP (15) (with D = 0). As in section 5.1, we may get rid of the variable τ and obtain the equivalent formulation p (29) minimize kH(x)k + ρ kxk2 + 1. This RSDP satisfies hypotheses H1–H3, so we conclude that the results of Theorem 4.3 hold. As in robust eigenvalue minimization, we can get improved results using [31, section 3, Thm. 3.1]. Theorem 5.4. The optimal value of the min-max problem (28) can be computed by solving the convex problem (29). For every ρ > 0, the solution is unique and Lipschitz stable with respect to perturbations in Hi , i = 0, . . . , m. When ρ → 0, the solution converges to the minimum norm solution of the nominal problem (27). Remark. As for the RSDP arising in robust eigenvalue minimization, the robust minimum norm minimization problem is a regularized version of the nominal problem, which belongs to the class of Tikhonov regularizations. We now consider the general case where each matrix Hi in (27) is perturbed in a structured manner. To be specific, we concentrate on the minimal norm extension problem mentioned above. In practice, the matrix H0 is obtained from measurement and is thus subject to error. We may assume that this matrix is constructed from an n × 1 vector h0 (δ) = h0 + δ, where δ is unknown but bounded. The perturbed matrix H0 is of the form H0 (∆) = H0 + L∆R, where L, R are given matrices (the exact form of which we do not detail), and ∆ ∈ D = {diag(δ1 I1 , . . . , δn In ) | δi ∈ R, i = 1, . . . , n} . (In the above, each δi corresponds to the uncertainty associated with the ith antidiagonal of H0 .) We address the min-max problem (30) where ρ ≥ 0 is given.

minimize

max

∆∈D, k∆k≤ρ

kH(x) + L∆Rk ,

46

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

This problem is amenable to the robustness analysis technique. Defining ª ∆  S = diag(S1 , . . . , Sn ) Si ∈ Ri×i , i = 1, . . . , n , we obtain the following result. Theorem 5.5. An upper bound on the objective value of the min-max problem (30) can be computed by solving the SDP in variables x, S, G:   λI − LSLT H(x) LG λI ρR   0. H(x)T inf λ subject to S = S T , G = −GT ∈ S,  T T T G L ρR S 5.7. Polynomial interpolation. This example, taken from [16], can be formulated as an RSDP with rational dependence. For given integers n ≥ 1, k, we seek a polynomial of degree n − 1 p(t) = x1 + · · · + xn tn−1 that interpolates given points (ai , bi ), i = 1, . . . , k, that is, p(ai ) = bi , i = 1, . . . , k. If we assume that (ai , bi ) are known exactly, we obtain a known x, with a Vandermonde structure:      x1 1 a1 . . . an−1 1  .. .. ..  ,  ..  =   . . .   .   1

ak

...

an−1 k

xn

linear equation in the un b1 ..  , .  bn

which can be solved via standard least-squares techniques. Now assume that the interpolation points are not known exactly. For instance, we may assume that the bi ’s are known, while the ai ’s are parameter dependent: ai (δ) = ai + δi , i = 1, . . . , k, where the δi ’s are unknown but bounded: |δi | ≤ ρ, i = 1, . . . , k, where ρ ≥ 0 is given. We seek a robust interpolant, that is, a solution x that minimizes (31)

max kA(δ)x − bk,

kδk∞ ≤ρ

where



1  .. A(δ) =  . 1

 a1 (δ)n−1  .. . . ak (δ) . . . ak (δ)n−1 a1 (δ) .. .

...

The above problem is an RSDP. Indeed, it can be shown that     A(δ) b = A(0) b + L∆(I − D∆)−1 R, where L=



k M  i=1

1

ai

...

an−2 i

 R1 k k M M    Di , ∆ = δi In−1 , , R =  ...  , D = i=1 i=1 Rk

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

and, for each i, i = 1, . . . , k,  0  ..  . Ri =   .  ..  0 0  ..  . Di =   .  .. 0

. . . an−2 i .. .. . . .. .. . . ai ... ... 0 1 1 ai . . . ain−3 .. .. .. .. . . . . .. .. . . a 1 .. .

...

ai .. .

...

0

i

47

    ∈ R(n−1)×n ,       ∈ R(n−1)×(n−1) .  

1

(Note that det(I − D∆) 6= 0, since D is strictly upper triangular.) With the above notation, if we define F(x, ∆) as in section 5, then problem (31) can be formulated as the RSDP (4). With the approach described in this paper, one can compute an upper bound for the minimizing value of (31), and a corresponding suboptimal x. We do not know if the problem can be solved exactly in polynomial time, e.g., using SDP. We conjecture (as the reviewers of this paper did) that the answer is no. To motivate this claim, note that the solution to the problem of computing (31) for arbitrary affine functions A is already NP-hard [16]. 5.8. Error-in-variables RSDPs. In many SDPs that arise in engineering, the variable x represents physical parameters that can be implemented with finite absolute precision only. A typical example is integer programming, where integer solutions to (linear) programs are sought. These problems (which are equivalent to integer programming) are NP-hard. We now show that we may find upper bounds on these problems using robustness analysis. Consider, for instance, the problem of finding a solution x to the feasibility SDP (32)

find an integer vector x such that F (x)  0.

Now, consider the robust SDP (33)

λI ≤ F0 +

m X

maximize λ subject to (xi + ∆xi )Fi for every ∆x, k∆xk∞ ≤ 1/2.

i=1

Assume there exists a feasible pair (xfeas , λ) to the above problem, with λ ≥ 0. By construction, xfeas satisfies F (xfeas )  0. Furthermore, any vector x chosen such that kx − xfeas k∞ ≤ 1/2 is guaranteed to satisfy F (x)  0. This is true, in particular, for xint , the integer closest to xfeas . Thus, if we know a positive lower bound λ, and corresponding feasible point for problem (33), then we can compute an integer solution to our original problem. Finding a lower bound for (33) and an associated feasible point can be done as follows. For i, 1 ≤ i ≤ m, define Fi = 2Li Ri , where Li , RiT ∈ Rn×ri , ri = Rank Fi . Let   ) ( R1 m M    ..  ∆xi Iri , ∆xi ∈ R . L = L1 . . . Lm , R =  .  , and D = ∆ = Rm

i=1

48

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

Problem (33) can be formulated as (34)

maximize λ subject to λI ≤ F (x) + L∆R + RT ∆T LT for every ∆ ∈ D, k∆k ≤ 1/2.

The above is a special instance of the structured problem examined in section 3.2. Define ª ∆  S = diag(S1 , . . . , Sm ) Si ∈ Rri ×ri , i = 1, . . . , n . Theorem 5.6. A sufficient condition for an integer solution to the feasibility SDP (32) is that the constraints   F (x) − λI − LSLT (1/2)RT + LG T T 0 λ ≥ 0, S = S ∈ S, G = − G ∈ S, S (1/2)R − GLT are feasible. If xfeas is feasible for the above constraints, then any integer vector closest to xfeas (in the maximum norm sense) is feasible for (32). 6. Conclusions. In this paper, we considered semidefinite programs subject to uncertainty. Assuming the latter is unknown but bounded, we have provided sufficient conditions that guarantee “robust” solutions to exist via SDPs. Under some conditions (detailed in section 4), the robust solution is unique, and not surprisingly, stable. The method can then be used to regularize possibly ill-conditioned problems. For some perturbation structures (as for unstructured perturbations), the conditions are also necessary. That is, there is no conservatism induced by the method. The paper raises several open questions. In our description, we have considered the problem of making the primal SDP robust, thereby obtaining upper bounds on an SDP subject to uncertainty. The dual point of view should be very interesting. One might be interested in applying the approach to the dual problem instead. Does this lead to lower bounds on the perturbed problem? Also, in some cases, the RSDP approach leads to a unique (and stable) primal solution. May we obtain a unique solution to the dual problem by making the latter robust? (This would lead to analyticity of the primal solution; see [32].) As seen in section 5.2 the notion of robust center has, certainly, connections with the well-known analytic center; is the latter related to some robustness characterization? It seems that the RSDP method could be useful for deriving fast and robust (stable) algorithms for solving SDPs (see section 5.4), especially in connection with maximum eigenvalue minimization. Finally, as said in section 2.2 (Lemma 2.1), an SDP with coefficient matrices depending rationally on a perturbation vector can always be represented by an LFR model. Now, this LFR model is not unique. However, the results given here (for example, Theorem 3.2) hinge on a particular linear-fractional representation for a perturbed SDP. Hence we have the question: are our results independent of the chosen representation? We partially answer this difficult question in Appendix B. Appendix A. Proof of Theorem 4.2. We take the notation of section 4. Let Y = diag(Z, µ) be dual variables associated with (xopt , τopt ) that are optimal (their existence is guaranteed by H1 and H2). Then, Y ∈ Y(yopt ). Let us show that condition (17) holds for this choice of Y .

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

49

Since the problem satisfies H1 and H2, the complementarity conditions hold; therefore, the (optimal) dual variable µ associated with the constraint τ = τopt is zero. Consequently the variable Z is nonzero (recall c 6= 0). Using TrY

∂G (yopt ) = dm+1 = 0, ∂ym+1

we obtain 2 TrLLT Z = TrR(xopt )T R(xopt )Z. τopt

From τopt 6= 0 (implied by H3(a)), and using hypothesis H3(b), we can show that TrLLT Z = 0 and TrR(xopt )T R(xopt )Z = 0 are impossible for Z  0, Z 6= 0. This yields TrR(xopt )T R(xopt )Z > 0. Now let ξ ∈ Rm and λ ∈ R, and define Φ(ξ, λ) = d2 G(xopt , τopt )[(ξ, λ), (ξ, λ)],    T ξ ξ 2 ∇yy L(yopt , Y ) = −TrZΦ(ξ, λ). φ(ξ, λ) = λ λ We have, for every i, j, 1 ≤ i, j ≤ m, 1 ∂G 1 ∂G = Fi − (R(x)T Ri + RiT R(x)), = −LLT + 2 R(x)T R(x), ∂xi τ ∂τ τ 1 ∂2G 1 ∂2G = − (RjT Ri + RiT Rj ), = 2 (R(x)T Ri + RiT R(x)), ∂xi ∂xj τ ∂xi ∂τ τ 2 ∂2G = − 3 R(x)T R(x). ∂τ 2 τ By summation, we have λ2 T 3 R(xopt ) R(xopt ) τopt τopt  λ ¡ − 2 R(xopt )T (R(ξ) − R(0)) + (R(ξ) − R(0))T R(xopt ) τopt 1 1 λ2 RT R + (R(ξ) − R(0))T (R(ξ) − R(0)), = 3 R(xopt )T R(xopt ) + τopt τopt τopt

−Φ(ξ, λ) =

2

(R(ξ) − R(0))T (R(ξ) − R(0)) + 2

where R = R(ξ) − R(0) − φ(ξ, λ)

=

λ R(xopt ). We obtain finally, τopt 1

TrZRT R +

λ2 T 3 TrZR(xopt ) R(xopt ) τopt

τopt 1 TrZ(R(ξ) − R(0))T (R(ξ) − R(0)). + τopt

If φ(ξ, λ) = 0, then λ = 0 (from TrR(xopt )T R(xopt )Z > 0), and thus TrZRT R = 0 with R = R(ξ) − R(0). Since Z  0, this means that every column of Z 1/2 belongs to the nullspace of R(ξ) − R(0). Now assume ξ 6= 0. By hypothesis H3(a), we obtain that every column of Z 1/2 also belongs to the nullspace of R(xopt ), which contradicts

50

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

TrR(xopt )T R(xopt )Z > 0. We conclude that ∇2yy L is positive definite at (yopt , Y ). Thus, problem (15) satisfies the QCG. Appendix B. Invariance with respect to the LFR model. In this section, we show that the sufficient conditions obtained in this paper are, in some sense, independent of the LFR model used to describe the perturbation structure. Consider a function F taking values in the set of symmetric matrices having an LFR such as that in section 5. This function can be written in a more symmetric form, (35)

˜T , ˜ ∆(I ˜ − D∆) ˜ −1 L F(∆) = F + L

where we have dropped the dependence on x for convenience, and     T   0 ∆ ˜ = ˜ = 0 D ˜ = L RT , D , ∆ . L D 0 ∆T 0 ˜ for ˜ T =∆ It is easy to check that, if an invertible matrix Z satisfies the relation Z ∆Z every ∆ ∈ D, then ˜ ∆) ˜ −1 (LZ) ˜ T. ˜ ∆(I ˜ − (Z T DZ) F(∆) = F + (LZ) ˜ ˜ can be used to represent F In other words, the “scaled” triple (F, (LZ), (Z T DZ)) ˜ ˜ instead of F, L, D in (35). By spanning valid scaling matrices Z, we span a subset of all LFR models that describe F. A valid scaling matrix Z can be constructed as follows. Let (S, T, G) ∈ B, and define    −1/2 0 I G T . Z= 0 I 0 S 1/2 ˜ for every ∆ ∈ D. ˜ T =∆ It turns out that such a Z satisfies the relation Z ∆Z Using the above facts, we can show that if condition (13) is true for the original LFR model F, L, R, D with appropriate S, T, G, then it is also true for the scaled LFR ˜ obtained using any scaling matrix Z such as that above, for appropriate matrices S, ˜ ˜ G, T . That is, the condition is independent of the scaling Z. In this sense, the conditions we obtained are independent of the LFR used to represent the perturbation structure. Acknowledgments. This paper has benefitted from many stimulating discussions with several colleagues, including Aharon Ben-Tal, Stephen Boyd, Arkadii Nemirovski, Michael Overton, and particularly, Lieven Vandenberghe (who pointed out a mistake just before the final version was sent). Last but not least, the authors would like to thank the editor and reviewers for their very helpful comments and revisions. REFERENCES [1] F. Alizadeh, Interior point methods in semidefinite programming with applications to combinatorial optimization, SIAM J. Optim., 5 (1995), pp. 13–51. [2] F. Alizadeh, J.-P. A. Haeberly, and M. L. Overton, Primal-dual interior-point methods for semidefinite programming: Convergence rates, stability and numerical results, SIAM J. Optim., 8 (1998), pp. 746–768. [3] H. Attouch, Viscosity Solutions of Optimization Problems, Tech. Report 07, D´ept. des Sciences Math´ematiques, Universit´e Montpellier 2, France, 1994.

ROBUST SOLUTIONS TO UNCERTAIN SEMIDEFINITE PROGRAMS

51

[4] A. Ben-Tal, L. E. Ghaoui, and A. Nemirovski, Robust semidefinite programming, in Semidefinite Programming and Applications, to appear. [5] A. Ben-Tal and A. Nemirovski, Robust Convex Programming, Tech. Report 1/95, Optimization Laboratory, Faculty of Industrial Engineering and Management, Technion, Israel Institute of Technology, Technion City, Haifa 32000, Israel, 1995; Math. Oper. Res., to appear. [6] A. Ben-Tal and A. Nemirovski, Robust Solutions to Uncertain Linear Programs, Tech. Report 6/95, Optimization Laboratory, Faculty of Industrial Engineering and Management, Technion, Israel Institute of Technology, Technion City, Haifa 32000, Israel, 1995; Oper. Res. Lett., to appear. [7] A. Ben-Tal and A. Nemirovski, Robust truss topology design via semidefinite programming, SIAM J. Optim., 7 (1997), pp. 991–1016. [8] A. Ben-Tal and A. Nemirovski, Robust convex programming, IMA J. Numer. Anal., 1998, to appear. [9] B. Bereanu, Some Numerical Methods in Stochastic Programming Under Risk and Uncertainty, Academic Press, New York, 1980, Ch. 11, pp. 169–205. [10] J. F. Bonnans, R. Cominetti, and A. Shapiro, Sensitivity Analysis of Optimization Problems under Second Order Regular Constraints, Tech. Report 2989, INRIA, 1996; Math. Oper. Res., to appear. [11] S. Boyd and L. El Ghaoui, Method of centers for minimizing generalized eigenvalues, Linear Algebra and Appl., special issue on Numerical Linear Algebra Methods in Control, Signals and Systems, 188 (1993), pp. 63–111. [12] S. Boyd, L. El Ghaoui, E. Feron, and V. Balakrishnan, Linear Matrix Inequalities in System and Control Theory, Studies in Applied Mathematics series, SIAM, Philadelphia, PA, 1994. [13] S. P. Boyd and L. Vandenberghe, Introduction to convex optimization with engineering applications, lecture notes for ee392x, Stanford University, Stanford, CA, 1996. Available via anonymous ftp at isl.stanford.edu/pub/boyd. [14] S. Chandrasekaran, G. H. Golub, M. Gu, and A. H. Sayed, An efficient algorithm for a bounded errors-in-variables model, SIAM J. Matrix Anal. Appl., to appear. [15] M. Dempster, Stochastic Programming, Academic Press, New York, 1980. [16] L. El Ghaoui and H. Lebret, Robust solutions to least-squares problems with uncertain data, SIAM J. Matrix Anal. Appl., 18 (1997), pp. 1035–1064. [17] M. K. H. Fan, A. L. Tits, and J. C. Doyle, Robustness in the presence of mixed parametric uncertainty and unmodeled dynamics, IEEE Trans. Automat. Control, 36 (1991), pp. 25– 38. [18] J. W. Helton and H. J. Woerdeman, Symmetric Hankel operators: Minimal norm extensions and eigenstructures, Linear Algebra Appl., 185 (1993), pp. 1–19. [19] M. Kojima, M. Shida, and S. Shindoh, Global and Local Convergence of Predictor-CorrectorInterior-Point Algorithm for Semidefinite Programming, Tech. Report B-308, Dept. of Information Sciences, Tokyo Institute of Technology, 2-12-1 Oh-Okayama, Meguru-ku, Tokyo 152, Japan, 1995. [20] H. Lebret, Synth` ese de diagrammes de r´ eseaux d’antennes par optimisation convexe, Ph.D. thesis, Universit´e de Rennes I, Nov. 1994. ´chal, F. Oustry, and C. Sagastiza ´bal, The U -Lagrangian of a convex function, [21] C. Lemare Trans. Amer. Math. Soc., to appear. [22] A. S. Lewis and M. L. Overton, Eigenvalue optimization, Acta Numerica, 5 (1996), pp. 149– 190. [23] M. S. Lobo, L. Vandenberghe, S. Boyd, and H. Lebret, Second-order cone programming: Interior-point methods and engineering applications, Linear Algebra Appl., submitted. [24] Y. Nesterov and A. Nemirovskii, Interior Point Polynomial Methods in Convex Programming: Theory and Applications, Studies in Applied Mathematics series, SIAM, Philadelphia, PA, 1994. [25] F. Oustry, The U-Lagrangian of the maximum eigenvalue function, SIAM J. Optim., to appear. [26] F. A. Potra and R. Sheng, On homogeneous interior-point algorithms for semidefinite programming, Optim. Methods Softw., 9 (1998), pp. 161–184. [27] S. Robinson, Stability theorems for systems of inequalities, part II: Differentiable nonlinear systems, SIAM J. Numer. Anal., 13 (1976), pp. 497–513. [28] R. T. Rockafellar, Convex Analysis, 2nd ed., Princeton Univ. Press, Princeton, NJ, 1970.

52

L. EL GHAOUI, F. OUSTRY, AND H. LEBRET

[29] R. Rubinstein and A. Shapiro, Discrete Event Systems, John Wiley, New York, 1993. [30] A. Shapiro, Perturbation theory of nonlinear programs when the set of optimal solutions is not a singleton, Appl. Math. Optim., 18 (1988), pp. 215–229. [31] A. Shapiro, Perturbation analysis of optimization problems in Banach spaces, Numer. Funct. Anal. Optim., 13 (1992), pp. 97–116. [32] A. Shapiro, First and second order analysis of nonlinear semidefinite programs. Semidefinite programming, Math. Programming Ser. B, 77 (1997), pp. 301–320. [33] A. Shapiro and M. K. H. Fan, On eigenvalue optimization, SIAM J. Optim., 5 (1995), pp. 552– 568. [34] A. Tikhonov and V. Arsenin, Solutions of Ill-Posed Problems, John Wiley, New York, 1977. [35] L. Vandenberghe and S. Boyd, Semidefinite programming, SIAM Rev., 38 (1996), pp. 49–95. [36] V. A. Yakubovich, The solution of certain matrix inequalities in automatic control theory, Soviet Math. Dokl., 3 (1962), pp. 620–623. [37] K. Zhou, J. Doyle, and K. Glover, Robust and Optimal Control, Prentice–Hall, Upper Saddle River, NJ, 1995.

Suggest Documents