Exploiting sparsity and symmetry in semidefinite programming with ...

5 downloads 5436 Views 236KB Size Report
Nov 29, 2006 - SEMIDEFINITE PROGRAMMING WITH APPLICATIONS IN ... One can identify a polynomial p(x) of degree d with its sequence of coefficients p ...
EXPLOITING SPARSITY AND SYMMETRY IN SEMIDEFINITE PROGRAMMING WITH APPLICATIONS IN POLYNOMIAL OPTIMIZATION Kartik Sivaramakrishnan Department of Mathematics North Carolina State University [email protected] http://www4.ncsu.edu/∼kksivara

Symbolic Computation Seminar North Carolina State University November 29, 2006

•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Contents • Semidefinite Programming • Applications in polynomial optimization • Difficulties in solving large semidefinite programs • Semidefinite programs with a block angular structure • Exploiting sparsity to get block-angular SDPs • Exploiting symmetry to get block-angular SDPs • Decomposition framework for block angular SDPs • Conclusions -1•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Semidefinite programming (SDP) max C • X s.t. A(X) = b X0

(SDD) min bT y s.t. AT y − S = C S0

• Notation - X, S, C ∈ S n, b ∈ Rm - A • B = trace(AB) =

Pn

i,j=1

Aij Bij (Frobenius inner product)

- The operator A : S n → Rm and its adjoint AT : Rm → S n are   m A1 • X X A(X) =  ...  , A>y = y i Ai Am • X i=1 where Ai ∈ S n, i = 1, . . . , m -2•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Polynomial preliminaries 1 1. p(x) =

X n α∈Z+

pα xα1 1

. . . xαnn

=

X

pα xα is a multivariate polynomial.

n α∈Z+

2. The degree of the monomial xα =

xα1 1

. . . xαnn is |α| =

n X

αi.

i=1

3. The degree of the polynomial p(x) is the maximum degree of a monomial xα for which pα 6= 0. 4. Example: x21x32x3 − x42 + 2x1x3 − 1 is a polynomial in 3 variables with degree 6.  5. Let Snd := {α ∈ Z+n : |α| ≤ d}, |Snd| = n+d . d 6. One can identify a polynomial p(x) of degree d with its sequence of coefficients p = (pα )α∈Snd . 7. The set Rd [x1, . . . , xn] of polynomials of degree ≤ d with real coeffin+d cients is isomorphic to R( d ). -3•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Polynomial preliminaries 2 1. A polynomial p(x) is nonnegative if p(x) ≥ 0, ∀x ∈ Rn. P Let Pn,d = {pα : p(x) = α pα xα ≥ 0, ∀x ∈ Rn, deg(p(x)) ≤ d}. 2. A polynomial p(x) has a sum of squares (SOS) if X p(x) = pi(x)2. i

for some polynomials pi. It is clear that p(x) has even degree 2d, and each pi(x) has degree P ≤ d. Let Σn,d = {pα : p(x) = α pα xα has an SOS, deg(p(x)) ≤ d}. 3. Example: x21 + x22 + 2x1x2 + x63 = (x1 + x2)2 + (x33)2 is an SOS. 4. The sets Pn,d and Σn,d are both closed convex cones.

-4•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

When does a nonnegative polynomial have an SOS decomposition? 1. An SOS polynomial is nonnegative but not every nonnegative polynomial is an SOS. 2. Hilbert in 1888 showed that Pn,d = Σn,d, i.e., every nonnegative polynomial of degree d in n variables is an SOS when • n = 1 (d ≥ 2 and even) • d = 2 (all n) • n = 2, d = 4 In all other cases, Σn,d ⊂ Pn,d (see Chapter 6 in Squares by Rajwade). 3. The Motzkin polynomial x41x22 + x21x42 − 3x21x22 + 1 is nonnegative but not a SOS.

-5•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Applications of semidefinite programming in polynomial optimization 1. Given a polynomial p(x) in n variables and even degree 2d. pmin =

inf p(x)

x∈Rn

= sup{t : p(x) − t ≥ 0 ∀x ∈ Rn}. The constraint requires p(x) − t ∈ Pn,2d. 2. A constrained version of this problem is psos = sup{t : p(x) − t is SOS}. Since p(x) − t ∈ Σn,2d ⊆ Pn,2d, we have pmin ≥ psos. 3. The second problem can be reformulated as a semidefinite programming  problem instandard form where the size of the matrix size is n+d and d m = n+2d (Parrilo, Lasserre). 2d 4. One can generate a hierarchy of SDPs (of increasing size) whose objective values converge to the global minimum value of the polynomial. -12•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Difficulties with solving large SDPs 1. Interior point methods (IPMs) can solve a semidefinite program to a prescribed accuracy in a polynomial number of arithmetic operations (Nesterov and Nemirovskii). 2. Excellent IPM software for solving semidefinite programs are currently available (CSDP, SeDuMi, SDPT3, SDPA). 3. In each iteration of an IPM, one has to (a) Compute the Schur matrix of size m; this requires O(mn3 + m2n2) flops. (b) Factorize and store this dense matrix; a Cholesky factorization requires O(m3) flops. 4. Limited to SDPs with matrix size n = 5000 and m = 10, 000. 5. Kartik is currently exploring decomposition techniques to solve large scale block-angular semidefinite programs in a parallel and distributed computing environment. -13•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Semidefinite programming with block angular structure Pr max Pi=1 Ci • Xi r s.t. i=1 Ai (Xi ) = b Xi ∈ Ci, i = 1, . . . , r • Notes - Xi, Ci ∈ S ni , b ∈ Rm. - Ci = {Xi : Bi(Xi) = di, Xi  0} are compact semidefinite feasibility sets (described by LMIs). - The objective function and coupling constraints are block separable. - If the coupling constraints were absent, then we only have to solve r independent problems of the form max Ci • Xi Xi ∈Ci

-14•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Exploiting sparsity to get block-angular SDPs: 1 Consider the SDP max L • X s.t. Xii = 1, i = 1, . . . , n, X  0, where L is the adjacency matrix of the graph 3

2

3

2

Cl2 = (2,3,4,6) Cl2 Cl4 = (5,6,7) 5

6

4

6

5 Cl4

4

Cl3

Cl1 = (1,5,7) Cl1 1

7 GRAPH

1

Cl3 = (4,6,7) 7

CHORDAL EXTENSION OF GRAPH

-15•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Exploiting sparsity to get block-angular SDPs: 2 Using matrix completion, one can reformulate the earlier SDP as max

4 X

(Lk • X k )

k=1

s.t.

1 4 X23 − X13 = 0, 2 3 X34 − X12 = 0, 3 4 X23 − X23 = 0,

Xiik = 1, i = 1, . . . , |Ck |, k = 1, . . . , 4, X k  0, k = 1, . . . , 4, which is in block-angular form. -16•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Exploiting sparsity to get block-angular SDPs: 3 1. Construct the aggregate sparsity graph G = (V, E) from data matrices A0 = C and Ai, i = 1, . . . , m of SDP. We have V = {1, . . . , n} and E = {(i, j) ∈ V × V : ∃k ∈ {0, 1, . . . , m} s.t. (Ak )ij 6= 0} 2. Construct a minimal chordal extension G0 = (V, E 0) of G = (V, E). 3. Find maximal cliques Cli, i = 1, . . . , k in G0 = (V, E 0). We have X  0 ⇔ XCli,Cli  0, i = 1, . . . , k 4. Each block in the block-diagonal SDP corresponds to a maximal clique. 5. Introduce additional equality constraints for common nodes and edges in the cliques ; for instance if cliques Clk and Cll share a common edge {i, j} then Xijk = Xijl etc. 6. In some special cases, the resulting block-diagonal SDP has a blockangular form. -17•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Exploiting symmetry to get block-angular SDPs: 1 1. We are given a finite group G, elements g ∈ G and its representation ρ(g) : G → GL(Rn) (the set of real invertible matrices of size n). Also, assume ρ(g) to be an orthogonal matrix. 2. Consider the associated representation σ(g) of G given by σ(g)X = ρ(g)T Xρ(g) where X is a symmetric psd matrix of size n (congruence transformation on X). 3. We say that a SDP is invariant w.r.t group action σ(g) if (a) If X is feasible in the SDP then σ(g)X is also feasible ∀g ∈ G. (b) We have C • X = C • σ(g)X, ∀g ∈ G. 4. In this case, one can add the additional linear constraints X = σ(g)X, i.e., ρ(g)X − Xρ(g) = 0, ∀g ∈ G (fixed point subspace) to the SDP without changing its optimal value (Gatermann-Parrilo). -18•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Exploiting symmetry: 2 1. Consider the SDP min −X11 + X22 + X33 s.t. X12 = X13 trace(X) = 1 X  0. 2. SDP unchanged under a simultaneous permutation of the last two rows and columns of X, i.e., invariant under the permutation group S2. 3. Add the additional constraint X22 − X33 = 0 (fixed point subspace) to the SDP without changing the optimal objective value. 4. Apply a symmetry-adapted orthogonal transformation (using irreducible representations of S2) to get an equivalent block-angular SDP 1 1 min −X11 + X22 + X2 s.t. trace(X 1) + X 2 = 1 X1  0 X2 ≥ 0

containing a semidefinite block of size 2 and a linear block of size 1. •First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

The Lagrangian dual problem • The Lagrangian dual problem is miny θ(y) = bT y +

Pr

i=1

θi(y)

where θi(y) = max (Ci − ATi y) • Xi Xi ∈Ci

• Dual is an unconstrained convex but nonsmooth problem. Pr • Given y k , we have θ(y k ) = bT y k + i=1(Ci − ATi y k ) • Xik Pr and a subgradient g(y k ) = (b − i=1 Ai(Xik )) where Xik = argmax (C − ATi y k ) • Xi Xi ∈Ci

(these can be computed in parallel!) -20•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Solving the Lagrangian dual 1. Construct a model θk (y) an underestimate for θ(y) Pr θk (y) = bT y + i=1 maxk (Ci − ATi y) • Xij j=1,...,J (i)

from the function values and subgradient information. 2. The regularized master problem then is k

miny θk (y) + u2 ||y − xk ||2 where uk ≥ 0 and xk is our current center (best iterate so far!) 3. The dual to this quadratic program (much easier to solve!) is k

max −

1 || 2uk

J (i) r X X

k

Ai(Xij )λji − b||2 +

J (i) r X X

((Ci − ATi xk ) • Xij )λji

i=1 j=1

i=1 j=1

J k (i)

s.t.

X

λji = 1, i = 1, . . . , r

j=1

λji ≥ 0, i = 1, . . . , r, j = 1, . . . , J k (i). -21•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Kartik’s approach in a nutshell (a) Solve large scale structured semidefinite programs (SDP) arising in polynomial optimization. Exploit the sparsity and the symmetry in these SDPs (arising from the underlying polynomial!) (b) The technique is to iteratively solve an SDP between a mixed conic master problem over linear and smaller semidefinite cones; and distributed subproblems (smaller SDPs) in a parallel high performance computing environment. (c) Improve the scalability of interior point methods (IPMs) by applying them instead on the smaller master problem, and subproblems (which are solved in parallel!) This is our conic interior point decomposition scheme. -22•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

theta(y) theta^3(y)

y

1

y

3

y

4

y

2

Figure 1: Solving the Lagrangian dual problem •First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

MAX sum(i=1)^r trace(C_iX_i) COUPLING CONSTRAINTS

sum(i=1)^r A_i(X_i) = b X_i in C_i, i=1,...,r

SUBPROBLEMS theta_1(y) =

MASTER PROBLEM MAX theta(y)

PROPOSALS X_i (PRIVATE VARIABLES)

MAX trace((C_1−A_i^Ty)X_1) s.t. X_1 in C_1

SUBSYSTEM 1

where

SUBSYSTEM r

theta(y) = b^Ty +

PRICES y

sum(i=1)^r theta_i(y)

theta_r(y) =

(SHARED VARIABLES)

MAX trace((C_r−A_r^Ty)X_r) s.t. X_r in C_r

C_i are convex sets defined by LMIs

Figure 2: Decomposition by prices •First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Thank you for your attention!. Questions, Comments, Suggestions ?

The slides from this talk are available online at

http://www4.ncsu.edu/∼kksivara/ publications/kartik-symb.pdf My publications and talks can be downloaded from

http://www4.ncsu.edu/∼kksivara/ publications

-25•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Bibliography: 1 1. K. Sivaramakrishnan, A parallel conic interior point decomposition for block angular semidefinite programs, submitted. http://www4.ncsu.edu/∼kksivara/publications/parallel-conic-blockangular.pdf 2. K. Sivaramakrishnan, G. Plaza, and T. Terlaky, A conic interior point decomposition approach for large scale semidefinite programming, submitted. http://www4.ncsu.edu/∼kksivara/publications/conic-ipm-decomposition.pdf ´ski, Nonlinear Optimization, Princeton University Press, 2006. 3. A. Ruszczyn 4. M. Fukuda, M. Kojima, K. Murota, and K. Nakata, Exploiting sparsity in semidefinite programming via matrix completion I: General framework, SIAM Journal on Optimization, 11(2000), pp. 647-674. 6. K. Fujisawa, M. Fukuda, and K. Nakata, Preprocessing sparse semidefinite programs via matrix completion, Optimization Methods and Software, 21(2006), pp. 17-39. 7. P.A. Parrilo, Semidefinite programming relaxations for semialgebraic problems, Mathematical Programming, 96(2003), pp. 293-320. 8. K. Gatermann and P.A. Parrilo, Symmetry groups, semidefinite programs, and sums of squares, Journal of Pure and Applied Algebra 192(2004), pp. 95-128.

-26-

•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Bibliography: 3 9. P.A. Parrilo, Exploiting algebraic structure in sums of squares programs, In Positive Polynomials in Control, LNCSI 312, Springer, 2005, pp. 181-194. 10. E. de Klerk, Exploiting algebraic symmetry in semidefinite programming: a review, Talk at Delft, the Netherlands, October, 2005. http://www4.ncsu.edu/∼kksivara/etienne-symmetry.pdf 11. H. Waki, S. Kim, M. Kojima, and M. Muramatsu, Sums of squares and semidefinite programming relaxations for polynomial problems with structured sparsity. http://www.optimization-online.org/DB HTML/2004/10/988.html 12. J.B. Lasserre, Global optimization with polynomials and the problem of moments, SIAM Journal on Optimization, 11(2001), pp. 796-817. 13. A. Rajwade, Squares, Volume 171 of London Math Society Lecture Note Series, Cambridge University Press, 1993. 14. S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge University Press 2004. 15. K. Sivaramakrishnan, Solving polynomial optimization problems via real algebraic geometry and semidefinite programming, Talk in the Symbolic Computation seminar, NCSU, January 2006. http://www4.ncsu.edu/∼kksivara/publications/ncsu-symb.pdf.

-27-

•First •Prev •Next •Last •Go Back •Full Screen •Close •Quit

Suggest Documents