Abstract
Clique tree conversion solves large-scale semidefinite programs by splitting an \(n\times n\) matrix variable into up to n smaller matrix variables, each representing a principal submatrix of up to \(\omega \times \omega \). Its fundamental weakness is the need to introduce overlap constraints that enforce agreement between different matrix variables, because these can result in dense coupling. In this paper, we show that by dualizing the clique tree conversion, the coupling due to the overlap constraints is guaranteed to be sparse over dense blocks, with a block sparsity pattern that coincides with the adjacency matrix of a tree. We consider two classes of semidefinite programs with favorable sparsity patterns that encompass the MAXCUT and MAX k-CUT relaxations, the Lovasz Theta problem, and the AC optimal power flow relaxation. Assuming that \(\omega \ll n\), we prove that the per-iteration cost of an interior-point method is linear O(n) time and memory, so an \(\epsilon \)-accurate and \(\epsilon \)-feasible iterate is obtained after \(O(\sqrt{n}\log (1/\epsilon ))\) iterations in near-linear \(O(n^{1.5}\log (1/\epsilon ))\) time. We confirm our theoretical insights with numerical results on semidefinite programs as large as \(n=13{,}659\).




Similar content being viewed by others
Notes
Throughout this paper, we denote the (i, j)-th element of the matrix X as X[i, j], and the submatrix of X formed by the rows in I and columns in J as X[I, J].
The symmetric matrices \(A_{1},A_{2},\dots ,A_{m}\) share an aggregate sparsity pattern E that contains at most \(\omega n\) nonzero elements (in the lower-triangular part). The set of symmetric matrices with sparsity pattern E is a linear subspace of \({\mathbb {R}}^{n\times n}\), with dimension at most \(\omega n\). Therefore, the number of linearly independent \(A_{1},A_{2},\dots ,A_{m}\) is at most \(\omega n\).
To keep our derivations simple, we perform the rank-1 update using the Sherman–Morrison—Woodbury (SMW) formula. In practice, the product-form Cholesky Factor (PFCF) formula of Goldfarb and Scheinberg [63] is more numerically stable and more widely used [59, 61]. Our complexity results remain valid in either cases because the PFCF is a constant factor of approximately two times slower than the SWM [63].
Since T is connected, we can always find a connected subset \(W'\) satisfying \(W\subseteq W'\subseteq V(T)\) and replace W by \(W'\).
References
Lovász, L.: On the Shannon capacity of a graph. IEEE Trans. Inf. Theory 25(1), 1–7 (1979)
Goemans, M.X., Williamson, D.P.: Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. J. ACM 42(6), 1115–1145 (1995)
Sherali, H.D., Adams, W.P.: A hierarchy of relaxations between the continuous and convex hull representations for zero-one programming problems. SIAM J. Discrete Math. 3(3), 411–430 (1990)
Lovász, L., Schrijver, A.: Cones of matrices and set-functions and 0–1 optimization. SIAM J. Optim. 1(2), 166–190 (1991)
Lasserre, J.B.: An explicit exact SDP relaxation for nonlinear 0-1 programs. In: International Conference on Integer Programming and Combinatorial Optimization, pp. 293–303, Springer, Berlin (2001)
Laurent, M.: A comparison of the Sherali-Adams, Lovász-Schrijver, and Lasserre relaxations for 0–1 programming. Math. Oper. Res. 28(3), 470–496 (2003)
Lasserre, J.B.: Global optimization with polynomials and the problem of moments. SIAM J. Optim. 11(3), 796–817 (2001)
Parrilo, P.A.: Structured semidefinite programs and semialgebraic geometry methods in robustness and optimization. Ph.D. thesis, California Institute of Technology (2000)
Nesterov, Y.: Introductory Lectures on Convex Optimization: A Basic Course, vol. 87. Springer, Berlin (2013)
Fukuda, M., Kojima, M., Murota, K., Nakata, K.: Exploiting sparsity in semidefinite programming via matrix completion I: general framework. SIAM J. Optim. 11(3), 647–674 (2001)
Molzahn, D.K., Holzer, J.T., Lesieutre, B.C., DeMarco, C.L.: Implementation of a large-scale optimal power flow solver based on semidefinite programming. IEEE Trans. Power Syst. 28(4), 3987–3998 (2013)
Madani, R., Sojoudi, S., Lavaei, J.: Convex relaxation for optimal power flow problem: Mesh networks. IEEE Trans. Power Syst. 30(1), 199–211 (2015)
Madani, R., Ashraphijuo, M., Lavaei, J.: Promises of conic relaxation for contingency-constrained optimal power flow problem. IEEE Trans. Power Syst. 31(2), 1297–1307 (2016)
Eltved, A., Dahl, J., Andersen, M.S.: On the robustness and scalability of semidefinite relaxation for optimal power flow problems. Optim. Eng. 21, 375–392 (2020). https://doi.org/10.1007/s11081-019-09427-4
Vandenberghe, L., Andersen, M.S., et al.: Chordal graphs and semidefinite optimization. Found. Trends Optim. 1(4), 241–433 (2015)
Sun, Y., Andersen, M.S., Vandenberghe, L.: Decomposition in conic optimization with partially separable structure. SIAM J. Optim. 24(2), 873–897 (2014)
Andersen, M.S., Hansson, A., Vandenberghe, L.: Reduced-complexity semidefinite relaxations of optimal power flow problems. IEEE Trans. Power Syst. 29(4), 1855–1863 (2014)
Löfberg, J.: Dualize it: software for automatic primal and dual conversions of conic programs. Optim. Methods Softw. 24(3), 313–325 (2009)
Griewank, A., Toint, P.L.: Partitioned variable metric updates for large structured optimization problems. Numerische Mathematik 39(1), 119–137 (1982)
Andersen, M., Dahl, J., Vandenberghe, L.: CVXOPT: A Python package for convex optimization. abel.ee.ucla.edu/cvxopt. (2013)
Löfberg, J.: Yalmip: A toolbox for modeling and optimization in matlab. In: Proceedings of the CACSD Conference, 3. Taipei, Taiwan (2004)
Fujisawa, K., Kim, S., Kojima, M., Okamoto, Y., Yamashita, M.: User’s manual for SparseCoLO: Conversion methods for sparse conic-form linear optimization problems. Technical report, Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, (2009). Research Report B-453
Kim, S., Kojima, M., Mevissen, M., Yamashita, M.: Exploiting sparsity in linear and nonlinear matrix inequalities via positive semidefinite matrix completion. Math. Program. 129(1), 33–68 (2011)
Andersen, M.S.: Opfsdr v0.2.3 (2018)
Madani, R., Kalbat, A., Lavaei, J.: ADMM for sparse semidefinite programming with applications to optimal power flow problem. In: IEEE 54th Annual Conference on Decision and Control (CDC), pp. 5932–5939. IEEE (2015)
Zheng, Y., Fantuzzi, G., Papachristodoulou, A., Goulart, P., Wynn, A.: Chordal decomposition in operator-splitting methods for sparse semidefinite programs. Math. Program. 180, 489–532 (2020). https://doi.org/10.1007/s10107-019-01366-3
Annergren, M., Pakazad, S.K., Hansson, A., Wahlberg, B.: A distributed primal-dual interior-point method for loosely coupled problems using ADMM. arXiv preprint arXiv:1406.2192 (2014)
Khoshfetrat Pakazad, S., Hansson, A., Andersen, M.S.: Distributed interior-point method for loosely coupled problems. IFAC Proc. Volumes 47(3), 9587–9592 (2014)
Zhang, R.Y., White, J.K.: Gmres-accelerated ADMM for quadratic objectives. SIAM J. Optim. 28(4), 3025–3056 (2018)
Andersen, M.S., Dahl, J., Vandenberghe, L.: Implementation of nonsymmetric interior-point methods for linear optimization over sparse matrix cones. Math. Program. Comput. 2(3), 167–201 (2010)
Duff, I.S., Reid, J.K.: The multifrontal solution of indefinite sparse symmetric linear. ACM Trans. Math. Softw. (TOMS) 9(3), 302–325 (1983)
Liu, J.W.: The multifrontal method for sparse matrix solution: theory and practice. SIAM Rev. 34(1), 82–109 (1992)
Pakazad, S.Khoshfetrat, Hansson, A., Andersen, M.S., Nielsen, I.: Distributed primal-dual interior-point methods for solving tree-structured coupled convex problems using message-passing. Optim. Methods Softw. 32(3), 401–435 (2017)
Khoshfetrat Pakazad, S., Hansson, A., Andersen, M.S., Rantzer, A.: Distributed semidefinite programming with application to large-scale system analysis. IEEE Trans. Autom. Control 63(4), 1045–1058 (2017)
Alizadeh, F., Haeberly, J.-P.A., Overton, M.L.: Complementarity and nondegeneracy in semidefinite programming. Math. Program. 77(1), 111–128 (1997)
Arnborg, S., Corneil, D.G., Proskurowski, A.: Complexity of finding embeddings in a k-tree. SIAM J. Algebraic Discrete Methods 8(2), 277–284 (1987)
Ye, Y., Todd, M.J., Mizuno, S.: An \(O(\sqrt{nL})\)-iteration homogeneous and self-dual linear programming algorithm. Math. Oper. Res. 19(1), 53–67 (1994)
Wolkowicz, H., Saigal, R., Vandenberghe, L. (eds.): Handbook of Semidefinite Programming: Theory, Algorithms, and Applications. Springer Science & Business Media (2012)
Permenter, F., Friberg, H.A., Andersen, E.D.: Solving conic optimization problems via self-dual embedding and facial reduction: a unified approach. SIAM J. Optim. 27(3), 1257–1282 (2017)
Frieze, A., Jerrum, M.: Improved approximation algorithms for MAX k-CUT and MAX BISECTION. Algorithmica 18(1), 67–81 (1997)
Pataki, G., Stefan S.: The DIMACS library of semidefinite-quadratic-linear programs. Technical Report. Preliminary draft, Computational Optimization Research Center, Columbia University, New York (2002)
Borchers, B.: Sdplib 1.2, a library of semidefinite programming test problems. Optim. Methods Softw. 11(1–4), 683–690 (1999)
Sun, Y.: Decomposition methods for semidefinite optimization. Ph.D. thesis, UCLA (2015)
Liu, J.W.: The role of elimination trees in sparse factorization. SIAM J. Matrix Anal. Appl. 11(1), 134–172 (1990)
Bodlaender, H.L., Gilbert, J.R., Hafsteinsson, H., Kloks, T.: Approximating treewidth, pathwidth, frontsize, and shortest elimination tree. J. Algorithms 18(2), 238–255 (1995)
Leighton, T., Rao, S.: An approximate max-flow min-cut theorem for uniform multicommodity flow problems with applications to approximation algorithms. In: Proceedings of the 29th Annual Symposium on Foundations of Computer Science, pp. 422–431. IEEE (1988)
Klein, P., Stein, C., Tardos, E.: Leighton-Rao might be practical: faster approximation algorithms for concurrent flow with uniform capacities. In: Proceedings of the twenty-second annual ACM symposium on Theory of computing, pp. 310–321. ACM (1990)
George, A., Liu, J.W.: The evolution of the minimum degree ordering algorithm. SIAM Rev. 31(1), 1–19 (1989)
Lipton, R.J., Rose, D.J., Tarjan, R.E.: Generalized nested dissection. SIAM J. Numer. Anal. 16(2), 346–358 (1979)
Rose, D.J., Tarjan, R.E., Lueker, G.S.: Algorithmic aspects of vertex elimination on graphs. SIAM J. Comput. 5(2), 266–283 (1976)
George, A., Liu, J.W.: Computer Solution of Large Sparse Positive Definite Systems. Prentice Hall, Englewood Cliffs, NJ (1981)
Grone, R., Johnson, C.R., Sá, E.M., Wolkowicz, H.: Positive definite completions of partial Hermitian matrices. Linear Algebra Appl. 58, 109–124 (1984)
Dancis, J.: Positive semidefinite completions of partial hermitian matrices. Linear Algebra Appl. 175, 97–114 (1992)
Laurent, M., Varvitsiotis, A.: A new graph parameter related to bounded rank positive semidefinite matrix completions. Math. Program. 145(1–2), 291–325 (2014)
Madani, R., Sojoudi, S., Fazelnia, G., Lavaei, J.: Finding low-rank solutions of sparse linear matrix inequalities using convex optimization. SIAM J. Optim. 27(2), 725–758 (2017)
Jiang, X.: Minimum rank positive semidefinite matrix completion with chordal sparsity pattern. Master’s thesis, UCLA (2017)
Kobayashi, K., Kim, S., Kojima, M.: Correlative sparsity in primal-dual interior-point methods for LP, SDP, and SOCP. Appl. Math. Optim. 58(1), 69–88 (2008)
Parter, S.: The use of linear graphs in Gauss elimination. SIAM Rev. 3(2), 119–130 (1961)
Sturm, J.F.: Using SeDuMi 1.02, a MATLAB toolbox for optimization over symmetric cones. Optim. Methods Softw. 11(1–4), 625–653 (1999)
Andersen, E.D.: Handling free variables in primal-dual interior-point methods using a quadratic cone. In: Proceedings of the SIAM Conference on Optimization, Toronto (2002)
Sturm, J.F.: Implementation of interior point methods for mixed semidefinite and second order cone optimization problems. Optim. Methods Softw. 17(6), 1105–1154 (2002)
Andersen, E.D., Roos, C., Terlaky, T.: On implementing a primal-dual interior-point method for conic quadratic optimization. Math. Program. 95(2), 249–277 (2003)
Goldfarb, D., Scheinberg, K.: Product-form Cholesky factorization in interior point methods for second-order cone programming. Math. Program. 103(1), 153–179 (2005)
Guo, J., Niedermeier, R.: Exact algorithms and applications for Tree-like Weighted Set Cover. J. Discrete Algorithms 4(4), 608–622 (2006)
Lewis, J.G., Peyton, B.W., Pothen, A.: A fast algorithm for reordering sparse matrices for parallel factorization. SIAM J. Sci. Stat. Comput. 10(6), 1146–1173 (1989)
Pothen, A., Sun, C.: Compact clique tree data structures in sparse matrix factorizations. In: Coleman, T.F., Li, Y. (eds.) Large-Scale Numerical Optimization, pp. 180–204. SIAM (1990)
Andersen, M.S., Dahl, J., Vandenberghe, L.: Logarithmic barriers for sparse matrix cones. Optim. Methods Softw. 28(3), 396–423 (2013)
George, A., Gilbert, J.R., Liu, J.W.H. (eds.): Graph Theory and Sparse Matrix Computation. Springer Science & Business Media (2012)
Zimmerman, R.D., Murillo-Sánchez, C.E., Thomas, R.J.: MATPOWER: Steady-state operations, planning, and analysis tools for power systems research and education. IEEE Trans. Power Syst. 26(1), 12–19 (2011)
Josz, C., Fliscounakis, S., Maeght, J., Panciatici, P.: AC power flow data in MATPOWER and QCQP format: iTesla, RTE snapshots, and PEGASE. arXiv preprint arXiv:1603.01533 (2016)
Mittelmann, H.D.: An independent benchmarking of SDP and SOCP solvers. Math. Program. 95(2), 407–430 (2003)
Frenk, H., Roos, K., Terlaky, T., Zhang, S. (eds.): High Performance Optimiztion. Springer Science & Business Media (2013)
Amestoy, P.R., Davis, T.A., Duff, I.S.: Algorithm 837: AMD, an approximate minimum degree ordering algorithm. ACM Trans. Math. Softw. (TOMS) 30(3), 381–388 (2004)
Lavaei, J., Low, S.H.: Zero duality gap in optimal power flow problem. IEEE Trans. Power Syst. 27(1), 92 (2012)
Nakata, K., Fujisawa, K., Fukuda, M., Kojima, M., Murota, K.: Exploiting sparsity in semidefinite programming via matrix completion II: implementation and numerical results. Math. Program. 95(2), 303–327 (2003)
Agler, J., Helton, W., McCullough, S., Rodman, L.: Positive semidefinite matrices with a given sparsity pattern. Linear Algebra Appl. 107, 101–149 (1988)
Vanderbei, R.J.: Linear Programming: Foundations and Extensions. Springer, Berlin (2015)
Nesterov, Y.E., Todd, M.J.: Primal-dual interior-point methods for self-scaled cones. SIAM J. Optim. 8(2), 324–364 (1998)
Sturm, J.F., Zhang, S.: Symmetric primal-dual path-following algorithms for semidefinite programming. Appl. Numer. Math. 29(3), 301–315 (1999)
Sturm, J.F., Zhang, S.: On a wide region of centers and primal-dual interior point algorithms for linear programming. Math. Oper. Res. 22(2), 408–431 (1997)
Todd, M.J., Toh, K.-C., Tütüncü, R.H.: On the nesterov-todd direction in semidefinite programming. SIAM J. Optim. 8(3), 769–796 (1998)
Alizadeh, F., Haeberly, J.-P.A., Overton, M.L.: Primal-dual interior-point methods for semidefinite programming: convergence rates, stability and numerical results. SIAM J. Optim. 8(3), 746–768 (1998)
Acknowledgements
The authors are grateful to Daniel Bienstock, Salar Fattahi, Cédric Josz, and Yi Ouyang for insightful discussions and helpful comments on earlier versions of this manuscript. We thank Frank Permenter for clarifications on various aspects of the homogeneous self-dual embedding for SDPs. Finally, we thank the Associate Editor and Reviewer 2 for meticulous and detailed comments that led to a significantly improved paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This work was supported by the ONR YIP Award, DARPA YFA Award, AFOSR YIP Award, NSF CAREER Award, and ONR N000141712933.
Appendices
A Linear independence and Slater’s conditions
In this section, we prove that (CTC) inherits the assumptions of linear independence and Slater’s conditions from (SDP). We begin with two important technical lemmas.
Lemma 10
The matrix \({\mathbf {N}}\) in (13) has full row rank, that is \(\det ({\mathbf {N}}{\mathbf {N}}^{T})\ne 0\).
Proof
We make \({\mathbf {N}}=[{\mathbf {N}}_{i,j}]_{i,j=1}^{\ell }\) upper-triangular by ordering its blocks topologically on T: each nonempty block row \({\mathbf {N}}_{j}\) contains a nonzero block at \({\mathbf {N}}_{j,j}\) and a nonzero block at \({\mathbf {N}}_{j,p(j)}\) where the parent node \(p(j)>j\) is ordered after j. Then, the claim follows because each diagonal block \({\mathbf {N}}_{j,j}\) implements a surjection and must therefore have full row-rank. \(\square \)
Lemma 11
(Orthogonal complement) Let \(d=\frac{1}{2}\sum _{j=1}^{\ell }|J_{j}|(|J_{j}|+1)\). Implicitly define the \(d\times \frac{1}{2}n(n+1)\) matrix \({\mathbf {P}}\) to satisfy
Then, (i) \({\mathbf {N}}{\mathbf {P}}=0\); (ii) every \(x\in {\mathbb {R}}^{d}\) can be decomposed as \(x={\mathbf {N}}^{T}u+{\mathbf {P}}v\).
Proof
For each \(x=[\mathrm {svec}\,(X_{j})]_{j=1}^{\ell }\in {\mathbb {R}}^{d}\), Theorem 4 says that there exists a Z satisfying \({\mathbf {P}}\,\mathrm {svec}\,(Z)=x\) if and only if \({\mathbf {N}}x=0\). Equivalently, \(x\in {\mathrm {span}}({\mathbf {P}})\) if and only if \(x\perp {\mathrm {span}}({\mathbf {N}}^{T})\). The “only if” part implies (i), while the “if” part implies (ii).
\(\square \)
Define the \(m\times \frac{1}{2}n(n+1)\) matrix \({\mathbf {M}}\) as the vectorization of the linear constraints in (SDP), as in
In reformulating (SDP) into (CTC), the splitting conditions (10) can be rewritten as the following
where \(c=[\mathrm {svec}\,(C_{j})]_{j=1}^{\ell }\) and \({\mathbf {A}}=[\mathrm {svec}\,(A_{i,j})^{T}]_{i,j=1}^{m,\ell }\) are the data for the vectorized verison of (CTC).
Proof of Lemma 1
We will prove that
(\(\implies \)) We must have \(u\ne 0\), because \({\mathbf {N}}\) has full row rank by Lemma 10, and so \({\mathbf {A}}^{T}0+{\mathbf {N}}^{T}v=0\) if and only if \(v=0\). Multiplying by \({\mathbf {P}}\) yields \({\mathbf {P}}^{T}({\mathbf {A}}^{T}u+{\mathbf {N}}^{T}v)={\mathbf {M}}^{T}u+0=0\) and so setting \(y=u\ne 0\) yields \({\mathbf {M}}^{T}y=0\). () We use Lemma 11 to decompose \({\mathbf {A}}^{T}y={\mathbf {P}}z+{\mathbf {N}}^{T}v\). If \(\mathrm {svec}\,(\sum _{i}y_{i}A_{i})={\mathbf {M}}^{T}y={\mathbf {P}}^{T}{\mathbf {A}}^{T}y=0,\) then \({\mathbf {P}}^{T}{\mathbf {P}}z={\mathbf {P}}^{T}{\mathbf {A}}^{T}y-{\mathbf {P}}^{T}{\mathbf {N}}^{T}v=0\) and so \({\mathbf {P}}z=0\). Setting \(u=-y\ne 0\) yields \({\mathbf {A}}^{T}u+{\mathbf {N}}^{T}v=0\). \(\square \)
Proof of Lemma 2
We will prove that
Define the chordal completion F as in (8). Observe that \({\mathbf {M}}\,\mathrm {svec}\,(X)={\mathbf {M}}\,\mathrm {svec}\,(Z)\) holds for all pairs of \(P_{F}(X)=Z\), because each \(A_{i}\in {\mathbb {S}}_{F}^{n}\) satisfies \(A_{i}\bullet X=A_{i}\bullet P_{F}(X)\). Additionally, the positive definite version of Theorem 3 is written
This result was first established by Grone et al. [52]; a succinct proof can be found in [15, Theorem 10.1]. (\(\implies \)) For every x satisfying \({\mathbf {N}}x=0\), there exists Z such that \({\mathbf {P}}\,\mathrm {svec}\,(Z)=x\) due to Lemma 11. If additionally \(x\in \mathrm {Int}({\mathcal {K}})\), then there exists \(X\succ 0\) satisfying \(Z=P_{F}(X)\) due to (39). We can verify that \({\mathbf {M}}\,\mathrm {svec}\,(X)={\mathbf {M}}\,\mathrm {svec}\,(Z)={\mathbf {A}}{\mathbf {P}}\,\mathrm {svec}\,(Z)={\mathbf {A}}x=b\). () For every \(X\succ 0,\) there exists Z satisfying \(Z=P_{F}(X)\) and \({\mathbf {P}}\,\mathrm {svec}\,(Z)\in \mathrm {Int}({\mathcal {K}})\) due to (39). Set \(u={\mathbf {P}}\,\mathrm {svec}\,(Z)\) and observe that \(u\in \mathrm {Int}({\mathcal {K}})\) and \({\mathbf {N}}u={\mathbf {N}}{\mathbf {P}}\,\mathrm {svec}\,(Z)=0\). If additionally \({\mathbf {M}}\,\mathrm {svec}\,(X)=b\), then \({\mathbf {A}}u={\mathbf {A}}{\mathbf {P}}\,\mathrm {svec}\,(Z)={\mathbf {M}}\,\mathrm {svec}\,(Z)=b\). \(\square \)
Proof of Lemma 2
We will prove that
Define the chordal completion F as in (8). Theorem 3 in (39) has a dual theorem
This result readily follows from the positive semidefinite version proved by Alger et al. [76]; see also [15, Theorem 9.2]. (\(\implies \)) For each \(h=c-{\mathbf {A}}^{T}u-{\mathbf {N}}^{T}v\), define \(S=C-\sum _{i}u_{i}A_{i}\) and observe that
If additionally \(h\in {\mathcal {K}}_{*}\), then \(S\succ 0\) due to (40). () For each \(S=C-\sum _{i}y_{i}A_{i}\succ 0\), there exists an \(h\in \mathrm {Int}({\mathcal {K}}_{*})\) satisfying \(\mathrm {svec}\,(S)={\mathbf {P}}^{T}h\) due to (40). We use Lemma 11 to decompose \(h={\mathbf {P}}u+{\mathbf {N}}^{T}v\). Given that \(\mathrm {svec}\,(S)={\mathbf {P}}^{T}h={\mathbf {P}}^{T}{\mathbf {P}}u+0\), we must actually have \({\mathbf {P}}u=c-{\mathbf {A}}^{T}y\) since \({\mathbf {P}}^{T}(c-{\mathbf {A}}^{T}y)=\mathrm {svec}\,(C)-{\mathbf {M}}^{T}y=\mathrm {svec}\,(S)\). Hence \(h=c-{\mathbf {A}}^{T}y+{\mathbf {N}}^{T}v\) and \(h\in \mathrm {Int}({\mathcal {K}}_{*})\). \(\square \)
B Extension to inequality constraints
Consider the modifying the equality constraint in (11) into an inequality constraint, as in
The corresponding dualization reads
where m denotes the number of rows in \({\mathbf {A}}\) and f now denotes the number of rows in \({\mathbf {N}}\). Embedding the equality constraint into a second-order cone, the associated normal equation takes the form
where \({\mathbf {D}}_{s}\) and \({\mathbf {D}}_{f}\) are comparable as before in (15) and (23), and \({\mathbf {D}}_{l}\) is a diagonal matrix with positive diagonal elements. This matrix has the same sparse-plus-rank-1 structure as (22), and can therefore be solved using the same rank-1 update
where \({\mathbf {H}}\) and \({\mathbf {q}}\) now read
The matrix \({\mathbf {H}}\) has the same block sparsity graph as the tree graph T, so we can evoke Lemma 5 to show that the cost of computing \(\varDelta y\) is again \(O(\omega ^{6}n)\) time and \(O(\omega ^{4}n)\) memory.
C Interior-point method complexity analysis
We solve the dualized problem (21) by solving its extended homogeneous self-dual embedding
where the data is given in standard form
and the residual vectors are defined
Here, \(\nu \) is the order of the cone \({\mathcal {C}}\), and \({\mathbf {1}}_{{\mathcal {C}}}\) is its identity element
Problem (41) has optimal value \(\theta ^{\star }=0\). Under the primal-dual Slater’s conditions (Assumption 2), an interior-point method is guaranteed to converge to an \(\epsilon \)-accurate solution with \(\tau >0\), and this yields an \(\epsilon \)-feasible and \(\epsilon \)-accurate solution to the dualized problem (21) by rescaling \(x/\tau \) and \(y=y/\tau \) and \(s=s/\tau \). The following result is adapted from [38, Lemma 5.7.2] and [77, Theorem 22.7].
Lemma 12
(\(\epsilon \)-accurate and \(\epsilon \)-feasible) If \((x,y,s,\tau ,\theta ,\kappa )\) satisfies (41b) and (41c) and
for constants \(\epsilon ,\gamma >0\), then the rescaled point \((x/\tau ,y/\tau ,s/\tau )\) satisfies
where K is a constant.
Proof
Note that (41b) implies \(\mu =\theta \) and
Hence, we obtain our desired result by upper-bounding \(1/\tau \). Let \((x^{\star },y^{\star },s^{\star },\tau ^{\star },\theta ^{\star },\kappa ^{\star })\) be a solution of (41), and note that for every \((x,y,s,\tau ,\theta ,\kappa )\) satisfying (41b) and (41c), we have the following via the skew-symmetry of (41b)
Rearranging yields
and hence
If (SDP) satisfies the primal-dual Slater’s condition, then (CTC) also satisfies the primal-dual Slater’s condition (Lemmas 2& 3). Therefore, the vectorized version (11) of (CTC) attains a solution \(({\hat{x}},{\hat{y}},{\hat{s}})\) with \({\hat{x}}^{T}{\hat{s}}=0\), and the following
with \(\theta ^{\star }=\kappa ^{\star }=0\) is a solution to (41). This proves the following upper-bound
Setting \(K=\max \{\Vert r_{p}\Vert K_{\tau },\Vert r_{d}\Vert K_{\tau },K_{\tau }^{2}\}\) yields our desired result. \(\square \)
We solve the homogeneous self-dual embedding (41) using the short-step method of Nesterov and Todd [78, Algorithm 6.1] (and also Sturm and S. Zhang [79, Section 5.1]), noting that SeDuMi reduces to it in the worst case; see [61] and [80]. Beginning at the following strictly feasible, perfectly centered point
with barrier parameter \(\mu =1\), we take the following steps
along the search direction defined by the linear system [81, Eqn. 9]
Here, F is the usual self-concordant barrier function on \({\mathcal {C}}\)
and \(w\in \mathrm {Int}({\mathcal {C}})\) is the unique scaling point satisfying \(\nabla ^{2}F(w)x=s\), which can be computed from x and s in closed-form. The following iteration bound is an immediate consequence of [78, Theorem 6.4]; see also [79, Theorem 5.1].
Lemma 13
(Short-Step Method) The sequence in (44) arrives at an iterate \((x,y,s,\tau ,\theta ,\kappa )\) satisfying the conditions of Lemma 12 with \(\gamma =9/10\) in at most \(O(\sqrt{\nu }\log (1/\epsilon ))\) iterations.
The cost of each interior-point iteration is dominated by the cost of computing the search direction in (45). Using elementary but tedious linear algebra, we can show that if
where \({\mathbf {D}}=\nabla ^{2}F(w)\) and \(d=-s-\mu ^{+}\nabla F(x)\), and
then
where \({\mathbf {D}}_{0}=\kappa /\tau \) and \(d_{0}=-\kappa +\mu ^{+}\tau ^{-1}\). Hence, the cost of computing the search direction is dominated by the cost of solving the normal equation for three different right-hand sides. Here, the normal matrix is written
where \(\sigma =\frac{1}{2}(w_{0}^{2}-w_{1}^{T}w_{1})>0\) and \(\otimes _{s}\) denotes the symmetric Kronecker product [82] implicitly defined to satisfy
Under the hypothesis on \({\mathbf {A}}\) stated in Theorem 5, the normal matrix satisfies the assumptions of Lemma 5, and can therefore be solved in linear O(n) time and memory.
Proof of Theorem 5
Combining Lemmas 12 and 13 shows that the desired \(\epsilon \)-accurate, \(\epsilon \)-feasible iterate is obtained after \(O(\sqrt{\nu }\log (1/\epsilon ))\) interior-point iterations. At each iteration we perform the following steps: (1) compute the scaling point w; (2) solve the normal equation (47a) for three right-hand sides; (3) back-substitute (47b)–(47f) for the search direction and take the step in (44). Note from the proof of Lemma 5 that the matrix \([{\mathbf {A}};{\mathbf {N}}]\) has at most \(O(\omega ^{2}n)\) rows under Assumption 1, and therefore \(\mathrm {nnz}\,({\mathbf {M}})=O(\omega ^{4}n)\) under the hypothesis of Theorem 5. Below, we show that the cost of each step is bounded by \(O(\omega ^{6}n)\) time and \(O(\omega ^{4}n)\) memory.
Scaling point We partition \(x=[x_{0};x_{1};\mathrm {svec}\,(X_{1});\ldots ;\mathrm {svec}\,(X_{\ell })]\) and similarly for s. Then, the scaling point w is given in closed-form [61, Section 5]
Noting that \(\mathrm {nnz}\,(w_{1})\le O(\omega ^{2}n)\), \(\ell \le n\) and each \(W_{j}\) is at most \(\omega \times \omega \), the cost of forming \(w=[w_{0};w_{1};\mathrm {svec}\,(W_{1});\ldots ;\mathrm {svec}\,(W_{\ell })]\) is at most \(O(\omega ^{3}n)\) time and \(O(\omega ^{2}n)\) memory. Also, since
the cost of each matrix-vector product with \({\mathbf {D}}\) and \({\mathbf {D}}^{-1}\) is also \(O(\omega ^{3}n)\) time and \(O(\omega ^{2}n)\) memory.
Normal equation The cost of matrix-vector products with \({\mathbf {M}}\) and \({\mathbf {M}}^{T}\) is \(\mathrm {nnz}\,({\mathbf {M}})=O(\omega ^{4}n)\) time and memory. Using Lemma 5, we form the right-hand sides and solve the three normal equations in (47a) in \(O(\omega ^{6}n)\) time and \(O(\omega ^{4}n)\) memory.
Back-substitution The cost of back substituting (47b)–(47f) and making the step (44) is dominated by matrix-vector products with \({\mathbf {D}}\), \({\mathbf {D}}^{-1}\), \({\mathbf {M}}\), and \({\mathbf {M}}^{T}\) at \(O(\omega ^{4}n)\) time and memory.\(\square \)
Rights and permissions
About this article
Cite this article
Zhang, R.Y., Lavaei, J. Sparse semidefinite programs with guaranteed near-linear time complexity via dualized clique tree conversion. Math. Program. 188, 351–393 (2021). https://doi.org/10.1007/s10107-020-01516-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10107-020-01516-y