A faster interior-point method for sum-of-squares optimization

Jiang, Shunhua; Natura, Bento; and Weinstein, Omri A faster interior-point method for sum-of-squares optimization. In: 49th EATCS International Conference on Automata, Languages, and Programming, ICALP 2022. Leibniz International Proceedings in Informatics, LIPIcs . Schloss Dagstuhl- Leibniz-Zentrum fur Informatik GmbH, Dagstuhl Publishing. ISBN 9783959772358
Copy

We present a faster interior-point method for optimizing sum-of-squares (SOS) polynomials, which are a central tool in polynomial optimization and capture convex programming in the Lasserre hierarchy. Let p = ∑ i qi2 be an n-variate SOS polynomial of degree 2d. Denoting by (Equation presented) and (Equation presented) the dimensions of the vector spaces in which qi's and p live respectively, our algorithm runs in time Õ(LU1.87). This is polynomially faster than state-of-art SOS and semidefinite programming solvers [16, 15, 27], which achieve runtime Õ(L0.5 min{U2.37, L4.24}). The centerpiece of our algorithm is a dynamic data structure for maintaining the inverse of the Hessian of the SOS barrier function under the polynomial interpolant basis [27], which efficiently extends to multivariate SOS optimization, and requires maintaining spectral approximations to low-rank perturbations of elementwise (Hadamard) products. This is the main challenge and departure from recent IPM breakthroughs using inverse-maintenance, where low-rank updates to the slack matrix readily imply the same for the Hessian matrix.

picture_as_pdf

picture_as_pdf
subject
Published Version
Available under Creative Commons: Attribution 4.0

Download

Atom BibTeX OpenURL ContextObject in Span OpenURL ContextObject Dublin Core MPEG-21 DIDL Data Cite XML EndNote HTML Citation METS MODS RIOXX2 XML Reference Manager Refer ASCII Citation
Export

Downloads