Towards a Theory of Non-Commutative Optimization: Geodesic 1st and 2nd Order Methods for Moment Maps and Polytopes

Type: Preprint

Publication Date: 2019-11-01

Citations: 59

DOI: https://doi.org/10.1109/focs.2019.00055

Download PDF

Abstract

This paper initiates a systematic development of a theory of non-commutative optimization, a setting which greatly extends ordinary (Euclidean) convex optimization. It aims to unify and generalize a growing body of work from the past few years which developed and analyzed algorithms for natural geodesically convex optimization problems on Riemannian manifolds that arise from the symmetries of non-commutative groups. More specifically, these are algorithms to minimize the moment map (a noncommutative notion of the usual gradient), and to test membership in moment polytopes (a vast class of polytopes, typically of exponential vertex and facet complexity, which quite magically arise from this apriori non-convex, non-linear setting). The importance of understanding this very general setting of geodesic optimization, as these works unveiled and powerfully demonstrate, is that it captures a diverse set of problems, many non-convex, in different areas of CS, math, and physics. Several of them were solved efficiently for the first time using noncommutative methods; the corresponding algorithms also lead to solutions of purely structural problems and to many new connections between disparate fields. In the spirit of standard convex optimization, we develop two general methods in the geodesic setting, a first order and a second order method, which respectively receive first and second order information on the "derivatives" of the function to be optimized. These in particular subsume all past results. The main technical work, again unifying and extending much of the previous work, goes into identifying the key parameters of the underlying group actions which control convergence to the optimum in each of these methods. These non-commutative analogues of "smoothness" in the commutative case are far more complex, and require significant algebraic and analytic machinery (much existing and some newly developed here). Despite this complexity, the way in which these parameters control convergence in both methods is quite simple and elegant. We also bound these parameters in several general cases. Our work points to intriguing open problems and suggests further research directions. We believe that extending this theory, namely understanding geodesic optimization better, is both mathematically and computationally fascinating; it provides a great meeting place for ideas and techniques from several very different research areas, and promises better algorithms for existing and yet unforeseen applications.

Locations

  • arXiv (Cornell University) - View - PDF
  • DataCite API - View

Similar Works

Action Title Year Authors
+ Non-commutative Optimization - Where Algebra, Analysis and Computational Complexity Meet 2022 Avi Wigderson
+ Geodesic Convex Optimization: Differentiation on Manifolds, Geodesics, and Convexity 2018 Nisheeth K. Vishnoi
+ Optimization, Complexity and Invariant Theory (Invited Talk) 2021 Peter BĂŒrgisser
+ Gauss–Newton method for convex composite optimizations on Riemannian manifolds 2011 Jinhua Wang
Jen‐Chih Yao
Chong Li
+ Preliminaries and Overview of Euclidean Optimization 2021 Hiroyuki Sato
+ An Introduction to Optimization on Smooth Manifolds 2023 Nicolas Boumal
+ Geodesic convexity 2023
+ Computing Brascamp-Lieb Constants through the lens of Thompson Geometry 2022 Melanie Weber
Suvrit Sra
+ Curvature and complexity: Better lower bounds for geodesically convex optimization 2023 Christopher Criscitiello
Nicolas Boumal
+ On the matrix square root via geometric optimization 2015 Suvrit Sra
+ On the matrix square root via geometric optimization 2015 Suvrit Sra
+ PDF Chat Riemannian Optimization via Frank-Wolfe Methods 2022 Melanie Weber
Suvrit Sra
+ Riemannian Optimization 2024 Andi Han
Pratik Jawanpuria
Bamdev Mishra
+ Riemannian optimization on tensor products of Grassmann manifolds: Applications to generalized Rayleigh-quotients 2010 O. Curtef
Gunther Dirr
Uwe Helmke
+ Computing the Gromov--Hausdorff distance using first-order methods 2023 Vladyslav Oles
+ On a class of geodesically convex optimization problems solved via Euclidean MM methods 2022 Suvrit Sra
Melanie Weber
+ Interior-point methods on manifolds: theory and applications 2023 Hiroshi Hirai
Harold Nieuwboer
Michael Walter
+ Interior-point methods on manifolds: theory and applications 2023 Hiroshi Hirai
Harold Nieuwboer
Michael Walter
+ Riemannian Optimization via Frank-Wolfe Methods 2017 Melanie Weber
Suvrit Sra
+ PDF Chat On the Matrix Square Root via Geometric Optimization 2016 Suvrit Sra