Constrained optimization
Posted: 2016-04-28 , Modified: 2016-04-28
Tags: convex optimization
Posted: 2016-04-28 , Modified: 2016-04-28
Tags: convex optimization
Consider \(\min_{Ax=b}f\).
This is unbounded below if there exist \(v,w\) such that \(Pv+A^Tw=0\), \(Av=0\), \(-q^Tv+b^Tw>0\) (left-multiply by \((v^T\,w^T)\)) to see that the equation above doesn’t have a solution).
Recall the dual function. Why do we care about it? \[g(\la,\nu) = \min_x f+\nu^T (Ax-b) = -\nu^Tb - f^*(-\nu^TA), \qquad f^*(y)=\max y^Tx - f(x).\] Equality constraints disappear in the dual. If the dual is nice, we can just solve an unconstrained problem \(\max_{\la \ge 0} g(\la,\nu)\).
Describe the Newton method starting with a feasible \(x, Ax=b\). The Newton step is the minimizer for the quadratic approximation under the constraint. (Finding the minimum of a quadratic amounts to solving a linear equation.) Using \[\matt{\nb^2 f}{A^T}A0 \coltwo{\De x_{nt}}w = \coltwo{-\nb f}{0}.\] Define \[\la(x) = (\De x_{nt}^T \nb^2 f \De x_{nt})^{\rc 2}\].
Note this is normal Newton if you use a parametrization.
(Convexity makes the KKT matrix invertible.)
Describe the infeasible start Newton method. Approximate \(f\) by a quadratic using \(P=\nb^2 f\) in \(\eqref{eq:kkt-mat}\) and find \(\De x_{nt}\) so that \(x+\De x_{nt}\) satisfies the KKT conditions. The equation for \(\De x_{nt}\) is \[\matt{\nb^2 f}{A^T}A0 \coltwo{\De x_{nt}}w = \coltwo{-\nb f}{b-Ax}.\]
Note here we’re just updating \(x\) by solving for \(\De x_{nt}\). Ech time \(w\) is treated just as an auxiliary variable. But \(w\) comes from the dual solution \(\nu\). Can we look at \((x,\nu)\) together as a primal-dual pair and update both of them?
Let the (dual and primal) residual be (recall KKT says we want \(\nb f + A^T\nu =0\)) \[\begin{align} r &= (\nb f + A^T \nu, Ax-b). \end{align}\]Instead of minimizing \(f\), we minimize the residual for the KKT conditions. The residual has a component for minimizing \(f\), and a component for trying to satisfy \(Ax=b\).
Here is the algorithm.