(Bar13) Convexity of the image of a quadratic map via the relative entropy distance
Posted: 2016-04-02 , Modified: 2016-04-02
Tags: none
Posted: 2016-04-02 , Modified: 2016-04-02
Tags: none
Theorem: Let \(q_1,\ldots, q_k:\R^n\to \R\) be positive definite quadratic forms, \(\psi = (q_1,\ldots, q_k)\). Let \(a\in \conv(\psi(\R^n))\cap \De\), where \(\De\) is the probability simplex. There exists \(b\in \psi(\R^n)\cap \De\) that is at most 4.8 away in KL divergence.
(Just for some motivation here)
When \(k=2\), \(\psi(\R^n)\) is convex. Proof: When system of linear equations in \(X\succeq 0\) have a solution, and the number of equations is small, there is in fact a low-rank solution. Here, look for a rank 1 solution.
Note KL-divergence is easier to deal with—the logs mean that scaling comes out as constants—and is stronger than \(L^p\) norms.
Prove Theorem 3.2 by SDP rounding. Let \(X\) be the optimal solution, write \(X=T^2\). In SDP rounding, take \(x\sim N(0,I)\), \(y=Tx\), and hope that \(y\) is a good solution to the original problem. (If \(X=xx^T\), then \(T\) would be a projection.) We use the Quadratic Sampling Lemma—the expectation of \(q(y)\) for any quadratic \(q\) is the same as the \(\an{Q,X}\).
Thus \(\E q_i(y) = \an{Q_i,X}=1\).
Now do some calculations (Lemma 2.1). Calculate that \[\begin{align} \E |\ln q_i(y)| & \le C_1 \implies \E\ab{\sumo ik a_i \ln q_i(y)}\le C\\ \Pj( \ve{y}^2 \ge C_2) & \le \ep_3 \end{align}\]for some \(C_1,C_2,\ep_3\). Use Markov’s inequality on the first inequality. Find there is a point with \(\ve{y}^2<C_2\), \(\sumo ik a_i\ln q_i(y)>-C_4\). Let \(y' = \wh y\) be the solution.