Posted 2019-06-05
Note: This post was inspired by Kunal Shah's question that came up at some point during one of our meetings: is there an efficient way of finding an ellipsoid which covers the intersections and unions of a bunch of other ellipsoids?
While this question has been explored somewhat extensively, the exposition is often more general than necessary and aimed at a relatively mathematical audience. Either way, if you're interested, both papers are fairly well-written—I highly recommend at least a quick skim!
The S-procedure is a well known lemma in control theory that seeks to answer the following question:
Let's say we have a bunch of quadratic functions . When is it true that
for ? (Recall that a quadratic is a function of the form for symmetric , , and ).
There are many reasons to attempt to answer this (surprisingly useful) question. The original motivations were to show stability of systems, though the domain of applications is certainly larger. We can use this to show anything from impossibility results (for example, many of the results of this paper can be recast in terms of the S-procedure) to, well, in our case, the construction of a small covering ellipsoid from a bunch of other ellipsoids, which is itself useful for things like filtering (for localizing drones from noisy measurements for example) along with many other applications.
If you're familiar with Lagrange duality, this is mostly an equivalent statement—except that this statement is in the special case of quadratics, where you can say a little more than with general functions.
We can fully and completely answer this question in the case that : there exists a nonnegative number such that
for all if, and only if, .
Why? Well, let say we have a that satisfies the above inequality. Then, if we have an such that , then
The converse is slightly trickier, so I will defer to B&V's Convex Optimization which has a very readable presentation of the proof (see B.1 and B.2 in the appendix).
The general case is really only a slight change from the case (except that the converse of the statement is not true). In particular, if there exist such that
then, . Showing this is nearly the same as the case,
So now we have a family of sufficient (but not necessary!) conditions for which we know when implies that .
Ellipsoids are a particularly nice family to work with since, as you may have guessed, they are the sets defined by
where is a convex quadratic. This definition gives us a way of translating statements about sets (inclusion, etc) into statements about the functions which generate them. In particular, if we have two ellipsoids defined by the convex quadratics , then
But wait a minute, we know exactly when this happens! By the previous section, we found that
if and only if there is some with . Also note that if we have a union of a bunch of ellipsoids (say ) that we want to cover with an ellipsoid , then this is the same as saying
or, that each ellipsoid is covered by the big one, .
Ok, to reiterate, we are looking for a small ellipsoid such that contains all of the other ellipsoids , where the and are convex quadratics. In other words, using the results of the previous subsection, we look for a quadratic such that
which we know happens only when there exists some with
Now remains the final question: given two quadratics, and and some number , how can we check if for all ? I won't prove this (though I have written a quick proof of this statement in my notes, found here), but, if we let and then for all if, and only if,
where we say two symmetric matrices satisfy whenever for all . A straightforward exercise it to verify that the set of matrices is a convex cone (almost universally called the positive semidefinite or PSD cone).
This rewriting is extremely useful, since we've turned a problem over a potentially difficult-to-handle space (the space of quadratics greater than or equal to another) into a problem that is easy to handle (the PSD cone). The best news, though, is that we have efficient algorithms to solve optimization problems whose constraints are PSD constraints.[1]
Finally, after enough background, we can get to the final goal: writing an efficiently-solvable optimization problem to give us a small bounding ellipsoid.
There are several ways of defining "small," in the case of ellipsoids, but one of the most common definitions is to pick the ellipsoid with the smallest volume. In the case that , the volume of this ellipsoid is given by the determinant, , of the matrix . So, we can write—using the conditions given above—an optimization problem corresponding to finding the smallest (in volume) ellipsoid which contains all ellipsoids, as
The only problem here (which we can easily fix) is that the determinant is not a convex function. On the other hand, the log determinant is (for a proof, see the Convex book, section 3.1.5), so we can write,
This is equivalent to the original problem since is an increasing function of .
Of course, any convex function (such as, for example the trace) would do here as well.
Ok, now we know how to solve the problem where we have a bunch of ellipsoids and we want to find an ellipsoid which covers all of them. How about the problem where we want to find an ellipsoid which also covers the sets for , which are, themselves, intersections of ellipsoids?
In particular, if is defined as
for some index set and some set of ellipsoids , each of which are defined as before (), we can perform a similar trick to the one above!
More generally, if we have an ellipsoid , then
(It's a worthwhile exercise to think about why, but it follows the same idea as before.) In other words, is a superset of only when a bunch of quadratic inequalities imply another. (Where have we seen this before...?)
In other words, we know that (by the first section), if there exist such that
then we immediately have that . Since the converse is not true, we are sadly not guaranteed to actually find the smallest bounding ellipsoid , but this is usually quite a good approximation (if it's not exact).
Following exactly the same steps as in the previous section and using the same definitions, we now get a new program for minimizing the volume for the union and intersection of ellipsoids:
As before, this program does not guarantee actually finding the minimal volume ellipsoid, but it is likely to be quite close! (That is, if it's not spot on, most of the time.)
[1] | For more information, see Boyd's Linear Matrix Inequalities book. |