Paper Summary: Project-free Online Learning
In this blog, I will give a summary of the paper “Project-free Online Learning” and some of my understandings. As its title suggests, this paper proposes an efficient projection-free algorithm for online convex optimization. As we discussed earlier, traditional online gradient descent method requires a projection step onto feasible set. In Euclidean space, this projection step is essentially a step to solve a convex quadratic programming problem on a convex set. This step can be computational expensive for certain constraint set
1. Convex sets on which projection is expensive
For a convex set
Moreover, a linear optimization step involves solving
for some vector
Bounded Trace Norm Matrices. The set of matrices with bounded trace norm is a common decision set for applications like matrix completion and collaborative filtering. Formally, the set is given by
The projection of a matrix
Flow polytope.
Given a directed acyclic graph
Matroid polytope
Given a matroid
Rotations
Rotation matrices are
2. Preliminaries
The online convex optimization setting is similar to previous blogs. In each round
is sub-linear in
A more interesting concept called smoothed functions is introduced in the paper. Given
where
3. The algorithm
The algorithm uses Frank-Wolfe technique, so the authors call it Online Frank-Wolfe (OFW). For initialization, we choose a input parameter
- Play
and observe . - Compute
. - Compute
. - Set
.
The linear optimization step shows in step 3. Define
where
Theorem 3.1 Assume that for
for both the following values of
and
In either case, this bound is obtained by setting a=d-b in the above algorithm.
With the above theorem, in the different settings, we can obtain regret bounds by setting different parameters. For example, if the cost functions are stochastic and smooth, then
for
In summary, the authors give the following table of regret bounds in different settings. Although all of regret bounds are sub-linear, the algorithm only achieves the optimal regret in the stochastic smooth setting. The additional regret costs in adversarial setting come from the projection-free operation.
Stochastic | Adversarial | |
---|---|---|
Smooth costs | ||
Non-smooth costs |
4. Conclusion
In the following section, the authors implemented the proposed algorithm on two standard datasets. The experiment results show the effectiveness of OFW over OGD. In conclusion, the authors proposed an an efficient algorithm scheme for online convex optimization that performs one linear optimization per iteration rather than one quadratic optimization. The advantages include computational efficiency, parameter free and efficient representation and sparsity. The regret analysis is performed in different settings. However, the regret bounds are not always optimal. So the open problem of how to improve the regret bounds are left as on future direction.