Frank-Wolfe

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search

Author: Adeline Tse, Jeff Cheung, Tucker Barrett (SYSEN 5800 Fall 2021)

Introduction

The Frank-Wolfe algorithm is an iterative first-order optimization algorithm for constrained convex optimization, first proposed by Marguerite Frank and Philip Wolfe from Princeton University in 1956.[1] It is also known as the “gradient and interpolation” algorithm or sometimes the “conditional gradient method” as it utilizes a simplex method of comparing an initial feasible point with a secondary basic feasible point to maximize the linear constraints in each iteration to find or approximate the optimal solution.

Advantages of the Frank-Wolfe algorithm include that it is simple to implement, it results in a projection-free computation (that is, it does not require projections back to the constraint set to ensure feasibility), and it generates solution iterations that are sparse with respect to the constraint set. However, one of the major drawbacks of Frank-Wolfe algorithm is the practical convergence, which can potentially be an extremely slow computation process. Nonetheless, new sub-method has been proposed to accelerate the algorithm by better aligning the descent direction while still preserving the projection-free property. [2] Therefore, despite the downside of Frank-Wolfe algorithm, the numerous benefits of this algorithm allow it to be utilized in artificial intelligence, machine learning, traffic assignment, signal processing, and many more applications.

Theory & Methodology

Figure 1. The Frank-Wolfe method optimizes by considering the linearization of the objective function f and moving the initial position x towards the minimizer of the linear function.[3]

The Frank-Wolfe algorithm uses step size and postulated convexity, which formulates a matrix of positive semidefinite quadratic form. Just like a convex function yields a global minimum at any local minimum on a convex set, by the definition of nonlinear programming, the concave quadratic function would yield a global maximum point at any local maximum on a convex constraint set.

The methodology of the Frank-Wolfe algorithm starts with obtaining the generalized Lagrange multipliers for a quadratic problem, PI. The found multipliers are considered solutions to a new quadratic problem,  PII. The simplex method is then applied to the new problem, PII, and the gradient and interpolation method is utilized to make incremental steps towards the solution of the original quadratic problem, PI. This is performed by taking an initial feasible point, obtaining a second basic feasible point along the gradient of the objective function at the initial point, and maximizing the objective function along the segment that joins these two points.  The found maximum along the segment is utilized as the starting point for the next iteration.[1] Using this method, the values of the objective function converge to zero and in one of the iterations, a secondary point will yield a solution.

Proof

  • is a matrix
  • is a matrix
  • is a matrix
  • and are and matrices respectively
  • is the gradient

PI is represented by the following equations using matrix notation:

Since PI is feasible, any local maximum found that is contained within the convex constraint set is a global maximum. Then by utilizing Kuhn and Tucker's generalizations of the Lagrange Multipliers of the maximization of the objective function , is considered the solution of PI.

By the Duality Theorem,

Therefore, the solution for PI is only valid if:

Since

Therefore, by the generalization of Lagrangian, is extracted with linear constraints:

The completion of the generalized Lagrangian shows that positive variables exist for PI, thus a feasible solution exists.

Based on the concavity of ,

If satisfies (where x is a solution if for some ), then

By the Boundedness Criterion and the Approximation Criterion,

The Frank-Wolfe algorithm then adds a non-negative variable to each of the inequalities to obtain the following:

  • and are and matrices respectively

Thus, the PII problem is to obtain search vectors to satisfy , and is considered a solution of PI if belongs to the convex cone spanned by the outward normals to the bounding hyperplanes on which lines as shown in Figure 1.

Algorithm

The Frank-Wolfe algorithm can generally be broken down into five steps as described below. Then the loop of iterations continues throughout Steps 2 to 5 until the minimum extreme point is identified.

Step 1. Define initial solution

If is the extreme point, the initial arbitrary basic feasible solution can be considered as:

The letter denotes the number of iterations.

Step 2. Determine search direction

Search direction, that is the direction vector is:

and are feasible extreme points and belong to where is convex

First-order Taylor expansion around and problem is now reformulated to LP:

Step 3. Determine step length

Step size is defined by the following formula where must be less than 1 to be feasible:

Step 4. Set new iteration point

Step 5. Stopping criterion

Check if is an approximation of , the extreme point. Otherwise, set and return to Step 2 for the next iteration.

In the Frank-Wolfe algorithm, the value of descends after each iteration and eventually decreases towards , the global minimum.

Numerical Example

Consider the non-linear problem min z

Step 1. Choose a starting point

Begin by choosing the feasible point (0,0)

Step 2. Determine a search direction The search direction is obtained by solving linear problem subject to the constraints defined earlier The solution of this linear problem can be obtained via any linear programming algorithm, ie. enumeration with the feasible region's extreme points, the simplex method or GAMS. The solution to the linear problem is

The direction for this iteration is found by

The next point is found by

Step 3. Determine Step Length We evaluate z(4t, 4t) to obtain t.

Step 4. Set New Iteration Point Therefore

Step 5. Perform Optimality Test If the following equation yields a zero, the optimal solution has been found for the problem.

ITERATION 2

Step 1. Utilize new iteration point as the starting point

Step 2. Determine a search direction subject to the constraints defined earlier The solution to the linear problem is Failed to parse (syntax error): At this point, because the gradient of the new starting point is [0,0], the optimality test will yield 0; therefore [1,1] is the optimum solution to the non-linear problem. == Applications == === Image and Video Co-localization === [[File:Co-localization with Frank-Wolfe Algorithm.png|thumb|586x586px|'''Figure 2.''' The process of co-localization that captures common objects by using bounding boxes.<ref name="Joulin">Joulin, A., Tang, K., & Li, F.F. (2014). [https://link.springer.com/chapter/10.1007/978-3-319-10599-4_17 Efficient image and video co-localization with frank-wolfe algorithm.] ''European Conference on Computer Vision'' (pp. 253-268). Springer, Cham.</ref>|alt=]] Co-localization is one of the programming methods in the machine learning field used to find a common object within a set of images or video frames. The process of co-localization involves first generating multiple candidate bounding boxes for each image/ video frame, then identifying the specific box in each image/ video frame that jointly contains the common object.<ref name="Harzallah">Harzallah, H., Jurie, F., & Schmid, C. (2009). [https://ieeexplore.ieee.org/document/5459257 Combining efficient object localization and image classification.] ''2009 IEEE 12th international conference on computer vision'' (pp. 237-244). IEEE.</ref> This technique is extremely important for analyzing online data. By being able to recognize the common objects, visual concepts can be tagged into specific labels, thus leveraging useful data for other marketing and product design purposes. Such development is especially useful for image and video hosting services such as Flickr and YouTube. The formulation of co-localization using the Frank-Wolfe algorithm involves minimizing a standard quadratic programming problem and reducing it to a succession of simple linear integer problems. The benefit of the algorithm is that it allows many more images to be analyzed simultaneously and more bounding boxes are generated in each image/ video frame for higher accuracy. By relaxing the discrete non-convex set to its convex hull, the constraints that define the original set can now be separated from one image/ video to another. In the case of co-localizing videos, the constraints use the shortest-path algorithm to reach optimization.<ref name="Joulin" /> === Traffic Assignment Problem === A typical transportation network, also known as flow network, is considered as a convex programming problem with linear constraints. With the vertices called nodes and the edges called arcs, a directed graph called a network is formed and the flow is monitored via the movement from one node, origin (O), to another node, destination (D) along the arc. The objective of this programming problem is to achieve user-equilibrium, which means system flow is set optimal without incurring unnecessarily high cost for users. The use of the Frank-Wolfe algorithm in this case means that the traffic assignment problem can be applied to nonlinear multicommodity network flow with flow dependent cost.<ref name="LeBlanc">LeBlanc, L., Morlok, E.K., & Pierskalla, W.P. (1975). [https://www.sciencedirect.com/science/article/abs/pii/0041164775900301 An efficient approach to solving the road network equilibrium traffic assignment problem.] ''Transportation Research'', 9, 309-318.</ref> The algorithm is also able to process a larger set of constraints with more O-D pairs. Since the Frank-Wolfe algorithm uses a descent method to search for the direction of extreme points, the technique only recognizes the sequence of the shortest route problems. Therefore, the Frank-Wolfe algorithm is known to solve the traffic assignment problem on networks with several hundred nodes efficiently with the minimal number of iterations. Due to the algorithm’s effectiveness in optimizing traffic assignment problem, it has been widely applied in many fields other than the estimation of the total demand for travel via road between a set of roads and zones. These applications include data-driven computer networks, fluids in pipes, currents in an electrical circuit and more. === Deep Neural Network Training === Although [[Stochastic gradient descent|Stochastic Gradient Descent]] (SGD) is still the conventional machine learning technique for deep learning, the Frank-Wolfe algorithm has been proven to be applicable for training neural networks as well. Instead of taking projection steps via SGD, the Frank-Wolfe algorithm adopts the simple projection-free first-order function by using the linear minimization oracle (LMO) approach for constrained optimization <ref name="Pokutta">Pokutta, S., Spiegel, C., & Zimmer, M. (2020). [https://arxiv.org/abs/2010.07243 Deep neural network training with Frank-Wolfe.] ''arXiv preprint arXiv:2010.07243''.</ref>: <math display=block>\theta_{t+1} = \theta_t + \alpha(\nu_t - \theta_t)

with [0,1] and being

Through LMO, non-unique solutions to the above equation that were deemed impossible by gradient-descent-based optimization would now be possible. In addition to the linearization approach, sparsity regularization is also applied to the scaling factors in the optimization of deep neural networks.[4] By forcing some factors to zero, the highly complex computational process of training deep neural networks is simplified. Since the other major characteristics of Frank-Wolfe algorithm is describing solutions with sparse structures, deep neural network training benefits greatly from the application of this algorithm.

Conclusion

The Frank-Wolfe algorithm is one of the key techniques in the machine learning field. The algorithm features favor linear minimization, simple implementation, and low complexity. By using the simplex method that requires little alteration to the gradient and interpolation in each iteration, the optimization of the linearized quadratic objective function can be performed in a significantly shorter runtime and larger datasets can be processed more efficiently. Therefore, even though the convergence rate of Frank-Wolfe algorithm is slow due to the search for naive descent directions in each iteration, the benefits that the algorithm brings still outweigh the disadvantages. With the main characteristics of being projection-free and the capability of producing solutions for sparse structures, the algorithm gained popularity with many use cases in machine learning.

References

  1. 1.0 1.1 Frank, M. and Wolfe, P. (1956). An algorithm for quadratic programming. Naval Research Logistics Quarterly, 3(1-2):95–110.
  2. Combettes, C.W.. and Pokutta, S. (2020). Boosting Frank-Wolfe by chasing gradients. International Conference on Machine Learning (pp. 2111-2121). PMLR.
  3. Jaggi, M. (2013). Revisiting Frank-Wolfe: Projection-free sparse convex optimization. Inerational Conference on Machine Learning (pp.427-435). PMLR.
  4. Huang, Z., and Wang, N. (2018). Data-driven sparse structure selection for deep neural networks. Proceedings of the European conference on computer vision (pp.304-320). ECCV.