Interior-point method for LP: Difference between revisions

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search
No edit summary
No edit summary
Line 31: Line 31:
('''Dual Problem formulation''') <br>
('''Dual Problem formulation''') <br>
→ maximize <math> b^{T}y </math>  Subject to: <math> A^{T}y + \lambda  = c </math> and </math> \lambda \geq 0 </math> ...................(2)<br>
→ maximize <math> b^{T}y </math>  Subject to: <math> A^{T}y + \lambda  = c </math> and </math> \lambda \geq 0 </math> ...................(2)<br>
'λ' vector introduced represents the slack variables.
'λ' vector introduced represents the slack variables.<br>
Now we use the "Barrier Logarithmic" function and form 2 Lagrangian equations for primal and dual forms mentioned above:<br>
Now we use the "Barrier Logarithmic" function and form 2 Lagrangian equations for primal and dual forms mentioned above:<br>
Lagrange function for Primal : <math> L_{p}(x,y) = c^{T}\cdot x + \mu \cdot \sum_{j=1}^{n}log(x_{j}) - y^{T}\cdot (Ax - b) </math> ......................................(3)<br>
Lagrange function for Primal : <math> L_{p}(x,y) = c^{T}\cdot x + \mu \cdot \sum_{j=1}^{n}log(x_{j}) - y^{T}\cdot (Ax - b) </math> ......................................(3)<br>
Lagrange function for Dual  : <math> L_{d}(x,y,\lambda ) = b^{T}\cdot y + \mu \cdot \sum_{j=1}^{n}log(\lambda _{j}) - x^{T}\cdot (A^{T}y -\lambda - c) </math> .........(4)<br>
Lagrange function for Dual  : <math> L_{d}(x,y,\lambda ) = b^{T}\cdot y + \mu \cdot \sum_{j=1}^{n}log(\lambda _{j}) - x^{T}\cdot (A^{T}y -\lambda - c) </math> .........(4)<br>


Taking the partial derivatives of L<sub>p</sub> and L<sub>d</sub> with respect to variables 'x', 'λ' , 'y', and forcing these terms to zero, we get the following equations: <br>
<math> Ax = b </math>  and  <math> x \geq 0 </math> ..........................................(5)<br>
<math> A^{T}y + \lambda  = c </math> and </math> \lambda \geq 0 </math> .......................(6)<br>
<math> x_{j}\cdot \lambda _{j} = \mu </math> for ''j''= 1.......''n''<br>
where, μ is strictly positive scaler parameter. For each μ > 0, the vectors in the set {x(μ), y(μ), λ(μ)} satisfying equations





Revision as of 20:14, 13 November 2020

Authors: Tomas Lopez Lauterio, Rohit Thakur and Sunil Shenoy
Steward: Dr. Fengqi You and Akshay Ajagekar

Introduction

Linear programming problems seeks to optimize linear functions given linear constraints. There are several applications of linear programming including inventory control, production scheduling, transportation optimization and efficient manufacturing processes. Simplex method has been a very popular method to solve these linear programming problems and has served these industries well for a long time. But over the past 40 years, there have been significant number of advances in different algorithms that can be used for solving these types of problems in more efficient ways, especially where the problems become very large scale in terms of variables and constraints. In early 1980s Karmarkar (1984) published a paper introducing interior point methods to solve linear-programming problems. A simple way to look at differences between simplex method and interior point method is that a simplex method moves along the edges of a polytope towards a vertex having a lower value of the cost function, whereas an interior point method begins its iterations inside the polytope and moves towards the lowest cost vertex without regard for edges. This approach reduces the number of iterations needed to reach that vertex, thereby reducing computational time needed to solve the problem.

Lagrange Function

Before getting too deep into description of Interior point method, there are a few concepts that are helpful to understand. First key concept to understand is related to Lagrange function. Lagrange function incorporates the constraints into a modified objective function in such a way that a constrained minimizer (x*) is connected to an unconstrained minimizer {x*, λ*} for the augmented objective function L(x,λ), where the augmentation is achieved with 'p' Lagrange multipliers.
To illustrate this point, if we consider a simple an optimization problem: minimize f(x) subject to: A·x = b, where A ε Rpxn is assumed to have a full row rank Lagrange function can be laid out as:


where, 'λ' introduced in this equation is called Lagrange Multiplier.

Newton's Method

Another key concept to understand is regarding solving linear and non-linear equations using Newton's methods. Assume you have an unconstrained minimization problem in the form:
minimize g(x) , where g(x) is a real valued function with n variables.
A local minimum for this problem will satisfy the following system of equations:

The Newton's iteration looks like:


Theory and Problem Formulation

We first start forming a primal-dual pair of linear programs and use the "Lagrangian function" and "Barrier function" methods to convert the constrained problems into unconstrained problems. These unconstrained problems are then solved using Newton's method as shown above.
Consider a combination of primal-dual problem below:
(Primal Problem formulation)
→ minimize Subject to: and .......................................(1)
(Dual Problem formulation)
→ maximize Subject to: and </math> \lambda \geq 0 </math> ...................(2)
'λ' vector introduced represents the slack variables.
Now we use the "Barrier Logarithmic" function and form 2 Lagrangian equations for primal and dual forms mentioned above:
Lagrange function for Primal : ......................................(3)
Lagrange function for Dual  : .........(4)

Taking the partial derivatives of Lp and Ld with respect to variables 'x', 'λ' , 'y', and forcing these terms to zero, we get the following equations:
and ..........................................(5)
and </math> \lambda \geq 0 </math> .......................(6)
for j= 1.......n

where, μ is strictly positive scaler parameter. For each μ > 0, the vectors in the set {x(μ), y(μ), λ(μ)} satisfying equations





Numerical Example

Applications

Conclusion

References