Column generation algorithms: Difference between revisions

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search
No edit summary
No edit summary
 
(84 intermediate revisions by 3 users not shown)
Line 1: Line 1:
Author: Lorena Garcia Fernandez (lgf572)
Author: Lorena Garcia Fernandez (lgf572) (SysEn 5800 Fall 2020)


= '''Introduction''' =
== Introduction ==
Column Generation techniques have the scope of setting up your Mixed integer Programming (MIP) problem by generating only the variables that will have an influence on the objective function. This is important for big problems with many variables where the formulation with these techniques would simplify the problem formulation, since not all the possibilities need to be listed.
Column Generation techniques have the scope of solving large linear optimization problems by generating only the variables that will have an influence on the objective function. This is important for big problems with many variables where the formulation with these techniques would simplify the problem formulation, since not all the possibilities need to be explicitly listed.<ref>Desrosiers, Jacques & Lübbecke, Marco. (2006). A Primer in Column Generation.p7-p14 10.1007/0-387-25486-2_1. </ref>


= '''Theory, methodology, and algorithmic discussions''' =
== Theory, methodology and algorithmic discussions ==
'''''Theory'''''
'''''Theory'''''


The way this method work is as follows; first, the original problem that is being solved needs to be split into two problems: the master problem and the sub-problem.
The way this method work is as follows; first, the original problem that is being solved needs to be split into two problems: the master problem and the sub-problem.


* The master problem is the original column-wise (i.e: one column at a time) formulation of the problem with only a subset of variables being considered.
* The master problem is the original column-wise (i.e: one column at a time) formulation of the problem with only a subset of variables being considered.<ref>
AlainChabrier, Column Generation techniques, 2019 URL: https://medium.com/@AlainChabrier/column-generation-techniques-6a414d723a64
</ref>


* The sub-problem is a new problem created to identify a new promising variable. The objective function of the sub-problem is the reduced cost of the new variable with respect to the current dual variables, and the constraints require that the variable obeys the naturally occurring constraints. The subproblem is also referred to as the RMP or “restricted master problem”. From this we can infer that this method will be a good fit for problems whose constraint set admit a natural breakdown (i.e: decomposition) into sub-systems representing a well understood combinatorial structure.
* The sub-problem is a new problem created to identify a new promising variable. The objective function of the sub-problem is the reduced cost of the new variable with respect to the current dual variables, and the constraints require that the variable obeys the naturally occurring constraints. The subproblem is also referred to as the RMP or “restricted master problem”. From this we can infer that this method will be a good fit for problems whose constraint set admit a natural breakdown (i.e: decomposition) into sub-systems representing a well understood combinatorial structure.<ref>
AlainChabrier, Column Generation techniques, 2019 URL: https://medium.com/@AlainChabrier/column-generation-techniques-6a414d723a64
</ref>


To execute that decomposition from the original problem into Master and subproblems there are different techniques. The theory behind this method relies on the Dantzig-Wolfe decomposition.
To execute that decomposition from the original problem into Master and subproblems there are different techniques. The theory behind this method relies on the Dantzig-Wolfe decomposition.<ref>Dantzig-Wolfe decomposition. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Dantzig-Wolfe_decomposition&oldid=50750</ref>


In summary, when the master problem is solved, we are able to obtain dual prices for each of the constraints in the master problem. This information is then utilized in the objective function of the subproblem. The subproblem is solved. If the objective value of the subproblem is negative, a variable with negative reduced cost has been identified. This variable is then added to the master problem, and the master problem is re-solved. Re-solving the master problem will generate a new set of dual values, and the process is repeated until no negative reduced cost variables are identified. The subproblem returns a solution with non-negative reduced cost, we can conclude that the solution to the master problem is optimal.
In summary, when the master problem is solved, we are able to obtain dual prices for each of the constraints in the master problem. This information is then utilized in the objective function of the subproblem. The subproblem is solved. If the objective value of the subproblem is negative, a variable with negative reduced cost has been identified. This variable is then added to the master problem, and the master problem is re-solved. Re-solving the master problem will generate a new set of dual values, and the process is repeated until no negative reduced cost variables are identified. The subproblem returns a solution with non-negative reduced cost, we can conclude that the solution to the master problem is optimal.<ref>Wikipedia, the free encyclopeda. Column Generation. URL: https://en.wikipedia.org/wiki/Column_generation</ref>


'''''Methodology'''''<ref>L.A. Wolsey, Integer programming. Wiley,Column Generation Algorithms p185-p189,1998</ref>
[[File:Column Generation.png|thumb|468x468px|Column generation schematics<ref name=":4">GERARD. (2005). Personnel and Vehicle scheduling, Column Generation, slide 12. URL: https://slideplayer.com/slide/6574/</ref>]]
Consider the problem in the form:


(IP)   
<math>z=max\left \{\sum_{k=1}^{K}c^{k}x^{k}:\sum_{k=1}^{K}A^{k}x^{k}=b,x^{k}\epsilon X^{k}\; \; \; for\; \; \; k=1,...,K  \right \}</math>


'''''Methodology in detail'''''


To illustrate the algorithm, we will use a common example: an algorithm for one-dimensional cutting stock problem.
Where <math>X^{k}=\left \{x^{k}\epsilon Z_{+}^{n_{k}}: D^{k}x^{k}\leq d^{_{k}}  \right \}</math> for <math>k=1,...,K</math>. Assuming that each set <math>X^{k}</math> contains a large but finite set of points <math>\left \{ x^{k,t} \right \}_{t=1}^{T_{k}}</math>, we have that <math>X^{k}=</math>:
 
''<u>Problem Overview</u>''
 
Given:
 
A set of item types I,
 
For every item type i ∈ I, its length Li and the number of pieces to be produced Ri ,
 
The length W of the starting objects to be cut,
 
Objective:
 
To find the minimum number of objects needed to satisfy the demand of all item types.
 
Model:
 
The problem can be modeled as follows:
 
<math>\begin{matrix} \\ \min(y)\sum_{s=1}^Sy_s  \\        \      s.t. \sum_kN_{ks}y_s\geq J_k    \forall    k\in K \\ y_s\in \Zeta_+\forall s\in S \end{matrix}
</math>
 
where:
 
S: set of all possible cutting patterns that can be used to obtain item types in I from the original objects of length W;
 
Nks : number of pieces of type k ∈ K in the cutting pattern s ∈ S .
 
ys : number of original objects to be cut with pattern s ∈ S.


<math>\left \{ x^{k}\epsilon R^{n_{k}}:x^{k}=\sum_{t=1}^{T_{k}}\lambda _{k,t}x^{k,t},\sum_{t=1}^{T_{k}}\lambda _{k,t}=1,\lambda _{k,t}\epsilon \left \{ 0,1 \right \}for \; \; k=1,...,K  \right \}</math>


The algorithm to solve this problem is built on the solution of the continuous relaxation of the above model, i.e., the model obtained by replacing constraints
Note that, on the assumption that each of the sets <math>X^{k}=</math> is bounded for <math>k=1,...,K</math> the approach will involve solving an equivalent problem of the form as below:


<math>\begin{matrix} \\ y_s\in \Zeta_+\forall s\in S \end{matrix}
<math>max\left \{ \sum_{k=1}^{K}\gamma ^{k}\lambda ^{k}: \sum_{k=1}^{K}B^{k}\lambda ^{k}=\beta ,\lambda ^{k}\geq 0\; \; integer\; \; for\; \; k=1,...,K \right \}</math>
</math>


with constraints...
where each matrix <math>B^{k}</math> has a very large number of columns, one for each of the feasible points in <math>X^{k}</math>, and each vector <math>\lambda ^{k}</math> contains the corresponding variables.


<math>\begin{matrix} \\    y_s\in    \\\Re_+  \forall s\in S    \end{matrix}
</math>


Sometimes |S| could be so large that enumerating all patterns would not be practical. For this purpose, the column generation below can be used:
Now, substituting for <math>x^{k}=</math> leads to an equivalent ''IP Master Problem (IPM)'':
Step 0: initialize the problem
Generate a subset of patterns S ′ for which the problem has a solution that is feasible (a typical initialization is that of starting with the |I| single-item cutting patterns)


Step 1: formulation and solution of the master problem
(IPM)
Solve the master problem (restricted to the patterns (i.e: variables) ysj with s ∈ S ′ )
<math>\begin{matrix}
z=max\sum_{k=1}^{K}\sum_{t=1}^{T_{k}}\left(c^{k}x^{k,t}\right )\lambda _{k,t} \\      \sum_{k=1}^{K}\sum_{t=1}^{T_{k}}\left ( A^{k}x^{k,t} \right )\lambda _{k,t}=b\\
\sum_{t=1}^{T_{k}}\lambda _{k,t}=1\; \; for\; \; k=1,...,K \\
\lambda _{k,t}\epsilon \left \{ 0,1 \right \}\; \; for\; \; t=1,...,T_{k}\; \; and\; \; k=1,...,K.
\end{matrix}</math>


<math>\begin{matrix} \\ \min(y)\sum_{s=1}^(S^') y_s  \\        \      s.t. \sum_kN_{ks}y_s\geq J_k    \forall    k\in K \\ y_s\in \Zeta_+\forall s\in S^' \end{matrix}
To solve the Master Linear Program, we use a column generation algorithm. This is in order to solve the linear programming relaxation of the Integer Programming Master Problem, called the ''Linear Programming Master Problem (LPM)'':
</math>


By solving this problem one can obtain first a primal optimal solution '''''y∗''''' and then a dual optimal solution '''''u''''' such that '''''y∗''''' and '''''u''''' satisfy the complementary slackness condition (for example, this could be done with the simplex method).  
(LPM)
<math>\begin{matrix}
z^{LPM}=max\sum_{k=1}^{K}\sum_{t=1}^{T_{k}}\left ( c^{k}x^{k,t} \right )\lambda _{k,t}\\
\sum_{k=1}^{K}\sum_{t=1}^{T_{k}}\left ( A^{k}x^{k,t} \right )\lambda _{k,t}=b \\
\sum_{t=1}^{T_{k}}\lambda _{k,t}=1\; \;for\; \; k=1,...,K \\
\lambda _{k,t} \geq 0\; \; for\; \; t=1,...,T_{k},\; k=1,...,K
\end{matrix}</math>


Step 2: solution of the subproblem  
Where there is a column <math>\begin{pmatrix}
Or in other words, the next step is to find the solution of the following integer linear programming problem (called subproblem or slave problem) with |K| variables and one constraint:
c^{k}x\\
A^{k}x\\
e_{k}
\end{pmatrix}</math> for each ''<math>x</math>'' ''<math display="inline">\in</math> <math display="inline">X^{k}</math>''. On the next steps of this method, we will use <math>\left \{ \pi _{i} \right \}_{i=1}^{m}</math> as the dual variables associated with the joint constraints, and <math>\left \{ \mu_{k} \right \}_{k=1}^{K}</math> as dual variables for the second set of constraints.The latter are also known as convexity constraints.
The idea is to solve the linear program by the primal simplex algorithm. However, the pricing step of choosing a column to enter the basis must be modified because of the very big number of columns in play. Instead of pricing the columns one at a time, the question of finding a column with the biggest reduced price is itself a set of <math>K</math> optimization problems.


<math>\begin{matrix} \\ \max(z)\sum_{k=1}^K u_k*z_k  \\        \      s.t. \sum_kL_kz_k \leq M  \\ z_k\in \Zeta_+\forall k\in K \end{matrix}
</math>


''Initialization:'' we suppose that a subset of columns (at least one for each <math>k</math>) is available, providing a feasible ''Restricted Linear Programming Master Problem'':


(RLPM)
<math>\begin{matrix}
z^{LPM}=max\tilde{c}\tilde{\lambda} \\
\tilde{A}\tilde{\lambda }=b \\
\tilde{\lambda }\geq 0
\end{matrix}</math>


For this problem, the optimal solution would be:
<math>\begin{matrix}  \\z^*\in \Zeta_+\end{matrix}
</math>
Step 3: optimality check
As it was previously highlighted, it is necessary to conduct an optimality check in order to decide if the optimal solution has been reached or not. Below is the condition:
If...
<math>\begin{matrix} \\\sum_Kz_k(optimal)u_k(optimal) \leq 1,  \end{matrix}
</math>


where <math>\tilde{b}=\begin{pmatrix}
b\\
1\\
\end{pmatrix}</math>, <math>\tilde{A}</math> is generated by the available set of columns and <math>\tilde{c}\tilde{\lambda }</math> are the corresponding costs and variables. Solving the RLPM gives an optimal primal solution <math>\tilde{\lambda ^{*}}</math> and an optimal dual solution <math>\left ( \pi ,\mu  \right )\epsilon\;  R^{m}\times R^{k}</math>




then STOP.
''Primal feasibility:'' Any feasible solution of ''RLMP'' is feasible for ''LPM''. More precisely, <math>\tilde{\lambda^{*} }</math> is a feasible solution of ''LPM'', and hence <math>\tilde{z}^{LPM}=\tilde{c}\tilde{\lambda ^{*}}=\sum_{i=1}^{m}\pi _{i}b_{i}+\sum_{k=1}^{K}\mu _{k}\leq z^{LPM}</math>


'''''y∗''''' is an ''optimal solution'' of the full continuous relaxation (including all patterns in '''S'''). Otherwise, ''update the master problem'' by including in '''''S′''''' the pattern γ defined by Nks= z∗k (this means that column '''''z∗''''' has to be included in the constraint matrix) and go to Step 1.
''Optimality check for LPM:'' It is required to check whether <math>\left ( \pi ,\mu  \right )</math>  is dual feasible for ''LPM''. This means checking for each column, that is for each <math>k</math>, and for each <math>x\; \epsilon \; X^{k}</math> if the reduced price <math>c^{k}x-\pi A^{k}x-\mu _{k}\leq 0</math>. Rather than examining each point separately, we treat all points in <math>X^{k}</math> implicitly, by solving an optimization subproblem:


Finally, one has to go from the optimal solution of the continuous relaxation to a heuristic (i.e., not necessarily optimal but hopefully good) solution of the original problem with integrality constraints. This can be done in at least two different ways:
<math>\zeta _{k}=max\left \{ \left (c^{k}-\pi A^{k}  \right )x-\mu _{k}\; :\; x\; \epsilon \; X^{k}\right \}.</math>


By rounding up the entries of '''''y∗''''' (this is a good choice if these entries are large: 335.4 is not very different from 336...); Is worth noticing that rounding down is not allowed due to the fact that it would provide infeasible integer solution;


By applying an integer linear programming method (for instance the Branch-and Bound) to the last master problem that was generated; we are taking a step which is equivalent to solving the original problem (with integrality constraints)but restricted to the “good” patterns (those in '''''S′''''' ) found in the above steps
''Stopping criteria:'' If <math>\zeta _{k}> 0</math> for <math>k=1,...,K</math> the solution <math>\left ( \pi ,\mu  \right )</math> is dual feasible for ''LPM'', and hence <math>z^{LPM}\leq \sum_{i=1}^{m}\pi _{i}b_{i}+\sum_{k=1}^{K}\mu _{k}</math>. As the value of the primal feasible solution <math>\tilde{\lambda }</math> equals that of this upper bound, <math>\tilde{\lambda }</math> is optimal for ''LPM''.




= '''Numerical example: The Cutting Stock problem''' =
''Generating a new column:'' If <math>\zeta _{k}> 0</math> for some <math>k</math>, the column corresponding to the optimal solution <math>\tilde{x}^{k}</math> of the subproblem has a positive reduced price. Introducing the column <math>\begin{pmatrix}
Suppose we want to solve a numerical example of the cutting stock problem that we have discussed during the theory section of this wiki, specifically a one-dimensional cutting stock problem
c^{k}x\\
A^{k}x\\
e_{k}
\end{pmatrix}</math> leads then to a Restricted Linear Programming Master Problem that can be easily reoptimized (e.g., by the primal simplex algorithm)


== Numerical example: The Cutting Stock problem<ref>L.A. Wolsey, Integer programming. Wiley,Column Generation Algorithms p185-p189,1998The Cutting Stock problem</ref> ==


Suppose we want to solve a numerical example of the cutting stock problem, specifically a one-dimensional cutting stock problem.


''<u>Problem Overview</u>''
''<u>Problem Overview</u>''


A company produces steel bars with diameter 45 millimeters and length 33 meters. The company also takes care of cutting the bars for their different customers, who each require different lengths. At the moment, the following demand forecast is expected and must be satisfied:  
A company produces steel bars with diameter <math>45</math> millimeters and length <math>33</math> meters. The company also takes care of cutting the bars for their different customers, who each require different lengths. At the moment, the following demand forecast is expected and must be satisfied:  
{| class="wikitable"
{| class="wikitable"
|+
|+
Line 156: Line 138:
'''Sets'''
'''Sets'''


K = {1, 2, 3, 4, 5}: set of item types;
<math>K=\left \{ 1,2,3,4,5 \right \}</math>: set of item types;


S: set of patterns (i.e., possible ways) that can be adopted to cut a given bar into portions of the need lengths.
''<math display="inline">S</math>:'' set of patterns (i.e., possible ways) that can be adopted to cut a given bar into portions of the need lengths.


'''Parameters'''
'''Parameters'''


M: bar length (before the cutting process)
<math display="inline">M</math>: bar length (before the cutting process);


L<sub>k</sub> : length of item k K;
<math display="inline">L_k</math>'':'' length of item ''<math display="inline">k</math>'' ''<math display="inline">\in</math> <math display="inline">K</math>'';


R<sub>s</sub> : number of pieces of type k K required;
<math display="inline">R_s</math> : number of pieces of type ''<math display="inline">k</math>'' ''<math display="inline">\in</math> <math display="inline">K</math>'' required;


N<sub>ks</sub> : number of pieces of type k K in pattern s S
<math display="inline">N_{k,s}</math> : number of pieces of type ''<math display="inline">k</math>'' ''<math display="inline">\in</math> <math display="inline">K</math>'' in pattern ''<math display="inline">s</math>'' ''<math display="inline">\in</math> <math display="inline">S</math>''.


'''Decision variables'''
'''Decision variables'''


Y<sub>s</sub> : number of bars that should be portioned using pattern s S  
<math display="inline">Y_s</math> : number of bars that should be portioned using pattern ''<math display="inline">s</math>'' ''<math display="inline">\in</math> <math display="inline">S</math>''.


'''Model'''  
'''Model'''  


<math>\begin{matrix} \\ \min(y)\sum_{s=1}^Sy_s  \\        \      s.t. \sum_kN_{ks}y_s\geq J_k    \forall    k\in K \\ y_s\in \Zeta_+\forall s\in S \end{matrix}
<math>\begin{matrix}\min(y)\sum_{s=1}^Sy_s  \\        \      s.t. \sum_kN_{ks}y_s\geq J_k    \forall    k\in K \\ y_s\in \Zeta_+\forall s\in S \end{matrix}
   
   
</math>
</math>
Line 182: Line 164:
''<u>Solving the problem</u>''
''<u>Solving the problem</u>''


The model assumes the availability of the set K and the parameters N<sub>ks</sub> . To generate this data, you would have to list all possible cutting patterns. However, the number of possible cutting patterns is a big number. This is why a direct implementation of the model above is not partical in real-world problems. In this case is when it makes sense to solve the continuous relaxation of the above model. This is because, in reality, the demand figures are so high that the number of bars to cut is also a large number, and therefore a good solution can be determined by rounding up to the next integer each variable y<sub>s</sub> found by solving the continuous relaxation. In addition to that, the solution of the relaxed problem will become the starting point for the application of an exact solution method (for instance, the Branch-and Bound).<blockquote><u>''Key take-away: In the next steps of this example we will analyze how to solve the continuous relaxation of the model.''</u></blockquote>As a starting point, we need any feasible solution. Such a solution can be constructed as follows:
The model assumes the availability of the set ''<math display="inline">K</math>'' and the parameters <math display="inline">N_{k,s}</math> . To generate this data, you would have to list all possible cutting patterns. However, the number of possible cutting patterns is a big number. This is why a direct implementation of the model above is not practical in real-world problems. In this case is when it makes sense to solve the continuous relaxation of the above model. This is because, in reality, the demand figures are so high that the number of bars to cut is also a large number, and therefore a good solution can be determined by rounding up to the next integer each variable <math>y_s
</math>found by solving the continuous relaxation. In addition to that, the solution of the relaxed problem will become the starting point for the application of an exact solution method (for instance, the Branch-and Bound).<blockquote><u>''Key take-away: In the next steps of this example we will analyze how to solve the continuous relaxation of the model.''</u></blockquote>As a starting point, we need any feasible solution. Such a solution can be constructed as follows:


# We consider any single-item cutting patterns, i.e., |K| configurations, each containing N<sub>ks</sub> = ⌊W/L<sub>k</sub>⌋ pieces of type k;
# We consider any single-item cutting patterns, i.e., <math>\|K\|
# Set y<sub>k</sub> = ⌈R<sub>s</sub>/N<sub>ks</sub>⌉ for pattern k (where pattern k is the pattern containing only pieces of type k).
</math> configurations, each containing <math display="inline">{\textstyle N_{k,s} } = \llcorner \frac{W}{L_k}\lrcorner
</math> pieces of type <math>k
</math>;
# Set <math display="inline">{\textstyle y_{k}} = \llcorner \frac{R_s}{N_{k,s}}\lrcorner
</math> for pattern <math>k
</math> (where pattern <math>k
</math> is the pattern containing only pieces of type <math>k
</math>).


This solution could also be arrived to by applying the simplex method to the model (without integrality constraints), considering only the decision variables that correspond to the above single-item patterns:  
This solution could also be arrived to by applying the simplex method to the model (without integrality constraints), considering only the decision variables that correspond to the above single-item patterns:  


''min        y<sub>1</sub>  +   y<sub>2</sub>  +   y<sub>3</sub>  +   y<sub>4</sub>  + y<sub>5</sub>''
<math>\begin{align}
 
\text{min} & ~~  y_{1}+y_{2}+y_{3}+y_{4}+y_{5}\\
''s.t   15y<sub>1</sub>                                                       ≥  144''
\text{s.t} & ~~ 15y_{1} \ge 144\\
\ & ~~ 6y_{2} \ge 105\\
\ & ~~ 6y_{3} \ge 72\\
\ & ~~ 6y_{4} \ge 30\\
\ & ~~ 3y_{5} \ge 24\\
\ & ~~ y_{1},y_{2},y_{3},y_{4},y_{5} \ge 0\\
\end{align}</math>


''6y<sub>2</sub>                                          ≥  105<br />                                6y<sub>3</sub>                            ≥  72''
In fact, if we solve this problem (for example, use CPLEX solver in GAMS) the solution is as below:
 
''6y<sub>4</sub>                ≥  30''
 
''3y<sub>5</sub>          ≥  24''
 
''y<sub>1 ,</sub> y<sub>2 ,</sub> y<sub>3 ,</sub> y<sub>4 ,</sub> y<sub>5</sub>      ≥  0''
 
In fact, if we solve this problem (for example, use CPLEX solver in GAMS) the solution is as below :  
{| class="wikitable"
{| class="wikitable"
|Y1
|Y1
Line 218: Line 214:
|24
|24
|}
|}
Next, a new possible pattern (number 6) will be consider. This pattern contains only one piece of item type number 5. So the question is if the new solution would remain optimal if this new pattern was allowed. Duality helps answer ths question. At every iteration of the simplex method, the outcome is a feasible basic solution (corresponding to some basis B) for the primal problem and a dual solution (the multipliers u T = c T BB−1 ) that satisfy the complementary slackness conditions. (Note: the dual solution will be feasible only when the last iteration is reached)  
Next, a new possible pattern (number <math>6</math>) will be considered. This pattern contains only one piece of item type number <math>5</math>. So the question is if the new solution would remain optimal if this new pattern was allowed. Duality helps answer ths question. At every iteration of the simplex method, the outcome is a feasible basic solution (corresponding to some basis <math>B</math>) for the primal problem and a dual solution (the multipliers <math>u^{t}=c^{t}BB^{-1}</math>) that satisfy the complementary slackness conditions. (Note: the dual solution will be feasible only when the last iteration is reached)  


The inclusion of new pattern "6" corresponds to including a new variable in the primal problem, with objective cost 1 (as each time pattern 6 is chosen, one bar is cut) and corresponding to the following column in the constraint matrix:  
The inclusion of new pattern <math>6</math> corresponds to including a new variable in the primal problem, with objective cost <math>1</math> (as each time pattern <math>6</math> is chosen, one bar is cut) and corresponding to the following column in the constraint matrix:  


<math>D_6= \begin{bmatrix}
<math>D_6= \begin{bmatrix}
Line 228: Line 224:
\ 0 \\  
\ 0 \\  
\ 1 \\  
\ 1 \\  
\end{bmatrix}
\end{bmatrix}</math>
   
   
</math>


These variables create a new dual constraint. We then have to check if this new constraint is violated by the current dual solution  ( or in other words, ''if the reduced cost of the new variable with respect to basis B is negative)''
These variables create a new dual constraint. We then have to check if this new constraint is violated by the current dual solution  (or in other words, ''if the reduced cost of the new variable with respect to basis <math>B</math> is negative)''


The new dual constraint is:
The new dual constraint is:<math>1\times u_{1}+0\times u_{2}+0\times u_{3}+0\times u_{4}+1\times u_{5}\leq 1</math>


''1*u<sub>1</sub>  +  0*u<sub>2</sub>  +  0*u<sub>3</sub>  +  0*u<sub>4</sub>  + 1*u<sub>5</sub>  ≤ 1''
The solution for the dual problem can be computed in different software packages, or by hand. The example below shows the solution obtained with GAMS for this example:


The solution for the dual problem can be computed in different software packages, or by hand. The example below shows the solution obtained with GAMS for this example:
(Note the solution for the dual problem would be: <math>u=c_{T}^{B}B^{-1}</math>)


(Note the solution for the dual problem would be:


<math>\begin{matrix} \\  u = c_T^B B^-1 \end{matrix}
</math>
{| class="wikitable"
{| class="wikitable"
|Dual  variable
|Dual  variable
Line 264: Line 255:
|0.333
|0.333
|}
|}
Since 0.2+1 = 1.2 > 1, the new constraint is violated.
Since <math>0.2+1=1.2> 1</math>, the new constraint is violated.


This means that the current primal solution (in which the new variable is y<sub>6</sub> = 0) may not be optimal anymore (although it is still feasible). The fact that the dual constraint is violated means the associated primal variable has negative reduced cost:  
This means that the current primal solution (in which the new variable is <math>y_{6}=0</math>) may not be optimal anymore (although it is still feasible). The fact that the dual constraint is violated means the associated primal variable has negative reduced cost:  


the norm of c<sub>6</sub> = c<sub>6</sub> - u<sup>T</sup>D<sub>6</sub> = 1 - 0.4 = 0.6
the norm of <math>c_6 = c_6-u^TD_6=1-0.4=0.6</math> 


<nowiki>****</nowiki>update formula*****
To help us solve the problem, the next step is to let <math>y_{6}</math> enter the basis. To do so, we modify the problem by inserting the new variable as below:


To help us solve the problem, the next step is t let y<sub>6</sub> enter the basis. To do so, we modify the problem by inserting the new variable as below:
<math>\begin{align}
\text{min} & ~~  y_{1}+y_{2}+y_{3}+y_{4}+y_{5}+y_{6}\\
\text{s.t} & ~~ 15y_{1} +y_{6}\ge 144\\
\ & ~~ 6y_{2} \ge 105\\
\ & ~~ 6y_{3} \ge 72\\
\ & ~~ 6y_{4} \ge 30\\
\ & ~~ 3y_{5}+y_{6} \ge 24\\
\ & ~~ y_{1},y_{2},y_{3},y_{4},y_{5},y_{6} \ge 0\\
\end{align}</math>


''min        y<sub>1</sub>  +  y<sub>2</sub>  +  y<sub>3</sub>  +  y<sub>4</sub>  +  y<sub>5</sub> +  y<sub>6</sub>''


''s.t    15y<sub>1</sub>                                                    +  y<sub>6</sub>  ≥  144''
If this problem is solved with the simplex method, the optimal solution is found, but restricted only to patterns <math>1</math> to <math>6</math>. If a new pattern is available, a decision should be made whether this new pattern should be used or not by proceeding as above. However, the problem is how to find a pattern (i.e., a variable; i.e, a column of the matrix) whose reduced cost is negative (i.e., which will mean it is convenient to include it in the formulation). At this point one can notice that number of possible patterns exponentially large,and all the patterns are not even known explicitly. The question then is:
 
''6y<sub>2</sub>                                                  ≥  105<br />                                6y<sub>3</sub>                                      ≥  72''
 
''6y<sub>4</sub>                          ≥  30''
 
''3y<sub>5</sub>    +  y<sub>6</sub>  ≥  24''
 
''y<sub>1 ,</sub> y<sub>2 ,</sub> y<sub>3 ,</sub> y<sub>4 ,</sub> y<sub>5 ,</sub> y<sub>6</sub>    ≥  0''
 
If this problem is solved with the simplex method, the optimal solution is found, but restricted only to patterns 1 to 6. If a new pattern is available, a decision should be made whether this new pattern should be used or not by proceeding as above. However, the problem is how to find a pattern (i.e., a variable; i.e, a column of the matrix) whose reduced cost is negative (i.e., which will mean it is convenient to include it in the formulation). At this point one can notice that number of possible patterns exponentially large,and all the patterns are not even known explicitly. The question then is:


''Given a basic optimal solution for the problem in which only some variables are included, how can we find (if any exists) a variable with negative reduced cost (i.e., a constraint violated by the current dual solution)?''  
''Given a basic optimal solution for the problem in which only some variables are included, how can we find (if any exists) a variable with negative reduced cost (i.e., a constraint violated by the current dual solution)?''  
Line 292: Line 280:
This question can be transformed into an optimization problem: in order to see whether a variable with negative reduced cost exists, we can look for the minimum of the reduced costs of all possible variables and check whether this minimum is negative:
This question can be transformed into an optimization problem: in order to see whether a variable with negative reduced cost exists, we can look for the minimum of the reduced costs of all possible variables and check whether this minimum is negative:


c ̅  = 1-u^T z
<math>\bar{c}=1-u^Tz</math>


Because every column of the constraint matrix corresponds to a cutting pattern, and every entry of the column says how many pieces of a certain type are in that pattern. In order for z to be a possible column of the constraint matrix, the following condition must be satisfied:
Because every column of the constraint matrix corresponds to a cutting pattern, and every entry of the column says how many pieces of a certain type are in that pattern. In order for <math>z
</math> to be a possible column of the constraint matrix, the following condition must be satisfied:


<math>\begin{matrix} \\ \\ z_k\in \Zeta_+\forall k\in K  \\        \      \sum_kL_kz_k \leq M  \end{matrix}
<math display="inline">\begin{matrix}z_k\in \Zeta_+\forall k\in K  \\        \      \sum_kL_kz_k \leq M  \end{matrix}
   
   
</math>
</math>


And by so doin, it enables the conversion of the problem of finding a variable with negative reduced cost into the integer linear programming problem below:
And by so doing, it enables the conversion of the problem of finding a variable with negative reduced cost into the integer linear programming problem below:


<math>\begin{matrix} \\ \min\  \bar{c} = 1 - sum_{k=1}^K u_k*z_k  \\        \      s.t. \sum_kL_kz_k \leq M  \\ z_k\in \Zeta_+\forall k\in K \end{matrix}
<math>\begin{matrix}\min\  \bar{c} = 1 - sum_{k=1}^K u_k \times z_k  \\        \      s.t. \sum_kL_kz_k \leq M  \\ z_k\in \Zeta_+\forall k\in K \end{matrix}
   
   
</math>
</math>


which, in turn, would be equivalent to the below formulation (we just write the objective in maximization form and ignore the additive constant 1):
which, in turn, would be equivalent to the below formulation (we just write the objective in maximization form and ignore the additive constant <math>1</math>):


<math>\begin{matrix} \\ \max\sum_{k=1}^K u_k*z_k  \\        \      s.t. \sum_kL_kz_k \leq M  \\ z_k\in \Zeta_+\forall k\in K \end{matrix}</math>
<math>\begin{matrix} \max\sum_{k=1}^K u_k \times z_k  \\        \      s.t. \sum_kL_kz_k \leq M  \\ z_k\in \Zeta_+\forall k\in K \end{matrix}</math>






The coefficients z<sub>k</sub> of a column with negative reduced cost can be found by solving the above integer [[wikipedia:Knapsack_problem|"knapsack"]] problem (which is a traditional type of problem that we find in integer programming).
The coefficients <math>z_k
</math> of a column with negative reduced cost can be found by solving the above integer [[wikipedia:Knapsack_problem|"knapsack"]] problem (which is a traditional type of problem that we find in integer programming).


In our example, if we start from the problem restricted to the five single-item patterns, the above problem reads as:
In our example, if we start from the problem restricted to the five single-item patterns, the above problem reads as:


''max        0.067z<sub>1</sub>  +   0.167z<sub>2</sub>  +   0.167z<sub>3</sub>  +   0.167z<sub>4</sub>  + z<sub>5</sub>''
<math>\begin{align}
\text{min} & ~~  0.067z_{1}+0.167z_{2}+0.167z_{3}+0.167z_{4}+z_{5}\\
\text{s.t} & ~~ 6z_{1} +13.5z_{2}+15z_{3}+16.5z_{4}+22.5z_{5}\le 33\\
\ & ~~ z_{1},z_{2},z_{3},z_{4},z_{5}\ge 0\\
\end{align}</math>


''s.t          6z<sub>1</sub>  +  13.5z<sub>2</sub>  +  15z<sub>3</sub>  +  16.5z<sub>4</sub>  +  22.5z<sub>5</sub>  ≤  33''


''z<sub>1</sub> <sub>,</sub> z<sub>2 ,</sub> z<sub>3 ,</sub> z<sub>4 ,</sub> z<sub>5</sub>  ≥  0''
which has the following optimal solution:    <math>z^T= [1 \quad 0\quad 0\quad 0\quad 1]</math>


which has the following optimal solution:
This matches the pattern we called <math>D6</math>, earlier on in this page.


''z<sup>T</sup> = [1  0  0  0  1]''


This matches the pattern called D6 earlier on in this page.
<u>Optimality test</u>


<u>Optimality test</u>
If :  <math display="inline">\sum_{k=1}^{K}z_{k}^{*}u_{k}^{*}\leq 1</math>


If :
then <math>y^*</math> is an optimal solution of the full continuous relaxed problem (that is, including all patterns in ''<math display="inline">S</math>'')


<math>\begin{matrix} \\\sum_Kz_k(optimal)u_k(optimal) \leq 1, \end{matrix}
If this condition is not true, we go ahead and update the master problem by including in ''<math display="inline">S^'</math>'' the pattern <math>\lambda</math> defined by <math>N_{s,\lambda}</math> (in practical terms this means that the column '''<math>y^*</math>''' needs to be included in the constraint matrix)
</math>


then '''y*''' is an optimal solution of the full continuous relaxed problem (that is, including all patterns in S)
For this example we find that the optimality test is met as <math>\sum_{k=1}^{K}z_{k}^{*}u_{k}^{*}=0.4 \leq 1</math> so we have have found an optimal solution of the relaxed continuous problem (if this was not the case we would have had to go back to reformulating and solving the master problem, as discussed in the methodology section of this page)  


If this condition is not true, we go ahead and update the master problem by including in S' the pattern λ defined by N<sub>sλ</sub> (in practical terms this means that the column '''y*''' needs to be included in the constraint matrix) Then,go to Step1.


For this example we find that the optimality test is met as 0.4 <1 so we have have found an optimal solution of the relaxed continuos problem (if this was not the case  we would have had to go back to Step 1 as descrbed in the algorithm discussion of this page)


<math>\begin{matrix} \\ z_k(optimal)u_k(optimal) = 0.4 \leq 1,  \end{matrix}
</math>


'''''Algorithm discussion'''''
'''''Algorithm discussion'''''


The column generation subproblem is the critical part of the method is Step 2, i.e., generating the new columns. It is not reasonable to compute the reduced costs of all variables y<sub>s</sub> for s = 1, . . . , S, otherwise this procedure would reduce to the simplex method. In fact, n can be very large (as in the cutting-stock problem) or, for some reason, it might not be possible or convenient to enumerate all decision variables. It is then necessary to study a specific column generation algorithm for each problem; ''only if such an algorithm exists (and is efficient)'', the method can be fully developed. In the one-dimensional cutting stock problem, we transformed the column generation subproblem into an easily solvable integer linear programming problem. In other cases, the computational effort required to solve the subproblem may be so high as to make the full procedure unpractical.
The column generation subproblem is the critical part of the method is generating the new columns. It is not reasonable to compute the reduced costs of all variables <math>y_s
</math> for <math>s=1,...,S</math>, otherwise this procedure would reduce to the simplex method. In fact, n<math>n</math> can be very large (as in the cutting-stock problem) or, for some reason, it might not be possible or convenient to enumerate all decision variables. This is when it would be necessary to study a specific column generation algorithm for each problem; ''only if such an algorithm exists (and is practical)'', the method can be fully applied. In the one-dimensional cutting stock problem, we transformed the column generation subproblem into an easily solvable integer linear programming problem. In other cases, the computational effort required to solve the subproblem is too high, such that appying this full procedure becomes unefficient.


= '''Applications''' =
== Applications ==
As previously mentioned, column generation techniques are most relevant when the problem that we are trying to solve has a high ratio of number of variables with respect to the number of constraints. As such some common applications are:
As previously mentioned, column generation techniques are most relevant when the problem that we are trying to solve has a high ratio of number of variables with respect to the number of constraints. As such some common applications are:


·       Network design
* Bandwith packing
* Bus driver scheduling
* Generally, column generation algorithms are used for large delivery networks, often in combination with other methods, helping to implement real-time solutions for on-demand logistics. We discuss a supply chain scheduling application below.
 
'''''Bandwidth packing'''''
 
The objective of this problem is to allocate bandwidth in a telecommunications network to maximize total revenue. The routing of a set of traffic demands between different users is to be decided, taking into account the capacity of the network arcs and the fact that the traffic between each pair of users cannot be split The problem can be formulated as an integer programming problem and the linear programming relaxation solved using column generation and the simplex algorithm. A branch and bound procedure which branches upon a particular path is used in this particular paper<ref name=":3">Parker, Mark & Ryan, Jennifer. (1993). A column generation algorithm for bandwidth packing. Telecommunication Systems. 2. 185-195. 10.1007/BF02109857. </ref> that looks into bandwidth routing, to solve the IP. The column generation algorithm greatly reduces the complexity of this problem.
 
'''''Bus driver scheduling'''''


·       Logistics – for example to determine an optimal path/routing for vehicles
Bus driver scheduling aims to find the minimum number of bus drivers to cover a published timetable of a bus company. When scheduling bus drivers, contractual working rules must be enforced, thus complicating the problem. A column generation algorithm can decompose this complicated problem into a master problem and a series of pricing subproblems. The master problem would select optimal duties from a set of known feasible duties, and the pricing subproblem would augment the feasible duty set to improve the solution obtained in the master problem.<ref name=":2">Dung‐Ying Lin, Ching‐Lan Hsu. Journal of Advanced Transportation. Volume50, Issue8, December 2016, Pages 1598-1615. URL: https://onlinelibrary.wiley.com/doi/abs/10.1002/atr.1417</ref>


·       Column generation algorithms are used for large delivery networks, often in combination with other methods, helping to implement real-time solutions for om-demand logistics.
'''''Supply Chain scheduling problem'''''


·       Supply Chain scheduling problems
A typical application is where we consider the problem of scheduling a set of shipments between different nodes of a supply chain network. Each shipment has a fixed departure time, as well as an origin and a destination node, which, combined, determine the duration of the associated trip. The aim is to schedule as many shipments as possible, while also minimizing the number of vehicles utilized for this purpose. This problem can be formulated by an integer programming model and an associated branch and price solution algorithm. The optimal solution to the LP relaxation of the problem can be obtained through column generation, solving the linear program a huge number of variables, without explicitly considering all of them. In the context of this application, the master problem schedules the maximum possible number of shipments using only a small set of vehicle-routes, and a column generation (colgen) sub-problem would generate cost-effective vehicle-routes to be fed fed into the master problem. After finding the optimal solution to the LP relaxation of the problem, the algorithm would branch on the fractional decision variables (vehicle-routes), in order to reach the optimal integer solution.<ref name=":1">Kozanidis, George. (2014). Column generation for scheduling shipments within a supply chain network with the minimum number of vehicles. OPT-i 2014 - 1st International Conference on Engineering and Applied Sciences Optimization, Proceedings. 888-898</ref>


= '''Conclusions''' =
== Conclusions ==
Column generation is a way of beginning with a small, manageable parts of a problem (specifically, a few of the variables), solving that part, analyzing that partial solution to discover the next part of the problem (specifically, one or more variables) to add to the model, and then resolving the extended model. Column generation repeats the algorithm steps until it achieves a optimal solution to entire problem.
Column generation is a way of starting with a small, manageable part of a problem (specifically, with some of the variables), solving that part, analyzing that interim solution to find the next part of the problem (specifically, one or more variables) to add to the model, and then solving the full or extended model. In the column generation method, the algorithm steps are repeated until an optimal solution to the entire problem is achieved.<ref> ILOG CPLEX 11.0 User's Manual > Discrete Optimization > Using Column Generation: a Cutting Stock Example > What Is Column Generation? 1997-2007. URL:http://www-eio.upc.es/lceio/manuals/cplex-11/html/usrcplex/usingColumnGen2.html#:~:text=In%20formal%20terms%2C%20column%20generation,method%20of%20solving%20the%20problem.&text=By%201960%2C%20Dantzig%20and%20Wolfe,problems%20with%20a%20decomposable%20structure</ref>


More formally, column generation is a way of solving a linear programming problem that adds columns (corresponding to constrained variables) during the pricing phase of the simplex method of solving the problem. In gross terms, generating a column in the primal simplex formulation of a linear programming problem corresponds to adding a constraint in its dual formulation.
This algorithm provides a way of solving a linear programming problem adding columns (corresponding to constrained variables) during the pricing phase of the problem solving phase, that would otherwise be very tedious to formulate and compute. Generating a column in the primal formulation of a linear programming problem corresponds to adding a constraint in its dual formulation.


Column generation provides an advantage to the simplex method as the solvers (when computing the solution with software) will not need to access all the variables of the problem simultaneously. In fact, a solver could begin work with only the basis (a particular subset of the constrained variables) and then use reduced cost to decide which other variables to access as needed.
== References ==

Latest revision as of 06:37, 21 December 2020

Author: Lorena Garcia Fernandez (lgf572) (SysEn 5800 Fall 2020)

Introduction

Column Generation techniques have the scope of solving large linear optimization problems by generating only the variables that will have an influence on the objective function. This is important for big problems with many variables where the formulation with these techniques would simplify the problem formulation, since not all the possibilities need to be explicitly listed.[1]

Theory, methodology and algorithmic discussions

Theory

The way this method work is as follows; first, the original problem that is being solved needs to be split into two problems: the master problem and the sub-problem.

  • The master problem is the original column-wise (i.e: one column at a time) formulation of the problem with only a subset of variables being considered.[2]
  • The sub-problem is a new problem created to identify a new promising variable. The objective function of the sub-problem is the reduced cost of the new variable with respect to the current dual variables, and the constraints require that the variable obeys the naturally occurring constraints. The subproblem is also referred to as the RMP or “restricted master problem”. From this we can infer that this method will be a good fit for problems whose constraint set admit a natural breakdown (i.e: decomposition) into sub-systems representing a well understood combinatorial structure.[3]

To execute that decomposition from the original problem into Master and subproblems there are different techniques. The theory behind this method relies on the Dantzig-Wolfe decomposition.[4]

In summary, when the master problem is solved, we are able to obtain dual prices for each of the constraints in the master problem. This information is then utilized in the objective function of the subproblem. The subproblem is solved. If the objective value of the subproblem is negative, a variable with negative reduced cost has been identified. This variable is then added to the master problem, and the master problem is re-solved. Re-solving the master problem will generate a new set of dual values, and the process is repeated until no negative reduced cost variables are identified. The subproblem returns a solution with non-negative reduced cost, we can conclude that the solution to the master problem is optimal.[5]

Methodology[6]

Column generation schematics[7]

Consider the problem in the form:

(IP)


Where for . Assuming that each set contains a large but finite set of points , we have that :

Note that, on the assumption that each of the sets is bounded for the approach will involve solving an equivalent problem of the form as below:

where each matrix has a very large number of columns, one for each of the feasible points in , and each vector contains the corresponding variables.


Now, substituting for leads to an equivalent IP Master Problem (IPM):

(IPM)

To solve the Master Linear Program, we use a column generation algorithm. This is in order to solve the linear programming relaxation of the Integer Programming Master Problem, called the Linear Programming Master Problem (LPM):

(LPM)

Where there is a column for each . On the next steps of this method, we will use as the dual variables associated with the joint constraints, and as dual variables for the second set of constraints.The latter are also known as convexity constraints. The idea is to solve the linear program by the primal simplex algorithm. However, the pricing step of choosing a column to enter the basis must be modified because of the very big number of columns in play. Instead of pricing the columns one at a time, the question of finding a column with the biggest reduced price is itself a set of optimization problems.


Initialization: we suppose that a subset of columns (at least one for each ) is available, providing a feasible Restricted Linear Programming Master Problem:

(RLPM)


where , is generated by the available set of columns and are the corresponding costs and variables. Solving the RLPM gives an optimal primal solution and an optimal dual solution


Primal feasibility: Any feasible solution of RLMP is feasible for LPM. More precisely, is a feasible solution of LPM, and hence

Optimality check for LPM: It is required to check whether is dual feasible for LPM. This means checking for each column, that is for each , and for each if the reduced price . Rather than examining each point separately, we treat all points in implicitly, by solving an optimization subproblem:


Stopping criteria: If for the solution is dual feasible for LPM, and hence . As the value of the primal feasible solution equals that of this upper bound, is optimal for LPM.


Generating a new column: If for some , the column corresponding to the optimal solution of the subproblem has a positive reduced price. Introducing the column leads then to a Restricted Linear Programming Master Problem that can be easily reoptimized (e.g., by the primal simplex algorithm)

Numerical example: The Cutting Stock problem[8]

Suppose we want to solve a numerical example of the cutting stock problem, specifically a one-dimensional cutting stock problem.

Problem Overview

A company produces steel bars with diameter millimeters and length meters. The company also takes care of cutting the bars for their different customers, who each require different lengths. At the moment, the following demand forecast is expected and must be satisfied:

Pieces needed Piece length(m) Type of item
144 6 1
105 13.5 2
72 15 3
30 16.5 4
24 22.5 5

The objective is to establish what is the minimum number of steel bars that should be used to satisfy the total demand.

A possible model for the problem, proposed by Gilmore and Gomory in the 1960ies is the one below:

Sets

: set of item types;

: set of patterns (i.e., possible ways) that can be adopted to cut a given bar into portions of the need lengths.

Parameters

: bar length (before the cutting process);

: length of item ;

 : number of pieces of type required;

 : number of pieces of type in pattern .

Decision variables

 : number of bars that should be portioned using pattern .

Model

Solving the problem

The model assumes the availability of the set and the parameters . To generate this data, you would have to list all possible cutting patterns. However, the number of possible cutting patterns is a big number. This is why a direct implementation of the model above is not practical in real-world problems. In this case is when it makes sense to solve the continuous relaxation of the above model. This is because, in reality, the demand figures are so high that the number of bars to cut is also a large number, and therefore a good solution can be determined by rounding up to the next integer each variable found by solving the continuous relaxation. In addition to that, the solution of the relaxed problem will become the starting point for the application of an exact solution method (for instance, the Branch-and Bound).

Key take-away: In the next steps of this example we will analyze how to solve the continuous relaxation of the model.

As a starting point, we need any feasible solution. Such a solution can be constructed as follows:

  1. We consider any single-item cutting patterns, i.e., configurations, each containing pieces of type ;
  2. Set for pattern (where pattern is the pattern containing only pieces of type ).

This solution could also be arrived to by applying the simplex method to the model (without integrality constraints), considering only the decision variables that correspond to the above single-item patterns:

In fact, if we solve this problem (for example, use CPLEX solver in GAMS) the solution is as below:

Y1 28.8
Y2 52.5
Y3 24
Y4 15
Y5 24

Next, a new possible pattern (number ) will be considered. This pattern contains only one piece of item type number . So the question is if the new solution would remain optimal if this new pattern was allowed. Duality helps answer ths question. At every iteration of the simplex method, the outcome is a feasible basic solution (corresponding to some basis ) for the primal problem and a dual solution (the multipliers ) that satisfy the complementary slackness conditions. (Note: the dual solution will be feasible only when the last iteration is reached)

The inclusion of new pattern corresponds to including a new variable in the primal problem, with objective cost (as each time pattern is chosen, one bar is cut) and corresponding to the following column in the constraint matrix:


These variables create a new dual constraint. We then have to check if this new constraint is violated by the current dual solution  (or in other words, if the reduced cost of the new variable with respect to basis is negative)

The new dual constraint is:

The solution for the dual problem can be computed in different software packages, or by hand. The example below shows the solution obtained with GAMS for this example:

(Note the solution for the dual problem would be: )


Dual variable Variable value
D1 0.067
D2 0.167
D3 0.167
D4 0.167
D5 0.333

Since , the new constraint is violated.

This means that the current primal solution (in which the new variable is ) may not be optimal anymore (although it is still feasible). The fact that the dual constraint is violated means the associated primal variable has negative reduced cost:

the norm of

To help us solve the problem, the next step is to let enter the basis. To do so, we modify the problem by inserting the new variable as below:


If this problem is solved with the simplex method, the optimal solution is found, but restricted only to patterns to . If a new pattern is available, a decision should be made whether this new pattern should be used or not by proceeding as above. However, the problem is how to find a pattern (i.e., a variable; i.e, a column of the matrix) whose reduced cost is negative (i.e., which will mean it is convenient to include it in the formulation). At this point one can notice that number of possible patterns exponentially large,and all the patterns are not even known explicitly. The question then is:

Given a basic optimal solution for the problem in which only some variables are included, how can we find (if any exists) a variable with negative reduced cost (i.e., a constraint violated by the current dual solution)?

This question can be transformed into an optimization problem: in order to see whether a variable with negative reduced cost exists, we can look for the minimum of the reduced costs of all possible variables and check whether this minimum is negative:

Because every column of the constraint matrix corresponds to a cutting pattern, and every entry of the column says how many pieces of a certain type are in that pattern. In order for to be a possible column of the constraint matrix, the following condition must be satisfied:

And by so doing, it enables the conversion of the problem of finding a variable with negative reduced cost into the integer linear programming problem below:

which, in turn, would be equivalent to the below formulation (we just write the objective in maximization form and ignore the additive constant ):


The coefficients of a column with negative reduced cost can be found by solving the above integer "knapsack" problem (which is a traditional type of problem that we find in integer programming).

In our example, if we start from the problem restricted to the five single-item patterns, the above problem reads as:


which has the following optimal solution:

This matches the pattern we called , earlier on in this page.


Optimality test

If :

then is an optimal solution of the full continuous relaxed problem (that is, including all patterns in )

If this condition is not true, we go ahead and update the master problem by including in the pattern defined by (in practical terms this means that the column needs to be included in the constraint matrix)

For this example we find that the optimality test is met as so we have have found an optimal solution of the relaxed continuous problem (if this was not the case we would have had to go back to reformulating and solving the master problem, as discussed in the methodology section of this page)



Algorithm discussion

The column generation subproblem is the critical part of the method is generating the new columns. It is not reasonable to compute the reduced costs of all variables for , otherwise this procedure would reduce to the simplex method. In fact, n can be very large (as in the cutting-stock problem) or, for some reason, it might not be possible or convenient to enumerate all decision variables. This is when it would be necessary to study a specific column generation algorithm for each problem; only if such an algorithm exists (and is practical), the method can be fully applied. In the one-dimensional cutting stock problem, we transformed the column generation subproblem into an easily solvable integer linear programming problem. In other cases, the computational effort required to solve the subproblem is too high, such that appying this full procedure becomes unefficient.

Applications

As previously mentioned, column generation techniques are most relevant when the problem that we are trying to solve has a high ratio of number of variables with respect to the number of constraints. As such some common applications are:

  • Bandwith packing
  • Bus driver scheduling
  • Generally, column generation algorithms are used for large delivery networks, often in combination with other methods, helping to implement real-time solutions for on-demand logistics. We discuss a supply chain scheduling application below.

Bandwidth packing

The objective of this problem is to allocate bandwidth in a telecommunications network to maximize total revenue. The routing of a set of traffic demands between different users is to be decided, taking into account the capacity of the network arcs and the fact that the traffic between each pair of users cannot be split The problem can be formulated as an integer programming problem and the linear programming relaxation solved using column generation and the simplex algorithm. A branch and bound procedure which branches upon a particular path is used in this particular paper[9] that looks into bandwidth routing, to solve the IP. The column generation algorithm greatly reduces the complexity of this problem.

Bus driver scheduling

Bus driver scheduling aims to find the minimum number of bus drivers to cover a published timetable of a bus company. When scheduling bus drivers, contractual working rules must be enforced, thus complicating the problem. A column generation algorithm can decompose this complicated problem into a master problem and a series of pricing subproblems. The master problem would select optimal duties from a set of known feasible duties, and the pricing subproblem would augment the feasible duty set to improve the solution obtained in the master problem.[10]

Supply Chain scheduling problem

A typical application is where we consider the problem of scheduling a set of shipments between different nodes of a supply chain network. Each shipment has a fixed departure time, as well as an origin and a destination node, which, combined, determine the duration of the associated trip. The aim is to schedule as many shipments as possible, while also minimizing the number of vehicles utilized for this purpose. This problem can be formulated by an integer programming model and an associated branch and price solution algorithm. The optimal solution to the LP relaxation of the problem can be obtained through column generation, solving the linear program a huge number of variables, without explicitly considering all of them. In the context of this application, the master problem schedules the maximum possible number of shipments using only a small set of vehicle-routes, and a column generation (colgen) sub-problem would generate cost-effective vehicle-routes to be fed fed into the master problem. After finding the optimal solution to the LP relaxation of the problem, the algorithm would branch on the fractional decision variables (vehicle-routes), in order to reach the optimal integer solution.[11]

Conclusions

Column generation is a way of starting with a small, manageable part of a problem (specifically, with some of the variables), solving that part, analyzing that interim solution to find the next part of the problem (specifically, one or more variables) to add to the model, and then solving the full or extended model. In the column generation method, the algorithm steps are repeated until an optimal solution to the entire problem is achieved.[12]

This algorithm provides a way of solving a linear programming problem adding columns (corresponding to constrained variables) during the pricing phase of the problem solving phase, that would otherwise be very tedious to formulate and compute. Generating a column in the primal formulation of a linear programming problem corresponds to adding a constraint in its dual formulation.

References

  1. Desrosiers, Jacques & Lübbecke, Marco. (2006). A Primer in Column Generation.p7-p14 10.1007/0-387-25486-2_1.
  2. AlainChabrier, Column Generation techniques, 2019 URL: https://medium.com/@AlainChabrier/column-generation-techniques-6a414d723a64
  3. AlainChabrier, Column Generation techniques, 2019 URL: https://medium.com/@AlainChabrier/column-generation-techniques-6a414d723a64
  4. Dantzig-Wolfe decomposition. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Dantzig-Wolfe_decomposition&oldid=50750
  5. Wikipedia, the free encyclopeda. Column Generation. URL: https://en.wikipedia.org/wiki/Column_generation
  6. L.A. Wolsey, Integer programming. Wiley,Column Generation Algorithms p185-p189,1998
  7. GERARD. (2005). Personnel and Vehicle scheduling, Column Generation, slide 12. URL: https://slideplayer.com/slide/6574/
  8. L.A. Wolsey, Integer programming. Wiley,Column Generation Algorithms p185-p189,1998The Cutting Stock problem
  9. Parker, Mark & Ryan, Jennifer. (1993). A column generation algorithm for bandwidth packing. Telecommunication Systems. 2. 185-195. 10.1007/BF02109857.
  10. Dung‐Ying Lin, Ching‐Lan Hsu. Journal of Advanced Transportation. Volume50, Issue8, December 2016, Pages 1598-1615. URL: https://onlinelibrary.wiley.com/doi/abs/10.1002/atr.1417
  11. Kozanidis, George. (2014). Column generation for scheduling shipments within a supply chain network with the minimum number of vehicles. OPT-i 2014 - 1st International Conference on Engineering and Applied Sciences Optimization, Proceedings. 888-898
  12. ILOG CPLEX 11.0 User's Manual > Discrete Optimization > Using Column Generation: a Cutting Stock Example > What Is Column Generation? 1997-2007. URL:http://www-eio.upc.es/lceio/manuals/cplex-11/html/usrcplex/usingColumnGen2.html#:~:text=In%20formal%20terms%2C%20column%20generation,method%20of%20solving%20the%20problem.&text=By%201960%2C%20Dantzig%20and%20Wolfe,problems%20with%20a%20decomposable%20structure