Optimization with absolute values: Difference between revisions

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search
No edit summary
 
(46 intermediate revisions by 5 users not shown)
Line 1: Line 1:
Authors: Matthew Chan (mdc297), Yilian Yin (yy896), Brian Amado (ba392), Peter (pmw99), Dewei Xiao (dx58) - SYSEN 5800 Fall 2020
Authors: Matthew Chan (mdc297), Yilian Yin (yy896), Brian Amado (ba392), Peter Williams (pmw99), Dewei Xiao (dx58) (SysEn 5800 Fall 2020)
 
Steward: Fengqi You


== Introduction ==
== Introduction ==
Absolute values can make it relatively difficult to determine the optimal solution when handled without first converting to standard form. This conversion of the objective function is a good first step in solving optimization problems with absolute values. As a result, one can go on to solve the problem using linear programing techniques.
Absolute values can make it relatively difficult to determine the optimal solution when handled without first converting to standard form. This conversion of the objective function is a good first step in solving optimization problems with absolute values. As a result, one can go on to solve the problem using linear programing techniques. With the addition of a new variable (ex: <math>\textstyle X^a </math>) in the objective function the problem is considered nonlinear. Additional constraints must be added to find the optimal solution.


== Method ==
== Method ==


=== Defining Absolute Values ===
=== Defining Absolute Values ===
An absolute value of a real number can be described as its distance away from zero, or the non-negative magnitude of the number. Thus,
An absolute value of a real number can be described as its distance away from zero, or the non-negative magnitude of the number. <ref> Mendelson, Elliott, Schaum's Outline of Beginning Calculus, McGraw-Hill Professional, 2008. https://books.google.com/books?id=A8hAm38zsCMC&pg=PA2#v=onepage&q&f=false </ref> Thus,


<math>|x| = \begin{cases} -x, & \text{if }x < 0 \\ x, & \text{if }x \ge 0 \end{cases} </math>
<math>\displaystyle |x|={\begin{cases}-x,&{\text{if }}x<0\\x,&{\text{if }}x\geq 0\end{cases}}</math>


Absolute values can exist in optimization problems in two primary instances: in constraints and in the objective function.
Absolute values can exist in linear optimization problems in two primary instances: in constraints and in the objective function. <ref> "Absolute Values." ''lp_solve'', http://lpsolve.sourceforge.net/. Accessed 20 Nov. 2020. </ref>


=== Absolute Values in Constraints ===
=== Absolute Values in Constraints ===
Within linear equations, linear constraints can exist in several forms.
Within constraints, absolute value relations can be transformed into one of the following forms:


The first form exists as <math>|X| = 0 </math>, where <math display="inline">X </math> is a linear combination of variables.
<math> \begin{align}
|X| &= 0 \\
|X| &\le C \\
|X| &\ge C
\end{align} </math>


In this case, the only solution is if <math>|X| = 0 </math>, simplifying the constraint to <math>X = 0  </math>. Note that this solution also occurs if the constraint is in the form <math>|X| <= 0  </math> due to the same conclusion (only solution <math>X = 0 </math>).
Where <math>\textstyle X</math> is a linear combination (<math>\textstyle ax_1 + bx_2 + ...</math> where <math>\textstyle a, b</math> are constants) and <math>\textstyle C</math> is a constant <math>\textstyle > 0</math>.


The second form a linear constraint can exist in is <math>|X| \le C </math> where <math display="inline">X </math> remains a linear combination of variables and constant <math display="inline">C > 0 </math>.
==== Form when <math>\displaystyle |X| = 0</math> ====
In this form, the only possible solution is if <math>\displaystyle X = 0</math> simplifying the constraint. Note that this solution also occurs if the constraint is in the form <math>\displaystyle |X| \le 0</math> due to the same conclusion that the only possible solution is <math>\textstyle X = 0</math>.


In this case, we can describe an equivalent feasible solution by splitting the inequality into
==== Form when <math>\displaystyle |X| \le C</math> ====
The second form a linear constraint can exist in is <math>\displaystyle |X|\leq C</math>. In this case, an equivalent feasible solution can be described by splitting the constraint into two:


<math>X \le C </math>
<math> \begin{align}
X &\leq C \\
-X &\leq C
\end{align} </math>


<math>-X \le C </math>
The solution can be understood visually since <math>\textstyle X</math> must lie between <math>\textstyle -C</math> and <math>\textstyle C</math>, as shown below:


We can understand this visually as the solution <math display="inline">X </math> must lie between <math display="inline">-C
</math> and <math display="inline">C </math>, as shown below:
[[File:Number Line X Less Than C.png|none|thumb]]
[[File:Number Line X Less Than C.png|none|thumb]]


The last case for linear constraints is when <math>|X| \ge C </math>.
==== Form when <math>\displaystyle |X| \ge C</math> ====
Visually, the solution space for the last form is the complement of the second solution above, resulting in the following representation:[[File:Number Line for X Greater Than C.png|none|thumb]]In expression form, the solutions can be written as:


Visually, the solution space is the complement of the second solution above, resulting in the following representation:
<math> \begin{align}
[[File:Number Line for X Greater Than C.png|none|thumb]]
X &\geq C \\
-X &\geq C
\end{align} </math>


In expression form, the solutions can be written as:
As seen visually, the feasible region has a gap and thus non-convex. The expressions also make it impossible for both to simultaneously hold true. This means that it is not possible to transform constraints in this form to linear equations. <ref> ''Optimization Methods in Management Science / Operations Research.'' Massachusetts Institute of Technology, Spring 2013, https://ocw.mit.edu/courses/sloan-school-of-management/15-053-optimization-methods-in-management-science-spring-2013/tutorials/MIT15_053S13_tut04.pdf. Accessed 20 Nov. 2020. </ref>


<math>X \ge C </math>
An approach to reach a solution for this particular case exists in the form of Mixed-Integer Linear Programming, where only one of the equations above is “active”.


<math>-X \ge C </math>
The inequality can be reformulated into the following:


As seen visually, the feasible region has a gap and thus non-convex. The expressions also make it impossible for both to simultaneously hold true. This means that it is not possible to transform constraints in this form to linear equations. An approach to reach a solution for this particular case exists in the form of [[Mixed-Integer Linear Programming]], where only one of the equations above is “active”.
<math> \begin{align}
&X + N*Y \ge C \\
-&X + N*(1-Y) \ge C \\
&Y = 0, 1
\end{align} </math>
 
With this new set of constraints, a large constant <math>\textstyle N</math> is introduced, along with a binary variable <math>\textstyle Y</math>. So long as <math>\textstyle N</math> is sufficiently larger than the upper bound of <math>\textstyle X + C</math>, the large constant multiplied with the binary variable ensures that one of the constraints must be satisfied. For instance, if <math>\textstyle Y = 0</math>, the new constraints will resolve to:
 
<math> \begin{align}
&X \ge C \\
-&X + N \ge C
\end{align} </math>
 
Since <math>\textstyle N</math> is sufficiently large, the latter constraint will always be satisfied, leaving only one relation active: <math>\textstyle X \ge C</math>. Functionally, this allows for the XOR logical operation of <math>\textstyle X \geq C</math> and <math>\textstyle -X \geq C</math>.


=== Absolute Values in Objective Functions ===
=== Absolute Values in Objective Functions ===
In objective functions, to leverage transformations of absolute functions, several conditions must be fulfilled:
In objective functions, to leverage transformations of absolute functions, all constraints must be linear.
 
Similar to the case of absolute values in constraints, there are different approaches to the reformation of the objective function, depending on the satisfaction of sign constraints. The satisfaction of sign constraints is when the coefficient signs of the absolute terms must all be either:
 
* Positive for a minimization problem
* Negative for a maximization problem


# First, the constraints must all be linear.
==== Sign Constraints are Satisfied ====
# Secondly, the coefficient signs of the absolute terms must all be:
At a high level, the transformation works similarly to the second case of absolute value in constraints – aiming to bound the solution space for the absolute value term with a new variable, <math>\textstyle Z</math>.
#* Positive for a minimization problem
#* Negative for a maximization problem


At a high level, the transformation works similarly to the second case of absolute value in constraints – we aim to bound the solution space for the absolute value term with a new variable, <math display="inline">Y </math>.
If <math>\textstyle |X|</math> is the absolute value term in our objective function, two additional constraints are added to the linear program:


If <math>|X| </math> is the absolute value term in our objective function, where <math display="inline">X </math> is a linear combination of variables,
<math> \begin{align}
&X\leq Z \\
-&X\leq Z
\end{align} </math>


Two additional constraints are added to the linear program:
The <math>\textstyle |X|</math> term in the objective function is then replaced by <math>\textstyle Z</math>, relaxing the original function into a collection of linear constraints.


<math>X \le Y </math>
==== Sign Constraints are Not Satisfied ====
In order to transform problems where the coefficient signs of the absolute terms do not fulfill the conditions above, a similar conclusion is reached to that of the last case for absolute values in constraints – the use of integer variables is needed to reach an LP format.


<math>-X \le Y </math>
The following constraints need to be added to the problem:


This is a relaxation of the original function into a collection of linear constraints.
<math> \begin{align}
&X + N*Y \ge Z \\
-&X + N*(1-Y) \ge Z \\
&X \le Z \\
-&X \le Z \\
&Y = 0, 1
\end{align} </math>


In order to transform problems where the coefficient signs of the absolute terms do not fulfill the conditions above, a similar conclusion is reached to that of the last case for absolute values in constraints – the use of integer variables is needed to reach an LP format.
Again, <math>\textstyle N</math> is a large constant, <math>\textstyle Z</math> is a replacement variable for <math>\textstyle |X|</math> in the objective function, and <math>\textstyle Y</math> is a binary variable. The first two constraints ensure that one and only one constraint is active while the other will be automatically satisfied, following the same logic as above. The third and fourth constraints ensure that <math>\textstyle Z</math> must be equal to <math>\textstyle |X|</math> and has either a positive or negative value. For instance, for the case of <math>\textstyle Y = 0</math>, the new constraints will resolve to:
 
<math> \begin{align}
&X \ge Z \\
-&X + N \ge Z \\
&X \le Z \\
-&X \le Z
\end{align} </math>
 
As <math>\textstyle N</math> is sufficiently large (<math>\textstyle N</math> must be at least <math>\textstyle 2|X|</math> for this approach), the second constraint must be satisfied. Since <math>\textstyle Z</math> is non-negative, the fourth constraint must also be satisfied. The remaining constraints, <math>\textstyle X \ge Z</math> and <math>\textstyle X \le Z</math> can only be satisfied when <math>\textstyle Z = X</math> and is of non-negative signage. Together, these constraints will allow for the selection of the largest <math>\textstyle |X|</math> for maximization problems (or smallest for minimization problems).
 
=== Absolute Values in Nonlinear Optimization Problems ===
The addition of a new variable <math> (X_a) </math> to an objective function with absolute value quantities forms a nonlinear optimization problem. The absolute value quantities would require that the problem be reformatted before proceeding. Additional constraints must be added to account for the added variable.


==Numerical Example==
==Numerical Example==
 
'''Example when All Sign Constraints are Satisfied'''
<math>\min{2|x_1| + 3|x_2| + |x_3|} </math>


<math>  \begin{align}
<math>  \begin{align}
\ s.t. x_1 + 2x_2 - 3x_3 \le 8 \\
\min \quad &{2|x_1| + 3|x_2| + |x_3|}  \\
2x_1 - x_2 + 4x_3= 14
s.t. \quad &x_1 + 2x_2 - 3x_3 \le 8 \\
&2x_1 - x_2 + 4x_3= 14
\end{align}</math>
\end{align}</math>


We replace the absolute value quantities with a single variable:
The absolute value quantities will be replaced with single variables:


<math>|x_1| = U_1 </math>
<math>|x_1| = U_1 </math>
Line 96: Line 142:
The problem has now been reformulated as a linear programming problem that can be solved normally:
The problem has now been reformulated as a linear programming problem that can be solved normally:


<math>\min{ 2U_1 + 3U_2 + U_3} </math>
<math> \begin{align}
\min \quad &{ 2U_1 + 3U_2 + U_3}       \\
s.t. \quad &x_1 + 2x_2 - 3x_3 \le 8 \\
&2x_1 - x_2 + 4x_3= 14          \\
-&U_1 \le x_1 \le U_1            \\
-&U_2 \le x_2 \le U_2          \\
-&U_3 \le x_3 \le U_3         
\end{align}</math>
 
The optimum value for the objective function is <math>3.5</math>, which occurs when <math>x_1 = 0 </math> and <math>x_2 = 0 </math> and <math>x_3 = 3.5 </math>.
 
'''Example when Sign Constraints are not Satisfied'''


<math>  \begin{align}
<math>  \begin{align}
\ s.t. x_1 + 2x_2 - 3x_3 \le 8 \\
\min \quad &{2|x_1| + 3|x_2| - |x_3|}  \\
2x_1 - x_2 + 4x_3= 14
s.t. \quad &x_1 + 2x_2 - 3x_3 \le 8 \\
&2x_1 - x_2 + 4x_3= 14
\end{align}</math>
\end{align}</math>


<math> -U_1 \le x_1 \le U_1 </math>
The absolute value quantities will be replaced with single variables:
 
<math>|x_1| = U_1 </math>
 
<math>|x_2| = U_2</math>
 
<math>|x_3| = U_3</math>
 
We must introduce additional constraints to ensure we do not lose any information by doing this substitution:
 
<math>  \begin{align}
-&U_1 \le x_1 \le U_1 \\
-&U_2 \le x_2 \le U_2 \\
&x_3 + M*Y \ge U_3 \\
-&x_3 + M*(1-Y) \ge U_3 \\
&x_3 \le U_3 \\
-&x_3 \le U_3 \\
&Y = 0,1
\end{align}</math>


<math> -U_2 \le x_2 \le U_2 </math>
The problem has now been reformulated as a linear programming problem that can be solved normally:
<ref> Shanno, David F., and Roman L. Weil. “'Linear' Programming with Absolute-Value Functionals.” Operations Research, vol. 19, no. 1, 1971, pp. 120–124. Accessed 13 Dec. 2020. JSTOR, www.jstor.org/stable/168871. </ref>


<math> -U_3 \le x_3 \le U_3 </math>
<math> \begin{align}
\min \quad &{ 2U_1 + 3U_2 - U_3}        \\
s.t. \quad &x_1 + 2x_2 - 3x_3 \le 8  \\
&2x_1 - x_2 + 4x_3= 14          \\
-&U_1 \le x_1 \le U_1            \\
-&U_2 \le x_2 \le U_2          \\
&x_3 + M*Y \ge U_3            \\
-&x_3 + M*(1-Y) \ge U_3         \\
&x_3 \le U_3                    \\
-&x_3 \le U_3                   \\
&Y = 0,1
\end{align}</math>


The optimum value for the objective function is <math>.5</math>, which occurs when <math>x_1 = 0 </math> and <math>x_2 = 0 </math> and <math>x_3 = 3.5 </math>.
The optimum value for the objective function is <math>-3.5</math>, which occur when <math>x_1 = 0 </math> and <math>x_2 = 0 </math> and <math>x_3 = 3.5 </math>.


== Applications ==
== Applications ==
There are no specific applications to Optimization with Absolute Values however it is necessary to account for at times when utilizing the simplex method.


Consider the problem Ax=b; max z= x c,jx,i. This problem cannot, in general, be solved with the simplex method. The problem has a simplex method solution (with unrestricted basis entry) only if c, are nonpositive (non-negative for minimizing problems).
 
Consider the problem <math>Ax=b; \quad max \quad z= x c,jx,i</math>. This problem cannot, in general, be solved with the simplex method. The problem has a simplex method solution (with unrestricted basis entry) only if c, are nonpositive (non-negative for minimizing problems).


The primary application of absolute-value functionals in linear programming has been for absolute-value or L(i)-metric regression analysis. Such application is always a minimization problem with all C(j) equal to 1 so that the required conditions for valid use of the simplex method are met.   
The primary application of absolute-value functionals in linear programming has been for absolute-value or L(i)-metric regression analysis. Such application is always a minimization problem with all C(j) equal to 1 so that the required conditions for valid use of the simplex method are met.   


By reformulating the original problem into a Mixed-Integer Linear Program (MILP), in most case we should be able to use GAMS/Pyomo/JuliaOPT to solve the problem.  
By reformulating the original problem into a Mixed-Integer Linear Program (MILP), we can utilize known programs to solve for the optimal solution(s).  


* '''Application in Financial:''' '''Portfolio Selection'''
=== Application in Financial: Portfolio Selection===
Under this topic, the same tricks played in the Numerical Example section to perform '''Reduction to a Linear Programming Problem''' will be applied here again, to reform the problem into a MILP in order to solve the problem. An example is given as below.  
Under this topic, the same tricks played in the Numerical Example section to perform '''Reduction to a Linear Programming Problem''' will be applied here again, to reform the problem into a MILP in order to solve the problem. An example is given as below.  






A portfolio is determined by what fraction of one's assets to put into each investment. It can be denoted as a collection of nonnegative numbers xj, where j = 1, 2,...,n. Because each xj stands for a portion of the assets, it sums to one. In order to get a highest reward through finding a right mix of assets, let <math>\mu</math>, the positive parameter, denote the importance of risk relative to the return, and Rj denote the return in the next time period on investment j, j = 1, 2,..., n.  The total return one would obtain from the investment is <math>R = \sum_{j}\!x_j\!R_j </math>. The expected return is <math>\mathbb{E}\!R = \sum_{j}\!x_j\mathbb{E}\!R_j </math>.  And the Mean Absolute Deviation from the Mean (MAD) is <math>\mathbb{E}\left\vert \!R - \mathbb{E}\!R \right\vert =  \mathbb{E}\left\vert \sum_{j}\!x_j\tilde{R}_j  \right\vert </math>.   
A portfolio is determined by what fraction of one's assets to put into each investment. <ref>  Vanderbei R.J. (2008) Financial Applications. In: Linear Programming. International Series in Operations Research & Management Science, vol 114. Springer, Boston, MA. <nowiki>https://doi.org/10.1007/978-0-387-74388-2_13</nowiki> https://link.springer.com/chapter/10.1007/978-0-387-74388-2_13 </ref> It can be denoted as a collection of nonnegative numbers <math>\textstyle x_j</math>, where <math> j = 1, 2,...,n </math>. Because each <math> \textstyle x_j </math>stands for a portion of the assets, it sums to one. In order to get a highest reward through finding a right mix of assets, let <math>\mu</math>, the positive parameter, denote the importance of risk relative to the return, and <math>/textstyle Rj</math> denote the return in the next time period on investment <math>j, j = 1, 2,..., n</math>.  The total return one would obtain from the investment is <math>R = \sum_{j}\!x_j\!R_j </math>. The expected return is <math>\mathbb{E}\!R = \sum_{j}\!x_j\mathbb{E}\!R_j </math>.  And the Mean Absolute Deviation from the Mean (MAD) is <math>\mathbb{E}\left\vert \!R - \mathbb{E}\!R \right\vert =  \mathbb{E}\left\vert \sum_{j}\!x_j\tilde{R}_j  \right\vert </math>.   


maximize      <math display="inline">\mu\sum_j\!x_j\mathbb{E}\!R_j - \mathbb{E}\left\vert  \sum_j \!x_j\tilde{R}_j  \right\vert </math>


subject to    <math>\sum_j\!x_j = 1</math>
maximize      <math display="inline">\mu\sum_j\!x_j\mathbb{E}\!R_j - \mathbb{E}\left\vert  \sum_j \!x_j\tilde{R}_j  \right\vert </math>  


<math>x_j \geq 0</math>     <math> j  = 1,2,..n.</math>
subject to    <math>\sum_j\!x_j = 1</math>
 
<math>x_j \geq 0</math> <math> j  = 1,2,..n.</math>  


where          <math>\tilde{R}_j  = \!R_j - \mathbb{E}\!R_j </math>
where          <math>\tilde{R}_j  = \!R_j - \mathbb{E}\!R_j </math>
Line 168: Line 257:
So finally, after some simplifications methods and some tricks applied, the original problem is converted into a linear programming which is easier to be solved further.
So finally, after some simplifications methods and some tricks applied, the original problem is converted into a linear programming which is easier to be solved further.


===Data Transfer Rate===
Another application of optimization with absolute values is data transfer rate. Faster-than-nyquist, or FTNS, is a framework to transmit signals beyond the Nyquist rate. The refence to this section proposed a 24.7% faster symbol rate by utilizing Sum-of-Absolute-Values optimization. <ref>Sasahara, Hampei & Hayashi, Kazunori & Nagahara, Masaaki. (2016). Symbol Detection for Faster-Than-Nyquist Signaling by Sum-of-Absolute-Values Optimization. IEEE Signal Processing Letters. PP. 1-1. 10.1109/LSP.2016.2625839.  https://www.researchgate.net/publication/309745511_Symbol_Detection_for_Faster-Than-Nyquist_Signaling_by_Sum-of-Absolute-Values_Optimization </ref>
The initial model is defined as follows:
<math>\displaystyle x_0 (t) = \sum^N_{n=1} x_{n,0} h_n (t), t \in [0,T] </math>
where t ∈ R denotes the continuous time index, N ∈ N is the number of transmitted symbols in each transmission period, T > 0 is the interval of one period, <math>x_{n,0}</math>  ∈ {+1, −1} are independent and identically distributed (i.i.d.) binary symbols [i.e., binary phase shift keying (BPSK)], and <math>h_n (t) (n = 1,...,N) </math> are the modulation pulses.
Reformulated as a convex optimization problem and repeating Newton’s method with absolute values, the solution approximates can be achieved:
<math>\displaystyle \min_{z \in R^N} (\lambda \Vert y - Hz \Vert^2_2 + \frac{1}{2} \Vert z - 1_N \Vert_1 + \frac{1}{2} \Vert z + 1_N \Vert_1 ) </math>


== Conclusion ==
== Conclusion ==
The presence of an absolute value within the objective function prevents the use of certain optimization methods. Solving these problems requires that the function be manipulated in order to continue with linear programming techniques like the simplex method.
The presence of an absolute value within the objective function prevents the use of certain optimization methods. Solving these problems requires that the function be manipulated in order to continue with linear programming techniques like the simplex method. The applications of optimization with absolute values range from the financial sector to the digital world where data transfer rates can be improved as well as improving portfolio returns. The way these problems are formulated, must take absolute values into account in order to model the problem correctly. The absolute values inherently make these problems non-linear so determining the most optimal solutions is only achievable after reformulating them into linear programs.


== References ==
== References ==
# "Absolute Values." ''lp_solve'', http://lpsolve.sourceforge.net/5.1/absolute.htm. Accessed 20 Nov. 2020.
<references />
# ''Optimization Methods in Management Science / Operations Research.'' Massachusetts Institute of Technology, Spring 2013, https://ocw.mit.edu/courses/sloan-school-of-management/15-053-optimization-methods-in-management-science-spring-2013/tutorials/MIT15_053S13_tut04.pdf. Accessed 20 Nov. 2020.
# https://www.ise.ncsu.edu/fuzzy-neural/wp-content/uploads/sites/9/2019/08/LP-Abs-Value.pdf
# Vanderbei R.J. (2008) Financial Applications. In: Linear Programming. International Series in Operations Research & Management Science, vol 114. Springer, Boston, MA. <nowiki>https://doi.org/10.1007/978-0-387-74388-2_13</nowiki>

Latest revision as of 06:31, 21 December 2020

Authors: Matthew Chan (mdc297), Yilian Yin (yy896), Brian Amado (ba392), Peter Williams (pmw99), Dewei Xiao (dx58) (SysEn 5800 Fall 2020)

Introduction

Absolute values can make it relatively difficult to determine the optimal solution when handled without first converting to standard form. This conversion of the objective function is a good first step in solving optimization problems with absolute values. As a result, one can go on to solve the problem using linear programing techniques. With the addition of a new variable (ex: ) in the objective function the problem is considered nonlinear. Additional constraints must be added to find the optimal solution.

Method

Defining Absolute Values

An absolute value of a real number can be described as its distance away from zero, or the non-negative magnitude of the number. [1] Thus,

Absolute values can exist in linear optimization problems in two primary instances: in constraints and in the objective function. [2]

Absolute Values in Constraints

Within constraints, absolute value relations can be transformed into one of the following forms:

Where is a linear combination ( where are constants) and is a constant .

Form when

In this form, the only possible solution is if simplifying the constraint. Note that this solution also occurs if the constraint is in the form due to the same conclusion that the only possible solution is .

Form when

The second form a linear constraint can exist in is . In this case, an equivalent feasible solution can be described by splitting the constraint into two:

The solution can be understood visually since must lie between and , as shown below:

Form when

Visually, the solution space for the last form is the complement of the second solution above, resulting in the following representation:

In expression form, the solutions can be written as:

As seen visually, the feasible region has a gap and thus non-convex. The expressions also make it impossible for both to simultaneously hold true. This means that it is not possible to transform constraints in this form to linear equations. [3]

An approach to reach a solution for this particular case exists in the form of Mixed-Integer Linear Programming, where only one of the equations above is “active”.

The inequality can be reformulated into the following:

With this new set of constraints, a large constant is introduced, along with a binary variable . So long as is sufficiently larger than the upper bound of , the large constant multiplied with the binary variable ensures that one of the constraints must be satisfied. For instance, if , the new constraints will resolve to:

Since is sufficiently large, the latter constraint will always be satisfied, leaving only one relation active: . Functionally, this allows for the XOR logical operation of and .

Absolute Values in Objective Functions

In objective functions, to leverage transformations of absolute functions, all constraints must be linear.

Similar to the case of absolute values in constraints, there are different approaches to the reformation of the objective function, depending on the satisfaction of sign constraints. The satisfaction of sign constraints is when the coefficient signs of the absolute terms must all be either:

  • Positive for a minimization problem
  • Negative for a maximization problem

Sign Constraints are Satisfied

At a high level, the transformation works similarly to the second case of absolute value in constraints – aiming to bound the solution space for the absolute value term with a new variable, .

If is the absolute value term in our objective function, two additional constraints are added to the linear program:

The term in the objective function is then replaced by , relaxing the original function into a collection of linear constraints.

Sign Constraints are Not Satisfied

In order to transform problems where the coefficient signs of the absolute terms do not fulfill the conditions above, a similar conclusion is reached to that of the last case for absolute values in constraints – the use of integer variables is needed to reach an LP format.

The following constraints need to be added to the problem:

Again, is a large constant, is a replacement variable for in the objective function, and is a binary variable. The first two constraints ensure that one and only one constraint is active while the other will be automatically satisfied, following the same logic as above. The third and fourth constraints ensure that must be equal to and has either a positive or negative value. For instance, for the case of , the new constraints will resolve to:

As is sufficiently large ( must be at least for this approach), the second constraint must be satisfied. Since is non-negative, the fourth constraint must also be satisfied. The remaining constraints, and can only be satisfied when and is of non-negative signage. Together, these constraints will allow for the selection of the largest for maximization problems (or smallest for minimization problems).

Absolute Values in Nonlinear Optimization Problems

The addition of a new variable to an objective function with absolute value quantities forms a nonlinear optimization problem. The absolute value quantities would require that the problem be reformatted before proceeding. Additional constraints must be added to account for the added variable.

Numerical Example

Example when All Sign Constraints are Satisfied

The absolute value quantities will be replaced with single variables:

We must introduce additional constraints to ensure we do not lose any information by doing this substitution:

The problem has now been reformulated as a linear programming problem that can be solved normally:

The optimum value for the objective function is , which occurs when and and .

Example when Sign Constraints are not Satisfied

The absolute value quantities will be replaced with single variables:

We must introduce additional constraints to ensure we do not lose any information by doing this substitution:

The problem has now been reformulated as a linear programming problem that can be solved normally: [4]

The optimum value for the objective function is , which occur when and and .

Applications

Consider the problem . This problem cannot, in general, be solved with the simplex method. The problem has a simplex method solution (with unrestricted basis entry) only if c, are nonpositive (non-negative for minimizing problems).

The primary application of absolute-value functionals in linear programming has been for absolute-value or L(i)-metric regression analysis. Such application is always a minimization problem with all C(j) equal to 1 so that the required conditions for valid use of the simplex method are met.

By reformulating the original problem into a Mixed-Integer Linear Program (MILP), we can utilize known programs to solve for the optimal solution(s).

Application in Financial: Portfolio Selection

Under this topic, the same tricks played in the Numerical Example section to perform Reduction to a Linear Programming Problem will be applied here again, to reform the problem into a MILP in order to solve the problem. An example is given as below.


A portfolio is determined by what fraction of one's assets to put into each investment. [5] It can be denoted as a collection of nonnegative numbers , where . Because each stands for a portion of the assets, it sums to one. In order to get a highest reward through finding a right mix of assets, let , the positive parameter, denote the importance of risk relative to the return, and denote the return in the next time period on investment . The total return one would obtain from the investment is . The expected return is . And the Mean Absolute Deviation from the Mean (MAD) is .


maximize

subject to

,

where


Very obviously, this problem is not a linear programming problem yet. Similar to the numerical example showed above, the right thing to do is to replace each absolute value with a new variable and impose inequality constraints to ensure that the new variable is the appropriate absolute value once an optimal value is obtained. To simplify the program, an average of the historical returns can be taken in order to get the mean expected return: . Thus the objective function is turned into:

Now, replace with a new variable and thus the problem can be rewrote as:


maximize

subject to . t = 1, 2,...,T

where

. j = 1, 2,...,n

. t = 1, 2,...,T


So finally, after some simplifications methods and some tricks applied, the original problem is converted into a linear programming which is easier to be solved further.


Data Transfer Rate

Another application of optimization with absolute values is data transfer rate. Faster-than-nyquist, or FTNS, is a framework to transmit signals beyond the Nyquist rate. The refence to this section proposed a 24.7% faster symbol rate by utilizing Sum-of-Absolute-Values optimization. [6]

The initial model is defined as follows:

where t ∈ R denotes the continuous time index, N ∈ N is the number of transmitted symbols in each transmission period, T > 0 is the interval of one period, ∈ {+1, −1} are independent and identically distributed (i.i.d.) binary symbols [i.e., binary phase shift keying (BPSK)], and are the modulation pulses.

Reformulated as a convex optimization problem and repeating Newton’s method with absolute values, the solution approximates can be achieved:

Conclusion

The presence of an absolute value within the objective function prevents the use of certain optimization methods. Solving these problems requires that the function be manipulated in order to continue with linear programming techniques like the simplex method. The applications of optimization with absolute values range from the financial sector to the digital world where data transfer rates can be improved as well as improving portfolio returns. The way these problems are formulated, must take absolute values into account in order to model the problem correctly. The absolute values inherently make these problems non-linear so determining the most optimal solutions is only achievable after reformulating them into linear programs.

References

  1. Mendelson, Elliott, Schaum's Outline of Beginning Calculus, McGraw-Hill Professional, 2008. https://books.google.com/books?id=A8hAm38zsCMC&pg=PA2#v=onepage&q&f=false
  2. "Absolute Values." lp_solve, http://lpsolve.sourceforge.net/. Accessed 20 Nov. 2020.
  3. Optimization Methods in Management Science / Operations Research. Massachusetts Institute of Technology, Spring 2013, https://ocw.mit.edu/courses/sloan-school-of-management/15-053-optimization-methods-in-management-science-spring-2013/tutorials/MIT15_053S13_tut04.pdf. Accessed 20 Nov. 2020.
  4. Shanno, David F., and Roman L. Weil. “'Linear' Programming with Absolute-Value Functionals.” Operations Research, vol. 19, no. 1, 1971, pp. 120–124. Accessed 13 Dec. 2020. JSTOR, www.jstor.org/stable/168871.
  5. Vanderbei R.J. (2008) Financial Applications. In: Linear Programming. International Series in Operations Research & Management Science, vol 114. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-74388-2_13 https://link.springer.com/chapter/10.1007/978-0-387-74388-2_13
  6. Sasahara, Hampei & Hayashi, Kazunori & Nagahara, Masaaki. (2016). Symbol Detection for Faster-Than-Nyquist Signaling by Sum-of-Absolute-Values Optimization. IEEE Signal Processing Letters. PP. 1-1. 10.1109/LSP.2016.2625839. https://www.researchgate.net/publication/309745511_Symbol_Detection_for_Faster-Than-Nyquist_Signaling_by_Sum-of-Absolute-Values_Optimization