Quadratic constrained quadratic programming: Difference between revisions

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search
No edit summary
Line 29: Line 29:
==Algorithm Discussion==
==Algorithm Discussion==


KKT: What is KKT +formulation


SDP: What is SDP +formulation


Selected Solvers:
Selected Solvers:
Line 48: Line 50:


== Numerical Example ==
== Numerical Example ==
'''Quadratically Constrained Quadratic Program''' ('''QCQP''') always has the form:
Quadratically Constrained Quadratic Program (QCQP) always has the form:


<math>
<math>
Line 68: Line 70:
</math> are all zero, then the constraints are in fact linear and the problem is a quadratic program.
</math> are all zero, then the constraints are in fact linear and the problem is a quadratic program.


=== Example 1 : KKT Approach ===
=== Example 1: KKT Approach ===
Considering the following numerical example:
Considering the following numerical example:


Line 79: Line 81:
</math>
</math>


=== Step 1: Formulate the Lagrangian ===
==== (1) Steps ====


* Formulate the [[wikipedia:Lagrangian_mechanics|Lagrangian]]
* Computed the Gradients
* Applied the Stationarity Conditions
* Determined the active constraints using complementary slackness
===== Lagrangian Formulation: =====
The Lagrangian <math>L</math> combines the objective function and the constraints, each multiplied by a Lagrange multiplier <math>\lambda_i</math>:
The Lagrangian <math>L</math> combines the objective function and the constraints, each multiplied by a Lagrange multiplier <math>\lambda_i</math>:


Line 91: Line 99:
- '''Complementary Slackness''':
- '''Complementary Slackness''':
   <math>
   <math>
  \lambda_i \geq 0, \quad \lambda_i f_i(x) = 0, \quad \text{for } i = 1, 2.
  \lambda_i \geq 0, \quad \lambda_i f_i(x) = 0, \quad \text{for } i = 1, 2.
  </math>
  </math>


- '''Primal Feasibility''':
- '''Primal Feasibility''':
   <math>
   <math>
  f_i(x) \leq 0 \quad \text{for } i = 1, 2.
  f_i(x) \leq 0 \quad \text{for } i = 1, 2.
  </math>
  </math>


=== Step 2: Compute the Gradient of the Lagrangian ===
=== Step 2: Compute the Gradient of the Lagrangian ===
Line 105: Line 113:
- '''Partial Derivative with respect to <math>x_1</math>''':
- '''Partial Derivative with respect to <math>x_1</math>''':
   <math>
   <math>
  \frac{\partial L}{\partial x_1} = 2(x_1 - 2) + 2\lambda_1 x_1 + 2\lambda_2 (x_1 - 1).
  \frac{\partial L}{\partial x_1} = 2(x_1 - 2) + 2\lambda_1 x_1 + 2\lambda_2 (x_1 - 1).
  </math>
  </math>


- '''Partial Derivative with respect to <math>x_2</math>''':
- '''Partial Derivative with respect to <math>x_2</math>''':
   <math>
   <math>
  \frac{\partial L}{\partial x_2} = 2x_2 + 2\lambda_1 x_2 + 2\lambda_2 x_2.
  \frac{\partial L}{\partial x_2} = 2x_2 + 2\lambda_1 x_2 + 2\lambda_2 x_2.
  </math>
  </math>


=== Step 3: Stationarity Conditions ===
=== Step 3: Stationarity Conditions ===
Line 119: Line 127:
- '''Equation (1)''':
- '''Equation (1)''':
   <math>
   <math>
  2(x_1 - 2) + 2\lambda_1 x_1 + 2\lambda_2 (x_1 - 1) = 0.
  2(x_1 - 2) + 2\lambda_1 x_1 + 2\lambda_2 (x_1 - 1) = 0.
  </math>
  </math>


- '''Equation (2)''':
- '''Equation (2)''':
   <math>
   <math>
  2x_2 + 2\lambda_1 x_2 + 2\lambda_2 x_2 = 0.
  2x_2 + 2\lambda_1 x_2 + 2\lambda_2 x_2 = 0.
  </math>
  </math>


From '''Equation (2)''', since <math>x_2 (1 + \lambda_1 + \lambda_2) = 0</math> and <math>\lambda_i \geq 0</math> for <math>i = 1, 2</math>, it follows that:
From '''Equation (2)''', since <math>x_2 (1 + \lambda_1 + \lambda_2) = 0</math> and <math>\lambda_i \geq 0</math> for <math>i = 1, 2</math>, it follows that:
Line 179: Line 187:
    
    
   <math>
   <math>
  \lambda_1 (x_1^2 - 1) = 1 \times (1 - 1) = 0.
  \lambda_1 (x_1^2 - 1) = 1 \times (1 - 1) = 0.
  </math>
  </math>


- '''Constraint 2''':
- '''Constraint 2''':
    
    
   <math>
   <math>
  \lambda_2 \left( (x_1 - 1)^2 + x_2^2 - 1 \right) = 0 \times (-1) = 0.
  \lambda_2 \left( (x_1 - 1)^2 + x_2^2 - 1 \right) = 0 \times (-1) = 0.
  </math>
  </math>


=== Step 6: Verify Primal Feasibility ===
=== Step 6: Verify Primal Feasibility ===
Line 193: Line 201:
    
    
   <math>
   <math>
  x_1^2 - 1 = 1 - 1 = 0 \leq 0.
  x_1^2 - 1 = 1 - 1 = 0 \leq 0.
  </math>
  </math>


- '''Constraint 2''':
- '''Constraint 2''':
    
    
   <math>
   <math>
  (x_1 - 1)^2 + x_2^2 - 1 = -1 \leq 0.
  (x_1 - 1)^2 + x_2^2 - 1 = -1 \leq 0.
  </math>
  </math>


=== Step 7: Conclusion ===
=== Step 7: Conclusion ===
Line 206: Line 214:
- '''Optimal Solution''':
- '''Optimal Solution''':
   <math>
   <math>
  x_1^* = 1, \quad x_2^* = 0.
  x_1^* = 1, \quad x_2^* = 0.
  </math>
  </math>


- '''Minimum Objective Value''':
- '''Minimum Objective Value''':
   <math>
   <math>
  f_0^*(x) = (1 - 2)^2 + 0 = 1.
  f_0^*(x) = (1 - 2)^2 + 0 = 1.
  </math>
  </math>
- Result: $x_1^*=1, x_2^*=0$, with a minimum objective value of $f_0\left(x^*\right)=1$.
 
- Performance: The KKT method efficiently handles the convex QCQP problem but requires careful initialization to ensure convergence, especially for non-convex problems.
 
2. SDP Relaxation:
 
- Approach: The constraints were relaxed into semidefinite constraints, allowing the problem to be solved as a convex optimization task.
 
- Result: The SDP relaxation provided an identical solution for this convex problem, confirming its accuracy and reliability.
 
- Performance: While SDP offers robust solutions, it is computationally more intensive, particularly for large-scale problems with numerous constraints.


=== Example 2 : SDP- Based QCQP ===
=== Example 2 : SDP- Based QCQP ===
Line 311: Line 331:
==External Links==
==External Links==


* [[wikipedia:Lagrangian_mechanics|Lagrangian mechanics]]
* [[wikipedia:Karush–Kuhn–Tucker_conditions|Karush–Kuhn–Tucker conditions]]
* [[wikipedia:Karush–Kuhn–Tucker_conditions|Karush–Kuhn–Tucker conditions]]
* [[wikipedia:Software-defined_perimeter|Software-defined perimeter]]
* [[wikipedia:Software-defined_perimeter|Software-defined perimeter]]

Revision as of 11:04, 11 December 2024

Author: Jialiang Wang (jw2697), Jiaxin Zhang (jz2289), Wenke Du (wd275), Yufan Zhu (yz2899), David Berroa (deb336) (ChemE 6800 Fall 2024)

Stewards: Nathan Preuss, Wei-Han Chen, Tianqi Xiao, Guoqing Hu

Introduction

Quadratic programming (QP) is one of the oldest topics in the field of optimization that researchers have studied in the twentieth century. The basic QP, where the objective function is quadratic and constraints are linear, paved the way for other forms, such as QCQPs, which also have quadratic constraints. QCQPs emerged as optimisation theory grew to address more realistic, complex problems of non-linear objectives and constraints.

A Quadratically Constrained Quadratic Program (QCQP) can be defined as an optimization problem where the objective function and the constraints are quadratic. In particular, the issue involves optimizing (or minimizing) a convex quadratic function of decision variables with quadratic constraints. This class of problems is well suited to finance (Zenios,1993), engineering, machine learning, and agriculture because it is easier to model the relationship between variables using quadratic functions.

  • The desire to study QCQPs stems from the fact that they can be used to model practical optimization problems that involve stochasticity in risk, resources, production, and decision-making. For example, in agriculture, using QCQPs can be useful in determining the best crop to grow based on the expected profits and the uncertainties of price changes and unfavourable weather conditions (Floudas,1995).
  • In finance, the QCQPs are applied in the portfolio's construction to maximize the portfolio's expected returns and the covariance between the assets. It is crucial to comprehend QCQPs and the ways to solve them, such as KKT (Karush-Kuhn Tucker) conditions and SDP (Semidefinite Programming) relaxations, to solve problems that linear models cannot effectively solve (Bao & Sahinidis,2011).
Name Brief info
KKT(Karush-Kuhn-Tucker) KKT is a mathematical optimization method used to solve constrained optimization problems(Ghojogh, Karray et al,2021).

It builds upon the method of Lagrange multipliers by introducing necessary conditions for optimality that incorporate primal and dual variables. The KKT conditions include stationarity, primal feasibility, dual feasibility, and complementary slackness, making it particularly effective for solving problems with nonlinear constraints.

SDP (Semidefinite Programming) SDP reformulates a QCQP problem into a semidefinite programming relaxation (Freund,2004).

By “lifting” the problem to a higher-dimensional space and applying SDP relaxation, this approach provides a tractable way to solve or approximate solutions to non-convex QCQP problems. It is widely used in areas where global optimization or approximations to non-convex issues are necessary.

In general, analyzing QCQPs is important in order to apply knowledge-based decision-making and enhance the performance and stability of optimization methods in different fields (Zenios,1993).

Algorithm Discussion

KKT: What is KKT +formulation

SDP: What is SDP +formulation

Selected Solvers:

Name Brief Info

Numerical Example

Quadratically Constrained Quadratic Program (QCQP) always has the form:

where are n-by-n matrices and x is the optimization variable. If are all positive semidefinite, then the problem is convex. If these matrices are neither positive nor negative semidefinite, the problem is non-convex. If where are all zero, then the constraints are in fact linear and the problem is a quadratic program.

Example 1: KKT Approach

Considering the following numerical example:

(1) Steps

  • Formulate the Lagrangian
  • Computed the Gradients
  • Applied the Stationarity Conditions
  • Determined the active constraints using complementary slackness
Lagrangian Formulation:

The Lagrangian combines the objective function and the constraints, each multiplied by a Lagrange multiplier :

For each constraint:

- Complementary Slackness:

 

- Primal Feasibility:

 

Step 2: Compute the Gradient of the Lagrangian

Compute the partial derivatives with respect to and :

- Partial Derivative with respect to :

 

- Partial Derivative with respect to :

 

Step 3: Stationarity Conditions

Set the gradients to zero:

- Equation (1):

 

- Equation (2):

 

From Equation (2), since and for , it follows that:

Substitute into the constraints:

Combining both constraints:

Step 4: Solve the problem Using Equation (1)

Substitute into Equation (1):

Assume (since Constraint 1 is active):

But from the feasible range, .

Substitute into the equation:

This is acceptable.

Assume because Constraint 2 is not active at .

Step 5: Verify Complementary Slackness

- Constraint 1:

 

- Constraint 2:

 

Step 6: Verify Primal Feasibility

- Constraint 1:

 

- Constraint 2:

 

Step 7: Conclusion

- Optimal Solution:

 

- Minimum Objective Value:

 

- Result: $x_1^*=1, x_2^*=0$, with a minimum objective value of $f_0\left(x^*\right)=1$.

- Performance: The KKT method efficiently handles the convex QCQP problem but requires careful initialization to ensure convergence, especially for non-convex problems.

2. SDP Relaxation:

- Approach: The constraints were relaxed into semidefinite constraints, allowing the problem to be solved as a convex optimization task.

- Result: The SDP relaxation provided an identical solution for this convex problem, confirming its accuracy and reliability.

- Performance: While SDP offers robust solutions, it is computationally more intensive, particularly for large-scale problems with numerous constraints.


Example 2 : SDP- Based QCQP

Consider the following Quadratically Constrained Quadratic Programming (QCQP) problem:

Interpretation:

The objective is the squared distance from the origin.A point is found in the feasible region that is as close as possible to the origin.

The constraint restricts to lie inside or on a circle of radius . The constraint defines a hyperbolic region. To satisfy , both variables must be sufficiently large in magnitude and have the same sign.

Step 1: Lifting and Reformulation

Introduce the lifted variable:

If , then (positive semidefinite) and is rank-1.

Rewrite the objective and constraints in terms of :

Objective: , where is the 2x2 identity matrix.

Constraint 1:

Constraint 2:

Step 2: SDP Relaxation

The original QCQP is non-convex due to the rank-1 condition on . Relax the rank constraint and consider only :

This is now a Semidefinite Program (SDP), which is convex and can be solved using standard SDP solvers.

Step 3: Solving the SDP and Recovering the Solution

Solving the SDP, the feasible solution found that achieves the minimum:

Check that is rank-1:

with .

Thus, from

Check feasibility:

All constraints are satisfied.

Step 4: Optimal Value

The optimal objective value is:

Conclusion

The SDP relaxation not only provides a lower bound but also recovers the global optimum for this particular QCQP. The optimal solution is with an objective value of .

This example demonstrates how an SDP relaxation can be used effectively to solve a non-convex QCQP and, in some cases, recover the exact optimal solution.

Application

Conclusion

In conclusion, Quadratically Constrained Quadratic Programs (QCQPs) are a significant class of optimization problems extending quadratic programming by incorporating quadratic constraints (Bao,Sahinidis,2011). These problems are essential for modeling complex real-world scenarios where both the objective function and the constraints are quadratic. QCQPs are widely applicable in areas such as agriculture, finance, production planning, and machine learning, where they help optimize decisions by balancing competing factors such as profitability, risk, and resource constraints.

The study and solution of QCQPs are critical due to their ability to capture complex relationships and non-linearities, offering a more realistic representation of many practical problems than simpler linear models. Techniques such as Karush-Kuhn Tucker (KKT) conditions and semidefinite programming (SDP) relaxations provide effective tools for solving QCQPs(Elloumi & Lambert,2019), offering both exact and approximate solutions depending on the problem’s structure. These methods allow for efficient handling of the challenges posed by quadratic constraints and non-linearities.

Looking forward, there are several potential areas for improvement in QCQP algorithms. One direction is the development of more efficient relaxation techniques for solving non-convex QCQPs, especially in large-scale problems where computational efficiency becomes critical. Additionally, there is ongoing research into hybrid methods that combine the strengths of different optimization techniques, such as SDP and machine learning, to improve the robustness and speed of solving QCQPs in dynamic environments. As optimization problems become increasingly complex and data-rich, advancements in QCQP algorithms will continue to play a crucial role in making informed, optimal decisions in diverse applications.

Reference

[1] Agarwal, D., Singh, P., & El Sayed, M. A. (2023). The Karush–Kuhn–Tucker (KKT) optimality conditions for fuzzy-valued fractional optimization problems. Mathematics and Computers in Simulation, 205, 861-877, DOI:10.1016/j.matcom.2022.10.024

[2] Bao, X., Sahinidis, N. V., & Tawarmalani, M. (2011). Semidefinite relaxations for quadratically constrained quadratic programming: A review and comparisons (PDF). Mathematical programming, 129, 129-157.

[3] Bose, S., Gayme, D. F., Chandy, K. M., & Low, S. H. (2015). Quadratically constrained quadratic programs on acyclic graphs with application to power flow. IEEE Transactions on Control of Network Systems, 2(3), 278-287,DOI: 10.1109/TCNS.2015.2401172

[4] Elloumi, S., & Lambert, A. (2019). Global solution of non-convex quadratically constrained quadratic programs. Optimization methods and software, 34(1), 98-114,doi: https://doi.org/10.1080/10556788.2017.1350675

[5] Freund, R. M. (2004). Introduction to semidefinite programming (SDP) (PDF). Massachusetts Institute of Technology, 8-11.

[6] Ghojogh, B., Ghodsi, A., Karray, F., & Crowley, M. (2021). KKT conditions, first-order and second-order optimization, and distributed optimization: tutorial and survey. arXiv preprint arXiv:2110.01858.

[5] Zenios, S. A. (Ed.). (1993). Financial optimization. Cambridge university press,doi:https://doi.org/10.1017/CBO9780511522130

Further Reading

In Algorithm:

In Finance Portfolio Construction:

External Links

Category: NonLinear Programming (NLP) - Quadratic programming