<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://optimization.cbe.cornell.edu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ty262</id>
	<title>Cornell University Computational Optimization Open Textbook - Optimization Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://optimization.cbe.cornell.edu/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ty262"/>
	<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Special:Contributions/Ty262"/>
	<updated>2026-05-02T18:55:58Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.0</generator>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5542</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5542"/>
		<updated>2021-12-15T02:58:07Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region. Cauchy step  &amp;lt;math&amp;gt;P_{K}^{C}&amp;lt;/math&amp;gt; is an unexpensive method( no matrix factorization) to solve trust-region subproblem. Furthermore, Cauchy point has been valued due to the fact that it can globally convergent. Following is a closed-form equations of the Cauchy point.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_k^c=-\tau _k\frac{\Delta k}{\left \| \ g_k \right \|}\ g_k&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;g_k^TB_kg_k\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _k=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _k= min\left ( \left \| \bigtriangledown g_k \right \|^{3}/\left ( \bigtriangleup _k\bigtriangledown g_k^TB_kg_k \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although it is unexpensive to apply Cauchy point, steepest descent methods sometimes performs poorly. Thus, we introduce some improving strategy. The improvement strategies is based on &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; where it contains valid curvature information about the function.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This method can be used if  &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; is a positive definite. The dogleg method finds an approximate solution by replacing the curved trajectory&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;p^*\left ( \bigtriangleup  \right )&amp;lt;/math&amp;gt; with a path consisting of two line segments. It chooses p to minimize the model m along this path, subject to the trust-region bound.&lt;br /&gt;
&lt;br /&gt;
First line segments  &amp;lt;math&amp;gt;p^U=-\frac{g^Tg}{g^TBg}g  &amp;lt;/math&amp;gt;, where  &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;runs from the origin to the minimizer of m along the steepest descent direction.&lt;br /&gt;
&lt;br /&gt;
While the second line segment run from &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;to &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt;, we donate this trajectory by &amp;lt;math&amp;gt;\tilde{p}\left ( \tau  \right )&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;\tau \in \left [ 0,2 \right ]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then a V-shaped trajectory can be determined by &lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^U  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^U+\left (\tau -1 \right )\left ( p^B-p^U \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;p^B &amp;lt;/math&amp;gt;=opitimal solution of quadratic model &lt;br /&gt;
&lt;br /&gt;
Although the dogleg strategy can be adapted to handle indefinite B, there is not much point in doing so because the full step  &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt; is not the unconstrained minimizer of m in this case. Instead, we now describe another strategy, which aims to include directions of negative&lt;br /&gt;
&lt;br /&gt;
curvature  in the space of  trust-region steps.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Conjugated Gradient Steihaug’s Method ( CG-Steihaug)&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This is the most widely used method for the approximate solution of the trust-region problem. The method works for quadratic models &amp;lt;math&amp;gt;m_{k}&amp;lt;/math&amp;gt; defined by an arbitrary symmetric &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; . Thus, it is not necessary for &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; to be positive. CG-Steihaug has the advantage of Cauchy point calculation and Dogleg method which is super-linear convergence rate and unexpensive costs .&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_0=0,r_0=g,d_0=-r_0    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_0 \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;j=0,1,2,...&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_j^TB_kd_j\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt; such that minimizes &amp;lt;math&amp;gt;m\left ( p \right )&amp;lt;/math&amp;gt; and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _j=r_j^Tr_j/d_j^TB_kd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_j+\alpha _jd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_j+\alpha _jB_kd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_j^Tr_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Global Convergence&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To study the convergence of trust region, we have to study how much reduction can we achieve at each&lt;br /&gt;
&lt;br /&gt;
iteration (similar to line search method). Thus, we derive an estimate in the following form:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( 0 \right )-m_k\left ( p_k \right )\geq c_1\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup k,\frac{\left \|  \bigtriangledown f_k\right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;c_1\in \left [ 0,1 \right ]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For cauchy point, &amp;lt;math&amp;gt;c_1&amp;lt;/math&amp;gt;=0.5&lt;br /&gt;
&lt;br /&gt;
that is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( 0 \right )-m_k\left ( p_k \right )\geq 0.5\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup k,\frac{\left \|  \bigtriangledown f_k\right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
we first consider the case of &amp;lt;math&amp;gt;\bigtriangledown f_k^TB_k\bigtriangledown f_k\leq 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( p_k^c \right )-m_k\left ( 0 \right )\geq m_k\left ( \bigtriangleup _k\bigtriangledown f_k/\left \| \bigtriangledown f_k \right \| \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;=-\frac{\bigtriangleup _k}{\left \| \bigtriangledown f_k \right \|}\left \| \bigtriangledown f_k \right \|^2+0.5\frac{\bigtriangleup _k^2}{\left \| \bigtriangledown f_k \right \|^2}\ \bigtriangledown f_k^TB_k\bigtriangledown f_k&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -\bigtriangleup _k\left \| \bigtriangledown f_k \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup _k,\frac{\left \| \bigtriangledown f_k \right \|}{B_k} \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For the next case, consider &amp;lt;math&amp;gt;\bigtriangledown f_k^TB_k\bigtriangledown f_k&amp;gt; 0&amp;lt;/math&amp;gt; and&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\frac{\left \| \bigtriangledown f_k \right \|^3}{\bigtriangleup _k\bigtriangledown f_k^TB_k\bigtriangledown f_k}\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
we then have &amp;lt;math&amp;gt;\tau =\frac{\left \| \bigtriangledown f_k \right \|^3}{\bigtriangleup _k\bigtriangledown f_k^TB_k\bigtriangledown f_k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
so&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( p_k^c \right )-m_k\left ( 0 \right )= -\frac{\left \| \bigtriangledown f_k \right \|^4}{\bigtriangledown f_k^TB_k\bigtriangledown f_k}+0.5\bigtriangledown f_k^TB_k\bigtriangledown f_k\frac{\left \| \bigtriangledown f_k \right \|^4}{\left ( \bigtriangledown f_k^TB_k\bigtriangledown f_k \right )^2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;=-0.5\frac{\left \| \bigtriangledown f_k \right \|^4}{\bigtriangledown f_k^TB_k\bigtriangledown f_k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -0.5\frac{\left \| \bigtriangledown f_k \right \|^4}{\left \| B_k \right \|\left \| \bigtriangledown f_k \right \|^2}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;=-0.5\frac{\left \| \bigtriangledown f_k \right \|^2}{\left \| B_k \right \|}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -0.5\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup _k,\frac{\left \| \bigtriangledown f_k \right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt;,&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;\frac{\left \| \bigtriangledown f_k \right \|^3}{\bigtriangleup _k\bigtriangledown f_k^TB_k\bigtriangledown f_k}\leq 1&amp;lt;/math&amp;gt; does not hold, thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\bigtriangledown f_k^TB_k\bigtriangledown f_k&amp;lt; \frac{\left \| \bigtriangledown f_k \right \|^3}{\bigtriangleup _k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the definition of &amp;lt;math&amp;gt;p_c^k&amp;lt;/math&amp;gt; , we have &amp;lt;math&amp;gt;\tau =1&amp;lt;/math&amp;gt;, therefore&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( p_k^c \right )-m_k\left ( 0 \right )= -\frac{\bigtriangleup _k}{\left \| \bigtriangledown f_k \right \|}\left \| \bigtriangledown f_k \right \|^2+0.5\frac{\bigtriangleup _k^2}{\left \| \bigtriangledown f_k \right \|^2}\bigtriangledown f_k^TB_k\bigtriangledown f_k&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -\bigtriangleup _k\left \| \bigtriangledown f_k \right \|^2+0.5\frac{\bigtriangleup _k^2}{\left \| \bigtriangledown f_k \right \|^2}\frac{\left \| \bigtriangledown f_k \right \|^3}{\bigtriangleup _k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;=-0.5\bigtriangleup _k\left \| \bigtriangledown f_k \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\leq -0.5\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup _k,\frac{\left \| \bigtriangledown f_k \right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5383</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5383"/>
		<updated>2021-12-14T22:42:11Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region. Cauchy step  &amp;lt;math&amp;gt;P_{K}^{C}&amp;lt;/math&amp;gt; is an unexpensive method( no matrix factorization) to solve trust-region subproblem. Furthermore, Cauchy point has been valued due to the fact that it can globally convergent. Following is a closed-form equations of the Cauchy point.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_k^c=-\tau _k\frac{\Delta k}{\left \| \ g_k \right \|}\ g_k&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;g_k^TB_kg_k\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _k=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _k= min\left ( \left \| \bigtriangledown g_k \right \|^{3}/\left ( \bigtriangleup _k\bigtriangledown g_k^TB_kg_k \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although it is unexpensive to apply Cauchy point, steepest descent methods sometimes performs poorly. Thus, we introduce some improving strategy. The improvement strategies is based on &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; where it contains valid curvature information about the function.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This method can be used if  &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; is a positive definite. The dogleg method finds an approximate solution by replacing the curved trajectory&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;p^*\left ( \bigtriangleup  \right )&amp;lt;/math&amp;gt; with a path consisting of two line segments. It chooses p to minimize the model m along this path, subject to the trust-region bound.&lt;br /&gt;
&lt;br /&gt;
First line segments  &amp;lt;math&amp;gt;p^U=-\frac{g^Tg}{g^TBg}g  &amp;lt;/math&amp;gt;, where  &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;runs from the origin to the minimizer of m along the steepest descent direction.&lt;br /&gt;
&lt;br /&gt;
While the second line segment run from &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;to &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt;, we donate this trajectory by &amp;lt;math&amp;gt;\tilde{p}\left ( \tau  \right )&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;\tau \in \left [ 0,2 \right ]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then a V-shaped trajectory can be determined by &lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^U  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^U+\left (\tau -1 \right )\left ( p^B-p^U \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;p^B &amp;lt;/math&amp;gt;=opitimal solution of quadratic model &lt;br /&gt;
&lt;br /&gt;
Although the dogleg strategy can be adapted to handle indefinite B, there is not much point in doing so because the full step  &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt; is not the unconstrained minimizer of m in this case. Instead, we now describe another strategy, which aims to include directions of negative&lt;br /&gt;
&lt;br /&gt;
curvature  in the space of  trust-region steps.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Conjugated Gradient Steihaug’s Method ( CG-Steihaug)&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This is the most widely used method for the approximate solution of the trust-region problem. The method works for quadratic models &amp;lt;math&amp;gt;m_{k}&amp;lt;/math&amp;gt; defined by an arbitrary symmetric &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; . Thus, it is not necessary for &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; to be positive. CG-Steihaug has the advantage of Cauchy point calculation and Dogleg method which is super-linear convergence rate and unexpensive costs .&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_0=0,r_0=g,d_0=-r_0    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_0 \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;j=0,1,2,...&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_j^TB_kd_j\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt; such that minimizes &amp;lt;math&amp;gt;m\left ( p \right )&amp;lt;/math&amp;gt; and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _j=r_j^Tr_j/d_j^TB_kd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_j+\alpha _jd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_j+\alpha _jB_kd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_j^Tr_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Global Convergence&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
To study the convergence of trust region, we have to study how much reduction can we achieve at each&lt;br /&gt;
&lt;br /&gt;
iteration (similar to line search method). Thus, we derive an estimate in the following form:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( 0 \right )-m_k\left ( p_k \right )\geq c_1\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup k,\frac{\left \|  \bigtriangledown f_k\right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;c_1\in \left [ 0,1 \right ]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
For cauchy point, &amp;lt;math&amp;gt;c_1&amp;lt;/math&amp;gt;=0.5&lt;br /&gt;
&lt;br /&gt;
that is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( 0 \right )-m_k\left ( p_k \right )\geq 0.5\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup k,\frac{\left \|  \bigtriangledown f_k\right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
first consider the case of &amp;lt;math&amp;gt;\bigtriangledown f_k^TB_k\bigtriangledown f_k\leq 0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( p_k^c \right )-m_k\left ( 0 \right )\geq m_k\left ( \bigtriangleup _k\bigtriangledown f_k/\left \| \bigtriangledown f_k \right \| \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5364</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5364"/>
		<updated>2021-12-14T20:24:59Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region. Cauchy step  &amp;lt;math&amp;gt;P_{K}^{C}&amp;lt;/math&amp;gt; is an unexpensive method( no matrix factorization) to solve trust-region subproblem. Furthermore, Cauchy point has been valued due to the fact that it can globally convergent. Following is a closed-form equations of the Cauchy point.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_k^c=-\tau _k\frac{\Delta k}{\left \| \ g_k \right \|}\ g_k&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;g_k^TB_kg_k\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _k=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _k= min\left ( \left \| \bigtriangledown g_k \right \|^{3}/\left ( \bigtriangleup _k\bigtriangledown g_k^TB_kg_k \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although it is unexpensive to apply Cauchy point, steepest descent methods sometimes performs poorly. Thus, we introduce some improving strategy. The improvement strategies is based on &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; where it contains valid curvature information about the function.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This method can be used if  &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; is a positive definite. The dogleg method finds an approximate solution by replacing the curved trajectory&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;p^*\left ( \bigtriangleup  \right )&amp;lt;/math&amp;gt; with a path consisting of two line segments. It chooses p to minimize the model m along this path, subject to the trust-region bound.&lt;br /&gt;
&lt;br /&gt;
First line segments  &amp;lt;math&amp;gt;p^U=-\frac{g^Tg}{g^TBg}g  &amp;lt;/math&amp;gt;, where  &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;runs from the origin to the minimizer of m along the steepest descent direction.&lt;br /&gt;
&lt;br /&gt;
While the second line segment run from &amp;lt;math&amp;gt;p^U&amp;lt;/math&amp;gt;to &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt;, we donate this trajectory by &amp;lt;math&amp;gt;\tilde{p}\left ( \tau  \right )&amp;lt;/math&amp;gt; for &amp;lt;math&amp;gt;\tau \in \left [ 0,2 \right ]&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Then a V-shaped trajectory can be determined by &lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^U  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^U+\left (\tau -1 \right )\left ( p^B-p^U \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;p^B &amp;lt;/math&amp;gt;=opitimal solution of quadratic model &lt;br /&gt;
&lt;br /&gt;
Although the dogleg strategy can be adapted to handle indefinite B, there is not much point in doing so because the full step  &amp;lt;math&amp;gt;p^B&amp;lt;/math&amp;gt; is not the unconstrained minimizer of m in this case. Instead, we now describe another strategy, which aims to include directions of negative&lt;br /&gt;
&lt;br /&gt;
curvature  in the space of  trust-region steps.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Conjugated Gradient Steihaug’s Method ( CG-Steihaug)&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This is the most widely used method for the approximate solution of the trust-region problem. The method works for quadratic models &amp;lt;math&amp;gt;m_{k}&amp;lt;/math&amp;gt; defined by an arbitrary symmetric &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; . Thus, it is not necessary for &amp;lt;math&amp;gt;B_k&amp;lt;/math&amp;gt; to be positive. CG-Steihaug has the advantage of Cauchy point calculation and Dogleg method which is super-linear convergence rate and unexpensive costs .&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_0=0,r_0=g,d_0=-r_0    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_0 \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p0&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;j=0,1,2,...&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_j^TBd_j\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt; such that minimizes &amp;lt;math&amp;gt;m\left ( p \right )&amp;lt;/math&amp;gt; and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _j=r_j^Tr_j/d_j^TBd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_j+\alpha _jd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_j+\alpha _jBd_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_j^Tr_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_j&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Global Convergence&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;m_k\left ( 0 \right )-m_k\left ( p_k \right )\geq c_1\left \| \bigtriangledown f_k \right \|min\left ( \bigtriangleup k,\frac{\left \|  \bigtriangledown f_k\right \|}{\left \| B_k \right \|} \right )&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5359</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5359"/>
		<updated>2021-12-14T15:37:45Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region. Cauchy step  &amp;lt;math&amp;gt;P_{K}^{C}&amp;lt;/math&amp;gt; is an unexpensive method( no matrix factorization) to solve trust-region subproblem. Furthermore, Cauchy point has been valued due to the fact that it can globally convergent. Following is a closed-form equations of the Cauchy point.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{k}^{c}=-\tau _{k}\frac{\Delta k}{\left \| \ g_{k} \right \|}\ g_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;g_{k}^{T}B_{k}g_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown g_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown g_{k}^{T}B_{k}g_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although it is unexpensive to apply Cauchy point, steepest descent methods sometimes performs poorly. Thus, we introduce some improving strategy. The improvement strategies is based on &amp;lt;math&amp;gt;B_{k}&amp;lt;/math&amp;gt; where it contains valid curvature information about the function.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This method can be used if  &amp;lt;math&amp;gt;B_{k}&amp;lt;/math&amp;gt; is a positive definite. The dogleg method finds an approximate solution by replacing the curved trajectory&lt;br /&gt;
&lt;br /&gt;
for &amp;lt;math&amp;gt;p^{*}\left ( \bigtriangleup  \right )&amp;lt;/math&amp;gt; with a path consisting of two line segments.&lt;br /&gt;
&lt;br /&gt;
First line segments  &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;, where  &amp;lt;math&amp;gt;p^{U}&amp;lt;/math&amp;gt;runs from the origin to the minimizer of m along the steepest descent direction.&lt;br /&gt;
&lt;br /&gt;
While the second line segment run from &amp;lt;math&amp;gt;p^{U}&amp;lt;/math&amp;gt;to &amp;lt;math&amp;gt;p^{B}&amp;lt;/math&amp;gt;,then a V-shaped trajectory can be determined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,dogleg method is limited to the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Global Convergence&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5284</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=5284"/>
		<updated>2021-12-14T03:27:16Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Methodology and theory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region. Cauchy step  &amp;lt;math&amp;gt;P_{K}^{C}&amp;lt;/math&amp;gt; is an unexpensive method( no matrix factorization) to solve trust-region subproblem. Furthermore, Cauchy point has been valued due to the fact that it can globally convergent. Following is a closed-form equations of the Cauchy point.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{k}^{c}=-\tau _{k}\frac{\Delta k}{\left \| \ g_{k} \right \|}\ g_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;g_{k}^{T}B_{k}g_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown g_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown g_{k}^{T}B_{k}g_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Although it is unexpensive to apply Cauchy point, steepest descent methods sometimes performs poorly. Thus, we introduce some improving strategy. The improvement strategies is based on &amp;lt;math&amp;gt;B_{k}&amp;lt;/math&amp;gt; where it contains valid curvature information about the function.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
This method can be used if  &amp;lt;math&amp;gt;B_{k}&amp;lt;/math&amp;gt; is a positive definite.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;(&amp;lt;math&amp;gt;p^{U}&amp;lt;/math&amp;gt;runs from the origin to the minimizer of m along the steepest descent direction),  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,dogleg method is limited to the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Global Convergence&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4860</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4860"/>
		<updated>2021-12-06T15:35:28Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16 (1,1).]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4859</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4859"/>
		<updated>2021-12-06T15:32:43Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:Trust-region method example trajectory.png|thumb|576x576px|Trust-region method trajectory of Rosenbrock function starting from (0,0). The data points represent the optimal solutions after each iteration, ending at iteration number 16.]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:Trust-region_method_example_trajectory.png&amp;diff=4858</id>
		<title>File:Trust-region method example trajectory.png</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:Trust-region_method_example_trajectory.png&amp;diff=4858"/>
		<updated>2021-12-06T15:26:14Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Trust-region method example trajectory&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4857</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4857"/>
		<updated>2021-12-06T15:25:06Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=4856</id>
		<title>File:Optimization trajectory of the example.jpg</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=4856"/>
		<updated>2021-12-06T15:24:13Z</updated>

		<summary type="html">&lt;p&gt;Ty262: Ty262 uploaded a new version of File:Optimization trajectory of the example.jpg&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Optimization trajectory of the example&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:TRM_Trajectory.jpg&amp;diff=4855</id>
		<title>File:TRM Trajectory.jpg</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:TRM_Trajectory.jpg&amp;diff=4855"/>
		<updated>2021-12-06T15:03:22Z</updated>

		<summary type="html">&lt;p&gt;Ty262: Ty262 uploaded a new version of File:TRM Trajectory.jpg&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Optimization trajectory of example&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4854</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4854"/>
		<updated>2021-12-06T14:58:08Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4846</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4846"/>
		<updated>2021-12-06T14:16:05Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960&amp;lt;ref&amp;gt;H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, &amp;lt;nowiki&amp;gt;https://doi.org/10.1093/comjnl/3.3.175&amp;lt;/nowiki&amp;gt;&amp;lt;/ref&amp;gt;, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[5] H. H. Rosenbrock, An Automatic Method for Finding the Greatest or Least Value of a Function, &#039;&#039;The Computer Journal&#039;&#039;, Volume 3, Issue 3, 1960, Pages 175–184, https://doi.org/10.1093/comjnl/3.3.175&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4845</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4845"/>
		<updated>2021-12-06T14:12:08Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s &amp;lt;code&amp;gt;fminunc&amp;lt;/code&amp;gt; as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4844</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4844"/>
		<updated>2021-12-06T14:05:43Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang (CHEME 6800, Fall 2021)&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4843</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4843"/>
		<updated>2021-12-06T14:04:01Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou, Ting-Guang Yeh, Yun-Chung Pan, Chen-Hua Wang, Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4470</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4470"/>
		<updated>2021-11-29T02:19:23Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
&lt;br /&gt;
Trust region method is a numerical optimization method that is employed to solve non-linear programming (NLP) problem. Instead of finding objective solution of the original function f, in each step the method defines a neighborhood around current best solution xk as a trust region f’ (typically by using quadratic model), which is capable of representing the f function appropriately, in order to derive the next point xk+1. Different from line search, the model selects the direction and step size simultaneously. For example, in a minimization problem, if the decrease in the value of optimal solution is not sufficient, we can conclude that the region is too large to get close to the minimizer of the objective function, so we will shrink the f’ to find the solution again. On the other hand, if such decrease is significant, it is believed that the model has adequate representation to the problem. Generally, the step direction depends on extent that the region is altered in the previous iteration.&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
Trusted region is a powerful method that can update the objective function in each step to ensure the model is always getting improved while keeping the previously learnt knowledge as the baseline. Nowadays, trust region algorithms are widely used in machine learning, applied mathematics, physics, chemistry, biology, etc. It is believed that the trust region method will have more far-reaching development in a wider range of fields in the near future.&lt;br /&gt;
&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] S. Boyd, L. Vandenberghe, Convex Optimization. Cambridge University Press, 2009&lt;br /&gt;
&lt;br /&gt;
[4] Trust region. (2020). Retrieved November 10, 2021, from &amp;lt;nowiki&amp;gt;https://en.wikipedia.org/wiki/Trust_region&amp;lt;/nowiki&amp;gt;.&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4455</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4455"/>
		<updated>2021-11-29T02:07:18Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;br /&gt;
[1] J. Nocedal, S. J. Wright, &#039;&#039;Numerical Optimization&#039;&#039;. Springer, 1999.&lt;br /&gt;
&lt;br /&gt;
[2] W. Sun and Y.-x. Yuan, Optimization theory and methods : nonlinear programming. New York: Springer, 2006.&lt;br /&gt;
&lt;br /&gt;
[3] Wikipedia page for Trust region.&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4447</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4447"/>
		<updated>2021-11-29T02:01:48Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Similar to line serach method which do not require optimal step lengths to be convergent, trust-region method is suffficient for global convergence purpose to find an approximate solution pk that lies within trust region.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The reason we improving on the couchy point is that we just implementing the steepest decent method with a particular choice of step length by using cauchy point. However, steepest decent performs poorly even if an optimial step length is used at every iteration.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The dogleg method chooses p to minimize the model m along this path.Furthermore,gogled method is limited to&lt;br /&gt;
&lt;br /&gt;
the trust-region bound. Since the dogleg path interacts the trust region at most one points and intersectin can be computed anlytically, there is no need to conduct a search.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal solution of quadratic model&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set   &amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt;\alpha _{j}=r_{j}^{T}r_{j}/d_{j}^{T}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set  &amp;lt;math&amp;gt; p_{j+1}=p_{j}+\alpha _{j}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\left \| p_{j+1} \right \|\geq \Delta&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau \geq 0&amp;lt;/math&amp;gt; such that &amp;lt;math&amp;gt;p=p_{j}+\tau d_{j}&amp;lt;/math&amp;gt; satisfies &amp;lt;math&amp;gt;\left \| p \right \|=\Delta &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;r_{j+1}=r_{j}+\alpha _{j}Bd_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt; \left \| r_{j+1} \right \|&amp;lt; \epsilon \left \| r_{0} \right \|&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return &amp;lt;math&amp;gt;p=p_{j+1}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;\beta _{j+1} = r_{j+1}^{T}r_{j+1}/r_{j}^{T}r_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set &amp;lt;math&amp;gt;d_{j+1}= r_{j+1}+ \beta _{j+1}d_{j}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
end(for)&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4430</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4430"/>
		<updated>2021-11-29T01:22:57Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Methodology and theory  */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal point&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set&amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if&amp;lt;math&amp;gt;\left \| r_{0} \right \|&amp;lt; \epsilon   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p=p0&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;d_{j}^{T}Bd_{j}\leq 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Find &amp;lt;math&amp;gt;\tau   &amp;lt;/math&amp;gt;such that minimizes m(p) and satisfies&amp;lt;math&amp;gt;\left \| p \right \|=\Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set&lt;br /&gt;
&lt;br /&gt;
Set&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4426</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4426"/>
		<updated>2021-11-29T01:19:04Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Methodology and theory  */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
otherewise, &amp;lt;math&amp;gt;\tau _{k}= min\left ( \left \| \bigtriangledown f_{k} \right \|^{3}/\left ( \bigtriangleup _{k}\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k} \right ),1 \right )  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;min m(p)=f+g^{T}p+\frac{1}{2}p^{T}B_{p}  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p\in R^{n}  &amp;lt;/math&amp;gt; s.t. &amp;lt;math&amp;gt;\left \| p \right \|\leq \Delta   &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}=\tau p^{U}  &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;0\leq \tau \leq 1  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\tilde{p}= p^{U}+\left (\tau -1 \right )\left ( p^{B}-p^{U} \right )   &amp;lt;/math&amp;gt;, when &amp;lt;math&amp;gt;1\leq \tau \leq 2  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
where &amp;lt;math&amp;gt;p^{U}=-\frac{g^{T}g}{g^{T}Bg}g  &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;p^{B}  &amp;lt;/math&amp;gt;=opitimal point&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Given&amp;lt;math&amp;gt;\epsilon &amp;gt; 0  &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Set&amp;lt;math&amp;gt;p_{0}=0,r_{0}=g,d_{0}=-r_{0}    &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
return&lt;br /&gt;
&lt;br /&gt;
for j=0,1,2,.....&lt;br /&gt;
&lt;br /&gt;
if&lt;br /&gt;
&lt;br /&gt;
Find such that minimizes m(p) and satisfies&lt;br /&gt;
&lt;br /&gt;
return p;&lt;br /&gt;
&lt;br /&gt;
Set&lt;br /&gt;
&lt;br /&gt;
Set&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4414</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4414"/>
		<updated>2021-11-29T01:02:53Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Methodology and theory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
&#039;&#039;&#039;Cauchy point calculation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;p_{c}^{K}=-\tau _{k}\frac{\Delta k}{\left \| \bigtriangledown f_{k} \right \|}\bigtriangledown f_{k}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
if &amp;lt;math&amp;gt;\bigtriangledown f_{k}^{T}B_{k}\bigtriangledown f_{k}\leq 0&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\tau _{k}=1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Improving on the couchy point&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Dogleg method&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Steihau&#039;s approach&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4408</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=4408"/>
		<updated>2021-11-29T00:44:11Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=Methodology and theory=&lt;br /&gt;
Cauchy point calculation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Improving on the couchy point&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Dogleg method&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Steihau&#039;s approach&lt;br /&gt;
&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
&#039;&#039;&#039;Approaches for a convex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
* Levenberg–Marquardt Steps&lt;br /&gt;
&lt;br /&gt;
The objective function is iteratively approximated by a quadratic surface, and the estimate is updated. However, this method may fail to converge if the initial guess is too large or  &lt;br /&gt;
&lt;br /&gt;
too small. The trust region methods help to control the trust region size, preventing each step from being too far from the optimum.&lt;br /&gt;
&lt;br /&gt;
* Powell Dogleg Steps&lt;br /&gt;
&lt;br /&gt;
Compared to Levenberg–Marquardt steps, the objective function is iteratively approximated by a linear surface, and the estimate is updated as well. Besides, the Levenberg–&lt;br /&gt;
&lt;br /&gt;
Marquardt steps provide an exact solution to the model problem, while the dogleg method provides only an approximate solution.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Approaches for a nonconvex model problem&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
If the exact Hessian is not positive definite, then a dogleg method that relies on Newton steps may be ascent steps or unbounded steps. Furthermore, the Levenberg–Marquardt method may not be flexible to exploit the size of the trust region. However, the trust region method can used to iteratively minimize a quadratic model of the Lagrangian subject to a possibly relaxed linearization of the problem constraints and a trust region constraint.&lt;br /&gt;
&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:CodeCog_1.gif&amp;diff=4161</id>
		<title>File:CodeCog 1.gif</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:CodeCog_1.gif&amp;diff=4161"/>
		<updated>2021-11-28T18:28:43Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;1&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3756</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3756"/>
		<updated>2021-11-28T00:43:20Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3519</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3519"/>
		<updated>2021-11-27T06:09:56Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting-Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3515</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3515"/>
		<updated>2021-11-27T06:04:01Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;[[File:TRM Trajectory.jpg|thumb|557x557px|Optimization trajectory of the example]]&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.25&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a good prediction, which increases the trust-region&#039;s size. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 7:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.765122406&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.560476539&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. The new iteration gives a poor prediction, which decreases the trust-region&#039;s size to improve the model&#039;s validity. The new optimal solution within the trust-region is &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;, &#039;&#039;&#039; &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 8:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.804352654&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.645444179&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;.The new iteration gives a poor prediction, therefore current best solution is unchanged and the radius for the trust-region is decreased.&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3512</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3512"/>
		<updated>2021-11-27T05:47:22Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm which uses the preconditioned conjugate method. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
[[File:TRM Trajectory.jpg|thumb|565x565px|Optimization trajectory of the example]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; The algorithm starts from the initial point of  &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;. The Rosenbrock function is visualized with a color coded map. For the first iteration, a full step was taken and the optimal solution (&amp;lt;math&amp;gt;x=0.25&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;) within the trust-region is denoted as a red dot. For each consecutive iteration, &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 2:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.263177536&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.061095029&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 3:&#039;&#039;&#039;  Start with  &amp;lt;math&amp;gt;x=0.371151679&lt;br /&gt;
&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;y=0.124075855&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 4:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.539493472&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.262714248&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 5:&#039;&#039;&#039;  Start with  &#039;&#039;&#039;&amp;lt;math&amp;gt;x=0.608557768&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=0.36557268&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
At the 16th iteration, the global optimal solution is found, &#039;&#039;&#039;&amp;lt;math&amp;gt;x=1&lt;br /&gt;
&amp;lt;/math&amp;gt;,  &amp;lt;math&amp;gt;y=1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/math&amp;gt;&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:TRM_Trajectory.jpg&amp;diff=3510</id>
		<title>File:TRM Trajectory.jpg</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:TRM_Trajectory.jpg&amp;diff=3510"/>
		<updated>2021-11-27T05:29:32Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Optimization trajectory of example&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=3509</id>
		<title>File:Optimization trajectory of the example.jpg</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=3509"/>
		<updated>2021-11-27T05:24:07Z</updated>

		<summary type="html">&lt;p&gt;Ty262: Ty262 uploaded a new version of File:Optimization trajectory of the example.jpg&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Optimization trajectory of the example&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=3508</id>
		<title>File:Optimization trajectory of the example.jpg</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=File:Optimization_trajectory_of_the_example.jpg&amp;diff=3508"/>
		<updated>2021-11-27T05:19:25Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Optimization trajectory of the example&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3507</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3507"/>
		<updated>2021-11-27T05:16:43Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iterations&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3506</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3506"/>
		<updated>2021-11-27T05:12:58Z</updated>

		<summary type="html">&lt;p&gt;Ty262: /* Numerical example */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
Here we will use the trust-region method to solve a classic optimization problem, the Rosenbrock function. The Rosenbrock function is a non-convex function, introduced by Howard H. Rosenbrock in 1960, which is often used as a performance test problem for optimization algorithms. This problem is solved using MATLAB&#039;s fminunc as the solver, with &#039;trust-region&#039; as the solving algorithm. &lt;br /&gt;
&lt;br /&gt;
The function is defined by&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\min f(x,y)=100(y-x^2)^2+(1-x)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The starting point chosen is &amp;lt;math&amp;gt;x=0&amp;lt;/math&amp;gt;  &amp;lt;math&amp;gt;y=0&lt;br /&gt;
&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration Process&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Iteration 1:&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+Summary of all iterations&lt;br /&gt;
!Iteration&lt;br /&gt;
!f(x)&lt;br /&gt;
!x&lt;br /&gt;
!y&lt;br /&gt;
!Norm of step&lt;br /&gt;
!First-order optimality&lt;br /&gt;
|-&lt;br /&gt;
|1&lt;br /&gt;
|1 &lt;br /&gt;
|0.25&lt;br /&gt;
|0&lt;br /&gt;
|1&lt;br /&gt;
|2&lt;br /&gt;
|-&lt;br /&gt;
|2&lt;br /&gt;
|0.953125&lt;br /&gt;
|0.263178&lt;br /&gt;
|0.061095&lt;br /&gt;
|0.25&lt;br /&gt;
|12.5&lt;br /&gt;
|-&lt;br /&gt;
|3&lt;br /&gt;
|0.549578 &lt;br /&gt;
|0.371152&lt;br /&gt;
|0.124076&lt;br /&gt;
|0.0625&lt;br /&gt;
|1.63&lt;br /&gt;
|-&lt;br /&gt;
|4&lt;br /&gt;
|0.414158&lt;br /&gt;
|0.539493&lt;br /&gt;
|0.262714&lt;br /&gt;
|0.125&lt;br /&gt;
|2.74&lt;br /&gt;
|-&lt;br /&gt;
|5&lt;br /&gt;
|0.292376 &lt;br /&gt;
|0.608558&lt;br /&gt;
|0.365573&lt;br /&gt;
|0.218082&lt;br /&gt;
|5.67&lt;br /&gt;
|-&lt;br /&gt;
|6&lt;br /&gt;
|0.155502&lt;br /&gt;
|0.765122&lt;br /&gt;
|0.560477&lt;br /&gt;
|0.123894&lt;br /&gt;
|0.954&lt;br /&gt;
|-&lt;br /&gt;
|7&lt;br /&gt;
|0.117347&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.25&lt;br /&gt;
|7.16&lt;br /&gt;
|-&lt;br /&gt;
|8&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.804353&lt;br /&gt;
|0.645444&lt;br /&gt;
|0.093587&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|9&lt;br /&gt;
|0.0385147&lt;br /&gt;
|0.836966&lt;br /&gt;
|0.69876&lt;br /&gt;
|0.284677&lt;br /&gt;
|0.308&lt;br /&gt;
|-&lt;br /&gt;
|10&lt;br /&gt;
|0.0268871&lt;br /&gt;
|0.90045&lt;br /&gt;
|0.806439&lt;br /&gt;
|0.0625&lt;br /&gt;
|0.351&lt;br /&gt;
|-&lt;br /&gt;
|11&lt;br /&gt;
|0.0118213&lt;br /&gt;
|0.953562&lt;br /&gt;
|0.90646&lt;br /&gt;
|0.125&lt;br /&gt;
|1.38&lt;br /&gt;
|-&lt;br /&gt;
|12&lt;br /&gt;
|0.0029522&lt;br /&gt;
|0.983251&lt;br /&gt;
|0.9659&lt;br /&gt;
|0.113247&lt;br /&gt;
|0.983&lt;br /&gt;
|-&lt;br /&gt;
|13&lt;br /&gt;
|0.000358233&lt;br /&gt;
|0.99749&lt;br /&gt;
|0.994783&lt;br /&gt;
|0.066442&lt;br /&gt;
|0.313&lt;br /&gt;
|-&lt;br /&gt;
|14&lt;br /&gt;
|1.04121e-05&lt;br /&gt;
|0.999902&lt;br /&gt;
|0.999799&lt;br /&gt;
|0.032202&lt;br /&gt;
|0.0759&lt;br /&gt;
|-&lt;br /&gt;
|15&lt;br /&gt;
|1.2959e-08&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.005565&lt;br /&gt;
|0.00213&lt;br /&gt;
|-&lt;br /&gt;
|16&lt;br /&gt;
|2.21873e-14&lt;br /&gt;
|1&lt;br /&gt;
|1&lt;br /&gt;
|0.000224&lt;br /&gt;
|3.59E-06&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3492</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=3492"/>
		<updated>2021-11-27T02:18:18Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;br /&gt;
=Introduction=&lt;br /&gt;
=Problem formulation=&lt;br /&gt;
=Numerical example=&lt;br /&gt;
=Applications=&lt;br /&gt;
=Conclusion=&lt;br /&gt;
=References=&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=2957</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=2957"/>
		<updated>2021-10-22T19:40:56Z</updated>

		<summary type="html">&lt;p&gt;Ty262: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou (cc2398), Ting Guang Yeh (ty262), Yun-Chung Pan (yp392), Chen-Hua Wang (cw893), Fall 2021&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
	<entry>
		<id>https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=2956</id>
		<title>Trust-region methods</title>
		<link rel="alternate" type="text/html" href="https://optimization.cbe.cornell.edu/index.php?title=Trust-region_methods&amp;diff=2956"/>
		<updated>2021-10-22T19:40:24Z</updated>

		<summary type="html">&lt;p&gt;Ty262: Created page with &amp;quot;Autor: Chun-Yu Chou(cc2398) Ting Guang Yeh(ty262) Yun-Chung Pan (yp392) Chen-Hua Wang (cw893) Fall 2021&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Autor: Chun-Yu Chou(cc2398) Ting Guang Yeh(ty262) Yun-Chung Pan (yp392) Chen-Hua Wang (cw893) Fall 2021&lt;/div&gt;</summary>
		<author><name>Ty262</name></author>
	</entry>
</feed>