Quick start guide to optimization

In this example, we perform the optimization of the Rosenbrock test function.

Let a, b\in\mathbb{R} be parameters. The Rosenbrock function is defined by

f(x_1, x_2) = (a-x_1)^2 + b(x_2 - x_1^2)^2

for any \mathbf{x}\in\mathbb{R}^2. This function is often used with a=1 and b=100. In this case, the function has a single global minimum at:

\mathbf{x}^\star = (1, 1)^T.

This function has a nonlinear least squares structure.

References

  • Rosenbrock, H.H. (1960). “An automatic method for finding the greatest or least value of a function”. The Computer Journal. 3 (3): 175–184.

Definition of the function

import openturns as ot
import openturns.viewer as viewer
from matplotlib import pylab as plt

ot.Log.Show(ot.Log.NONE)
rosenbrock = ot.SymbolicFunction(["x1", "x2"], ["(1-x1)^2+100*(x2-x1^2)^2"])
x0 = [-1.0, 1.0]
xexact = ot.Point([1.0, 1.0])
lowerbound = [-2.0, -2.0]
upperbound = [2.0, 2.0]

Plot the iso-values of the objective function

rosenbrock = ot.MemoizeFunction(rosenbrock)
graph = rosenbrock.draw(lowerbound, upperbound, [100] * 2)
graph.setTitle("Rosenbrock function")
view = viewer.View(graph)
Rosenbrock function

We see that the minimum is on the top right of the picture and the starting point is on the top left of the picture. Since the function has a long valley following the curve x_2 - x^2=0, the algorithm generally have to follow the bottom of the valley.

Create and solve the optimization problem

problem = ot.OptimizationProblem(rosenbrock)
algo = ot.Cobyla(problem)
algo.setMaximumRelativeError(1.0e-1)  # on x
algo.setMaximumEvaluationNumber(50000)
algo.setStartingPoint(x0)
algo.run()
result = algo.getResult()
xoptim = result.getOptimalPoint()
xoptim
class=Point name=Unnamed dimension=2 values=[0.992501,0.985019]


delta = xexact - xoptim
absoluteError = delta.norm()
absoluteError
0.016752861417341336

We see that the algorithm found an accurate approximation of the solution.

result.getOptimalValue()  # f(x*)
class=Point name=Unnamed dimension=1 values=[5.63863e-05]


result.getEvaluationNumber()
10520
graph = rosenbrock.draw(lowerbound, upperbound, [100] * 2)
cloud = ot.Cloud(ot.Sample([x0, xoptim]))
cloud.setColor("black")
cloud.setPointStyle("bullet")
graph.add(cloud)
graph.setTitle("Rosenbrock function")
view = viewer.View(graph)
Rosenbrock function

We see that the algorithm had to start from the top left of the banana and go to the top right.

graph = result.drawOptimalValueHistory()
view = viewer.View(graph)
Optimal value history

The function value history make the path of the algorithm clear. In the first step, the algorithm went in the valley, which made the function value decrease rapidly. Once there, the algorithm had to follow the bottom of the valley so that the function decreased but slowly. In the final steps, the algorithm found the neighbourhood of the minimum so that the local convergence could take place.

In order to see where the function was evaluated, we use the getInputSample method.

inputSample = result.getInputSample()
graph = rosenbrock.draw(lowerbound, upperbound, [100] * 2)
graph.setTitle("Rosenbrock function solved with Cobyla")
cloud = ot.Cloud(inputSample)
graph.add(cloud)
view = viewer.View(graph)
Rosenbrock function solved with Cobyla

We see that the algorithm made lots of evaluations in the bottom of the valley before getting in the neighbourhood of the minimum.

Solving the problem with NLopt

We see that the Cobyla algorithm required lots of function evaluations. This is why we now use the NLopt class with the LBFGS algorithm. However, the algorithm may use input points which are far away from the input domain we used so far. This is why we had bounds to the problem, so that the algorithm never goes to far away from the valley.

bounds = ot.Interval(lowerbound, upperbound)
problem = ot.OptimizationProblem(rosenbrock)
problem.setBounds(bounds)
algo = ot.NLopt(problem, "LD_LBFGS")
algo.setStartingPoint(x0)
algo.run()
result = algo.getResult()
xoptim = result.getOptimalPoint()
xoptim
class=Point name=Unnamed dimension=2 values=[1,1]


delta = xexact - xoptim
absoluteError = delta.norm()
absoluteError
7.740583643426769e-12

We see that the algorithm found an extremely accurate approximation of the solution.

result.getOptimalValue()  # f(x*)
class=Point name=Unnamed dimension=1 values=[1.77616e-23]


result.getEvaluationNumber()
44

This number of iterations is much less than the previous experiment.

inputSample = result.getInputSample()
graph = rosenbrock.draw(lowerbound, upperbound, [100] * 2)
graph.setTitle("Rosenbrock function solved with NLopt/LD_LBFGS")
cloud = ot.Cloud(inputSample)
graph.add(cloud)
view = viewer.View(graph)

plt.show()
Rosenbrock function solved with NLopt/LD_LBFGS