Python Scipy Optimization.minimize using SLSQP showing maximized results

14.2k views Asked by At

I am learning to optimize a multivariate constrained nonlinear problem with scipy.optimize.minimize,but received strange results.

My problem:

minimize objfun

objfun   x*y

constraints 0<=x<=5,  0<=y<=5,  x+y==5

My code:

from scipy import optimize
def func(x):

    return x[0]*x[1]

bnds=((0,100),(0,5))

cons=({'type':'eq','fun':lambda x:x[0]+x[1]-5})
x0=[0,0]
res= optimize.minimize(func,x0,method='SLSQP',bounds=bnds,constraints=cons)

Received results:

status: 0 success: True njev: 2 nfev: 8 fun: 6.2499999999999991 x: array([ 2.5, 2.5]) message: 'Optimization terminated successfully.' jac: array([ 2.5, 2.5, 0. ]) nit: 2

I am expecting the fun to be 0 or significantly close to 0 and x or y to be 0

1

There are 1 answers

1
gg349 On BEST ANSWER

I think you are hitting a edge case. If you try with a guess that is not symmetric, you converge to the right solution.

Just change x0=[0,0] to something else, like x0=[.2,.9].

EDIT: expanding after @pv comment.

[x,y]=[2.5,2.5] is a local maximum of the constrained function. After jumping to this local maximum, the algorithm calculates again the direction it should take to minimize the target.

It does so by calculating the value at [ 2.50000001 2.5 ] and at [ 2.5 2.50000001]. It finds that this direction is (-1,-1). This direction is however orthogonal to the constraint, and it then stops.

The problem arises because the target and the constraint are symmetric with respect to x=y, and that we are starting with the guess exactly on x=y.