Search All of the Math Forum:

Views expressed in these public forums are not endorsed by NCTM or The Math Forum.

Notice: We are no longer accepting new posts, but the forums will continue to be readable.

Topic: How to define Hessian for fminunc.
Replies: 3   Last Post: Oct 15, 2013 8:22 AM

 Messages: [ Previous | Next ]
 Alan Weiss Posts: 1,430 Registered: 11/27/08
Re: How to define Hessian for fminunc.
Posted: Oct 15, 2013 8:22 AM

On 10/14/2013 12:24 PM, Céldor wrote:
> Alan_Weiss <aweiss@mathworks.com> wrote in message
> <l3gtao\$mqh\$1@newscl01ah.mathworks.com>...
> ...

>> When you start at a point where the gradient of the objective is
>> zero, fminunc sees that the gradient is zero and realizes that it is
>> at a local extremum (min, max, or saddle point). It stops because the
>> first-order optimality measure is zero (the norm of the gradient). It
>> does not check second-order conditions in this case.
>>
>> I hope this helps. You can read about the fminunc algorithms here:
>> http://www.mathworks.com/help/optim/ug/unconstrained-nonlinear-optimization-algorithms.html
>>
>>
>> Alan Weiss
>> MATLAB mathematical toolbox documentation

>
> Hi Alan Weiss,
>
> Thank you for your replay.
> It is not as it supposed to be. Is there any other method in this
> toolbox which would check both the first and the second derivative and
> return the minimum even if the gradient norm is 0?

You can try patternsearch (requires a Global Optimization Toolbox
they are slower solvers, and fminsearch is less reliable.

It seems to me that you might want to use fminunc and restart any run
that doesn't change from the initial point, and nudge the initial point
to see if the solver returns or goes to a new optimum. See
http://www.mathworks.com/help/optim/ug/when-the-solver-succeeds.html

Alan Weiss
MATLAB mathematical toolbox documentation

Date Subject Author
10/12/13 Zebbik
10/14/13 Alan Weiss
10/14/13 Zebbik
10/15/13 Alan Weiss