octave: Solvers

 
 20.1 Solvers
 ============
 
 Octave can solve sets of nonlinear equations of the form
 
      F (x) = 0
 
 using the function ‘fsolve’, which is based on the MINPACK subroutine
 ‘hybrd’.  This is an iterative technique so a starting point must be
 provided.  This also has the consequence that convergence is not
 guaranteed even if a solution exists.
 
  -- : fsolve (FCN, X0, OPTIONS)
  -- : [X, FVEC, INFO, OUTPUT, FJAC] = fsolve (FCN, ...)
      Solve a system of nonlinear equations defined by the function FCN.
 
      FCN should accept a vector (array) defining the unknown variables,
      and return a vector of left-hand sides of the equations.
      Right-hand sides are defined to be zeros.  In other words, this
      function attempts to determine a vector X such that ‘FCN (X)’ gives
      (approximately) all zeros.
 
      X0 determines a starting guess.  The shape of X0 is preserved in
      all calls to FCN, but otherwise it is treated as a column vector.
 
      OPTIONS is a structure specifying additional options.  Currently,
      ‘fsolve’ recognizes these options: "FunValCheck", "OutputFcn",
      "TolX", "TolFun", "MaxIter", "MaxFunEvals", "Jacobian", "Updating",
      "ComplexEqn" "TypicalX", "AutoScaling" and "FinDiffType".
 
      If "Jacobian" is "on", it specifies that FCN, called with 2 output
      arguments also returns the Jacobian matrix of right-hand sides at
      the requested point.  "TolX" specifies the termination tolerance in
      the unknown variables, while "TolFun" is a tolerance for equations.
      Default is ‘1e-7’ for both "TolX" and "TolFun".
 
      If "AutoScaling" is on, the variables will be automatically scaled
      according to the column norms of the (estimated) Jacobian.  As a
      result, TolF becomes scaling-independent.  By default, this option
      is off because it may sometimes deliver unexpected (though
      mathematically correct) results.
 
      If "Updating" is "on", the function will attempt to use Broyden
      updates to update the Jacobian, in order to reduce the amount of
      Jacobian calculations.  If your user function always calculates the
      Jacobian (regardless of number of output arguments) then this
      option provides no advantage and should be set to false.
 
      "ComplexEqn" is "on", ‘fsolve’ will attempt to solve complex
      equations in complex variables, assuming that the equations possess
      a complex derivative (i.e., are holomorphic).  If this is not what
      you want, you should unpack the real and imaginary parts of the
      system to get a real system.
 
      For description of the other options, see ‘optimset’.
 
      On return, FVAL contains the value of the function FCN evaluated at
      X.
 
      INFO may be one of the following values:
 
      1
           Converged to a solution point.  Relative residual error is
           less than specified by TolFun.
 
      2
           Last relative step size was less that TolX.
 
      3
           Last relative decrease in residual was less than TolF.
 
      0
           Iteration limit exceeded.
 
      -3
           The trust region radius became excessively small.
 
      Note: If you only have a single nonlinear equation of one variable,
      using ‘fzero’ is usually a much better idea.
 
      Note about user-supplied Jacobians: As an inherent property of the
      algorithm, a Jacobian is always requested for a solution vector
      whose residual vector is already known, and it is the last accepted
      successful step.  Often this will be one of the last two calls, but
      not always.  If the savings by reusing intermediate results from
      residual calculation in Jacobian calculation are significant, the
      best strategy is to employ OutputFcn: After a vector is evaluated
      for residuals, if OutputFcn is called with that vector, then the
      intermediate results should be saved for future Jacobian
      evaluation, and should be kept until a Jacobian evaluation is
      requested or until OutputFcn is called with a different vector, in
      which case they should be dropped in favor of this most recent
      vector.  A short example how this can be achieved follows:
 
           function [fvec, fjac] = user_func (x, optimvalues, state)
           persistent sav = [], sav0 = [];
           if (nargin == 1)
             ## evaluation call
             if (nargout == 1)
               sav0.x = x; # mark saved vector
               ## calculate fvec, save results to sav0.
             elseif (nargout == 2)
               ## calculate fjac using sav.
             endif
           else
             ## outputfcn call.
             if (all (x == sav0.x))
               sav = sav0;
             endif
             ## maybe output iteration status, etc.
           endif
           endfunction
 
           ## ...
 
           fsolve (@user_func, x0, optimset ("OutputFcn", @user_func, ...))
 
      See also: Seefzero XREFfzero, Seeoptimset XREFoptimset.
 
    The following is a complete example.  To solve the set of equations
 
      -2x^2 + 3xy   + 4 sin(y) = 6
       3x^2 - 2xy^2 + 3 cos(x) = -4
 
 you first need to write a function to compute the value of the given
 function.  For example:
 
      function y = f (x)
        y = zeros (2, 1);
        y(1) = -2*x(1)^2 + 3*x(1)*x(2)   + 4*sin(x(2)) - 6;
        y(2) =  3*x(1)^2 - 2*x(1)*x(2)^2 + 3*cos(x(1)) + 4;
      endfunction
 
    Then, call ‘fsolve’ with a specified initial condition to find the
 roots of the system of equations.  For example, given the function ‘f’
 defined above,
 
      [x, fval, info] = fsolve (@f, [1; 2])
 
 results in the solution
 
      x =
 
        0.57983
        2.54621
 
      fval =
 
        -5.7184e-10
         5.5460e-10
 
      info = 1
 
 A value of ‘info = 1’ indicates that the solution has converged.
 
    When no Jacobian is supplied (as in the example above) it is
 approximated numerically.  This requires more function evaluations, and
 hence is less efficient.  In the example above we could compute the
 Jacobian analytically as
 
      function [y, jac] = f (x)
        y = zeros (2, 1);
        y(1) = -2*x(1)^2 + 3*x(1)*x(2)   + 4*sin(x(2)) - 6;
        y(2) =  3*x(1)^2 - 2*x(1)*x(2)^2 + 3*cos(x(1)) + 4;
        if (nargout == 2)
          jac = zeros (2, 2);
          jac(1,1) =  3*x(2) - 4*x(1);
          jac(1,2) =  4*cos(x(2)) + 3*x(1);
          jac(2,1) = -2*x(2)^2 - 3*sin(x(1)) + 6*x(1);
          jac(2,2) = -4*x(1)*x(2);
        endif
      endfunction
 
 The Jacobian can then be used with the following call to ‘fsolve’:
 
      [x, fval, info] = fsolve (@f, [1; 2], optimset ("jacobian", "on"));
 
 which gives the same solution as before.
 
  -- : fzero (FUN, X0)
  -- : fzero (FUN, X0, OPTIONS)
  -- : [X, FVAL, INFO, OUTPUT] = fzero (...)
      Find a zero of a univariate function.
 
      FUN is a function handle, inline function, or string containing the
      name of the function to evaluate.
 
      X0 should be a two-element vector specifying two points which
      bracket a zero.  In other words, there must be a change in sign of
      the function between X0(1) and X0(2).  More mathematically, the
      following must hold
 
           sign (FUN(X0(1))) * sign (FUN(X0(2))) <= 0
 
      If X0 is a single scalar then several nearby and distant values are
      probed in an attempt to obtain a valid bracketing.  If this is not
      successful, the function fails.
 
      OPTIONS is a structure specifying additional options.  Currently,
      ‘fzero’ recognizes these options: "FunValCheck", "OutputFcn",
      "TolX", "MaxIter", "MaxFunEvals".  For a description of these
      options, see Seeoptimset XREFoptimset.
 
      On exit, the function returns X, the approximate zero point and
      FVAL, the function value thereof.
 
      INFO is an exit flag that can have these values:
 
         • 1 The algorithm converged to a solution.
 
         • 0 Maximum number of iterations or function evaluations has
           been reached.
 
         • -1 The algorithm has been terminated from user output
           function.
 
         • -5 The algorithm may have converged to a singular point.
 
      OUTPUT is a structure containing runtime information about the
      ‘fzero’ algorithm.  Fields in the structure are:
 
         • iterations Number of iterations through loop.
 
         • nfev Number of function evaluations.
 
         • bracketx A two-element vector with the final bracketing of the
           zero along the x-axis.
 
         • brackety A two-element vector with the final bracketing of the
           zero along the y-axis.
 
      See also: Seeoptimset XREFoptimset, Seefsolve XREFfsolve.