NLS#
purpose#
The nls
package uses a regularization method to find a (local) unconstrained
minimizer of a differentiable weighted sum-of-squares objective function
See Section 4 of $GALAHAD/doc/nls.pdf for additional details.
terminology#
The gradient \(\nabla_x f(x)\) of a function \(f(x)\) is the vector whose \(i\)-th component is \(\partial f(x)/\partial x_i\). The Hessian \(\nabla_{xx} f(x)\) of \(f(x)\) is the symmetric matrix whose \(i,j\)-th entry is \(\partial^2 f(x)/\partial x_i \partial x_j\). The Hessian is sparse if a significant and useful proportion of the entries are universally zero.
The algorithm used by the package is iterative. From the current best estimate of the minimizer \(x_k\), a trial improved point \(x_k + s_k\) is sought. The correction \(s_k\) is chosen to improve a model \(m_k(s)\) of the objective function \(f(x_k+s)\) built around \(x_k\). The model is the sum of two basic components, a suitable approximation \(t_k(s)\) of \(f(x_k+s)\), %another approximation of \((\rho/r) \|x_k+s\|_r^r\) (if \(\rho > 0\)), and a regularization term \((\sigma_k/p) \|s\|_{S_k}^p\) involving a weight \(\sigma_k\), power \(p\) and a norm \(\|s\|_{S_k} := \sqrt{s^T S_k s}\) for a given positive definite scaling matrix \(S_k\) that is included to prevent large corrections. The weight \(\sigma_k\) is adjusted as the algorithm progresses to ensure convergence.
The model \(t_k(s)\) is a truncated Taylor-series approximation, and this relies on being able to compute or estimate derivatives of \(c(x)\). Various models are provided, and each has different derivative requirements. We denote the \(m\) by \(n\) residual Jacobian \(J(x) \equiv \nabla_x c(x)\) as the matrix whose \(i,j\)-th component
the first-order Taylor approximation \(f(x_k) + g(x_k)^T s\), where \(g(x) = J^T(x) W c(x)\),
a barely second-order approximation \(f(x_k) + g(x_k)^T s + \frac{1}{2} s^T W s\),
the Gauss-Newton approximation \(\frac{1}{2} \| c(x_k) + J(x_k) s\|^2_W\),
the Newton (second-order Taylor) approximation
\(f(x_k) + g(x_k)^T s + \frac{1}{2} s^T [ J^T(x_k) W J(x_k) + H(x_k,W c(x_k))] s\), and
the tensor Gauss-Newton approximation \(\frac{1}{2} \| c(x_k) + J(x_k) s + \frac{1}{2} s^T \cdot P(x_k,s) \|^2_W\), where the \(i\)-th component of \(s^T \cdot P(x_k,s)\) is shorthand for the scalar \(s^T H_i(x_k) s\), where \(W\) is the diagonal matrix of weights \(w_i\), \(i = 1, \ldots m\)0.
method#
An adaptive regularization method is used. In this, an improvement to a current estimate of the required minimizer, \(x_k\) is sought by computing a step \(s_k\). The step is chosen to approximately minimize a model \(t_k(s)\) of \(f_{\rho,r}(x_k+s)\) that includes a weighted regularization term \(\frac{\sigma_k}{p} \|s\|_{S_k}^p\) for some specified positive weight \(\sigma_k\). The quality of the resulting step \(s_k\) is assessed by computing the “ratio” \((f(x_k) - f(x_k + s_k))/(t_k(0) - t_k(s_k))\). The step is deemed to have succeeded if the ratio exceeds a given \(\eta_s > 0\), and in this case \(x_{k+1} = x_k + s_k\). Otherwise \(x_{k+1} = x_k\), and the weight is increased by powers of a given increase factor up to a given limit. If the ratio is larger than \(\eta_v \geq \eta_d\), the weight will be decreased by powers of a given decrease factor again up to a given limit. The method will terminate as soon as \(f(x_k)\) or \(\|\nabla_x f(x_k)\|\) is smaller than a specified value.
A choice of linear, quadratic or quartic models \(t_k(s)\) is available (see the previous section), and normally a two-norm regularization will be used, but this may change if preconditioning is employed.
If linear or quadratic models are employed, an appropriate,
approximate model minimizer is found using either a direct approach
involving factorization of a shift of the model Hessian \(B_k\) or an
iterative (conjugate-gradient/Lanczos) approach based on approximations
to the required solution from a so-called Krlov subspace. The direct
approach is based on the knowledge that the required solution
satisfies the linear system of equations \((B_k + \lambda_k I) s_k
= - \nabla_x f(x_k)\) involving a scalar Lagrange multiplier \(\lambda_k\).
This multiplier is found by uni-variate root finding, using a safeguarded
Newton-like process, by RQS
. The iterative approach
uses GLRT
, and is best accelerated by preconditioning with
good approximations to the Hessian of the model using PSLS
. The
iterative approach has the advantage that only Hessian matrix-vector products
are required, and thus the Hessian \(B_k\) is not required explicitly.
However when factorizations of the Hessian are possible, the direct approach
is often more efficient.
When a quartic model is used, the model is itself of least-squares form, and the package calls itself recursively to approximately minimize its model. The quartic model often gives a better approximation, but at the cost of more involved derivative requirements.
references#
The generic adaptive cubic regularization method is described in detail in
C. Cartis, N. I. M. Gould and Ph. L. Toint, ``Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results’’ Mathematical Programming 127(2) (2011) 245–295,
and uses ``tricks’’ as suggested in
N. I. M. Gould, M. Porcelli and Ph. L. Toint, ``Updating the regularization parameter in the adaptive cubic regularization algorithm’’. Computational Optimization and Applications 53(1) (2012) 1–22.
The specific methods employed here are discussed in
N. I. M. Gould, J. A. Scott and T. Rees, ``Convergence and evaluation-complexity analysis of a regularized tensor-Newton method for solving nonlinear least-squares problems’’. Computational Optimization and Applications 73(1) (2019) 1–35.
matrix storage#
The unsymmetric \(m\) by \(n\) Jacobian matrix \(J = J(x)\) and the residual-Hessians-vector product matrix \(P(x,v)\) may be presented and stored in a variety of convenient input formats. Let \(A\) be \(J\) or \(P\) as appropriate.
Dense storage format: The matrix \(A\) is stored as a compact dense matrix by rows, that is, the values of the entries of each row in turn are stored in order within an appropriate real one-dimensional array. In this case, component \(n \ast i + j\) of the storage array A_val will hold the value \(A_{ij}\) for \(0 \leq i \leq m-1\), \(0 \leq j \leq n-1\).
Dense by columns storage format: The matrix \(A\) is stored as a compact dense matrix by columns, that is, the values of the entries of each column in turn are stored in order within an appropriate real one-dimensional array. In this case, component \(m \ast j + i\) of the storage array A_val will hold the value \(A_{ij}\) for \(0 \leq i \leq m-1\), \(0 \leq j \leq n-1\).
Sparse co-ordinate storage format: Only the nonzero entries of the matrices are stored. For the \(l\)-th entry, \(0 \leq l \leq ne-1\), of \(A\), its row index i, column index j and value \(A_{ij}\), \(0 \leq i \leq m-1\), \(0 \leq j \leq n-1\), are stored as the \(l\)-th components of the integer arrays A_row and A_col and real array A_val, respectively, while the number of nonzeros is recorded as A_ne = \(ne\).
Sparse row-wise storage format: Again only the nonzero entries are stored, but this time they are ordered so that those in row i appear directly before those in row i+1. For the i-th row of \(A\) the i-th component of the integer array A_ptr holds the position of the first entry in this row, while A_ptr(m) holds the total number of entries. The column indices j, \(0 \leq j \leq n-1\), and values \(A_{ij}\) of the nonzero entries in the i-th row are stored in components l = A_ptr(i), \(\ldots\), A_ptr(i+1)-1, \(0 \leq i \leq m-1\), of the integer array A_col, and real array A_val, respectively. For sparse matrices, this scheme almost always requires less storage than its predecessor.
Sparse column-wise storage format: Once again only the nonzero entries are stored, but this time they are ordered so that those in column j appear directly before those in column j+1. For the j-th column of \(A\) the j-th component of the integer array A_ptr holds the position of the first entry in this column, while A_ptr(n) holds the total number of entries. The row indices i, \(0 \leq i \leq m-1\), and values \(A_{ij}\) of the nonzero entries in the j-th columnsare stored in components l = A_ptr(j), \(\ldots\), A_ptr(j+1)-1, \(0 \leq j \leq n-1\), of the integer array A_row, and real array A_val, respectively. As before, for sparse matrices, this scheme almost always requires less storage than the co-ordinate format.
The symmetric \(n\) by \(n\) matrix \(H = H(x,y)\) may be presented and stored in a variety of formats. But crucially symmetry is exploited by only storing values from the lower triangular part (i.e, those entries that lie on or below the leading diagonal).
Dense storage format: The matrix \(H\) is stored as a compact dense matrix by rows, that is, the values of the entries of each row in turn are stored in order within an appropriate real one-dimensional array. Since \(H\) is symmetric, only the lower triangular part (that is the part \(H_{ij}\) for \(0 \leq j \leq i \leq n-1\)) need be held. In this case the lower triangle should be stored by rows, that is component \(i * i / 2 + j\) of the storage array H_val will hold the value \(H_{ij}\) (and, by symmetry, \(H_{ji}\)) for \(0 \leq j \leq i \leq n-1\).
Sparse co-ordinate storage format: Only the nonzero entries of the matrices are stored. For the \(l\)-th entry, \(0 \leq l \leq ne-1\), of \(H\), its row index i, column index j and value \(H_{ij}\), \(0 \leq j \leq i \leq n-1\), are stored as the \(l\)-th components of the integer arrays H_row and H_col and real array H_val, respectively, while the number of nonzeros is recorded as H_ne = \(ne\). Note that only the entries in the lower triangle should be stored.
Sparse row-wise storage format: Again only the nonzero entries are stored, but this time they are ordered so that those in row i appear directly before those in row i+1. For the i-th row of \(H\) the i-th component of the integer array H_ptr holds the position of the first entry in this row, while H_ptr(n) holds the total number of entries. The column indices j, \(0 \leq j \leq i\), and values \(H_{ij}\) of the entries in the i-th row are stored in components l = H_ptr(i), …, H_ptr(i+1)-1 of the integer array H_col, and real array H_val, respectively. Note that as before only the entries in the lower triangle should be stored. For sparse matrices, this scheme almost always requires less storage than its predecessor.
Diagonal storage format: If \(H\) is diagonal (i.e., \(H_{ij} = 0\) for all \(0 \leq i \neq j \leq n-1\)) only the diagonals entries \(H_{ii}\), \(0 \leq i \leq n-1\) need be stored, and the first n components of the array H_val may be used for the purpose.
Multiples of the identity storage format: If \(H\) is a multiple of the identity matrix, (i.e., \(H = \alpha I\) where \(I\) is the n by n identity matrix and \(\alpha\) is a scalar), it suffices to store \(\alpha\) as the first component of H_val.
The identity matrix format: If \(H\) is the identity matrix, no values need be stored.
The zero matrix format: The same is true if \(H\)7 is the zero matrix.
functions#
- nls.initialize()#
Set default option values and initialize private data
Returns:
- optionsdict
- dictionary containing default control options:
- errorint
error and warning diagnostics occur on stream error.
- outint
general output occurs on stream out.
- print_levelint
the level of output required. Possible values are
<= 0
gives no output.
1
gives a one-line summary for every iteration.
2
gives a summary of the inner iteration for each iteration.
>=3
gives increasingly verbose (debugging) output.
- start_printint
any printing will start on this iteration.
- stop_printint
any printing will stop on this iteration.
- print_gapint
the number of iterations between printing.
- maxitint
the maximum number of iterations performed.
- alive_unitint
removal of the file alive_file from unit alive_unit terminates execution.
- alive_filestr
see alive_unit.
- jacobian_availableint
is the Jacobian matrix of first derivatives available (\(\geq\) 2), is access only via matrix-vector products (=1) or is it not available (\(\leq\) 0) ?.
- hessian_availableint
is the Hessian matrix of second derivatives available (\(\geq\) 2), is access only via matrix-vector products (=1) or is it not available (\(\leq\) 0) ?.
- modelint
the model used. Possible values are
0
dynamic (not yet implemented)
1
first-order (no Hessian)
2
barely second-order (identity Hessian)
3
Gauss-Newton (\(J^T J\) Hessian)
4
second-order (exact Hessian)
5
Gauss-Newton to Newton transition
6
tensor Gauss-Newton treated as a least-squares model
7
tensor Gauss-Newton treated as a general model
8
tensor Gauss-Newton transition from a least-squares to a general mode.
- normint
the regularization norm used. The norm is defined via \(\|v\|^2 = v^T S v\), and will define the preconditioner used for iterative methods. Possible values for \(S\) are
-3
user’s own regularization norm
-2
\(S\) = limited-memory BFGS matrix (with
PSLS_options.lbfgs_vectors
history) (not yet implemented)-1
identity (= Euclidan two-norm)
0
automatic (not yet implemented)
1
diagonal, \(S\) = diag( max( \(J^TJ\) Hessian,
PSLS_options.min_diagonal
) )2
diagonal, \(S\) = diag( max( Hessian,
PSLS_options.min_diagonal
) )3
banded, \(S\) = band( Hessian ) with semi-bandwidth
PSLS_options.semi_bandwidth
4
re-ordered band, \(S\) = band(order(A)) with semi-bandwidth
PSLS_options.semi_bandwidth
5
full factorization, \(S\) = Hessian, Schnabel-Eskow modification
6
full factorization, \(S\) = Hessian, GMPS modification (not yet implemented)
7
incomplete factorization of Hessian, Lin-More’
8
incomplete factorization of Hessian, HSL_MI28
9
incomplete factorization of Hessian, Munskgaard (not yet implemented)
10
expanding band of Hessian (not yet implemented).
- non_monotoneint
non-monotone <= 0 monotone strategy used, anything else non-monotone strategy with this history length used.
- weight_update_strategyint
define the weight-update strategy: 1 (basic), 2 (reset to zero when very successful), 3 (imitate TR), 4 (increase lower bound), 5 (GPT).
- stop_c_absolutefloat
overall convergence tolerances. The iteration will terminate when \(||c(x)||_2 \leq \max(\)
stop_c_absolute,
stop_c_relative
* \(\|c(x_{\mbox{initial}})\|_2\) or when the norm of the gradient, \(g = J^T(x) c(x) / \|c(x)\|_2\), of ||c(x)||_2 satisfies \(\|g\|_2 \leq \max(\)stop_g_absolute,
stop_g_relative
* \(\|g_{\mbox{initial}}\|_2\), or if the step is less thanstop_s
.- stop_c_relativefloat
see stop_c_absolute.
- stop_g_absolutefloat
see stop_c_absolute.
- stop_g_relativefloat
see stop_c_absolute.
- stop_sfloat
see stop_c_absolute.
- powerfloat
the regularization power (<2 means chosen according to the model).
- initial_weightfloat
initial value for the regularization weight (-ve means \(1/\|g_0\|)\)).
- minimum_weightfloat
minimum permitted regularization weight.
- initial_inner_weightfloat
initial value for the inner regularization weight for tensor GN (-ve means 0).
- eta_successfulfloat
potential iterate will only be accepted if the actual decrease f - f(x_new) is larger than
eta_successful
times that predicted by a quadratic model of the decrease. The regularization weight will be decreaed if this relative decrease is greater thaneta_very_successful
but smaller thaneta_too_successful
.- eta_very_successfulfloat
see eta_successful.
- eta_too_successfulfloat
see eta_successful.
- weight_decrease_minfloat
on very successful iterations, the regularization weight will be reduced by the factor
weight_decrease
but no more thanweight_decrease_min
while if the iteration is unsucceful, the weight will be increased by a factorweight_increase
but no more thanweight_increase_max
(these are delta_1, delta_2, delta3 and delta_max in Gould, Porcelli and Toint, 2011).- weight_decreasefloat
see weight_decrease_min
- weight_increasefloat
see weight_decrease_min
- weight_increase_maxfloat
see weight_decrease_min
- reduce_gapfloat
expert parameters as suggested in Gould, Porcelli and Toint, “Updating the regularization parameter in the adaptive cubic regularization algorithm” RAL-TR-2011-007, Rutherford Appleton Laboratory, England (2011), http://epubs.stfc.ac.uk/bitstream/6181/RAL-TR-2011-007.pdf (these are denoted beta, epsilon_chi and alpha_max in the paper).
- tiny_gapfloat
see reduce_gap.
- large_rootfloat
see reduce_gap.
- switch_to_newtonfloat
if the Gauss-Newto to Newton model is specified, switch to Newton as soon as the norm of the gradient g is smaller than switch_to_newton.
- cpu_time_limitfloat
the maximum CPU time allowed (-ve means infinite).
- clock_time_limitfloat
the maximum elapsed clock time allowed (-ve means infinite).
- subproblem_directbool
use a direct (factorization) or (preconditioned) iterative method to find the search direction.
- renormalize_weightbool
should the weight be renormalized to account for a change in scaling?.
- magic_stepbool
allow the user to perform a “magic” step to improve the objective.
- print_objbool
print values of the objective/gradient rather than ||c|| and its gradient.
- space_criticalbool
if
space_critical
is True, every effort will be made to use as little space as possible. This may result in longer computation time.- deallocate_error_fatalbool
if
deallocate_error_fatal
is True, any array/pointer deallocation error will terminate execution. Otherwise, computation will continue.- prefixstr
all output lines will be prefixed by the string contained in quotes within
prefix
, e.g. ‘word’ (note the qutoes) will result in the prefix word.- rqs_optionsdict
default control options for RQS (see
rqs.initialize
).- glrt_optionsdict
default control options for GLRT (see
glrt.initialize
).- psls_optionsdict
default control options for PSLS (see
psls.initialize
).- bsc_optionsdict
default control options for BSC (see
bsc.initialize
).- roots_optionsdict
default control options for ROOTS (see
roots.initialize
).- subproblem_optionsdict
- default control options for the step-finding subproblem:
- errorint
error and warning diagnostics occur on stream error.
- outint
general output occurs on stream out.
- print_levelint
the level of output required. Possible values are:
<= 0
gives no output.
1
gives a one-line summary for every iteration.
2
gives a summary of the inner iteration for each iteration.
>=3
gives increasingly verbose (debugging) output.
- start_printint
any printing will start on this iteration.
- stop_printint
any printing will stop on this iteration.
- print_gapint
the number of iterations between printing.
- maxitint
the maximum number of iterations performed.
- alive_unitint
removal of the file alive_file from unit alive_unit terminates execution.
- alive_filestr
see alive_unit.
- jacobian_availableint
is the Jacobian matrix of first derivatives available (\(\geq\) 2), is access only via matrix-vector products (=1) or is it not available (\(\leq\) 0) ?.
- hessian_availableint
is the Hessian matrix of second derivatives available (\(\geq\) 2), is access only via matrix-vector products (=1) or is it not available (\(\leq\) 0) ?.
- modelint
the model used. Possible values are
0
dynamic (not yet implemented)
1
first-order (no Hessian)
2
barely second-order (identity Hessian)
3
Gauss-Newton (\(J^T J\) Hessian)
4
second-order (exact Hessian)
5
Gauss-Newton to Newton transition
6
tensor Gauss-Newton treated as a least-squares model
7
tensor Gauss-Newton treated as a general model
8
tensor Gauss-Newton transition from a least-squares to a general mode.
- normint
the regularization norm used. The norm is defined via \(\|v\|^2 = v^T S v\), and will define the preconditioner used for iterative methods. Possible values for \(S\) are
-3
user’s own regularization norm
-2
\(S\) = limited-memory BFGS matrix (with
PSLS_options.lbfgs_vectors
history) (not yet implemented)
-1
identity (= Euclidan two-norm)
0
automatic (not yet implemented)
1
diagonal, \(S\) = diag( max( \(J^TJ\) Hessian,
PSLS_options.min_diagonal
) )
2
diagonal, \(S\) = diag( max( Hessian,
PSLS_options.min_diagonal
) )3
banded, \(S\) = band( Hessian ) with semi-bandwidth
PSLS_options.semi_bandwidth
4
re-ordered band, P=band(order(A)) with semi-bandwidth
PSLS_options.semi_bandwidth
5
full factorization, \(S\) = Hessian, Schnabel-Eskow modification
6
full factorization, \(S\) = Hessian, GMPS modification (not yet implemented)
7
incomplete factorization of Hessian, Lin-More’
8
incomplete factorization of Hessian, HSL_MI28
9
incomplete factorization of Hessian, Munskgaard (not yet implemented)
10
expanding band of Hessian (not yet implemented).
- non_monotoneint
non-monotone <= 0 monotone strategy used, anything else non-monotone strategy with this history length used.
- weight_update_strategyint
define the weight-update strategy: 1 (basic), 2 (reset to zero when very successful), 3 (imitate TR), 4 (increase lower bound), 5 (GPT).
- stop_c_absolutefloat
overall convergence tolerances. The iteration will terminate when \(||c(x)||_2 \leq \max(\)
stop_c_absolute,
stop_c_relative
* \(\|c(x_{\mbox{initial}})\|_2\) or when the norm of the gradient, \(g = J^T(x) c(x) / \|c(x)\|_2\), of ||c(x)||_2 satisfies \(\|g\|_2 \leq \max(\)stop_g_absolute,
stop_g_relative
* \(\|g_{\mbox{initial}}\|_2\), or if the step is less thanstop_s
.- stop_c_relativefloat
see stop_c_absolute.
- stop_g_absolutefloat
see stop_c_absolute.
- stop_g_relativefloat
see stop_c_absolute.
- stop_sfloat
see stop_c_absolute.
- powerfloat
the regularization power (<2 means chosen according to the model).
- initial_weightfloat
initial value for the regularization weight (-ve means \(1/\|g_0\|)\)).
- minimum_weightfloat
minimum permitted regularization weight.
- initial_inner_weightfloat
initial value for the inner regularization weight for tensor GN (-ve means 0).
- eta_successfulfloat
potential iterate will only be accepted if the actual decrease f - f(x_new) is larger than
eta_successful
times that predicted by a quadratic model of the decrease. The regularization weight will be decreaed if this relative decrease is greater thaneta_very_successful
but smaller thaneta_too_successful
.- eta_very_successfulfloat
see eta_successful.
- eta_too_successfulfloat
see eta_successful.
- weight_decrease_minfloat
on very successful iterations, the regularization weight will be reduced by the factor
weight_decrease
but no more thanweight_decrease_min
while if the iteration is unsucceful, the weight will be increased by a factorweight_increase
but no more thanweight_increase_max
(these are delta_1, delta_2, delta3 and delta_max in Gould, Porcelli and Toint, 2011).- weight_decreasefloat
see weight_decrease_min
- weight_increasefloat
see weight_decrease_min
- weight_increase_maxfloat
see weight_decrease_min
- reduce_gapfloat
expert parameters as suggested in Gould, Porcelli and Toint, “Updating the regularization parameter in the adaptive cubic regularization algorithm” RAL-TR-2011-007, Rutherford Appleton Laboratory, England (2011), http://epubs.stfc.ac.uk/bitstream/6181/RAL-TR-2011-007.pdf (these are denoted beta, epsilon_chi and alpha_max in the paper).
- tiny_gapfloat
see reduce_gap.
- large_rootfloat
see reduce_gap.
- switch_to_newtonfloat
if the Gauss-Newto to Newton model is specified, switch to Newton as soon as the norm of the gradient g is smaller than switch_to_newton.
- cpu_time_limitfloat
the maximum CPU time allowed (-ve means infinite).
- clock_time_limitfloat
the maximum elapsed clock time allowed (-ve means infinite).
- subproblem_directbool
use a direct (factorization) or (preconditioned) iterative method to find the search direction.
- renormalize_weightbool
should the weight be renormalized to account for a change in scaling?.
- magic_stepbool
allow the user to perform a “magic” step to improve the objective.
- print_objbool
print values of the objective/gradient rather than ||c|| and its gradient.
- space_criticalbool
if
space_critical
is True, every effort will be made to use as little space as possible. This may result in longer computation time.- deallocate_error_fatalbool
if
deallocate_error_fatal
is True, any array/pointer deallocation error will terminate execution. Otherwise, computation will continue.- prefixstr
all output lines will be prefixed by
prefix(2:LEN(TRIM(.prefix))-1)
whereprefix
contains the required string enclosed in quotes, e.g. “string” or ‘string’.- rqs_optionsdict
default control options for RQS (see
rqs.initialize
).- glrt_optionsdict
default control options for GLRT (see
glrt.initialize
).- psls_optionsdict
default control options for PSLS (see
psls.initialize
).- bsc_optionsdict
default control options for BSC (see
bsc.initialize
).- roots_optionsdict
default control options for ROOTS (see
roots.initialize
).
- nls.load(n, m, J_type, J_ne, J_row, J_col, J_ptr, H_type, H_ne, H_row, H_col, H_ptr, P_type, P_ne, P_row, P_col, P_ptr, w, options=None)#
Import problem data into internal storage prior to solution.
Parameters:
- nint
holds the number of variables.
- mint
holds the number of residuals.
- J_typestring
specifies the unsymmetric storage scheme used for the Jacobian \(J = J(x)\). It should be one of ‘coordinate’, ‘sparse_by_rows’ or ‘dense’; lower or upper case variants are allowed.
- J_neint
holds the number of entries in \(J\) in the sparse co-ordinate storage scheme. It need not be set for any of the other two schemes.
- J_rowndarray(J_ne)
holds the row indices of \(J\) in the sparse co-ordinate storage scheme. It need not be set for any of the other two schemes, and in this case can be None.
- J_colndarray(J_ne)
holds the column indices of \(J\) in either the sparse co-ordinate, or the sparse row-wise storage scheme. It need not be set when the dense storage scheme is used, and in this case can be None.
- J_ptrndarray(m+1)
holds the starting position of each row of \(J\), as well as the total number of entries, in the sparse row-wise storage scheme. It need not be set when the other schemes are used, and in this case can be None.
- H_typestring, optional
specifies the symmetric storage scheme used for the Hessian \(H = H(x,y)\). It should be one of ‘coordinate’, ‘sparse_by_rows’, ‘dense’ or ‘diagonal’; lower or upper case variants are allowed. This and the following H_* arguments are only required if a Newton approximation or tensor Gauss-Newton approximation model is required (see control.model = 4,…,8).
- H_neint, optional
holds the number of entries in the lower triangular part of \(H\) in the sparse co-ordinate storage scheme. It need not be set for any of the other three schemes.
- H_rowndarray(H_ne), optional
holds the row indices of the lower triangular part of \(H\) in the sparse co-ordinate storage scheme. It need not be set for any of the other three schemes, and in this case can be None.
- H_colndarray(H_ne), optional
holds the column indices of the lower triangular part of \(H\) in either the sparse co-ordinate, or the sparse row-wise storage scheme. It need not be set when the dense or diagonal storage schemes are used, and in this case can be None.
- H_ptrndarray(n+1), optional
holds the starting position of each row of the lower triangular part of \(H\), as well as the total number of entries, in the sparse row-wise storage scheme. It need not be set when the other schemes are used, and in this case can be None.
- P_typestring, optional
specifies the unsymmetric storage scheme used for the residual Hessian-product matrix \(P = P(x,v)\). It should be one of ‘sparse_by_columns’ or ‘dense_by_columns’ (with the intention that ‘coordinate’ will be added at some time);; lower or upper case variants are allowed. This and the following P_* arguments are only required if a tensor Gauss-Newton approximation model is required (see control.model = 6,7,8).
- P_neint, optional
holds the number of entries in \(P\) in the sparse co-ordinate storage scheme. It need not be set for any of the other two schemes.
- P_rowndarray(P_ne), optional
holds the row indices of \(P\) in the sparse co-ordinate storage scheme. It need not be set for any of the other two schemes, and in this case can be None.
- P_colndarray(P_ne), optional
holds the column indices of \(P\) in either the sparse co-ordinate, or the sparse row-wise storage scheme. It need not be set when the dense storage scheme is used, and in this case can be None.
- P_ptrndarray(n+1), optional
holds the starting position of each column of \(P\), as well as the total number of entries, in the sparse column-wise storage scheme. It need not be set when the other schemes are used, and in this case can be None.
- wndarray(m), optional
holds the vector of weights \(w\). If w is not provided, weights of one will be presumed.
- optionsdict, optional
dictionary of control options (see
nls.initialize
).
- nls.solve(n, m, x, eval_c, J_ne, eval_j, H_ne, eval_h, P_ne, eval_hprod)#
Find an approximate local unconstrained minimizer of a given least-squares function using a regularization method.
Parameters:
- nint
holds the number of variables.
- mint
holds the number of residuals.
- xndarray(n)
holds the values of optimization variables \(x\).
- eval_ccallable
a user-defined function that must have the signature:
c = eval_c(x)
The components of the residual \(c(x)\) evaluated at \(x\) must be assigned to
c
.- J_neint
holds the number of entries in the Jacobian \(J = J(x)\).
- eval_jcallable
a user-defined function that must have the signature:
j = eval_(x)
The components of the nonzeros in the Jacobian \(J(x)\) of the objective function evaluated at \(x\) must be assigned to
j
in the same order as specified in the sparsity pattern innls.load
.- H_neint, optional
holds the number of entries in the lower triangular part of the Hessian \(H = H(x,y\). This and the following eval_h argument are only required if a Newton approximation or tensor Gauss-Newton approximation model is required (see control.model = 4,…,8).
- eval_hcallable, optional
a user-defined function that must have the signature:
h = eval_h(x,y)
The components of the nonzeros in the lower triangle of the Hessian \(H(x,y)\) evaluated at \(x\) and \(y\) must be assigned to
h
in the same order as specified in the sparsity pattern innls.load
.- P_neint, optional
holds the number of entries in the residual Hessian-product matrix \(P = P(x,v)\). This and the following eval_hprod argument are only required if a Newton approximation or tensor Gauss-Newton approximation model is required (see control.model = 6,7,8).
- eval_hprodcallable, optional
a user-defined function that must have the signature:
p = eval_hprod(x,v)
The components of the nonzeros in the Hessian producr matrix \(P(x,y)\) evaluated at \(x\) and \(v\) must be assigned to
p
in the same order as specified in the sparsity pattern innls.load
.Returns:
- xndarray(n)
holds the value of the approximate minimizer \(x\) after a successful call.
- cndarray(m)
holds the value of the residuals \(c(x)\).
- gndarray(n)
holds the gradient \(\nabla f(x)\) of the objective function.
- [optional] nls.information()
Provide optional output information
Returns:
- informdict
- dictionary containing output information:
- statusint
return status. Possible values are:
0
The run was successful.
-1
An allocation error occurred. A message indicating the offending array is written on unit options[‘error’], and the returned allocation status and a string containing the name of the offending array are held in inform[‘alloc_status’] and inform[‘bad_alloc’] respectively.
-2
A deallocation error occurred. A message indicating the offending array is written on unit options[‘error’] and the returned allocation status and a string containing the name of the offending array are held in inform[‘alloc_status’] and inform[‘bad_alloc’] respectively.
-3
The restriction n > 0 or m > 0 or requirement that type contains its relevant string ‘dense’, ‘coordinate’, ‘sparse_by_rows’, ‘diagonal’ or ‘absent’ has been violated.
-9
The analysis phase of the factorization failed; the return status from the factorization package is given by inform[‘factor_status’].
-10
The factorization failed; the return status from the factorization package is given by inform[‘factor_status’].
-11
The solution of a set of linear equations using factors from the factorization package failed; the return status from the factorization package is given by inform[‘factor_status’].
-15
The preconditioner \(S(x)\) appears not to be positive definite.
-16
The problem is so ill-conditioned that further progress is impossible.
-18
Too many iterations have been performed. This may happen if options[‘maxit’] is too small, but may also be symptomatic of a badly scaled problem.
-19
The CPU time limit has been reached. This may happen if options[‘cpu_time_limit’] is too small, but may also be symptomatic of a badly scaled problem.
-82
The user has forced termination of the solver by removing the file named options[‘alive_file’] from unit options[‘alive_unit’].
- alloc_statusint
the status of the last attempted allocation/deallocation.
- bad_allocstr
the name of the array for which an allocation/deallocation error occurred.
- bad_evalstr
the name of the user-supplied evaluation routine for which an error occurred.
- iterint
the total number of iterations performed.
- cg_iterint
the total number of CG iterations performed.
- c_evalint
the total number of evaluations of the residual function c(x).
- j_evalint
the total number of evaluations of the Jacobian J(x) of c(x).
- h_evalint
the total number of evaluations of the scaled Hessian H(x,y) of c(x).
- factorization_maxint
the maximum number of factorizations in a sub-problem solve.
- factorization_statusint
the return status from the factorization.
- max_entries_factorslong
the maximum number of entries in the factors.
- factorization_integerlong
the total integer workspace required for the factorization.
- factorization_reallong
the total real workspace required for the factorization.
- factorization_averagefloat
the average number of factorizations per sub-problem solve.
- objfloat
the value of the objective function \(\frac{1}{2}\|c(x)\|^2_W\) at the best estimate the solution, x, determined by NLS_solve.
- norm_cfloat
the norm of the residual \(\|c(x)\|_W\) at the best estimate of the solution x, determined by NLS_solve.
- norm_gfloat
the norm of the gradient of \(\|c(x)\|_W\) of the objective function at the best estimate, x, of the solution determined by NLS_solve.
- weightfloat
the final regularization weight used.
- timedict
- dictionary containing timing information:
- totalfloat
the total CPU time spent in the package.
- preprocessfloat
the CPU time spent preprocessing the problem.
- analysefloat
the CPU time spent analysing the required matrices prior to factorization.
- factorizefloat
the CPU time spent factorizing the required matrices.
- solvefloat
the CPU time spent computing the search direction.
- clock_totalfloat
the total clock time spent in the package.
- clock_preprocessfloat
the clock time spent preprocessing the problem.
- clock_analysefloat
the clock time spent analysing the required matrices prior to factorization.
- clock_factorizefloat
the clock time spent factorizing the required matrices.
- clock_solvefloat
the clock time spent computing the search direction.
- rqs_informdict
inform parameters for RQS (see
rqs.information
).- glrt_informdict
inform parameters for GLTR (see
glrt.information
).- psls_informdict
inform parameters for PSLS (see
psls.information
).- bsc_informdict
inform parameters for BSC (see
bsc.information
).- roots_informdict
inform parameters for ROOTS (see
roots.information
).- subproblem_informdict
- inform parameters for subproblem:
- statusint
return status. Possible values are:
0
The run was successful.
-1
An allocation error occurred. A message indicating the offending array is written on unit options[‘error’], and the returned allocation status and a string containing the name of the offending array are held in inform[‘alloc_status’] and inform[‘bad_alloc’] respectively.
-2
A deallocation error occurred. A message indicating the offending array is written on unit options[‘error’] and the returned allocation status and a string containing the name of the offending array are held in inform[‘alloc_status’] and inform[‘bad_alloc’] respectively.
-3
The restriction n > 0 or m > 0 or requirement that type contains its relevant string ‘dense’, ‘coordinate’, ‘sparse_by_rows’, ‘diagonal’ or ‘absent’ has been violated.
-9
The analysis phase of the factorization failed; the return status from the factorization package is given by inform[‘factor_status’].
-10
The factorization failed; the return status from the factorization package is given by inform[‘factor_status’].
-11
The solution of a set of linear equations using factors from the factorization package failed; the return status from the factorization package is given by inform[‘factor_status’].
-15
The preconditioner \(S(x)\) appears not to be positive definite.
-16
The problem is so ill-conditioned that further progress is impossible.
-18
Too many iterations have been performed. This may happen if options[‘maxit’] is too small, but may also be symptomatic of a badly scaled problem.
-19
The CPU time limit has been reached. This may happen if options[‘cpu_time_limit’] is too small, but may also be symptomatic of a badly scaled problem.
-82
The user has forced termination of the solver by removing the file named options[‘alive_file’] from unit options[‘alive_unit’].
- alloc_statusint
the status of the last attempted allocation/deallocation.
- bad_allocstr
the name of the array for which an allocation/deallocation error occurred.
- bad_evalstr
the name of the user-supplied evaluation routine for which an error occurred.
- iterint
the total number of iterations performed.
- cg_iterint
the total number of CG iterations performed.
- c_evalint
the total number of evaluations of the residual function c(x).
- j_evalint
the total number of evaluations of the Jacobian J(x) of c(x).
- h_evalint
the total number of evaluations of the scaled Hessian H(x,y) of c(x).
- factorization_maxint
the maximum number of factorizations in a sub-problem solve.
- factorization_statusint
the return status from the factorization.
- max_entries_factorslong
the maximum number of entries in the factors.
- factorization_integerlong
the total integer workspace required for the factorization.
- factorization_reallong
the total real workspace required for the factorization.
- factorization_averagefloat
the average number of factorizations per sub-problem solve.
- objfloat
the value of the objective function \(\frac{1}{2}\|c(x)\|^2_W\) at the best estimate the solution, x, determined by NLS_solve.
- norm_cfloat
the norm of the residual \(\|c(x)\|_W\) at the best estimate of the solution x, determined by NLS_solve.
- norm_gfloat
the norm of the gradient of \(\|c(x)\|_W\) of the objective function at the best estimate, x, of the solution determined by NLS_solve.
- weightfloat
the final regularization weight used.
- timedict
- dictionary containing timing information:
- totalfloat
the total CPU time spent in the package.
- preprocessfloat
the CPU time spent preprocessing the problem.
- analysefloat
the CPU time spent analysing the required matrices prior to factorization.
- factorizefloat
the CPU time spent factorizing the required matrices.
- solvefloat
the CPU time spent computing the search direction.
- clock_totalfloat
the total clock time spent in the package.
- clock_preprocessfloat
the clock time spent preprocessing the problem.
- clock_analysefloat
the clock time spent analysing the required matrices prior to factorization.
- clock_factorizefloat
the clock time spent factorizing the required matrices.
- clock_solvefloat
the clock time spent computing the search direction.
- rqs_informdict
inform parameters for RQS (see
rqs.information
).- glrt_informdict
inform parameters for GLTR (see
glrt.information
).- psls_informdict
inform parameters for PSLS (see
psls.information
).- bsc_informdict
inform parameters for BSC (see
bsc.information
).- roots_informdict
inform parameters for ROOTS (see
roots.information
).
- nls.terminate()#
Deallocate all internal private storage.
example code#
from galahad import nls
import numpy as np
np.set_printoptions(precision=4,suppress=True,floatmode='fixed')
print("\n** python test: nls")
# allocate internal data and set default options
options = nls.initialize()
# set some non-default options
options['print_level'] = 0
options['jacobian_available'] = 2
options['hessian_available'] = 2
options['model'] = 6
#print("options:", options)
# set parameters
p = 1.0
n = 2
m = 3
# set Jacobian sparsity
J_type = 'coordinate'
J_ne = 5
J_row = np.array([0,1,1,2,2])
J_col = np.array([0,0,1,0,1])
J_ptr = None
# set Hessian sparsity
H_type = 'coordinate'
H_ne = 2
H_row = np.array([0,1])
H_col = np.array([0,1])
H_ptr = None
# set Hessian product sparsity
P_type = 'sparse_by_columns'
P_ne = 2
P_row = np.array([0,1])
P_col = None
P_ptr = np.array([0,1,2,2])
w = np.array([1.0,1.0,1.0])
# load data (and optionally non-default options)
nls.load(n, m,
J_type, J_ne, J_row, J_col, J_ptr,
H_type, H_ne, H_row, H_col, H_ptr,
P_type, P_ne, P_row, P_col, P_ptr,
w, options)
# define residual function and its derivatives
def eval_c(x):
return np.array([(x[0])**2 + p, x[0] + (x[1])**2, x[0] - x[1]])
def eval_j(x):
return np.array([2.0 * x[0], 1.0, 2.0 * x[1], 1.0, - 1.0])
def eval_h(x,y):
return np.array([2.0 * y[0], 2.0 * y[1]])
def eval_hprod(x,v):
return np.array([2.0 * v[0], 2.0 * v[1]])
# set starting point
x = np.array([1.5,1.5])
# find optimum
x, c, g = nls.solve(n, m, x, eval_c, J_ne, eval_j, H_ne, eval_h,
P_ne, eval_hprod)
print(" x:",x)
print(" c:",c)
print(" g:",g)
# get information
inform = nls.information()
#print(inform)
print(" f: %.4f" % inform['obj'])
print('** nls exit status:', inform['status'])
# deallocate internal data
nls.terminate()
This example code is available in $GALAHAD/src/nls/Python/test_nls.py .