aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'sci-libs/levmar/levmar-2.5/lmbc_core.c')
-rw-r--r--sci-libs/levmar/levmar-2.5/lmbc_core.c948
1 files changed, 948 insertions, 0 deletions
diff --git a/sci-libs/levmar/levmar-2.5/lmbc_core.c b/sci-libs/levmar/levmar-2.5/lmbc_core.c
new file mode 100644
index 000000000..ca0fdb054
--- /dev/null
+++ b/sci-libs/levmar/levmar-2.5/lmbc_core.c
@@ -0,0 +1,948 @@
+/////////////////////////////////////////////////////////////////////////////////
+//
+// Levenberg - Marquardt non-linear minimization algorithm
+// Copyright (C) 2004-05 Manolis Lourakis (lourakis at ics forth gr)
+// Institute of Computer Science, Foundation for Research & Technology - Hellas
+// Heraklion, Crete, Greece.
+//
+// This program is free software; you can redistribute it and/or modify
+// it under the terms of the GNU General Public License as published by
+// the Free Software Foundation; either version 2 of the License, or
+// (at your option) any later version.
+//
+// This program is distributed in the hope that it will be useful,
+// but WITHOUT ANY WARRANTY; without even the implied warranty of
+// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+// GNU General Public License for more details.
+//
+/////////////////////////////////////////////////////////////////////////////////
+
+#ifndef LM_REAL // not included by lmbc.c
+#error This file should not be compiled directly!
+#endif
+
+
+/* precision-specific definitions */
+#define FUNC_STATE LM_ADD_PREFIX(func_state)
+#define LNSRCH LM_ADD_PREFIX(lnsrch)
+#define BOXPROJECT LM_ADD_PREFIX(boxProject)
+#define LEVMAR_BOX_CHECK LM_ADD_PREFIX(levmar_box_check)
+#define LEVMAR_BC_DER LM_ADD_PREFIX(levmar_bc_der)
+#define LEVMAR_BC_DIF LM_ADD_PREFIX(levmar_bc_dif)
+#define LEVMAR_FDIF_FORW_JAC_APPROX LM_ADD_PREFIX(levmar_fdif_forw_jac_approx)
+#define LEVMAR_FDIF_CENT_JAC_APPROX LM_ADD_PREFIX(levmar_fdif_cent_jac_approx)
+#define LEVMAR_TRANS_MAT_MAT_MULT LM_ADD_PREFIX(levmar_trans_mat_mat_mult)
+#define LEVMAR_L2NRMXMY LM_ADD_PREFIX(levmar_L2nrmxmy)
+#define LEVMAR_COVAR LM_ADD_PREFIX(levmar_covar)
+#define LMBC_DIF_DATA LM_ADD_PREFIX(lmbc_dif_data)
+#define LMBC_DIF_FUNC LM_ADD_PREFIX(lmbc_dif_func)
+#define LMBC_DIF_JACF LM_ADD_PREFIX(lmbc_dif_jacf)
+
+#ifdef HAVE_LAPACK
+#define AX_EQ_B_LU LM_ADD_PREFIX(Ax_eq_b_LU)
+#define AX_EQ_B_CHOL LM_ADD_PREFIX(Ax_eq_b_Chol)
+#define AX_EQ_B_QR LM_ADD_PREFIX(Ax_eq_b_QR)
+#define AX_EQ_B_QRLS LM_ADD_PREFIX(Ax_eq_b_QRLS)
+#define AX_EQ_B_SVD LM_ADD_PREFIX(Ax_eq_b_SVD)
+#define AX_EQ_B_BK LM_ADD_PREFIX(Ax_eq_b_BK)
+#else
+#define AX_EQ_B_LU LM_ADD_PREFIX(Ax_eq_b_LU_noLapack)
+#endif /* HAVE_LAPACK */
+
+/* find the median of 3 numbers */
+#define __MEDIAN3(a, b, c) ( ((a) >= (b))?\
+ ( ((c) >= (a))? (a) : ( ((c) <= (b))? (b) : (c) ) ) : \
+ ( ((c) >= (b))? (b) : ( ((c) <= (a))? (a) : (c) ) ) )
+
+#define _POW_ LM_CNST(2.1)
+
+#define __LSITMAX 150 // max #iterations for line search
+
+struct FUNC_STATE{
+ int n, *nfev;
+ LM_REAL *hx, *x;
+ void *adata;
+};
+
+static void
+LNSRCH(int m, LM_REAL *x, LM_REAL f, LM_REAL *g, LM_REAL *p, LM_REAL alpha, LM_REAL *xpls,
+ LM_REAL *ffpls, void (*func)(LM_REAL *p, LM_REAL *hx, int m, int n, void *adata), struct FUNC_STATE state,
+ int *mxtake, int *iretcd, LM_REAL stepmx, LM_REAL steptl, LM_REAL *sx)
+{
+/* Find a next newton iterate by backtracking line search.
+ * Specifically, finds a \lambda such that for a fixed alpha<0.5 (usually 1e-4),
+ * f(x + \lambda*p) <= f(x) + alpha * \lambda * g^T*p
+ *
+ * Translated (with minor changes) from Schnabel, Koontz & Weiss uncmin.f, v1.3
+
+ * PARAMETERS :
+
+ * m --> dimension of problem (i.e. number of variables)
+ * x(m) --> old iterate: x[k-1]
+ * f --> function value at old iterate, f(x)
+ * g(m) --> gradient at old iterate, g(x), or approximate
+ * p(m) --> non-zero newton step
+ * alpha --> fixed constant < 0.5 for line search (see above)
+ * xpls(m) <-- new iterate x[k]
+ * ffpls <-- function value at new iterate, f(xpls)
+ * func --> name of subroutine to evaluate function
+ * state <--> information other than x and m that func requires.
+ * state is not modified in xlnsrch (but can be modified by func).
+ * iretcd <-- return code
+ * mxtake <-- boolean flag indicating step of maximum length used
+ * stepmx --> maximum allowable step size
+ * steptl --> relative step size at which successive iterates
+ * considered close enough to terminate algorithm
+ * sx(m) --> diagonal scaling matrix for x, can be NULL
+
+ * internal variables
+
+ * sln newton length
+ * rln relative length of newton step
+*/
+
+ register int i, j;
+ int firstback = 1;
+ LM_REAL disc;
+ LM_REAL a3, b;
+ LM_REAL t1, t2, t3, lambda, tlmbda, rmnlmb;
+ LM_REAL scl, rln, sln, slp;
+ LM_REAL tmp1, tmp2;
+ LM_REAL fpls, pfpls = 0., plmbda = 0.; /* -Wall */
+
+ f*=LM_CNST(0.5);
+ *mxtake = 0;
+ *iretcd = 2;
+ tmp1 = 0.;
+ if(!sx) /* no scaling */
+ for (i = 0; i < m; ++i)
+ tmp1 += p[i] * p[i];
+ else
+ for (i = 0; i < m; ++i)
+ tmp1 += sx[i] * sx[i] * p[i] * p[i];
+ sln = (LM_REAL)sqrt(tmp1);
+ if (sln > stepmx) {
+ /* newton step longer than maximum allowed */
+ scl = stepmx / sln;
+ for(i=0; i<m; ++i) /* p * scl */
+ p[i]*=scl;
+ sln = stepmx;
+ }
+ for(i=0, slp=0.; i<m; ++i) /* g^T * p */
+ slp+=g[i]*p[i];
+ rln = 0.;
+ if(!sx) /* no scaling */
+ for (i = 0; i < m; ++i) {
+ tmp1 = (FABS(x[i])>=LM_CNST(1.))? FABS(x[i]) : LM_CNST(1.);
+ tmp2 = FABS(p[i])/tmp1;
+ if(rln < tmp2) rln = tmp2;
+ }
+ else
+ for (i = 0; i < m; ++i) {
+ tmp1 = (FABS(x[i])>=LM_CNST(1.)/sx[i])? FABS(x[i]) : LM_CNST(1.)/sx[i];
+ tmp2 = FABS(p[i])/tmp1;
+ if(rln < tmp2) rln = tmp2;
+ }
+ rmnlmb = steptl / rln;
+ lambda = LM_CNST(1.0);
+
+ /* check if new iterate satisfactory. generate new lambda if necessary. */
+
+ for(j=__LSITMAX; j>=0; --j) {
+ for (i = 0; i < m; ++i)
+ xpls[i] = x[i] + lambda * p[i];
+
+ /* evaluate function at new point */
+ (*func)(xpls, state.hx, m, state.n, state.adata); ++(*(state.nfev));
+ /* ### state.hx=state.x-state.hx, tmp1=||state.hx|| */
+#if 1
+ tmp1=LEVMAR_L2NRMXMY(state.hx, state.x, state.hx, state.n);
+#else
+ for(i=0, tmp1=0.0; i<state.n; ++i){
+ state.hx[i]=tmp2=state.x[i]-state.hx[i];
+ tmp1+=tmp2*tmp2;
+ }
+#endif
+ fpls=LM_CNST(0.5)*tmp1; *ffpls=tmp1;
+
+ if (fpls <= f + slp * alpha * lambda) { /* solution found */
+ *iretcd = 0;
+ if (lambda == LM_CNST(1.) && sln > stepmx * LM_CNST(.99)) *mxtake = 1;
+ return;
+ }
+
+ /* else : solution not (yet) found */
+
+ /* First find a point with a finite value */
+
+ if (lambda < rmnlmb) {
+ /* no satisfactory xpls found sufficiently distinct from x */
+
+ *iretcd = 1;
+ return;
+ }
+ else { /* calculate new lambda */
+
+ /* modifications to cover non-finite values */
+ if (!LM_FINITE(fpls)) {
+ lambda *= LM_CNST(0.1);
+ firstback = 1;
+ }
+ else {
+ if (firstback) { /* first backtrack: quadratic fit */
+ tlmbda = -lambda * slp / ((fpls - f - slp) * LM_CNST(2.));
+ firstback = 0;
+ }
+ else { /* all subsequent backtracks: cubic fit */
+ t1 = fpls - f - lambda * slp;
+ t2 = pfpls - f - plmbda * slp;
+ t3 = LM_CNST(1.) / (lambda - plmbda);
+ a3 = LM_CNST(3.) * t3 * (t1 / (lambda * lambda)
+ - t2 / (plmbda * plmbda));
+ b = t3 * (t2 * lambda / (plmbda * plmbda)
+ - t1 * plmbda / (lambda * lambda));
+ disc = b * b - a3 * slp;
+ if (disc > b * b)
+ /* only one positive critical point, must be minimum */
+ tlmbda = (-b + ((a3 < 0)? -(LM_REAL)sqrt(disc): (LM_REAL)sqrt(disc))) /a3;
+ else
+ /* both critical points positive, first is minimum */
+ tlmbda = (-b + ((a3 < 0)? (LM_REAL)sqrt(disc): -(LM_REAL)sqrt(disc))) /a3;
+
+ if (tlmbda > lambda * LM_CNST(.5))
+ tlmbda = lambda * LM_CNST(.5);
+ }
+ plmbda = lambda;
+ pfpls = fpls;
+ if (tlmbda < lambda * LM_CNST(.1))
+ lambda *= LM_CNST(.1);
+ else
+ lambda = tlmbda;
+ }
+ }
+ }
+ /* this point is reached when the iterations limit is exceeded */
+ *iretcd = 1; /* failed */
+ return;
+} /* LNSRCH */
+
+/* Projections to feasible set \Omega: P_{\Omega}(y) := arg min { ||x - y|| : x \in \Omega}, y \in R^m */
+
+/* project vector p to a box shaped feasible set. p is a mx1 vector.
+ * Either lb, ub can be NULL. If not NULL, they are mx1 vectors
+ */
+static void BOXPROJECT(LM_REAL *p, LM_REAL *lb, LM_REAL *ub, int m)
+{
+register int i;
+
+ if(!lb){ /* no lower bounds */
+ if(!ub) /* no upper bounds */
+ return;
+ else{ /* upper bounds only */
+ for(i=0; i<m; ++i)
+ if(p[i]>ub[i]) p[i]=ub[i];
+ }
+ }
+ else
+ if(!ub){ /* lower bounds only */
+ for(i=0; i<m; ++i)
+ if(p[i]<lb[i]) p[i]=lb[i];
+ }
+ else /* box bounds */
+ for(i=0; i<m; ++i)
+ p[i]=__MEDIAN3(lb[i], p[i], ub[i]);
+}
+
+/*
+ * This function seeks the parameter vector p that best describes the measurements
+ * vector x under box constraints.
+ * More precisely, given a vector function func : R^m --> R^n with n>=m,
+ * it finds p s.t. func(p) ~= x, i.e. the squared second order (i.e. L2) norm of
+ * e=x-func(p) is minimized under the constraints lb[i]<=p[i]<=ub[i].
+ * If no lower bound constraint applies for p[i], use -DBL_MAX/-FLT_MAX for lb[i];
+ * If no upper bound constraint applies for p[i], use DBL_MAX/FLT_MAX for ub[i].
+ *
+ * This function requires an analytic Jacobian. In case the latter is unavailable,
+ * use LEVMAR_BC_DIF() bellow
+ *
+ * Returns the number of iterations (>=0) if successful, LM_ERROR if failed
+ *
+ * For details, see C. Kanzow, N. Yamashita and M. Fukushima: "Levenberg-Marquardt
+ * methods for constrained nonlinear equations with strong local convergence properties",
+ * Journal of Computational and Applied Mathematics 172, 2004, pp. 375-397.
+ * Also, see K. Madsen, H.B. Nielsen and O. Tingleff's lecture notes on
+ * unconstrained Levenberg-Marquardt at http://www.imm.dtu.dk/pubdb/views/edoc_download.php/3215/pdf/imm3215.pdf
+ */
+
+int LEVMAR_BC_DER(
+ void (*func)(LM_REAL *p, LM_REAL *hx, int m, int n, void *adata), /* functional relation describing measurements. A p \in R^m yields a \hat{x} \in R^n */
+ void (*jacf)(LM_REAL *p, LM_REAL *j, int m, int n, void *adata), /* function to evaluate the Jacobian \part x / \part p */
+ LM_REAL *p, /* I/O: initial parameter estimates. On output has the estimated solution */
+ LM_REAL *x, /* I: measurement vector. NULL implies a zero vector */
+ int m, /* I: parameter vector dimension (i.e. #unknowns) */
+ int n, /* I: measurement vector dimension */
+ LM_REAL *lb, /* I: vector of lower bounds. If NULL, no lower bounds apply */
+ LM_REAL *ub, /* I: vector of upper bounds. If NULL, no upper bounds apply */
+ int itmax, /* I: maximum number of iterations */
+ LM_REAL opts[4], /* I: minim. options [\mu, \epsilon1, \epsilon2, \epsilon3]. Respectively the scale factor for initial \mu,
+ * stopping thresholds for ||J^T e||_inf, ||Dp||_2 and ||e||_2. Set to NULL for defaults to be used.
+ * Note that ||J^T e||_inf is computed on free (not equal to lb[i] or ub[i]) variables only.
+ */
+ LM_REAL info[LM_INFO_SZ],
+ /* O: information regarding the minimization. Set to NULL if don't care
+ * info[0]= ||e||_2 at initial p.
+ * info[1-4]=[ ||e||_2, ||J^T e||_inf, ||Dp||_2, mu/max[J^T J]_ii ], all computed at estimated p.
+ * info[5]= # iterations,
+ * info[6]=reason for terminating: 1 - stopped by small gradient J^T e
+ * 2 - stopped by small Dp
+ * 3 - stopped by itmax
+ * 4 - singular matrix. Restart from current p with increased mu
+ * 5 - no further error reduction is possible. Restart with increased mu
+ * 6 - stopped by small ||e||_2
+ * 7 - stopped by invalid (i.e. NaN or Inf) "func" values. This is a user error
+ * info[7]= # function evaluations
+ * info[8]= # Jacobian evaluations
+ * info[9]= # linear systems solved, i.e. # attempts for reducing error
+ */
+ LM_REAL *work, /* working memory at least LM_BC_DER_WORKSZ() reals large, allocated if NULL */
+ LM_REAL *covar, /* O: Covariance matrix corresponding to LS solution; mxm. Set to NULL if not needed. */
+ void *adata) /* pointer to possibly additional data, passed uninterpreted to func & jacf.
+ * Set to NULL if not needed
+ */
+{
+register int i, j, k, l;
+int worksz, freework=0, issolved;
+/* temp work arrays */
+LM_REAL *e, /* nx1 */
+ *hx, /* \hat{x}_i, nx1 */
+ *jacTe, /* J^T e_i mx1 */
+ *jac, /* nxm */
+ *jacTjac, /* mxm */
+ *Dp, /* mx1 */
+ *diag_jacTjac, /* diagonal of J^T J, mx1 */
+ *pDp; /* p + Dp, mx1 */
+
+register LM_REAL mu, /* damping constant */
+ tmp; /* mainly used in matrix & vector multiplications */
+LM_REAL p_eL2, jacTe_inf, pDp_eL2; /* ||e(p)||_2, ||J^T e||_inf, ||e(p+Dp)||_2 */
+LM_REAL p_L2, Dp_L2=LM_REAL_MAX, dF, dL;
+LM_REAL tau, eps1, eps2, eps2_sq, eps3;
+LM_REAL init_p_eL2;
+int nu=2, nu2, stop=0, nfev, njev=0, nlss=0;
+const int nm=n*m;
+
+/* variables for constrained LM */
+struct FUNC_STATE fstate;
+LM_REAL alpha=LM_CNST(1e-4), beta=LM_CNST(0.9), gamma=LM_CNST(0.99995), gamma_sq=gamma*gamma, rho=LM_CNST(1e-8);
+LM_REAL t, t0;
+LM_REAL steptl=LM_CNST(1e3)*(LM_REAL)sqrt(LM_REAL_EPSILON), jacTeDp;
+LM_REAL tmin=LM_CNST(1e-12), tming=LM_CNST(1e-18); /* minimum step length for LS and PG steps */
+const LM_REAL tini=LM_CNST(1.0); /* initial step length for LS and PG steps */
+int nLMsteps=0, nLSsteps=0, nPGsteps=0, gprevtaken=0;
+int numactive;
+int (*linsolver)(LM_REAL *A, LM_REAL *B, LM_REAL *x, int m)=NULL;
+
+ mu=jacTe_inf=t=0.0; tmin=tmin; /* -Wall */
+
+ if(n<m){
+ fprintf(stderr, LCAT(LEVMAR_BC_DER, "(): cannot solve a problem with fewer measurements [%d] than unknowns [%d]\n"), n, m);
+ return LM_ERROR;
+ }
+
+ if(!jacf){
+ fprintf(stderr, RCAT("No function specified for computing the Jacobian in ", LEVMAR_BC_DER)
+ RCAT("().\nIf no such function is available, use ", LEVMAR_BC_DIF) RCAT("() rather than ", LEVMAR_BC_DER) "()\n");
+ return LM_ERROR;
+ }
+
+ if(!LEVMAR_BOX_CHECK(lb, ub, m)){
+ fprintf(stderr, LCAT(LEVMAR_BC_DER, "(): at least one lower bound exceeds the upper one\n"));
+ return LM_ERROR;
+ }
+
+ if(opts){
+ tau=opts[0];
+ eps1=opts[1];
+ eps2=opts[2];
+ eps2_sq=opts[2]*opts[2];
+ eps3=opts[3];
+ }
+ else{ // use default values
+ tau=LM_CNST(LM_INIT_MU);
+ eps1=LM_CNST(LM_STOP_THRESH);
+ eps2=LM_CNST(LM_STOP_THRESH);
+ eps2_sq=LM_CNST(LM_STOP_THRESH)*LM_CNST(LM_STOP_THRESH);
+ eps3=LM_CNST(LM_STOP_THRESH);
+ }
+
+ if(!work){
+ worksz=LM_BC_DER_WORKSZ(m, n); //2*n+4*m + n*m + m*m;
+ work=(LM_REAL *)malloc(worksz*sizeof(LM_REAL)); /* allocate a big chunk in one step */
+ if(!work){
+ fprintf(stderr, LCAT(LEVMAR_BC_DER, "(): memory allocation request failed\n"));
+ return LM_ERROR;
+ }
+ freework=1;
+ }
+
+ /* set up work arrays */
+ e=work;
+ hx=e + n;
+ jacTe=hx + n;
+ jac=jacTe + m;
+ jacTjac=jac + nm;
+ Dp=jacTjac + m*m;
+ diag_jacTjac=Dp + m;
+ pDp=diag_jacTjac + m;
+
+ fstate.n=n;
+ fstate.hx=hx;
+ fstate.x=x;
+ fstate.adata=adata;
+ fstate.nfev=&nfev;
+
+ /* see if starting point is within the feasile set */
+ for(i=0; i<m; ++i)
+ pDp[i]=p[i];
+ BOXPROJECT(p, lb, ub, m); /* project to feasible set */
+ for(i=0; i<m; ++i)
+ if(pDp[i]!=p[i])
+ fprintf(stderr, RCAT("Warning: component %d of starting point not feasible in ", LEVMAR_BC_DER) "()! [%g projected to %g]\n",
+ i, pDp[i], p[i]);
+
+ /* compute e=x - f(p) and its L2 norm */
+ (*func)(p, hx, m, n, adata); nfev=1;
+ /* ### e=x-hx, p_eL2=||e|| */
+#if 1
+ p_eL2=LEVMAR_L2NRMXMY(e, x, hx, n);
+#else
+ for(i=0, p_eL2=0.0; i<n; ++i){
+ e[i]=tmp=x[i]-hx[i];
+ p_eL2+=tmp*tmp;
+ }
+#endif
+ init_p_eL2=p_eL2;
+ if(!LM_FINITE(p_eL2)) stop=7;
+
+ for(k=0; k<itmax && !stop; ++k){
+ /* Note that p and e have been updated at a previous iteration */
+
+ if(p_eL2<=eps3){ /* error is small */
+ stop=6;
+ break;
+ }
+
+ /* Compute the Jacobian J at p, J^T J, J^T e, ||J^T e||_inf and ||p||^2.
+ * Since J^T J is symmetric, its computation can be sped up by computing
+ * only its upper triangular part and copying it to the lower part
+ */
+
+ (*jacf)(p, jac, m, n, adata); ++njev;
+
+ /* J^T J, J^T e */
+ if(nm<__BLOCKSZ__SQ){ // this is a small problem
+ /* J^T*J_ij = \sum_l J^T_il * J_lj = \sum_l J_li * J_lj.
+ * Thus, the product J^T J can be computed using an outer loop for
+ * l that adds J_li*J_lj to each element ij of the result. Note that
+ * with this scheme, the accesses to J and JtJ are always along rows,
+ * therefore induces less cache misses compared to the straightforward
+ * algorithm for computing the product (i.e., l loop is innermost one).
+ * A similar scheme applies to the computation of J^T e.
+ * However, for large minimization problems (i.e., involving a large number
+ * of unknowns and measurements) for which J/J^T J rows are too large to
+ * fit in the L1 cache, even this scheme incures many cache misses. In
+ * such cases, a cache-efficient blocking scheme is preferable.
+ *
+ * Thanks to John Nitao of Lawrence Livermore Lab for pointing out this
+ * performance problem.
+ *
+ * Note that the non-blocking algorithm is faster on small
+ * problems since in this case it avoids the overheads of blocking.
+ */
+ register int l, im;
+ register LM_REAL alpha, *jaclm;
+
+ /* looping downwards saves a few computations */
+ for(i=m*m; i-->0; )
+ jacTjac[i]=0.0;
+ for(i=m; i-->0; )
+ jacTe[i]=0.0;
+
+ for(l=n; l-->0; ){
+ jaclm=jac+l*m;
+ for(i=m; i-->0; ){
+ im=i*m;
+ alpha=jaclm[i]; //jac[l*m+i];
+ for(j=i+1; j-->0; ) /* j<=i computes lower triangular part only */
+ jacTjac[im+j]+=jaclm[j]*alpha; //jac[l*m+j]
+
+ /* J^T e */
+ jacTe[i]+=alpha*e[l];
+ }
+ }
+
+ for(i=m; i-->0; ) /* copy to upper part */
+ for(j=i+1; j<m; ++j)
+ jacTjac[i*m+j]=jacTjac[j*m+i];
+ }
+ else{ // this is a large problem
+ /* Cache efficient computation of J^T J based on blocking
+ */
+ LEVMAR_TRANS_MAT_MAT_MULT(jac, jacTjac, n, m);
+
+ /* cache efficient computation of J^T e */
+ for(i=0; i<m; ++i)
+ jacTe[i]=0.0;
+
+ for(i=0; i<n; ++i){
+ register LM_REAL *jacrow;
+
+ for(l=0, jacrow=jac+i*m, tmp=e[i]; l<m; ++l)
+ jacTe[l]+=jacrow[l]*tmp;
+ }
+ }
+
+ /* Compute ||J^T e||_inf and ||p||^2. Note that ||J^T e||_inf
+ * is computed for free (i.e. inactive) variables only.
+ * At a local minimum, if p[i]==ub[i] then g[i]>0;
+ * if p[i]==lb[i] g[i]<0; otherwise g[i]=0
+ */
+ for(i=j=numactive=0, p_L2=jacTe_inf=0.0; i<m; ++i){
+ if(ub && p[i]==ub[i]){ ++numactive; if(jacTe[i]>0.0) ++j; }
+ else if(lb && p[i]==lb[i]){ ++numactive; if(jacTe[i]<0.0) ++j; }
+ else if(jacTe_inf < (tmp=FABS(jacTe[i]))) jacTe_inf=tmp;
+
+ diag_jacTjac[i]=jacTjac[i*m+i]; /* save diagonal entries so that augmentation can be later canceled */
+ p_L2+=p[i]*p[i];
+ }
+ //p_L2=sqrt(p_L2);
+
+#if 0
+if(!(k%100)){
+ printf("Current estimate: ");
+ for(i=0; i<m; ++i)
+ printf("%.9g ", p[i]);
+ printf("-- errors %.9g %0.9g, #active %d [%d]\n", jacTe_inf, p_eL2, numactive, j);
+}
+#endif
+
+ /* check for convergence */
+ if(j==numactive && (jacTe_inf <= eps1)){
+ Dp_L2=0.0; /* no increment for p in this case */
+ stop=1;
+ break;
+ }
+
+ /* compute initial damping factor */
+ if(k==0){
+ if(!lb && !ub){ /* no bounds */
+ for(i=0, tmp=LM_REAL_MIN; i<m; ++i)
+ if(diag_jacTjac[i]>tmp) tmp=diag_jacTjac[i]; /* find max diagonal element */
+ mu=tau*tmp;
+ }
+ else
+ mu=LM_CNST(0.5)*tau*p_eL2; /* use Kanzow's starting mu */
+ }
+
+ /* determine increment using a combination of adaptive damping, line search and projected gradient search */
+ while(1){
+ /* augment normal equations */
+ for(i=0; i<m; ++i)
+ jacTjac[i*m+i]+=mu;
+
+ /* solve augmented equations */
+#ifdef HAVE_LAPACK
+ /* 6 alternatives are available: LU, Cholesky, 2 variants of QR decomposition, SVD and LDLt.
+ * Cholesky is the fastest but might be inaccurate; QR is slower but more accurate;
+ * SVD is the slowest but most accurate; LU offers a tradeoff between accuracy and speed
+ */
+
+ issolved=AX_EQ_B_BK(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_BK;
+ //issolved=AX_EQ_B_LU(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_LU;
+ //issolved=AX_EQ_B_CHOL(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_CHOL;
+ //issolved=AX_EQ_B_QR(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_QR;
+ //issolved=AX_EQ_B_QRLS(jacTjac, jacTe, Dp, m, m); ++nlss; linsolver=(int (*)(LM_REAL *A, LM_REAL *B, LM_REAL *x, int m))AX_EQ_B_QRLS;
+ //issolved=AX_EQ_B_SVD(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_SVD;
+
+#else
+ /* use the LU included with levmar */
+ issolved=AX_EQ_B_LU(jacTjac, jacTe, Dp, m); ++nlss; linsolver=AX_EQ_B_LU;
+#endif /* HAVE_LAPACK */
+
+ if(issolved){
+ for(i=0; i<m; ++i)
+ pDp[i]=p[i] + Dp[i];
+
+ /* compute p's new estimate and ||Dp||^2 */
+ BOXPROJECT(pDp, lb, ub, m); /* project to feasible set */
+ for(i=0, Dp_L2=0.0; i<m; ++i){
+ Dp[i]=tmp=pDp[i]-p[i];
+ Dp_L2+=tmp*tmp;
+ }
+ //Dp_L2=sqrt(Dp_L2);
+
+ if(Dp_L2<=eps2_sq*p_L2){ /* relative change in p is small, stop */
+ stop=2;
+ break;
+ }
+
+ if(Dp_L2>=(p_L2+eps2)/(LM_CNST(EPSILON)*LM_CNST(EPSILON))){ /* almost singular */
+ stop=4;
+ break;
+ }
+
+ (*func)(pDp, hx, m, n, adata); ++nfev; /* evaluate function at p + Dp */
+ /* ### hx=x-hx, pDp_eL2=||hx|| */
+#if 1
+ pDp_eL2=LEVMAR_L2NRMXMY(hx, x, hx, n);
+#else
+ for(i=0, pDp_eL2=0.0; i<n; ++i){ /* compute ||e(pDp)||_2 */
+ hx[i]=tmp=x[i]-hx[i];
+ pDp_eL2+=tmp*tmp;
+ }
+#endif
+ if(!LM_FINITE(pDp_eL2)){
+ stop=7;
+ break;
+ }
+
+ if(pDp_eL2<=gamma_sq*p_eL2){
+ for(i=0, dL=0.0; i<m; ++i)
+ dL+=Dp[i]*(mu*Dp[i]+jacTe[i]);
+
+#if 1
+ if(dL>0.0){
+ dF=p_eL2-pDp_eL2;
+ tmp=(LM_CNST(2.0)*dF/dL-LM_CNST(1.0));
+ tmp=LM_CNST(1.0)-tmp*tmp*tmp;
+ mu=mu*( (tmp>=LM_CNST(ONE_THIRD))? tmp : LM_CNST(ONE_THIRD) );
+ }
+ else
+ mu=(mu>=pDp_eL2)? pDp_eL2 : mu; /* pDp_eL2 is the new pDp_eL2 */
+#else
+
+ mu=(mu>=pDp_eL2)? pDp_eL2 : mu; /* pDp_eL2 is the new pDp_eL2 */
+#endif
+
+ nu=2;
+
+ for(i=0 ; i<m; ++i) /* update p's estimate */
+ p[i]=pDp[i];
+
+ for(i=0; i<n; ++i) /* update e and ||e||_2 */
+ e[i]=hx[i];
+ p_eL2=pDp_eL2;
+ ++nLMsteps;
+ gprevtaken=0;
+ break;
+ }
+ }
+ else{
+
+ /* the augmented linear system could not be solved, increase mu */
+
+ mu*=nu;
+ nu2=nu<<1; // 2*nu;
+ if(nu2<=nu){ /* nu has wrapped around (overflown). Thanks to Frank Jordan for spotting this case */
+ stop=5;
+ break;
+ }
+ nu=nu2;
+
+ for(i=0; i<m; ++i) /* restore diagonal J^T J entries */
+ jacTjac[i*m+i]=diag_jacTjac[i];
+
+ continue; /* solve again with increased nu */
+ }
+
+ /* if this point is reached, the LM step did not reduce the error;
+ * see if it is a descent direction
+ */
+
+ /* negate jacTe (i.e. g) & compute g^T * Dp */
+ for(i=0, jacTeDp=0.0; i<m; ++i){
+ jacTe[i]=-jacTe[i];
+ jacTeDp+=jacTe[i]*Dp[i];
+ }
+
+ if(jacTeDp<=-rho*pow(Dp_L2, _POW_/LM_CNST(2.0))){
+ /* Dp is a descent direction; do a line search along it */
+ int mxtake, iretcd;
+ LM_REAL stepmx;
+
+ tmp=(LM_REAL)sqrt(p_L2); stepmx=LM_CNST(1e3)*( (tmp>=LM_CNST(1.0))? tmp : LM_CNST(1.0) );
+
+#if 1
+ /* use Schnabel's backtracking line search; it requires fewer "func" evaluations */
+ LNSRCH(m, p, p_eL2, jacTe, Dp, alpha, pDp, &pDp_eL2, func, fstate,
+ &mxtake, &iretcd, stepmx, steptl, NULL); /* NOTE: LNSRCH() updates hx */
+ if(iretcd!=0) goto gradproj; /* rather inelegant but effective way to handle LNSRCH() failures... */
+#else
+ /* use the simpler (but slower!) line search described by Kanzow et al */
+ for(t=tini; t>tmin; t*=beta){
+ for(i=0; i<m; ++i){
+ pDp[i]=p[i] + t*Dp[i];
+ //pDp[i]=__MEDIAN3(lb[i], pDp[i], ub[i]); /* project to feasible set */
+ }
+
+ (*func)(pDp, hx, m, n, adata); ++nfev; /* evaluate function at p + t*Dp */
+ for(i=0, pDp_eL2=0.0; i<n; ++i){ /* compute ||e(pDp)||_2 */
+ hx[i]=tmp=x[i]-hx[i];
+ pDp_eL2+=tmp*tmp;
+ }
+ if(!LM_FINITE(pDp_eL2)) goto gradproj; /* treat as line search failure */
+
+ //if(LM_CNST(0.5)*pDp_eL2<=LM_CNST(0.5)*p_eL2 + t*alpha*jacTeDp) break;
+ if(pDp_eL2<=p_eL2 + LM_CNST(2.0)*t*alpha*jacTeDp) break;
+ }
+#endif
+ ++nLSsteps;
+ gprevtaken=0;
+
+ /* NOTE: new estimate for p is in pDp, associated error in hx and its norm in pDp_eL2.
+ * These values are used below to update their corresponding variables
+ */
+ }
+ else{
+gradproj: /* Note that this point can also be reached via a goto when LNSRCH() fails */
+
+ /* jacTe is a descent direction; make a projected gradient step */
+
+ /* if the previous step was along the gradient descent, try to use the t employed in that step */
+ /* compute ||g|| */
+ for(i=0, tmp=0.0; i<m; ++i)
+ tmp+=jacTe[i]*jacTe[i];
+ tmp=(LM_REAL)sqrt(tmp);
+ tmp=LM_CNST(100.0)/(LM_CNST(1.0)+tmp);
+ t0=(tmp<=tini)? tmp : tini; /* guard against poor scaling & large steps; see (3.50) in C.T. Kelley's book */
+
+ for(t=(gprevtaken)? t : t0; t>tming; t*=beta){
+ for(i=0; i<m; ++i)
+ pDp[i]=p[i] - t*jacTe[i];
+ BOXPROJECT(pDp, lb, ub, m); /* project to feasible set */
+ for(i=0; i<m; ++i)
+ Dp[i]=pDp[i]-p[i];
+
+ (*func)(pDp, hx, m, n, adata); ++nfev; /* evaluate function at p - t*g */
+ /* compute ||e(pDp)||_2 */
+ /* ### hx=x-hx, pDp_eL2=||hx|| */
+#if 1
+ pDp_eL2=LEVMAR_L2NRMXMY(hx, x, hx, n);
+#else
+ for(i=0, pDp_eL2=0.0; i<n; ++i){
+ hx[i]=tmp=x[i]-hx[i];
+ pDp_eL2+=tmp*tmp;
+ }
+#endif
+ if(!LM_FINITE(pDp_eL2)){
+ stop=7;
+ goto breaknested;
+ }
+
+ for(i=0, tmp=0.0; i<m; ++i) /* compute ||g^T * Dp|| */
+ tmp+=jacTe[i]*Dp[i];
+
+ if(gprevtaken && pDp_eL2<=p_eL2 + LM_CNST(2.0)*LM_CNST(0.99999)*tmp){ /* starting t too small */
+ t=t0;
+ gprevtaken=0;
+ continue;
+ }
+ //if(LM_CNST(0.5)*pDp_eL2<=LM_CNST(0.5)*p_eL2 + alpha*tmp) break;
+ if(pDp_eL2<=p_eL2 + LM_CNST(2.0)*alpha*tmp) break;
+ }
+
+ ++nPGsteps;
+ gprevtaken=1;
+ /* NOTE: new estimate for p is in pDp, associated error in hx and its norm in pDp_eL2 */
+ }
+
+ /* update using computed values */
+
+ for(i=0, Dp_L2=0.0; i<m; ++i){
+ tmp=pDp[i]-p[i];
+ Dp_L2+=tmp*tmp;
+ }
+ //Dp_L2=sqrt(Dp_L2);
+
+ if(Dp_L2<=eps2_sq*p_L2){ /* relative change in p is small, stop */
+ stop=2;
+ break;
+ }
+
+ for(i=0 ; i<m; ++i) /* update p's estimate */
+ p[i]=pDp[i];
+
+ for(i=0; i<n; ++i) /* update e and ||e||_2 */
+ e[i]=hx[i];
+ p_eL2=pDp_eL2;
+ break;
+ } /* inner loop */
+ }
+
+breaknested: /* NOTE: this point is also reached via an explicit goto! */
+
+ if(k>=itmax) stop=3;
+
+ for(i=0; i<m; ++i) /* restore diagonal J^T J entries */
+ jacTjac[i*m+i]=diag_jacTjac[i];
+
+ if(info){
+ info[0]=init_p_eL2;
+ info[1]=p_eL2;
+ info[2]=jacTe_inf;
+ info[3]=Dp_L2;
+ for(i=0, tmp=LM_REAL_MIN; i<m; ++i)
+ if(tmp<jacTjac[i*m+i]) tmp=jacTjac[i*m+i];
+ info[4]=mu/tmp;
+ info[5]=(LM_REAL)k;
+ info[6]=(LM_REAL)stop;
+ info[7]=(LM_REAL)nfev;
+ info[8]=(LM_REAL)njev;
+ info[9]=(LM_REAL)nlss;
+ }
+
+ /* covariance matrix */
+ if(covar){
+ LEVMAR_COVAR(jacTjac, covar, p_eL2, m, n);
+ }
+
+ if(freework) free(work);
+
+#ifdef LINSOLVERS_RETAIN_MEMORY
+ if(linsolver) (*linsolver)(NULL, NULL, NULL, 0);
+#endif
+
+#if 0
+printf("%d LM steps, %d line search, %d projected gradient\n", nLMsteps, nLSsteps, nPGsteps);
+#endif
+
+ return (stop!=4 && stop!=7)? k : LM_ERROR;
+}
+
+/* following struct & LMBC_DIF_XXX functions won't be necessary if a true secant
+ * version of LEVMAR_BC_DIF() is implemented...
+ */
+struct LMBC_DIF_DATA{
+ int ffdif; // nonzero if forward differencing is used
+ void (*func)(LM_REAL *p, LM_REAL *hx, int m, int n, void *adata);
+ LM_REAL *hx, *hxx;
+ void *adata;
+ LM_REAL delta;
+};
+
+static void LMBC_DIF_FUNC(LM_REAL *p, LM_REAL *hx, int m, int n, void *data)
+{
+struct LMBC_DIF_DATA *dta=(struct LMBC_DIF_DATA *)data;
+
+ /* call user-supplied function passing it the user-supplied data */
+ (*(dta->func))(p, hx, m, n, dta->adata);
+}
+
+static void LMBC_DIF_JACF(LM_REAL *p, LM_REAL *jac, int m, int n, void *data)
+{
+struct LMBC_DIF_DATA *dta=(struct LMBC_DIF_DATA *)data;
+
+ if(dta->ffdif){
+ /* evaluate user-supplied function at p */
+ (*(dta->func))(p, dta->hx, m, n, dta->adata);
+ LEVMAR_FDIF_FORW_JAC_APPROX(dta->func, p, dta->hx, dta->hxx, dta->delta, jac, m, n, dta->adata);
+ }
+ else
+ LEVMAR_FDIF_CENT_JAC_APPROX(dta->func, p, dta->hx, dta->hxx, dta->delta, jac, m, n, dta->adata);
+}
+
+
+/* No Jacobian version of the LEVMAR_BC_DER() function above: the Jacobian is approximated with
+ * the aid of finite differences (forward or central, see the comment for the opts argument)
+ * Ideally, this function should be implemented with a secant approach. Currently, it just calls
+ * LEVMAR_BC_DER()
+ */
+int LEVMAR_BC_DIF(
+ void (*func)(LM_REAL *p, LM_REAL *hx, int m, int n, void *adata), /* functional relation describing measurements. A p \in R^m yields a \hat{x} \in R^n */
+ LM_REAL *p, /* I/O: initial parameter estimates. On output has the estimated solution */
+ LM_REAL *x, /* I: measurement vector. NULL implies a zero vector */
+ int m, /* I: parameter vector dimension (i.e. #unknowns) */
+ int n, /* I: measurement vector dimension */
+ LM_REAL *lb, /* I: vector of lower bounds. If NULL, no lower bounds apply */
+ LM_REAL *ub, /* I: vector of upper bounds. If NULL, no upper bounds apply */
+ int itmax, /* I: maximum number of iterations */
+ LM_REAL opts[5], /* I: opts[0-4] = minim. options [\mu, \epsilon1, \epsilon2, \epsilon3, \delta]. Respectively the
+ * scale factor for initial \mu, stopping thresholds for ||J^T e||_inf, ||Dp||_2 and ||e||_2 and
+ * the step used in difference approximation to the Jacobian. Set to NULL for defaults to be used.
+ * If \delta<0, the Jacobian is approximated with central differences which are more accurate
+ * (but slower!) compared to the forward differences employed by default.
+ */
+ LM_REAL info[LM_INFO_SZ],
+ /* O: information regarding the minimization. Set to NULL if don't care
+ * info[0]= ||e||_2 at initial p.
+ * info[1-4]=[ ||e||_2, ||J^T e||_inf, ||Dp||_2, mu/max[J^T J]_ii ], all computed at estimated p.
+ * info[5]= # iterations,
+ * info[6]=reason for terminating: 1 - stopped by small gradient J^T e
+ * 2 - stopped by small Dp
+ * 3 - stopped by itmax
+ * 4 - singular matrix. Restart from current p with increased mu
+ * 5 - no further error reduction is possible. Restart with increased mu
+ * 6 - stopped by small ||e||_2
+ * 7 - stopped by invalid (i.e. NaN or Inf) "func" values. This is a user error
+ * info[7]= # function evaluations
+ * info[8]= # Jacobian evaluations
+ * info[9]= # linear systems solved, i.e. # attempts for reducing error
+ */
+ LM_REAL *work, /* working memory at least LM_BC_DIF_WORKSZ() reals large, allocated if NULL */
+ LM_REAL *covar, /* O: Covariance matrix corresponding to LS solution; mxm. Set to NULL if not needed. */
+ void *adata) /* pointer to possibly additional data, passed uninterpreted to func.
+ * Set to NULL if not needed
+ */
+{
+struct LMBC_DIF_DATA data;
+int ret;
+
+ //fprintf(stderr, RCAT("\nWarning: current implementation of ", LEVMAR_BC_DIF) "() does not use a secant approach!\n\n");
+
+ data.ffdif=!opts || opts[4]>=0.0;
+
+ data.func=func;
+ data.hx=(LM_REAL *)malloc(2*n*sizeof(LM_REAL)); /* allocate a big chunk in one step */
+ if(!data.hx){
+ fprintf(stderr, LCAT(LEVMAR_BC_DIF, "(): memory allocation request failed\n"));
+ return LM_ERROR;
+ }
+ data.hxx=data.hx+n;
+ data.adata=adata;
+ data.delta=(opts)? FABS(opts[4]) : (LM_REAL)LM_DIFF_DELTA;
+
+ ret=LEVMAR_BC_DER(LMBC_DIF_FUNC, LMBC_DIF_JACF, p, x, m, n, lb, ub, itmax, opts, info, work, covar, (void *)&data);
+
+ if(info){ /* correct the number of function calls */
+ if(data.ffdif)
+ info[7]+=info[8]*(m+1); /* each Jacobian evaluation costs m+1 function calls */
+ else
+ info[7]+=info[8]*(2*m); /* each Jacobian evaluation costs 2*m function calls */
+ }
+
+ free(data.hx);
+
+ return ret;
+}
+
+/* undefine everything. THIS MUST REMAIN AT THE END OF THE FILE */
+#undef FUNC_STATE
+#undef LNSRCH
+#undef BOXPROJECT
+#undef LEVMAR_BOX_CHECK
+#undef LEVMAR_BC_DER
+#undef LMBC_DIF_DATA
+#undef LMBC_DIF_FUNC
+#undef LMBC_DIF_JACF
+#undef LEVMAR_BC_DIF
+#undef LEVMAR_FDIF_FORW_JAC_APPROX
+#undef LEVMAR_FDIF_CENT_JAC_APPROX
+#undef LEVMAR_COVAR
+#undef LEVMAR_TRANS_MAT_MAT_MULT
+#undef LEVMAR_L2NRMXMY
+#undef AX_EQ_B_LU
+#undef AX_EQ_B_CHOL
+#undef AX_EQ_B_QR
+#undef AX_EQ_B_QRLS
+#undef AX_EQ_B_SVD
+#undef AX_EQ_B_BK