adding weights to nonlinear least squares algorithm

Note: this uses MathML code produced by OpenOffice. It doesn't seem to work on Internet Explorer. It's also been rejected by Chrome. But it works on Safari, Chromium, and Firefox.

Nonlinear least-square fits are done with a Jacobian Matrix, which describes the linearized dependencies of the model, evaluated at each of the points, on each of the model parameters.  In this case I show two parameters, τ1 and τ2.  I like writing these things out, rather than using index notation, because index notation is a bit abstract.  In the following, f is a function of time t representing the model to be fit.

A = f τ 1 t 1 f τ 2 t 1 f τ 1 t 2 f τ 2 t 2 f τ 1 t N 1 f τ 2 t N 1 f τ 1 t N f τ 2 t N
 

With the unweighted nonlinear least-squares fit, the transpose of the Jocobian matrix is then taken:

A T = f τ 1 t 1 f τ 1 t 2 f τ 1 t N 1 f τ 1 t N f τ 2 t 1 f τ 2 t 2 f τ 2 t N 1 f τ 2 t N
 

This then creates a linear equation describing the iteration of the solution.  In the following, Pi are the points to be fit, of which there are N.  Point Pi is sampled at time ti.

A T A Δ τ 1 Δ τ 2 = A T P 1 f t 1 P 1 f t 1 P N 1 f t N 1 P N f t N
 

To weight this, I replace the AT matrix with the following:

A W T = w 1 f τ 1 t 1 w 2 f τ 1 t 2 w N 1 f τ 1 t N 1 w N f τ 1 t N w 1 f τ 2 t 1 w 2 f τ 2 t 2 w N 1 f τ 2 t N 1 w N f τ 2 t N
 

Then I solve the following modified equation for Δτ1 and Δτ2:

A w T A Δ τ 1 Δ τ 2 = A w T P 1 f t 1 P 1 f t 1 P N 1 f t N 1 P N f t N
 

On the right side of this equation, each Pi − f|ti are multiplied by a weighting factor wi, consistent with points being duplicated.  Similarly, on the left side of the equation, the normalization term is weighted, as it must be since all that matters is the relative weights, not the absolute weights.

It's good to check this with simple cases.  One is to reduce the number of parameters to one: τ.  Then additionally I'll reduce the number of data points to a single point: t1.

Then the above equation becomes:

w 1 f τ 2 Δ τ = w 1 f τ P 1 f t 1

Note the weights cancel, as they must, since there's only one point. It's easy to see from the equation that this is the correct result: it's Newton's method in one dimension.

If I change back to two parameters and two points, I get:

w 1 f τ 1 2 t 1 + w 2 f τ 1 2 t 2 Δ τ 1 + w 1 f τ 1 f τ 2 t 1 + w 2 f τ 1 f τ 2 t 2 Δ τ 2 = w 1 f τ 1 t 1 P 1 f t 1 + w 2 f τ 1 t 2 P 2 f t 2
and:
w 1 f τ 2 2 t 1 + w 2 f τ 2 2 t 2 Δ τ 2 + w 1 f τ 1 f τ 2 t 1 + w 2 f τ 1 f τ 2 t 2 Δ τ 1 = w 1 f τ 2 t 1 P 1 f t 1 + w 2 f τ 2 t 2 P 2 f t 2

Every term is multiplied by a single weight and so only the ratios of weights matters, as expected.

If I set w2 to zero while w1 remains positive, the above two equations collapse into the following simplified equation:

f τ 1 t 1 Δ τ 1 + f τ 2 t 1 Δ τ 2 = P 1 f t 1

This is clearly correct yet it is underconstrained: two unknowns for a single condition. But it shows the weights are working as expected: putting the emphasis on the points with the higher weighting coefficients.

Appendix: the weights can be represented as a diagonal square matrix, the weights on the diagonal. Then I can map ATA to AT W A, where W is the weight matrix, A is the Jacobian matrix described by Wolfram, and AT is the transpose. Then I simply use W AT where in the unweighted case I'd used AT.

Comments

Popular posts from this blog

Proposed update to the 1-second gap rule: 3-second gap

Post-Election Day

Marin Avenue (Berkeley)