Consider the biquadratic representation of an ellipse:
Given n noisy points ( ), we want to estimate the coefficients of the ellipse: . Due to the homogeneity, we set .
For each point , we thus have one scalar equation:
where
Hence, can be estimated by minimizing the following objective function (weighted least-squares optimization)
where 's are positive weights.
Assume that each point has the same error distribution with mean zero and covariance . The covariance of is then given by
where
Thus we have
The weights can then be chosen to the inverse proportion of the variances. Since multiplication by a constant does not affect the result of the estimation, we set
The objective function (16) can be rewritten as
which is a quadratic form in unit vector . Let
The solution is the eigenvector of associated to the smallest eigenvalue.
If each point is perturbed by noise of with
the matrix is perturbed accordingly: , where is the unperturbed matrix. If , then the estimate is statistically unbiased; otherwise, it is statistically biased, because following the perturbation theorem the bias of , i.e., .
Let , then . We have
If we carry out the Taylor development and ignore quantities of order higher than 2, it can be shown that the expectation of is given by
It is clear that if we define
then is unbiased, i.e., , and hence the unit eigenvector of associated to the smallest eigenvalue is an unbiased estimate of the exact solution .
Ideally, the constant c should be chosen so that , but this is impossible unless image noise characteristics are known. On the other hand, if , we have
because takes its absolute minimum 0 for the exact solution in the absence of noise. This suggests that we require that at each iteration. If for the current c and , , we can update c by such that
That is,
To summarize, the renormalization procedure can be described as:
associated to the smallest eigenvalue, which is denoted by .
and recompute using the new .
Remark 1: This implementation is different from that described in the paper of Kanatani [9]. This is because in his implementation, he uses the N-vectors to represent the 2-D points. In the derivation of the bias, he assumes that the perturbation in each N-vector, i.e., in his notations, has the same magnitude . This is an unrealistic assumption. In fact, to the first order, thus . Hence, , where we assume the perturbation in image plane is the same for each point (with mean zero and standard deviation ).
Remark 2: This method is optimal only in the sense of unbiasness. Another criterion of optimality, namely the minimum variance of estimation, is not addressed in this method.