Ellipse Detection/Localization

In this article I’ll discuss multiple ways to localize an ellipse in an image.

“DUAL CONIC” method

This method is from Hebert09. I think it’s akin to the “opencv” checker localization algorithm in that it’s a linear algorithm that operates on the image gradients.

Anyway, to understand this method, you need to understand what a conic section is. A conic section is a curve obtained as the intersection of the surface of a cone with a plane. The possible conic sections are a hyperbola, parabola, and ellipse. It turns out that a conic can be represented as a matrix:

 [Aq]=  \begin{bmatrix} A & B/2 & D/2 \\ B/2 & C & E/2 \\ D/2 & E/2 & F \end{bmatrix}

and points, represented in homogeneous coordinates as \vec{x} = [x\  y\  z]^T , lie on the conic if:

\vec{x}^T [Aq]\vec{x} = 0

Now, it turns out there’s a “dual” to this in that a line, represented in homogeneous coordinates as \vec{l} = [a\ b\ c]^T , is tangent to the conic if:

\vec{l}^T [Aq]^{*}\vec{l} = 0

Where [Aq]^{*} = [Aq]^{-1}. [Aq]^{*} is known as the “dual conic”.

Proving this is a little mathy, so I’ll just leave this as a “black box” for now, but there are plenty of resources out there that explain why this is the case. Anyway, Hebert09 made the observation that lines orthogonal to the image gradients should be tangent to the ellipse, as shown below:

So these lines will satisfy the dual conic constraint. In addition to this, the dual conic constraint is linear, so it can be solved with a linear solver.

The first thing to note is the homogeneous form for the point slope formula is:

\vec{l}^T = \begin{bmatrix}m & -1 & y-mx\end{bmatrix}

Where m is the slope and x and y are a point on the line. The slope of the gradient is \nabla I_y(\vec{p})/\nabla I_x(\vec{p}) . The orthogonal slope is -\nabla I_x(\vec{p})/\nabla I_y(\vec{p}) . We then obtain:

\vec{l}^T = \begin{bmatrix}-\nabla I_x(\vec{p})/\nabla I_y(\vec{p}) & -1 & y+(\nabla I_x(\vec{p})/\nabla I_y(\vec{p}))x\end{bmatrix}

If you multiple through by -\nabla I_y(\vec{p}), we arrive at:

\vec{l}^T = \begin{bmatrix}\nabla I_x(\vec{p}) & \nabla I_y(\vec{p}) & -(\nabla I_x(\vec{p})x + \nabla I_y(\vec{p})y)\end{bmatrix}

These lines must satisfy \vec{l}^T [Aq]^{*}\vec{l} = 0. If you multiply this out, you obtain:

A^{*}a^2 + B^{*}ab + C^{*}b^2 + D^{*}ac + E^{*}bc + F^{*}c^2 = 0

We can form the matrix equation:

\begin{bmatrix}a^2 & ab & b^2 & ac  & bc & c^2 \end{bmatrix} \begin{bmatrix}A^{*} \\ B ^{*}\\ C^{*} \\ D^{*} \\ E^{*} \\ F^{*} \end{bmatrix}= 0

From \vec{l}^T [Aq]^{*}\vec{l} = 0, it’s apparent that [Aq]^{*} can have any arbitrary scaling, so an additional constraint needs to be added. According to Hebert09, a constraint of F^{*} = 1 ensures the conic is an ellipse. If applied, this results in:

\begin{bmatrix}a^2 & ab & b^2 & ac & bc \end{bmatrix} \begin{bmatrix}A^{*} \\ B^{*} \\ C^{*} \\ D^{*} \\ E^{*} \end{bmatrix}= \begin{bmatrix}-c^2\end{bmatrix}

Now, we can use a linear solver to solve for \begin{bmatrix}A^{*} \ B^{*} \ C^{*}\ D^{*} \ E^{*} \end{bmatrix} and hence [Aq]^{*}. The original conic, [Aq], which is what we’re really after, can be recovered by taking the inverse of [Aq]^{*}.

I guess the last thing to do is to convert the conic to a more tractable form which usually includes the center point, minor and major axes, and the rotation of the major axis. For that, I used this stack overflow post.

If you’d like to see an example in action, you can download my camera calibration toolbox. The functions which implements this are alg.fit_conic(), alg.refine_ellipse_dualconic(), and alg.conic2ellipse() and examples can be found in the unit tests folder.

“Edges” method

I tried to make an analogous function to the “edges” method from the checker detection article. If we revisit the equation used for the checker method, it turns out that the first term of the equation is equivalently:

{h_1}e^{-h_2^2(distance\ function)^2}

Where:

  •  h_1 – magnitude of the gradient peak
  •  h_2 – variance (i.e. width) of the gaussian distribution

This “distance function” is easy to compute for lines. But, for ellipses, it’s quite difficult. After googling around for a bit, I couldn’t find any satisfactory solutions (or any non-iterative closed form solutions which are differentiable; if a solution does exist, please leave a comment). So, I decided to use an approximation to the distance function as shown below:

Computing distance this way is reasonable as long as the minor axis isn’t super small compared to the major axis (i.e. the ellipse is still “circle-ish”). You can compute this by starting out with the general equation for an ellipse:

((x-h)cos(\alpha)+(y-k)sin(\alpha))^2/a^2 + ((x-h)sin(\alpha)-(y-k)cos(\alpha))^2/b^2 = 1

I plugged this into matlab’s symbolic toolbox and then substituted (x-h) for rcos(\theta) and (y-k) for rsin(\theta). Then, I substituted \theta for  \arctan((y-k)/(x-h)). I then solve()‘d for the roots, r. If you select the positive root, r_+, it represents a function where any (x,y) coordinate will return the distance from the center of the ellipse to the edge of the ellipse on a line which goes through (x,y). To compute d, all you need is the distance from the center of the ellipse to the point, and then subtract r_+ from this value. This will be the d shown in the figure above, which, again, is an approximation to the distance function.

After computing the derivatives symbolically (and getting a pretty massive set of ugly equations), optimization is possible. Below is an example of the optimization process:

Note that the final optimized edge function actually matches the array gradient for an ellipse very well.

If you’d like to see an example in action, you can download my camera calibration toolbox. The function which implements it is alg.refine_ellipse_edges() and an example can be found in the unit tests folder.

“dot” method

I wanted to try to model an ellipse directly instead of using the image gradients like in the “dualconic” and “edges” methods. It turns out you can use a sigmoid-like function to do this. The specific function I needed was a step function convolved with a gaussian. It turns out this function exists and is called the “error function” (or erf). The function can be constructed similar to the “edges” method:

h_1/2(1+erf((distance\ function)/h_2)

Where:

  •  h_1 – magnitude of edge peak
  •  h_2 – “blur” of edge

I used the same distance function approximation in the “edges” method section and just plugged it into the equation above. Again, to make things easy, I computed the derivatives symbolically and then was able to use nonlinear optimization. Below is an example of the optimization process:

The optimized “dot” function matches the array magnitude pretty well. The downsides with this method are: you have to know beforehand if the dot is black and the background is white (or vice-versa) and it will be affected by non-uniform lighting. These issues aren’t a problem with the gradient based approaches. However, not having to compute the gradient also has it’s advantages, as sometimes computing gradients can be unstable if the image is resampled to a higher resolution.

If you’d like to see an example in action, you can download my camera calibration toolbox. The function which implements it is alg.refine_ellipse_dot() and an example can be found in the unit tests folder.

Gradient ascent with backtracking method

This method maximizes the sum of sampled points along an ellipse on the array gradient magnitude. For example:

The cost we want to maximize is:

 \underset{\vec{e}}{\arg\max}\ c(\vec{e}) = \sum_{}I(x(\vec{e}),y(\vec{e}))

Where I is the image gradient magnitude and \vec{e} is a parameterization of an ellipse. To maximize it, I’ll use gradient ascent with backtracking. Backtracking is a robust way to ensure the cost function increases at every iteration. In addition, it provides a bounded way to ensure that larger steps result in a larger increase in cost. This acts to ensure the nearest-ish local maximum is found, which may be good or bad depending on the application. For this ellipse fitting technique, it’s assumed a reasonable initial guess is provided, so we actually do want to fit to the nearest-ish local maximum.

The figure below gives a good intuition for backtracking:

The backtracking condition is to pick a smaller step if the following condition exists:

 c(\vec{e}+\Delta{\vec{e}}) < c(\vec{e}) + 1/2(\partial{c}/\partial{\vec{e}}) \Delta{\vec{e}}

From the figure above, that means if the cost computed at the next step is below the second dashed line, then the step size is decreased. The decrease is done by multiplying the previous step size by a \lambda parameter which is less than 1.

For our case, the step is selected to be equal to the gradient times a \lambda parameter:

 \Delta{\vec{e}} = \lambda(\partial{c}/\partial{\vec{e}})

So the condition becomes:

 c(\vec{e}+\lambda(\partial{c}/\partial{\vec{e}})) < c(\vec{e}) + 1/2\lambda(\partial{c}/\partial{\vec{e}})^2

To compute the gradient, \partial{c}/\partial{\vec{e}}, we first need to parameterize x(\vec{e}) and y(\vec{e}):

 x(\vec{e}) = a( cos(\alpha) cos(\theta)) - b(sin(\alpha) sin(\theta)) + h

 y(\vec{e}) = a(sin(\alpha) cos(\theta)) - b(cos(\alpha) sin(\theta)) + k

where  \vec{e} = [h\ k\ a\ b\ \alpha]. For \theta, I chose to use 100 sampling points. The gradient is then:

 \partial{c}/\partial{\vec{e}} = \sum_{}(\partial{I/\partial{x}) (\partial{x}/\partial{\vec{e}}) + (\partial{I/\partial{y}) (\partial{y}/\partial{\vec{e}})

\partial{I/\partial{x} and \partial{I/\partial{y} are computed from the gradient of the image gradient magnitude and \partial{x}/\partial{\vec{e}} and \partial{y}/\partial{\vec{e}} are computed by taking the partial derivatives of the parameterizations of x(\vec{e}) and y(\vec{e}) in the equations above.

An example of the optimization is shown below:

This is kind of an extreme case, since the initial guess is poor and the ellipse has a large aspect ratio, but after quite a few iterations (70) it does appear to fit the array gradient magnitude pretty well, which demonstrates the robustness of the method.

If you’d like to see an example in action, you can download my camera calibration toolbox. The function which implements this is alg.fp_detect_blobs() and an example can be found in the unit tests folder.

Leave a Reply

Your email address will not be published.