Triangulation (computer vision)

Triangulation (computer vision)

In computer vision triangulation refers to the process of determining a point in 3D space given its projections onto two, or more, images. In order to solve this problem it is necessary to know the parameters of the camera projection function from 3D to 2D for the cameras involved, in the simplest case represented by the camera matrices. Triangulation is sometimes also referred to as reconstruction.

The triangulation problem is in theory trivial, each point in an image corresponds to a line in 3D space such that all points on the line are projected to that first point in the image. If a pair of corresponding points in two, or more images, can be found it must be the case that they are the projection of a common 3D point x. The set of lines generated by the image points must intersect at x and the algebraic formulation of the coordinates of x can be computed in a variety of ways, as is presented below.

In practice, however, the coordinates of image points cannot be measured with arbitrary accuracy. Instead, various types of noise, such as geometric noise from lens distortion or interest point detection error, lead to inaccuracies in the measured image coordinates. As a consequence, the lines generated by the corresponding image points do not always intersect in 3D space. The problem, then, is to find a 3D point which optimally fits the measured image points. In the literature there are multiple proposals for how to define optimality and how to find the optimal 3D point. Since they are based on different optimality criteria, the various methods produce different estimates of the 3D point x when noise is involved.

Introduction

In the following, it is assumed that triangulation is made on corresponding image points from two views generated by pinhole cameras. Generalization from these assumptions are discussed here.

The image to the left illustrates the epipolar geometry of a pair of stereo cameras of pinhole model. A point x in 3D space is projected onto the respective image plane along a line (green) which goes through the camera's focal point, mathbf{O}_{1} and mathbf{O}_{2} , resulting in the two corresponding image points mathbf{y}_{1} and mathbf{y}_{2} . If mathbf{y}_{1} and mathbf{y}_{2} is given and the geometry of the two cameras are known, the two projection lines can be determined and it must be the case that they intersect at point x. Using basic linear algebra that intersection point can be determined in a straight-forward way.

The image to the right shows the real case. The position of the image points mathbf{y}_{1} and mathbf{y}_{2} cannot be measured exactly. The reason is a combination of factors such as

* Geometric distortion, for example lens distortion, which means that the 3D to 2D mapping of the camera deviates from the pinhole camera model. To some extent these errors can be compensated for, leaving a residual geometric error.
* A single ray of light from x is dispersed in the lens system of the cameras according to a point spread function. The recovery of the corresponding image point from measurements of the dispersed intensity function in the images gives errors.
* In digital camera the image intensity function is only measured in discrete sensor elements. Inexact interpolation of the discrete intensity function have to be used to recover the true one.
* The image points used for triangulation are often found using various types of feature extractors, for example of corners or interest points in general. There is an inherent localization error for any type of feature extraction based on neighborhood operations.

As a consequence, the measured image points are mathbf{y}'_{1} and mathbf{y}'_{2} instead of mathbf{y}_{1} and mathbf{y}_{2} . However, their projection lines (blue) do not have to intersect in 3D space or come close to x. In fact, these lines intersect if and only if mathbf{y}'_{1} and mathbf{y}'_{2} satisfy the epipolar constraint defined by the fundamental matrix. Given the measurement noise in mathbf{y}'_{1} and mathbf{y}'_{2} it is rather likely that the epipolar constraint is not satisfied and the projection lines do not intersect.

This observation leads to the problem which is solved in triangulation. Which 3D point xest is the best estimate of x given mathbf{y}'_{1} and mathbf{y}'_{2} and the geometry of the cameras? The answer is often found by defining an error measure which depends on xest and then minimize this error. In the following some of the various methods for computing xest presented in the literature are briefly described.

All triangulation methods produce xest = x in the case that mathbf{y}_{1} = mathbf{y}'_{1} and mathbf{y}_{2} = mathbf{y}'_{2} , that is, when the epipolar constraint is satisfied (except for singular points, see below). It is what happens when the constraint is not satisfied which differs between the methods.

Properties of triangulation methods

A triangulation method can be described in terms of a function au , such that

: mathbf{x} sim au(mathbf{y}'_{1}, mathbf{y}'_{2}, mathbf{C}_{1}, mathbf{C}_{2})

where mathbf{y}'_{1}, mathbf{y}'_{2} are the homogeneous coordinates of the detected image points and mathbf{C}_{1}, mathbf{C}_{2} are the camera matrices. x is the homogeneous representation of the resulting 3D point. The sim , sign implies that au , is only required to produce a vector which is equal to x up to a multiplication by a non-zero scalar since homogeneous vectors are involved.

Before looking at the specific methods, that is, specific functions au , , there are some general concepts related to the methods that need to be explained. Which triangulation method is chosen for a particular problem depends to some extent on these characteristics.

Singularities

Some of the methods fail to correctly compute an estimate of x if it lies in a certain subset of the 3D space, correspondig to some combination of mathbf{y}'_{1}, mathbf{y}'_{1}, mathbf{C}_{1}, mathbf{C}_{2} . A point in this subset is then a "singularity" of the triangulation method. The reason for the failure can be that some equation system to be solved is under-determined or that the projective representation of xest becomes the zero vector for the singular points, .

Invariance

It is in some applications desirable that the triangulation is independent of the coordinate system used to represent 3D points; if the triangulation problem is formulated in one coordinate system and then transformed into another the resulting estimate xest should transform in the same way. This property is commonly referred to as "invariance". Not every triangulation method assures invariance, at least not for general types of coordinate transformations.

For a homogeneous representation of 3D coordinates, the most general transformation is a projective transformation, represented by a 4 imes 4 matrix mathbf{T} . If the homogeneous coordinates are transformed according to

: mathbf{ar x} sim mathbf{T} , mathbf{x}

then the camera matrices must transform as

: mathbf{ar C}_{k} sim mathbf{C}_{k} , mathbf{T}^{-1}

to produce the same homogeneous image coordinates

: mathbf{y}_{k} sim mathbf{ar C}_{k} , mathbf{ar x} = mathbf{C}_{k} , mathbf{x}

If the triangulation function au is invariant to mathbf{T} then the following relation must be valid

: mathbf{ar x}_{ m est} sim mathbf{T} , mathbf{x}_{ m est}

from which follows that

: au(mathbf{y}'_{1}, mathbf{y}'_{2}, mathbf{C}_{1}, mathbf{C}_{2}) sim mathbf{T}^{-1} , au(mathbf{y}'_{1}, mathbf{y}'_{2}, mathbf{C}_{1} , mathbf{T}^{-1}, mathbf{C}_{2} , mathbf{T}^{-1}), for all mathbf{y}'_{1}, mathbf{y}'_{2}

For each triangulation method, it can be determined if this last relation is valid. If it is, it may be satisfied only for a subset of the projective transformations, for example, rigid or affine transformations.

Computational complexity

The function au is only an abstract representation of a computation which, in practice, may be relatively complex. Some methods result in a au which is a closed-form continuous function while others need to be decomposed into a series of computational steps involving, for example, SVD or finding the roots of a polynomial. Yet another class of methods results in au which must rely on iterative estimation of some parameters. This means that both the computation time and the complexity of the operations involved may vary between the different methods.

Some triangulation methods found in the literature

Mid-point method

Each of the two image points mathbf{y}'_{1} and mathbf{y}'_{2} has a corresponding projection line (blue in the right image above), here denoted as mathbf{L}'_{1} and mathbf{L}'_{2} , which can be determined given the camera matrices mathbf{C}_{1}, mathbf{C}_{2} . Let d, be a distance function between a 3D line and a 3D point such that

: d(mathbf{L}, mathbf{x}) = the Euclidean distance between mathbf{L} and mathbf{x} .

The mid-point method finds the point xest which minimizes

: d(mathbf{L}'_{1}, mathbf{x})^{2} + d(mathbf{L}'_{2}, mathbf{x})^{2}

It turns out that xest lies exactly at the middle of the shortest line segment which joins the two projection lines.

Direct linear transformation

Via the essential matrix

Optimal triangulation

References

* cite book
author=Richard Hartley and Andrew Zisserman
title=Multiple View Geometry in computer vision
publisher=Cambridge University Press
year=2003
id=ISBN 978-0-521-54051-3


Wikimedia Foundation. 2010.

Игры ⚽ Нужен реферат?

Look at other dictionaries:

  • Triangulation (disambiguation) — Triangulation refers to measurement by using triangles. The term triangulation may also refer to:Mathematics and computer science*Subdivisions of spaces into triangles or higher dimensional simplices: **triangulation (advanced geometry), division …   Wikipedia

  • Scanner 3D — Traduction à relire 3D scanner → Scanner 3D …   Wikipédia en Français

  • 3D scanner — A 3D scanner is a device that analyzes a real world object or environment to collect data on its shape and possibly its appearance (i.e. color). The collected data can then be used to construct digital, three dimensional models useful for a wide… …   Wikipedia

  • Parallax — For other uses, see Parallax (disambiguation). A simplified illustration of the parallax of an object against a distant background due to a perspective shift. When viewed from Viewpoint A , the object appears to be in front of the blue square.… …   Wikipedia

  • Alhazen — For the Moon crater, see Alhazen (crater). For the asteroid, see 59239 Alhazen. Alhazen Alhazen (Ibn al Haytham) …   Wikipedia

  • Laser rangefinder — A laser rangefinder is a device which uses a laser beam in order to determine the distance to a reflective object. The most common form of laser rangefinder operates on the time of flight principle by sending a laser pulse in a narrow beam… …   Wikipedia

  • Range imaging — is the name for a collection of techniques which are used to produce a 2D image showing the distance to points in a scene from a specific point, normally associated with some type of sensor device.The resulting image, the range image , has pixel… …   Wikipedia

  • Marr Prize — The Marr Prize is a prestigious award in computer vision given by the committee of the International Conference on Computer Vision. Named after David Marr, the Marr Prize is considered one of the top honors for a computer vision researcher.… …   Wikipedia

  • Epipolar geometry — refers to the geometry of stereo vision. When two cameras view a 3D scene from two distinct positions, there are a number of geometric relations between the 3D points and their projections onto the 2D images that lead to constraints between the… …   Wikipedia

  • Image rectification — is a transformation process used to project multiple images onto a common image surface. It is used to correct a distorted image into a standard coordinate system. *It is used in computer stereo vision to simplify the problem of finding matching… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”