HOME

TheInfoList



OR:

Image rectification is a transformation process used to project images onto a common image plane. This process has several degrees of freedom and there are many strategies for transforming images to the common plane. Image rectification is used in
computer stereo vision Computer stereo vision is the extraction of 3D information from digital images, such as those obtained by a CCD camera. By comparing information about a scene from two vantage points, 3D information can be extracted by examining the relative positi ...
to simplify the problem of finding matching points between images (i.e. the
correspondence problem The correspondence problem refers to the problem of ascertaining which parts of one image correspond to which parts of another image, where differences are due to movement of the camera, the elapse of time, and/or movement of objects in the photo ...
), and in geographic information systems to merge images taken from multiple perspectives into a common map coordinate system.


In computer vision

Computer stereo vision Computer stereo vision is the extraction of 3D information from digital images, such as those obtained by a CCD camera. By comparing information about a scene from two vantage points, 3D information can be extracted by examining the relative positi ...
takes two or more images with known relative camera positions that show an object from different viewpoints. For each pixel it then determines the corresponding scene point's depth (i.e. distance from the camera) by first finding matching pixels (i.e. pixels showing the same scene point) in the other image(s) and then applying triangulation to the found matches to determine their depth. Finding matches in stereo vision is restricted by
epipolar geometry Epipolar geometry is the geometry of stereo vision. When two cameras view a 3D scene from two distinct positions, there are a number of geometric relations between the 3D points and their projections onto the 2D images that lead to constraints b ...
: Each pixel's match in another image can only be found on a line called the epipolar line. If two images are coplanar, i.e. they were taken such that the right camera is only offset horizontally compared to the left camera (not being moved towards the object or rotated), then each pixel's epipolar line is horizontal and at the same vertical position as that pixel. However, in general settings (the camera does move towards the object or rotate) the epipolar lines are slanted. Image rectification warps both images such that they appear as if they have been taken with only a horizontal displacement and as a consequence all epipolar lines are horizontal, which slightly simplifies the stereo matching process. Note however, that rectification does not fundamentally change the stereo matching process: It searches on lines, slanted ones before and horizontal ones after rectification. Image rectification is also an equivalent (and more often used) alternative to perfect camera coplanarity. Even with high-precision equipment, image rectification is usually performed because it may be impractical to maintain perfect coplanarity between cameras. Image rectification can only be performed with two images at a time and simultaneous rectification of more than two images is generally impossible.


Transformation

If the images to be rectified are taken from camera pairs without geometric
distortion In signal processing, distortion is the alteration of the original shape (or other characteristic) of a signal. In communications and electronics it means the alteration of the waveform of an information-bearing signal, such as an audio signa ...
, this calculation can easily be made with a
linear transformation In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pre ...
. X & Y rotation puts the images on the same plane, scaling makes the image frames be the same size and Z rotation & skew adjustments make the image pixel rows directly line up. The rigid alignment of the cameras needs to be known (by calibration) and the calibration coefficients are used by the transform. In performing the transform, if the cameras themselves are calibrated for internal parameters, an
essential matrix In computer vision, the essential matrix is a 3 \times 3 matrix, \mathbf that relates corresponding points in stereo images assuming that the cameras satisfy the pinhole camera model. Function More specifically, if \mathbf and \mathbf' ...
provides the relationship between the cameras. The more general case (without camera calibration) is represented by the fundamental matrix. If the fundamental matrix is not known, it is necessary to find preliminary point correspondences between stereo images to facilitate its extraction.


Algorithms

There are three main categories for image rectification algorithms: planar rectification, cylindrical rectification and polar rectification.


Implementation details

All rectified images satisfy the following two properties: * All epipolar lines are parallel to the horizontal axis. * Corresponding points have identical vertical coordinates. In order to transform the original image pair into a rectified image pair, it is necessary to find a
projective transformation In projective geometry, a homography is an isomorphism of projective spaces, induced by an isomorphism of the vector spaces from which the projective spaces derive. It is a bijection that maps lines to lines, and thus a collineation. In general, ...
''H''. Constraints are placed on ''H'' to satisfy the two properties above. For example, constraining the epipolar lines to be parallel with the horizontal axis means that epipoles must be mapped to the infinite point '' ,0,0sup>T'' in
homogeneous coordinates In mathematics, homogeneous coordinates or projective coordinates, introduced by August Ferdinand Möbius in his 1827 work , are a system of coordinates used in projective geometry, just as Cartesian coordinates are used in Euclidean geometry. ...
. Even with these constraints, ''H'' still has four degrees of freedom. It is also necessary to find a matching ''H' '' to rectify the second image of an image pair. Poor choices of ''H'' and ''H' '' can result in rectified images that are dramatically changed in scale or severely distorted. There are many different strategies for choosing a projective transform ''H'' for each image from all possible solutions. One advanced method is minimizing the disparity or least-square difference of corresponding points on the horizontal axis of the rectified image pair. Another method is separating ''H'' into a specialized projective transform, similarity transform, and shearing transform to minimize image distortion. One simple method is to rotate both images to look perpendicular to the line joining their collective optical centers, twist the optical axes so the horizontal axis of each image points in the direction of the other image's optical center, and finally scale the smaller image to match for line-to-line correspondence. This process is demonstrated in the following example.


Example

Our model for this example is based on a pair of images that observe a 3D point ''P'', which corresponds to ''p'' and ''p' '' in the pixel coordinates of each image. ''O'' and ''O' '' represent the optical centers of each camera, with known camera matrices M=K ~ 0/math> and M'=K' ~ T/math> (we assume the world origin is at the first camera). We will briefly outline and depict the results for a simple approach to find a ''H'' and ''H' '' projective transformation that rectify the image pair from the example scene. First, we compute the epipoles, ''e'' and ''e' '' in each image: : e=M \begin O' \\ 1 \end =M\begin -R^T T \\ 1 \end = K ~ 0begin -R^T T \\ 1 \end = -KR^T T : e'=M'\begin O \\ 1 \end = M'\begin 0 \\ 1 \end = K' ~Tbegin 0 \\ 1 \end = K'T Second, we find a projective transformation ''H1'' that rotates our first image to be parallel to the baseline connecting ''O'' and ''O' '' (row 2, column 1 of 2D image set). This rotation can be found by using the cross product between the original and the desired optical axes. Next, we find the projective transformation ''H2'' that takes the rotated image and twists it so that the horizontal axis aligns with the baseline. If calculated correctly, this second transformation should map the ''e'' to infinity on the x axis (row 3, column 1 of 2D image set). Finally, define H=H_2H_1 as the projective transformation for rectifying the first image. Third, through an equivalent operation, we can find ''H' '' to rectify the second image (column 2 of 2D image set). Note that ''H'1'' should rotate the second image's optical axis to be parallel with the transformed optical axis of the first image. One strategy is to pick a plane parallel to the line where the two original optical axes intersect to minimize distortion from the reprojection process. In this example, we simply define ''H' '' using the rotation matrix ''R'' and initial projective transformation ''H'' as H' = HR^T. Finally, we scale both images to the same approximate resolution and align the now horizontal epipoles for easier horizontal scanning for correspondences (row 4 of 2D image set). Note that it is possible to perform this and similar algorithms without having the camera parameter matrices ''M'' and ''M' ''. All that is required is a set of seven or more image to image correspondences to compute the fundamental matrices and epipoles.


In geographic information system

Image rectification in GIS converts images to a standard map coordinate system. This is done by matching ground control points (GCP) in the mapping system to points in the image. These GCPs calculate necessary image transforms. Primary difficulties in the process occur *when the accuracy of the map points are not well known *when the images lack clearly identifiable points to correspond to the maps. The maps that are used with rectified images are non-topographical. However, the images to be used may contain distortion from terrain. Image orthorectification additionally removes these effects. Image rectification is a standard feature available with GIS software packages.


See also

* Binocular disparity *
Correspondence problem The correspondence problem refers to the problem of ascertaining which parts of one image correspond to which parts of another image, where differences are due to movement of the camera, the elapse of time, and/or movement of objects in the photo ...
*
Epipolar geometry Epipolar geometry is the geometry of stereo vision. When two cameras view a 3D scene from two distinct positions, there are a number of geometric relations between the 3D points and their projections onto the 2D images that lead to constraints b ...
* Geographic information system *
Georeferencing Georeferencing means that the internal coordinate system of a map or aerial photo image can be related to a geographic coordinate system. The relevant coordinate transforms are typically stored within the image file ( GeoPDF and GeoTIFF are exam ...
*
Homography In projective geometry, a homography is an isomorphism of projective spaces, induced by an isomorphism of the vector spaces from which the projective spaces derive. It is a bijection that maps lines to lines, and thus a collineation. In general, ...
*
Image registration Image registration is the process of transforming different sets of data into one coordinate system. Data may be multiple photographs, data from different sensors, times, depths, or viewpoints. It is used in computer vision, medical imaging, milit ...
* Photogrammetry *
Point set registration In computer vision, pattern recognition, and robotics, point-set registration, also known as point-cloud registration or scan matching, is the process of finding a spatial transformation (''e.g.,'' scaling, rotation and translation) that aligns t ...
*
Rubbersheeting In cartography and geographic information systems, rubbersheeting is a form of coordinate transformation that warps a vector dataset to match a known geographic space. This is most commonly needed when a dataset has systematic positional error ...
*
Stereo camera A stereo camera is a type of camera with two or more lenses with a separate image sensor or film frame for each lens. This allows the camera to simulate human binocular vision, and therefore gives it the ability to capture three-dimensional ...
*
Stereo vision Stereopsis () is the component of depth perception retrieved through binocular vision. Stereopsis is not the only contributor to depth perception, but it is a major one. Binocular vision happens because each eye receives a different image becaus ...
*
Structure from motion Structure from motion (SfM) is a photogrammetric range imaging technique for estimating three-dimensional structures from two-dimensional image sequences that may be coupled with local motion signals. It is studied in the fields of computer visio ...


References


Further reading


Computing Rectifying Homographies for Stereo Vision
by
Charles Loop Charles is a masculine given name predominantly found in English and French speaking countries. It is from the French form ''Charles'' of the Proto-Germanic name (in runic alphabet) or ''*karilaz'' (in Latin alphabet), whose meaning was " ...
and Zhengyou Zhang (April 8, 1999) Microsoft Research
Computer Vision: Algorithms and Applications, Section 11.1.1 "Rectification"
by Richard Szeliski (September 3, 2010) Springer {{DEFAULTSORT:Image Rectification Geographic information systems Geometry in computer vision Image processing