Introduction to image registration

Tutorial on image registration
Image registration (IR) is a process of overlaying images (two or more) of the same scene taken at different times from different viewpoints, and/or by different sensors. The registration geometrically aligns two images.

Approaches:
The reviewed approaches are classified according to their nature
1. Area based
2. Feature based

Image registration procedures:
  • Feature detection
  • Feature matching
  • Mapping function design
  • Image transformation and re-sampling
Typically IR is required in multi spectral classification, environmental monitoring, change detection, image mosaing, weather forecasting,, creating super-resolution images, geographical information systems (GIS) etc (basically capturing image with remote sensing).
In this post we will look into following

  • Various aspects and problem of image registration.
  • Both, area based and feature based approaches to feature selection.
  • Review of existing algorithm for feature matching.
  • Methods for mapping function design.
Image registration methodology
IR widely used in remote sensing, medical imaging, computer vision. In general its application can be divided into four main groups.

Different viewpoints – (multi view analysis):
Images of same scene are acquired from different viewpoints. The aim is to gain larger 2D view or a 3D representation of the scanned scene.

Different times – (multi temporal analysis)
Images of same scene are acquired from different times often on regular basis, possibly under different condition. The aim is to find and evaluate changes in the scene which appeared between the consecutive image acquisitions

Different sensors – (multi modal analysis)
Images of same scene are acquired from different sensors. The aim is to integrate the information obtained from different source streams to gain more complex and detailed scene representation.

Scene to model registration
Images of a scene and a model of the scene are registered. The model can be computer representation of the scene (for instance maps).

Registration methods

Majority of registration method consists of following four steps.

  • Feature detection
  • Feature matching
  • Transform model estimation
  • Image re-sampling
Feature detection:
Salient and distinctive objects like closed boundary regions, edges, contours, line intersections, corners etc are manually or preferably automatically detected. For further processing these features can be represented by their point representative’s like center of gravity, line endings, distinctive points, which are called control points (CPs) in the literature.

Feature matching:
In this step, the correspondence between the features detected in the sensed images and those detected in reference images is established. Various feature descriptors and similarity measures along with spatial relationships among the features are sued for that purpose.

Transform model estimations:
The type and parameters of the so-called mapping functions, aligning the sensed image with the reference image, are estimated. The parameters of the mapping functions are computed by means of established feature correspondence.

Image re-sampling and transformation:
The scene image is transformed by means of mapping functions. Image values in non-integer coordinates are computed by the appropriate interpolation technique.

Implementation:
The implementation of each registration step has its typical problems
First, we have to decide what kind of features is appropriate for the given task. The features should be distinctive objects, which are frequently spread over the images and which are easily detectable. The detection should have good localization accuracy and should not be sensitive to the assumed image degradation.

Mapping function should be chosen according to the prior known information about the acquisition process and expected image degradation.

Feature detection: Feature based methods

Region features – Closed boundary regions of appropriate size
-The regions are often represented by their center of gravity which is invariant w.r.t rotation, scaling and skewing and stable under random noise and gray level variation.
- Region features are detected by means of segmentation methods. The accuracy of segmentation can significantly influence the resulting registration.

Line features:
The LF can be representations of general line segments, object contours, costal lines, and roads.
Standard edge detection methods, like canny detector or a detector based on the laplacian of Gaussian are employed for the line feature detection.

Point features:
The point features group consists of methods working with line intersection, road crossings centroids of water regions..etc high variances points, local curvature discontinuous detected using gobar wavelet, inflection points of curves. The core algorithms of feature detectors in most cases follow the definitions of the ‘point’ as line intersection, centroids closed-boundary region or local modules maxima of the wavelet transform.

Feature based methods is recommended if the images contain enough distinctive and easily detectable objects.

Feature matching:
The detected features in the reference and sensed images can be matched by means of the image intensity values in their close neighborhood the feature special distribution or the feature symbolic description. Some methods, while looking for the feature correspondence, simultaneously estimate the parameters of mapping function and merge the feature matching and transform mode estimation registration steps.

Area based methods:
Some time called correlation-like methods or template matching. These methods deal with the images without attempting to detect salient objects. Classical area based methods like cross correlation exploit for matching directly image intensity, without any structural analysis consequently they are sensitive to the intensity changes introduced for instance by none varying illumination.

Correlation like methods
The classical representative of the area based method is normalized CC and its modification.

Fourier methods:
If an acceleration of the computation speed is needed or if the images were acquired under varying conditions or they are corrupted by frequency dependent noise, then Fourier methods is preferred rather than the correlation like methods. They exploit representation of the image in the frequency domain. The phase correlation methods based on the Fourier shift theorem and were originally proposed for the registration of translated images.

It computed the cross power spectrum of the sense and reference image and looks for the location of the peak in its inverse. The method shows strong robustness against the correlated and frequency dependent noise and non-uniform time varying illumination disturbance.

Mutual information methods:
The MI methods are the last group of the area based methods to be reviewed. They have appeared recently and represent the leading tech in multimodal registration. Registration of multimodal images is the difficult tasks, but often necessary to solve, especially in medical imaging. The MI, originating from the information theory, is a measure of statistical dependency between two data sets and it is particularly suitable for registration of images from different modalisation. MI was maximized using the gradient descent optimization method.

Feature based methods:
We assume that two sets of features in the reference and sensed images represented by the CP’s have been detected.
Methods using spatial registration:
Points method based primarily on the spatial relation among the features are usually applied if detected features are ambiguous or if their neighborhood are locally distorted.
Few more methods: Methods using invariant descriptions, relax methods, pyramid and wavelets.

Transform model estimation:
After the feature correspondence has been established the mapping function is constructed . It should transform the sensed image to overlay it over the reference one. The correspondence of the CP’s from the sensed and reference image together with the fact that the corresponding CP Paris should be as close as possible after the sense image transformation are employed in the mapping function design.

Mapping methods:
  • Global Mapping model
  • Local mapping models
  • Elastic registration

Image file formats and file hedaers

In this post we will discuss about various file formats used for representing Image data. Firstly, we will start with basic understanding of types of file formats and brief details about each format. Later on we will discuss the file header details on each of the image file formats.
Image File formats
1. BMP
2. GIF
3. PNG
4. TIFF
5. JPEG
What each format stands for and its application

BMP:

BMP is standard windows image format on DOS and windows compatible systems. BMP format supports RGB, Indexed Color, Grayscale, and bitmap color modes.BMP images are generally written bottom to top, however, you can select the Flip Row Order option to write them from top to bottom

GIF:

Graphics Interchange Format (GIF) is the file format commonly used to display indexed-color graphics and images in hypertext mark-up language (HTML) documents over the World Wide Web and other online services. GIF is an LZW-compressed format designed to minimize file size and electronic transfer time. GIF format preserves transparency in indexed-color images; however, it does not support alpha channels.

PNG:

Developed as a patent-free alternative to GIF, Portable Network Graphics (PNG) format is used for lossless compression and for display of images on the World Wide Web. Unlike GIF, PNG supports 24-bit images and produces background transparency without jagged edges; however, some Web browsers do not support PNG images. PNG format supports RGB, indexed-color, gray scale, and Bitmap-mode images without alpha channels. PNG preserves transparency in gray scale and RGB images.

TIFF:

Tagged-Image File Format (TIFF) is used to exchange files between applications and computer platforms. TIFF is a flexible bitmap image format supported by virtually all paint, image-editing, and page-layout applications. Also, virtually all desktop scanners can produce TIFF images. TIFF format supports CMYK, RGB, Lab, indexed-color, and gray scale images with alpha channels and Bitmap-mode images without alpha channels. Photoshop can save layers in a TIFF file; however, if you open the file in another application, only the flattened image is visible. Photoshop can also save annotations, transparency, and multiresolution pyramid data in TIFF format.

JPEG:

Joint Photographic Experts Group (JPEG) format is commonly used to display photographs and other continuous-tone images in hypertext markup language (HTML) documents over the World Wide Web and other online services. JPEG format supports CMYK, RGB, and Grayscale color modes, and does not support alpha channels. Unlike GIF format, JPEG retains all color information in an RGB image but compresses file size by selectively discarding data. A JPEG image is automatically decompressed when opened. A higher level of compression results in lower image quality, and a lower level of compression results in better image quality. In most cases, the Maximum quality option produces a result indistinguishable from the original.


Day2 we will discuss about each of the file headers discussed above


Fixed math - Fixed point multiplication

Fixed point Multiplication:

For fixed point multiplication, two operands need not be in same Q-format.

QIResult = QIA + QIB.
QFResult = QFA + QFB.


For example if two operands are in same format (QI) the resultant value after multiplication will be 2 X QI (twice of QI).

Consider two fixed point values which are in Q2.6 and Q1.7 format. The integer part of the product will be sum of integer bits taken by each operand, fractional part will be sum of fractional bits of each operand.

Therefore => Q2.6 X Q1.7
=> Q3.13

Word length:

WL required to store the product will be the sum of the WL’s of each operand. In the above example WL of each operand taken as 8-bit so the resultant will be twice the WL of any of the operands WL i.e. WLR = 16-bit

Saturation:

If two operands are of B-bits the resultant product will be of (2 x B) bits. If you want to store (2 x B) bits in X bits data type, you need to saturate the result and then store it in the given data type.

Examples to come

Related Posts

Twitter Updates

Random Posts

share this post
Bookmark and Share
| More
Share/Save/Bookmark Share