Digital Image Processing Unit 2 ppt
Digital Image Processing Unit 2 ppt
CS3EA14
Swati Vaidya
Unit II
Imaging Geometry, Digital Geometry, Image Acquisition Systems, Different types of digital images Introduction to
Fourier Transform and DFT, Properties of 2D Fourier Transform, FFT, Separable Image Transforms, Walsh –
Hadamard, Discrete Cosine Transform, Haar, Slant – Karhunen – Loeve transforms.
Imaging Geometry
• Imaging geometry in digital image fundamentals refers to the study of relationship
between the image plane and the object being imaged. It involves understanding
the position, orientation, and size of the object in the image, as well as the imaging
system parameters such as focal length and sensor size.
• Imaging geometry plays a crucial role in determining the quality of the digital
image. If the imaging geometry is not properly calibrated or aligned, it can lead to
distortions, blurring or misalignment in the image. This can result in a loss of
detail, reduced accuracy in measurements and overall degradation of image quality.
• The Key parameters of imaging geometry include focal length, sensor size, image
resolution, field of view, and camera position. Imaging geometry is a widely used
in various applications of digital Image Processing. some common applications
include image rectification, and registration, 3D reconstruction, object recognition
and tracking, augmented reality and computer vision.
Imaging Geometry includes-
1. Imaging geometry
2. Geometrical operations
3. Geometrical coordinates
4. Geometrical operation- Translation
5. Geometrical operation- Scaling
6. Geometrical operation- Rotation
7. Image formation process
Digital Coordinates-
- 2D-(x,y)
- 3D-(x,y,z)
Translation
• Translation displaces an image by a certain amount of pixels about the x and y axis.
This operation translates the image in a way such that every pixel in the image will
be shifted to a new position maintaining the shape and size of the image.
• Object is moving or Camera is moving from A to B, there occurs a displacement
(Xo, Yo, Zo) which can be represented as
In this, the response of each sensor is proportional to the integral of the light energy projected onto the
surface of the sensor. Noise reduction is achieved by letting the sensor integrate the input light signal over
minutes or ever hours.
Advantage: Since sensor array is 2D, a complete image can be obtained by focusing the energy pattern onto
the surface of the array.
A normal grayscale image contains 8 bits/pixel data, which has 256 different grey
levels. In medical images and astronomy, 12 or 16 bits/pixel images are used.
3. COLOUR IMAGES
Colour images are three band monochrome images in which, each band contains a different
color and the actual information is stored in the digital image. The color images contain gray
level information in each spectral band.
The images are represented as red, green and blue (RGB images). And each color image has 24
bits/pixel means 8 bits for each of the three color band(RGB).
8-BIT COLOR FORMAT
8-bit color is used for storing image information in a computer's memory or in a file of an
image. In this format, each pixel represents one 8 bit byte. It has 0-255 range of colors, in
which 0 is used for black, 255 for white and 127 for gray color. The 8-bit color format is also
known as a grayscale image. Initially, it was used by the UNIX operating system.
16-BIT COLOR FORMAT
The 16-bit color format is also known as high color format. It has 65,536 different color
shades. It is used in the system developed by Microsoft. The 16-bit color format is further
divided into three formats which are Red, Green, and Blue also known as RGB format.
In RGB format, there are 5 bits for R, 6 bits for G, and 5 bits for B. One additional bit is added
in green because in all the 3 colors green color is soothing to eyes.
24-BIT COLOR FORMAT
The 24-bit color format is also known as the true color format. The 24-bit color format is also
distributed in Red, Green, and Blue. As 24 can be equally divided on 8, so it is distributed equally
between 3 different colors like 8 bits for R, 8 bits for G and 8 bits for B.
What is Transform? Why it is required?
LINEARITY:
Addition of two functions corresponding to the addition of the two frequency spectrum is called the
linearity. If we multiply a function by a constant, the Fourier transform of the resultant function is
multiplied by the same constant. The Fourier transform of sum of two or more functions is the sum of the
Fourier transforms of the functions.
Case I.
If h(x) -> H(f) then ah(x) -> aH(f)
Case II.
If h(x) -> H(f) and g(x) -> G(f) then h(x)+g(x) -> H(f)+G(f)
SCALING:
Scaling is the method that is used to the change the range of the independent variables or features of data.
If we stretch a function by the factor in the time domain then squeeze the Fourier transform by the same
factor in the frequency domain.
If f(t) -> F(w) then f(at) -> (1/|a|)F(w/a)
DIFFERENTIATION:
Differentiating function with respect to time yields to the constant multiple of the initial function.
If f(t) -> F(w) then f'(t) -> jwF(w)
CONVOLUTION:
It includes the multiplication of two functions. The Fourier transform of a convolution of two functions is the
point-wise product of their respective Fourier transforms.
If f(t) -> F(w) and g(t) -> G(w)
then f(t)*g(t) -> F(w)*G(w)
FREQUENCY SHIFT:
Frequency is shifted according to the co-ordinates. There is a duality between the time and frequency domains and
frequency shift affects the time shift.
If f(t) -> F(w) then f(t)exp[jw't] -> F(w-w')
TIME SHIFT:
The time variable shift also effects the frequency function. The time shifting property concludes that a linear
displacement in time corresponds to a linear phase factor in the frequency domain.
If f(t) -> F(w) then f(t-t') -> F(w)exp[-jwt']
FAST FOURIER TRANSFORM
It is an algorithm which plays a very important role in the computation of the Discrete Fourier
Transform of a sequence. It converts a space or time signal to signal of the frequency domain.
The DFT signal is generated by the distribution of value sequences to different frequency component.
Working directly to convert on Fourier transform is computationally too expensive. So, Fast Fourier
transform is used as it rapidly computes by factorizing the DFT matrix as the product of sparse factors.
As a result, it reduces the DFT computation complexity from O(n2) to O(N log N). And this is a huge
difference when working on a large dataset. Also, FFT algorithms are very accurate as compared to the
DFT definition directly, in the presence of round-off error.
SEPARABLE IMAGE TRANSFORMS
where 𝐻1(𝑥)H 1(x) is the filter applied along the rows, and 𝐻2( 𝑦)H 2(y) is the filter applied along the
columns.
•Signal Processing: Used for edge detection and feature extraction in signals.
1. Slant Transform
• The Slant Transform is a simple and efficient orthogonal transform
primarily used for image processing. It is designed to approximate the
Karhunen–Loève Transform (KLT) for signals with linear trends or
slopes. The Slant Transform captures the overall linear structure of an
image or signal, making it useful for processing signals with sharp edges
or rapid changes.
MATHEMATICAL REPRESENTATION: 75
KEY FEATURES OF SLANT – KARHUNEN –
LOEVE TRANSFORMS 76
•Handles slants: Good for detecting linear structures, such as edges in images.
• The Slant Transform matrix S is designed with specific rows that capture
constant signals and others that capture linear or slanted patterns. The matrix
is of size N×N, where N is the number of data points in the signal.
For a 1D signal x, the transform is:
X = S⋅x
where S is the Slant Transform matrix, and x is the signal vector.
For 2D (images), the Slant Transform is applied in two dimensions:
Karhunen–Loève Transform
Feature Slant Transform
(KLT)
Computationally expensive
Efficiency Computationally efficient (requires eigen-
decomposition)
Optimal decorrelation and
Optimality Sub-optimal compared to KLT
energy compaction
Data compression,
Image processing, edge
Applications dimensionality reduction,
detection, simple compression
feature extraction
More complex to compute
Complexity Simple to implement
(especially for large datasets)
Data-dependent (requires
Fixed transform matrix (not
Data-Dependence covariance matrix
data-dependent)
computation)
THANK YOU!....................