0% found this document useful (0 votes)
42 views65 pages

AD8703 BCV Unit I 2023

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
42 views65 pages

AD8703 BCV Unit I 2023

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 65

Please read this disclaimer before proceeding:

This document is confidential and intended solely for the educational purpose of
RMK Group of Educational Institutions. If you have received this document
through email in error, please notify the system manager. This document
contains proprietary information and is intended only to the respective group /
learning community as intended. If you are not the addressee you should not
disseminate, distribute or copy through e-mail. Please notify the sender
immediately by e-mail if you have received this document by mistake and delete
this document from your system. If you are not the intended recipient you are
notified that disclosing, copying, distributing or taking any action in reliance on
the contents of this information is strictly prohibited.
AD8703
BASICS OF COMPUTER
VISION
UNIT I
Department: AI&DS

Batch/Year : 2020 - 2024 /IV

Created by : Dr. V. Seethalakshmi

Date : 26.07.2023
Table of Contents

S
CONTENTS PAGE NO
NO

1 Contents 1

2 Course Objectives 6
8
3 Pre Requisites (Course Names with Code)

4 Syllabus (With Subject Code, Name, LTPC details) 10

5 Course Outcomes 12

6 CO- PO/PSO Mapping 14

7 Lecture Plan 16

8 Activity Based Learning 18

9 Lecture Notes 20

Lecture Slides 37

Lecture Videos 39

10 Assignments 41

11 Part A (Q & A) 43

12 Part B Qs 47

13 Supportive Online Certification Courses 49

Real time Applications in day to day life and to 51


14
Industry

15 Contents Beyond the Syllabus 53

16 Assessment Schedule 55

17 Prescribed Text Books & Reference Books 57

18 Mini Project Suggestions 59


Course Objectives
AD8703 BASICS OF COMPUTER VISION

COURSE OBJECTIVES
To review image processing techniques for computer vision.
To understand various features and recognition techniques
To learn about histogram and binary vision
Apply three-dimensional image analysis techniques
Study real world applications of computer vision algorithms
Prerequisite
CS 404 OPERATING SYSTEM

PREREQUISITE
Syllabus
AD8703 -BASICS OF COMPUTER VISION

SYLLABUS 3003

UNIT I INTRODUCTION

Image Processing, Computer Vision ,What is Computer Vision - Low-level, Mid-level,


High-level Fundamentals of Image Formation, Transformation: Orthogonal, Euclidean,
Affine, Projective, Fourier Transform, Convolution and Filtering, Image Enhancement,
Restoration, Histogram Processing.

UNIT II FEATURE EXTRACTION AND FEATURE SEGMENTATION

Feature Extraction -Edges - Canny, LOG, DOG; Line detectors (Hough Transform),
Corners -Harris and Hessian Affine, Orientation Histogram, SIFT, SURF, HOG, GLOH,
Scale-Space Analysis- Image Pyramids and Gaussian derivative filters, Gabor Filters
and DWT. Image Segmentation -Region Growing, Edge Based approaches to
segmentation, Graph-Cut, Mean-Shift, MRFs, Texture Segmentation.

UNIT III IMAGES, HISTOGRAMS, BINARY VISION

Simple pinhole camera model – Sampling – Quantisation – Colour images – Noise –


Smoothing –1D and 3D histograms - Histogram/Image Equalisation - Histogram
Comparison - Back-projection - k-means Clustering – Thresholding - Threshold
Detection Methods - Variations on Thresholding -Mathematical Morphology –
Connectivity.

UNIT IV 3D VISION AND MOTION

Methods for 3D vision – projection schemes – shape from shading – photometric


stereo – shape from texture – shape from focus – active range finding – surface
representations – point-based representation – volumetric representations – 3D object
recognition – 3D reconstruction – introduction to motion – triangulation – bundle
adjustment – translational alignment – parametric motion–spline-based motion- optical
flow – layered motion.

UNIT V APPLICATIONS

Overview of Diverse Computer Vision Applications: Document Image Analysis,


Biometrics, Object Recognition, Tracking, Medical Image Analysis, Content-Based
Image Retrieval, Video Data Processing , Virtual Reality and Augmented Reality.
Course Outcomes
COURSE OUTCOMES

CO1: Recognise and describe how mathematical and scientific concepts are
applied in computer vision.

CO2: Identify and interpret appropriate sources of information relating to


computer vision.

CO3: Apply knowledge of computer vision to real life scenarios..

CO4: Reflect on the relevance of current and future computer vision


applications

CO5: Discuss principles of computer vision using appropriate language and


terminology. Implement various I/O and file management techniques.
CO – PO/ PSO Mapping
CO-PO MAPPING

PO’s/PSO’s
COs
PO PO PO PO PO PO PO PO PO PO PO PO PSO PSO PSO
1 2 3 4 5 6 7 8 9 10 11 12 1 2 3

CO1
3 2 2 - - - - - - - - - 2 - -
CO2
3 3 2 - - - - - - - - - 2 - -
CO3
2 2 1 - - - - - - - - - 2 - -
CO4
3 3 1 - - - - - - - - - 2 - -
CO5
3 3 1 - - - - - - - - - 3 1 -

1 – Low, 2 – Medium, 3 – Strong


Lecture Plan
LECTURE PLAN

No of Taxo Mode
Actual
S No Topics period Pertain nom of
Proposed date Lecture
s ing CO y delivery
Date
level
Introduction to
CO1
1 Image Processing, 1 7.08.2023 K1 Lecture
Computer Vision
- Low-level, Mid-
level, High-level CO1
2 1 9.08.2023 K2 Lecture
Fundamentals of
Image Formation
Transformation:
3 Orthogonal, 1 10.8.2023 CO1 K2 Lecture
Euclidean
Affine, Projective 10.08.202
4 1 CO1 K2 Lecture
3

16.08.202 Lectur
5 Fourier Transform 1 CO1 K2
3 e

Convolution and 16.08.202


6 1 CO1 K2 Lectur
Filtering 3
e
Image Enhancement
17.08.202 CO1 Lectur
7 1 K2
3 e

Restoration, Lectur
8 Histogram 1 17.08.202 CO1 K2 e
Processing. 3
Revision – Quiz 23.08.202 Lectur
9 Activity 1 3 CO1 K2 e
Activity Based Learning
ACTIVITY BASED LEARNING

Using an on-prem Linux CV Server deployment, you can also


use the UiPath Screen OCR engine with this activity
https://docs.uipath.com/activities/other/latest/ui-
automation/computer-vision-activities
Lecture Notes
1. Image processing:

Image processing is a fundamental component of computer vision,


focusing on the manipulation and analysis of digital images to extract
useful information, enhance image quality, and enable subsequent higher-
level vision tasks. It involves applying a series of operations and
algorithms to images to improve their visual appearance and facilitate the
understanding of their content. Some key aspects of image processing in
computer vision include:

Image Preprocessing: Image preprocessing is the initial step in image


processing, which involves various techniques to enhance the quality of
images and make them more suitable for analysis. Preprocessing
techniques may include resizing, denoising, color normalization, histogram
equalization, and contrast adjustment.

 Filtering: Filtering operations are used to remove noise from images


and smooth out sharp edges. Common filters include Gaussian filters,
median filters, and mean filters.

 Edge Detection: Edge detection is the process of identifying


boundaries or edges between different objects or regions in an image.
Edge detection algorithms highlight areas of significant intensity
changes, which often correspond to object boundaries or outlines.

 Corner Detection: Corner detection algorithms identify points in an


image where two or more edges meet. Corners are useful for tasks
like object tracking and image registration.

21
 Image Enhancement: Image enhancement techniques aim to
improve the visual quality of an image, making it easier for humans
and machines to perceive relevant details.

 Histogram Manipulation: Histogram operations, such as histogram


equalization and stretching, are used to adjust the distribution of pixel
intensity values in an image, thereby enhancing contrast and improving
visibility.

 Morphological Operations: Morphological operations, including


dilation and erosion, are used for analyzing and processing binary
images to manipulate the shapes and structures present.

 Color Space Conversion: Converting images between different color


spaces (e.g., RGB, HSV, LAB) can help extract specific information or
highlight particular image characteristics.

 Image Segmentation: Image segmentation divides an image into


meaningful regions or segments based on common properties such as
color, texture, or intensity. Segmentation is crucial for understanding
the different objects or regions present in an image.

 Feature Extraction: Feature extraction involves identifying and


extracting relevant patterns or characteristics from an image. These
features serve as input to higher-level vision tasks, such as object
recognition and classification.

 Geometric Transformations: Geometric transformations, such as


rotation, scaling, and translation, are used to modify the spatial
arrangement of objects within an image.
1.1 Introduction to Computer Vision:
Computer Vision is a fascinating field at the intersection of computer
science and artificial intelligence, which aims to enable machines to
interpret and understand visual information from the world around us. It
involves the development of algorithms and techniques that allow
computers to perceive, analyze, and make decisions based on digital
images or videos.
Importance of Computer Vision: Computer Vision plays a crucial role in
various applications, transforming the way we interact with technology
and the world. Some key applications include:
1. Image and Video Analysis: Object detection, recognition,
and tracking in images and videos, scene understanding, and image
segmentation.
2. Medical Imaging: Diagnosing diseases, identifying tumors,
and analyzing medical images like X-rays, MRIs, and CT scans.
3. Autonomous Vehicles: Enabling cars, drones, and robots to
perceive their surroundings, detect obstacles, and navigate safely.
4. Augmented Reality (AR): Integrating virtual objects into the
real world, enhancing user experiences in gaming, entertainment, and
other fields.
5. Robotics: Providing robots with the ability to interact with their
environment and perform tasks autonomously.

1.2 Core Concepts in Computer Vision:

1. Image Representation: Understanding how digital images are


represented in computers, typically as matrices or tensors of pixel
values.

2. Image Processing: Applying various techniques to manipulate


and enhance images, including filtering, smoothing, and edge detection.

3. Feature Extraction: Identifying relevant patterns or features in


an image that can be used for further analysis and recognition tasks.

23
4. Object Detection and Recognition: Locating and identifying specific
objects or patterns within an image or video.

5. Machine Learning and Deep Learning: Utilizing machine learning


algorithms, particularly deep neural networks, to learn patterns and make
predictions from visual data.

6. Convolutional Neural Networks (CNNs): Specialized deep learning


architectures designed for computer vision tasks, which have shown
remarkable success in image recognition.

7. Image Segmentation: Dividing an image into meaningful segments


or regions for detailed analysis.

8. Optical Character Recognition (OCR): Converting text within images


into machine-readable text.

1.3 Challenges in Computer Vision:

Despite significant advancements, computer vision still faces some


challenges:

1. Variability in Data: Images can have different lighting conditions,


viewpoints, and occlusions, making it challenging to create robust
algorithms.

2. Scale and Complexity: The sheer amount of visual data and the
complexity of real-world scenes require powerful and efficient algorithms.

3. Interpretation and Context: Understanding the context of visual


information, which is often intuitive for humans, remains challenging for
machines.

4. Semantic Gap: Bridging the gap between low-level pixel data and
high-level semantic understanding.
2
5

Computer vision tasks can be broadly categorized into three levels


based on the complexity and depth of understanding required:

 Low-level Computer Vision:

1. Focus: Low-level computer vision deals with basic image


processing and feature extraction tasks.

2. Objectives: The main objectives are to preprocess and enhance


images, extract simple features (e.g., edges, corners, textures), and
manipulate pixel-level information.

3. Applications: Low-level computer vision is used in tasks such as


image denoising, edge detection, corner detection, image
smoothing, and basic filtering operations.

 Mid-level Computer Vision:

1. Focus: Mid-level computer vision involves more advanced feature


extraction and grouping techniques.

2. Objectives: The main objectives are to extract higher-level


features (e.g., segments, regions) and group them into coherent
structures to facilitate understanding.

3. Applications: Mid-level computer vision is applied in tasks such as


image segmentation, object recognition (with the focus on individual
objects or regions), and optical flow estimation.
 High-level Computer Vision:
1. Focus: High-level computer vision focuses on advanced understanding and
interpretation of visual data.
2. Objectives: The main objectives are to recognize complex patterns, infer
relationships between objects, and understand the context of the visual
information.
3. Applications: High-level computer vision is used in tasks such as object
detection (identifying and localizing multiple objects in an image), scene
understanding, visual reasoning, image captioning, visual question answering
(VQA), and 3D scene reconstruction.

1.4 Image formation


Image formation refers to the process of capturing, creating, or rendering
visual representations of objects or scenes. It involves understanding the
fundamentals of how images are generated, recorded, and displayed. The
levels of image formation can be categorized as low-level, mid-level, and high-
level. Let's explore each level:
2
7

1.4.1 Low-level Image Formation:

Low-level image formation deals with basic concepts related to the


acquisition and representation of images. It focuses on the physical
aspects of image formation, such as light, color, and pixel-level operations.
Key topics at this level include:

a. Light and Optics:

Understanding the behavior of light, including reflection, refraction, and


the interaction of light with objects and surfaces. It involves concepts like
illumination, shading, and specular highlights.

b. Color Spaces: Different color models and spaces used to


represent and interpret colors in images, such as RGB (Red, Green, Blue),
CMYK (Cyan, Magenta, Yellow, Black), and HSL/HSV (Hue, Saturation,
Lightness/Value).

c. Image Sampling and Quantization: Techniques for capturing


and digitizing continuous real-world images into discrete digital
representations. This includes concepts like resolution, pixel sampling, and
quantization.

d. Image Filtering: Operations that modify images by altering their


spatial or frequency characteristics. Examples include blurring, sharpening,
noise reduction, and edge detection.

1. 4.2 Mid-level Image Formation:

Mid-level image formation focuses on extracting meaningful


information from images. It involves analyzing and interpreting image
content to extract features, objects, or structures. Key topics at this level
include:
2
8

a. Image Enhancement:

Techniques to improve image quality, visibility, and perception. This


includes contrast enhancement, histogram equalization, and adaptive
filtering.

b. Image Segmentation:

The process of partitioning an image into meaningful regions or


objects. It involves techniques like thresholding, region growing, edge-
based segmentation, and clustering.

c. Feature Extraction:

Identifying and describing distinctive image attributes or patterns, such


as corners, edges, textures, or color descriptors. This forms the basis for
higher-level analysis and recognition tasks.

d. Image Registration:

Aligning and matching multiple images or image sequences to establish


correspondences or create composite images. Applications include image
stitching, motion analysis, and medical image alignment.

1.4.3. High-level Image Formation:

High-level image formation involves advanced analysis and


interpretation of images, aiming to extract semantic information and
understand the content at a higher level. Key topics at this level include:

a. Object Recognition and Classification: Identifying and


categorizing objects or patterns in images using machine learning,
computer vision, and pattern recognition techniques. This includes object
detection, image classification, and image understanding.
2
9
b. Scene Understanding: Inferring higher-level meaning and context from
images, such as recognizing scenes, understanding relationships between
objects, and interpreting actions or events.

c. Image Synthesis and Rendering: Creating new images from scratch


or modifying existing images to generate realistic or artistic renderings. This
includes techniques like computer graphics, image-based rendering, and
generative models.

d. Image Retrieval and Indexing: Efficiently searching and retrieving


images from large databases based on their content or similarity. This
involves indexing techniques, feature matching, and content-based image
retrieval algorithms.

These levels of image formation represent a progression from the basic


physical aspects of image creation to the extraction of meaningful information
and interpretation at a higher semantic.

2. Transformation

Transformation refers to the process of mapping points or objects from one


coordinate system to another. In computer vision and image processing,
various types of transformations are used to manipulate and manipulate
images. Let's explore four common types of transformations: orthogonal,
Euclidean, affine, and projective.
3
0

2.1. Orthogonal Transformation:

An orthogonal transformation, also known as an isometric transformation,


preserves distances, angles, and parallel lines. It includes rotation and
reflection operations. Key characteristics of orthogonal transformations
include:

- Rotation: Rotates the image around a fixed point by a certain angle.

- Reflection: Flips the image along a line of symmetry, such as a vertical


or horizontal axis.

Orthogonal transformations do not change the shape or size of objects in


the image, only their orientation or position.

2.2 Euclidean Transformation:

Euclidean transformations are similar to orthogonal transformations, but


they also include translation operations. Euclidean transformations preserve
distances, angles, parallel lines, and the orientation of objects. Key
characteristics include:

- Rotation: Same as in orthogonal transformation.

- Translation: Shifts the image along the x and y axes by specific


distances.

Euclidean transformations can change the position and orientation of


objects but preserve their shape and size.
3
1
2. 3. Affine Transformation:

Affine transformations are more flexible and general than Euclidean


transformations. They include rotation, translation, scaling, and shearing
operations. Key characteristics include:

- Rotation: Same as in orthogonal and Euclidean transformations.

- Translation: Same as in Euclidean transformations.

- Scaling: Enlarges or shrinks the image by different factors along the x


and y axes.

- Shearing: Skews the image along the x or y axis.

Affine transformations can change the position, orientation, shape, and size
of objects but preserve parallel lines.

2.4. Projective Transformation:

Projective transformations, also known as perspective transformations,


are the most general type of transformation. They can represent more
complex distortions and non-planar mappings. Key characteristics include:

- Rotation: Same as in orthogonal, Euclidean, and affine transformations.

- Translation: Same as in Euclidean and affine transformations.

- Scaling: Same as in affine transformations.

- Shearing: Same as in affine transformations.

- Perspective Distortion: Represents the effect of viewing a three-


dimensional scene from a particular viewpoint.
Projective transformations can distort the shape, size, and perspective of
objects, including the transformation from a 3D scene to a 2D image.

These types of transformations provide a range of tools for manipulating


and manipulating images in computer vision and image processing
tasks. Each type has its own characteristics and applications depending
on the desired modifications to the image or scene.

3. FOURIER TRANSFORM

The Fourier transform is a mathematical operation that transforms a


function from the time domain to the frequency domain. It decomposes
a signal into its constituent frequencies, revealing the frequency content
of the signal. The transform is named after Jean-Baptiste Joseph
Fourier, a French mathematician who introduced the concept in the early
19th century.

The Fourier transform is widely used in various fields, including


mathematics, physics, engineering, signal processing, and data analysis.
It has applications in areas such as image processing, audio signal
processing, telecommunications, and quantum mechanics.

Mathematically, the Fourier transform of a continuous-time function f(t)


is defined as follows:

F(ω) = ∫[f(t) * exp(-jωt)] dt


Where F(ω) represents the frequency domain representation of the function, ω is
the angular frequency, j is the imaginary unit (√-1), and the integration is
performed over the entire time domain.

For a discrete-time function, the Fourier transform is defined as follows:

F(ω) = Σ[f[n] * exp(-jωn)]

where F(ω) represents the frequency domain representation of the function, f[n]
is the discrete function, ω is the angular frequency, and the summation is
performed over all discrete time instances.

The Fourier transform converts a function from a time representation to a


frequency representation, providing information about the amplitudes and
phases of the different frequency components present in the signal. The inverse
Fourier transform can also be applied to obtain the original time-domain signal
from its frequency-domain representation.

The Fast Fourier Transform (FFT) is an efficient algorithm for computing the
discrete Fourier transform (DFT) or its inverse. It allows the Fourier
transform to be computed quickly, especially for large datasets.

The Fourier transform has revolutionized many fields by enabling the analysis,
manipulation, and synthesis of signals in the frequency domain. It is an essential
tool in understanding and processing various types of signals and has
widespread applications in science, technology, and everyday life.
3
4
4. Convolution and Filtering

Convolution and Filtering are fundamental concepts in signal processing


and image processing. They are closely related and often used together to
process and analyze signals and images.

4.1 Convolution:

Convolution is a mathematical operation that combines two signals to


produce a third signal. In the context of signal processing, convolution
involves sliding a function (known as the kernel or filter) over another
function (known as the input signal) and calculating the integral of their
pointwise multiplication at each position. The result of convolution is a new
signal that represents the blending or mixing of the two input signals.

Mathematically, the convolution of two continuous signals f(t) and g(t) is


denoted as (f * g)(t) and is defined as:

(f * g)(t) = ∫[f(τ) * g(t - τ)] dτ

In the discrete domain, convolution is performed on discrete signals using


summation instead of integration. If we have two discrete signals f[n] and
g[n], the convolution of f[n] and g[n] is denoted as (f * g)[n] and is
defined as:

(f * g)[n] = Σ[f[k] * g[n - k]]

Convolution is a powerful tool in signal processing as it allows us to extract


information from signals by applying filters or performing operations such
as blurring, sharpening, edge detection, and noise reduction.
3
5

4.2 FILTERING

Filtering is the process of modifying or enhancing a signal by removing or


altering certain components of the signal. Filters are used to selectively
pass or attenuate different frequencies or spatial components of a signal.
Filtering is commonly employed to remove noise, extract specific features,
or enhance desired characteristics of a signal.

In the context of signal processing, filters are usually implemented using


convolution. The filter itself is represented by a kernel, which defines the
weights or coefficients used in the convolution operation. The filter kernel
determines how the input signal is transformed or modified.

There are various types of filters used in signal processing, including low-
pass filters, high-pass filters, band-pass filters, and notch filters. Each filter
has its own characteristics and is designed to address specific requirements
in signal processing applications.

In image processing, filtering is often used to enhance or modify images by


applying spatial filters. Spatial filters are convolution kernels that operate on
the pixel values of an image. They can be used for tasks such as smoothing
(blurring), sharpening, edge detection, noise reduction, and image
enhancement.

Overall, convolution and filtering are fundamental techniques in signal


processing and image processing that allow us to analyze, modify, and
extract useful information from signals and images.
5. Image enhancement

Image enhancement refers to a set of techniques and processes used to


improve the visual quality or interpretability of an image. The goal of image
enhancement is to highlight important details, reduce noise or artifacts,
improve contrast, and make the image more visually pleasing or suitable for
further analysis.

There are various methods and algorithms for image enhancement, and the
choice of technique depends on the specific requirements and characteristics
of the image. Here are some commonly used image enhancement
techniques:

5.1 Histogram Equalization: Histogram equalization is a technique that


redistributes the pixel intensities in an image to improve contrast. It stretches
the intensity range of the image to cover the full dynamic range, resulting in
a more balanced distribution of intensities.

5.2 Contrast Stretching: Contrast stretching is a simple technique that


expands the range of pixel intensities in an image. It maps the original
intensity values to a new range to enhance the contrast between different
image regions.
3
7
5.3 Spatial Filtering: Spatial filtering involves applying convolution filters
to an image to modify its pixel values. Filters can be designed to perform
operations such as blurring (smoothing), sharpening, edge detection, noise
reduction, and detail enhancement. Common spatial filters include Gaussian
filters, Laplacian filters, and Sobel filters.

5.4. Noise Removal: Image noise can degrade the quality of an image
and reduce its clarity. Various noise reduction techniques, such as median
filtering, Gaussian filtering, and Wiener filtering, can be employed to
suppress noise and enhance image details.

5.5. Image Fusion: Image fusion combines multiple images of the same
scene captured under different conditions or using different sensors to
create a single enhanced image. Fusion methods aim to extract the most
salient information from each input image and generate an output image
with improved quality and more comprehensive information.

5.6. Retinex Algorithms: Retinex-based algorithms simulate the human


visual system's ability to perceive and adjust for changes in illumination
conditions. They enhance image details by separating the illumination and
reflectance components of the image and adjusting them independently.

5.7. Super-Resolution: Super-resolution techniques aim to enhance the


resolution or level of detail in an image beyond its original quality. These
methods utilize information from multiple low-resolution images or employ
sophisticated algorithms to estimate and reconstruct high-resolution details.

These are just a few examples of image enhancement techniques. The


choice of method depends on the specific image characteristics, the desired
enhancement goals, and the application requirements. In practice, a
combination of different techniques may be used to achieve the desired
image enhancement results.
3
8
6. RESTORATION

Image restoration is a branch of image processing that focuses on


recovering or improving the quality of images that have been degraded or
corrupted by various factors such as noise, blur, compression artifacts,
motion blur, or other types of distortions. The goal of image restoration is
to estimate the original image as accurately as possible.

Image restoration techniques aim to reverse or compensate for the


degradation process and restore the lost or degraded image details. Here
are some commonly used image restoration methods:

6.1 Noise Reduction: Image denoising techniques are used to reduce the
effect of noise in an image. These methods aim to preserve image details
while suppressing the unwanted noise. Common denoising techniques
include median filtering, Gaussian filtering, wavelet denoising, and total
variation denoising.

6.2 Deblurring: Deblurring techniques are employed to reduce or remove


blur from images. Image blur can be caused by factors such as out-of-focus
capture, camera shake, or motion blur. Deblurring methods utilize
algorithms that estimate the blur kernel and then perform inverse filtering
or deconvolution to restore the sharpness and details in the image.

6.3. Super-Resolution:

Super-resolution techniques aim to enhance the resolution and level of


detail in images that are low-resolution or degraded. These methods utilize
algorithms that estimate the high-resolution details based on information
from multiple low-resolution images or through advanced interpolation and
reconstruction techniques.
3
9
6.4 Inpainting: Image inpainting techniques are used to fill in missing or
corrupted regions in an image. Inpainting algorithms can estimate the
missing information based on the surrounding image content and
structures. It is useful for restoring images with missing parts or removing
unwanted objects from images.

6.5 Compression Artifact Reduction: Compression artifacts can occur


when images are heavily compressed and result in blocky or blurry regions.
Image restoration techniques can be applied to reduce compression
artifacts and restore the original image quality. Methods such as deblocking
filters and post-processing algorithms can be used to mitigate the artifacts
introduced by compression.

6.6 Restoration with Deep Learning: Deep learning-based approaches


have shown promising results in image restoration tasks. Convolutional
neural networks (CNNs) can be trained to learn the mapping between
degraded and clean images, enabling them to restore image details and
reduce the effects of various degradations.

The choice of image restoration technique depends on the specific


degradation types, the available information about the degradation process,
and the desired restoration goals. In some cases, a combination of different
techniques may be used to achieve the best restoration results.
7. HISTOGRAM PROCESSING
Histogram processing also known as histogram manipulation or histogram
modification, is a technique used in image processing to modify the
distribution of pixel intensities in an image's histogram. The histogram of an
image represents the frequency of occurrence of different pixel intensities.
By adjusting the histogram, we can alter the contrast, brightness, and overall
appearance of an image. Histogram processing techniques are often used for
image enhancement, equalization, and normalization. Here are some
commonly used histogram processing operations:

7.1. Histogram Equalization: Histogram equalization is a technique that


redistributes the pixel intensities in an image to improve contrast. It stretches
the intensity range of the image to cover the full dynamic range. This is
achieved by mapping the original histogram to a more uniform histogram
distribution. Histogram equalization can enhance images with low contrast
and improve their visual appearance.

7.2. Contrast Stretching: Contrast stretching is a simple technique that


expands the range of pixel intensities in an image. It maps the original
intensity values to a new range to enhance the contrast between different
image regions. By increasing the dynamic range, the dark regions become
darker, and the bright regions become brighter, resulting in an image with
enhanced contrast.
7.3. Histogram Specification (Matching): Histogram specification is a
technique used to modify an image's histogram to match a desired histogram
or a reference histogram. It is often used to transfer the statistical properties
of one image to another. By matching histograms, we can adjust the image's
appearance to resemble a specific reference image or to achieve a desired
target histogram distribution.
7.4. Histogram Shifting: Histogram shifting involves modifying the
histogram by shifting its values to a new intensity range. This operation can be
used to adjust the brightness level of an image. By shifting the histogram, we
can make the image darker or brighter without significantly affecting the
contrast.

7.5. Histogram Smoothing: Histogram smoothing techniques are used to


reduce the noise or fluctuations in the histogram. This can be achieved by
applying filters or averaging operations to the histogram values. Smoothing
the histogram can help alleviate the effects of noise or artifacts and produce a
more stable representation of the image's intensity distribution.

Histogram processing operations can be applied globally to the entire image or


locally to specific regions of interest. The choice of histogram processing
technique depends on the specific image characteristics, the desired
enhancement goals, and the application requirements. Histogram processing is
a powerful tool for adjusting the tonal distribution of an image and improving
its visual quality.
Lecture Slides
Lecture Slides

Lecture Slides
Lecture Videos
Lecture Videos

Lecture Videos
Assignment
Assignment
 Feature Extraction and Matching Vision and Image Processing

 Provide an example of an affine transformation applied to an image

 Describe two types of filters commonly used in image processing and their
respective applications.

 Provide an example of histogram processing to improve the contrast of a


grayscale image.
Part A Q & A
4
9
PART -A

1. What is Computer Vision?


Computer Vision is a field that involves the processing and analysis of
digital images or videos to extract meaningful information and enable
computers to interpret and understand the visual world.

2.Define Low-level Vision.


Low-level Vision deals with basic image processing tasks, such as noise
reduction, edge detection, and simple image transformations like
scaling and rotation.

3. Give an example of an affine transformation.


An affine transformation includes operations like translation, rotation,
scaling (uniform or non-uniform), and shearing. An example would be
resizing an image.

4. What is the purpose of image segmentation?


Image segmentation is the process of dividing an image into
meaningful regions or objects. Its purpose is to separate different
objects or parts of an image for further analysis.

5. Explain the concept of the Fourier Transform.


The Fourier Transform is a mathematical technique used to analyze the
frequency components of an image. It's commonly used in image
processing for filtering and compression tasks.
5
0
PART -A

6.How does convolution relate to image processing?


Convolution involves applying a mathematical kernel (filter) to an image. It's used
for various operations like blurring, sharpening, and edge detection in image
processing.

7. Describe the goal of image enhancement.


Image enhancement aims to improve the visual quality of an image by adjusting
parameters such as brightness, contrast, and other image properties.

8. What is the primary purpose of image restoration?


Image restoration techniques are used to remove noise, artifacts, or other
degradations from images, aiming to recover the original clean image.

9. How is the histogram used in image processing?


Histogram processing involves analyzing and modifying the distribution of pixel
intensities in an image. It's used for tasks like enhancing contrast or equalizing the
histogram.

10.What is the role of High-level Vision in Computer Vision?


High-level Vision focuses on advanced interpretation and understanding of images,
including complex tasks like scene understanding, object tracking, and human
behavior analysis.
11. Give an example of a task in Mid-level Vision. Answer: Object recognition, which
involves identifying and categorizing objects within images, is an example of a task
in Mid-level Vision.

12. What is the Fourier Transform used for in image processing? Answer: The Fourier
Transform is used to analyze the frequency components of an image, aiding tasks
like filtering and compression.
Part B Q
5
2

PART -B

1.What is Computer Vision? Explain its applications.


2. Geometric 2D transformation in image formation or
two-dimensional convolution
operation with the required equation.
3. What is convolution? Explain the process of image
convolution with example.
4.What is Digitization of the image and Explain
photometric image formation in detail.
5.What is histogram? Explain histogram equalization
algorithm. Write Matlab code for calculation of histogram
and histogram equalization.
6.Differentiate between locally adaptive histogram
equalization and block histogram equalization methods.
7.Differentiate between low pass filtering and high pass
filtering.
8.How do you perform filtering process in frequency
domain? Show step by step process with clear diagram.
9.Discuss wiener filter in detail.
10.Explain Butterworth Low Pass filter in frequency
domain.
Supportive Online
Certification courses
SUPPORTIVE ONLINE COURSES

Course
S No Course title Link
provider

https://www.udemy.co
Computer vision applies m/topic/computer-
machine learning.
1 Udemy vision/

https://www.udacity.co

Introduction to Computer m/course/introduction-


2 Udacity Vision to-computer-vision--
ud810

https://www.coursera.o
Advanced Computer Vision rg/learn/advanced-
3 Coursera with TensorFlow
computer-vision-with-
tensorflow
Computer Vision and https://www.edx.org/lear
Image Processing n/computer-
Fundamentals programming/ibm-
edX
4 computer-vision-and-
image-processing-
fundamentals?webview=
false&campaign=Comput
er+Vision+and+Image+
Processing+Fundamental
s&source=edx&product_
category=course&placem
ent_url=https%3A%2F%
2Fwww.edx.org%2Flearn
%2Fcomputer-vision

54
Real life Applications in
day to day life and to
Industry
5
6

REAL TIME APPLICATIONS IN DAY TO DAY LIFE

AND TO INDUSTRY

1.Explain the role of an Computer vision in Health care system.(k4,CO1)


Content beyond
Syllabus
5
8

Contents beyond the Syllabus

Basics of Computer Vision

Reference Video – Content Beyond Syllabus

https://www.youtube.com/watch?v=2w8XIskzdFw
Assessment Schedule
ASSESSMENT SCHEDULE

FIAT
Proposed date :24.08.2023
Prescribed Text books &
Reference books
PRESCRIBED TEXT BOOKS AND REFERENCE BOOKS

TEXT BOOKS
D. A. Forsyth, J. Ponce, “Computer Vision: A Modern Approach”,
Pearson Education,
2003.
2. Richard Szeliski, “Computer Vision: Algorithms and Applications”,
Springer Verlag London Limited,2011.
REFERENCE BOOKS
B. K. P. Horn -Robot Vision, McGraw-Hill.
Simon J. D. Prince, Computer Vision: Models, Learning, and
Inference, Cambridge University Press, 2012.
Mark Nixon and Alberto S. Aquado, Feature Extraction & Image
Processing for Computer Vision, Third Edition, Academic Press,
2012.
E. R. Davies, (2012), “Computer & Machine Vision”, Fourth Edition,
Academic Press.
Concise Computer Vision: An Introduction into Theory and
Algorithms, by Reinhard Klette,2014
Mini Project
Suggestions
MINI PROJECT SUGGESTIONS

1. License Plate Recognition with OpenCV and Tesseract OCR


2. Detect and Recognize Car License Plate from a video in real-
time
Thank you

Disclaimer:

This document is confidential and intended solely for the educational purpose of RMK Group of
Educational Institutions. If you have received this document through email in error, please notify the
system manager. This document contains proprietary information and is intended only to the
respective group / learning community as intended. If you are not the addressee you should not
disseminate, distribute or copy through e-mail. Please notify the sender immediately by e-mail if you
have received this document by mistake and delete this document from your system. If you are not
the intended recipient you are notified that disclosing, copying, distributing or taking any action in
reliance on the contents of this information is strictly prohibited.

You might also like