Unit 5
Unit 5
MOTION ESTIMATION
Chap 6,7,8
Introduction
• Refers to the process of calculating the motion or movement of
objects or features within a sequence of images or frames over time.
• The primary goal of motion estimation is to estimate the motion
field, which describes how pixels or features move from one frame to
another.
• This information is crucial for understanding the dynamics of a scene and can
be used to make predictions about future movements.
• In summary, motion estimation is a crucial aspect of computer vision,
enabling the understanding and analysis of the dynamics and movements
within a sequence of images or video frames. It has broad applications in
fields like video processing, surveillance, robotics, and autonomous
vehicles. This technique is fundamental for various applications,
including video compression, object tracking, optical flow estimation,
image stabilization, action recognition, and more.
Types
• Global Motion Estimation:
• Global motion models assume that the entire image or a large portion of it undergoes a
uniform motion. Common global motion models include translations, rotations, scaling,
and affine transformations. These models are useful for estimating camera motion or
large rigid motions in a scene.
• Local Motion Estimation:
• Local motion estimation focuses on estimating motion at a smaller scale, typically
within small regions or patches of an image. This is often achieved by estimating the
optical flow, which provides the motion vector for each pixel or feature between
consecutive frames. Optical flow can be dense (computed for all pixels) or sparse
(computed for selected feature points).
ALGORITHMS USED
• Block matching, Lucas-Kanade method, Horn-Schunck method, Farnebäck
method
• Deep learning-based approaches like CNNs
Feature Based Alignment (Chap-6)
• 2D and 3D feature-based alignment
• Pose estimation
• Geometric intrinsic calibration
SfM Techniques(Chap-7)
• Introduction & Concepts-Sfm ppt
• Triangulation
• Factorization
• Bundle Adjustment
• Constrained structure and motion
Triangulation
• The problem of determining a point’s 3D position from a set of
corresponding image locations and known camera positions is known as
triangulation.
• fundamental in computer vision and 3D reconstruction, enabling
the reconstruction of a 3D scene or object from multiple 2D views.
If the points xL and xR are known, their projection lines are also known. If the
two image points correspond to the same 3D point X the projection lines must
intersect precisely at X. This means that X can be calculated from the
coordinates of the two image points, a process called triangulation.
• A point x (3D point) in 3D space is projected onto
the respective image plane along a line (green)
which goes through the camera's focal point,o1 and
o2 resulting in two points y1 and y2.
• If y1 and y2 are given and the geometry of the two
cameras are known, the two projection lines (green
lines) can be determined and it must be the case
The ideal case of epipolar geometry. A 3D point x is projected onto two that they intersect at point x (3D point).
camera images through lines (green) which intersect with each camera's • Using basic linear algebra that intersection point can
focal point, O1 and O2. The resulting image points are y1 and y2. The be determined in a straightforward way.
green lines intersect at x. • Due reasons like distortions, camera postions etc
the measured image points are y1’ and y2’. Their
projection lines (blue) do not have to intersect in 3D
space or come close to x.
• Triangulation solves this problem
where A is a matrix of size nX2 containing the image gradient components evaluated for each pixel of
the window W:
v is the vector representing the optical flow of the window W that we are going to estimate
• The least-square solution is:
Aperture Problem: It is challenging to determine the true motion direction when only a limited
portion of the object is visible in the image.
Illumination Changes: Changes in lighting conditions can violate the brightness constancy
assumption.
OF for Object Tracking using OpenCV
• https://mpolinowski.github.io/docs/IoT-and-Machine-Learning/ML/2
021-12-10--opencv-optical-flow-tracking/2021-12-10/
• Shi-Tomasi corner detection to get key points of corners of
objects and pass it to LK method.
• ()