ImageAnalysisForArcGIS10 2010
ImageAnalysisForArcGIS10 2010
\ .
|
|
| |
i 1 =
q
F
--------------------------------------------- =
86 Performing Spatial Enhancement
Source: Modified from Jensen 1996; Schowengerdt 1983
The sum of the coefficients (F) is used as the denominator of the
equation above so that the output values are in relatively the same
range as the input values. Because F cannot equal 0 (division by 0 is
not defined), F is set to 1 if the sum is 0.
Zero Sum Kernels Zero sum kernels are kernels in which the sum of all coefficients in the
kernel equals 0. When a zero sum kernel is used, the sum of the
coefficients is not used in the convolution equation, as shown above. In
this case, no division is performed (F = 1), because division by 0 is not
defined.
This generally causes the output values to be:
Zero in areas where all input values are equal (no edges)
Low in areas of low spatial frequency
Extreme in areas of high spatial frequency (high values become
much higher, low values become much lower)
Therefore, a zero sum kernel is an edge detector, which usually
smooths out or zeros out areas of low spatial frequency and creates a
sharp contrast where spatial frequency is high, which is at the edges
between homogeneous (homogeneity is low spatial frequency) groups
of pixels. The resulting image often consists of only edges and zeros.
Zero sum kernels can be biased to detect edges in a particular direction.
For example, this 3 x 3 kernel is biased to the south (Jensen 1996).
High-Frequency Kernels A high-frequency kernel, or high-pass kernel, has the effect of
increasing spatial frequency.
High-frequency kernels serve as edge enhancers because they bring
out the edges between homogeneous groups of pixels. Unlike edge
detectors (such as zero sum kernels), they highlight edges and do not
necessarily eliminate other features.
-1 -1 -1
1 -2 1
1 1 1
-1 -1 -1
-1 16 -1
-1 -1 -1
Performing Spatial Enhancement 87
When a high-frequency kernel is used on a set of pixels in which a
relatively low value is surrounded by higher values, like this....
...the low value gets lower. Inversely, when the high-frequency kernel is
used on a set of pixels in which a relatively high value is surrounded by
lower values...
...the high value becomes higher. In either case, spatial frequency is
increased by this kernel.
Low-Frequency Kernels Below is an example of a low-frequency kernel, or low-pass kernel,
which decreases spatial frequency.
This kernel averages the values of the pixels, causing them to be more
homogeneous. The resulting image looks either more smooth or more
blurred.
Figure 18: Convolution with High-Pass Filtering
BEFORE AFTER
204 200 197 - - -
201 106 209 - 10 -
198 200 210 - - -
BEFORE AFTER
64 60 57 - - -
61 125 69 - 188 -
58 60 70 - - -
1 1 1
1 1 1
1 1 1
88 Performing Spatial Enhancement
Applying Convolution To apply convolution, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spatial
Enhancement, and then select Convolution to open the Convolution
dialog.
2. Select a file from the Input Image dropdown list, or navigate to the
directory where the file is stored.
3. Select a kernel to use from the Kernel dropdown list.
4. Click either the Reflection or Background Fill button to specify the
way to handle edges in the image. See Using Convolution for more
information.
5. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
Convolution with High Pass
3
4
5
6
2
Performing Spatial Enhancement 89
6. Click OK to close the Convolution dialog.
Using Convolution
Convolution lets you perform image-enhancement operations such as
averaging and high-pass or low-pass filtering.
Reflection fills in the area beyond the edge of the image with a reflection
of the values at the edge. Background Fill uses zeros to fill in the kernel
area beyond the edge of the image.
Each data file value of the new output file is calculated by centering the
kernel over a pixel and multiplying the original values of the center pixel
and the appropriate surrounding pixels by the corresponding
coefficients from the matrix. These numbers are summed and then
divided by the sum of the coefficients to ensure the output values are
within the general range of the input values. If the sum is zero, the
division is not performed.
Non-Directional
Edge
The non-directional edge function averages the results of two
orthogonal first derivative edge detectors, using the Sobel and Prewitt
filters. The filters are based on a calculation of the 1st derivative, or
slope, in both the X and Y directions. Both use orthogonal kernels
convolved separately with the original image and then combined.
The non-directional edge function is based on the Sobel zero sum
convolution kernel. Most of the standard image processing filters are
implemented as a single-pass moving window (kernel) convolution.
Examples include low-pass, edge-enhance, edge-detection, and
summary filters.
For this model, a Sobel filter is used. To convert this model to the
Prewitt filter calculation, change the kernels according to the example
below.
1 0 1
2 0 2
1 0 1
Vertical
1 2 1
0 0 0
1 2 1
Horizontal
Sobel=
1 0 1
1 0 1
1 0 1
Vertical
1 1 1
0 0 0
1 1 1
Horizontal
Prewitt=
90 Performing Spatial Enhancement
Figure 19: Image of Seattle before Non-Directional Edge
Figure 20: Image of Seattle after Non-Directional Edge
Performing Spatial Enhancement 91
Applying Non-Directional
Edge
To apply non-directional edge, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spatial
Enhancement, and then select Non-Directional Edge to open the
Non-Directional Edge dialog.
2. Select a file from the Input Image dropdown list, or navigate to the
directory where the file is stored.
3. Click either the Sobel or Prewitt button to specify the filter to use.
4. Click either the Reflection or Background Fill button to specify the
way to handle edges in the image. For more information, see Using
Non-Directional Edge.
5. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
6. Click OK to close the Non-Directional Edge dialog.
Using Non-Directional Edge
In step 4 in the previous section, Reflection fills in the area beyond the
edge of the image with a reflection of the values at the edge.
Background Fill uses zeros to fill in the kernel area beyond the edge of
the image.
3
4
5
6
2
92 Performing Spatial Enhancement
Focal Analysis The focal analysis function lets you perform one of several types of
analysis on class values in an image file using a process similar to
convolution filtering.
This model (Median Filter) is useful for reducing noise such as random
spikes in data sets, dead sensor striping, and other impulse
imperfections in any type of image. It is also useful for enhancing
thematic images.
Focal analysis evaluates the region surrounding the pixel of interest
(center pixel). The operations you can perform on the pixel of interest
include:
Standard Deviation (measure of texture)
Sum
Mean (good for despeckling radar data)
Median (despeckle radar)
Min
Max
These functions let you select the size of the surrounding region to
evaluate by selecting the window size.
Figure 21: Image before Focal Analysis
Performing Spatial Enhancement 93
Figure 22: Image after Focal Analysis
Applying Focal Analysis To apply focal analysis, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spatial
Enhancement, and then select Focal to open the Focal Analysis
dialog.
2. Select a file from the Input Image dropdown list, or navigate to the
directory where the file is stored.
3. Select the function to use from the Focal Function dropdown list.
4. Select a shape from the Neighborhood Shape dropdown list.
5. Select a matrix size from the Neighborhood Definition - Matrix Size
dropdown list.
3
4
6
7
2
5
94 Performing Spatial Enhancement
6. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
7. Click OK to close the Focal Analysis dialog.
Using Focal Analysis
Focal analysis is similar to convolution in the process it uses. With focal
analysis, you can perform several different types of analysis on the pixel
values in an image file.
Resolution Merge The resolution of a specific sensor can refer to radiometric, spatial,
spectral, or temporal resolution. This function merges imagery of
differing spatial resolutions.
Landsat TM sensors have seven bands with a spatial resolution of 28.5
m. SPOT panchromatic has one broad band with a very good spatial
resolution of 10 m. Combining these two images to yield a seven-band
data set with 10 m resolution provides the best characteristics of both
sensors.
A number of models have been suggested to achieve this image merge:
Welch and Ehlers (1987) use forward-reverse RGB to IHS
transforms, replacing I (from transformed TM data) with the SPOT
panchromatic image. However, this technique is limited to three
bands (RGB).
Chavez (1991), among others, uses the forward-reverse principal
components transforms with the SPOT image, replacing PC-1.
In the above two techniques, it is assumed that the intensity component
(PC-1 or I) is spectrally equivalent to the SPOT panchromatic image,
and that all the spectral information is contained in the other PCs or in
H and S. Because SPOT data does not cover the full spectral range that
TM data does, this assumption does not strictly hold. It is unacceptable
to resample the thermal band (TM6) based on the visible (SPOT
panchromatic) image.
Another technique (Schowengerdt 1980) additively combines a high-
frequency image derived from the high-spatial resolution data (that is,
SPOT panchromatic) with the high-spectral resolution Landsat TM
image.
Brovey Transform The resolution merge function uses the Brovey Transform method of
resampling low spatial resolution data to a higher spatial resolution
while retaining spectral information.
In the Brovey Transform, three bands are used according to the
following formula:
Performing Spatial Enhancement 95
DNB1_new = [DNB1 / DNB1 + DNB2 + DNB3]
[DNhigh res. image]
DNB2_new = [DNB2 / DNB1 + DNB2 + DNB3]
[DNhigh res. image]
DNB3_new = [DNB3 / DNB1 + DNB2 + DNB3]
[DNhigh res. image]
Where:
B = band
The Brovey Transform was developed to visually increase contrast in
the low and high ends of an images histogram (that is, to provide
contrast in shadows, water, and high-reflectance areas such as urban
features). Brovey Transform is good for producing RGB images with a
higher degree of contrast in the low and high ends of the image
histogram and for producing visually appealing images.
Because the Brovey Transform is intended to produce RGB images,
you should merge only three bands at a time from the input
multispectral scene, such as bands 3, 2, 1 from a SPOT or Landsat TM
image or 4, 3, 2 from a Landsat TM image. You should display the
resulting merged image with bands 1, 2, 3 to RGB.
Applying Resolution
Merge
To apply resolution merge, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spatial
Enhancement, and then select Resolution Merge to open the
Resolution Merge dialog.
2. Select a file from the High Resolution Image dropdown list, or navigate
to the directory where the file is stored.
3. Select a file from the Multi-Spectral Image dropdown list, or navigate to
the directory where the file is stored.
4. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
3
4
2
5
96 Performing Spatial Enhancement
5. Click OK to close the Resolution Merge dialog.
Using Resolution Merge
Use resolution merge to integrate imagery of different spatial
resolutions (pixel size). The following images display the inputs and
result of the resolution merge function. The first image is a high-
resolution image, the second image is a multi-spectral image, and the
bottom image in a resolution merge.
Figure 23: High-Resolution, Multi-spectral, and Resolution Merge
Images
97 Using Radiometric Enhancement
Using Radiometric Enhancement 97
Using Radiometric Enhancement
Radiometric enhancement deals with the individual values of pixels in
an image. It differs from spatial enhancement, which takes into account
the values of neighboring pixels.
Radiometric enhancement contains functions to enhance your image
by using the values of individual pixels in each band. Depending on the
points and the bands in which they appear, radiometric enhancements
that are applied to one band might not be appropriate for other bands.
Therefore, the radiometric enhancement of a multiband image can
usually be considered as a series of independent, single-band
enhancements (Faust 1989).
IN THIS CHAPTER
LUT Stretch
Histogram Equalization
Histogram Matching
Brightness Inversion
98 Using Radiometric Enhancement
LUT Stretch LUT stretch creates an output image that contains the data values as
modified by a lookup table (LUT). The output is three bands.
Contrast Stretch When radiometric enhancements are performed on the display device,
the transformation of data file values into brightness values is illustrated
by the graph of a lookup table.
Contrast stretching involves taking a narrow input range and stretching
the output brightness values for those same pixels over a wider range.
This process is done using the Layer Properties dialog in ArcGIS.
Linear and Nonlinear The terms linear and nonlinear, when describing types of spectral
enhancement, refer to the function applied to data to perform the
enhancement. A piecewise linear stretch uses a polyline function to
increase contrast to varying degrees over different ranges of the data.
Linear Contrast Stretch A linear contrast stretch is a simple way to improve the visible contrast
of an image. It is often necessary to contrast stretch raw image data so
you can see it on the display device.
In most raw data, the data file values fall within a narrow rangeusually
a range much narrower than the display device is capable of displaying.
That range can be expanded to utilize the total range of the display
device (usually 0 to 255).
Nonlinear Contrast
Stretch
A nonlinear spectral enhancement gradually increases or decreases
contrast over a range, instead of applying the same amount of contrast
(slope) across the entire image. Usually, nonlinear enhancements bring
out the contrast in one range while decreasing the contrast in other
ranges.
Piecewise Linear
Contrast Stretch
A piecewise linear contrast stretch allows for the enhancement of a
specific portion of data by dividing the lookup table into three sections:
low, middle, and high. It lets you create a number of straight-line
segments that can simulate a curve. You can enhance the contrast or
brightness of any section in a single color gun at a time. This technique
is very useful for enhancing image areas in shadow or other areas of
low contrast.
A piecewise linear contrast stretch normally follows two rules:
1. The data values are continuous; there can be no break in the values
between high, middle, and low. Range specifications adjust in relation
to any changes to maintain the data value range.
2. The data values specified can go only in an upward, increasing
direction.
Using Radiometric Enhancement 99
The contrast value for each range represents a percentage of the
available output range that particular range occupies. Because rules 1
and 2 above are enforced, the contrast and brightness values may
affect the contrast and brightness of other ranges as they are changed.
For example, if the contrast of the low range increases, it forces the
contrast of the middle range to decrease.
Contrast Stretch on the
Display
Usually, a contrast stretch is performed on the display device only, so
that the data file values are not changed. Lookup tables are created that
convert the range of data file values to the maximum range of the
display device. You can then edit and save the contrast stretch values
and lookup tables as part of the raster data image file. These values are
loaded into view as the default display values the next time the image
is displayed.
The statistics in the image file contain the mean, standard deviation,
and other statistics on each band of data. The mean and standard
deviation are used to determine the range of data file values to translate
into brightness values or new data file values. You can specify the
number of standard deviations from the mean to use in the contrast
stretch. Usually the data file values that are two standard deviations
above and below the mean are used. If the data has a normal
distribution, this range represents approximately 95 percent of the data.
The mean and standard deviation are used instead of the minimum and
maximum data file values because the minimum and maximum data file
values do not usually represent most of the data. A notable exception
occurs when the feature being sought is in shadow. The shadow pixels
are usually at the low extreme of the data file values, outside the range
of two standard deviations from the mean.
Varying the Contrast
Stretch
There are variations of the contrast stretch you can use to change the
contrast of values over a specific range, or by a specific amount. By
manipulating the lookup tables as in the following illustration, you can
bring out the maximum contrast in the features of an image.
100 Using Radiometric Enhancement
This figure shows how the contrast stretch manipulates the histogram
of the data, increasing contrast in some areas and decreasing it in
others. This is also a good example of a piecewise linear contrast
stretch, which is created by adding breakpoints to the histogram.
Figure 24: Contrast Stretch
Applying a LUT Stretch To apply a LUT stretch, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Radiometric
Enhancement, and then select LUT Stretch to open the LUT Stretch
dialog.
2. Select the file you want to use from the Input Image dropdown list, or
navigate to the directory where it is stored.
3
2
5
Using Radiometric Enhancement 101
3. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
4. Set the output type to TIFF.
5. Click OK to close the LUT Stretch dialog.
Using LUT Stretch
LUT stretch provides a means of producing an output image that has
the stretch built into the pixel values to use with packages that have no
stretching capabilities.
Histogram
Equalization
Histogram equalization is a nonlinear stretch that redistributes pixel
values so that there is approximately the same number of pixels with
each value in a range. The result approximates a flat histogram.
Therefore, contrast is increased at the peaks of the histogram and
lessened at the tails.
Histogram equalization can also separate pixels into distinct groups if
there are few output values over a wide range. This can have the visual
effect of a crude classification.
Original Histogram
After Equalization
Peak
Tail
Pixels at peak are spread
apartcontrast is gained
Pixels at
tail are
grouped
contrast
is lost
102 Using Radiometric Enhancement
When performing a histogram equalization, the pixel values of an image
(either data file values or brightness values) are reassigned to a certain
number of bins, which are numbered sets of pixels. The pixels are then
given new values based on the bins to which they are assigned.
The total number of pixels is divided by the number of bins, equaling the
number of pixels per bin, as shown in the following equation:
Where:
N = Number of bins
T = The total number of pixels in the image
A = The equalized number of pixels per bin
The pixels of each input value are assigned to bins, so that the number
of pixels in each bin is as close to A as possible. Consider the following:
There are 240 pixels represented by this histogram. To equalize it to 10
bins, there would be:
240 pixels / 10 bins = 24 pixels per bin = A
A
T
N
---- =
0 1 2 3 4 5 6 7 8 9
5 5
10
15
60 60
40
30
10
5
N
u
m
b
e
r
o
f
P
i
x
e
l
s
Data File Values
A = 24
Using Radiometric Enhancement 103
The following equation is used to assign pixels to bins:
Where:
A =Equalized number of pixels per bin (see above)
H
i
=The number of values with the value i (histogram)
int=Integer function (truncating real numbers to integer)
B
i
=Bin number for pixels with value i
Source: Modified from Gonzalez and Wintz 1977
The 10 bins are rescaled to the range 0 to M. In this example, M = 9
because the input values ranged from 0 to 9 so that the equalized
histogram can be compared to the original. The output histogram of this
equalized image looks like the following illustration:
Effect on Contrast By comparing the original histogram of the example data with the last
one, you can see that the enhanced image gains contrast in the peaks
of the original histogram. For example, the input range of 3 to 7 is
stretched to the range 1 to 8. However, data values at the tails of the
original histogram are grouped together. Input values 0 through 2 all
have the output value of 0. So, contrast among the tail pixels, which
usually make up the darkest and brightest regions of the input image, is
lost.
B
i
int
H
k
k 1 =
i 1
\ .
|
|
| |
H
i
2
------ +
A
---------------------------------------- =
0 1 2 3 4 5 6 7 8 9
15
60 60
40
30
N
u
m
b
e
r
o
f
P
i
x
e
l
s
Output Data File Values
A = 24
20
15
0
1
2
3
4 5
7
8
9
6
Numbers inside bars are input data file values
0 0 0
104 Using Radiometric Enhancement
The resulting histogram is not exactly flat because pixels rarely group
together into bins with an equal number of pixels. Sets of pixels with the
same value are never split up to form equal bins.
Performing Histogram
Equalization
To perform histogram equalization, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Radiometric
Enhancement, and then select Histogram Equalization to open the
Histogram Equalization dialog.
2. Select the file you want to use from the Input Image dropdown list, or
navigate to the directory where it is stored.
3. Type the number of bins in the Number of Bins field.
4. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
5. Click OK to close the Histogram Equalization dialog.
3
5
4
2
Using Radiometric Enhancement 105
Using Histogram Equalization
The histogram equalization process works by redistributing pixel values
so that there are approximately the same number of pixels with each
value within a range.
Histogram equalization can also separate pixels into distinct groups if
there are few output values over a wide range. This process can have
the effect of a crude classification.
Histogram
Matching
Histogram matching is the process of determining a lookup table that
converts the histogram of one image so that it resembles the histogram
of another. Histogram matching is useful for matching data of the same
or adjacent scenes that were collected on separate days, or are slightly
different because of sun angle or atmospheric effects. This is especially
useful for mosaicking or change detection.
The two input images should have similar characteristics to achieve
good results with histogram matching:
The general shape of the histogram curves should be similar.
Relative dark and light features in the image should be the same.
For some applications, the spatial resolution of the data should be
the same.
The relative distributions of land covers should be about the same,
even when matching scenes that are not of the same area.
106 Using Radiometric Enhancement
When matching the histograms, a lookup table is mathematically
derived, which serves as a function for converting one histogram to the
other as illustrated here.
Figure 25: Histogram Matching
Performing Histogram
Matching
To perform histogram matching, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Radiometric
Enhancement, and then select Histogram Match to open the
Histogram Match dialog.
Source histogram (A), mapped through the lookup table (B),
approximates model histogram (C).
F
r
e
q
u
e
n
c
y
Input
0 255
F
r
e
q
u
e
n
c
y
Input
0 255
F
r
e
q
u
e
n
c
y
Input
0 255
+
=
(A) (B)
(C)
3
4
2
Using Radiometric Enhancement 107
2. Select the file you want to use from the Input Image dropdown list, or
navigate to the directory where it is stored.
3. select the file you want to use from the Match Image dropdown list, or
navigate to the directory where it is stored.
4. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
5. Click OK to close the Histogram Match dialog.
Using Histogram Matching
Histogram matching mathematically determines a lookup table that
converts the histogram of one image to resemble the histogram of
another, and is particularly useful for mosaicking images or change
detection.
Perform histogram matching when using matching data of the same or
adjacent scenes that were gathered on different days and have
differences due to the angle of the sun or atmospheric effects.
Brightness
Inversion
The brightness inversion function produces images that have the
opposite contrast of the original image. Dark detail becomes light, and
light detail becomes dark.
Inverse emphasizes detail that might be lost in the darkness of low DN
pixels. This function applies the following algorithm:
DN
out
= 1.0 if 0.0 < DN
in
< 0.1
DN
out
=
Figure 26: Image before Brightness Inversion
0.1
DN
in
if 0.1 < DN < 1
108 Using Radiometric Enhancement
Figure 27: Image after Brightness Inversion
Using Radiometric Enhancement 109
Applying Brightness
Inversion
To apply brightness inversion, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Radiometric
Enhancement, and then select Brightness Inversion to open the
Brightness Inversion dialog.
2. Select the file you want to use from the Input Image dropdown list, or
navigate to the directory where it is stored.
3. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
4. Click OK to close the Brightness Inversion dialog.
3
4
2
110 Using Radiometric Enhancement
111 Applying Spectral Enhancement
Applying Spectral Enhancement 111
Applying Spectral Enhancement
Spectral enhancement enhances images by transforming the values of
each pixel on a multiband basis. All of the techniques in this chapter
require more than one band of data. You can use them to:
Extract new bands of data that are more interpretable to the eye
Apply mathematical transforms and algorithms
Display a wider variety of information in the three available color
guns
You can use the features of spectral enhancement to study patterns
that can occur with deforestation or crop rotation and to see images in
a more natural state. You can also use spectral enhancement to view
images in different ways, such as changing the bands in an image from
red, green, and blue (RGB) to intensity, hue, and saturation (IHS).
IN THIS CHAPTER
RGB to IHS
IHS to RGB
Vegetative Indices
Color IR to Natural Color
112 Applying Spectral Enhancement
RGB to IHS The color monitors used for image display on image processing
systems have three color guns that correspond to RGB, the additive
primary colors. When displaying three bands of a multiband data set,
the viewed image is considered to be in RGB space.
However, it is possible to define an alternate color space that uses IHS
as the three positioned parameters (in lieu of RGB). This system is
advantageous in that it presents colors more closely as perceived by
the human eye:
Intensity Refers to the overall brightness of the scene (like PC-1)
and varies from 0 (black) to 1 (white).
Saturation Represents the purity of color and also varies linearly
from 0 to 1.
Hue Represents the color or dominant wavelength of the pixel. It
varies from 0 at the red midpoint through green and blue back to the
red midpoint at 360. It is a circular dimension. In the following
image, 0 to 255 is the selected range; it can be defined as any data
range. However, hue must vary from 0 to 360 to define the entire
sphere (Buchanan 1979).
Figure 28: Variance of Intensity and Hue in RGB to IHS
Saturation
Hue
I
n
t
e
n
s
i
t
y
Applying Spectral Enhancement 113
The following algorithm was used in the Image Analysis for ArcGIS
RGB to IHS transform (Conrac 1980):
Where:
R, G, B= Are each in the range of 0 to 1.0
r, g, b=Are each in the range of 0 to 1.0
M= Largest value, r, g, or b
m= Least value, r, g, or b
At least one of the R, G, or B values is 0, corresponding to the color with
the largest value, and at least one of the R, G, or B values is 1,
corresponding to the color with the least value.
The equation for calculating intensity in the range of 0 to 1.0 is:
The equations for calculating saturation in the range of 0 to 1.0 are:
If M = m, S = 0
If I 0.5,
If I > 0.5,
R
M r
M m
--------------- =
G
M g
M m
--------------- =
B
M b
M m
--------------- =
I
M m +
2
---------------- =
S
M m
M m +
---------------- =
S
M m
2 M m
------------------------ =
114 Applying Spectral Enhancement
The equations for calculating hue in the range of 0 to 360 are:
If M = m, H = 0
If R = M, H = 60 (2 + b - g)
If G = M, H = 60 (4 + r - b)
If B = M, H = 60 (6 + g - r)
Where:
R, G, B =Are each in the range of 0 to 1.0
M = Largest value, R, G, or B
m = Least value, R, G, or B
Converting RGB to IHS To convert RGB to IHS, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spectral
Enhancement, and then select RGB to IHS to open the RGB to IHS
dialog.
2. Type the name of the input image in the Input Image field, or navigate
to the directory where it is stored.
3. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
4. Click OK to close the RGB to IHS dialog.
Using RGB to IHS
Using RGB to IHS applies an algorithm that transforms RGB values to
IHS values.
4
2
3
Applying Spectral Enhancement 115
IHS to RGB IHS to RGB is intended as a complement to the standard RGB to IHS
transform. In the IHS to RGB algorithm, a min-max stretch is applied to
either intensity (I), saturation (S), or both, so that they more fully utilize
the 0 to 1 value range. The values for hue (H), a circular dimension, are
0 to 360. However, depending on the dynamic range of the DN values
of the input image, it is possible that I or S or both occupy only a part of
the 0 to 1 range. In this model, a min-max stretch is applied to either I,
S, or both, so that they more fully utilize the 0 to 1 value range. After
stretching, the full IHS image is retransformed back to the original RGB
space. As the parameter hue is not modified, it largely defines what we
perceive as color, and the resulting image looks very much like the input
image.
It is not essential that the input parameters (IHS) to this transform are
derived from an RGB to IHS transform. You can define I or S as other
parameters, set hue at 0 to 360, and then transform to RGB space. This
is a method of color coding other data sets.
In another approach (Daily 1983), H and I are replaced by low- and
high-frequency radar imagery. You can also replace I with radar
intensity before the IHS to RGB transform (Holcomb 1993). Chavez
evaluates the use of the IHS to RGB transform to resolution merge
Landsat TM with SPOT panchromatic imagery (Chavez 1991).
The algorithm used by Image Analysis for ArcGIS for the IHS to RGB
function is (Conrac 1980):
Given: H in the range of 0 to 360; I and S in the range of 0 to 1.0
If I 0.5,
If I > 0.5,
The equations for calculating R in the range of 0 to 1.0 are:
If H < 60,
If 60 H < 180,
If 180 H < 240,
If 240 H 360,
M I 1 S + ( ) =
M I S I S ( ) + =
m 2 1 M =
R m M m ( )
H
60
------
\ .
| |
+ =
R M =
R m M m ( )
240 H
60
-------------------
\ .
| |
+ =
R m =
116 Applying Spectral Enhancement
The equations for calculating G in the range of 0 to 1.0 are:
If H < 120,
If 120 H < 180,
If 180 H < 300,
If 300 H 360,
The equations for calculating B in the range of 0 to 1.0 are:
If H < 60,
If 60 H < 120,
If 120 H < 240,
If 240 H < 300,
If 300 H 360,
Converting IHS to RGB To convert IHS to RGB, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spectral
Enhancement, and then select IHS to RGB to open the IHS to RGB
dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the input image is stored.
3. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
4. Click OK to close the IHS to RGB dialog.
Using IHS to RGB
Using IHS to RGB applies an algorithm that transforms IHS values to
RGB values.
G m =
G m M m ( )
H 120
60
-------------------
\ .
| |
+ =
G M =
G m M m ( )
360 H
60
-------------------
\ .
| |
+ =
B M =
B m M m ( )
120 H
60
-------------------
\ .
| |
+ =
B M =
B m M m ( )
H 240
60
-------------------
\ .
| |
+ =
B M =
4
2
3
Applying Spectral Enhancement 117
Vegetative Indices Mapping vegetation is a common application of remotely sensed
imagery. To help you find vegetation quickly and easily, Image Analysis
for ArcGIS includes a vegetative indices feature.
Indices are used to create output images by mathematically combining
the DN values of different bands. These may be simplistic:
(Band X - Band Y)
Or more complex:
In many instances, the indices are ratios of band DN values:
These ratio images are derived from the absorption/reflection spectra of
the material of interest. The absorption is based on the molecular bonds
in the (surface) material. Thus, the ratio often gives information on the
chemical composition of the target.
Applications The main applications of vegetative analysis are as follows:
Indices are used extensively in vegetation analysis to bring out
small differences between various vegetation classes. Often,
judiciously chosen indices can highlight and enhance differences
that you cannot observe in the display of the original color bands.
Indices are also used to minimize shadow effects in satellite and
aircraft multispectral images. You can generate black-and-white
images of individual indices, or a color combination of three ratios.
Examples The following are examples of indices that are preprogrammed in
Image Analysis for ArcGIS:
IR/R (infrared/red)
SQRT (IR/R)
Vegetation Index = IR-R
Normalized Difference Vegetation Index (NDVI) =
Transformed NDVI (TNDVI) =
Source: Modified from Sabins 1987; Jensen 1996; Tucker 1979
The following table shows the infrared (IR) and red (R) band for some
common sensors (Tucker 1979, Jensen 1996):
BandX BandY
BandX BandY +
------------------------------------------
BandX
BandY
-----------------
IR R
IR R +
----------------
IR R
IR R +
---------------- 0.5 +
118 Applying Spectral Enhancement
Image Algebra Image algebra is a general term used to describe operations that
combine the pixels of two or more raster layers in mathematical
combinations. For example, the calculation:
(infrared band) - (red band)
DN
ir
- DN
red
yields a simple, yet very useful, measure of the presence of vegetation.
Band ratios are also commonly used. These are derived from the
absorption spectra of the material of interest. The numerator is a
baseline of background absorption, and the denominator is an
absorption peak. NDVI is a combination of addition, subtraction, and
division.
NDVI =
Table 5: IR and R Bands of Common Sensors
Sensor IR Band R Band
Landsat MSS 4 2
SPOT XS 3 2
Landsat TM 4 3
NOAA AVHRR 2 1
IR R
IR R +
----------------
Applying Spectral Enhancement 119
Applying Vegetative
Indices
To apply vegetative indices, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spectral
Enhancement, and then select Vegetative Indices to open the
Vegetative Indices dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the input image is stored.
3. Select an appropriate layer from the Near Infrared Band dropdown list.
4. Select an appropriate layer from the Visible Red Band dropdown list.
5. Select an appropriate index from the Desired Index dropdown list.
6. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
7. Click OK to close the Vegetative Indices dialog.
2
7
3
4
6
5
120 Applying Spectral Enhancement
Color IR to Natural
Color
This function lets you simulate natural colors from the bands of data
from an infrared image so that the output is a fair approximation of a
natural color image. You cannot apply this feature to images having
only one band of data (for example, grayscale images). Its for use when
you have data only from the near infrared, visible red, and visible green
segments of the spectrum.
When an image is displayed in natural color, the bands are arranged to
approximate the most natural representation of the image in the real
world. Vegetation becomes green in color, and water becomes dark in
color. Certain bands of data are assigned to the red, green, and blue
color guns of your computer monitor to create natural color.
Figure 29: Infrared Image of a Golf Course
Figure 30: Natural Colors after Color IR to Natural Color
Applying Spectral Enhancement 121
Changing Color IR to
Natural Color
To change Color IR to Natural Color, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Spectral
Enhancement, and then select Color IR to Natural Color to open the
Color IR to Natural Color dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the input image is stored.
3. Select the appropriate band from the Near Infrared Band dropdown list.
4. Select the appropriate band from the Visible Red Band dropdown list.
5. Select the appropriate band from the Visible Green Band dropdown list.
6. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
7. Click OK to close the Color IR to Natural Color dialog.
2
4
3
7
6
5
122 Applying Spectral Enhancement
123 Performing GIS Analysis
Performing GIS Analysis 123
Performing GIS Analysis
A geographic information system (GIS) is a unique system designed to
enter, store, retrieve, manipulate, and analyze layers of geographic
data to produce interpretable information. A GIS should also create
reports and maps (Marble 1990). The GIS database might include
images, maps, statistical data, or any other data needed in a study. This
chapter is about using the different analysis functions in Image Analysis
for ArcGIS to better use the images, data, maps, and so on located in a
GIS.
The central purpose of a GIS is to turn geographic data into useful
informationthe answers to real-life questionssuch as:
How to redraw political districts in a growing metropolitan area?
How to monitor the influence of global climatic changes on the
Earths resources?
What areas to protect to ensure the survival of endangered
species?
Although the term GIS is commonly used to describe software
packages, a true GIS includes knowledgeable staff, a training program,
budgets, marketing, hardware, data, and software (Walker and Miller
1990). You can use GIS technology in almost any geography-related
discipline, from landscape architecture to natural resource
management to transportation routing.
IN THIS CHAPTER
Information Versus Data
Neighborhood Analysis
Thematic Change
Summarize Areas
Recode
124 Performing GIS Analysis
Information Versus
Data
Information, as opposed to data, is independently meaningful. It is
relevant to a particular problem or question:
Data The land cover at coordinate N875250, E757261 has a data
file value of 8.
Information Land cover with a value of 8 are on slopes too steep
for development.
You can enter data into a GIS and produce information. The information
you want to derive determines the type of data you must enter. For
example, if you are looking for a suitable refuge for bald eagles, zip
code data is probably not needed, while land cover data might be
useful.
For this reason, the first step in any GIS project is usually an
assessment of the scope and goals of the study. Once the project is
defined, you can begin the process of building the database. Although
software and data are commercially available, you must create a
custom database for the particular project and study area. Design the
database to meet the needs and objectives of the organization.
A major step in successful GIS implementation is analysis. In the
analysis phase, data layers are combined and manipulated to create
new layers and to extract meaningful information from them.
Once the database (layers and attribute data) is assembled, the layers
are analyzed and new information is extracted. Some information is
extracted by looking at the layers and visually comparing them to other
layers. However, you can retrieve new information by combining and
comparing layers.
Performing GIS Analysis 125
Neighborhood
Analysis
Neighborhood analysis applies to any image processing technique that
takes surrounding pixels into consideration, such as convolution
filtering and scanning. This is similar to the convolution filtering
performed on continuous data. You can perform several types of
analyses, such as boundary, density, mean, and sum.
With a process similar to the convolution filtering of continuous raster
layers, you can also filter thematic raster layers. The GIS filtering
process is sometimes referred to as scanning, but is not to be confused
with data capture via a digital camera. Neighborhood analysis is based
on local or neighborhood characteristics of the data (Star and Estes
1990).
Every pixel is analyzed spatially, according to the pixels that surround
it. The number and the location of the surrounding pixels are
determined by a scanning window, which is defined by you. These
operations are known as focal operations.
Neighborhood analysis creates a new thematic layer. There are several
types of analyses that you can perform on each window of pixels, as
described below:
Density Produces the number of pixels that have the same class
value as the center (analyzed) pixel. It also measures homogeneity
(sameness) based upon the analyzed pixel. This is often useful in
assessing vegetation crown closure.
Diversity Produces the number of class values that are in the
window. Diversity is also a measure of heterogeneity (difference).
Majority Produces the class value that represents the majority of
the class values in the window. This option operates like a low-
frequency filter to clean up a salt-and-pepper layer.
Maximum Produces the greatest class value in the window. You
can use this to emphasize classes with the higher class values or to
eliminate linear features or boundaries.
Minimum Produces the least or smallest class value in the
window. You can use this option to emphasize classes with low-
class values.
Minority Produces the least common of the class values in the
window. You can use this option to identify the least common
classes or to highlight disconnected linear features.
Rank Produces the number of pixels in the scan window whose
value is less than the center pixel.
126 Performing GIS Analysis
Sum Totals the class values. In a file where class values are
ranked, totaling lets you further rank pixels based on their proximity
to high-ranking pixels.
Performing
Neighborhood Analysis
To perform neighborhood analysis, follow these steps:
1. Click the Image Analysis dropdown arrow, point to GIS Analysis, and
then select Neighborhood to open the Neighborhood Analysis dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the input image is stored.
3. Select the function you want to use from the Neighborhood Function
dropdown list.
4. Select Rectangle from the Neighborhood Shape dropdown list.
5. Select the size you want to use from the Neighborhood Definition -
Matrix Size dropdown list.
6. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
7. Click OK to close the Neighborhood Analysis dialog.
2
3
4
7
5
6
Performing GIS Analysis 127
Thematic Change Thematic change identifies areas that have undergone change over
time. Typically, you use thematic change after performing a
classification of your data. By using the categorizations of Before
Theme and After Theme in the Thematic Change dialog, you can
quantify both the amount and the type of changes that take place over
time. Image Analysis for ArcGIS produces a thematic image that has all
the possible combinations of change.
Thematic change creates an output image from two input classified
raster files. The class values of the two input files are organized into a
matrix. The first input file specifies the columns of the matrix, and the
second one specifies the rows. Zero or background values are not
treated special in any way. The number of classes in the output file is
the product of the number of classes from the two input files. The
classes should be the same in each image and in the same order.
Figure 31: Input Image from 1973
Figure 32: Input Image from 1994
Figure 33: Thematic Change Result
128 Performing GIS Analysis
Performing Thematic
Change
To perform thematic change, follow these steps:
1. Click the Image Analysis dropdown arrow, point to GIS Analysis, and
then select Thematic Change to open the Thematic Change dialog.
2. Click the browse button for the Before Theme field and navigate to the
directory where the before theme image is stored.
3. Click the browse button for the After Theme field and navigate to the
directory where the after theme image is stored.
4. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
5. Click OK to close the Thematic Change dialog.
Note: You must use the output from this exercise in the next exercise,
Summarize Areas on page 130. Please keep the image displayed in
your view if you are going on to that exercise.
2
5
3
4
Performing GIS Analysis 129
The following illustration is an example of the previous image after
undergoing thematic change. In the table of contents, you see the
combination of classes from the before and after images.
Figure 34: Image Showing Changes between 1973 and 1994
Summarize Areas Image Analysis for ArcGIS also provides summarize areas as a method
of assessing change in thematic data. Once you complete the thematic
change analysis, you can use summarize areas to limit the analysis to
only a portion of the image and derive quantitative information about
that area.
Summarize areas works by using a feature theme to compile
information about that area in tabular format. Summarize areas
produces cross-tabulation statistics that compare class value areas
between two thematic files using a single thematic change image, and
includes a number of points in common, number of acres (or hectares
or square miles) in common, and percentages.
An example of using summarize areas is to assist a regional planning
office in preparing a study of urban change for certain counties within
the jurisdiction or even within one county or city. A file containing a
polygonal boundary of the area to inventory is analyzed against a file for
the same geographical area containing the land cover categories. The
summary report indicates the amount of urban change in that particular
area of a larger thematic change.
130 Performing GIS Analysis
Applying Summarize
Areas
This exercise uses the output of the thematic change exercise in
Thematic Change on page 128 as the input. Display the thematic
change image in your view before starting this exercise.
To apply summarize areas, follow these steps:
1. Click the Image Analysis dropdown arrow, point to GIS Analysis, and
then select Summarize Areas to open the Summarize Areas dialog.
2. Select the vector theme you want to use from the Zone Theme
dropdown list, or navigate to the directory where it is stored.
3. Select the attribute you want to summarize from the Zone Attribute
dropdown list.
4. Select the class theme from the Class Theme dropdown list, or navigate
to the directory where it is stored. This is the thematic theme you
generated in Thematic Change on page 128.
5. Click the browse button for the Summarize Results Table field to specify
a name for the new summarize areas table that is created.
6. Click OK to close the Summarize Areas dialog.
When the process completes, the resulting table is added to ArcMap.
7. Click the Source tab in the ArcMap table of contents to see the new
table.
4
3
2
5
Performing GIS Analysis 131
Recode Recoding involves assigning new values to one or more classes of an
existing file. Recoding is used to:
Reduce the number of classes
Combine classes
Assign different class values to existing classes
Write class name and color changes to the Attribute table
When an ordinal, ratio, or interval class numbering system is used, you
can use recoding to assign classes to appropriate values. Recoding is
often performed to make later steps easier. For example, in creating a
model that produces good, better, and best areas, it might be beneficial
to recode the input layers so all of the best classes have the highest
class values.
You can also use recode to save any changes made to the color
scheme or class names of a classified image to the Attribute table for
later use. Just saving an image does not record these changes.
Recoding an image involves two major steps:
1. First, you arrange the discrete classes into common groups.
2. Secondly, you perform the actual recoding process, which rewrites the
Attribute table using the information from your grouping process.
The three recoding methods described in the exercises that follow this
section are more accurately described as three methods of grouping
the classified image to get it ready for the recode process. These
methods are:
Recoding by class name
Recoding by symbology
Recoding a previously grouped image
132 Performing GIS Analysis
The following is a thematic image of South Carolina soil types before
recode by class name.
Figure 35: Thematic Image before Recode by Class Name
The following is a thematic image of South Carolina soil types after the
recode. The changed and grouped class names are listed in the table
of contents.
Figure 36: Thematic Image after Recode by Class Name
Performing GIS Analysis 133
Recoding by Class Name You must first group the classified image in the ArcMap table of
contents and then perform the recode.
To recode by class name, follow these steps:
1. Click the Add Data button and add a classified image to your view.
2. Identify the classes you want to group in the table of contents.
3. Right-click the image name and select Properties.
2
3
134 Performing GIS Analysis
The Layer Properties dialog opens.
4. Click the Symbology tab.
5. Click the Unique Values category.
6. Rename the classes in the Label column in the table that appears in the
middle of the tab so that all of the classes you group together have the
same class name.
7. Double-click the color bar in the Symbol column for each class and
change the colors to reflect the color scheme you want.
The image in your view should now look the way you want.
Note: The Attribute table is not updated with the new class names.
8. Click Apply, and then click OK to apply your changes and close the
Layer Properties dialog.
9. Click the Image Analysis dropdown arrow, point to GIS Analysis, and
then select Recode.
4
5
7
7
Performing GIS Analysis 135
The Recode dialog opens.
10. Type the name of the output image in the Output Image field, or click
the browse button and navigate to the directory where you want the
output image stored and type a name.
11. Click OK to close the Recode dialog.
Recoding by Symbology This process shows you how to recode by symbology. You see
similarities with recoding by class name, but be aware of some different
procedures.
To record by symbology, follow these steps:
1. Click the Add Data button and display a classified image in your
view.
2. Identify the classes you want to group together.
10
11
136 Performing GIS Analysis
3. Double-click the image name in the table of contents.
The Layer Properties dialog opens.
4. Click the Symbology tab.
5. Click Unique Values in the Show list on the left side of the tab to
expand the category.
3
6
4
5
10
Performing GIS Analysis 137
6. Press the Ctrl key while clicking the first set of classes you want to
group together.
7. Right-click the selected classes and select Group Values.
8. Click in the Label column and type a new name for the class.
9. Follow steps steps 5 - 8 to group the rest of your classes.
10. Click Apply and OK to close the Layer Properties dialog.
11. Click the Image Analysis dropdown arrow, point to GIS Analysis, and
then select Recode to open the Recode dialog.
12. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
13. Click OK to close the Recode dialog.
Recoding with
Previously Grouped
Image
You may need to open an image that has been classified and grouped
in another program such as ERDAS IMAGINE
Band A
Data File Values
B
a
n
d
B
D
a
t
a
F
i
l
e
V
a
l
u
e
s
154 Understanding Classification
Considering the regular, arbitrary assignment of the initial cluster
means, the first iteration of the ISODATA algorithm always gives results
similar to those in this illustration.
For the second iteration, the means of all clusters are recalculated,
causing them to shift in feature space. The entire process is repeated
each candidate pixel is compared to the new cluster means and
assigned to the closest cluster mean.
Percentage Unchanged After each iteration, the normalized percentage of pixels whose
assignments are unchanged since the last iteration displays in the
dialog. When this number reaches T (the convergence threshold), the
program terminates.
1
Cluster
2
Cluster
3
Cluster
4
Cluster
5
Band A
Data File Values
B
a
n
d
B
D
a
t
a
F
i
l
e
V
a
l
u
e
s
Cluster
Band A
Data File Values
B
a
n
d
B
D
a
t
a
F
i
l
e
V
a
l
u
e
s
1
Cluster
2
Cluster
3
Cluster
4
Cluster
5
Cluster
Understanding Classification 155
Performing
Unsupervised
Classification
To perform unsupervised classification, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Classification, and
then select Unsupervised/Categorize to open the Unsupervised
Classification dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the input file is stored.
3. Type the number of classes you want in the Desired Number of Classes
field.
4. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
5. Click OK to close the Unsupervised Classification dialog.
Supervised
Classification
Supervised classification requires a priori (already known) information
about the data, such as:
What type of classes need to be extracted? Soil type? Land use?
Vegetation?
What classes are most likely to be present in the data? That is,
which types of land cover, soil, or vegetation (or whatever) are
represented by the data?
In supervised training, you rely on your own pattern recognition skills
and a priori knowledge of the data to help the system determine the
statistical criteria (signatures) for data classification. You should know
some informationeither spatial or spectralabout the pixels that you
want to classify to select reliable samples.
3
2
4
5
156 Understanding Classification
The location of a specific characteristic, such as a land cover type, may
be known through ground truthing. Ground truthing refers to the
acquisition of knowledge about the study area from field work, analysis
of aerial photography, personal experience, and so on. Ground truth
data is considered to be the most accurate (true) data available about
the area of study. It should be collected at the same time as the
remotely sensed data, so that the data corresponds as much as
possible (Star and Estes 1990). However, some ground data may not
be very accurate due to a number of errors and inaccuracies.
Performing Supervised
Classification
To perform supervised classification, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Classification, and
then select Supervised to open the Supervised Classification dialog.
2. Click the browse button for the Input Image field and navigate to the
directory where the file is stored.
3. Click the browse button for the Signature Features field and navigate to
the directory where the file is stored.
4. Select the field that contains the class names from the Class Name
Field dropdown list.
5. Click either the All Features or Selected Features button to specify
which features to use during classification.
6. Select the rule you want to use from the Classification Rule dropdown
list.
Note: For more information about each option, see Classification
Decision Rules on page 157.
3
2
6
7
4
5
8
Understanding Classification 157
7. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
8. Click OK to close the Supervised Classification dialog.
Classification
Decision Rules
Once you create and evaluate a set of reliable signatures, the next step
is to perform a classification of the data. Each pixel is analyzed
independently. The measurement vector for each pixel is compared to
each signature according to a decision rule or algorithm. Pixels that
pass the criteria and are established by the decision rule are then
assigned to the class for that signature. Image Analysis for ArcGIS lets
you classify the data parametrically with statistical representation.
Parametric Rules Image Analysis for ArcGIS provides these commonly used decision
rules for parametric signatures:
Minimum Distance
Mahalanobis Distance
Maximum Likelihood (With Bayesian Variation)
Nonparametric Rule Image Analysis for ArcGIS provides only one decision rule for
nonparametric signatures:
Parallelepiped
Minimum Distance The minimum distance decision rule (also called spectral distance)
calculates the spectral distance between the measurement vector for
the candidate pixel and the mean vector for each signature.
In this illustration, spectral distance is illustrated by the lines from the
candidate pixel to the means of the three signatures. The candidate
pixel is assigned to the class with the closest mean.
B3
B2
B1
A1
A2
A3
u
u
u
3
Band A
Data File Values
B
a
n
d
B
D
a
t
a
F
i
l
e
V
a
l
u
e
s
Candidate Pixel
o
o
158 Understanding Classification
The equation for classifying by spectral distance is based on the
equation for Euclidean distance.
Where:
n =Number of bands (dimensions)
i = A particular band
c = A particular class
X
xyi
= Data file value of pixel x,y in band i
ci
=Mean of data file values in band i for the sample for class c
SD
xyc
= Spectral distance from pixel x,y to the mean of class c
Source: Swain and Davis 1978
When spectral distance is computed for all possible values of c (all
possible classes) the class of the candidate pixel is assigned to the
class for which SD is the lowest.
Maximum Likelihood The Maximum Likelihood decision rule is based on the probability that
a pixel belongs to a particular class. The basic equation assumes that
these probabilities are equal for all classes, and that the input bands
have normal distributions.
Note: The Maximum Likelihood algorithm assumes that the histograms
of the bands of data have normal distributions. If this is not the case, you
might have better results with the minimum distance decision rule.
The equation for the Maximum Likelihood/Bayesian Classifier is as
follows:
Where:
D = Weighted distance (likelihood)
c =A particular class
X = The measurement vector of the candidate pixel
M
c
= The mean vector of the sample of class c
a
c
= Percent probability that any candidate pixel is a member of
class c (defaults to 1.0, or is entered from a priori data)
Cov
c
= The covariance matrix of the pixels in the sample of class c
|Cov
c
|=Determinant of Cov
c
(matrix algebra)
SD
xyc
ci
X
xyi
( )
2
i 1 =
n
=
a
c
( ) 0.5 Cov
c
( ) ln | | 0.5 X M
c
( )T Cov
c
1
\ .
| |
X M
c
( ) ln
Understanding Classification 159
Cov
c
-1 = Inverse of Cov
c
(matrix algebra)
ln = Natural logarithm function
T = Transposition function (matrix algebra)
Mahalanobis Distance Mahalanobis distance is similar to minimum distance, except that the
covariance matrix is used in the equation. Variance and covariance are
figured in so that clusters that are highly varied lead to similarly varied
classes, and vice versa. For example, when classifying urban areas
typically a class whose pixels vary widelycorrectly classified pixels
may be farther from the mean than those of a class for water, which is
usually not a highly varied class (Swain and Davis 1978).
Note: The Mahalanobis distance algorithm assumes that the
histograms of the bands have normal distributions. If this is not the
case, you might have better results with the parallelepiped or minimum
distance decision rule, or by performing a first-pass parallelepiped
classification.
The equation for the Mahalanobis distance classifier is as follows:
Where:
D= Mahalanobis distance
c= A particular class
X= The measurement vector of the candidate pixel
M
c
= The mean vector of the signature of class c
Cov
c
= The covariance matrix of the pixels in the signature of
class c
Cov
c
-1
= Inverse of Cov
c
T
= Transposition function
The pixel is assigned to the class, c, for which D is the lowest.
Parallelepiped Image Analysis for ArcGIS provides the Parallelepiped decision rule as
its nonparametric decision rule. In the Parallelepiped decision rule, the
data file values of the candidate pixel are compared to upper and lower
limits, which are the minimum and maximum data file values of each
band in the signature.
There are high and low limits for every signature in every band. When
a pixels data file values are between the limits for every band in a
signature, the pixel is assigned to that signatures class. If a pixel falls
into more than one class, the first class is the one assigned. If a pixel
falls into no class boundaries, it is labeled unclassified.
D X M
c
( )
T
Cov
c
1
\ .
| |
X M
c
( ) =
160 Understanding Classification
147 Using Conversion
Using Conversion 147
Using Conversion
The conversion feature gives you the ability to convert shapefiles to
raster images and raster images to shapefiles. This tool is very helpful
when you need to isolate or highlight certain parts of a raster image or
when you have a shapefile and you need to view it as a raster image.
Possible applications include viewing deforestation patterns, urban
sprawl, and shore erosion.
The Image Info tool that is discussed in Applying Data Tools on
page 51 is also an important part of raster/feature conversion. The
ability to assign certain pixel values as NoData is very helpful when
converting images.
IN THIS CHAPTER
Conversion
Converting Raster to Features
Converting Features to Raster
148 Using Conversion
Conversion Always be aware of how the raster dataset represents features when
converting points, polygons, or polylines to a raster, and vice versa.
There is a trade-off when working with a cell-based system. Although
points don't have area, cells do even if points are represented by a
single cell. The smaller the cell size, the smaller the area, and thus a
closer representation of the point feature.
Points with area have an accuracy of plus or minus half the cell size. For
many users, having all data types in the same format and being able to
use them interchangeably in the same language is more important than
a loss of accuracy.
Linear data is represented by a polyline that is also comprised of cells.
Therefore, it has area although, by definition, lines do not. Because of
this, the accuracy of representation varies according to the scale of the
data and the resolution of the raster dataset.
With polygonal or areal data, problems can occur from trying to
represent smooth polygon boundaries with square cells. The accuracy
of the representation is dependent upon the scale of the data and the
size of the cell. The finer the cell resolution and the greater the number
of cells that represent small areas, the more accurate the
representation.
Converting Raster
to Features
During a conversion of a raster representing polygonal features to a
polygonal vector file or feature dataset, the polygons are built from
groups of contiguous cells having the same cell values. Arcs are
created from cell borders in the raster. Continuous cells with the same
value are grouped together to form polygons. Cells that are NoData in
the input raster do not become features in the output polygon feature.
When a raster that represents linear features is converted to a polyline
vector file or feature dataset, a polyline is created from each cell in the
input raster, passing through the center of each cell. Cells that are
NoData in the input raster do not become features in the output polyline
feature.
When you convert a raster representing point features to point vector
file or feature dataset, a point is created in the output for each cell of the
input raster. Each point is positioned at the center of the cell it
represents. NoData cells are not transformed into points.
When you choose Convert Raster to Feature, the Raster to Feature
dialog gives you the choice of a field to specify from the image in the
conversion. You are also given the choice of an output geometry type
so you can specify whether the feature is a point, a polygon, or a
polyline according to the field and data youre using. You can specify
Generalize Lines to smooth out ragged or sharp edges in the new
feature file. You should note that regardless of which field you select,
the category is not populated in the Attribute table after conversion.
The images below show a raster image before conversion and a raster
image after conversion to a shapefile using Value as the field.
Using Conversion 149
Figure 42: Raster Image before Conversion
Figure 43: Raster Image after Conversion
Performing Raster to
Features Conversion
To perform raster to features conversion, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Convert, and then
select Convert Raster to Features to open the Raster to Features
dialog.
150 Using Conversion
2. Click the browse button for the Input Raster field and navigate to the
directory where the file is stored.
3. Select a field to use from the Field dropdown list.
4. Select Point, Polygon, or Polyline from the Output Geometry Type
dropdown list.
5. Check the Generalize Lines check box if you want to smooth out sharp
edges in the image.
6. Type the file name of the shapefile in the Output Features field, or
navigate to the directory where you want it stored.
7. Click OK to close the Raster to Features dialog.
3
4
5
6
7
2
Using Conversion 151
Converting
Features to Raster
You can convert polygons, polylines, or points from any source file to a
raster. You can convert features using both string and numeric fields.
Each unique string in a string field is assigned a unique value to the
output raster. A field is added to the table of the output raster to hold the
original string value from the features.
When you convert points, cells are given the value of the points found
within each cell. Cells that do not contain a point are given the value of
NoData. You can specify a cell size to use in the Features to Raster
dialog. Specify a cell size based on these factors:
The resolution of the input data
The output resolution needed to perform your analysis
The need to maintain a rapid processing speed
Polylines are features that, at certain resolutions, appear as lines
representing streams or roads. When you convert polylines, cells are
given the value of the line that intersects each cell. Cells that are not
intersected by a line are given the value NoData. If more than one line
is found in a cell, the cell is given the value of the first line encountered
while processing. Using a smaller cell size during conversion alleviates
this.
Polygons are used for buildings, forests, fields, and many other features
that are best represented by a series of connected cells. When you
convert polygons, the cells are given the value of the polygon found at
the center of each cell.
152 Using Conversion
Performing Features to
Raster Conversion
To perform features to raster conversion, follow these steps:
1. Click the Image Analysis dropdown arrow, point to Convert, and then
select Convert Features to Raster to open the Features to Raster
dialog.
2. Click the browse button for the Input features field and navigate to the
directory where the file is stored.
3. Select a field to use from the Field dropdown list.
4. Type an output cell size in the Output Cell Size field.
5. Click the browse button for the Output Image field and navigate to the
directory where you want the output image stored.
6. Click OK to close the Features to Raster dialog.
4
3
2
5
6
153 Applying GeoCorrection Tools
Applying GeoCorrection Tools 153
Applying GeoCorrection Tools
The tools and methods in this chapter describe the process of
geometrically correcting the distortions in images caused by sensors
and the curvature of the Earth. Even images of seemingly flat areas are
distorted, but you can correct, or rectify, these images so they are
represented on a planar surface, conform to other images, and have the
integrity of a map.
The terms geocorrection and rectification are used synonymously when
discussing geometric correction. Rectification is the process of
transforming data from one grid system into another grid system using
a geometric transformation. Because the pixels of a new grid may not
align with the pixels of the original grid, you must resample the pixels.
Resampling is the process of extrapolating data values for the pixels on
the new grid from the values of the source pixels.
Orthorectification is a form of rectification that corrects for terrain
displacement and is used if there is a DEM of the study area. It is based
on collinearity equations, which is derived by using 3D Ground GCPs.
In relatively flat areas, orthorectification is not necessary, but is
recommended in mountainous areas (or on aerial photographs of
buildings) where a high degree of accuracy is required.
IN THIS CHAPTER
Rectification
GeoCorrection
SPOT
Polynomial Transformation
Rubber Sheeting
Camera
IKONOS, QuickBird, and RPC Properties
Landsat
154 Applying GeoCorrection Tools
Rectification Rectification is necessary in cases where you must change the pixel
grid of an image to fit a map projection system or a reference image.
There are several reasons for rectifying image data:
Comparing pixels scene-to-scene in applications, such as change
detection or thermal inertia
Mapping (day and night comparison)
Developing GIS databases for GIS modeling
Identifying training samples according to map coordinates prior to
classification
Creating accurate, scaled photomaps
Overlaying an image with vector data
Comparing images that are originally at different scales
Extracting accurate distance and area measurements
Mosaicking images
Performing any other analyses requiring precise geographic
locations
Before rectifying the data, consider the primary use for the database
before selecting the optimum map projection and appropriate
coordinate system. If you are doing a government project, the projection
is often predetermined. A commonly used projection in the United
States government is state plane. Use an equal area projection for
thematic or distribution maps and conformal or equal area projections
for presentation maps.
Consider the following before selecting a map projection:
How large or small an area is mapped? Different projections are
intended for different size areas.
Where on the globe is the study area? Polar regions and equatorial
regions require different projections for maximum accuracy.
What is the extent of the study area? Circular, north-south, east-
west, and oblique areas may all require different projection systems
(ESRI 1992).
Applying GeoCorrection Tools 155
Disadvantages of
Rectification
During rectification, you must resample the data file values of rectified
pixels to fit into a new grid of pixel rows and columns. Although some of
the algorithms for calculating these values are highly reliable, you can
lose some spectral integrity of the data during rectification. If map
coordinates or map units are not needed in the application, it might be
wiser not to rectify the image. An unrectified image is more spectrally
correct than a rectified image.
Georeferencing Georeferencing refers to the process of assigning map coordinates to
image data. The image data might already project onto the plane, but
not reference the proper coordinate system. Rectification, by definition,
involves georeferencing, because all map projection systems are
associated with map coordinates. Image-to-image registration involves
georeferencing only if the reference image is already georeferenced.
Georeferencing, by itself, involves changing only the map coordinate
information in the image file. The grid of the image does not change.
Geocoded data are images that are rectified to a particular map
projection and pixel size, usually with radiometric corrections applied. It
is possible to purchase image data that is already geocoded. You
should rectify geocoded data only if it must conform to a different
projection system or be registered to other rectified data.
Georeferencing Only Rectification is unnecessary if there is no distortion in the image. For
example, if an image file is produced by scanning or digitizing a paper
map that is in the projection system you want, the image is already
planar and does not require rectification unless there is some skew or
rotation of the image. Scanning or digitizing produces images that are
planar, but do not contain any map coordinate information. You can
georeference these images, which is a much simpler process than
rectification. In many cases, you need only to update the image header
with new map coordinate information. This involves redefining:
The map coordinate of the upper-left corner of the image
The cell size (the area represented by each pixel)
This information is usually the same for each layer of an image file,
although it can be different. For example, the cell size of band 6 of
Landsat TM data is different from the cell size of the other bands.
Ground Control Points GCPs are specific pixels in an image for which the output map
coordinates (or other output coordinates) are known. GCPs consist of
two X,Y pairs of coordinates:
Source Coordinates Usually data file coordinates in the image
you are rectifying
Reference Coordinates The coordinates of the map or reference
image to which the source image is being registered
156 Applying GeoCorrection Tools
The term map coordinates is sometimes used loosely to apply to
reference coordinates and rectified coordinates. These coordinates are
not limited to map coordinates. For example, map coordinates are
unnecessary in image-to-image registration.
Entering GCP
Coordinates
Accurate GCPs are essential for an accurate rectification. From the
GCPs, the rectified coordinates for all other points in the image are
extrapolated. Select many GCPs throughout the scene. The more
dispersed the GCPs are, the more reliable the rectification. GCPs for
large-scale imagery might include the intersection of two roads, airport
runways, utility corridors, towers, or buildings. For small-scale imagery,
you can use larger features such as urban areas or geologic features.
Do not use landmarks that can vary (edges of lakes, other water bodies,
vegetation, and so on).
You can enter the source and reference coordinates of the GCPs in the
following ways:
Enter a known a priori at the keyboard.
Use your mouse to select a pixel from an image in the view. With
both the source and destination views open, enter source
coordinates and reference coordinates for image-to-image
registration.
Use a digitizing tablet to register an image to a hardcopy map.
Tolerance of RMSE Acceptable RMSE is determined by the end use of the database, the
type of data used, and the accuracy of the GCPs and ancillary data
used. For example, GCPs acquired from GPS should have an accuracy
of about 10 m, but GCPs from 1:24,000-scale maps should have an
accuracy of about 20 m.
It is important to remember that RMSE is reported in pixels. Therefore,
if you are rectifying Landsat TM data and want the rectification accurate
to within 30 meters, the RMSE should not exceed 1.00. Acceptable
accuracy depends on the image area and the project.
Classification Some analysts recommend classification before rectification because
the classification is then based on the original data values. Another
benefit is that a thematic file has only one band to rectify instead of the
multiple bands of a continuous file. On the other hand, it might benefit
you to rectify the data first, especially when using GPS data for the
GCPs. Because this data is very accurate, the classification might be
more accurate if the new coordinates help to locate better training
samples.
Thematic Files Nearest neighbor is the only appropriate resampling method for
thematic files, which is a drawback in some applications. The available
resampling methods are discussed in detail in Options Dialog on page
62.
Applying GeoCorrection Tools 157
Orthorectification Orthorectification is a form of rectification that corrects for terrain
displacement and is used if there is a DEM of the study area. It is based
on collinearity equations, which is derived by using 3D GCPs. In
relatively flat areas, orthorectification is not necessary, but in
mountainous areas (or on aerial photographs of building), where a high
degree of accuracy is required, orthorectification is recommended.
GeoCorrection You can perform georeferencing, which is the process of assigning map
coordinates to image data by using various sensor models, using the
GeoCorrection properties dialogs and the Add Links tool.
The GeoCorrection
Properties Dialog
The individual GeoCorrection tools have their own dialog. It displays
whenever you select a model type for an image on the Image Analysis
toolbar and click the GeoCorrection Properties button. Some of the
dialogs for these tools contain tabs pertaining to that specific tool, but
they all have several tabs in common. Every GeoCorrection tool dialog
has a General tab and a Links tab, and all but Polynomial properties and
Rubber Sheeting properties have an Elevation tab.
General Tab The General tab has the following settings:
Link Coloring Lets you set a threshold and select or change link
colors.
Displayed Units Lets you view the horizontal and vertical units if
they are known. Often only one is known so it might display Meters
for vertical units and Unknown for horizontal units. Display units do
not have an effect on the original data in latitude/longitude format.
The image in the view does not show the changes either.
To specify settings on the General tab, follow these steps:
1. Type a number in the Threshold field, and then click the Within
Threshold and Over Threshold color bar arrows to change the link
colors.
2. View the measurement of the vertical units in the Displayed Units box.
Links Tab The Links tab (this display is also called a CellArray) displays
information about the links in your image, including reference points
and RMSE. If you added links to your image, they are listed on this tab.
The program is interactive between the image and the Links tab, so
when you add links in an image or between two images, information is
automatically updated in the CellArray. You can edit and delete
information displayed in the CellArray. For example, if you want to
experiment with coordinates other than the ones youve been given,
you can plug your own coordinates into the CellArray.
158 Applying GeoCorrection Tools
Note: Before adding links or editing the links table, you must select the
coordinate system in which you want to store the link coordinates.
To select a coordinate system, follow these steps:
1. Right-click in the view and select Data Frame Properties to open the
Data Frame Properties dialog.
2. Click the Coordinate System tab.
3. Select the appropriate predefined coordinate system in the Select a
Coordinate System box if your link coordinates are predefined.
Note: Expand the Layers folder and select that layer if you want to use
the coordinate system from a specific layer.
4. Click OK to close the Data Frame Properties dialog.
3
4
2
Applying GeoCorrection Tools 159
To perform a few additional checks you need to make before
proceeding, follow these steps:
1. Make sure that the correct layer displays in the Layers dropdown list on
the Image Analysis toolbar.
2. Select your model type from the Model Types dropdown list.
3. Click the Add Links button to set your new links.
To proof and edit the coordinates of the links as you enter them, follow
these steps:
1. Click the GeoCorrection Properties button.
2. Click the Links tab.
The coordinates display in the CellArray on this tab.
3. Click in a cell and edit the contents.
4. Click the Export Links to Shapefile button and save the new shapefile.
1 2
3
2 3
4
160 Applying GeoCorrection Tools
Elevation Tab The Elevation tab is in all GeoCorrection Model properties dialogs
except for Polynomial and Rubber Sheeting. The default settings on the
Elevation tab let you choose a file to use as an elevation source.
Figure 44: Elevation Source File Settings
If you do not have an elevation file, click the Constant button to change
the settings in the Elevation Source box and specify the elevation value
and units. Use the constant value that is the average ground elevation
for the entire scene.
Figure 45: Elevation Source Constant Settings
Note: You can also check the Account for Earths curvature check box
if you want to use this option as part of the elevation.
Applying GeoCorrection Tools 161
The following steps take you through the Elevation tab. The first set of
instructions uses File as the elevation source. The second set uses
Constant as the elevation source.
To use a file value as the elevation source, follow these steps:
1. Click the File button in the Elevation Source box.
2. Type the name of the file in the Elevation File field, or navigate to the
directory where it is stored.
3. Select Feet or Meters from the Elevation Units dropdown list.
4. Check the Account for Earths Curvature check box.
5. Click Apply to set the elevation source.
6. Click OK to close the dialog.
To use a constant value as the elevation source, follow these steps:
1. Click the Constant button in the Elevation Source box.
2. Type the elevation value in the Elevation Value field.
3. Select Feet or Meters from the Elevation Units dropdown list.
4. Check the Account for Earths Curvature check box.
5. Click Apply to set the elevation source.
6. Click OK to close the dialog.
1
2
3
4
1
4
2
3
162 Applying GeoCorrection Tools
SPOT The first SPOT satellite, developed by the French Centre National
dEtudes Spatiales (CNES), was launched in early 1986. The second
SPOT satellite was launched in 1990, and the third was launched in
1993. The sensors operate in two modes, multispectral and
panchromatic. SPOT is commonly referred to as a pushbroom scanner,
which means that all scanning parts are fixed, and scanning is
accomplished by the forward motion of the scanner. SPOT pushes
3000/6000 sensors along its orbit. This is different from Landsat, which
scans using 16 detectors perpendicular to its orbit.
The SPOT satellite can observe the same area on the globe once every
26 days. The SPOT scanner normally produces nadir views, but it does
have off-nadir viewing capability. Off-nadir refers to any point that is not
directly beneath the detectors, but off to an angle. Using this off-nadir
capability, you can view one area on the Earth as often as every three
days.
Off-nadir viewing is programmable from the ground control station. It is
useful for collecting data in a region not directly in the path of the
scanner, or where timeliness of data acquisition is crucial in the event
of a natural or man-made disaster. It is also very useful in collecting
stereo data from which you can extract elevation data.
The width of the swath observed varies between 60 km for nadir viewing
and 80 km for off-nadir viewing at a height of 832 km (Jensen 1996).
Panchromatic SPOT Panchromatic (meaning sensitive to all visible colors) has 10
10 m spatial resolution, contains 1 band0.51 to 0.73 mmand is
similar to a black-and-white photograph. It has a radiometric resolution
of 8 bits (Jensen 1996).
Applying GeoCorrection Tools 163
XS SPOT XS, or multispectral, has 20 20 m spatial resolution, 8-bit
radiometric resolution, and contains 3 bands (Jensen 1996).
Table 6: SPOT XS Bands and Wavelengths
Figure 46: SPOT Panchromatic versus SPOT XS
Stereoscopic Pairs The panchromatic scanner can make two observations on successive
days, so that the two images are acquired at angles on either side of the
vertical, resulting in stereoscopic imagery. Stereoscopic imagery is also
achieved by using one vertical scene and one off-nadir scene. This type
of imagery can be used to produce a single image, or topographic and
planimetric maps (Jensen 1996).
Topographic maps indicate elevation. Planimetric maps correctly
represent horizontal distances between objects (Star and Estes 1990).
Band Wavelength
(Microns)
Comments
1, Green 0.50 to 0.59
m
This band corresponds to the green
reflectance of healthy vegetation.
2, Red 0.61 to 0.68
m
This band is useful for discriminating between
plant species. It is also useful for soil boundary
and geological boundary delineations.
3,
Reflective
IR
0.79 to 0.89
m
This band is especially responsive to the
amount of vegetation biomass present in a
scene. It is useful for crop identification and
emphasizes soil/crop and land/water contrasts.
P
a
n
chro
m
a
tic
X
S
1 Band
3 Bands
1 Pixel =
10 m x 10 m
1 Pixel =
20 m x 20 m
Radiometric
Resolution
0-255
164 Applying GeoCorrection Tools
SPOT 4 The SPOT 4 satellite was launched in 1998. SPOT 4 carries High
Resolution Visible Infrared (HR VIR) instruments that obtain information
in the visible and near-infrared spectral bands. It orbits the Earth at 822
km above the Equator and has two sensors on board: a multispectral
sensor, and a panchromatic sensor. The multispectral scanner has a
pixel size of 20 20 m, and a swath width of 60 km. The panchromatic
scanner has a pixel size of 10 10 m, and a swath width of 60 km.
Table 7: SPOT 4 Bands and Wavelengths
The Spot Properties
Dialog Box
In addition to the General, Links, and Elevation tabs, the Spot
Properties dialog also contains a Parameters tab. Most of the
GeoCorrection Properties dialogs contain a Parameters tab, but each
one offers different options.
To specify settings on the Parameters tab on the Spot Properties
dialog, follow these steps:
1. Select Spot from the Model Types dropdown list on the Image Analysis
toolbar.
2. Click the GeoCorrection Properties button to open the Spot Properties
dialog.
Band Wavelength
1, Green 0.50 to 0.59 m
2, Red 0.61 to 0.68 m
3, (near-IR) 0.78 to 0.89 m
4, (mid-IR) 1.58 to 1.75 m
Panchromatic 0.61 to 0.68 m
Applying GeoCorrection Tools 165
3. Click the Parameters tab.
4. Click the XS/XI or Pan button to specify the sensor type.
5. Type the number of iterations in the Number of Iterations field.
6. Type a number for the incidence angle in the Incidence Angle field.
7. Type the pixel value, which forms the background, in the Value field.
8. Type the number of the layer in the Layer field.
9. Click OK to close the Spot Properties dialog.
Polynomial
Transformation
Polynomial equations are used to convert source file coordinates to
rectified map coordinates. Depending on the distortion in the imagery,
complex polynomial equations might be required to express the needed
transformation. The degree of complexity of the polynomial is
expressed as the order of the polynomial. The order of transformation
is the order of the polynomial used in the transformation. Image
Analysis for ArcGIS allows 1st through nth order transformations.
Usually, 1st or 2nd order transformations are used.
4
7
5
3
9
6
8
166 Applying GeoCorrection Tools
Transformation Matrix A transformation matrix is computed from the GCPs. The matrix
consists of coefficients that are used in polynomial equations to convert
the coordinates. The size of the matrix depends on the order of
transformation. The goal in calculating the coefficients of the
transformation matrix is to derive the polynomial equations for which
there is the least possible amount of error when used to transform the
reference coordinates of the GCPs into the source coordinates. It is not
always possible to derive coefficients that produce no error. For
example, in the figure below, GCPs are plotted on a graph and
compared to the curve that is expressed by a polynomial.
Every GCP influences the coefficients, even if there isnt a perfect fit of
each GCP to the polynomial that the coefficients represent. The
distance between the GCP reference coordinate and the curve is called
RMSE, which is discussed later in Camera on page 177.
Linear Transformations A 1st order transformation is a linear transformation. It can change:
Location in X or Y
Scale in X or Y
Skew in X or Y
Rotation
You can use 1st order transformations to project raw imagery to a
planar map projection, convert a planar map projection to another
planar map projection, and rectify relatively small image areas. You can
perform simple linear transformations to an image in a view or to the
transformation matrix itself. Linear transformations may be required
before collecting GCPs for the displayed image. You can reorient
skewed Landsat TM data, rotate scanned quad sheets according to the
angle of declination stated in the legend, and rotate descending data so
that north is up.
Source X Coordinate
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
GCP
Polynomial Curve
Applying GeoCorrection Tools 167
You can also use a 1st order transformation for data already projected
onto a plane. For example, SPOT and Landsat Level 1B data is already
transformed to a plane, but might not be rectified to the map projection
you want. When doing this type of rectification, it is not advisable to
increase the order of transformation if a high RMSE occurs first.
Examine other factors, such as the GCP source and distribution, and
then look for systematic errors.
The transformation matrix for a 1st order transformation consists of six
coefficientsthree for each coordinate (X and Y):
Coefficients are used in a 1st order polynomial as follows:
Where:
x and y are source coordinates (input)
x
0
and y
0
are rectified coordinates (output)
The coefficients of the transformation matrix are as above.
Nonlinear
Transformations
You can use 2nd order transformations to convert Lat/Lon data to a
planar projection, for data covering a large area (to account for the
Earths curvature), and with distorted data (for example, due to camera
lens distortion). Use 3rd order transformations with distorted aerial
photographs, on scans of warped maps, and with radar imagery. You
can use 4th order transformations on very distorted aerial photographs.
The transformation matrix for a transformation of order t contains this
number of coefficients:
It is multiplied by two for the two sets of coefficientsone set for X and
one for Y.
An easier way to arrive at the same number is:
Clearly, the size of the transformation matrix increases with the order of
the transformation.
a
0
a
1
a
2
b
0
b
1
b
2
x
0
a
0
a
1
x a
2
y + + =
y
0
b
0
b
1
x b
2
y + + =
2 i
i 0 =
t 1 +
t 1 + ( )x t 2 + ( )
168 Applying GeoCorrection Tools
High-Order Polynomials The polynomial equations for a t order transformation take this form:
Where:
t is the order of the polynomial
a and b are coefficients
The subscript k in a and b is determined by:
Effects of Order The computation and output of a higher polynomial equation are more
complex than a lower-order polynomial equation. Therefore, you should
use higher-order polynomials to perform more complicated image
rectifications. It is helpful to see the output of various orders of
polynomials to understand the effects of different orders of
transformation in image rectification.
The following example uses only one coordinate (X) instead of the two
(X,Y) used in the polynomials for rectification. This lets you draw two-
dimensional graphs that illustrate the way higher orders of
transformation affect the output image. Because only the X coordinate
is used in these examples, the number of GCPs used is less than the
number required to perform the different orders of transformation.
Coefficients like those in this example are generally calculated by the
least squares regression method. Suppose there are GCPs with the
following X coordinates:
These GCPs allow a 1st order transformation of the X coordinates,
which is satisfied by this equation (the coefficients are in parentheses):
x
o
t
i o =
\ .
|
|
| |
i
j o =
\ .
|
|
| |
= a
k
x
i j
y
j
y
o
t
i o =
\ .
|
|
| |
i
j o =
\ .
|
|
| |
= b
k
x
i j
y
j
k
i i j +
2
------------------ j + =
Source X Coordinate
(Input)
Reference X Coordinate
(Output)
1 17
2 9
3 1
x
r
25 ( ) 8 ( )x
i
+ =
Applying GeoCorrection Tools 169
Where:
x
r
= Reference X coordinate
x
i
= Source X coordinate
This equation takes on the same format as the equation of a line
(y = mx + b). In mathematical terms, a 1st order polynomial is linear.
Therefore, a 1st order transformation is also known as a linear
transformation.
This equation is graphed below:
0 1 2 3 4
0
4
8
12
16
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
Source X Coordinate
x
r
= (25) + (-8)x
i
170 Applying GeoCorrection Tools
However, what if the second GCP were changed as follows?
These points are plotted against each other below:
A line cannot connect these points, which illustrates why they are not
expressed by a 1st order polynomial like the graph on the left. In this
case, a 2nd order polynomial equation expresses these points.
Polynomials of the 2nd order or higher are nonlinear. The graph of this
curve is drawn below:
Source X
Coordinate
(Input)
Reference X
Coordinate
(Output)
1 17
2 7
3 1
0 1 2 3 4
0
4
8
12
16
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
Source X Coordinate
x
r
31 ( ) 16 ( )x
i
2 ( )x
i
2
+ + =
0 1 2 3 4
0
4
8
12
16
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
Source X Coordinate
x
r
= (31) + (-16)
xi
+ (2)x
i
2
Applying GeoCorrection Tools 171
What if you added one more GCP to the list?
As illustrated in the graph above, this fourth GCP does not fit on the
curve of the 2nd order polynomial equation. You must increase the
order of the transformation to the 3rd order to ensure that all the GCPs
fit. The equation and graph below are possible results:
Source X
Coordinate
(Input)
Reference X
Coordinate
(Output)
1 17
2 7
3 1
4 5
0 1 2 3 4
0
4
8
12
16
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
Source X Coordinate
x
r
= (31) + (-16)
xi
+ (2)x
i
2
(4,5)
x
r
25 ( ) 5 ( )x
i
4 ( )x
i
2
1 ( )x
i
2
+ + + =
0 1 2 3 4
0
4
8
12
16
R
e
f
e
r
e
n
c
e
X
C
o
o
r
d
i
n
a
t
e
Source X Coordinate
x
r
= (25) + (-5)x
i
+ (-4)x
i
2
+ (1)x
i
3
172 Applying GeoCorrection Tools
The figure above illustrates a 3rd order transformation. However, this
equation may be unnecessarily complex. Performing a coordinate
transformation with this equation can cause unwanted distortions in the
output image for the sake of a perfect fit for all GCPs. In this example,
a 3rd order transformation probably is too high, because the output
pixels in the X direction are arranged in a different order than the input
pixels in the X direction.
In this case, a higher order of transformation probably does not produce
the results you want.
Source X
Coordinate
(Input)
Reference X
Coordinate
(Output)
1
2
3
4
x
0
1 ( ) 17 =
x
0
2 ( ) 7 =
x
0
3 ( ) 1 =
x
0
4 ( ) 5 =
x
0
1 ( ) x
0
2 ( ) x
0
4 ( ) x
0
3 ( ) > > >
17 7 5 1 > > >
1 2 3 4
1 2 3 4
Input Image
X Coordinates
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
3 4 2 1
Output Image
X Coordinates
Applying GeoCorrection Tools 173
Minimum Number of
GCPs
You can use higher orders of transformation to correct more
complicated types of distortion. However, more GCPs are needed to
use a higher order of transformation. For example, three points define
a plane. Therefore, to perform a 1st order transformation, which is
expressed by the equation of a plane, at least three GCPs are needed.
Similarly, the equation in a 2nd order transformation is the equation of
a paraboloid. Six points are required to define a paraboloid. Therefore,
at least six GCPs are required to perform a 2nd order transformation.
The minimum number of points required to perform a transformation of
order t equals:
Use more than the minimum number of GCPs whenever possible.
Although it is possible to get a perfect fit, it is rare, no matter how many
GCPs are used.
For 1st through 10th order transformations, the minimum number of
GCPs required to perform a transformation is listed in this table:
Table 8: Number of GCPs
Order of
Transformation
Minimum GCPs
Required
1 3
2 6
3 10
4 15
5 21
6 28
7 36
8 45
9 55
10 66
t 1 + ( ) t 2 + ( ) ( )
2
--------------------------------------
174 Applying GeoCorrection Tools
The Polynomial
Properties Dialog Box
The Polynomial Properties dialog has a Parameters tab in addition to
the General and Links tabs. It does not have an Elevation tab. The
General tab and the Links tab are the same as the ones featured at the
beginning of this chapter.
The Parameters tab contains a CellArray that shows the transformation
coefficients table. The table is populated when the model is solved.
To specify settings on the Polynomial Properties Parameters tab, follow
these steps:
1. Type the number for the polynomial order in the Polynomial Order field.
2. Click OK.
2
1
Applying GeoCorrection Tools 175
Rubber Sheeting Triangle-based finite element analysis is a powerful tool for solving
complicated computation problems that can be approached by small
simpler pieces. It is widely used as a local interpolation technique in
geographic applications. For image rectification, known control points
can be triangulated into many triangles. Each triangle has three control
points as its vertices. You can then use the polynomial transformation
to establish mathematical relationships between source and destination
systems for each triangle.
Because the transformation passes through each control point and is
not in a uniform manner, finite element analysis is also called rubber
sheeting. You can also call it the triangle-based rectification because
the transformation and resampling for image rectification are performed
on a triangle-by-triangle basis. Use this triangle-based technique when
other rectification methods such as polynomial transformation and
photogrammetric modeling cannot produce acceptable results.
Triangulation It is necessary to triangulate the control points into a mesh of triangles
to perform the triangle-based rectification. Watson (1994) summarily
listed four kinds of triangulation, including the Arbitrary, Optimal,
Greedy, and Delaunay triangulation. Of the four kinds, the Delaunay
triangulation is most widely used and is adopted because of the smaller
angle variations of the resulting triangles.
You can construct the Delaunay triangulation by the empty circumcircle
criterion. The circumcircle formed from three points of any triangle does
not have any other points inside. The triangles defined this way are the
most equiangular possible.
Triangle-Based
Rectification
Once the triangle mesh is generated and the spatial order of the control
points is available, you can perform the geometric rectification on a
triangle-by-triangle basis. This triangle-based method is appealing
because it breaks the region into smaller subsets. If the geometric
problem of the region is very complicated, the geometry of each subset
is much simpler and modeled through simple transformation.
For each triangle, you can use the polynomials as the general
transformation form between source and destination systems.
Linear Transformation The easiest and fastest transformation is the linear transformation with
the 1st order polynomials:
Additional information is not necessary because there are three known
conditions in each triangle and three unknown coefficients for each
polynomial.
xo a
0
a
1
x a
2
y + + =
yo b
0
b
1
x b
2
y + + =
j 0 =
i
i 0 =
5
=
y
0
b
k
x
i j
y
j
j 0 =
i
i 0 =
5