Taiga Manual
Taiga Manual
August 2009
J. Ronald Eastman
Clark Labs
Clark University
950 Main Street
Worcester, MA
01610-1477 USA
tel: +1-508-793-7526
fax: +1-508-793-8842
email: [email protected]
web: www.clarklabs.org
IDRISI Production
©1987-2009
Clark University
License Agreement
The software described in this document is furnished under a license, and may be used or copied only in accordance with
the terms of this license.
The IDRISI Taiga software described in this document is protected by the United States Copyright Law and International
Treaty provisions. This software remains the property of Clark Labs, Clark University. However, Clark Labs grants the
purchaser non-exclusive license to use this software subject to the terms outlined in this statement.
The purchaser of a single-user license of IDRISI Taiga is licensed to install and use the software on no more than one sin-
gle-user computer system. The purchaser is also permitted to make a backup copy of the IDRISI Taiga distribution media
for the sole purpose of protecting the purchaser’s investment from loss. The purchaser may not rent or lease the software,
but may transfer the license to another user upon written agreement from Clark Labs. The user may not reverse-engineer,
decompile, or disassemble the IDRISI Taiga software or any of its associated software programs contained on the distri-
bution media.
The PDF manuals that accompany this software are also protected by United States Copyright Law and International
Treaty provisions. All rights are reserved. No part of the manuals may be reproduced, stored in a retrieval system, or
transmitted, in any form or by any means, electronic, mechanical, photocopy, microfilm, recording, or otherwise without
written consent of Clark Labs. Direct translation of any part of the manuals is also prohibited without the written permis-
sion of the copyright holder.
Warranty
This software is sold “as is,” and the information in this manual is subject to change without notice. Furthermore, the
Clark Labs assumes no responsibility for any errors that may appear in this document, or in the software it describes.
CLARK LABS DISCLAIMS ALL OTHER WARRANTIES, EXPRESSED OR IMPLIED, INCLUDING BUT NOT
LIMITED TO, IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PUR-
POSE WITH RESPECT TO DEFECTS IN THE CD-ROM AND/OR DOCUMENTATION, AND THE PRO-
GRAM LICENSE GRANTED HEREIN IN PARTICULAR, AND WITHOUT LIMITING OPERATION OF THE
PROGRAM LICENSE WITH RESPECT TO ANY PARTICULAR APPLICATION, USE, OR PURPOSE. IN NO
EVENT SHALL CLARK LABS BE LIABLE FOR ANY LOSS OF PROFIT OR ANY OTHER COMMERCIAL
DAMAGE, INCLUDING BUT NOT LIMITED TO SPECIAL, INCIDENTAL, CONSEQUENTIAL OR OTHER
DAMAGES.
Trademarks
IDRISI and CartaLinx are registered trademarks of Clark University. Windows and Access are trademarks of Microsoft
Corporation. ArcInfo and ArcView are trademarks of Environmental Systems Research Institute, Inc. MapInfo is a regis-
tered trademark of MapInfo Corporation. Adobe and the Adobe logo are trademarks of Adobe Systems Incorporated.
Acrobat Reader Copyright 1987-2009 Adobe Systems Incorporated. All rights reserved. GfK Macon’s digital maps are
protected by copyright (Copyright GfK Macon GmbH). All other product names mentioned in this volume may be trade-
Introduction
Thank you for choosing IDRISI Taiga and welcome to Clark Labs.
IDRISI is the industry leader in raster analytical functionality, covering the full spectrum of GIS and remote sensing needs
from database query, to spatial modeling, to image enhancement and classification. Special facilities are included for envi-
ronmental monitoring and natural resource management, including land change modeling and time series analysis, multi-
criteria and multi-objective decision support, uncertainty and risk analysis, simulation modeling, surface interpolation and
statistical characterization. Yet, despite the highly sophisticated nature of these capabilities, the system is very easy to use.
IDRISI consists of a main interface program (containing the menu and toolbar system) and a collection of nearly 300 pro-
gram modules that provide facilities for the input, display and analysis of geographic data. See the IDRISI Modules
chapter for an overview of the menu structure and a listing of all modules and their capabilities. Detailed information
about each module, as well as a variety of other technical information, may be found in the on-line Help System.
Along with this manual, the IDRISI software package also includes a set of tutorial exercises and data that guide the new
user through the concepts of GIS and Image Processing while also introducing the features of IDRISI. The tutorial exer-
cises are appropriate for use in either self-training or in classroom settings.
Components of a GIS
Although we think of a GIS as a single piece of software, it is typically made up of a variety of different components. Fig-
ure 2-1 gives a broad overview of the software components typically found in a GIS. Not all systems have all of these ele-
ments, but to be a true GIS, an essential group must be found.
Images
SpatialAttribute
Data Data
Base Base
Geographic Database
Analysis Manage-
System ment
System
Cartographic Statistics
Display
System Tabular Data
Maps
Figure 1
Vector
With vector representation, the boundaries or the course of the features are defined by a series of points that, when joined
with straight lines, form the graphic representation of that feature. The points themselves are encoded with a pair of num-
bers giving the X and Y coordinates in systems such as latitude/longitude or Universal Transverse Mercator grid coordi-
nates. The attributes of features are then stored with a traditional database management (DBMS) software program. For
example, a vector map of property parcels might be tied to an attribute database of information containing the address,
owner's name, property valuation and landuse. The link between these two data files can be a simple identifier number
that is given to each feature in the map (Figure 2).
Raster
The second major form of representation is known as raster. With raster systems, the graphic representation of features
and the attributes they possess are merged into unified data files. In fact, we typically do not define features at all. Rather,
the study area is subdivided into a fine mesh of grid cells in which we record the condition or attribute of the earth's sur-
face at that point (Figure 2). Each cell is given a numeric value which may then represent either a feature identifier, a qual-
itative attribute code or a quantitative attribute value. For example, a cell could have the value "6" to indicate that it
belongs to District 6 (a feature identifier), or that it is covered by soil type 6 (a qualitative attribute), or that it is 6 meters
above sea level (a quantitative attribute value). Although the data we store in these grid cells do not necessarily refer to
phenomena that can be seen in the environment, the data grids themselves can be thought of as images or layers, each
depicting one type of information over the mapped region. This information can be made visible through the use of a ras-
ter display. In a raster display, such as the screen on your computer, there is also a grid of small cells called pixels. The word
pixel is a contraction of the term picture element. Pixels can be made to vary in their color, shape or grey tone. To make an
image, the cell values in the data grid are used to regulate directly the graphic appearance of their corresponding pixels.
Thus in a raster system, the data directly controls the visible form we see.
1. The basic data structure of vector systems can best be described as a network. As a result, it is not surprising to find that vector systems have excellent
capabilities for the analysis of network space. Thus the difference between raster and vector is less one of inherent ability than one of the difference in
the types of space they describe.
roads
soils
elevation
Figure 3
Raster systems also use this map-like logic, but usually divide data sets into unitary layers. A layer contains all the data for a
single attribute. Thus one might have a soils layer, a roads layer and a landuse layer. A few raster systems, including
IDRISI, can link a feature identifier layer (a layer that contains the identifiers of the features located at each grid cell) with
attribute tables. More commonly, separate layers exist for each attribute and on-screen displays and paper maps are pro-
duced from these, either singly or in combination.
Although there are subtle differences, for all intents and purposes, raster layers and vector coverages can be thought of as
simply different manifestations of the same concept—the organization of the database into elementary map-like themes.
Layers and coverages differ from traditional paper maps, however, in an important way. When map data are encoded in
digital form (digitized), scale differences are removed. The digital data may be displayed or printed at any scale. More
importantly, digital data layers that were derived from paper maps of different scales, but covering the same geographic
area, may be combined.
In addition, many GIS packages, including IDRISI, provide utilities for changing the projection and reference system of
digital layers. This allows multiple layers, digitized from maps having various projections and reference systems, to be con-
verted to a common system.
With the ability to manage differences of scale, projection and reference system, layers can be merged with ease, eliminat-
ing a problem that has traditionally hampered planning activities with paper maps. It is important to note, however, that
the issue of resolution of the information in the data layers remains. Although features digitized from a poster-sized world
map could be combined in a GIS with features digitized from a very large scale local map, such as a city street map, this
would normally not be done. The level of accuracy and detail of the digital data can only be as good as that of the original
maps.
Georeferencing
All spatial data files in a GIS are georeferenced. Georeferencing refers to the location of a layer or coverage in space as
defined by a known coordinate referencing system. With raster images, a common form of georeferencing is to indicate
the reference system (e.g., latitude/longitude), the reference units (e.g., degrees) and the coordinate positions of the left,
right, top and bottom edges of the image. The same is true of vector data files, although the left, right, top and bottom
edges now refer to what is commonly called the bounding rectangle of the coverage—a rectangle which defines the limits of
Analysis in GIS
The organization of the database into layers is not simply for reasons of organizational clarity. Rather, it is to provide rapid
access to the data elements required for geographic analysis. Indeed, the raison d'être for GIS is to provide a medium for
geographic analysis.
The analytical characteristics of GIS can be looked at in two ways. First, one can look at the tools that GIS provides. Then
one can look at the kinds of operations that GIS allows. Regardless of whether we are using a raster or a vector system, we
will tend to find that the tools fall into four basic groups and that the operations undertaken fall into three.
Analytical Tools
Database Query
The most fundamental of all tools provided by a GIS are those involved with Database Query. Database query simply
asks questions about the currently-stored information. In some cases, we query by location—what landuse is at this location?
In other cases, we query by attribute—what areas have high levels of radon gas? Sometimes we undertake simple queries such as
those just illustrated, and at other times we ask about complex combinations of conditions—show me all wetlands that are
larger than 1 hectare and that are adjacent to industrial lands.
In most systems, including IDRISI, these query operations are undertaken in two steps. The first step, called a reclassifica-
tion, creates a new layer of each individual condition of interest (Figure 2-4). For example, consider a query to find residen-
tial areas on bedrock associated with high levels of radon gas. The first step would be to create a layer of residential areas
alone by reclassifying all landuse codes into only two—a 1 whenever an area is residential and a 0 for all other cases. The
resulting layer is known as a Boolean layer since it shows only those areas that meet the condition (1 = true, residential) and
those that don't (0 = false, not residential). Boolean layers are also called logical layers since they show only true/false rela-
tionships. They are also sometimes called binary layers since they contain only zeros and ones. We will avoid using that
term, however, since it also describes a particular kind of data storage format. Here we will call them Boolean layers.
Once the residential layer has been created, a geology layer is then also reclassified to create a Boolean layer showing areas
with bedrock associated with high levels of radon gas. At this point we can combine the two conditions using an overlay
operation (Figure 4). As mentioned previously, it is only a GIS that can combine conditions such as this that involve fea-
tures with different geographies. Typically, an overlay operation in GIS will allow the production of new layers based on
some logical or mathematical combination of two or more input layers. In the case of database query, the key logical oper-
ations of interest are the AND and OR relational operators, also known as the INTERSECTION and UNION opera-
tions respectively. Here we are looking for cases of residential land AND high radon gas—the logical intersection of our
2. The bounding rectangle is defined by the study region of interest and does not necessarily refer to the actual minimum and maximum coordinates in
the data file.
Soils Boundaries
Reclassification
Overlay
Figure 4
Map Algebra
The second set of tools that a GIS will typically provide is that for combining map layers mathematically. Modeling in par-
ticular requires the ability to combine layers according to various mathematical equations. For example, we might have an
equation that predicts mean annual temperature as a result of altitude. Or, as another example, consider the possibility of
creating a soil erosion potential map based on factors of soil erodability, slope gradient and rainfall intensity. Clearly we
need the ability to modify data values in our map layers by various mathematical operations and transformations and to
combine factors mathematically to produce the final result.
The Map Algebra tools will typically provide three different kinds of operations:
1. the ability to mathematically modify the attribute data values by a constant (i.e., scalar arithmetic);
2. the ability to mathematically transform attribute data values by a standard operation (such as the trigonomet-
ric functions, log transformations and so on);
3. the ability to mathematically combine (such as add, subtract, multiply, divide) different data layers to produce
a composite result.
This third operation is simply another form of overlay—mathematical overlay, as opposed to the logical overlay of data-
base query.
To illustrate this, consider a model for snow melt in densely forested areas:3
M = (0.19T + 0.17D)
where M is the melt rate in cm/day, T is the air temperature and D is the dewpoint temperature. Given layers of the air
temperatures and dewpoints for a region of this type, we could clearly produce a snow melt rate map. To do so would
require multiplying the temperature layer by 0.19 (a scalar operation), the dewpoint layer by 0.17 (another scalar opera-
3. Equation taken from Dunne, T., and Leopold, L.B., (1978) Water in Environmental Planning, (W.H. Freeman and Co.: San Francisco), 480.
Distance Operators
The third tool group provided by GIS consists of the Distance Operators. As the name suggests, these are a set of tech-
niques where distance plays a key role in the analysis undertaken. Virtually all systems provide the tools to construct buf-
fer zones—areas within a specified distance of designated target features. Some can also evaluate the distance of all
locations to the nearest of a set of designated features, while others can even incorporate frictional effects and barriers in
distance calculations (Figure 5).
When frictional effects are incorporated, the distance calculated is often referred to as a cost distance. This name is used
because movement through space can be considered to incur costs, either in money, time or effort. Frictions increase
those costs. When the costs of movement from one or more locations are evaluated for an entire region, we often refer to
the result as a cost surface (Figure 5). In this case, areas of low cost (presumably near to the starting point) can be seen as
valleys and areas of high cost as hills. A cost surface thus has its lowest point(s) at the starting location(s) and its highest
point(s) at the locations that are farthest away (in the sense of the greatest accumulated cost).4
Figure 5
There may be cases in which frictions do not affect the cost of movement the same way in all directions. In other words,
they act anisotropically. For example, going up a steep slope might incur a cost that would be higher than going down the
same steep slope. Thus the direction of movement through the friction is important, and must be taken into account
when developing the cost surface. IDRISI provides modules to model this type of cost surface which are explained in
detail in the Anisotropic Cost Analysis chapter.
Given the concept of a cost surface, Geographic Information Systems also commonly offer least-cost path analysis—another
important distance operation. As the name suggests, our concern is to evaluate the least-cost path between two locations.
The cost surface provides the needed information for this to be evaluated (Figure 5).
Regardless of how distance is evaluated, by straight line distance or by cost distance, another commonly provided tool is
allocation. With allocation, we assign locations to the nearest of a set of designated features. For example, we might estab-
lish a set of health facilities and then wish to allocate residents to their nearest facility, where "near" might mean linear dis-
tance, or a cost distance such as travel time.
4. It should be noted here that a cost surface as just described can only be evaluated with a raster system. For vector systems, the closest equivalent
would be cost distances evaluated over a network. Here we see a simple, but very powerful, illustration of the differences between raster and vector sys-
tems in how they conceive of space.
Figure 6
A second good example of a context operator is a digital filter. Digital filters operate by changing values according to the
character of neighboring values. For example, a surface of terrain heights can be smoothed by replacing values with the
average of the original height and all neighboring heights. Digital filters have a broad range of applications in GIS and
remote sensing, ranging from noise removal to the visual enhancement of images.
Because of their simple and uniform data structure, raster systems tend to offer a broad range of context operators. In
IDRISI, for example, these include surface analysis and digital filtering, identification of contiguous areas, watershed anal-
ysis, viewshed analysis (an evaluation of all areas in view of one or more designated features) and a special supply/demand
modeling procedure where demands are satisfied by taking supplies in a radial fashion from neighboring locations.
Analytical Operations
Given these basic tools, a broad range of analytical operations can be undertaken. However, it would appear that most of
these fall into one of three basic groups: Database Query, Derivative Mapping and Process Modeling.
Database Query
With database query, we are simply selecting out various combinations of variables for examination. The tools we use are
largely the database query tools previously discussed (hence the name), but also include various measurement and statisti-
cal analysis procedures. The key thing that distinguishes this kind of analysis is that we have taken out no more than we
have put into the system. While we may extract combinations we have never examined before, the system provides us
with no new information—we are simply making a withdrawal from a data bank we have built up.
One of the key activities in database query is pattern seeking. Typically we are looking for spatial patterns in the data that
may lead us to hypothesize about relationships between variables.
Process Modeling
Database query and derivative mapping make up the bulk of GIS analysis undertaken today. However, there is a third area
that offers incredible potential—Process or Simulation Modeling.
With process modeling, we also bring something new to the database—knowledge of process. Process refers to the causal
chain by which some event takes place. For example, a simple model of fuel-wood demand satisfaction might run as fol-
lows:
1. Take all the wood you need (if you can) from your present location.
2. If your demand is satisfied or if you have traveled more than 10 kilometers from home, go to step 4.
3. If your demand is not met, move to an immediately adjacent location not already visited and repeat step 1.
4. Stop.
Process modeling is a particularly exciting prospect for GIS. It is based on the notion that in GIS, our database doesn't
simply represent an environment, it is an environment! It is a surrogate environment, capable of being measured, manipu-
lated and acted upon by geographic and temporal processes. Our database thus acts as a laboratory for the exploration of
processes in a complex environment. Traditionally, in science, we have had to remove that complexity in order to under-
stand processes in isolation. This has been an effective strategy and we have learned much from it. However, technologies
such as GIS now offer the tools to reassemble those simple understandings in order to gain an understanding and appre-
ciation of how they act in the full complexity of a real environmental setting. Often even very simple understandings yield
complex patterns when allowed to interact in the environment.
A different sort of process, the decision making process, may also be supported and in some ways modeled with the use
of GIS. GIS technology is becoming more important as a tool for decision support. Indeed, even the simplest database
query results may prove to be invaluable input to the decision maker. However, the more complex process of decision mak-
ing, in which decision makers often think in terms of multiple criteria, soft boundaries (non-Boolean) and levels of
acceptable risk, may also be modeled using GIS. IDRISI provides a suite of decision support modules to help decision
Definition
Remote sensing can be defined as any process whereby information is gathered about an object, area or phenomenon
without being in contact with it. Our eyes are an excellent example of a remote sensing device. We are able to gather infor-
mation about our surroundings by gauging the amount and nature of the reflectance of visible light energy from some
external source (such as the sun or a light bulb) as it reflects off objects in our field of view. Contrast this with a thermom-
eter, which must be in contact with the phenomenon it measures, and thus is not a remote sensing device.
Given this rather general definition, the term remote sensing has come to be associated more specifically with the gauging of
interactions between earth surface materials and electromagnetic energy. However, any such attempt at a more specific
definition becomes difficult, since it is not always the natural environment that is sensed (e.g., art conservation applica-
tions), the energy type is not always electromagnetic (e.g., sonar) and some procedures gauge natural energy emissions
(e.g., thermal infrared) rather than interactions with energy from an independent source.
Fundamental Considerations
Energy Source
Sensors can be divided into two broad groups—passive and active. Passive sensors measure ambient levels of existing
sources of energy, while active ones provide their own source of energy. The majority of remote sensing is done with pas-
sive sensors, for which the sun is the major energy source. The earliest example of this is photography. With airborne
cameras we have long been able to measure and record the reflection of light off earth features. While aerial photography
is still a major form of remote sensing, newer solid state technologies have extended capabilities for viewing in the visible
and near infrared wavelengths to include longer wavelength solar radiation as well. However, not all passive sensors use
energy from the sun. Thermal infrared and passive microwave sensors both measure natural earth energy emissions. Thus
G reen
R ed
Blue
UV Near - infrared
Visible
(1mm) (1m)
Wavelength (μm)
10 −6 10 −5 10 −4 10 −3 10 −2 10 −1 1 10 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 Wavelength
(μm)
C M
id The Te
γr
M
os
N ible let
U
V irav
X
an lev
ea
lt
ic
ay
m s io -I rm
ra
ro
R a
r- U V
ic d isio
s
ys
R
w
ad
ra lI
IR
av
ys n
io
e
a
R
(
)
Wavelength
As indicated, most remote sensing devices make use of electromagnetic energy. However, the electromagnetic spectrum is
very broad and not all wavelengths are equally effective for remote sensing purposes. Furthermore, not all have significant
interactions with earth surface materials of interest to us. Figure 3-1 illustrates the electromagnetic spectrum. The atmo-
sphere itself causes significant absorption and/or scattering of the very shortest wavelengths. In addition, the glass lenses
of many sensors also cause significant absorption of shorter wavelengths such as the ultraviolet (UV). As a result, the first
significant window (i.e., a region in which energy can significantly pass through the atmosphere) opens up in the visible
wavelengths. Even here, the blue wavelengths undergo substantial attenuation by atmospheric scattering, and are thus
often left out in remotely sensed images. However, the green, red and near infrared (IR) wavelengths all provide good
opportunities for gauging earth surface interactions without significant interference by the atmosphere. In addition, these
regions provide important clues to the nature of many earth surface materials. Chlorophyll, for example, is a very strong
absorber of red visible wavelengths, while the near infrared wavelengths provide important clues to the structures of plant
leaves. As a result, the bulk of remotely sensed images used in GIS-related applications are taken in these regions.
Extending into the middle and thermal infrared regions, a variety of good windows can be found. The longer of the mid-
dle infrared wavelengths have proven to be useful in a number of geological applications. The thermal regions have
proven to be very useful for monitoring not only the obvious cases of the spatial distribution of heat from industrial activ-
ity, but a broad set of applications ranging from fire monitoring to animal distribution studies to soil moisture conditions.
After the thermal IR, the next area of major significance in environmental remote sensing is in the microwave region. A
number of important windows exist in this region and are of particular importance for the use of active radar imaging.
The texture of earth surface materials causes significant interactions with several of the microwave wavelength regions.
This can thus be used as a supplement to information gained in other wavelengths, and also offers the significant advan-
tage of being usable at night (because as an active system it is independent of solar radiation) and in regions of persistent
cloud cover (since radar wavelengths are not significantly affected by clouds).
Light Source
reflection
absorption transmission
Figure 3-2
B G R B G R B G R
purple white dark gray
Figure 3-3
The eye is able to sense spectral response patterns because it is truly a multi-spectral sensor (i.e., it senses in more than
one place in the spectrum). Although the actual functioning of the eye is quite complex, it does in fact have three separate
types of detectors that can usefully be thought of as responding to the red, green and blue wavelength regions. These are
the additive primary colors, and the eye responds to mixtures of these three to yield a sensation of other hues. For example,
the color perceived by the third spectral response pattern in Figure 3-3 would be a yellow—the result of mixing a red and
green. However, it is important to recognize that this is simply our phenomenological perception of a spectral response
pattern. Consider, for example, the fourth curve. Here we have reflectance in both the blue and red regions of the visible
spectrum. This is a bimodal distribution, and thus technically not a specific hue in the spectrum. However, we would per-
ceive this to be a purple! Purple (a color between violet and red) does not exist in nature (i.e., as a hue—a distinctive dom-
inant wavelength). It is very real in our perception, however. Purple is simply our perception of a bimodal pattern
involving a non-adjacent pair of primary hues.
In the early days of remote sensing, it was believed (more correctly hoped) that each earth surface material would have a
distinctive spectral response pattern that would allow it to be reliably detected by visual or digital means. However, as our
common experience with color would suggest, in reality this is often not the case. For example, two species of trees may
have quite a different coloration at one time of the year and quite a similar one at another.
Finding distinctive spectral response patterns is the key to most procedures for computer-assisted interpretation of
remotely sensed imagery. This task is rarely trivial. Rather, the analyst must find the combination of spectral bands and the
time of year at which distinctive patterns can be found for each of the information classes of interest.
For example, Figure 3-4 shows an idealized spectral response pattern for vegetation along with those of water and dry
bare soil. The strong absorption by leaf pigments (particularly chlorophyll for purposes of photosynthesis) in the blue and
red regions of the visible portion of the spectrum leads to the characteristic green appearance of healthy vegetation. How-
ever, while this signature is distinctively different from most non-vegetated surfaces, it is not very capable of distinguishing
between species of vegetation—most will have a similar color of green at full maturation. In the near infrared, however,
we find a much higher return from vegetated surfaces because of scattering within the fleshy mesophyllic layer of the
leaves. Plant pigments do not absorb energy in this region, and thus the scattering, combined with the multiplying effect
of a full canopy of leaves, leads to high reflectance in this region of the spectrum. However, the extent of this reflectance
will depend highly on the internal structure of leaves (e.g., broadleaf versus needle). As a result, significant differences
between species can often be detected in this region. Similarly, moving into the middle infrared region we see a significant
dip in the spectral response pattern that is associated with leaf moisture. This is, again, an area where significant differ-
ences can arise between mature species. Applications looking for optimal differentiation between species, therefore, will
typically involve both the near and middle infrared regions and will use imagery taken well into the development cycle.
Relative Reflectance
0.4 0.8 1.2 1.6 2.0 2.4
Wavelength (μm)
Figure 3-5
Sensor/Platform Systems
Given recent developments in sensors, a variety of platforms are now available for the capture of remotely sensed data.
Here we review some of the major sensor/platform combinations that are typically available to the GIS user community.
Color Photography
Normal color photographs are produced from a composite of three film layers with intervening filters that act to isolate,
in effect, red, green, and blue wavelengths separately to the different film layers. With color infrared film, these wave-
lengths are shifted to the longer wavelengths to produce a composite that has isolated reflectances from the green, red
and near infrared wavelength regions. However, because the human eye cannot see infrared, a false color composite is
produced by making the green wavelengths appear blue, the red wavelengths appear green, and the infrared wavelengths
Aerial Videography
Light, portable, inexpensive video cameras and recorders can be carried in chartered aircraft. In addition, a number of
smaller aerial mapping companies offer videography as an output option. By using several cameras simultaneously, each
with a filter designed to isolate a specific wavelength range, it is possible to isolate multispectral image bands that can be
used individually, or in combination in the form of a color composite. For use in digital analysis, special graphics hardware
boards known as frame grabbers can be used to freeze any frame within a continuous video sequence and convert it to digi-
tal format, usually in one of the more popular exchange formats such as TIF or TARGA. Like small-format photography,
aerial videography cannot be used for detailed mapping, but provides a useful overview for reconnaissance surveys, and
can be used in conjunction with ground point sampling.
Image Restoration
Remotely sensed images of the environment are typically taken at a great distance from the earth's surface. As a result,
there is a substantial atmospheric path that electromagnetic energy must pass through before it reaches the sensor.
Depending upon the wavelengths involved and atmospheric conditions (such as particulate matter, moisture content and
turbulence), the incoming energy may be substantially modified. The sensor itself may then modify the character of that
data since it may combine a variety of mechanical, optical and electrical components that serve to modify or mask the
measured radiant energy. In addition, during the time the image is being scanned, the satellite is following a path that is
subject to minor variations at the same time that the earth is moving underneath. The geometry of the image is thus in
constant flux. Finally, the signal needs to be telemetered back to earth, and subsequently received and processed to yield
the final data we receive. Consequently, a variety of systematic and apparently random disturbances can combine to
degrade the quality of the image we finally receive. Image restoration seeks to remove these degradation effects.
Broadly, image restoration can be broken down into the two sub-areas of radiometric restoration and geometric restoration.
Radiometric Restoration
Radiometric restoration refers to the removal or diminishment of distortions in the degree of electromagnetic energy reg-
istered by each detector. A variety of agents can cause distortion in the values recorded for image cells. Some of the most
common distortions for which correction procedures exist include:
uniformly elevated values, due to atmospheric haze, which preferentially scatters short wavelength bands (particularly
the blue wavelengths);
striping, due to detectors going out of calibration;
random noise, due to unpredictable and unsystematic performance of the sensor or transmission of the data; and
Geometric Restoration
For mapping purposes, it is essential that any form of remotely sensed imagery be accurately registered to the proposed
map base. With satellite imagery, the very high altitude of the sensing platform results in minimal image displacements
due to relief. As a result, registration can usually be achieved through the use of a systematic rubber sheet transformation
process1 that gently warps an image (through the use of polynomial equations) based on the known positions of a set of
widely dispersed control points. This capability is provided in IDRISI through the module RESAMPLE.
With aerial photographs, however, the process is more complex. Not only are there systematic distortions related to tilt
and varying altitude, but variable topographic relief leads to very irregular distortions (differential parallax) that cannot be
removed through a rubber sheet transformation procedure. In these instances, it is necessary to use photogrammetric rec-
tification to remove these distortions and provide accurate map measurements2. Failing this, the central portions of high
altitude photographs can be resampled with some success.
RESAMPLE is a module of major importance, and it is essential that one learn to use it effectively. Doing so also requires
a thorough understanding of reference systems and their associated parameters such as datums and projections. The
chapter on Georeferencing provides an in-depth discussion of these issues.
Image Enhancement
Image enhancement is concerned with the modification of images to make them more suited to the capabilities of human
vision. Regardless of the extent of digital intervention, visual analysis invariably plays a very strong role in all aspects of
remote sensing. While the range of image enhancement techniques is broad, the following fundamental issues form the
backbone of this area:
Contrast Stretch
Digital sensors have a wide range of output values to accommodate the strongly varying reflectance values that can be
found in different environments. However, in any single environment, it is often the case that only a narrow range of val-
ues will occur over most areas. Grey level distributions thus tend to be very skewed. Contrast manipulation procedures are
thus essential to most visual analyses. Figure 3-6 shows TM Band 3 (visible red) and its histogram. Note that the values of
the image are quite skewed. The right image of the figure shows the same image band after a linear stretch between values
12 and 60 has been applied. In IDRISI, this type of contrast enhancement may be performed interactively through Com-
poser’s Layer Properties while the image is displayed. This is normally used for visual analysis only—original data values
1. Satellite-based scanner imagery contains a variety of inherent geometric problems such as skew (caused by rotation of the earth underneath the satel-
lite as it is in the process of scanning a complete image) and scanner distortion (caused by the fact that the instantaneous field of view (IFOV) covers
more territory at the ends of scan lines, where the angle of view is very oblique, than in the middle). With commercially-marketed satellite imagery, such
as Landsat, IRS and SPOT, most elements of systematic geometric restoration associated with image capture are corrected by the distributors of the
imagery. Thus, for the end user, the only geometric operation that typically needs to be undertaken is a rubber-sheet resampling in order to rectify the
image to a map base. Many commercial distributors will perform this rectification for an additional fee.
2. Photogrammetry is the science of taking spatial measurements from aerial photographs. In order to provide a full rectification, it is necessary to have
stereoscopic images—photographs which overlap enough (e.g., 60% in the along-track direction and 10% between flight lines) to provide two independent
images of each part of the landscape. Using these stereoscopic pairs and ground control points of known position and height, it is possible to fully re-
create the geometry of the viewing conditions, and thereby not only rectify measurements from such images, but also derive measurements of terrain
height. The rectified photographs are called orthophotos. The height measurements may be used to produce digital elevation models.
Linear Stretch
Figure 3-6
Composite Generation
For visual analysis, color composites make fullest use of the capabilities of the human eye. Depending upon the graphics
system in use, composite generation ranges from simply selecting the bands to use, to more involved procedures of band
combination and associated contrast stretch. Figure 3-7 shows several composites made with different band combinations
from the same set of TM images. (See Figure 3-5 for TM band definitions.) The IDRISI module COMPOSITE is used to
construct three-band 24-bit composite images for visual analysis.
Digital Filtering
One of the most intriguing capabilities of digital analysis is the ability to apply digital filters. Filters can be used to provide
edge enhancement (sometimes called crispening), to remove image blur, and to isolate lineaments and directional trends, to
mention just a few. The IDRISI module FILTER is used to apply standard filters and to construct and apply user-defined
filters.
Pansharpening
Pansharpening is the process of merging lower resolution multispectral imagery with the higher resolution panchromatic
image. Typically, the panchromatic band associated with most systems, e.g., SPOT, IKONOS or QuickBird, are captured
across the visible range of the spectrum affording them a higher resolution, commensurately giving them better detail in
shape and texture. But what they gain in clarity penalizes them in regards to their spectral properties, unlike the multispec-
tral bands. Merging the two results in increasing the resolution of the multispectral imagery while preserving its spectral
Figure 3-8
In IDRISI the module PANSHARPEN is used to pansharpen multispectral images using the panchromatic band. Three
methods are available: color space transformation, principal components analysis, and local regression.
Image Classification
Image classification refers to the computer-assisted interpretation of remotely sensed images. The procedures involved
are treated in detail in the chapter titled Classification of Remotely Sensed Imagery. This section provides a brief over-
view.
Although some procedures are able to incorporate information about such image characteristics as texture and context,
the majority of image classification is based solely on the detection of the spectral signatures (i.e., spectral response pat-
terns) of landcover classes. The success with which this can be done will depend on two things: 1) the presence of distinc-
tive signatures for the landcover classes of interest in the band set being used; and 2) the ability to reliably distinguish
these signatures from other spectral response patterns that may be present.
There are two general approaches to image classification: supervised and unsupervised. They differ in how the classification is
performed. In the case of supervised classification, the software system delineates specific landcover types based on sta-
tistical characterization data drawn from known examples in the image (known as training sites). With unsupervised clas-
sification, however, clustering software is used to uncover the commonly occurring landcover types, with the analyst
providing interpretations of those cover types at a later stage.
Supervised Classification
The first step in supervised classification is to identify examples of the information classes (i.e., landcover types) of inter-
est in the image. These are called training sites. The software system is then used to develop a statistical characterization of
the reflectances for each information class. This stage is often called signature analysis and may involve developing a charac-
terization as simple as the mean or the range of reflectances on each band, or as complex as detailed analyses of the mean,
variances and covariances over all bands.
Once a statistical characterization has been achieved for each information class, the image is then classified by examining
the reflectances for each pixel and making a decision about which of the signatures it resembles most. There are several
techniques for making these decisions, called classifiers. Most image processing software will offer several, based on varying
decision rules. IDRISI offers a wide range of options falling into three groups, depending upon the nature of the output
desired and the nature of the input bands.
Hard Classifiers
The distinguishing characteristic of hard classifiers is that they all make a definitive decision about the landcover class to
which any pixel belongs. IDRISI offers a host of supervised classifiers in this group. Some like parallelepiped (PIPED),
minimum distance to means (MINDIST), maximum likelihood (MAXLIKE), linear discriminant analysis (FISHER), and
Fuzzy ARTMAP neural network only output one hard classified map. Others can output both a hard and soft outputs.
Unsupervised Classification
In contrast to supervised classification, where we tell the system about the character (i.e., signature) of the information
classes we are looking for, unsupervised classification requires no advance information about the classes of interest.
Rather, it examines the data and breaks it into the most prevalent natural spectral groupings, or clusters, present in the
data. The analyst then identifies these clusters as landcover classes through a combination of familiarity with the region
and ground truth visits. For example, the system might identify classes for asphalt and cement which the analyst might
later group together, creating an information class called pavement.
IDRISI includes several unsupervised techniques. The module CLUSTER performs classification based on a set of input
images using a multi-dimensional histogram peak technique. While attractive conceptually, unsupervised classification has
Accuracy Assessment
A vital step in the classification process, whether supervised or unsupervised, is the assessment of the accuracy of the
final images produced. This involves identifying a set of sample locations (such as with the SAMPLE module) that are vis-
ited in the field. The landcover found in the field is then compared to that which was mapped in the image for the same
location. Statistical assessments of accuracy may then be derived for the entire study area, as well as for individual classes
(using ERRMAT).
In an iterative approach, the error matrix produced (sometimes referred to as a confusion matrix), may be used to identify
particular cover types for which errors are in excess of that desired. The information in the matrix about which covers are
being mistakenly included in a particular class (errors of commission) and those that are being mistakenly excluded (errors of
omission) from that class can be used to refine the classification approach.
Image Transformation
Digital Image Processing offers a limitless range of possible transformations on remotely sensed data. IDRISI includes
many transformations including: Principal Components Analysis (PCA) allowing both standardized and unstandardized
transformations; Canonical Components Analysis (CCA); Minimum Noise Fraction (MNF) that maximizes the signal to
noise ratio; Temporal Fourier Analysis (TFA) that performs harmonic analysis on temporal images; Color Space Transfor-
mation (COLSPACE); texture calculations (TEXTURE); blackbody thermal transformations (THERMAL); green vegeta-
tion indices (VEGINDEX); and Tasseled Cap (TASSCAP). In addition, a wide variety of ad hoc transformations (such as
image ratioing) that can be most effectively accomplished with the Image Calculator utility. Two transformations are men-
tioned specifically below (VEGINDEX and PCA), because of their special significance in environmental monitoring
applications.
Vegetation Indices
There are a variety of vegetation indices that have been developed to help in the monitoring of vegetation. Most are based
on the very different interactions between vegetation and electromagnetic energy in the red and near infrared wave-
lengths. Refer back to Figure 3-4, which includes a generalized spectral response pattern for green broad leaf vegetation.
As can be seen, reflectance in the red region (about 0.6 - 0.7μ) is low because of absorption by leaf pigments (principally
chlorophyll). The infrared region (about 0.8 - 0.9 μ), however, characteristically shows high reflectance because of scatter-
ing by the cell structure of the leaves. A very simple vegetation index can thus be achieved by comparing the measure of
infrared reflectance to that of the red reflectance.
Figure 3-9
This kind of calculation is quite simple for a raster GIS or image processing software system, and the result has been
shown to correlate well with ground measurements of biomass. Although NDVI needs specific calibration to be used as
an actual measure of biomass, many agencies have found the index to be useful as a relative measure for monitoring pur-
poses. For example, the United Nations Food and Agricultural Organization (FAO) Africa Real Time Information System
(ARTEMIS) and the USAID Famine Early Warning System (FEWS) programs both use continental scale NDVI images
derived from the NOAA-AVHRR system to produce vegetation index images for the entire continent of Africa every ten
days.3
While the NDVI measure has proven to be useful in a variety of contexts, a large number of alternative indices have been
proposed to deal with special environments, such as arid lands. IDRISI offers a wide variety of these indices (19) in the
VEGINDEX module. The chapter on Vegetation Indices offers a detailed discussion of their characteristics and poten-
tial application.
3. An archive dataset of monthly NDVI images for Africa is available on CD from Clark Labs. The Africa NDVI data CD contains monthly NDVI max-
imum value composite images (1982-1999), average and standard deviation of monthly NDVI images for each month over the same time period,
monthly NDVI anomaly images, and ancillary data (DEM, landuse and landcover, country boundaries and coast line) for Africa in IDRISI format. Con-
tact Clark Labs for more information.
Conclusions
Remotely sensed data is important to a broad range of disciplines. This will continue to be the case and will likely grow
with the greater availability of data promised by an increasing number of operational systems. The availability of this data,
coupled with the computer software necessary to analyze it, provides opportunities for environmental scholars and plan-
ners, particularly in the areas of landuse mapping and change detection, that would have been unheard of only a few
decades ago.
The inherent raster structure of remotely sensed data makes it readily compatible with raster GIS. Thus, while IDRISI
provides a wide suite of image processing tools, they are completely integrated with the broader set of raster GIS tools the
system provides.
System Operation
The IDRISI Application Window
When IDRISI is open, the application window will completely occupy the screen (in Windows terminology, it is automat-
ically maximized). Though not required, it is recommended that it be kept maximized because many of the dialog boxes
and images you will display will require a substantial amount of display space.2 The IDRISI application window includes
the menu, the toolbar, IDRISI Explorer, and the status bar.
The Toolbar
Just below the menu is a set of buttons that are collectively known as the toolbar. Each button represents either a program
module or an interactive operation that can be selected by clicking on that button with the mouse. Some of these buttons
toggle between an active and an inactive state. When active, the button appears to remain depressed after being clicked. In
these cases, the button can be released by clicking it again. You will also notice that some buttons may not always be avail-
able for use. This is indicated when the button icon appears grey. Hold the cursor over an icon to cause the name of the
function or module represented by that icon to momentarily appear. The set of icons represents interactive display func-
tions as well as some of the most commonly-used modules.
1. For an in-depth discussion of raster and vector data structures, see the chapter Introduction to GIS in this volume.
2. For this reason, we recommend that the operating system taskbar be set to "autohide" so that it does not constantly occupy much-needed screen
space. To do so, go to Start/Settings/Taskbar and toggle on the autohide option.
IDRISI
Explorer
IDRISI Explorer
IDRISI Explorer is a general purpose utility to manage and explore IDRISI files and projects. Use IDRISI Explorer to set
your project environment, manage your group files, review metadata, display files, and organize your data. Tools are pro-
vided to copy, delete, rename, and move. IDRISI Explorer can also be used to view the structure of IDRISI file formats
and to drag and drop files into IDRISI dialog boxes. IDRISI Explorer is permanently docked to the left of the IDRISI
desktop. It cannot be moved but it can be minimized, horizontally resized and closed.
Program Modules
Program modules may be accessed in three ways:
3. A Representative Fraction expresses the scale of a map as a fractional proportion. Thus, an RF of 1/10,000 indicates that the image on the screen is
one ten-thousandth of the actual size of the corresponding features on the ground. IDRISI automatically determines your screen resolution and dimen-
sions, which, in combination with the minimum and maximum X and Y coordinates of the image or vector layer, are used to determine the RF for any
display window.
Projects
Program modules act upon data—map layers and tabular data stored in data files. These files are stored in folders (also
known as directories) and sub-folders (sub-directories) on drives in the computer. The location of any specific data file is
thus designated by a name consisting of the filename plus the drive, folder and sub-folder locations. Collectively, these are
known as a Project (since they specify the route to a particular collection of data). For example, "c:\massachusetts\middle-
sex\census_tracts.vct" might designate a vector layer of census tracts contained in the Middlesex County sub-folder of the
Massachusetts folder on hard disk drive C.
IDRISI can work with files from any folder (including network folders). However, it can often be a nuisance to continu-
ally specify folder and sub-folder names, particularly when a specific project typically accesses data in a limited set of fold-
ers. To simplify matters, IDRISI allows you to specify these folders as belonging to a specific project. When the paths to
these folders are designated in a Project, IDRISI finds data very easily and with minimal user intervention.
Setting Projects
Establishing a Project for the Working and Resource Folders can be set in IDRISI Explorer from the File menu. Once in
IDRISI Explorer use the Projects tab to set your folders. You can also open IDRISI Explorer by clicking the leftmost
icon on the toolbar.
Pick Lists
Whenever a dialog box requires the name of an input data file (such as a raster image), you have two choices for input of
that value. The first is simply to type the name into the input box. Alternatively, you can activate a Pick List of choices by
either double-clicking into the input box, or clicking the Pick List button to the right of the input box (the small button
with the ellipses characters "..."). The Pick List is organized as a tree directory of all files of the required type that exist in
the Working and Resource Folders. The Working folder is always at the top of the list, followed by each Resource Folder
in the order in which they are listed in the Project Environment. The files displayed in the Pick List depend upon the par-
ticular input box from which the Pick List was launched. All IDRISI file types are designated by unique filename exten-
sions. If an input box requires a raster image, for example, the Pick List launched from that input box will display only
those files that have an .rst extension.
You may select a file from the Pick List by first highlighting it with a click of the left mouse button, and then selecting it
by clicking the OK button or pressing the Enter key. Alternatively, double-clicking the highlighted entry will also cause it
to be selected. The Pick List will then disappear, and the name of that file will appear in the input box. Note that you can
also choose a file that is not in the current project environment by clicking on the Browse button at the bottom of the
Pick List.
4. The documentation file for output raster and vector data files includes in the lineage field the command line from which the layer was created. This
may be viewed with the Metadata utility in IDRISI Explorer.
5. The automatically-generated output file prefix may be changed in the User Preferences option of the File menu.
Getting Help
While using IDRISI, help is always close at hand. IDRISI contains an extensive on-line Help System that can be accessed
in a variety of ways. The most general means of accessing help is to select it from the main IDRISI menu. However, each
program module also contains a help button. In this case, help is context sensitive, accessing the section in Help associ-
ated with that particular module. In either case, it is worth becoming familiar with the many options the Help System pro-
vides. To do so, select the Using Help option from the main Help menu.
Map Layers
Map layers are the fundamental units of display and analysis in GIS. A map layer is an elementary theme, a single phenom-
enon that can be mapped across space. Examples would include a landuse layer, a roads layer, an elevation layer, a soils
layer, and so on. Map layers, then, are somewhat different from traditional maps. Traditional maps usually consist of sev-
eral map layers. For example, a topographic map sheet would typically consist of a contour layer, a forest cover layer, an
administrative boundaries layer, a roads layer and a settlements layer. In IDRISI, this same concept is applied. A map is a
graphic representation of space that is composed of one or more map layers using a tool called Composer.
This breakdown of the map into its elementary constituents offers important advantages. Clearly it allows for the simple
production of highly customized maps; we simply pull together the layers we wish to see together. However, the more
important reason for storing data this way is that layers are the basic variables used by the GIS in analytical modeling.
Thus, for example, we might create a map of soil erosion as a mathematical function of soil type, slope, landcover, and
precipitation layers. Indeed, the breakdown of geographic data into layers allows for the production of an extraordinary
range of mathematical and logical models.
1. In IDRISI, the terms raster layer, raster image layer and image are all used interchangeably.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 41
folders). It is also not necessary to specify their filename extensions (since IDRISI looks for and creates specific exten-
sions automatically).
Consistent with the 32-bit variants of the Windows operating system, layer names may contain any letter of the alphabet
as well as numerals, spaces and most symbols (excluding \ / : * ? " < > and |). For example, layers might have names such
as "soils" or "landuse 1996." In principle, filenames can be up to 255 characters in length. However, in practical use, it is
recommended that they be only as long as necessary so the entire filename will be visible in the Pick Lists and dialog input
boxes.
In normal use, no distinction is made between image and vector layer names. However, their type is always known by the
system—by the context in which they were created and by the file extension that the system uses internally to store each
layer. Thus a raster image layer named "soils" would be stored as "soils.rst" internally, while a vector layer named "soils"
would be stored as "soils.vct". As a result, it is possible to have raster and vector layers of the same name. However, most
users of IDRISI allow this to occur only in cases where the two files store vector and image manifestations of the same
underlying data.
When filenames are input into IDRISI dialog boxes, the system always knows the appropriate layer type in use. Indeed,
only those files of the appropriate type for that operation will display in the Pick List.
2. Feature definition vector files and image files describe the geography of features but not their attributes. Rather, the features have identifier values that
are used to associate attributes stored in database tables or attribute values files.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 42
2. an associated attribute database;
3. a specific table within that database; and
4. the database field that contains the identifiers which can associate database records with vector features (the link
field).
In Database Workshop, the Establish Display Link utility is used to construct these link files. Only vector point, line or
polygon layers can be specified as a feature definition vector file. The database files and associated tables must be Micro-
soft Access Database Engine files with an ".mdb" extension.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 43
you choose one, IDRISI will enter its full reference (using the dot convention) into the input box. In some cases, a group
filename may itself be selected as input for a dialog box.
Attribute Files
An attribute file contains information about the characteristics of features, but not the geographic definition of those fea-
tures. IDRISI supports two forms of attribute files, database tables and ASCII values files.
Data Tables
IDRISI incorporates the Microsoft Access Database Engine within its Database Workshop facility. Thus, this form of an
IDRISI attribute files are tables and have ".mdb" extensions, and can also be used and modified with Microsoft Access or
any other Access-compatible system. The linked layer collection can only be created with this type of attribute file. Micro-
soft Access Database files may contain multiple tables, each one of which can be linked to form a layer collection. See the
chapter Database Workshop for more information about using database tables with IDRISI.
Values Files
The second form of attribute file that IDRISI uses is an ASCII values file. This has a very simple file structure that con-
tains the values for a single attribute. It is stored in ASCII text format and consists of two columns of data separated by
one or more spaces. The first column contains an identifier that can be used to associate the value with a feature (either
raster or vector), while the second column contains the attribute value. Attribute values files have an ".avl" extension.
Some modules create these as output files while others require (or can use) them as input for analyses. Attribute values
files may be imported or exported to or from a database table using Database Workshop. In addition, IDRISI provides a
procedure (ASSIGN) to assign those values to a set of features included in a raster image or vector file, and a companion
procedure (EXTRACT) to create a new values file as a statistical summary of the values in another image for the features
of a feature definition image.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 44
instance, an image consisting of 3 rows by 5 columns is stored as a single column of 15 numbers. It is the image's docu-
mentation file that allows IDRISI modules to reconstruct the grid from this list. An image that looks like this:
10 15 9 10
1 14 10 11
14 13 11 10
has an image file that looks like this (if one could see it):
10
15
9
10
1
14
10
11
14
13
11
10
The raster documentation file, containing the number of rows and columns, allows the image to be correctly recreated for
display and analysis.
As mentioned earlier, the major sub-types of raster images are differentiated on the basis of the data types that are used to
represent cell values. IDRISI recognizes four raster data types: integer, byte, real, and RGB24.
1. Integers are numbers that have no fractional part and lie within the range of -32768 to +32767. Integer files are
sometimes called 16-bit integer files since they require 16 bits (i.e., 2 bytes) of memory per value (and therefore per pixel).
Integer values can be used to represent quantitative values or can be used as codes for categorical data types. For example,
a soils map may record three soil types in a particular region. Since IDRISI images are stored in numeric format, these
types can be given integer codes 1, 2 and 3. The documentation file records a legend for these, on the relationship
between the integer codes and the actual soil types.
2. Byte values are positive integer numbers ranging from 0 to 255. The byte data type is thus simply a subrange of
the integer type. It is used in cases where the number range is more limited. Since this more limited range only requires 8
bits of memory per value for storage (i.e., 1 byte), only half as much hard disk or memory space for each image is needed
compared to integer. This data type is probably the most commonly used in GIS since it provides an adequate range to
describe most qualitative map data sets and virtually all remotely-sensed data.
3. Real numbers have a fractional part such as 3.14. Real numbers are used whenever a continuous (rather than dis-
crete) data variable is being stored with great precision, or whenever the data range exceeds that of the integer data type.
The real data type of IDRISI raster images is known as single precision real numbers. Thus it can store values within a
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 45
range of ± 1 x 1038 with a precision of 7 significant figures. Single precision values provide a nice balance between range,
precision and data volume. Each number (and thus each pixel) requires 4 bytes of storage. Note that real numbers have no
discrete representation (e.g., 4.0 is the same as 3.99999999999999999999 at some level of precision). Thus, while it is pos-
sible to store whole numbers in a real data type, they cannot then be used as feature identifiers (e.g., for ASSIGN or
EXTRACT) or in modules that require byte or integer values (e.g., GROUP).
4. RGB24 data files use 24 bits (3 bytes) for each pixel to encode full color images. The internal structure is band-
interleaved-by-pixel and can encode over 16 million separate colors. RGB24 images are also constructed using the COM-
POSITE module (in the Display menu). The IDRISI display system allows for interactive and independent contrast
manipulation of the three input bands of a displayed RGB24 image. They are the preferred format for the display of full
color images.
The documentation file associated with each image records the file data type. When a new image is created, it maintains
the prevailing data type of the input image, or it produces a data type that is logical based upon standard mixed arithmetic
rules. Thus, dividing one integer data image by a second integer data image yields a real data image. IDRISI accepts inte-
ger, byte and real data types for almost all operations that would logically allow this. Some modules, though, do not make
sense with real data. For example, the GROUP operation that extracts contiguous groups can only do so for categorical
data coded with integer numbers. If the data type in an image is incorrect for a particular operation, an error message will
be displayed.
The CONVERT module can be used at any time to convert between the integer, byte, and real data types. In the case of
conversion from real numbers to either the integer or byte formats, CONVERT offers the option of converting by
rounding or by truncation.
The data type described above indicates the type of numbers stored in an image. Another parameter, the file type, indi-
cates how these numbers are stored. As with the data type, the file type of an image is recorded in its documentation file.
Image files may be stored in ASCII, binary or packed binary formats, although only the binary form is recommended (for
reasons of efficiency). Binary files are those which are stored in the native binary encoding format of the operating system
in use. In the case of IDRISI, this will be one of the Windows operating system variants. Thus, for example, the binary
coding of a real number is that adopted by Windows and the Intel hardware platform.
Binary files are efficient in both hard disk space utilization and processing time. However, the format is not universal and
not always very accessible. As a consequence, IDRISI also provides limited support for data recorded in ASCII format. A
file in ASCII format is also referred to as a text file, and can be viewed directly with any text editor (such as the Edit mod-
ule in IDRISI). The ASCII file type is primarily used to transfer files to and from other programs, since the coding system
is a recognized standard (ASCII = American Standard Code for Information Interchange). ASCII files are not an efficient
means of storing data and they cannot be displayed. You must convert your files to binary format before continuing with
your analyses. The CONVERT module converts files from ASCII to binary or binary to ASCII. Although binary files
cannot be examined directly, IDRISI provides strong data examination facilities through IDRISI Explorer on the File
menu. Binary files may be viewed as a matrix of numbers using the Show Structure utility of IDRISI Explorer. There is
rarely a need then to convert images to ASCII form.
The packed binary format is a special data compression format for binary integer or byte data, known as run-length
encoding. It played a special role in earlier MS-DOS versions of IDRISI before file compression was available at the oper-
ating system level. However, it is of limited use now that Windows takes an active role in data compression. Its support is
largely for purposes of backward compatibility with earlier versions of IDRISI. Like ASCII files, most IDRISI modules
still support the packed binary data type. However, also like ASCII files, packed binary files cannot be directly displayed.
Thus the use of this file type is also discouraged. As with other conversions, converting raster images to or from packed
binary is undertaken with the CONVERT module.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 46
tion file. Raster documentation files are automatically assigned an ".rdc" file extension by IDRISI.
Documentation files are always stored in ASCII format. They can be viewed and modified using the Metadata utility in
IDRISI Explorer from the toolbar icon or File menu. After specifying the file you wish to view from the file list, the con-
tents of its documentation file will then be displayed.
The documentation file consists of a series of lines containing vital information about the corresponding image file. The
first 14 characters describe the contents of the line, while the remaining characters contain the actual data. For example,
the documentation file for a soils image (soils.rst) might look like this:
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 47
comment : based on a stratified random sample of 37 points.
This file contains information on major soils groups and is in byte format stored as a binary file. The image contains 512
columns and 480 rows, for a total of 245,760 values. The image is georeferenced with a reference system named
US83TM183 with each coordinate unit representing 1 meter. The minimum and maximum X and Y coordinate values
indicate the reference system coordinates of the left, right, top and bottom edges.
The position error indicates how close a feature's actual position is to its mapped position in the image. It is marked in the
example as unknown. If known, this field should record the RMS (Root Mean Square) error of locational references
derived from the bounding rectangle coordinates. This field is for documentation purposes only and is not currently used
analytically by any module.
The resolution refers to the inherent resolution of the image. In most cases, it should correspond with the result of divid-
ing the range of reference coordinates in X (or Y) by the number of columns (or rows) in the image. However, there are
some rare instances where it might differ from this result. A good example is the case of Landsat Band 6 (Thermal) imag-
ery. The resolution of these data is actually 120 meters and would be recorded as such in this field. However, the data is
distributed in an apparent 30 meter format to make them physically match the dimensions of the other bands in the
image. This is done by duplicating each 120 meter pixel 4 times in X and then each row 4 times in Y. The resolution field
is a way of correctly indicating the underlying resolution of these data.
In most images, the resolution in X and Y will be equal (i.e., pixels will be square). However, when this is not the case, the
coarser resolution is recorded in the documentation file.4 For example, with Landsat MSS data, the pixel resolution is 80
meters in X and 60 meters in Y. In this case, the documentation file would show a resolution value of 80. Rectangular pix-
els will display correctly, however, preserving the true resolution in both X and Y. In addition, all analytical modules using
measures of distance or area calculate resolution in X and Y independently. They, therefore, do not rely on the resolution
value recorded in the documentation file.
The minimum and maximum value fields record the actual range of data values that occurs in image cells, while the dis-
play min and display max values designate what are sometimes called the saturation points of the data range. Values less than
or equal to display min are assigned the lowest color in the palette sequence, while values greater than or equal to display
max are assigned the highest color in the palette sequence. All values in between are assigned palette colors according to
their position in the range (i.e., with a linear stretch). It is very common for the display min and max to match the min and
max values. However, in cases where one wishes to alter the brightness and contrast of the image, altering the saturation
points can produce a significant enhancement.5 Saturation points may be manipulated interactively in the display system
(see the chapter Display System in this volume).
The value units read classes in this example to indicate that the numbers are simply qualitative codes for soil classes, not
quantitative values. The value units field is informational only, therefore any description can be used. It is suggested that
the term classes be used for all qualitative data sets. However, when standard linear units are appropriate, use the same
abbreviations that are used for reference units (m, ft, mi, km, deg, rad).
The value error field is very important and should be filled out whenever possible. It records the error in the data values
that appear in image cells. For qualitative data, this should be recorded as a proportional error. In the example, the error is
recorded as 0.15, indicating cell accuracy of 85% (i.e., what is mapped is expected to be found on the ground 85% of the
3. The reference system indicated in the documentation file is the name of a reference system parameter file (.ref). The parameters of this file are detailed
in the chapter Georeferencing in this volume.
4. Earlier versions of IDRISI calculated and reported the resolution field as that of the X dimension only.
5. Note that one would normally only change the saturation points with quantitative data. In the example illustrated here, the numeric values represent
qualitative classes. In these cases, the saturation points should match the actual minimum and maximum values.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 48
time). For quantitative data, the value here should be an RMS error figure. For example, for an elevation image, an RMS
error of 3 would indicate 68% of all values will be within ± 3 meters of the mapped value, that approximately 95% will be
within ± 6 meters, and so on. This field is analytical for some modules (e.g., PCLASS) and is intended to be incorporated
into more modules in the future.
The flag value and flag definition fields can be used to indicate any special meaning that certain cell values might carry.
The flag value indicates the numeric value that indicates a flag while the flag definition describes the nature of the flagged
areas. The most common data flags are those used to indicate background cells and missing data cells. These flag defini-
tion field entries are specifically recognized and used analytically by some modules (e.g., SURFACE). Other terms would
be considered only informational at this stage. Note that some modules also output data flags. For example, when SUR-
FACE is used to create an aspect image, aspects derived from a slope of zero are flagged with a -1 (to indicate that the
aspect cannot be evaluated). In the output documentation file, the flag value field reads -1 while the flag definition field
reads "aspect not evaluated, slope=0".
The legend cats field records the number of legend captions that are recorded in the file. Following this, each legend cap-
tion is described, including the symbol code (a value from 0-255) and the text caption.
Finally, the image documentation file structure allows for any number of occurrences in four optional fields: comment,
lineage, consistency and completeness. At present, these fields are for information only and are not read by IDRISI mod-
ules (although some modules will write them). Note that the lineage, consistency and completeness fields are intended to
meet the recommendations of the U.S. National Committee for Digital Cartographic Data Standards (NCDCDS). Along
with the positional (pos'n) error and value error fields, they provide a means of adhering to the current standards for
reporting digital cartographic data quality.6 Multiple occurrences of any of these field types can occur (but only at the end of
the file) as long as they are correctly indicated in the 14 character descriptive field to the left. The lineage field can be used
to record the history of an image. The command line used to generate a new file is recorded in a lineage field of the doc-
umentation file. The user may add any number of additional lineage lines. The consistency field is used to report the logi-
cal consistency of the file; it has particular application for vector files where issues of topological errors would be
reported. The completeness field refers to the degree to which the file comprehensively describes the subject matter indi-
cated. It might record, for example, the minimum mapping unit by which smaller features were eliminated. Finally, the
comment field can be used for any informational purpose desired. Note that the Metadata utility in IDRISI Explorer and
CONVERT module both read and maintain these optional fields. The Metadata utility also allows one to enter, delete or
update fields of this nature (as well as any documentation file fields).
6. For further information about the NCDCDS standard, refer to the January 1988 issue of The American Cartographer, Volume 15, No. 1.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 49
and can be stored either as integer or real numbers (although identifiers must be integers).7
A special feature of vector files is that their data types (integer or real) are less specific in their meaning than raster files.
The integer type has a very broad range, and is compatible with both the byte and integer type of raster layers as well as
the long integer type commonly used for identifiers in database tables. Very approximately, the vector integer type covers
the range of whole numbers from -2,000,000,000 to +2,000,000,000. Similarly, the vector real data type is compatible with
the single precision real numbers of raster layers, but is in fact stored as a double precision real number with a minimum
of 15 significant figures. IDRISI's conversion procedures from vector to raster handle these translations automatically, so
there is little reason to be concerned about this detail. Simply recognize that integer values represent whole numbers while
real numbers include fractional parts.
Point Files
Point files are used to represent features for which only the location (as a single point location designation) is of impor-
tance. Examples include meteorological station data, fire hydrants, and towns and cities (when their areal extent is not of
concern). Each point feature is described with an attribute value that can be integer or real, and an X,Y coordinate pair.
Line Files
Line files describe linear features such as rivers or roads. Each line feature in a layer is described by an attribute value that
can be integer or real, a count of the number of points that make up the line, and a series of X,Y coordinate pairs (one for
each point).
Polygon Files
Polygon files describe areal features such as forest stands or census tracts. Each polygon feature in a polygon layer is
described by an attribute value that can be integer or real, a count of the number of parts in the polygon, and for each part,
a list of the points (by means of an internal index) that make up that part. This is then followed by an indexed list of the
X,Y coordinate pairs of all points in the polygon. The parts of a polygon are concerned with the issue of holes. A polygon
with one part has no holes, whereas one with two parts has one hole. The first part listed is always the polygon which
encloses the holes.8
Text Files
Text vector files represent text captions that can be displayed as a layer on a map. They store the caption text, their posi-
tion and orientation, and a symbol code that can be used to link them to a text symbol file. Text vector files can be created
with the on-screen digitizing feature in IDRISI or by exporting text from a CartaLinx coverage. Text symbol files are cre-
ated with Symbol Workshop.
7. All features in a single vector layer should be encoded with numeric values of the same type. Thus, if integer identifiers are used, all features must have
integer identifiers. Similarly, if real number attributes are encoded, all features in that layer must have real number attributes.
8. Note that islands are treated as separate polygons by IDRISI. Thus, if one were to take the case of the State of Hawaii as an example, one might con-
sider it to be a single polygon with five parts. In IDRISI, though, these would be five separate polygons. However, they can each have the same identifier
allowing them to be linked to a common data entry in a database table.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 50
A sample vector documentation file appears as follows:
As can be seen, the structure of a vector documentation file is virtually identical to that for a raster layer. The only differ-
ences are:
1. the row and column information is absent;
2. data type has been replaced by id type, although the intention is identical. Choices here are integer or real. The rea-
son for the slightly different wording is that coordinates are always stored as double precision real numbers. The id's,
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 51
though, can vary in their data type according to whether they encode identifiers or numeric attribute values.
3. the file type is always binary. (ASCII format for vector files is supported through the Vector Export Format.)
4. the minimum and maximum X and Y values recorded in the documentation file do not necessarily refer to the
minimum and maximum coordinates associated with any feature but to the boundary or limits of the study area. Thus
they correspond to the BND (boundary) coordinates in vector systems such as Arc/Info.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 52
A sample attribute documentation file appears as follows:
This example shows the documentation file for an ASCII attribute values file. This type of file always has two fields. Data-
base tables will normally have many more fields. For each field, the parameters shown are repeated.
In this version of IDRISI, the following file types are supported:
ascii
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 53
Simple 2-column ASCII format files where the left field contains integer feature identifiers and the right field
contains data about those features. These values files have an ".avl" file extension.
access
A database file in Microsoft Access format having an ".mdb" extension. This format is the current resident data-
base format supported by IDRISI, and is supported by Database Workshop. It is also the format used in the cre-
ation and display of vector collections.
In the case of the simple ASCII form (".avl" file), format information is unimportant, and thus reads 0 in this example.
Spaces or tabs can be used to separate fields. With fixed length ASCII and database files, format information is essential.
The format line simply indicates the number of character positions occupied by the field.
byte and integer data
A single number to indicate the number of character positions to be used.
For most other entries, the interpretation is the same as for raster image layers. Valid data types include byte, integer, real and
string (for character data). However, for data tables (.mdb), many of the data types recognized by the Microsoft Access Jet
Engine are supported. These include:
real (-3.402823E38 to +3.402823E38 real numbers)
byte (0-255, whole numbers)
integer (-32768 to 32767, whole numbers)
longint (long integer, -2,147,483,648 to 2,147,483,647, whole numbers)
text (character strings)
Boolean (true or false)
IDRISI will document these types automatically, and knows how to convert them when they are used in the context of
display (in a vector collection) or when creating an ASCII values file.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 54
Symbol and Palette Files (.sm0, .sm1, .sm2, .smt, .smp)
In order to display map layers, it is necessary to set up an association between vector features or image values and partic-
ular graphic renditions. This is done through the use of Symbol Files and Palette Files. A set of symbol and palette files
is included with IDRISI, and may meet most user needs. However, for final output, it is often desirable to create custom
symbol and palette files. This is done with the Symbol Workshop utility under the Display Menu. See the Display System
chapter in this volume for more information about Symbol Workshop.
Symbol files indicate the manner in which vector features should be symbolized (such as the type, thickness and color of
lines or the font, style, size and color of text). Each symbol file lists the characteristics of up to 256 symbols that are iden-
tified by index numbers from 0 to 255. In all, four kinds of symbol files are used, one each for the vector feature types:
point, line, polygon and text. They are stored with file extensions of ".sm0," ".sm1," ".sm2" and ".smt" respectively. As
usual, IDRISI always knows the appropriate symbol file type to use. As a result, you never need to specify a symbol file
with its extension.
For raster images, graphic renditions are specified by the use of palette files. Like symbol files, palette files also define up
to 256 renditions identified by index numbers from 0 to 255. However, in this case, only the color mixture (defined by the
relative amounts of red, green and blue primaries) is specified. Palette files are stored with an ".smp" extension.
Chapter 5 Map Layers, Raster Group Files, Vector Collections and Data Structures 55
Display System
Introduction
The previous chapter described map layers as each representing a single elementary theme. By combining map layers and
giving them graphic renditions, we create a map. Map compositions may contain as few as one layer and as many as 32 lay-
ers. Map compositions are created as a natural consequence of working with the IDRISI display system. As you work,
IDRISI keeps track of all changes and additions you make to the composition. You can then save the composition at any
time. The composition is stored in a file with a ".map" extension and is simply called a map file.
The display system in IDRISI consists of several separate but interdependent program components, each of which is
described in this chapter.
DISPLAY Launcher
DISPLAY Launcher is used to open a new display window. It begins the map composition process, and is always the first
operation required to create a new map display. DISPLAY Launcher can be accessed from its toolbar icon (shown above)
or by choosing it from the Display menu. Doing so opens a dialog box with options to display a raster layer, a vector layer,
or an existing map composition.
When you select a raster or a vector layer, IDRISI uses a set of decision rules based on the values in the layer to suggest an
appropriate palette or symbol file. You may change this selection. You must also specify if the layer should be displayed
with a direct relationship between numeric values and symbol codes, or should be autoscaled (see below). In the case of a
map composition, you will only be required to specify its name, since all display parameters are stored in the map file. See
the chapter System Overview in this volume for options on choosing filenames from the Pick List or typing them in
You should first indicate whether your data express quantitative or qualitative variations. Quantitative data express differ-
ences of degree (such as elevation) while qualitative data express differences of kind (such as landcover categories). Alter-
natively, you can indicate that you wish to symbolize all features with a uniform symbol. You will also need to decide on
the color logic. For each color logic, you are provided four choices which can be selected by pressing the appropriate but-
ton. For Qualitative and Uniform symbol schemes, the choices of the color logic will be straightforward. However, the
choices for quantitative data warrant additional explanation:
Autoscaling
Autoscaling concerns the relationship between raster cell values and palette indices (and, similarly, vector ID’s or linked
attribute values and symbol indices). By default, a direct relationship between them is assumed (i.e., a cell with a numeric
value of 12 should be displayed with the color defined by palette index 12). However, not all images contain integer values
that fall nicely within the allowable range of values for palette indices (0-255). As a result, it is often necessary to scale the
actual range of values into this more limited range. For example, we might have an image layer of temperature anomalies
ranging from (-7.2) degrees to (+4.6) degrees. These values cannot be directly displayed because there is no palette index
of (-7.2). Autoscaling offers a solution to this problem.
Four options for autoscaling are provided: Off (Direct), Equal Intervals, Quantiles and Standard Scores.
• Off (Direct). When autoscaling is off, it is assumed that there is a direct relationship between the numeric values
of pixels or features in the layer and the numeric codes in palettes and symbol files. Thus if a pixel contains the number 5
in a layer, it is symbolized with the palette color 5. This option is only permitted in cases where all pixels or features have
integer values between 0-255.
• Equal Intervals. This is the default for all cases where direct scaling is not possible - i.e., for cases where some
data values are less than zero or greater than 255, and all cases where the layer contains real numbers. In this case, the
range of numeric values is automatically scaled (hence the name autoscaling) into a series of equal-width data classes that
are subsequently assigned to symbols. Thus, for example, a layer containing elevations from 0 to 5000 meters might be
automatically scaled into 10 classes, 500 meters in width. Therefore, the values from 0-499 would be assigned to symbol 0,
500-999 to symbol 1, 1000-1499 to symbol code 2, etc. DISPLAY Launcher will make an initial determination of the
number of classes. However, you have the ability to specify any number from 2-256. Remember, however, that all numeric
series start from 0 in IDRISI. Thus if you were to choose 256 classes, they would be understood to be symbolized by sym-
bols codes 0 through 255. Note that this process of dividing the numeric data range into a series of symbolization classes
is also referred to as classification by cartographers.
• Quantiles. A quantiles classification scheme places an equal number of pixels or features into each class, first by
rank ordering the pixels or features and then assigning them to classes in rank order. Some quantiles schemes are known
by special names. Thus, for example, a quantiles scheme with four classes is known as a quartiles scheme. Quantiles are a
good choice whenever the data are strongly skewed or disproportionately loaded on a small number of equal interval
classes.
• Standardized. A standardized classification scheme divides the data into symbol classes based on standard devia-
tion units. With an even number of classes, the mean will represent the central boundary between classes, while with an
odd number of classes, the mean will be at the center of the middle class. All classes are one standard deviation in width
and the end classes always include all cases below its defining boundary. Thus, for example, the default case of 6 classes
will be defined as:
<=-2sd
<=-1.5sd
-1.5sd to -0.5sd
-0.5sd to +0.5sd
+0.5sd to +1.5sd
>=+1.5sd
Standardized classes are appropriate whenever the data are approximately normally distributed and one wishes to differ-
entiate unusual from common cases.
There are several further issues that should be noted about autoscaling. First, Equal Intervals autoscaling is automatically
invoked whenever real numbers or integer values outside the 0-255 range must be displayed. Second, while the maximum
range of palette and symbol indices is 0-255, some palette or symbol files may use a more limited range of values for the
purpose of autoscaling (e.g., 1-100). The autoscaling range of a palette or symbol file can be inspected and changed with
Symbol Workshop. Third, while Equal Intervals autoscaling works well for many images, those with extremely skewed
distributions of data values (i.e., with a very small number of extremely high or extremely low values) may be rendered
with poor contrast. In these cases, you can alter the display min and display max values (known as the saturation points)
using the Layer Properties option of Composer (see below). Alternatively, Quantiles autoscaling can be used. Finally, note
that autoscaling does not change the values stored in a data file; it only alters the display. To create a new raster image with
contrast-stretched values, use the module STRETCH.
Automatic Display
IDRISI includes an automatic display feature that can be enabled or disabled from User Preferences, under the File menu.
With automatic display, the results of analytical operations are displayed immediately after the operation has finished. The
system will determine whether to use the Default Qualitative or Quantitative palette (both of which are user-defined in
the Preferences dialog) for display and if autoscaling should be invoked. The artificial intelligence that is used to make
these determinations is not fool-proof, however. Palette and autoscaling choices may be quickly corrected in the Layer
Properties dialog, if necessary. The automatic display feature is intended as a quick-look facility, not as a substitute for
DISPLAY Launcher.
1. To drag a component, place the cursor over the component (not on the sizing buttons) and hold the left mouse button down while moving the cursor
to the desired position. "Drop" the component in place by releasing the mouse button.
Add Layer
To add a layer, click the Add Layer3 button on Composer and choose the desired file.4 You will then be presented with a
similar dialog to indicate the palette or symbol file to be used and whether the new layer should be autoscaled. All layers in
a map composition must share a common reference system. If this is not the case, a warning message will appear indicat-
ing this and warning that the layer may not display properly. The bounds of files do not need to match to be displayed
together in a map composition.
When a new layer is added, it is automatically given the highest priority (i.e., it is drawn on top of the other layers). The
first layer displayed will thus, by default, have the lowest priority (priority 0) and will appear to be on the bottom of the
2. This is only one of many settings in the Windows system display that is defined when you choose a color scheme or define the display characteristics
of individual Windows components.
3. Alternatively, pressing Ctrl-R when a map window has focus will bring up the Add Layer dialog set for adding a raster layer, and pressing Ctrl-V will
bring up the Add Layer dialog set for adding a vector layer.
4. Only one layer frame is present in any map window, so all layers are added to the same layer frame. Note that layers may be added from several paths
for display. However, if the composition is to be saved as a map composition file, all the layers must exist in the Working Folder and/or Resource Fold-
ers of the project from which it is displayed again.
right is the Transparency button. If the highlighted layer is raster, clicking it will cause the background color (whatever is
color 0 in the palette) to become transparent. Note that layers can be both blended and transparent as illustrated below.
This sequence shows a hillshading layer and a digital elevation model being blended (third frame). In the fourth frame a
mask layer has been added with black as color 0 and grey as color 1. In the fifth frame, the mask is made transparent
allowing the elevation model to show through the background color. Finally, in the last frame, the mask is also set to blend
allowing the topography to be partially visible through the grey mask.
The remaining buttons are for creating various color composites. The Red, Green and Blue buttons allow you to desig-
nate layers as the red, green and blue primary color components of a full color image. Note that the layers need to be adja-
cent to each other for this to work, but that the specific order is unimportant. The Cyan button (second from the left) is
used to create three dimensional anaglyphs. The creation of an anaglyph requires stereo images -- two views of the same
landscape taken from different locations, and a set of anaglyphic 3-D glasses. When these are superimposed using Add
Layer, assign red primary to the left view and the cyan primary to the right view (this order may need to be reversed
depending upon your glasses). Again, the layers should be next to each other in their priority for this to work.
Raster
Vector Point
Vector Line
Vector Polygon
Vector Text
Layer Properties
The properties of any layer, including key documentation information and display parameters, are shown in the Layer
Properties dialog. To view this, highlight the layer concerned with a single mouse click on the layer name in Composer,
and then click the Layer Properties button. This will display the Layer Properties dialog associated with that layer.
5. Note that the lowest and highest symbols or colors in the sequence are those set as the autoscale min and max in the symbol file or palette. These may
be adjusted in Symbol Workshop.
6. If you always wish to display a layer with particular display min/max values that are not the actual min/max values of the layer, use the Metadata utility
in IDRISI Explorer to set the display min and display max fields of the layer documentation file to the desired values. Whenever the file is displayed,
these values will be used to set the range of values covered by the palette or symbol file. The actual data values remain unchanged.
While the Layer Properties dialog displays information about a single map layer, the Map Properties dialog describes the
entire map composition. It is used to set the visibility and characteristics of map components such as layer legends, north
arrow and so forth.
The Map Properties dialog can be brought forward by clicking the Map Properties button on Composer or by clicking the
right mouse button in a map window. The right-click method is context sensitive. For example, a right-click over a legend
launches the Map Properties dialog with the legend options visible.
Map Properties is a tabbed dialog. Simply click on the appropriate tab to access the options associated with a particular
map component. Ten tabs are provided: Legends, Georeferencing, Map Grid, North Arrow, Scale Bar, Text Inset,
Graphic Insets, Titles, Background and Placemarks. The options for each are described below.
You can enlarge the Map Window at any time to increase the size of your graphic page and make space for new map com-
ponents. In all cases, the components must be set to Visible to be displayed. All components are movable and many may
be interactively resized. To do so, double-click on the component, make the necessary changes, then click any other ele-
ment to release move/resize mode.
Legends
Up to five layers may have displayed legends in a map composition. Any raster layer or vector point, line or polygon layer
can have a legend. The legend text entries are stored in the layer documentation file and may be entered or altered with
the Metadata utility in IDRISI Explorer. Up to 256 entries are allowed. By default, the legend will display up to 20 catego-
ries (this can be changed in User Preferences under the File menu). However, whenever there are more than 20 legend
entries in the documentation file, a vertical scrollbar will be attached to the legend, and you may scroll through the
remaining entries.
For a quantitative data layer with byte or integer data type and a range of values of 20 or less, a legend will be formed and
labeled automatically (i.e., no legend entries are needed in the documentation file). If the data are real or the data range is
greater than 20, a special continuous legend will be displayed with automatically generated representative category labels.
Georeferencing
The Georeferencing tab shows the reference system and units of the current composition, as well as its bounding rectan-
gle. It also shows the actual bounding rectangle of all features within the composition. The former can be changed using
Map Grid
A map grid can be placed on the layer frame. The intervals may be specified as well as the starting values in X and Y (all in
reference units). You can also choose the type of grid, color and width of the grid lines and the label font. The grid will be
automatically labeled.
North Arrow
A north arrow can be added to a composition using several default styles, or one can create their own and import them as
bitmaps or enhanced metafiles. You set the declination of the arrow and may specify any text to be displayed to the left
and right of the arrow. This would allow you, for example, to change the text to read "Magnetic North" rather than "Grid
North," or to change to a different language.
Scale Bar
Adding a scale bar can also be achieved with the Map Properties dialog. You set the length (in map reference system
units), the number of divisions, color properties and text label of the scale bar. The scale bar automatically adjusts in width
as you zoom.7
Text Inset
The text inset can be used to display blocks of text. This is done by specifying the name of an ASCII text file. The text
inset frame is sizeable and incorporates automatic wordwrap. Font style and background color may be set.
Graphic Insets
Graphic insets can be used to hold inset maps, pictures, logos, or any other graphic image. Any Windows Bitmap
(".bmp") or Metafile (".wmf" or ".emf") may be used as a graphic inset. All graphic insets are resizeable. However, if the
Stretchable option is disabled, resizing is controlled to preserve the original aspect ratio of the inset.
Titles
A title, subtitle, and caption may be added to a map composition. These are not associated with any particular map layer,
but rather belong to the map composition as a whole. However, the title of the base layer, read from its documentation
file, will appear as the title text by default. This may be edited and font properties may be set.
Background
The Background tab allows you to change the color of the backgrounds for the Layer Frame and the Map Window. The
7. This can sometimes prove to be a nuisance since the scale bar may need to expand to be greater than the width of the map window. In these cases,
simply set its visibility to "off" using Map Properties. Otherwise, use Map Properties to reset the size of the scale bar to a smaller length.
Placemarks
Finally, the Placemarks tab keeps track of placemarks associated with the map composition. A placemark is a particular
window of the map layers. The Placemarks tab allows you to define new placemarks, delete existing ones, and go to any
specific placemark. Placemarks are described further below in the section on Interactive Display Features.
Feature Properties
The Feature Properties button on Composer invokes the Feature Properties cursor mode, which brings up a table of
properties associated with features identified with the mouse. This may also be invoked with the Feature Properties tool-
bar icon. See the section below on Interactive Display Features for a complete description.
Save Composition
At any point in your use of Composer, it is possible to save your map composition in several formats. The Save Composi-
tion button brings up a dialog presenting the format choices. The first is to save the map composition as a MAP file. A
MAP file contains a complete record of the composition in a file with a ".map" extension.8 This can then be redisplayed
at any time by starting DISPLAY Launcher and indicating that you wish to display a map composition file. With this form
of composition storage, it is possible to restore the composition to its exact previous state, and then continue with the
composition process.
The next three options allow you to save the composition to either a BMP, WMF or EMF graphics file format. These file
types are used in desktop publishing, spreadsheet and word processing software programs. The BMP is a raster bitmap,
produced by simply copying the screen. This is often called a screen dump, since it copies the exact appearance of the map
window, without copying the details of how the composition was formed (i.e., the names of the layers, their symboliza-
tion, and so on). The EMF format is the latest 32-bit Windows "Enhanced Metafile" structure. The WMF format is the
"Windows Metafile" structure used through Windows 3.1. Both store the Windows instructions that can be used to
reconstruct both the vector and raster elements of the map window. These can then be imported into a desktop publish-
ing or graphics program. Whenever you have a choice between WMF and EMF, choose the EMF format.
The next option copies the map window to the Windows clipboard rather than to a file. This facilitates copying composi-
tions immediately into other software programs.
Finally, the last option allows you to save the currently windowed region of the active layer as a new IDRISI layer. If the
layer is raster, only the portion of the raster image that is currently displayed will be written to the new raster file. This is
thus an interactive version of the WINDOW module. However, if the active layer is vector, the entire file will be copied,
but the bounding coordinates will be altered to match the current window. The new file will display the window region,
but all the original data outside that window still exists in the file.
Print Composition
To print a map composition, first display it. Ensure that the map window has focus, then click the Print Composition but-
ton on Composer. In the Print Composition dialog, you will be able to select the desired printer, access its setup options,
8. More information about the MAP file structure may be found in the chapter Map Layers, Collections and Data Structures. Note especially that
the MAP file contains the instructions to construct the map composition, but does not include the map data layers themselves. These separate files are
used to recreate the composition.
Stretch
At the bottom of Composer are several image stretch buttons. These allow you to stretch the display of the active image
in a map window for optimal display. The inherent values are not altered, just the display minimum and maximum values
stored in the files metadata. All three options perform a saturation stretch, the first option applies it to the entire image,
the second option applies the stretch symetrically around zero, assuming you have values above and below zero. The last
option is used to stretch within a zoomed in display. It will stretch using the values only within the zoomed window, but it
will apply it to the entire image. When you zoom back to the default display, you can click the first icon to optimize for the
entire image.
Symbol Workshop
The map layers in a composition are rendered by means of symbol and palette files. While a set of symbol files is included
with IDRISI, commonly you will want to develop specific symbol files to optimize the impact of the information pre-
sented on your final maps.
Symbol and palette files are created and modified using Symbol Workshop,9 available under the Display menu and
through its toolbar icons. In all, five types of files can be created: point symbol files, line symbol files, polygon symbol
files, text symbol files and palette files. Symbol files record the graphic renditions for up to 256 symbols, indexed 0-255.
For example, a text symbol file might indicate that features assigned symbol index 5 are to have their names rendered
using bold, italic, 10 point Times New Roman text in a red color.
From the Symbol Workshop File menu, you can choose to open an existing file or create a new one. If you choose the lat-
ter, you will need to indicate the symbol type: point, line, polygon, text, or palette. The 256 symbols are then arrayed in a
16 x 16 grid. To change a symbol, simply click within its symbol grid cell. A symbol-specific dialog will then appear, allow-
ing you to alter any of the following settings:
9. IDRISI for Windows had a separate utility named Palette Workshop for the manipulation of palettes. This was incorporated into Symbol Workshop.
A very important feature of Symbol Workshop is the ability to copy or blend attributes. For example, imagine creating a
point symbol file of graduated circles. Click on symbol 0 and set its properties to yield a very small yellow circle. Then
move to symbol 255 and set it to be a large red circle. Now set the blend end points to be 0 and 255 and click the blend
button. You will now have a sequence of circles smoothly changing in both size and color.
As a companion to the blend option, Symbol Workshop also provides a copy function. With both the blend and copy
functions, it is possible to set options that will cause them only to copy or blend specific attributes (such as color or size).
To get a sense of how the symbols will look on a particular background, you may also set a background color for the Sym-
bol Workshop display.
Finally, the autoscaling range of a symbol file may be specified or altered in Symbol Workshop. All symbol files contain
definitions for 256 symbols. However, by altering the autoscale range, it is possible to create sequences of more limited
range. For example, to set up a palette with 8 colors, set the autoscale min and max to be 0 and 7 respectively. Then define
these 8 colors. In use, it will then appear that the palette has only 8 colors if the layer with which it is used is autoscaled.
Media Viewer
Media Viewer is a facility for creating and displaying video images composed of a series of IDRISI images. Media Viewer
is located under the Display menu. When activated, it will present the basic control dialog with its own separate menu.
Click on the File menu to create a new video or to open an existing video. If you choose to create a new video, you will be
presented with a new dialog that will require the name of either a Raster Group file (".rgf") or a Time Series file (".ts") that
defines the images to be used and their order in the video. You will also be asked to specify a palette to be used and the
time delay to be used between each successive image. The result of the operation will be an ".avi" multi-media video file
that can be displayed at any time using the Media Viewer controls. Note that the viewer can be sized by dragging its bor-
ders. The image can be sized to fit within the viewer by selecting the Fit to Window option under the Properties menu. Also
note that you can place a video into a continuous loop.
Feature Properties
Cursor Inquiry Mode allows one to view the values of features for a single layer at a time.
For layers that are members of a collection, the Feature Properties option allows one to
look at a simple tabular view of the values for all members of the collection at the query
location. To activate this feature, either click the Feature Properties button on Composer or
click the Feature Properties button on the tool bar. Click it again if you wish to turn it off.
For multi-layer query mode to work, the layer must have been displayed or added to the
map window as a part of a collection. This is accomplished by either typing in the layer
name using the "dot logic" described in the chapter Map Layers, Collections and Data
Structures in this volume, or by choosing the layer name from within the list of members
under the collection filename in the Pick List. When Feature Properties is used with layers
that are not members of a group, some simple information regarding the layer is presented
in the Feature Properties table along with the value and position of the location queried.
Clearly the major advantage of the Feature Properties table is that one can see the values of
all group members at one time. In addition, it can display these as a graph. Note that con-
trols are also provided to change the position of the separator line in table mode. To turn
off Feature Properties mode, click on the Feature Properties button again, or click its corresponding icon on the tool bar.
Cursor Inquiry mode is automatically invoked whenever Feature Properties is active.
Zoom Window
All zoom and pan operations take place within the context of the layer frame. See the earlier section of this chapter on
Map Windows and Layer Frames for information about enlarging both of these.
Group Link
When a raster layer is part of a group, it is possible to have all displayed members of the group zoom and pan simultane-
ously and identically. To activate this, click the Group Link button on the tool bar. Notice that the button depresses and
remains so until clicked again. This feature is especially useful when ground truthing remotely sensed imagery, particularly
when used with the GPS link (see below).
The Group Link works only when layers are launched with their complete collection reference. For layers belonging to
raster group files, these layers exist in isolation as well as being referenced as part of a group. Thus, if you wish to use the
group link feature, be sure to enter the full collection reference (e.g., "spotxs.band3" rather than simply "band3").
Placemarks
Placemarks are the spatial equivalent of bookmarks. Any particular view of a composition can be saved as a placemark. To
do so, either launch the Map Properties dialog and click the Placemarks tab or click the Placemarks icon on the tool bar.
Then choose the appropriate option to name, rename, delete, or go to any placemark. Note that placemarks are saved
with map compositions. They will not be saved if the composition is not saved.
Measure Length
Distances can be measured in map compositions. With an image displayed, click on the Measure Length icon. To begin
measuring the distance, move the mouse to the desired start location and click the left mouse button. A balloon above the
cursor will appear which will indicate the length as you move the mouse. If length is non-linear, click the left mouse but-
ton to indicate inflection points as you measure. End measuring by right clicking the mouse.
Measure Zone
Circular zones can be measured in map compositions. With an image displayed, click on the Measure Zone icon. To begin
measuring circular zones, move the mouse to the desired start location and click the left mouse button. A balloon above
the cursor will appear which will indicate the length from the center of the zone. A circular zone will also be drawn out as
you move the mouse. End measuring by right clicking the mouse.
How It Works
When you click the GPS button, IDRISI checks to see if there is a map layer with focus and with a valid reference system.
This can be any system with a Reference System Parameter file that has a projection supported by PROJECT (this does
10. Actually, the native reference system is a three-dimensional datum known as WGS84. However, the NMEA interface for communication of fixes
converts these to latitude/longitude and elevation readings.
11. See the chapter Georeferencing in this volume for a discussion of projections and reference systems.
Digitize
Delete Feature
To digitize on screen, make sure that the appropriate map window has focus, and then click on the Digitize button. If the
active layer (the one highlighted in Composer) is a raster layer, you will be presented with a dialog box in which you can
define the new vector layer to be created. If the active layer is a vector layer, you will first be asked whether you wish to
create a new vector layer for the digitized features or append the new features to those existing in the active layer. If you
choose to append to an existing layer, you will notice that the file information is already filled out and you only need enter
the ID or value. If you are digitizing a new file, you will be asked to specify the following:
Data Type
The data type refers to the attribute stored for each feature. This can be either integer or real. If you are digitizing identifi-
ers to be associated with data in a table, this must be integer.13 In all other cases, use integer for coding qualitative attri-
butes and real for recording quantitative data (where a fractional number may be required), or whenever the value is
expected to exceed the integer range.14
12. This is the same system used by CartaLinx. The GPS server program can simultaneously serve multiple applications. Thus, if it is already loaded for
use by CartaLinx, IDRISI does not load another copy, but simply registers itself as another user.
13. With vector data, no distinction is made between different storage formats of integer. Vector integer data can have values between ±2,147,483,647.
Automatic Index
This option is for those cases where you are storing identifiers for features (rather than some attribute directly). When
checked, this option will automatically increment the identifier from one feature to the next. This offers speed when digi-
tizing features whose ID’s increment in this simple manner.
Deleting Features
The Delete Feature button to the immediate right of the Digitize button on the toolbar allows you to select and delete
vector features from the active vector layer. First click on the Delete Feature icon, then select the feature to be deleted (the
cursor will become a hand). When the feature is selected, it will turn red. To delete it from the file, press the Delete key on
the keyboard.
14. Although vector layers can carry integer attribute values within a ±2,147,483,647 range, integer raster layers are restricted to a range of ±32,767. Thus,
if you intend to rasterize your digitized data at a later stage, you may wish to specify real if the range is outside the ±32,767 range.
15. When polygons are chosen, an option to digitize a flood polygon is presented. This is useful in image classification for delineating training sites. See
the Classification of Remotely Sensed Imagery chapter for more information.
A Final Note
Perhaps the best way to learn more about the IDRISI display system is to work through the first few tutorial exercises in
the Tutorial. These give step-by-step instructions and guide you through the basics of map layer display and map compo-
sition in IDRISI.
Software-Specific Formats
SHAPEIDR, ARCRASTER, and ARCIDRIS for ESRI Shape files, ArcInfo Raster Exchange, and ArcInfo GENER-
ATE/UNGEN file formats, respectively;
ATLIDRIS; for Atlas*GIS BNA files;
ECWIDRIS for ECW files;
ENVIIDRIS for ENVI files;
ERDIDRIS for Erdas LAN and GIS files;
ERMIDRIS for ER Mapper files;
GRASSIDR for GRASS raster files;
MAPIDRIS for Map Analysis Package files;
MIFIDRIS for MapInfo Interchange files;
PALIDRIS for palette files
SPLUSIDRIS for SPLUS statistical files;
SRFIDRIS for Surfer GRD files; and
STATIDRIS for STATISTICA files.
.IDRISI Vector Export for IDRISI vector export format files (.vxp).
The IDRISI File Conversion (16/32) utility converts between earlier 16-bit versions of IDRISI files and the current 32-
bit version.
To analyze change over multiple dates, the following four modules may be used.
PROFILE, creates profiles over space by querying the values along a linear transect across an image, or over time by que-
rying the value of the same location across multiple images.
TFA performs temporal Fourier analysis on image time-series.
CORRELATE calculates the Pearson Product Moment Coefficient of Correlation between a set of values in an attribute
values file and the values through a time series of images for each pixel of an image.
Interpolation Submenu
The first of the Surface Analysis submenus is Interpolation. The issues surrounding surface interpolation as well as the
options available in IDRISI are discussed in detail in the chapter Surface Interpolation.
INTERPOL interpolates a distance-weighted average or a potential model surface given an input set of points.
INTERCON interpolates a surface from a set of digitized contour lines.
Geostatistics Submenu
The second submenu in the Surface Analysis submenu is Geostatistics. The field of geostatistics has a broad range of
applications to many types of data and analyses. The chapter on Geostatistics gives a broad overview of Geostatistics
and the functions available in IDRISI through the three interfaces to Gstat (see footnote on previous page).
In the Spatial Dependence Modeler interface, the user employs a wide range of tools to learn about the patterns of spa-
tial dependence in the sample data set. In the Model Fitting interface, the user defines mathematical models to describe
the covariance relationships among sample data. In the Kriging and Conditional Simulation interface, full raster sur-
faces may be created from sample data and the models developed through the other interfaces.
1. IDRISI provides a graphical user interface to Gstat, a program for geostatistical modeling, prediction and simulation written by Edzer J. Pebesma
(Department of Physical Geography, Utrecht University). Gstat is freely available under the GNU General Public License from www.gstat.org. Clark
Labs' modifications of the Gstat code are available from the downloads section of the Clark Labs Web site at www.clarklabs.org.
Restoration Submenu
Image Restoration is the manipulation of remotely sensed images in an attempt to remove known value distortions. Res-
toration can be geometric or radiometric. The first two modules perform geometric corrections which are used to reduce
distortion at the edges of the image and to register the image to a coordinate system. The other modules perform radio-
metric restoration for the removal or diminishment of distortions in the data values of the images.
Enhancement Submenu
Image enhancement is the modification of image values to highlight information within the image. Most often these
enhanced images are used in visual analysis only, while the original images are used for automated analyses. The IDRISI
display system includes some facilities for enhancement of the screen display. These include the ability to interactively set
the endpoints used in applying the color palette to images. No new files are created with the Display tools, however. To
create new images that are enhanced, the following three modules are often used.
STRETCH increases the contrast in an image for the enhancement of visual interpretation.
COMPOSITE produces a 24-bit color composite image from three bands of imagery.
FILTER applies 3 by 3, 5 by 5, 7 by 7, or user-defined kernels to calculates new values using a mathematical operation on
the original cell value and its neighbors. The following filters are available: mean, Gaussian, median, standard deviation,
adaptive box, mode, Laplacian edge enhancement, high pass, Sobel edge detection, and user-defined.
PANSHARPEN performs a panchromatic merge using color space transformation, principal component transforma-
tion, and local regression transformation techniques.
Transformation Submenu
PCA provides both standardized and unstandardized principal components analysis.
CCA performs a canonical components analysis transformation.
MNF (Minimum Noise Fraction) maximizes the signal to noise ratio for a set of images.
TFA (Temporal Fourier Analysis) performs harmonic analysis on temporal images;
COLSPACE performs Hue/Lightness/Saturation (HLS) to Red/Green/Blue (RGB) and vice versa color space transfor-
mations.
TEXTURE calculate measures of variability (fractional dimension, class frequency, and edge analysis, and others), using
a 3 by 3, 5 by 5, or 7 by 7 template.
Window List
The Window List menu item provides a listing of all open windows. Open dialogs are listed with the module name (e.g.,
Composer) and open map windows are listed with the filename that appears in the banner of the map display window.
Clicking on a window name in the list will bring that window into focus. All open map layers and dialogs can be closed
from the Window List menu.
1. For more about collections, see the chapter Map Layers, Raster Group Files, Vector Collections and Data Structures.
2. This may not be evident since each map window will exactly overlay the previous one. We recommend moving each new window to an unused area of
the screen as it is created so that all can be seen.
Query by Location
When a map window contains an active layer (the one highlighted in Composer) linked to a database, you can use Cursor
Inquiry Mode (from the IDRISI toolbar) to perform database query by location. When you click on a feature in the map
display, Database Workshop will automatically locate the corresponding record in the database. The located record is indi-
cated with a triangular marker at the left edge of the record in the table display.
Sorting
To sort the records of the database according to the values of a particular field, click the mouse into any record of that
field then click either the ascending or descending sort button on the Database Workshop toolbar.
Query by Attribute
Query by attribute may be performed several ways, depending on the geography of the layers. If you are working with a
single geography (e.g. farm fields, provinces) defined by a vector file for which you have multiple attributes in a database,
the database query may be accomplished entirely in Database Workshop using an SQL filter. The results may then be
linked to a vector file for display or they may be assigned to a raster feature definition image for subsequent display.
For example, if you had a map of the countries of the world, and multiple attributes for each country stored in a database,
then you could perform a query such as, "Find all the countries where the median per capita annual income is less than
$5000, but the literacy rate is higher than 60%." The query conditions could be used in an SQL filter in Database Work-
shop, and the result linked for display to the original vector feature definition file in DISPLAY Launcher. If a new raster
image file should be made, the attributes may be assigned to a raster feature definition image from Database Workshop.
However, if the geographies of the attributes of interest are not the same, or if the attributes exist only as image layers,
then two steps are involved. First, the features meeting the conditions specified are selected in each layer. This normally
involves the use of RECLASS or ASSIGN. Then those selected data are used in an overlay operation, provided through
the module OVERLAY, to find the locations that meet all the conditions. (Both steps may be carried out with a single
command in Image Calculator, but behind the interface, the individual steps are still carried out in sequence.)
For example, you might ask, "Where are all the locations that have residential land use and are within a half mile of the
primary path of planes taking off from the proposed airport?" In this case, the geography of land use and that of the flight
paths for the airport are not the same. A Boolean image (zeros and ones only) would be made for each condition using
RECLASS or ASSIGN, then these would be combined in an OVERLAY multiply operation. The resulting image would
have the value of one only where both conditions are found:
0 0 0
1 0 0
0 1 0
1 1 1
Reclassification and overlay are fundamental to query by attribute in GIS. In IDRISI, RECLASS and ASSIGN are the
tools used to perform database queries on single attributes, and may be used to produce Boolean images either directly or
through the Image Calculator.
All of the logical operations can be achieved similarly through simple operations on Boolean images. For example, the
Boolean XOR (exclusive OR) operation can be performed with an addition operation, followed by a reclassification of all
values not equal to 1 to 0. In developing models that require this kind of logic, it is often helpful to construct a table such
as the ones above in order to determine the type of IDRISI operations needed.
In Image Calculator, these analyses are built as logical expressions. While Image Calculator often provides a faster and
easier interface, there are advantages, particularly to those new to GIS, to using the modules directly and performing each
step individually. Doing so allows each step in the process to be evaluated so that errors in logic may be detected more
easily. Using the modules individually also allows the user to become more familiar with their operation, facilitating the
use of these modules outside the limits of database query.
Query by Location
Query by location is most easily accomplished in IDRISI with the Cursor Inquiry tool in the Display System. Select the
Cursor Inquiry icon and with your cursor placed on the location in question, click the left mouse button. The underlying
data value for that location will be displayed on the screen.
Query by location can be extended to include query across multiple raster files by simply creating a raster image group file
Figure 1
Model Construction
To work with Macro Modeler, either click on its icon on the tool bar, or select it from the Modeling Menu. This yields a
special workspace in the form of a graphic page along with a separate menu and tool bar. Constructing a model involves
placing symbols for data files and modules on the graphic page, then linking these model elements with connectors. To
place an element, click on its icon (shown in Figure 2) or choose the element type from the Macro Modeler menu, then
choose the specific file or operation from the supplied pick list. (Note that not all IDRISI modules are available for use in
the Macro Modeler.1) To connect elements, click on the connector icon, then click on one of the elements and drag the
cursor onto the other element and release. It is necessary to connect elements in the proper order because the Macro
Modeler assumes that the process flows from the element you first clicked to the element upon which you released the cur-
1. The modules of the Image Processing menu as well as any module that does not create an output file, such as REGRESS, do not work in the Macro
Modeler.
Whenever a module is placed in the model, it is placed with the appropriate output data element for that operation already
linked. Default temporary filenames are assigned to output files. Right-click on a data file symbol to change the filename.
Long filenames are left-justified. To see the entire filename, pause the cursor over the element graphic to cause a balloon
to appear with the entire element name. Filenames are stored without paths. Input files may exist in any folder of the cur-
rent project (specified with IDRISI Explorer) while intermediate and final output files are always written to the Working
Folder.
To specify the parameters for the module, right-click on the module symbol. This opens the module parameters dialog.
Click on any parameter to see and choose other options. Access the Help System for detailed information about the vari-
ous options by clicking the Help button on the parameters dialog.
Submodels are user-constructed models that are subsequently encapsulated into a single command element. When a
model is saved as a submodel, a parameters dialog is created for the submodel in which the user provides captions for all
the model inputs and outputs. A submodel consists of a submodel parameter file (.ims) and its macro model (.imm) file.
When a submodel is placed as a command element in a model, the Macro Modeler knows how many and what types of
input files are required to run the submodel and what types of output files should be produced. For example, you might
create a submodel that stretches a single image then exports it to a JPG image. The model would include both STRETCH
and JPGIDRIS, but the submodel would only require the input image and the output image. The user may change the
input and output data elements of a submodel through the parameters dialog, but settings for module commands that are
part of a submodel must be changed by opening the model from which the submodel was created, making the changes,
then resaving the submodel. Submodels not only simplify multi-step processes, but are important elements when models
have sections that should loop. This is discussed in further detail below.
Models are saved to an IDRISI Macro Model file (.imm). This file preserves all aspects of the model, including the
graphic layout. The file may be opened and modified with the Macro Modeler. The graphic description of the model may
be copied (as a .bmp file) to the operating system clipboard then pasted into other word processing and graphics software.
The graphic may also be printed. The toolbar icons for Macro Modeler file management are shown in Figure 3.
Figure 3
Models may be run at any stage of completion. The Macro Modeler first checks to see if any of the output images already
exist. If so, it shows a message indicating which file exists and asks if you wish to continue. Answering Yes (or Yes to All)
will cause the existing output file to be deleted before the model runs. Answering No will prevent the Macro Modeler
from running the model. As the model runs, the module element that is currently being processed turns green. To stop a
model at any point while it is running, click the stop icon. The model will finish the process it is currently doing and will
then stop.
DynaGroups
Two distinctive capabilities of the Macro Modeler are its use of DynaGroups to facilitate running the same model on mul-
tiple data layers and DynaLinks to construct iterative models in which the output of a process becomes an input for the
next iteration of the process.
A DynaGroup is a raster group file or time series file that is used in the Macro Modeler so each member of the group is
used to produce an output. Contrast this with the use of a group file as a regular data input to the Macro Modeler in
which a group file used as input produces a single output as with the module COUNT. More than one DynaGroup may
be used in a process. The module OVERLAY, for example, requires two raster images as input and produces a single ras-
ter image output. If DynaGroups are used as the two inputs, then Macro Modeler will first verify that the same number of
members is in each group, then it will run OVERLAY using corresponding images from the two group files. If a single
raster layer is used as one input and a DynaGroup is used as the other input to OVERLAY, then the modeler will run
OVERLAY with the single raster image and each group member in turn. In all cases the number of output images is equal
to the number of members in the DynaGroup file. Several options for naming the output files are described in the Help
System. In addition, a group file of the output files is automatically constructed. In the model shown in Figure 5, two
DynaGroups are used as input to an overlay operation. The output will be the number of raster images contained in the
input DynaGroups and a raster group file.
Figure 5
DynaLinks
DynaLinks are used to run a model iteratively. An output data element is joined to an initial input data element with a
DynaLink. This indicates that when the model performs the second iteration, the linked output filename should be substi-
tuted in place of the linked input filename. A DynaLink may only be connected to an initial data layer, i.e., one which has
no connector joining to its left-hand side. Macro Modeler asks how many iterations are to be performed and whether each
terminal output of the model or only the terminal output of the final iteration should be displayed. The final output of a
model that includes a DynaLink is a single raster layer. The model shown in Figure 6 uses a DynaLink. When the model is
Figure 6
Note that the entire model is run during each iteration, even if the substitution occurs near the end of the model. To run
just a portion of a model iteratively, use a submodel that contains a DynaLink (see below).
Figure 7
To iterate or loop only a portion of the model, create a submodel for the portion that should loop. When you run the
model, a dialog box for each submodel will appear asking for the number of iterations for that submodel.
Image Calculator
Image Calculator is a calculator tool for quickly evaluating equations using raster image layers. Equations can be saved and
edited at a later time. It is extremely easy to use since it works exactly like a scientific calculator and offers the functionality
typically associated with calculators. The only special requirement is that all image files must be specified using square
brackets as delimiters. For instance, in the example shown in Figure 8, a calibrated image band is created by multiplying its
values by a gain of 0.9421 and then adding an offset of 0.0168. It can be accessed through its icon on the tool bar, or from
several menu entries: under the Mathematical Operators and Database Query sections of the GIS Analysis menu, and the
Figure 8
Introduction
As illustrated in the Introduction to GIS chapter, the database is at the center of the Geographic Information System. In
fact, it provides fuel for the GIS. The tasks of finding, creating, assembling and integrating these data may collectively be
termed database development. While it is seldom the most interesting part, database development can easily consume 80 to
90 percent of the time and resources allocated to any project. This chapter discusses many of the issues encountered in
database development and also presents a summary of the database development tools included in IDRISI. Also pre-
sented are techniques and tips for importing raster data, particularly satellite imagery.
Collecting Data
The first stage in database development is typically the identification of the data layers necessary for the project. While it
is tempting to acquire every available data layer that coincides with the study area, it is usually more efficient to identify
what is needed prior to assessing what is available. This helps to avoid the problems of data-driven project definitions and
encourages a question-driven approach.
In addition to identifying the themes that are necessary (e.g., elevation, landcover), it is also important to determine what
resolution (precision) and what level of accuracy are required. These issues will be discussed more thoroughly below.
Once the necessary data requirements for the project have been specified, the search for data can begin. There are five
main ways to get data into the database:
1. Find data in digital format and import it;
2. Find data in hard-copy format and digitize it;
3. Collect data yourself in the field, then enter it;
4. Substitute an existing data layer as a surrogate;
5. Derive new data from existing data.
Physical Media
Data might come on a CD-ROM or DVD. When acquiring data, be sure to ask about the medium on which it will arrive
and make sure you have the ability to read that medium. The ability to download large data files electronically has lessened
this problem substantially, but even with network download, access to the network and appropriate software, such as an
FTP client, are necessary.
Data Formats
You must also determine the file format of the data. In general terms, you should know which category it falls within. For
example, a data layer of Superfund sites could be stored as a raster, vector, database or spreadsheet data file. More specif-
ically, you will need to know if it is in a particular software format, an agency format, or some sort of data interchange for-
mat. IDRISI provides for direct import of a number of standard raster, vector and attribute data formats.
If the particular file format you have is not directly supported by IDRISI import routines, then you may still be able to use
it. Raster data tends to have a simpler structure than vector data and it is often the case that a raster file can be imported
into IDRISI using a low-level tool, such as GENERICRASTER, to re-order the pixels. Vector data are more complex,
and unless the structure is extremely simplistic, it is unlikely that you will be able to import vector data using low-level
tools.
For both raster and vector data, translation software can provide the bridge between the current data format and IDRISI.
Translation software for GIS file formats vary in the number of formats they support and in price. Your goal in using a
translation software is to turn the file you have into something that you can import with IDRISI.
You must also consider data compression when acquiring data. Compression allows files to be stored using less memory
than they would normally require. There are two types of compression of which to be aware. The first is compression that
is used by the GIS software itself. IDRISI, for example, has a run-length encoded packed binary file type. While IDRISI
can use these files analytically, other GIS software will not be able to use them. It is best to avoid sharing files that are
compressed with software-specific routines between different GIS software systems.
The second type of compression, sometimes referred to as "zipping", is used for files that are not currently being used.
This compression is used to make the transfer and storage of files easier. The files must be uncompressed or "unzipped"
prior to use.
Always inquire about the compression software used. If you don't have the same software, get the file in an uncompressed
format, or as a self-extracting executable file (a compressed file that will uncompress itself without the compression soft-
ware installed on your computer).
Information about importing specific file formats is available in the IDRISI on-line Help System.
Considerations
Resolution
In all database development tasks, no matter the source of data, there are several issues to consider. The first consider-
ation is the resolution of the data. How much detail do you need? Resolution affects storage and processing time if too
fine and limits the questions you can ask of the data if too coarse.
Resolution may refer to the spatial, temporal or attribute components of the database. Spatial resolution refers to the size
of the pixel in raster data or the scale of the map that was digitized to produce vector data. Temporal resolution refers to
the currency of the data and whether the images in the time series are frequent enough and spaced appropriately. Attri-
1. Several GPS companies (e.g., Trimble, Magellan) have excellent Web sites explaining how positions are calculated from the satellite signals.
Accuracy
Accuracy is the second consideration. While accuracy does have a relationship to resolution, it is also a function of the
methods and care with which the data were collected and digitized. Unfortunately, it is not always easy to evaluate the
accuracy of data layers. However, most government mapping agencies do have mapping standards that are available. The
IDRISI metadata (documentation) file structure includes fields for accuracy information, but many other file formats
carry no accuracy information. In addition, even when such information is reported, the intelligent use of accuracy infor-
mation in GIS analysis is still an avenue of active research. The capabilities of IDRISI in this area are discussed in the
chapter on Decision Support.
Georeferencing
The third consideration is georeferencing. Data layers from various sources will often be georeferenced using different
reference systems.2 For display and GIS analysis, all data layers that are to be used together must use the same reference
system. Providing that the details of the reference systems are known, they can usually be converted with the IDRISI
module PROJECT.
However, it may be possible to download graphic files from the Web, for example, that are not georeferenced. Also, while
IDRISI supports many projections, you may find data that are in an unsupported projection. In both cases, you will not
be able to use PROJECT. It is sometimes possible to georeference an unreferenced file or a file with an unsupported pro-
jection through resampling (the IDRISI module RESAMPLE) if points of known locations can be found on the unrefer-
enced image. All users are encouraged to read the chapter on Georeferencing for more information about this key issue.
Also, see the section on Data Integration below.
Cost
Finally, the fourth consideration in database development is cost. This must be assessed in terms of both time and money.
Other considerations include whether the data will be used once or many times, the accuracy level necessary for the par-
ticular (and future) uses of the data, and how often it must be updated to remain useful. Don't underestimate the amount
of labor required to develop a good database. As stated in the introduction, database development quite commonly con-
sumes a large portion of the labor allocated to a GIS project.
2. A reference system is defined by a projection, datum and grid system. See the chapter on Georeferencing for more information.
Band-interleaved-by-pixel (BIP)
with the bands 12 columns by 4 rows
Band-interleaved-by-line (BIL)
with three bands 4 columns by 12 rows
Figure 1
With BSQ format files, the following two import routines apply:
1. If a header is attached to the data file, use the GENERICRASTER module, specify BSQ and no header. GENERI-
CRASTER will remove the header, rename the data file to have an .RST extension, and create a documentation file.
GENERICRASTER will ask for the number of rows and columns, the minimum information necessary to create the
documentation file. GENERICRASTER will ask for additional information as well, such as the reference system parame-
ters which you should be able to find in the header. When in doubt, you can try the following values: Minimum X=0,
Maximum X=1 (or # of columns), Minimum Y=0, Maximum Y=1 (or # of rows), Reference System=Plane.
2. If the header file is separate from the data file, then run GENERICRASTER and use the BSQ option and specify one
band. Then specify the appropriate reference information as in the step above. Again, you must know the number of col-
umns and rows in order to complete this step, but you should also enter any reference system information that is available
from the header. If the values are unknown, try entering the values noted in the previous paragraph, with the addition of
3. The documentation for the SPOT panchromatic band will indicate that it is in BIL format, but it contains only one band, so there is no interleaving.
This is the same as BSQ format.
Figure 2
Conclusions
Database development is seldom quick and easy. The quality of the database, in many ways, limits the quality of any result-
ing analysis to be carried out with the data. There are a significant number of technical pitfalls that are typically encoun-
tered during database development and it is easy to become overwhelmed by these. However, despite these difficulties, it
4. The reference system projection, datum and grid system must all match for the reference system to match.
Geodesy
Geodesy is that field of study which is concerned with the measurement of the size and shape of the earth and positions
upon it. The most fundamental problem that geodesists face is the fact that the earth's surface is irregular in shape. For
example, imagine two locations (A and B) at either end of a thin straight beach bordered by a steep cliff (Figure 1). Clearly
the distance one would determine between the two locations would depend upon the route chosen to undertake the mea-
surement. Measuring the distance by the cliff route would clearly be longer than that determined along the coast. Thus
irregularities (hills and valleys) along the measuring surface can cause ambiguities in distance (and thereby, location). To
alleviate this situation, it has long been a common practice to reduce all measurements to a more regular measuring sur-
face—a reference surface.
Figure 1
Geoid
The oldest reference surface used for mapping is known as the geoid. The geoid can be thought of as mean sea level, or
where mean sea level would be if the oceans could flow under the continents. More technically, the geoid is an equipotential
surface of gravity defining all points in which the force of gravity is equivalent to that experienced at the ocean's surface.
Since the earth spins on its axis and causes gravity to be counteracted by centrifugal force progressively towards the equa-
tor, one would expect the shape of the geoid to be an oblate spheroid—a sphere-like object with a slightly fatter middle
and flattened poles. In other words, the geoid would have the nature of an ellipse of revolution—an ellipsoid.
As a reference surface, the geoid has several advantages—it has a simple physical interpretation (and an observable posi-
tion along the coast), and it defines the horizontal for most traditional measuring instruments. Thus, for example, leveling
a theodolite or sextant is, by definition, a process of referring the instrument to the geoid.
earth surface
geoid
reference ellipsoid
Figure 2
Geodetic Datums
Selecting a specific reference ellipsoid to use for a specific area and orienting it to the landscape, defines what is known in
Geodesy as a datum (note that the plural of datum in geodesy is datums, not data!). A datum thus defines an ellipsoid (itself
defined by the major and minor semi-axes), an initial location, an initial azimuth (a reference direction to define the direc-
tion of north), and the distance between the geoid and the ellipsoid at the initial location. Establishing a datum is the task
of geodetic surveyors, and is done in the context of the establishment of national or international geodetic control survey
networks. A datum is thus intended to establish a permanent reference surface, although recent advances in survey tech-
nology have led many nations to redefine their current datums.
Most datums only attempt to describe a limited portion of the earth (usually on a national or continental scale). For exam-
ple, the North American Datum (NAD) and the European Datum each describe large portions of the earth, while the
Kandawala Datum is used for Sri Lanka alone. Regardless, these are called local datums since they do not try to describe
the entire earth. By contrast, we are now seeing the emergence of World Geodetic Systems (such as WGS84) that do try to
1. The major and minor semi-axes are also commonly referred to as the semi-major and semi-minor axes.
Cartographic Transformation
Once a logic has been established for referring to earth locations and a set of measurements has been made, a means of
storing and analyzing those positions is required. Traditionally, maps have been the preferred medium for both storage
and analysis, while today that format has been supplemented by digital storage and analysis. Both, however, share a com-
mon trait—they are most commonly flat! Just as flat maps are a more manageable medium than map globes, plane coor-
dinates are a more workable medium than spherical (ellipsoidal) coordinates for digital applications. As a result, surveyed
locations are commonly transformed to a plane grid referencing system before use.
Projection
The process of transforming spheroidal geodetic coordinates to plane coordinate positions is known as projection, and
falls traditionally within the realm of cartography. Originally, the concern was only with a one-way projection of geodetic
coordinates to the plane coordinates of a map sheet. With the advent of GIS, however, this concern has now broadened
to include the need to undertake transformations in both directions in order to develop a unified database incorporating
maps that are all brought to a common projection. Thus, for example, a database developed on a Transverse Mercator
projection might need to incorporate direct survey data in geodetic coordinates along with map data in several different
projections. Back projecting digitized data from an existing projection to geodetic coordinates and subsequently using a for-
ward projection to bring the data to the final projection is thus a very common activity in GIS. The PROJECT module in
IDRISI supports both kinds of transformation.
Projection of spheroidal positions to a flat plane simply cannot be done without some (and oftentimes considerable) dis-
tortion. The stretching and warping required to make the transformation work leads to continuous differences in scale
over the face of the map, which in turn leads to errors in distance, area and angular relationships. However, while distor-
Georeferencing in IDRISI
The logic that IDRISI uses for georeferencing is quite simple, and is based on the issues previously described.
2. Note, however, that geodetic coordinates are not considered to be identical to a system based on the Plate Carrée since no new numbering scheme has
been undertaken. PROJECT supports transformations based on the Plate Carrée as well, where it is understood that a true plane coordinate system with
its own special numbering scheme has been superimposed onto the projected geodetic coordinates.
The first line is simply a title and has no analytical use. The second line indicates the projection. In the current release of
IDRISI, the following projections are supported:
Mercator
Transverse Mercator
Gauss-Kruger (Gauss-Krueger spelling also accepted.)
Lambert Conformal Conic
Plate Carrée
Hammer Aitoff
Lambert North Polar Azimuthal Equal Area
Lambert South Polar Azimuthal Equal Area
Lambert Transverse Azimuthal Equal Area
Lambert Oblique Azimuthal Equal Area
North Polar Stereographic
South Polar Stereographic
Transverse Stereographic
Oblique Stereographic
Albers Equal Area Conic
Sinusoidal
none (i.e., geodetic coordinates)
Note that each of the names above are keywords and must appear exactly in this form to be understood by the PROJECT module.
Note that the latitude of the "lower" standard line (i.e., that which is nearer the equator) is listed first and that of the lati-
tudinally "higher" standard line is listed second. Future additions of other projections may dictate the need for additional
special parameters. However, in this version, only those reference systems based on the Lambert Conformal Conic and
Alber's Equal Area Conic projections require special parameters.
Further Reading
Geodesy is not a familiar topic for many in GIS. However, as one can see from the material in this chapter, it is essential
to effective database development. The following readings provide a good overview of the issues involved:
3. Note that while a set of constants exist for the North American Datum as a whole, Molodensky constants also exist for more specific portions of the
regions covered to enhance the precision of the transformation.
Smith, J.R., (1988) Basic Geodesy, (Rancho Cordova, CA: Landmark Enterprises).
For projections, there are few texts that can match the scope and approachability of:
Maling, D.H., (1973) Coordinate Systems and Map Projections, (London: George Phillip and Son).
Snyder, J.P., (1987) Map Projections: A Working Manual. U.S. Geological Survey Professional Paper 1395, (Wash-
ington, DC: US Government Printing Office).
Finally, with respect to the procedures used for datum transformations, please refer to:
DMA, (1987) Department of Defense World Geodetic System 1984: Its Definition and Relationships with Local Geo-
detic Datums, DMA TR 8350.2, (Washington, DC: The Defense Mapping Agency).
Dewhurst, W.T., (1990) NADCON: The Application of Minimum Curvature-Derived Surfaces in the Transformation
of Positional Data from the North American Datum of 1927 to the North American Datum of 1983, NOAA
Technical Memorandum NOS NGS-50, (Rockville, MD: National Geodetic Information Center).
Introduction1
Decision Theory is concerned with the logic by which one arrives at a choice between alternatives. What those alterna-
tives are varies from problem to problem. They might be alternative actions, alternative hypotheses about a phenomenon,
alternative objects to include in a set, and so on. In the context of GIS, it is useful to distinguish between policy decisions
and resource allocation decisions. The latter involves decisions that directly affect the utilization of resources (e.g., land)
while the former is only intended to influence the decision behavior of others who will in turn make resource commit-
ments. GIS has considerable potential in both arenas.
In the context of policy decisions, GIS is most commonly used to inform the decision maker. However, it also has poten-
tial (almost entirely unrealized at this time) as a process modeling tool, in which the spatial effects of predicted decision
behavior might be simulated. Simulation modeling, particularly of the spatial nature of socio-economic issues and their
1. The introductory material in this chapter is adapted from Eastman, J.R., 1993. Decision Theory and GIS, Proceedings, Africa GIS '93, UNITAR,
Geneva. (out of print)
Definitions
Decision
A decision is a choice between alternatives. The alternatives may represent different courses of action, different hypothe-
ses about the character of a feature, different classifications, and so on. We call this set of alternatives the decision frame.
Thus, for example, the decision frame for a zoning problem might be [commercial residential industrial]. The decision
frame, however, should be distinguished from the individuals to which the decision is being applied. We call this the candi-
date set. For example, extending the zoning example above, the set of all locations (pixels) in the image that will be zoned is
the candidate set. Finally, a decision set is that set of all individuals that are assigned a specific alternative from the decision
frame. Thus, for example, all pixels assigned to the residential zone constitute one decision set. Similarly, those belonging
to the commercial zone constitute another. Therefore, another definition of a decision would be to consider it the act of
assigning an individual to a decision set. Alternatively, it can be thought of as a choice of alternative characterizations for
an individual.
Criterion
A criterion is some basis for a decision that can be measured and evaluated. It is the evidence upon which an individual
can be assigned to a decision set. Criteria can be of two kinds: factors and constraints, and can pertain either to attributes
of the individual or to an entire decision set.
Factors
A factor is a criterion that enhances or detracts from the suitability of a specific alternative for the activity under consider-
ation. It is therefore most commonly measured on a continuous scale. For example, a forestry company may determine
that the steeper the slope, the more costly it is to transport wood. As a result, better areas for logging would be those on
shallow slopes — the shallower the better. Factors are also known as decision variables in the mathematical programming lit-
erature (see Feiring, 1986) and structural variables in the linear goal programming literature (see Ignizio, 1985).
Constraints
A constraint serves to limit the alternatives under consideration. A good example of a constraint would be the exclusion
from development of areas designated as wildlife reserves. Another might be the stipulation that no development may
Decision Rule
The procedure by which criteria are selected and combined to arrive at a particular evaluation, and by which evaluations
are compared and acted upon, is known as a decision rule. A decision rule might be as simple as a threshold applied to a
single criterion (such as, all regions with slopes less than 35% will be zoned as suitable for development) or it may be as
complex as one involving the comparison of several multi-criteria evaluations.
Decision rules typically contain procedures for combining criteria into a single composite index and a statement of how
alternatives are to be compared using this index. For example, we might define a composite suitability map for agriculture
based on a weighted linear combination of information on soils, slope, and distance from market. The rule might further
state that the best 5000 hectares are to be selected. This could be achieved by choosing that set of raster cells, totaling
5000 hectares, in which the sum of suitabilities is maximized. It could equally be achieved by rank ordering the cells and
taking enough of the highest ranked cells to produce a total of 5000 hectares. The former might be called a choice function
(known as an objective function or performance index in the mathematical programming literature—see Diamond and Wright,
1989) while the latter might be called a choice heuristic.
Choice Function
Choice functions provide a mathematical means of comparing alternatives. Since they involve some form of optimization
(such as maximizing or minimizing some measurable characteristic), they theoretically require that each alternative be
evaluated in turn. However, in some instances, techniques do exist to limit the evaluation only to likely alternatives. For
example, the Simplex Method in linear programming (see Feiring, 1986) is specifically designed to avoid unnecessary eval-
uations.
Choice Heuristic
Choice heuristics specify a procedure to be followed rather than a function to be evaluated. In some cases, they will pro-
duce an identical result to a choice function (such as the ranking example above), while in other cases they may simply
provide a close approximation. Choice heuristics are commonly used because they are often simpler to understand and
easier to implement.
Objective
Decision rules are structured in the context of a specific objective. The nature of that objective, and how it is viewed by
the decision makers (i.e., their motives) will serve as a strong guiding force in the development of a specific decision rule.
An objective is thus a perspective that serves to guide the structuring of decision rules.2 For example, we may have the
Evaluation
The actual process of applying the decision rule is called evaluation.
Multi-Criteria Evaluations
To meet a specific objective, it is frequently the case that several criteria will need to be evaluated. Such a procedure is
called Multi-Criteria Evaluation (Voogd, 1983; Carver, 1991). Another term that is sometimes encountered for this is model-
ing. However, this term is avoided here since the manner in which the criteria are combined is very much influenced by the
objective of the decision.
Multi-criteria evaluation (MCE) is most commonly achieved by one of two procedures. The first involves Boolean overlay
whereby all criteria are reduced to logical statements of suitability and then combined by means of one or more logical
operators such as intersection (AND) and union (OR). The second is known as Weighted Linear Combination (WLC)
wherein continuous criteria (factors) are standardized to a common numeric range, and then combined by means of a
weighted average. The result is a continuous mapping of suitability that may then be masked by one or more Boolean con-
straints to accommodate qualitative criteria, and finally thresholded to yield a final decision.
While these two procedures are well established in GIS, they frequently lead to different results, as they make very differ-
ent statements about how criteria should be evaluated. In the case of Boolean evaluation, a very extreme form of decision
making is used. If the criteria are combined with a logical AND (the intersection operator), a location must meet every cri-
terion for it to be included in the decision set. If even a single criterion fails to be met, the location will be excluded. Such
a procedure is essentially risk-averse, and selects locations based on the most cautious strategy possible—a location suc-
ceeds in being chosen only if its worst quality (and therefore all qualities) passes the test. On the other hand, if a logical
OR (union) is used, the opposite applies—a location will be included in the decision set even if only a single criterion
passes the test. This is thus a very gambling strategy, with (presumably) substantial risk involved.
Now compare these strategies with that represented by weighted linear combination (WLC). With WLC, criteria are per-
mitted to tradeoff their qualities. A very poor quality can be compensated for by having a number of very favorable quali-
ties. This operator represents neither an AND nor an OR—it lies somewhere in between these extremes. It is neither risk
averse nor risk taking.
For reasons that have largely to do with the ease with which these approaches can be implemented, the Boolean strategy
dominates vector approaches to MCE, while WLC dominates solutions in raster systems. But clearly neither is better—
they simply represent two very different outlooks on the decision process—what can be called a decision strategy. IDRISI
also includes a third option for multi-criteria evaluation, known as an Ordered Weighted Average (OWA) (Eastman and
Jiang, 1996). This method offers a complete spectrum of decision strategies along the primary dimensions of degree of
tradeoff involved and degree of risk in the solution.
Multi-Objective Evaluations
While many decisions we make are prompted by a single objective, it also happens that we need to make decisions that
satisfy several objectives. A multi-objective problem is encountered whenever we have two candidate sets (i.e., sets of enti-
ties) that share members. These objectives may be complementary or conflicting in nature (Carver, 1991: 322).
2. It is important to note here that we are using a somewhat broader definition of the term objective than would be found in the goal programming lit-
erature (see Ignizio, 1985). In goal programming, the term objective is synonymous with the term objective function in mathematical programming and choice
function used here.
Database Uncertainty
Database uncertainty is that which resides in our assessments of the criteria which are enumerated in the decision rule.
Measurement error is the primary source of such uncertainty. For example, a slope of 35% may represent an important
threshold. However, because of the manner in which slopes are determined, there may be some uncertainty about
whether a slope that was measured as 34% really is 34%. While we may have considerable confidence that it is most likely
around 34%, we may also need to admit that there is some finite probability that it is as high as 36%. Our expression of
database uncertainty is likely to rely upon probability theory.
Decision Risk
Decision Risk may be understood as the likelihood that the decision made will be wrong.3 Risk arises as a result of uncer-
tainty, and its assessment thus requires a combination of uncertainty estimates from the various sources involved (data-
base and decision rule uncertainty) and procedures, such as Bayesian Probability theory, through which it can be
determined. Again, this topic will be discussed more thoroughly in Part B of this chapter.
3. Note that different fields of science define risk in different ways. For example, some disciplines modify the definition given here to include a measure
of the cost or consequences of a wrong decision (thus allowing for a direct relationship to cost/benefit analysis). The procedures developed in IDRISI
do not preclude such an extension. We have tried here to present a fairly simple perspective that can be used as a building block for more specific inter-
pretations.
Figure 1
Decisions may be characterized as single- or multi-objective in nature, based on either a single criterion or multiple criteria. While
one is occasionally concerned with single criterion problems, most problems approached with a GIS are multi-criteria in
nature. For example, we might wish to identify areas of concern for soil erosion on the basis of slope, landuse, soil type
and the like. In these instances, our concern lies with how to combine these criteria to arrive at a composite decision. As a
consequence, the first major area of concern in GIS with regard to Decision Theory is Multi-Criteria Evaluation.
Most commonly, we deal with decision problems of this nature from a single perspective. However, in many instances, the
problem is actually multi-objective in nature (Diamond and Wright, 1988). Multi-objective problems arise whenever the
same resources belong to more than one candidate set. Thus, for example, a paper company might include all forest areas
in its candidate set for consideration of logging areas, while a conservation group may include forest areas in a larger can-
didate set of natural areas to be protected. Any attempt, therefore, to reconcile their potential claims to this common set
of resources presents a multi-objective decision problem.
Despite the prevalence of multi-objective problems, current GIS software is severely lacking in techniques to deal with
this kind of decision. To date, most examples of multi-objective decision procedures in the literature have dealt with the
problem through the use of linear programming optimization (e.g., Janssen and Rietveld 1990; Carver, 1991; Campbell et.
al., 1992; Wright et. al., 1983). However, in most cases, these have been treated as choice problems between a limited num-
ber (e.g., less than 20) of candidate sites previously isolated in a vector system. The volume of data associated with raster
applications (where each pixel is a choice alternative) clearly overwhelms the computational capabilities of today's com-
puting environment. In addition, the terminology and procedures of linear programming are unknown to most decision
makers and are complex and unintuitive by nature. As a consequence, the second major area of Decision Theory of
importance to GIS is Multi-Objective Land Allocation. Here, the focus will be on a simple decision heuristic appropriate
to the special needs of raster GIS.
Criterion Scores
Because of the different scales upon which criteria are measured, it is necessary that factors be standardized4 before com-
bination using the formulas above, and that they be transformed, if necessary, such that all factors maps are positively cor-
related with suitability.5 Voogd (1983: 77-84) reviews a variety of procedures for standardization, typically using the
minimum and maximum values as scaling points. The simplest is a linear scaling such as:
xi = (Ri-Rmin) / (Rmax-Rmin) * standardized_range where R = raw score
However, if we recognize that continuous factors are really fuzzy sets, we easily recognize this as just one of many possible
set membership functions. In IDRISI, the module named FUZZY is provided for the standardization of factors using a
whole range of fuzzy set membership functions. The module is quick and easy to use, and provides the option of stan-
dardizing factors to either a 0-1 real number scale or a 0-255 byte scale. This latter option is recommended because the
MCE module has been optimized for speed using a 0-255 level standardization. Importantly, the higher value of the stan-
dardized scale must represent the case of being more likely to belong to the decision set.
A critical issue in the standardization of factors is the choice of the end points at which set membership reaches either 0.0
or 1.0 (or 0 and 255). Our own research has suggested that blindly using a linear scaling (or indeed any other scaling)
between the minimum and maximum values of the image is ill advised. In setting these critical points for the set member-
ship function, it is important to consider their inherent meaning. Thus, for example, if we feel that industrial development
should be placed as far away from a nature reserve as possible, it would be dangerous to implement this without careful
consideration. Taken literally, if the map were to cover a range of perhaps 100 km from the reserve, then the farthest point
away from the reserve would be given a value of 1.0 (or 255 for a byte scaling). Using a linear function, then, a location 5
km from the reserve would have a standardized value of only 0.05 (13 for a byte scaling). And yet it may be that the pri-
mary issue was noise and minor disturbance from local citizens, for which a distance of only 5 kilometers would have
been equally as good as being 100 km away. Thus the standardized score should really have been 1.0 (255). If an MCE
were undertaken using the blind linear scaling, locations in the range of a few 10s of km would have been severely deval-
ued when it fact they might have been quite good. In this case, the recommended critical points for the scaling should
have been 0 and 5 km. In developing standardized factors using FUZZY, then, careful consideration should be given to
the inherent meaning of the end points chosen.
Criterion Weights
A wide variety of techniques exist for the development of weights. In very simple cases, assigning criteria weights may be
accomplished by dividing 1.0 among the criteria. (It is sometimes useful for people to think about "spending" one dollar,
for example, among the criteria). However, when the number of criteria is more than a few, and the considerations are
4. In using the term standardization, we have adopted the terminology of Voogd (1983), even though this process should more properly be called normal-
ization.
5. Thus, for example, if locations near to a road were more advantageous for industrial siting than those far away, a distance map would need to be trans-
formed into one expressing proximity.
In developing the weights, an individual or group compares every possible pairing and enters the ratings into a pairwise
comparison matrix (Figure 3). Since the matrix is symmetrical, only the lower triangular half actually needs to be filled in.
The remaining cells are then simply the reciprocals of the lower triangular half (for example, since the rating of slope gra-
dient relative to town proximity is 4, the rating of town proximity relative to slope gradient will be 1/4). Note that where
empirical evidence exists about the relative efficacy of a pair of factors, this evidence can also be used.
Road Proximity 1
Slope Gradient 1 4 1
RoadProx 0.33
TownProx 0.08
Slope 0.34
SmalHold 0.07
Since the complete pairwise comparison matrix contains multiple paths by which the relative importance of criteria can be
assessed, it is also possible to determine the degree of consistency that has been used in developing the ratings. Saaty
(1977) indicates the procedure by which an index of consistency, known as a consistency ratio, can be produced (Figure 4).
The consistency ratio (CR) indicates the probability that the matrix ratings were randomly generated. Saaty indicates that
matrices with CR ratings greater than 0.10 should be re-evaluated. In addition to the overall consistency ratio, it is also
possible to analyze the matrix to determine where the inconsistencies arise. This has also been developed as part of the
WEIGHT module in IDRISI.
Evaluation
Once the criteria maps (factors and constraints) have been developed, an evaluation (or aggregation) stage is undertaken
to combine the information from the various factors and constraints. The MCE module offers three logics for the evalu-
ation/aggregation of multiple criteria: Boolean intersection, weighted linear combination (WLC), and the ordered
weighted average (OWA).
6. The logic of the Boolean AND and OR is implemented with fuzzy sets as the minimum and maximum. Thus, as we are considering continuous factor
scores rather than Boolean 0-1 images in this discussion, the logical AND is evaluated as the minimum value for a pixel across all factors and the logical
OR is evaluated as the maximum value for a pixel across all factors.
The first set of order weights in the table is [1, 0, 0]. The weight of factor B (the factor with the minimum value in the set
[B, A, C]) will receive all possible weight while factors A and C will be given no weight at all. Such a set of order weights
make irrelevant the factor weights. Indeed, the order weights have altered the evaluation such that no tradeoff is possible.
As can be seen in the table, this has the effect of applying a minimum operator to the factors, thus producing the tradi-
tional intersection operator (AND) of fuzzy sets.
Similarly, the last set of order weights [0, 0, 1] has the effect of a maximum operator, the traditional union operator (OR)
of fuzzy sets. Again, there is no tradeoff and the factor weights are not employed.
Another important example from the table is where the order weights are equal, [.33, .33, .33]. Here all ranked positions
get the same weight; this makes tradeoff fully possible and locates the analysis exactly midway between AND and OR.
Equal order weights produce the same result as WLC.
In all three cases, the order weights have determined not only the level of tradeoff but have situated the analysis on a con-
tinuum from (risk-averse, minimum, AND) to (risk-taking, maximum, OR).
As seen in the table, the order weights in the OWA option of MCE are not restricted to these three possibilities, but
instead can be assigned any combination of values that sum to 1.0. Any assignment of order weights results in a decision
rule that falls somewhere in a triangular decision strategy space that is defined by the dimensions of risk and tradeoff as
shown in Figure 5.
TRADEOFF
decision
strategy
space
no
tradeoff
risk-averse RISK risk-taking
(AND) (OR)
Figure 5
Whether most of the order weight is assigned to the left, right or center of the order weights determines the position in
the risk dimension. The logical AND operator is the most risk-averse combination and the logical OR is the most risk-
taking combination. When order weights are predominantly assigned to the lower-ranked factors, there is greater risk
aversion (more of an AND approach). When order weights are more dominant for the higher-ranked factors, there is
greater risk taking (more of an OR approach). As discussed above, equal order weights yield a solution at the middle of
the risk axis.
The degree of tradeoff is governed by the relative distribution of order weights between the ranked factors. Thus, if the
sum of the order weights is evenly spread between the factors, there is strong tradeoff, whereas if all the weight is assigned
to a single factor rank, there is no tradeoff. (It may be helpful to think of this in terms of a graph of the order weights,
with rank order on the X axis and the order weight value on the Y axis. If the graph has a sharp peak, there is little trad-
eoff. If the graph is relatively flat, there is strong tradeoff.)
Thus, as seen from the table, the order weights of [0.5 0.3 0.2] would indicate a strong (but not perfect) degree of risk
aversion (because weights are skewed to the risk-averse side of the risk axis) and some degree of tradeoff (because the
weights are spread out over all three ranks). Weights of [0 1 0], however, would imply neither risk aversion nor acceptance
(exactly in the middle of the risk axis), and no tradeoff (because all the weight is assigned to a single rank).
The OWA method is particularly interesting because it provides this continuum of aggregation procedures. At one
extreme (the logical AND), each criterion is considered necessary (but not sufficient on its own) for inclusion in the deci-
sion set. At the other extreme (the logical OR), each criterion is sufficient on its own to support inclusion in the decision
set without modification by other factors. The position of the weighted linear combination operator halfway between
these extremes is therefore not surprising. This operator considers criteria as neither necessary nor sufficient—strong
support for inclusion in the decision set by one criterion can be equally balanced by correspondingly low support by
another. It thus offers full tradeoff.
Using OWA
Given this introduction, it is worth considering how one would use the OWA option of MCE. Some guidelines are as fol-
lows:
1. Divide your criteria into three groups: hard constraints, factors that should not tradeoff, and factors that should trad-
eoff. For example, factors with monetary implications typically tradeoff, while those associated with some safety concern
typically do not.
2. If you find that you have factors that both tradeoff and do not tradeoff, separate their consideration into two stages of
analysis. In the first, aggregate the factors that tradeoff using the OWA option. You can govern the degree of tradeoff by
Complementary Objectives
As indicated earlier, the case of complementary objectives can be dealt with quite simply by means of a hierarchical exten-
sion of the multi-criteria evaluation process (e.g., Carver, 1991). Here a set of suitability maps, each derived in the context
of a specific objective, serve as the factors for a new evaluation in which the objectives are themselves weighted and com-
bined by linear summation. Since the logic which underlies this is multiple use, it also makes sense to multiply the result by
all constraints associated with the component objectives.
Conflicting Objectives
With conflicting objectives, land can be allocated to one objective but not more than one (although hybrid models might
combine complementary and conflicting objectives). As was indicated earlier, one possible solution lies with a prioritiza-
tion of objectives (Rosenthal, 1985). After the objectives have been ordered according to priority, the needs of higher pri-
ority objectives are satisfied (through rank ordering of cells and reclassification to meet areal goals) before those of lower
priority ones. This is done by successively satisfying the needs of higher priority objectives and then removing (as a new
constraint) areas taken by that objective from consideration by all remaining objectives. A prioritized solution is easily
achieved with the use of the RANK, RECLASS and OVERLAY modules in IDRISI. However, instances are rare where a
prioritized solution makes sense. More often a compromise solution is required.
As noted earlier, compromise solutions to the multi-objective problem have most commonly been approached through
the use of mathematical programming tools outside GIS (e.g., Diamond and Wright, 1988; Janssen and Rietveld, 1990;
Campbell, et. al., 1992). Mathematical programming solutions (such as linear or integer programming) can work quite well
in instances where only a small number of alternatives are being addressed. However, in the case of raster GIS, the mas-
sive data sets involved will typically exceed present-day computing power. In addition, the concepts and methodology of
linear and integer programming are not particularly approachable to a broad range of decision makers. As a result, we
have sought a solution to the problem of multi-objective land allocation under conditions of conflicting objectives such
that large raster datasets may be handled using procedures that have an immediate intuitive appeal.
255 255
conflict- Ideal point for -
non-conflict region desired by Objective 2
allocated to obj. 2 both
non-conflict
region
Objective 2
Objective 2
allocated
unsuitable to obj. 1
choices
Figure 6a Figure 6b
Every raster cell in the image can be located within this decision space according to its suitability level on each of the
objectives. To find the best x hectares of land for Objective 1, we simply need to move a decision line down from the top
(i.e., far right) of the Objective 1 suitability axis until enough of the best raster cells are captured to meet our area target.
We can do the same with the Objective 2 suitability axis to capture the best y hectares of land for it. As can be seen in Fig-
ure 6a, this partitions the decision space into four regions—areas best for Objective 1 and not suitable for Objective 2,
areas best for Objective 2 and not suitable for Objective 1, areas not suitable for either, and areas judged best for both.
The latter represents areas of conflict.
To resolve these areas of conflict, a simple partitioning of the affected cells is used. As can be seen in Figure 6b, the deci-
sion space can also be partitioned into two further regions: those closer to the ideal point for Objective 1 and those closer
to that for Objective 2. The ideal point represents the best possible case—a cell that is maximally suited for one objective
and minimally suited for anything else. To resolve the conflict zone, the line that divides these two regions is overlaid onto
it and cells are then allocated to their closest ideal point. Since the conflict region will be divided between the objectives,
both objectives will be short on achieving their area goals. As a result, the process will be repeated with the decision lines
being lowered for both objectives to gain more territory. The process of resolving conflicts and lowering the decision lines
is iteratively repeated until the exact area targets are achieved.
It should be noted that a 45-degree line between a pair of objectives assumes that they are given equal weight in the reso-
lution of conflicts. However, unequal weighting can be given. Unequal weighting has the effect of changing the angle of
this dividing line. In fact, the tangent of that angle is equal to the ratio of the weights assigned to those objectives.
It should also be noted that just as it was necessary to standardize criteria for multi-criteria evaluation, it is also required
for multi-objective evaluation. The process involves a matching of the histograms for the two suitability maps. In cases
where the distributions are normal, conversion to standard scores (using the module named STANDARD) would seem
appropriate. However, in many cases, the distributions are not normal. In these cases, the matching of histograms is most
easily achieved by a non-parametric technique known as histogram equalization. This is a standard option in many image
processing systems such as IDRISI. However, it is also the case that the ranked suitability maps produced by the RANK
module are also histogram equalized (i.e., a histogram of a rank map is uniform). This is fortuitous since the logic outlined
in Figure 12-6a is best achieved by reclassification of ranked suitability maps.
A Worked Example
To illustrate these multi-criteria/multi-objective procedures, we will consider the following example of developing a zon-
ing map to regulate expansion of the carpet industry (one of the largest and most rapidly growing industries in Nepal)
within agricultural areas of the Kathmandu Valley of Nepal. The problem is to zone 1500 hectares of current agricultural
land outside the ring road of Kathmandu for further expansion of the carpet industry. In addition, 6000 hectares will be
zoned for special protection of agriculture. The problem clearly falls into the realm of multi-objective/multi-criteria deci-
sion problems. In this case, we have two objectives: to protect lands that are best for agriculture, and at the same time find
other lands that are best suited for the carpet industry. Since land can be allocated to only one of these uses at any one
time, the objectives must be viewed as conflicting (i.e., they may potentially compete for the same lands). Furthermore,
the evaluation of each of these objectives can be seen to require multiple criteria.
In the illustration that follows, a solution to the multi-objective/multi-criteria problem is presented as developed with a
group of Nepalese government officials as part of an advanced seminar in GIS.7 While the scenario was developed purely
for the purpose of demonstrating the techniques used, and while the result does not represent an actual policy decision, it
is one that incorporates substantial field work and the perspectives of knowledgeable decision makers. The procedure fol-
lows a logic in which each of the two objectives is first dealt with as a separate multi-criteria evaluation problem. The
result consists of two separate suitability maps (one for each objective) which are then compared to arrive at a single solu-
tion that balances the needs of the two competing objectives.
7. The seminar was hosted by UNITAR at the International Center for Integrated Mountain Development (ICIMOD) in Nepal, September 28-October
2, 1992.
Proximity to
Roads Factor
Slope Gradient
Factor
Proximity to
Power Factor
Proximity to Mar-
ket Factor
Current
Landuse
Constraint
Ring Road
Constraint
Factor Factor
Weight
Figure 8
Figure 9 Composite Suitability images for Carpet Industry (left) and Agriculture (right).
Suitability scale corresponds to that in Figure 7.
Figure 10 Final allocation to the carpet industry (red) and agriculture (green) objectives.
A Closing Comment
The decision support tools provided in IDRISI are still under active development. We therefore welcome written com-
ments and observations to further improve the modules and enhance their application in real-world situations.
A Typology of Uncertainty
Uncertainty includes any known or unknown error, ambiguity or variation in both the database and the decision rule.
Thus, uncertainty may arise from such elements as measurement error, inherent variability, instability, conceptual ambigu-
ity, over-abstraction, or simple ignorance of important model parameters.
Considering the decision making process as a set membership problem is a useful perspective from which to understand
the source and role of uncertainty in decision making. As previously defined, a decision frame contains all the alternatives (or
hypotheses) under consideration, and evidence is that information through which set membership of a location in the deci-
sion set (the set of chosen alternatives) can be evaluated. Thus, the decision making process contains three basic elements
within which uncertainty can occur—the evidence, the decision set, and the relation that associates the two.
Error Assessment
The assessment of measurement error is normally achieved by selecting a sample of sites to visit on the ground, remeasur-
ing the attribute at those locations using some more accurate instrument, and then comparing the new measurements to
those in the data layer. To assist this procedure, IDRISI provides the SAMPLE and ERRMAT modules.
SAMPLE has the ability to lay out a sample of points (in vector format) according to a random, systematic or stratified
random scheme. The latter is usually preferred since it combines the best qualities of the other two—the unbiased charac-
ter of the random sampling scheme with the even geographic coverage of the systematic scheme.
The size of the sample (n) to be used is determined by multiplying an estimate of the standard error of the evaluation sta-
tistic being calculated by the square of the standard score (z) required for the desired level of confidence (e.g., 1.96 for
95% confidence), and dividing the result by the square of the desired confidence interval (e) (e.g., 0.01 for ±10%). For esti-
mates of the sample size required for estimating an RMS error, this formula simplifies to:
n = z2 s2 / 2e2 where s is the estimated RMS.
For estimates of the proportional error in categorical data, the formula becomes:
n = z2 pq / e2 where p is the estimated proportional error and q = (1-p).
Note that the term stratified in stratified random means that it is spatially stratified according to a systematic division of the
area into rectangular regions. In cases where some other stratification is desired, and/or where the region to be sampled is
not rectangular in shape, the following procedure can be used:
1. Determine the area of the stratum or irregular region using the AREA module and divide it by the area of the total
image. This will indicate the proportional area of the stratum or irregular region.
2. Divide the desired sample size by the proportional area. This will indicate a new (and larger) sample size that will be
required to ensure that the desired number of sample points will fall within the area of interest.
3. Run SAMPLE with the new sample size and use only those points that fall within the area of interest.
Once the ground truth has been undertaken at the sample points, the characteristic error can be assessed. In the case of
assessing quantitative data using RMS, the standard formula for RMS derived from a sample can be used:
Σ ( xi – t )
RMS = --------------------
n–1
where xi = a measurement
t = true value
True
Conifers 24 0 0 3 27 0.11
Water 2 0 0 14 16 0.12
Total 29 36 44 17 126
errors of omission
In addition to the basic error matrix, ERRMAT also reports the overall and per category Kappa Index of Agreement
(KIA) values. The Kappa Index of Agreement is similar to a proportional accuracy figure (and thus the complement of
proportional error), except that it adjusts for chance agreement.
Error Propagation
When uncertainty exists in data layers, that error will propagate through any analysis and combine with the error from
other sources. Specific formulas do exist for the expected error propagation arising from typical GIS mathematical opera-
tions (such as those involved with SCALAR and OVERLAY). Appendix 1 contains a representative set of such formulae.
In addition, IDRISI contains two modules that under certain circumstances will propagate error information automati-
cally using such procedures. The first is the MCE module described earlier in this chapter while the second is SURFACE.
If all of the input factors presented to MCE contain error (RMS) information in the value error field of their documenta-
tion files, MCE will determine the propagated output error and place it in the documentation file of the result. However,
1. Monte Carlo Simulation relies upon the use of a very large set of simulations to derive its characterizations. In cases such as this where each cell pro-
vides a new simulation, the total composite of cells can provide such a large sample. Results are improved by repeated runs of such an analysis and an
averaging of results.
True
Value
Distribution of Threshold
f measurements about
the true value
Probability that the
value exceeds the
threshold
RMS
Figure 2
With PCLASS we have the soft equivalent of a hard RECLASS operation. For example, consider the case of finding areas
that will be inundated by a rise in sea level as a result of global warming. Traditionally, this would be evaluated by reclassi-
fying the heights in a digital elevation model into two groups—those below the projected sea level and those above. With
PCLASS, however, recognition is made of the inherent error in the measurement of heights so that the output map is not
a hard Boolean map of zeros and ones, but a soft probability map that ranges continuously from zero to one. Figure 3, for
example, illustrates the output from PCLASS after evaluating the probability that heights are less than a new projected sea
level of 1.9 meters above the current level in Boston Harbor in the USA. Given this continuous probability map, a final
decision can be made by reclassifying the probability map according to the level of decision risk one is willing to assume.
Figures 4 and 5, for example, show the difference between the original coastline and that associated with the new sea level
while accepting only a 5% chance of being wrong compared to that of accepting a 25% chance. Clearly, the Digital Eleva-
tion Model (DEM) used in this assessment is not very precise. However, this illustrates the fact that even poor data can be
used effectively if we know how poor they are.
Figure 3 Probability of being Figure 4 Flood zone at 5% risk. Figure 5 Flood zone at 25% risk.
flooded.
Fuzzy Sets
Fuzzy sets are sets (or classes) without sharp boundaries; that is, the transition between membership and nonmembership
of a location in the set is gradual (Zadeh, 1965; Schmucker, 1982). A fuzzy set is characterized by a fuzzy membership
grade (also called a possibility) that ranges from 0.0 to 1.0, indicating a continuous increase from nonmembership to com-
plete membership. For example, in evaluating whether a slope is steep, we may define a fuzzy membership function such
that a slope of 10% has a membership of 0, and a slope of 25% has a membership of 1.0. Between 10% and 25%, the
fuzzy membership of a slope gradually increases on the scale from 0 to 1 (Figure 6). This contrasts with the classic crisp set
which has distinct boundaries. However, a crisp set can also be seen as a special case of fuzzy set where fuzzy membership
changes instantaneously from 0 or 1.
possibility
0.5 Fuzzy Set
Crisp Set
0.0
0 5 10 15 20 25 30 35
slope gradient (%)
Fuzzy set theory provides a rich mathematical basis for understanding decision problems and for constructing decision
rules in criteria evaluation and combination. In use, the FUZZY module in IDRISI is designed for the construction of
Fuzzy set membership functions, while the OWA option of the MCE module offers a range of appropriate aggregation
operators. FUZZY offers four types of membership function:
1. Sigmoidal: The sigmoidal ("s-shaped") membership function is perhaps the most commonly used function in fuzzy set
theory. It is produced here using a cosine function as described in the on-line Help System. In use, FUZZY requires the
positions (along the X axis) of 4 inflection points governing the shape of the curve. These are indicated in Figure 7 as points
a, b, c and d, and represent the inflection points as the membership function rises above 0, approaches 1, falls below 1
again, and finally approaches 0. The right-most function of Figure 7 shows all four inflection points as distinct. However,
this same function can take different forms. Figure 7 shows all possibilities. Beginning at the left, the monotonically increasing
function shape rises from 0 to 1 then never falls. The previously mentioned concept of steep slopes is a good example
here where the first inflection point a would be 10%, and the second b would be 25%. Since it never falls again, inflection
points c and d would be given the same value as b (FUZZY understands this convention). However, the FUZZY interface
facilitates data input in this case by requesting values only for inflection points a and b. The second curve of Figure 7
shows a monotonically decreasing function that begins at 1 then falls and stays at 0. In this case where the membership func-
tion starts at 1 and falls to 0 but never rises, a and b would be given identical values to c (the point at which it begins to
fall), and d would be given the value of the point at which it reaches 0. The FUZZY interface only requires inflection
points c and d for this type of function. The last two functions shown are termed symmetric as they rise then fall again. In
the case where the function rises and then immediately falls (the third curve in Figure 7), points b and c take on the same
value. Finally, where it rises, stays at 1 for a while, and then falls, all four values are distinct. In both cases, the FUZZY
interface requires input of all four inflection points. Note that there is no requirement of geometric symmetry for sym-
metric functions, only that the curve rise then fall again. It is quite likely that the shape of the curve between a and b and
that between c and d would be different, as illustrated in the right-most curve of Figure 7.
b,c,d a,b,c b,c b c
a d a d a d
Figure 7 Sigmoidal Membership Function
2. J-Shaped: The J-Shaped function is also quite common, although in most cases it would seem that a sigmoidal function
would be better. Figure 8 shows the different possibilities of J-shaped functions and the positions of the inflection points.
It should be pointed out that with the J-shaped function, the function approaches 0 but only reaches it at infinity. Thus the inflection points a
a d a d a d
3. Linear: Figure 9 shows the linear function and its variants, along with the position of the inflection points. This func-
tion is used extensively in electronic devices advertising fuzzy set logic, in part because of its simplicity, but also in part
because of the need to monitor output from essentially linear sensors.
a d a d a d
4. User-defined: When the relationship between the value and fuzzy membership does not follow any of the above three
functions, the user-defined function is most applicable. An unlimited number of control points may be used in this func-
tion to define the fuzzy membership curve. The fuzzy membership between any two control points is linearly interpo-
lated, as in Figure 10.
control points
In Multi-Criteria Evaluation, fuzzy set membership is used in the standardization of criteria. Exactly which function
should be used will depend on the understanding of the relationship between the criterion and the decision set, and on
the availability of information to infer fuzzy membership. In most cases, either the sigmoidal or linear functions will be
sufficient.
Dempster-Shafer Theory
Dempster-Shafer theory, an extension of Bayesian probability theory, allows for the expression of ignorance in uncer-
tainty management (Gordon and Shortliffe, 1985; Lee et al., 1987). The basic assumptions of Dempster-Shafer theory are
that ignorance exists in the body of knowledge, and that belief for a hypothesis is not necessarily the complement of belief
for its negation.
First, Dempster-Shafer theory defines hypotheses in a hierarchical structure (Figure 13-11) developed from a basic set of
hypotheses that form the frame of discernment.2 For example, if the frame of discernment includes three basic hypotheses:
{A, B, C}, the structure of hypotheses for which Dempster-Shafer will accept evidence includes all possible combinations,
[A], [B], [C], [A, B], [A, C], [B, C], and [A, B, C]. The first three are called singleton hypotheses as each contains only one
basic element. The rest are non-singleton hypotheses containing more than one basic element. Dempster-Shafer recog-
nizes these hierarchical combinations because it often happens that the evidence we have supports some combinations of
hypotheses without the ability to further distinguish the subsets. For example, we may wish to include classes of [decidu-
ous] and [conifer] in a landcover classification, and find that evidence from a black and white aerial photograph can distin-
2. The frame of discernment in Dempster-Shafer theory has essentially the same meaning as the term decision frame as used in this paper—i.e., the set of alter-
native hypotheses or classes that can be substantiated or assigned to entities. Dempster-Shafer considers these hypotheses to be exhaustive. Thus, state-
ments of support for any hierarchical combination of classes represents a degree of inability to commit to one of the singleton hypotheses in the frame
of discernment. However, in practice, Dempster-Shafer does treat these hierarchical combinations as additional hypotheses. In addition, in a GIS and
Remote Sensing context, there may be good reason to treat some unresolvable commitment to one of these hierarchical combinations as truly evidence
of an independent class/hypothesis to which entities might be assigned. For example, with a frame of discernment that includes [forest] and [wetland],
the presence of commitment to a [forest wetland] combination may in fact represent the presence of a "forested wetland" class that cannot be resolved
by attaining better evidence. As a result, we recognize here that the analyst may wish to consider the decision frame as containing all of the hierarchical
combinations, and not just the more limited set of singletons that forms the Dempster-Shafer frame of discernment. This does not violate the logic of
Dempster-Shafer, since we are simply making the post-analysis judgement that certain combinations represent new classes and thus may form a decision
set.
[A,B,C]
In expressing commitment to any of these hypotheses, Dempster-Shafer theory recognizes six important concepts: basic
probability assignment (BPA), ignorance, belief, disbelief, plausibility, and belief interval.
A basic probability assignment (BPA) represents the support that a piece of evidence provides for one of these hypothe-
ses and not its proper subsets. Thus a BPA for [A, B] represents that mass of support for [A,B], but not [A] or [B]—i.e.,
that degree of support for some indistinguishable combination of [A] and [B]. This is usually symbolized with the letter
"m" (for mass), e.g.,:
m(A,B) = basic probability assignment to [A, B]
The basic probability assignment for a given hypothesis may be derived from subjective judgment or empirical data. Since
a BPA is a fuzzy measure, the FUZZY module can also be used in IDRISI to develop a BPA from a given data set.
The sum of all BPAs will equal 1.0 at all times. Thus, the BPA for the ultimate superset ([A, B, C] in this example) will
equal the complement of the sum of all other BPAs. This quantity thus represents ignorance—the inability to commit to
any degree of differentiation between the elements in the frame of discernment.
Belief represents the total support for an hypothesis, and will be drawn from the BPAs for all subsets of that hypothesis,
i.e.,:
BEL ( X ) = Σm ( Y ) when Y ⊆ X
Thus the belief in [A, B] will be calculated as the sum of the BPAs for [A, B], [A], and [B]. In this example, belief repre-
sents the probability that an entity is A or B. Note that in the case of singleton hypotheses, the basic probability assign-
ment and belief are identical.
In contrast to belief, plausibility represents the degree to which an hypothesis cannot be disbelieved. Unlike the case in
Bayesian probability theory, disbelief is not automatically the complement of belief, but rather, represents the degree of
support for all hypotheses that do not intersect with that hypothesis. Thus:
PL ( X ) = 1 – BEL ( X ) where X = not X
thus PL ( X ) = Σm ( Y ) when Y ∩ X ≠ φ
Interpreting these constructs, we can say that while belief represents the degree of hard evidence in support of an hypoth-
esis, plausibility indicates the degree to which the conditions appear to be right for that hypothesis, even though hard evi-
dence is lacking. For each hypothesis, then, belief is the lower boundary of our commitment to that hypothesis, and
plausibility represents the upper boundary. The range between the two is called the belief interval, and represents the degree
of uncertainty in establishing the presence or absence of that hypothesis. As a result, areas with a high belief interval are
those in which new evidence will supply the greatest degree of information. Dempster-Shafer is thus very useful in estab-
lishing the value of information and in designing a data gathering strategy that is most effective in reducing uncertainty.
Σm 1 ( X ) • m 2 ( Y ) when ( X ∩ Y ) = Z
m ( Z ) = ---------------------------------------------------- ----------------------------------------------
1 – Σm 1 ( X ) • m 2 ( Y ) when ( X ∩ Y ) = φ
m ( Z ) = Σm 1 ( X ) • m 2 ( Y ) for X ∩ Y = Z .
The final belief, plausibility, and belief interval for each of the hypotheses can then be calculated based on the basic prob-
ability assignment calculated using the above equations. Ignorance for the whole set can also be derived. In most cases,
after adding new evidence, the ignorance is reduced.
Working with Dempster-Shafer Theory: Belief
In IDRISI, the Belief module can be used to implement the Dempster-Shafer logic. Belief constructs and stores the cur-
rent state of knowledge for the full hierarchy of hypotheses formed from a frame of discernment. In addition, it has the
ability to aggregate new evidence with that knowledge to create a new state of knowledge, that may be queried in the form
of map output for the belief, plausibility or belief interval associated with any hypothesis.
Belief first requires that the basic elements in the frame of discernment be defined. As soon as the basic elements are
entered, all hypotheses in the hierarchical structure will be created in the hypothesis list. For each line of evidence entered,
basic probability assignment images (in the form of real number images with a 0 - 1 range) are required with an indication
of their supported hypothesis. The BUILD KNOWLEDGE BASE item in the ANALYSIS menu then incorporates this
new evidence by recalculating the state of knowledge using the Dempster-Shafer rule of combination, from which sum-
mary images in the form of belief, plausibility or belief interval statements for each hypothesis can be selected. All the
information entered can be saved in a knowledge base file for later use when more evidence is obtained.
The Dempster-Shafer rule of combination provides an important approach to aggregating indirect evidence and incom-
plete information. Consider, for example, the problem of estimating where an archaeological site of a particular culture
might be found. The decision frame includes two basic elements, [site] and [non-site].3 Four pieces of evidence are used:
3. The total number of hypotheses that Dempster-Shafer generates in the full hierarchy is 2n-1. Implicitly, there is an extra hypothesis that is the null set,
which is assumed by Dempster-Shafer to be automatically false. Thus in this example, the [non-site] hypothesis is not the null set, nor is it automatically
assumed by Dempster-Shafer. In this example it was entered as a positive hypothesis, and member of the frame of discernment.
Figure 12 Basic Probability Assignment Images used in Aggregating Evidence for Archaeological Sites.
From left to right, the BPA’s support the hypothesis [Site] based on distance from known sites, [Site] based
on frequency of surface artifacts, [Non-Site] based on distance from permanent water, and [Non-Site]
based on slope. In all cases, darker areas represent a higher BPA.
The module Belief combines information from all four sources and has been used to produce belief, plausibility and belief
interval images for the [site] hypothesis as illustrated in Figure 13. The belief interval image is particularly interesting in
that it shows us where we have substantial uncertainty. Further sampling of evidence in these areas might prove profitable
since the conditions support the plausibility of a site, even though concrete evidence is poor.
Using Belief
1. You may find it difficult to decide whether a particular piece of evidence should be used to support the belief of an
hypothesis or, alternatively, the complement of that image should be used to support its disbelief. The latter is actually a
statement in support of the plausibility of an hypothesis, but not its belief, and is very common in GIS. For example, in
the case above, proximity to permanent water was treated as a distance image in support of disbelief in the possibility of a
site. The reason for this is that if one were near to water there is no reason to believe that a site would or would not be
present, but if one were far from water, there is excellent reason to assume that a site could not have existed. In deciding
how to treat lines of evidence, consider carefully whether the data provide true evidence in support of an hypothesis, or
simply support for its plausibility (i.e., the inability to deny its possibility).
2. To enter a disbelief, indicate that the evidence supports the collection of all hypotheses that do not include the one of
concern. In the archaeology example, distance from water was entered as evidence for [non-site]. In a case with three
Figure 13 Belief (left), Plausibility (middle) and Belief Interval (right) images for the pres-
ence of archaeological sites after Dempster-Shafer combination of evidence.
Radiometric Restoration
Sensor Calibration
The detectors on sensors can vary between instruments (such as on successive satellites in a series, such as the NOAA
TIROS-N weather satellites) and within an instrument over time or over the face of the image if multiple detectors are
used (as is commonly the case). Sensor calibration is thus concerned with ensuring uniformity of output across the face of
the image, and across time.
Radiance Calibration
Pixel values in satellite imagery typically express the amount of radiant energy received at the sensor in the form of uncal-
ibrated relative values simply called Digital Numbers (DN). Sometimes these DN are referred to as the brightness values.
For many (perhaps most) applications in remote sensing (such as classification of a single-date image using supervised
classification), it is not necessary to convert these values. However, conversion of DN to absolute radiance values is a nec-
essary procedure for comparative analysis of several images taken by different sensors (for example, Landsat-2 versus
Landsat-5). Since each sensor has its own calibration parameters used in recording the DN values, the same DN values in
two images taken by two different sensors may actually represent two different radiance values.
Usually, detectors are calibrated so that there is a linear relationship between DN and spectral radiance. This linear func-
tion is typically described by three parameters: the range of DN values in the image, and the lowest (Lmin) and highest
(Lmax) radiances measured by a detector over the spectral bandwidth of the channel. Most commonly, the data are dis-
tributed in 8-bit format corresponding to 256 DN levels. Lmin is the spectral radiance corresponding to the minimum
DN value (usually 0). Lmax is the radiance corresponding to the maximum DN (usually 255). Not only each sensor, but
each band within the same sensor, has its own Lmin and Lmax. The information about sensor calibration parameters
(Lmin and Lmax) is usually supplied with the data or is available elsewhere.1 The equation2 relating DN in remotely
sensed data to radiance is:
1. The Landsat satellites calibration parameters can be found in: EOSAT Landsat Technical Notes No. 1, August 1986, or the Landsat Data User's
Handbook.
2. For an explanation of radiance computation from DN, you may wish to consult: Lillesand, T.M., R.W. Kiefer and J.W. Chipman, 2004. Remote
Sensing and Image Interpretation. Fifth Edition. John Wiley and Sons.
Either the CALIBRATE or RADIANCE modules in IDRISI can be used to convert raw DN values to calibrated radi-
ances. The RADIANCE module has the most extensive options. It contains a lookup table of Lmin and Lmax for both
the MSS and TM sensors on Landsat 1-5. For other satellite systems, it permits the user to enter system-specific Lmin/
Lmax, or Offset/Gain values. CALIBRATE is more specifically geared towards brightness level matching, but does allow
for adjustment to a specific offset and gain. In either case, special care must be taken that the calibration coefficients are
correctly matched to the output units desired. The most common expression of radiance is in mWcm-2sr-1 mm-1 (i.e.,
milliWatts per square centimeter per steradian per micron). However, it is also common to encounter Wm-2sr-1 mm-1
(i.e., watts per square meter per steradian per micron).
Band Striping
Striping or banding is systematic noise in an image that results from variation in the response of the individual detectors
used for a particular band. This usually happens when a detector goes out of adjustment and produces readings that are
consistently much higher or lower than the other detectors for the same band. In the case of MSS data, there are 6 detec-
tors per band which scan in a horizontal direction. If one of the detectors is miscalibrated, then horizontal banding occurs
repetitively on each 6th line. Similarly, in the case of TM data with 16 detectors per band, each 16th line in the image will
be affected. Multispectral SPOT data have a pushbroom scanner with 3000 detectors for each band, one detector for each
pixel in a row. Since detectors in a pushbroom scanner are arranged in a line perpendicular to the satellite orbit track, mis-
calibration in SPOT detectors produces vertical banding. Since the SPOT satellite has an individual detector for each col-
umn of data, there is no repetitive striping pattern in the image.
The procedure that corrects the values in the bad scan lines is called destriping. It involves the calculation of the mean (or
median) and standard deviation for the entire image and then for each detector separately. Some software packages offer
an option for applying a mask to the image to exclude certain areas from these calculations (for example, clouds and cloud
shadows should be excluded). Also, sometimes only a portion of the image, usually a homogeneous area such as a body of
water, is used for these calculations. Then, depending on the algorithm employed by a software system, one of the follow-
ing adjustments is usually made:
1. The output from each detector is scaled to match the mean and standard deviation of the entire image. In this case, the
value of each pixel in the image is altered.
2. The output from the problem detector is scaled to resemble the mean and standard deviation of the other detectors. In
this case, the values of the pixels in normal data lines are not altered.
Figure 1 Figure 2
Mosaicing
Mosaicing refers to the process of matching the radiometric characteristics of a set of images that fit together to produce
a larger composite. In IDRISI, the MOSAIC module facilitates this process. The basic logic is to equalize the means and
variances of recorded values across the set of images, based on an analysis of comparative values in overlap areas. The first
image specified acts as the master, to which all other images are adjusted.
Atmospheric Correction
The atmosphere can affect the nature of remotely sensed images in a number of ways. At the molecular level, atmospheric
gases cause Rayleigh scattering that progressively affects shorter wavelengths (causing, for example, the sky to appear
blue). Further, major atmospheric components such as oxygen, carbon dioxide, ozone and water vapor (particularly these
latter two) cause absorption of energy at selected wavelengths. Aerosol particulates (an aerosol is a gaseous suspension of
fine solid or liquid particles) are the primary determinant of haze, and introduce a largely non-selective (i.e., affecting all
wavelength equally) Mie scattering. Atmospheric effects can be substantial (see Figure 3). Thus remote sensing specialists
have worked towards the modeling and correction of these effects. IDRISI offers several approaches to atmospheric cor-
rection, with the most sophisticated being the module ATMOSC.
Cos(t) Model
One of the difficulties with atmospheric correction is that the data necessary for a full accommodation are often not avail-
able. The Cos(t) model was developed by Chavez (1996) as a technique for approximation that works well in these
instances. It is also available in the ATMOSC module and incorporates all of the elements of the Dark Object Subtraction
model (for haze removal) plus a procedure for estimating the effects of absorption by atmospheric gases and Rayleigh
scattering. It requires no additional parameters over the Dark Object Subtraction model and estimates these additional
elements based on the cosine of the solar zenith angle (90 - solar elevation).
Figure 3 Figure 4
Topographic Effects
Topographic effect3 is defined simply as the difference in radiance values from inclined surfaces compared to horizontal ones.
The interaction of the angle and azimuth of the sun's rays with slopes and aspects produce topographic effects resulting in
variable illumination. Images are often taken in the early morning hours or late afternoon, when the effect of sun angle on
slope illumination can be extreme. In mountainous environments, reflectances of slopes facing away from the sun are
considerably lower than the overall reflectance or mean of an image area. In extreme terrain conditions, some areas may
be shadowed to the extent that meaningful information is lost altogether.
Shadowing and scattering effects exaggerate the difference in reflectance information coming from similar earth materi-
als. The signature of the same landcover type on opposite facing slopes may not only have a different mean and variance,
but may even have non-overlapping reflectance ranges. In the classification process, the highly variable relationship
between slope, landcover, and sun angle can lead to a highly exaggerated number of reflectance groups that make final
interpretation of data layers more costly, difficult, and time consuming. Even when landcovers are not the same on oppo-
site sides of the mountain (which is often the case since slope and aspect help determine the cover type present), variable
illumination nonetheless makes it difficult to derive biomass indices or perform other comparisons between landcover
classes.
Several techniques for mitigating topographic effect have evolved in recent years. However, many tend to be only appro-
priate for the specific environment in which they were developed, or they require high-detail ancillary data that is often
unavailable. The three most accessible techniques used are band ratioing, partitioning an image into separate areas for
classification, and illumination modeling based on a DEM. More sophisticated techniques (which are not discussed here)
involve the modeling of such illumination effects as backscattering and indirect diffusion effects.
Band Ratioing
In band ratioing, one band image is divided by another.
BandA
------------------
BandB
The resulting output image is then linearly stretched back to the 0 to 255 value range, and used for image classification.
Band ratioing is based on the principle that terrain variations (in slope and aspect) cause variations in illumination that are
3. The Topographic Effects section is condensed from the "Mitigating Topographic Effects in Satellite Imagery" exercise in Schneider and Robbins,
1995. UNITAR Explorations in GIS Technology, Volume 5, GIS and Mountain Environments, UNITAR, Geneva, also available from the Clark Labs..
Image Partitioning
Image partitioning works from the simple assumption that since different areas within an image are affected differently by
illumination effects resulting from slope and aspect, these distinct areas should be classified separately. Using a digital ele-
vation model to produce mask images, the bands are subdivided according to different elevation, slope, and aspect catego-
ries. These sub-scenes are then classified separately and the results recombined after classification. Like band ratioing, the
technique is simple and intuitive, but is effective only under the right conditions. Such partitioning works best where land-
cover conditions are stratified environmentally. Otherwise there is the potential to create hundreds of meaningless clusters
when using an unsupervised classification or to misclassify pixels when applying supervised techniques.
The thresholds set for slope, aspect, and elevation are dependent upon the known sun angle and azimuth. Without solar
information, the significant thresholds of topographic effect may be imprecisely determined by analyzing the shadow
effect visually. The registration of the DEM to the satellite data must be as precise as possible, otherwise the inexact
nature of thresholds will further increase the possibility of less meaningful classifications.
Illumination Modeling
The analytical tools associated with most raster GIS software systems offer a very effective technique for modeling illumi-
nation effects. The steps, using IDRISI, are as follows:
1. Use HISTO to calculate the mean of the image band to be corrected.
2. Using a Digital Elevation Model (DEM) for the image area, use the HILLSHADE (SURFACE) module to create a map
of analytical hillshading. This will be the model of illumination effects for all bands and simply needs to be calibrated for
use.
3. Use REGRESS to calculate the linear relationship between the hillshading map and the image to be corrected. Use the
image as the dependent variable and the hillshading as the independent variable.
4. Use CALIBRATE to apply the offset (the intercept of the regression equation) and gain (the slope of the regression
equation) to the hillshading map. The result is a model of the terrain-induced illumination component.
5. Use Image Calculator to subtract the result of the previous step from the original image and then add the mean calcu-
lated in the first step. The result is a reasonable estimate of what the image would have looked like if the ground had been
flat.
Noise
Noise in images occurs because of any number of mechanical or electronic interferences in the sensing system that lead to
transmission errors. Noise either degrades the recorded signal or it virtually eliminates all radiometric information. Noise
can be systematic, such as the periodic malfunctioning of a detector, resulting in striping or banding in the imagery. Or it
can be more random in character, causing radiometric variations described as having a "salt-and-pepper" appearance. In
RADAR imagery, "speckle" occurs because the signal's interaction with certain object edges or buildings produces a
highly elevated recording, which when frequent, has a similar effect as "salt-and-pepper" noise.
"Salt-and-Pepper" Noise
Random noise often produces values that are abnormally high or low relative to surrounding pixel values. Given the
assumption that noisy reflectance values show abrupt changes in reflectance from pixel to pixel, it is possible to use filter-
ing operations to replace these values with another value generated from the interpretation of surrounding pixels. FIL-
TER in IDRISI provides several options for this purpose. A 3 x 3 or 5 x 5 median filter commonly are applied. The noisy
pixels are replaced by the median value selected from the neighbors of the specified window. Because all pixels are pro-
cessed by median filtering, some detail and edges may be lost. This is especially problematic with RADAR imagery
because of the particularly high level of speckle that can occur. Therefore, we have included an Adaptive Box filter that is
an extension of the common Lee filter. The Adaptive Box filter determines locally within a specified window (3 x 3, 5 x 5,
or 7 x 7) the mean and the min/max value range based on a user-specified standard deviation. If the center window value
is outside the user-specified range, then it is assumed to be noise and the value is replaced by an average of the surround-
ing neighbors. You may choose the option of replacing the value with a zero. The filter also allows the user to specify a
minimum threshold variance in order to protect pixels in areas of very low variation. See the on-line Help System of the
FILTER module to learn more about this highly flexible approach.
Geometric Restoration
As stated in the chapter Introduction to Remote Sensing and Image Processing, most elements of geometric restora-
tion associated with image capture are corrected by the distributors of the imagery, most importantly skew correction and
scanner distortion correction. Distributors also sell imagery already georeferenced. Georeferencing is not only a restoration tech-
nique but a method of reorienting the data to satisfy the specific desires and project requirements of the data user. As
such, it is particularly important that georeferenced imagery meets the data user's standards and registers well with other
data in the same projection and referencing system.
It is our experience that even if one's standards are not very stringent for the particular imaging task at hand, it is well
worth the time one takes to georeference the imagery oneself rather than having the distributor do so. This is true for a
number of reasons. First, certain radiometric corrections become more difficult (if not impossible) to perform if the data
are already georeferenced. Of particular concern is the ability to reduce the effects of banding, scan line drop, and topo-
graphic effects on illumination. If the geometric orientation of the effects is altered, then standard restoration techniques
are rendered useless. Given that the severity of these effects is not usually known prior to receiving the data, georeferenc-
ing the data oneself is important in maintaining control over the image restoration process.
Another reason to georeference imagery oneself is to gain more control over the spatial uncertainties produced by the
georeferencing process. Only then is it possible to know how many control points are used, where they are located, what
the quality of each is individually, and what the most satisfying combination of control points is to select. The RESAM-
PLE module in IDRISI provides the user with significant control over this process. The user may freely drop and add
References
Chavez, P.S., (1996) "Image-Based Atmospheric Corrections - Revisited and Improved", Photogrammetric Engineering
and Remote Sensing, 62, 9, 1025-1036.
Forster, B.C., (1984) "Derivation of atmospheric correction procedures for Landsat MSS with particular reference to
urban data", International Journal of Remote Sensing, 5, 5, 799-817.
Lillesand, T.M. and R.W. Kiefer, (1994) Remote Sensing and Image Interpretation. Third Edition. John Wiley and Sons.
Turner, R.E., and Spencer, M.M., (1972) "Atmospheric Model for Correction of Spacecraft Data", Proceedings, Eighth
International Symposium on Remote Sensing of the Environment, Vol. II, 895-934.
Figure 1 Figure 2
Fourier Analysis uses this logic in reverse. It starts with a complex wave form and assumes that this is the result of the
additive effects of simple sine waves of varying frequency, amplitude and phase. Frequency refers to the number of complete
wavelengths that occur over a specified distance. Figure 3 shows a series of sine waves that differ only in their frequency.
Amplitude refers to the height or strength of the wave. Figure 4 shows a series of sine waves that vary only in amplitude.
Finally, phase refers to the offset of the first wave crest from origin. Figure 5 shows a series of waves that vary only in
phase. In the decomposition of digital images, a finite set of waves are assumed, ranging from the lowest frequency wave
which completes a single wavelength over the extent of the image in X and Y, to one that completes 2 wavelengths over
that extent, to 3 wavelengths, and so on, up to the highest frequency wave with a wavelength equal to twice the pixel reso-
lution (known as the Nyquist frequency). The task of the Fourier Analysis process is then simply one of estimating the phase
0.5
0.5
0
0
-0.5
-0.5
-1
-1
0 1 2 3 4 5 6 7 8 9 10 11 12 13
0 1 2 3 4 5 6 7 8 9 10 11 12 13
Figure 3 Figure 4
1
0.5
-0.5
-1
0 1 2 3 4 5 6 7 8 9 10 11 12 13
Figure 5
1. The use of a sine/cosine pair is identical in concept to describing locations using a pair of coordinates—X and Y. In both cases, the reference pair are
known as basis vectors. In plane two-dimensional space, any location can be defined by its X and Y coordinates. Similarly, in the frequency domain, any
wave can be defined by its sine and cosine coordinates.
where f(x) is the value of the function at position x, ω is equal to 2π/T where T is the period (the length of the series), and
k is the harmonic.2 The coefficients a and b are determined by means of the Fourier Transform.
The Fourier Transform itself makes use of Euler's Formula:
– i2πux
e = cos 2πux – i sin 2πux
where i2 is -1, leading to the following formulation for the case of discrete data:
1 – ikωx
F ( u ) = ---- Σf ( x )e
N
and an inverse formula of:
ikωx
f ( x ) = ΣF ( u )e
It is not critical that these mathematical expressions be fully understood in order to make productive use of the Fourier
Transform. However, it can be appreciated from the above that:
1. these formulas express the forward and inverse transforms for one-dimensional data. Simple extensions make these
applicable to two-dimensional data.
2. the implementation of the Fourier Transform uses complex-valued inputs and outputs. A complex number is one with
both real and imaginary parts of the form a+bi, where i2 = -1. In the case considered here, for input into the FOURIER
module, the image grey-level values make up the real part, while the imaginary part is set to zero (this is done automati-
cally by IDRISI) since it doesn't exist. Thus the input to the Fourier Transform is a single image.
3. the resulting output of the forward transform thus consists of two images—a real and an imaginary part. The real part
expresses the cosine component of the Fourier series while the imaginary part expresses the sine component. These are
the amplitudes a and b of the cosine and sine components expressed in the first formula in this section. From these, the
amplitude and phase of the wave can be determined as follows:
2 2
Amplitude = a +b
Phase = tan–1 ( b ⁄ a )
Together, the real and imaginary parts express the frequency spectrum of an image. While both the amplitude and phase
can be readily calculated from these parts (Image Calculator can be used in IDRISI to do this), neither is commonly used.
More commonly, the power spectrum is calculated and the phase is ignored. The power spectrum is simply the square of
the amplitude. However, for purposes of visual examination, it is commonly expressed as a power spectrum image as fol-
lows:
2
PowerSpectrum = ln ( 1 + amplitude )
This is the formula used in IDRISI. Thus the forward transform produced by the FOURIER module yields three out-
puts—a real part image, an imaginary part image, and a power image. These are commonly referred to as frequency
2. The term "harmonic" used here refers to the relation of quantities whose reciprocals are in arithmetic progression (e.g., 1, 1/2, 1/3, 1/4, etc.); or to
points, lines, functions, etc. involving such a relation.
3. The phase information is not important for visual analysis, and is lost in the production of the power spectrum image. However, all mathematical
manipulations are undertaken on the real and imaginary part images, which together contain complete amplitude and phase information.
IDRISI supplies a variety of facilities for developing frequency domain filters. One would most commonly use FIL-
TERFQ, which offers 26 filters, each of which can be controlled for the specific characteristics of the data being manipu-
lated as well as for the specific purpose of the filter. The next most commonly used module is DRAWFILT, an interactive
filter development tool in which one literally draws the areas (i.e., frequencies) to be removed. Finally, for even further
flexibility, IDRISI offers a module named FREQDIST. As the name suggests, this module creates a frequency distance
image (as measured from the center point of the power spectrum). This can then be used as the basic input to a variety of
filter shaping tools such as RECLASS or FUZZY. The frequency distance image can also be submitted to SURFACE to
create an aspect image which can then be shaped by RECLASS or FUZZY to create directional filters.
Regardless of how the filter is created, however, application of that filter is achieved the same way in all cases—by simple
multiplication using OVERLAY with both the real and imaginary part images. As it turns out, multiplication in the fre-
quency domain is the equivalent of convolution in the spatial domain.
Introduction
Classification is the process of developing interpreted maps from remotely sensed images. As a consequence, classifica-
tion is perhaps the most important aspect of image processing to GIS. Traditionally, classification was achieved by visual
interpretation of features and the manual delineation of their boundaries. However, with the advent of computers and dig-
ital imagery, attention has focused on the use of computer-assisted interpretation. Although the human eye still brings a
superior set of capabilities to the classification process, the speed and consistency of digital procedures make them very
attractive. As a consequence, the majority of classification projects today make use of digital classification procedures,
guided by human interpretation.
Supervised Classification
General Logic
There is a consistent logic to all of the supervised classification routines in IDRISI, regardless of whether they are hard or
soft classifiers. In addition, there is a basic sequence of operations that must be followed no matter which of the super-
vised classifiers is used. This sequence is described here. The Tutorial manual also contains worked examples of this pro-
cess.
1. IDRISI offers two procedures for the digitizing of training site polygons. With the default procedure, one digitizes a set of points that form the
boundary of the training site polygon. The second procedure creates the polygon by aggregating together all contiguous pixels surrounding a designated
point that fall within a specified tolerance of the spectral characteristics of the central pixel. This is called a flood polygon since it is analogous to the
concept of water flowing outward from the designated point. You will also note that with this option a maximum distance can be specified to limit how
far this growth procedure will spread. Note that the system also allows you to define training sites by means of a raster image. In some instances, it may
make sense to define these locations by direct reference to ground locations (such as by means of point locations gathered with a GPS). However, this
requires very exact prior georeferencing of the image, and a confidence that positional errors will not include unwanted pixels in the training sites. Some
georeferencing procedures also alter image characteristics which may be undesirable. It is for these reasons that classification is commonly undertaken
on ungeoreferenced imagery using on-screen digitizing of training sites. The final classified image is then georeferenced at a later stage.
2. Extract Signatures
After the training site areas have been digitized, the next step will be to create statistical characterizations of each informa-
tional class. These are called signatures in IDRISI. This is usually achieved with the MAKESIG module.2 MAKESIG will
ask for the name of the vector or raster file that contains the training sites for one or more informational classes, and the
bands to be used in the development of signatures. It will then ask for a name for each of the included classes. These
names should be suitable as IDRISI file names since they will be used to create signatures files (.sig extension) for each
informational class. If you used more than one vector file to store your training site polygons, run MAKESIG for each of
these files. Your goal is to create a SIG file for every informational class.
SIG files contain a variety of information about the landcover classes they describe.3 These include the names of the
image bands from which the statistical characterization was taken, the minimum, maximum and mean values on each
band, and the full variance/covariance matrix associated with that multispectral image band set for that class. To examine
the contents of this file in detail, use SIGCOMP. Note also that the SEPSIG module can be used to assess the distinctive-
ness of your set of signatures.
2. The FUZSIG module offers an interesting, but quite different logic for extracting signatures from impure training sites. This is discussed further in
the section on fuzzy signatures below. The ENDSIG module also creates signatures for use with the UNMIX classifier.
3. Each SIG file also has a corresponding SPF file that contains the actual pixel values used to create the SIG file. It is used only by HISTO in displaying
histograms of signatures.
4. This second possibility does not exist if one uses the BAYCLASS module or FUZCLASS with normalized output. Both of these soft classification
procedures assume that the classes considered are the only ones possible. It is for this reason that the BELCLASS procedure is recommended for com-
parison to MAXLIKE and the FUZCLASS with un-normalized output is recommended for comparison to MINDIST.
5. Generalization
The fifth stage is optional and frequently omitted. After classification, there may be many cases of isolated pixels that
belong to a class that differs from the majority that surround them. This may be an accurate description of reality, but for
mapping purposes, a very common post-processing operation is to generalize the image and remove these isolated pixels.
This is done by passing a mode filter over the result (using the FILTER module in IDRISI). The mode filter replaces each
pixel with the most frequently occurring class within a 3x3 window around each pixel. This effectively removes class
patches of one or a few pixels and replaces them with the most common neighboring class. Use this operation with
care—it is a generalization that truly alters the classified result.
6. Accuracy Assessment
The final stage of the classification process usually involves an accuracy assessment. Traditionally this is done by generat-
ing a random set of locations (using the stratified random option of the SAMPLE module in IDRISI) to visit on the
ground for verification of the true landcover type. A simple values file is then made to record the true landcover class (by
its integer index number) for each of these locations. This values file is then used with the vector file of point locations to
create a raster image of the true classes found at the locations examined. This raster image is then compared to the classi-
fied map using ERRMAT. ERRMAT tabulates the relationship between true landcover classes and the classes as mapped.
It also tabulates errors of omission and errors of commission as well as the overall proportional error.
The size of the sample (n) to be used in accuracy assessment can be estimated using the following formula:
n = z2 pq / e2
where
z is the standard score required for the desired level of confidence (e.g., 1.96 for 95% confidence, 2.58 for 99%,
etc.) in the assessment
e is the desired confidence interval (e.g., 0.01 for ±10%)
p is the a priori estimated proportional error, and
q=1-p
Hard Classifiers
The hard classifiers are so named because they all reach a hard (i.e., unequivocal) decision about the class to which each
pixel belongs. They are all based on a logic that describes the expected position of a class (based on training site data) in
what is known as band space, and then gauging the position of each pixel to be classified in the same band space relative to
these class positions. From this perspective, the easiest classifier to understand is the MINDIST procedure.
Minimum-Distance-to-Means
The MINDIST module implements a Minimum-Distance-to-Means classifier. Based on training site data, MINDIST charac-
terizes each class by its mean position on each band. For example, if only two bands were to be used, Figure 1 might char-
Sand
Deciduous
Band 2
Conifer = Class Mean
0 255
Figure 1 Band 1
Here each axis indicates reflectance on one of the bands. Thus, using the mean reflectance on these bands as X,Y coordi-
nates, the position of the mean can be placed in this band space. Similarly, the position of any unclassified pixel can also
be placed in this space by using its reflectance on the two bands as its coordinates.
To classify an unknown pixel, MINDIST then examines the distance from that pixel to each class and assigns it the iden-
tity of the nearest class. For example, the unclassified pixel shown in Figure 2 would be assigned the "sand" class since this
is the class mean to which it is closest.
255
Sand
Deciduous
Band 2
Unclassified pixel
Conifer
0 255
Figure 2 Band 1
Despite the simplicity of this approach, it actually performs quite well. It is reasonably fast and can employ a maximum
distance threshold which allows for any pixels that are unlike any of the given classes to be left unclassified. However, the
approach does suffer from problems related to signature variability. By characterizing each class by its mean band reflec-
tances only, it has no knowledge of the fact that some classes are inherently more variable than others. This, in turn, can
lead to misclassification. For example, consider the case of a highly variable deciduous class and a very consistent sand
class in classifying the unclassified pixel in Figure 3.
Sand
Deciduous
Band 2
Unclassified pixel
Conifer
0 255
Figure 3 Band 1
The circles in this figure illustrate the variability of each of these two classes. If we assume that these circles represent a
distance of two standard deviations from the mean, we can see that the pixel lies within the variability range of the decid-
uous category, and outside that of sand. However, we can also see that it is closer to the mean for sand. In this case, the
classifier would misclassify the pixel as sand when it should really be considered to be deciduous forest.
This problem of variability can be overcome if the concept of distance is changed to that of standard scores. This trans-
formation can be accomplished with the following equation:
standardized distance = ( original distance - mean ) / standard deviation
The MINDIST procedure in IDRISI offers this option of using standardized distances, which is highly recommended. In
the example above, the pixel would be correctly classified as deciduous since its standardized distance from the mean for
deciduous would be less than 2 (perhaps 1.95 in this illustration), while that for sand would be greater than 2 (probably
close to 4 in this illustration).
Our experience with MINDIST has been that it can perform very well when standardized distances are used. Indeed, it
often outperforms a maximum likelihood procedure whenever training sites have high variability.
Parallelepiped
The PIPED module implements the parallelepiped procedure for image classification. The parallelepiped procedure char-
acterizes each class by the range of expected values on each band. This range may be defined by the minimum and maxi-
mum values found in the training site data for that class, or (more typically) by some standardized range of deviations
from the mean (e.g., ± 2 standard deviations). With multispectral image data, these ranges form an enclosed box-like poly-
gon of expected values known as a parallelepiped. Unclassified pixels are then given the class of any parallelepiped box they
fall within. If a pixel does not fall within any box, it is left unassigned. Figure 4 illustrates this effect.
255
Sand
Deciduous
Band 2
Conifer
0 255
Figure 4 Band 1
255
Overlap Zone
Band 2
Non-Representative Zone
0 255
Figure 5 Band 1
Clearly the MINDIST procedure would not encounter this problem, since the line of separation between these classes
would fall in between these two distributions. However, in this context of correlation between bands (which is virtually
guaranteed), the parallelepiped procedure produces both zones of overlap and highly non-representative areas that really
should not be included in the class. In general, then, the parallelepiped procedure should be avoided, despite the fact that
it is the fastest of the supervised classifiers.5
Maximum Likelihood
To compensate for the main deficiencies of both the Parallelepiped and Minimim-Distance-to-Means procedures, the
Maximum Likelihood procedure, provided by the MAXLIKE module in IDRISI, is used. The Maximum Likelihood pro-
cedure is based on Bayesian probability theory. Using the information from a set of training sites, MAXLIKE uses the
mean and variance/covariance data of the signatures to estimate the posterior probability that a pixel belongs to each
class.
In many ways, the MAXLIKE procedure is similar to MINDIST with the standardized distance option. The difference is
that MAXLIKE accounts for intercorrelation between bands. By incorporating information about the covariance
between bands as well as their inherent variance, MAXLIKE produces what can be conceptualized as an elliptical zone of
characterization of the signature. In actuality, it calculates the posterior probability of belonging to each class, where the
probability is highest at the mean position of the class, and falls off in an elliptical pattern away from the mean, as shown
in Figure 6.
5. In the early days of image processing when computing resources were poor, this classifier was commonly used as a quick look classifier because of its
speed.
Deciduous Sand
Band 2
Conifer
0 255
Figure 6 Band 1
Soft Classifiers
Unlike hard classifiers, soft classifiers defer making a definitive judgment about the class membership of any pixel in favor
of a group of statements about the degree of membership of that pixel in each of the possible classes. Like traditional
supervised classification procedures, each uses training site information for the purpose of classifying each image pixel.
However, unlike traditional hard classifiers, the output is not a single classified landcover map, but rather a set of images
(one per class) that express for each pixel the degree of membership in the class in question. In fact, each expresses the
degree to which each pixel belongs to the set identified by a signature according to one of the following set membership
metrics:
BAYCLASS based on Bayesian probability theory,
BELCLASS based on Dempster-Shafer theory,
MAHALCLASS based on Mahalanobis distance,
FUZCLASS based on Fuzzy Set theory, and
Classification Uncertainty
In addition to these set membership images, each of these soft classifiers outputs an image that expresses the degree of
classification uncertainty it has about the class membership of any pixel. Classification uncertainty measures the degree to
which no class clearly stands out above the others in the assessment of class membership of a pixel. In the case of BAY-
CLASS, BELCLASS and FUZCLASS, it is calculated as follows:
sum
max – ----------
n
ClassificationUncertainty = 1 – ----------------------------
1---
1–
n
where
max = the maximum set membership value for that pixel
sum = the sum of the set membership values for that pixel
n = the number of classes (signatures) considered
The logic of this measure is as follows:
- The numerator of the second term expresses the difference between the maximum set membership value and the total
dispersion of the set membership values over all classes.
- The denominator of the second term expresses the extreme case of the difference between a maximum set membership
value of 1 (and thus total commitment to a single class) and the total dispersion of that commitment over all classes.
- By taking the ratio of these two quantities, one develops a measure that expresses the degree of commitment to a spe-
cific class relative to the largest possible commitment that can be made. Classification uncertainty is thus the complement
of this ratio.
In spirit, the measure of uncertainty developed here is similar to the entropy measure used in Information Theory. How-
ever, it differs in that it is concerned not only with the degree of dispersion of set membership values between classes, but
also the total amount of commitment present. Following are some examples that can clarify this concept.
Examples
Assuming a case where three classes are being evaluated, consider those with the following allocations of set membership:
(0.0 0.0 0.0) Classification Uncertainty = 1.00
(0.0 0.0 0.1) Classification Uncertainty = 0.90
(0.1 0.1 0.1) Classification Uncertainty = 1.00
where :
p(h|e) = the probability of the hypothesis being true given the evidence (posterior probability)
p(e|h) = the probability of finding that evidence given the hypothesis being true
p(h) = the probability of the hypothesis being true regardless of the evidence (prior probability)
In this context, the variance/covariance matrix derived from training site data is that which allows one to assess the multi-
variate conditional probability p(e|h). This quantity is then modified by the prior probability of the hypothesis being true
and then normalized by the sum of such considerations over all classes. This latter step is important in that it makes the
assumption that the classes considered are the only classes that are possible as interpretations for the pixel under consid-
eration. Thus even weak support for a specific interpretation may appear to be strong if it is the strongest of the possible
choices given.
This posterior probability p(h|e) is the same quantity that MAXLIKE evaluates to determine the most likely class, and
indeed, if the output images of BAYCLASS were to be submitted directly to HARDEN, the result would be identical to
that of MAXLIKE. In essence, BAYCLASS is a confident classifier. It assumes that the only possible interpretation of a
pixel is one of those classes for which training site data have been provided. It therefore admits to no ignorance. As a
result, lack of evidence for an alternative hypothesis constitutes support for the hypotheses that remain. In this context, a
pixel for which reflectance data only very weakly support a particular class is treated as unequivocally belonging to that
class (p = 1.0) if no support exists for any other interpretation.
The prime motivation for the use of BAYCLASS is sub-pixel classification—i.e., to determine the extent to which mixed
pixels exist in the image and their relative proportions. It is also of interest to observe the underlying basis of the MAX-
LIKE procedure. However, for In-Process Classification Assessment (IPCA), the BELCLASS procedure is generally preferred
because of its explicit recognition that some degree of ignorance may surround the classification process.
In the context of mixture analysis, the probabilities of BAYCLASS are interpreted directly as statements of proportional
representation. Thus if a pixel has posterior probabilities of belonging to deciduous and conifer of 0.68 and 0.32 respec-
tively, this would be interpreted as evidence that the pixel contains 68% deciduous species and 32% conifers. Note, how-
ever, that this requires several important assumptions to be true. First, it requires that the classes for which training site
data have been provided are exhaustive (i.e., that there are no other possible interpretations for that pixel). Second, it
assumes that the conditional probability distributions p(e|h) do not overlap in the case of pure pixels. In practice, these
conditions may be difficult to meet.
In testing at Clark Labs, we have found that while BAYCLASS is effective in determining the constituent members of
6. The logic of the decomposition process is detailed in the module description for BELCLASS in the on-line Help System.
7. The filename for the classification uncertainty image in BELCLASS is composed by concatenating the prefix supplied by the user and the letter string
"clu".
8. Clearly these are sets of classes. However, since evidence may support one of these sets without further distinction about which members of the set
are supported, the set itself can be thought of as a class.
Belief
Both BELCLASS and BELCALC deconstruct the evidence to infer belief and plausibility for each class. One of the moti-
vations for doing so is that it allows the user to combine ancillary information with that determined from the reflectance
data. New evidence can be combined with existing knowledge with the Belief module. Belief is described more fully in the
chapter on Decision Support: Uncertainty Management, and is used in exactly the same manner for the data
described here.
f1(24) + f2(56) = 32
f1(86) + f2(98) = 89
where f1 and f2 represent the fractions (proportions) of the two end members. Such a system of simultaneous equations
can be solved using matrix algebra to yield a best fit estimate of f1 and f2 (0.75 and 0.25 in this example). In addition, the
sum of squared residuals between the fitted signature values and the actual values can be used as a measure of the uncer-
tainty in the fit.
The primary limitation of this approach is that the number of end members cannot exceed the number of bands. This can
be a severe limitation in the case of SPOT imagery, but of little consequence with hyperspectral imagery. IDRISI thus
offers three approaches (in UNMIX) to Linear Spectral Unmixing:
1. The standard linear spectral unmixing approach (as indicated above) for cases where sufficient bands are available.
2. A probability guided option for cases where insufficient bands exist. Although the total number of possible end mem-
bers may be large, the number that coexist within a single pixel is typically small (e.g., 2-3). This approach thus uses a first
stage based on the BAYCLASS module to determine the most likely constituents (up to the number of bands), with a sec-
ond stage linear spectral unmixing to determine their fractions. Experiments at Clark Labs have shown this to produce
excellent results.
3. An exhaustive search option for cases where insufficient bands exist. In this instance, one specifies the number of con-
stituents to consider (up to the total number of bands). It then tests all possible combinations of that many end members
and reports the fractions of that combination with the lowest sum of squared residuals. This approach is considerably
slower than the other options. In addition, experiments at Clark Labs have shown that this approach yields inferior results
to the probability guided procedure in cases where the end member signatures are drawn from training sites. Pure end
member signatures, created with ENDSIG, should be used.
End member signatures are specified using standard signature (.sig) files, created using either MAKESIG or ENDSIG.
The former is used in the case where end members are derived from training sites, while the latter is used in cases where
pure end member values are known (such as from a spectral library). Note that only the mean value on each band is used
from these signature files—variance/covariance data are ignored.
Hardeners
Once a soft classifier has been applied to a multispectral image set, the soft results can be re-evaluated to produce a hard
classification by using one of the following hardeners from the module HARDEN:
BAYCLASS
Using the results from BAYCLASS, this option determines the class possessing the maximum posterior probability for
each cell, given a set of probability images. Up to four levels of abstraction can be produced. The first is the most likely
class, just described. The second outputs the class of the second highest posterior probability, and so on, up to the fourth
highest probability.
BELCLASS
Using the results from BELCLASS, this option is essentially identical to the BAYCLASS hardener, except that it is
designed for use with Dempster-Shafer beliefs.
FUZCLASS
Using the results from FUZCLASS, this option is essentially identical to the BAYCLASS hardener, except that it is
designed for use with Fuzzy Sets.
MAHALCLASS
Using the results from MAHALCLASS, this option is essentially identical to the BAYCLASS hardener, except that it is
designed for use with the typicalities produced by MAHALCLASS.
Unsupervised Classification
General Logic
Unsupervised classification techniques share a common intent to uncover the major landcover classes that exist in the
image without prior knowledge of what they might be. Generically, such procedures fall into the realm of cluster analysis,
since they search for clusters of pixels with similar reflectance characteristics in a multi-band image. They are also all gen-
eralizations of landcover occurrence since they are concerned with uncovering the major landcover classes, and thus tend
to ignore those that have very low frequencies of occurrence. However, given these broad commonalities, there is little
else that they share in common. There are almost as many approaches to clustering as there are image processing systems
on the market. IDRISI is no exception. The primary unsupervised procedure IDRISI offers is unique (CLUSTER). How-
ever, IDRISI also offers a commonly used procedure (ISODATA) a variant on this (ISOCLUST). As implemented here,
ISOCLUST is really an iterative combination of unsupervised and supervised procedures, as is also the case with the third
procedure offered, MAXSET. IDRISI also has a true K-means clustering classifier. IDRISI offers additional supervised
modules that also have unsupervised capabilities: SOM and Fuzzy ARTMAP, act in this fashion.
CLUSTER
The CLUSTER module in IDRISI implements a special variant of a Histogram Peak cluster analysis technique (Richards,
1993). The procedure can best be understood from the perspective of a single band. If one had a single band of data, a
histogram of the reflectance values on that band would show a number of peaks and valleys. The peaks represent clusters
of more frequent values associated with commonly occurring cover types.
The CLUSTER procedure thus searches for peaks by looking for cases where the frequency is higher than that of its
immediate neighbors on either side. In the case of two bands, these peaks would be hills, while for three bands they would
be spheres, and so on. The concept can thus be extended to any number of bands. Once the peaks have been located,
each pixel in the image can then be assigned to its closest peak, with each such class being labeled as a cluster. It is the ana-
lyst's task to then identify the landcover class of each cluster by looking at the cluster image and comparing it to ground
features.
CLUSTER offers two levels of generalization. With the broad level of generalization, clusters must occur as distinct peaks
in the multi-dimensional histogram as outlined above. However, with the fine level of generalization, CLUSTER also rec-
ognizes shoulders in the curve as cluster peaks. Shoulders occur when two adjacent clusters overlap to a significant extent.
Peaks and shoulders are identified in the histogram shown in Figure 7.
Figure 7
The CLUSTER procedure in IDRISI has been modified and tailored to work with the special case of three bands as
described by an 8-bit color composite image created with the COMPOSITE module. The reason for doing so is based
largely on the fact that the procedure involved in creating an 8-bit color composite image is essentially the same as the
first stage of multi-dimensional histogram generation in the clustering algorithm. Since it is not uncommon to experiment
with various clusterings of a single multi-band image, speed is greatly enhanced by not repeating this histogram generation
step. While it may seem that the restriction of working with a three-band composite is limiting, bear in mind that the
underlying "bandness" of a multi-spectral image in the visible-through-middle infrared is rarely more than 2 or 3 (to con-
firm this, try running a Principal Components Analysis on a higher spectral resolution multi-band image set). In most
environments, creating composite images using the red and near infrared bands, along with a middle-infrared band (such
as Landsat Band 5) will essentially capture all of the information in the image.
Experience in using the CLUSTER routine has shown that it is fast and is capable of producing excellent results. How-
ever, we have learned that the following sequence of operations is particularly useful.
1. Run CLUSTER using the most informative bands available (generally, these include the red visible band, the near infra-
red band, and a middle infrared band—e.g., Landsat TM bands 3, 4 and 5 respectively). Use the linear stretch with satura-
tion option with 1% saturation the "fine generalization level" and " retain all clusters" options.
2. Display a histogram of this image. This histogram shows the frequency of pixels associated with each of the clusters
that can be located in the image. Many of these clusters have very small frequencies, and thus are somewhat insignificant.
Figure 8 presents an example of just such a histogram.
As can be seen, there are three clusters that dominate the image. Then there is a sharp break with a second group of
strong clusters through to Cluster 12. Then there is a third group that follows until Cluster 25, followed by a small group
of very insignificant clusters. Experience suggests then that a good generalization of the data would be to extract the first
12 clusters, with a more detailed analysis focusing on the first 25.
3. Once the number of clusters to be examined has been determined, run CLUSTER again, but this time choose the "fine
generalization" and set the "maximum number of clusters" to the number you determined (e.g., 12).
4. Display the resulting image with a qualitative color palette and then try to identify each of the clusters in turn. You can
use the interactive legend editing option to change the legend caption to record your interpretation. You may also wish to
change the color of that category to match a logical color scheme. Also, remember that you may highlight all pixels
belonging to a category by holding down the mouse button over a legend category color box.
5. At the end of the identification process, you may need to combine several categories. For example, the cluster analysis
may have uncovered several pavement categories, such as asphalt and concrete, that you may wish to merge into a single
category. The simplest way of doing this is to use Edit to create a values file containing the integer reassignments. This file
has two columns. The left column should record the original cluster number, while the right column should contain the
new category number to which it should be reassigned. After this file has been created, run ASSIGN to assign these new
category indices to the original cluster data.
The CLUSTER procedure in IDRISI is fast and remarkably effective in uncovering the basic landcover structure of the
image. It can also be used as a preliminary stage to a hybrid unsupervised/supervised process whereby the clusters are
used as the training sites to a second classification stage using the MAXLIKE classifier.9 This has the advantage of allow-
ing the use of a larger number of raw data bands, as well as providing a stronger classification stage of pixels to their most
similar cluster. In fact, it is this basic logic that underlies the ISOCLUST procedure described below.
ISOCLUST
The ISOCLUST module is an iterative self-organizing unsupervised classifier based on a concept similar to the well-
known ISODATA routine of Ball and Hall (1965) and cluster routines such as the H-means and K-means procedures.
The typical logic is as follows:
9. This is possible because MAKESIG can create signatures based on training sites defined by either a vector file or an image. In this case, the image
option is used.
ISODATA
The ISODATA module is a commonly used unsupervised technique that employs the so-called iterative self-organizing
data analysis algorithm to partition n-dimensional imagery into a number of clusters according to a specified value. In
general ISODATA begins by initializing a number of centroids using given parameters, then assigns each pixel to the clus-
ter whose centroid is the nearest. It then updates the cluster centroids, and splits and merges clusters whenever splitting
and merging criteria apply. ISODATA uses a Euclidean distance for calculating the distances between pixels and cluster
centroids. The performance of ISODATA depends on the initial estimation of the partition and the parameters specified.
See the classification chapter in: Richards, J.A., and X. Jia, 1999. Remote Sensing Digital Image Analysis (New York:
Springer), for more detail.
MAXSET
As previously described, MAXSET is a hard classifier that assigns to each pixel the class with the greatest degree of com-
mitment from the full Dempster-Shafer class hierarchy that describes all classes and their hierarchical combination.
Although it is run as if it were a supervised classifier (it requires training site data), ultimately it behaves as if it were an
unsupervised classifier in that it can assign a pixel to a class for which no exclusive training data have been supplied.
MAXSET is very similar in concept to the PIPED, MINDIST, and MAXLIKE classifiers in that it makes a hard determi-
nation of the most likely class to which each pixel belongs according to its own internal logic of operation. MAXSET is
different, however, in that it recognizes that the best assignment of a pixel might be to a class that is mixed rather than
unique. For example, it might determine that a pixel more likely belongs to a class of mixed conifers and deciduous forest
than it does to either conifers or deciduous exclusively. The logic that it uses in doing so is derived from Dempster-Shafer
theory, a special variant of Bayesian probability theory that is described more fully in the section on the BELCLASS soft
classifier above. Dempster-Shafer theory provides a logic for expressing one's belief in the degree to which an item
Segmentation Classification
IDRISI provides three modules for classification from image segments. Together they provide a hybrid methodology
between pixel-based and segment-based classification. The module SEGMENTATION creates an image of segments.
The module SEGTRAIN interactively develops training sites and signatures based on the segments from SEGMENTA-
TION. And the module SEGCLASS is a majority rule classifier based on the majority class within a segment. The major-
ity class within a segment is derived from a previously classified image, typically from a pixel-based classifier such as
MAXLIKE or MLP. SEGCLASS can improve the accuracy of the pixel-based classification and produce a smoother
map-like classification result while preserving the boundaries between segments.
Segmentation is a process by which pixels are grouped that share a homogeneous spectral similarity. The module SEG-
MENTATION groups adjacent pixels into image segments according to their spectral similarity. Specifically, SEGMEN-
TATION employs a watershed delineation approach to partition input imagery based on their variance. A derived
variance image is treated as a surface image allocating pixels to particular segments based on variance similarity. Across
space and over all input bands, a moving window assesses this similarity and segments are defined according to a stated
similarity threshold. The smaller the threshold, the more homogeneous the segments. A larger threshold will cause a more
heterogeneous and generalized segmentation result.
See the Help for SEGMENTATION for complete details on the algorithm.
10. The contents of these files are also known as image spectra.
11. The contents of these files are also known as library spectra.
PROFILE
The PROFILE module offers an additional tool for exploring hyperspectral data. A profile generated over a hyperspectral
series will graphically (or numerically) show how the reflectance at a location changes from one band to the next across
the whole series. Thus the result is a spectral response pattern for the particular locations identified.
PROFILE requires a raster image of the sample spots to be profiled. Up to 15 profiles can be generated simultaneously,12
corresponding to sample sites with index values 1-15. A sample site can consist of one or many pixels located in either a
contiguous grouping or in several disconnected groups.
The second piece of information that PROFILE will require is the name of a file that contains the names of the IDRISI
image files that comprise the hyperspectral series. You have two choices here. You may use either an image group file
(.rgf) or a sensor band file (.sbf), since both contain this information. In either case, you will need to have created one of
these before running PROFILE. You may wish to choose the sensor band file option, since it can be used with other
operations as well (such as HYPERSIG).
12. The display can become quite difficult to read whenever more than just a few profiles are generated at the same time. Under normal use, you may
wish to limit the profiling to no more than 5 sites.
255
Band 2
0
Figure 9 Band 1 255
Thus, in the presence of significant illumination variations, it would be anticipated that a traditional distance-based classi-
fier would have some difficulty in identifying the feature in all cases. The Spectral Angle Mapper thus uses a minimum-
angle approach. In essence, it treats each signature as a vector. Then by comparing the angle formed by an unknown pixel,
the origin, and a class mean, and comparing that to all other classes, the class that will be assigned to the unknown pixel is
that with the minimum angle, as illustrated in Figure 10.
255
Class 2 Signature
Band 2
Unknown Pixel
β α Class 1 Signature
0
Figure 10 Band 1 255
In the above figure, the unknown pixel would be assigned to Class 1 since the angle it subtends with the unknown pixel
(α) is smaller than that with Class 2 (β).
HYPERMIN
HYPERMIN is a minimum-distance classifier for hyperspectral data that is specifically intended for use with image-based
signatures developed from training sites. It uses a logic that is identical to that of the multispectral hard classifier MIND-
IST using standardized distances.
Introduction
Analysis of vegetation and detection of changes in vegetation patterns are keys to natural resource assessment and moni-
toring. Thus it comes as no surprise that the detection and quantitative assessment of green vegetation is one of the major
applications of remote sensing for environmental resource management and decision making.
Healthy canopies of green vegetation have a very distinctive interaction with energy in the visible and near infrared
regions of the electromagnetic spectrum. In the visible regions, plant pigments (most notably chlorophyll) cause strong
absorption of energy, primarily for the purpose of photosynthesis. This absorption peaks in the red and blue areas of the
visible spectrum, thus leading to the characteristic green appearance of most leaves. In the near infrared, however, a very
different interaction occurs. Energy in this region is not used in photosynthesis, and it is strongly scattered by the internal
structure of most leaves, leading to a very high apparent reflectance in the near infrared. It is this strong contrast, then,
most particularly between the amount of reflected energy in the red and near infrared regions of the electromagnetic spec-
trum, that has been the focus of a large variety of attempts to develop quantitative indices of vegetation condition using
remotely sensed imagery.
The aim of this chapter is to present a set of vegetation index (VI) models designed to provide a quantitative assessment
of green vegetation biomass. The proposed VIs are applicable to both low and high spatial resolution satellite images,
such as NOAA AVHRR, Landsat TM and MSS, SPOT HRV/XS, and any others similar to these that sense in the red and
near infrared regions. They have been used in a variety of contexts to assess green biomass and have also been used as a
proxy to overall environmental change, especially in the context of drought (Kogan, 1990; Tripathy et al., 1996; Liu and
Kogan, 1996) and land degradation risk assessment. As a consequence, special interest has been focused on the assess-
ment of green biomass in arid environments where soil background becomes a significant component of the signal
detected.
This chapter reviews the character of over 20 VIs that are provided by the TASSCAP and VEGINDEX modules in the
IDRISI system software. They are provided to facilitate the use of these procedures and to further the debate concerning
this very important environmental index. We welcome both your comments on the VIs currently included in IDISI as
well as your suggestions for future additions to the set.
intercept
a b
Figure 1
In contrast to the slope-based group, the distance-based group measures the degree of vegetation present by gauging the
difference of any pixel's reflectance from the reflectance of bare soil. A key concept here is that a plot of the positions of
bare soil pixels of varying moisture levels in a bi-spectral plot will tend to form a line (known as a soil line). As vegetation
canopy cover increases, this soil background will become progressively obscured, with vegetated pixels showing a ten-
dency towards increasing perpendicular distance from this soil line (Figure 1b). All of the members of this group (such as
the Perpendicular Vegetation Index—PVI) thus require that the slope and intercept of the soil line be defined for the
image being analyzed.
To these two groups of vegetation indices, a third group can be added called orthogonal transformation VIs. Orthogonal indi-
ces undertake a transformation of the available spectral bands to form a new set of uncorrelated bands within which a
green vegetation index band can be defined. The Tasseled Cap transformation is perhaps the most well-known of this
group.
A Special Note About Measurement Scales: IDRISI differs from most other GIS and image processing software in
that it supports real number images. Thus the descriptions that follow describe these vegetation indices without rescaling
to suit more limited data types. However, in most implementations, a subsequent rescaling is required to make the index
suitable for expression in an integer form (e.g., a rescaling of values from a -1.0 to +1.0 real number range to a 0-255 8-bit
integer range). In IDRISI, this is not required, and thus the indices are produced and described in their purest form.
The Normalized Difference Vegetation Index (NDVI) was also introduced by Rouse et al. (1974) in order to produce
a spectral VI that separates green vegetation from its background soil brightness using Landsat MSS digital data. It is
expressed as the difference between the near infrared and red bands normalized by the sum of those bands, i.e.:
NIR – RED
NDVI = ------------------------------
NIR + RED
This is the most commonly used VI as it retains the ability to minimize topographic effects while producing a linear mea-
surement scale. In addition, division by zero errors are significantly reduced. Furthermore, the measurement scale has the
desirable property of ranging from -1 to 1 with 0 representing the approximate value of no vegetation. Thus negative val-
ues represent non-vegetated surfaces.
The Transformed Vegetation Index (TVI) (Deering et al., 1975) modifies the NDVI by adding a constant of 0.50 to all
its values and taking the square root of the results. The constant 0.50 is introduced in order to avoid operating with nega-
tive NDVI values. The calculation of the square root is intended to correct NDVI values that approximate a Poisson dis-
tribution and introduce a normal distribution. With these two elements, the TVI takes the form:
NIR – RED-⎞
⎛ -----------------------------
TVI = + 0.5
⎝ NIR + RED⎠
However, the use of TVI requires that the minimum input NDVI values be greater than -0.5 to avoid aborting the opera-
tion. Negative values still will remain if values less than -0.5 are found in the NDVI. Moreover, there is no technical differ-
ence between NDVI and TVI in terms of image output or active vegetation detection.
The Corrected Transformed Vegetation Index (CTVI) proposed by Perry and Lautenschlager (1984) aims at correct-
ing the TVI. Clearly adding a constant of 0.50 to all NDVI values does not always eliminate all negative values as NDVI
values may have the range -1 to +1. Values that are lower than -0.50 will leave small negative values after the addition
operation. Thus, the CTVI is intended to resolve this situation by dividing (NDVI + 0.50) by its absolute value
ABS(NDVI + 0.50) and multiplying the result by the square root of the absolute value (SQRT[ABS(NDVI + 0.50)]). This
suppresses the negative NDVI. The equation is written:
NDVI + 0.5
CTVI = ----------------------------------------------- × ABS ( NDVI + 0.5 )
ABS ( NDVI + 0.5 )
Given that the correction is applied in a uniform manner, the output image using CTVI should have no difference with
the initial NDVI image or the TVI whenever TVI properly carries out the square root operation. The correction is
intended to eliminate negative values and generate a VI image that is similar to, if not better than, the NDVI. However,
Thiam (1997) indicates that the resulting image of the CTVI can be very "noisy" due to an overestimation of the green-
ness. He suggests ignoring the first term of the CTVI equation in order to obtain better results. This is done by simply
taking the square root of the absolute values of the NDVI in the original TVI expression to have a new VI called Thiam’s
Transformed Vegetation Index (TTVI).
TTVI = ABS ( NDVI + 0.5 )
Figure 3
PVI<0
water
PVI=0
soil line
red band
Rgg5,Rgg7
PVI>0
Rp5,Rp7 vegetation
Figure 4 The Perpendicular Vegetation Index (from Richardson and Wiegand, 1977)
1. Check the Help System for each VI to determine which band should be used as the dependent and independent variables. In this example, for the
PVI, red is dependent and infrared is independent.
a –b
0 0
Rgg7 = ----------------
b1 – a1
4) Find the distance between the intersection (Rgg5,Rgg7) and the pixel coordinate (Rp5,Rp7) using the Pythagorean
Theorum.
2 2
PVI = ( Rgg5 – Rp5 ) + ( Rgg7 – Rp7 )
Attempts to improve the performance of the PVI have yielded three others suggested by Perry and Lautenschlager
(1984), Walther and Shabaani (1991), and Qi, et al. (1994). In order to avoid confusion, the derived PVIs are indexed 1 to
3 (PVI1, PVI2, PVI3).
PVI1 was developed by Perry and Lautenschlager (1984) who argued that the original PVI equation is computationally
intensive and does not discriminate between pixels that fall to the right or left side of the soil line (i.e., water from vegeta-
tion). Given the spectral response pattern of vegetation in which the infrared reflectance is higher than the red reflec-
tance, all vegetation pixels will fall to the right of the soil line (e.g., pixel 2 in Figure 18-5). In some cases, a pixel
representing non-vegetation (e.g., water) may be equally far from the soil line, but lies to the left of that line (e.g., pixel 1 in
Figure 5). In the case of PVI, that water pixel will be assigned a high vegetation index value. PVI1 assigns negative values
to those pixels lying to the left of the soil line.
soil line
1 d2
red d1
2
infrared
where
NIR = reflectance in the near infrared band
RED = reflectance in the visible red band
a = intercept of the soil line
b = slope of the soil line
PVI2 (Walther and Shabaani, 1991; Bannari, et al., 1996) weights the red band with the intercept of the soil line and is
NIR − a ∗ Red + b
PVI 2 =
1+ a 2
where
NIR = reflectance in the near infrared band
RED = reflectance in the visible red band
a = slope of the soil line
b = intercept of the soil line
PVI3, presented by Qi, et al (1994), is written:
PVI3 = apNIR - bpRED
where
pNIR = reflectance in the near infrared band
pRED = reflectance in the visible red band
a = intercept of the soil line
b = slope of the soil line
Difference Vegetation Index (DVI) is also suggested by Richardson and Wiegand (1977) as an easier vegetation index
calculation algorithm. The particularity of the DVI is that it weights the near infrared band by the slope of the soil line. It
is written:
DVI = g MSS7 - MSS5
where
g = the slope of the soil line
MSS7 = reflectance in the near infrared 2 band
MSS5 = reflectance in the visible red band
Similar to the PVI1, with the DVI, a value of zero indicates bare soil, values less than zero indicate water, and those
greater than zero indicate vegetation.
The Ashburn Vegetation Index (AVI) (Ashburn, 1978) is presented as a measure of green growing vegetation. The val-
ues in MSS7 are multiplied by 2 in order to scale the 6-bit data values of this channel to match with the 8-bit values of
MSS5. The equation is written:
AVI = 2.0MSS7 - MSS5
This scaling factor would not apply wherever both bands are 7-bit or 8-bit and the equation is rewritten as a simple sub-
traction.
The Soil-Adjusted Vegetation Index (SAVI) is proposed by Huete (1988). It is intended to minimize the effects of soil
2. In Bannari, et al. (1996), a is used to designate the slope and b is used to designate the intercept. More commonly in linear regression, a is the inter-
cept and b the slope of the fitted line. This has been corrected here.
Figure 6 Influence of light and dark soil on the SAVI values of cotton as a function of the shifted correc
factor L (from Huete, 1988).
ρ nir – ρ red
SAVI = ---------------------------------------- ⋅ ( 1 + L )
( ρ nir + ρ red + L )
where
ρnir = near infrared band (expressed as reflectances)
ρred = visible red band (expressed as reflectances)
L = soil adjustment factor
The Transformed Soil-Adjusted Vegetation Index (TSAVI1) was defined by Baret, et al. (1989) who argued that the
SAVI concept is exact only if the constants of the soil line are a=1 and b=0 (note the reversal of these common symbols).
Because this is not generally the case, they transformed SAVI. By taking into consideration the PVI concept, they pro-
posed a first modification of TSAVI designated as TSAVI1. The transformed expression is written:
a ( NIR − a * Red − b)
TSAVI1 =
(Red + a ∗ NIR − a ∗ b)
where
NIR = reflectance in the near infrared band (expressed as reflectances)
RED = reflectance in the visible red band (expressed as reflectances)
a = slope of the soil line
The Modified Soil-Adjusted Vegetation Indices (MSAVI1 and MSAVI2) suggested by Qi, et al. (1994) are based on a
modification of the L factor of the SAVI. Both are intended to better correct the soil background brightness in different
vegetation cover conditions.
With MSAVI1, L is selected as an empirical function due to the fact that L decreases with decreasing vegetation cover as
is the case in semi-arid lands (Qi, et al., 1994). In order to cancel or minimize the effect of the soil brightness, L is set to
be the product of NDVI and WDVI (described below). Therefore, it uses the opposite trends of NDVI and WDVI. The
full expression of MSAVI1 is written:
NIR – RED
MSAVI1 = ---------------------------------------- ⋅ ( 1 + L )
NIR + RED + L
where
NIR = reflectance in the near infrared band (expressed as reflectances)
RED = reflectance in the visible red band (expressed as reflectances)
L = 1 - 2 γ NDVI * WDVI
where
NDVI = Normalized Difference Vegetation Index
WDVI = Weighted Difference Vegetation Index
γ = slope of the background soil line
2 = used to increase the L dynamic range
range of L = 0 to 1
The second modified SAVI, MSAVI2, uses an inductive L factor to:
1. remove the soil "noise" that was not canceled out by the product of NDVI by WDVI, and
2. correct values greater than 1 that MSAVI1 may have due to the low negative value of NDVI*WDVI. Thus, its use is
limited for high vegetation density areas.
The general expression of MSAVI2 is:
2
2ρ nir + 1 – ( 2ρ nir + 1 ) – 8 ( ρ nir – ρ red )
MSAVI 2 = -----------------------------------------------------------------------------------------------------
2
where
ρnir = reflectance of the near infrared band (expressed as reflectances)
ρred = reflectance of the red band (expressed as reflectances)
The Green Vegetation Index (GVI) of the Tasseled Cap is the second of the four new bands that Kauth and Thomas
(1976) extracted from raw MSS images. The GVI provides global coefficients that are used to weight the original MSS
digital counts to generate the new transformed bands. The TASSCAP module in IDRISI is specifically provided to calcu-
late the Tasseled Cap bands from Landsat MSS or TM images. The output from TASSCAP corresponding to GVI is
xxgreen (xx = the two character prefix entered by the user) by default. The expression of the green vegetation index band,
The PCA module in IDRISI generates eigenvectors as well as factor loadings with the component images. A site-specific
MGVI image can then be produced with Image Calculator by using the appropriate eigenvector values. The following
equation would be used to produce the MGVI image for the example shown in Table 18-2:
MGVI90 = (-0.507MSS4) + (- 0.400MSS5) + (0.275MSS6) + (0.712MSS7)
Summary
The use of any of these transformations depends on the objective of the investigation and the general geographic charac-
teristics of the application area. In theory, any of them can be applied to any geographic area, regardless of their sensitivity
to various environmental components that might limit their effectiveness. In this respect, one might consider applying the
slope-based indices as they are simple to use and yield numerical results that are easy to interpret. However, including the
well known NDVI, they all have the major weakness of not being able to minimize the effects of the soil background.
This means that a certain proportion of their values, negative or positive, represents the background soil brightness. The
effect of the background soil is a major limiting factor to certain statistical analyses geared towards the quantitative assess-
ment of above-ground green biomass.
References
Ashburn, P., 1978. The vegetative index number and crop identification, The LACIE Symposium Proceedings of the Technical
Session, 843-850.
Bannari, A., Huete, A. R., Morin, D., and Zagolski, 1996. Effets de la Couleur et de la Brillance du Sol Sur les Indices de
Végétation, International Journal of Remote Sensing, 17(10): 1885-1906.
Baret, F., Guyot, G., and Major, D., 1989. TSAVI: A Vegetation Index Which Minimizes Soil Brightness Effects on LAI
and APAR Estimation, 12th Canadian Symposium on Remote Sensing and IGARSS’90, Vancouver, Canada, 4.
Baret, F., and Guyot, G., 1991. Potentials and Limits of Vegetation Indices for LAI and APAR Assessment, Remote Sensing
and the Environment, 35: 161-173.
Deering, D. W., Rouse, J. W., Haas, R. H., and Schell, J. A., 1975. Measuring “Forage Production” of Grazing Units From
Landsat MSS Data, Proceedings of the 10th International Symposium on Remote Sensing of Environment, II, 1169-1178.
Fung, T., and LeDrew, E., 1988. The Determination of Optimal Threshold Levels for Change Detection Using Various
Accuracy Indices, Photogrammetric Engineering and Remote Sensing, 54(10): 1449-1454.
Huete, A. R., 1988. A Soil-Adjusted Vegetation Index (SAVI), Remote Sensing and the Environment, 25: 53-70.
Jackson, R. D., 1983. Spectral Indices in n-Space, Remote Sensing and the Environment, 13: 409-421.
Kauth, R. J., and Thomas, G. S., 1976. The Tasseled Cap - A Graphic Description of the Spectral Temporal Development
of Agricultural Crops As Seen By Landsat. Proceedings of the Symposium on Machine Processing of Remotely Sensed Data, Perdue
University, West Lafayette, Indiana, 41-51.
Kogan, F. N., 1990. Remote Sensing of Weather Impacts on Vegetation in Nonhomogeneous Areas, International Journal of
Remote Sensing, 11(8): 1405-1419.
Liu, W. T., and Kogan, F. N., 1996. Monitoring Regional Drought Using the Vegetation Condition Index, International Jour-
nal of Remote Sensing 17(14): 2761-2782.
Misra, P. N., and Wheeler, S.G., 1977. Landsat Data From Agricultural Sites - Crop Signature Analysis, Proceedings of the
11th International Symposium on Remote Sensing of the Environment, ERIM.
1. Other RADAR frequencies include X - band 12.5-8.0 Ghz (2.4 - 3.75 cm), L - band : 2.0-1.0 Ghz (15-30 cm) and P-band 1.0-0.3 Ghz (30-100 cm). The frequency of the
C-band is 8.0-4.0 Ghz (3.75-7.5 cm).
2. Other options are vertical transmit, vertical receive (VV), horizontal transmit, vertical receive (HV) and vertical transmit, horizontal receive (VH).
One other method suggested for further reducing speckled effects is multi-look processing (Lillesand and Kiefer, 1987). This
simply involves averaging (with OVERLAY or Image Calculator) scenes from the same area, acquired at different inci-
dence angles, to produce a smoother image.
References
Eliason, E. M., and McEwen, A. S., 1990. Adaptive Box Filters for Removal of Random Noise from Digital Images, Photo-
grammetric Engineering and Remote Sensing, 56(4): 453-458.
Lillesand, T., Kiefer, R. W., and J.W. Chipman, 2004. Remote Sensing and Image Interpretation, John Wiley and Sons, New
York.
RADARSAT International, 1995. RADARSAT Guide to Products and Services, RADARSAT International, Richmond, B.C.,
Canada.
Pairwise Comparisons
With pairwise comparisons we can break down the techniques according to whether they are suitable for quantitative or
qualitative data. Quantitative data has values that indicate an amount or measurement, such as NDVI, rainfall or reflec-
tance. Qualitative data has values that indicate different categories, such as census tract ID’s or landuse classes.
Quantitative Data
Image Differencing
With quantitative data, the simplest form of change analysis is image differencing. In IDRISI, this can be achieved with the
OVERLAY module through a simple subtraction of one image from the other. However, a second stage of analysis is
often required since the difference image will typically contain a wide range of values. Both steps are included in the mod-
ule IMAGEDIFF, which produces several common image difference products: a simple difference image (later - earlier),
a percentage change image (later-earlier/earlier), a standardized difference image (Z-scores), or a classified standardized
difference image (z-scores divided into 6 classes). Mask images that limit the study area may also be specified.
Care must be taken in choosing a threshold to distinguish true change from natural variability in any of these difference
Image Ratioing
While image differencing looks at the absolute difference between images, image ratioing looks at the relative difference.
Again, this could be achieved with OVERLAY using the ratio option. However, because the resulting scale of relative
change is not symmetric about 1 (the no change value), it is recommended that a logarithmic transformation be under-
taken before thresholding the image. The module IMAGERATIO offers both a simple ratio and a log ratio result.
Regression Differencing
A third form of differencing is called regression differencing. This technique should be used whenever it is suspected that the
measuring instrument (e.g., a satellite sensor) has changed its output characteristics between the two dates being com-
pared. Here the earlier image is used as the independent variable and the later image as the dependent variable in a linear
regression. The intercept and slope of this regression expresses the offset and gain required to adjust the earlier image to
have comparable measurement characteristics to the later. In effect, we create a predicted later image in which the values
are what we would expect if there were no change other than the offset and gain caused by the changes in the sensor. The
equation is:
predicted later image = (earlier image * gain) + offset
With the sensor differences accounted for, the predicted later image and the actual later image may then be analyzed for
change. Note that this technique requires that the overall numeric characteristics of the two images be equal except for
sensor changes. The technique may not be valid if the two images represent conditions that are overall very different
between the two dates.
The module CALIBRATE automates the image adjustment process. The input image (the one to calibrate) is used as the
independent variable and the reference image is used as the dependent variable in the regression. The output image is
adjusted to the characteristics of the reference image and thus can be used in a standard comparison operation (such as
IMAGEDIFF or IMAGERATIO) with any image also based on this reference, including the reference image itself.
Note that CALIBRATE also offers options to adjust an image by entering offset and gain values or by entering mean and
standard deviation values.
Band 2
Band 2 magnitude
. Date1
0
direction
. Date1
.
Date2
270 .
Date2
90
Qualitative Data
Crosstabulation / Crossclassification
With qualitative data, CROSSTAB should be used for change analysis between image pairs and there are several types of
output that can be useful. The crosstabulation table shows the frequencies with which classes have remained the same
(frequencies along the diagonal) or have changed (off-diagonal frequencies). The Kappa Index of Agreement (KIA) indi-
cates the degree of agreement between the two maps, both in an overall sense and on a per-category basis. Finally, the
crossclassification image can readily be reclassified into either a change image or an agreement image. Note that the
numeric values of data classes must be identical on both maps for the output from CROSSTAB to be meaningful.
GEOMOD
GEOMOD is a landuse change simulation model that predicts the transition from one landuse state to another landuse
state, i.e., the location of grid cells that change over time from one state to another. GEOMOD simulates the change
between exactly two categories, state 1 and state 2. For example, GEOMOD could be used to predict areas likely to
change from forest (state 1) to non-forest (state 2) over a given time. The simulation can occur either forward or back-
ward in time. The simulation is based on:
-specification of the beginning time, ending time and time step for the simulation,
-an image showing the location of landuse states 1 and 2 at the beginning time,
-an optional mask image distinguishing between areas in and out of the study region,
Cellular Automata
One of the basic spatial elements that underlies the dynamics of many change events is proximity: areas will have a higher
tendency to change to a class when they are near existing areas of the same class (i.e., an expansion phenomenon). These
can be very effectively modeled using cellular automata. A cellular automaton is a cellular entity that independently varies
its state based on its previous state and that of its immediate neighbors according to a specific rule. Clearly there is a sim-
ilarity here to a Markovian process. The only difference is application of a transition rule that depends not only upon the
previous state, but also upon the state of the local neighborhood.
Many cellular automata transition rules can be implemented through a combination of FILTER and RECLASS. Take, for
example, the case of Conway's Game of Life. In this hypothetical illustration, the automata live or die according to the fol-
lowing criteria:
The critical element of this rule is the use of the 10 multiplier in the central cell. As a result of the filter step, you know
that the central cell is occupied if the result is 10 or greater. The CELLATOM module can be used to implement this kind
of Cellular Automaton rule. However, a cellular automaton procedure very specific to the context of predictive landcover
change modeling is implemented with the CA_MARKOV module.
CA_MARKOV takes as input the name of the landcover map from which changes should be projected, the transition
areas file produced by MARKOV from analysis of that image and an earlier one, and a collection (.rgf) of suitability
images that express the suitability of a pixel for each of the landcover types under consideration. It then begins an iterative
process of reallocating landcover until it meets the area totals predicted by the MARKOV module. The logic it uses is this:
- The total number of iterations is based on the number of time steps set by the user. For example, if the projection is for
10 years into the future, the user might choose to complete the model in 10 steps.
- Within each iteration, every landcover class will typically lose some of its land to one or more of the other classes (and it
may also gain land from others). Thus within the consideration of each host within each iteration, claimant classes select
land from the host based on the suitability map for the claimant class. Since there will commonly be competition for spe-
cific land parcels, this process of land allocation is undertaken using a multi-objective allocation procedure (the MOLA
module).
- The Cellular Automaton component arises in part from the iterative process of land allocation, and in part from a filter-
ing stage with each iteration that reduces the suitability of land away from existing areas of that type. By default, the mod-
ule uses a 5x5 mean filter to achieve this contiguity constraint. By filtering a Boolean mask of the class being considered,
the mean filter yields a value of 1 when it is entirely within the existing class and 0 when it is entirely outside it. However,
when it crosses the boundary, it will yield values that quickly transition from 1 to 0. This result is then multiplied by the
suitability image for that class, thereby progressively downweighting the suitabilities as one moves away from existing
instances of that class. Note that it is possible to apply a different filter by specifying an alternative filter file (.fil). Also
note that class masks are defined at each step to incorporate new areas of growth.
Model Validation
An important stage in the development of any predictive change model is validation. Typically, one gauges one's under-
standing of the process, and the power of the model, by using it to predict some period of time when the landcover con-
ditions are known. This is then used as a test for validation. IDRISI supplies a pair of modules to assist in the validation
process.
The first is called VALIDATE, and provides a comparative analysis on the basis of the Kappa Index of Agreement.
Kappa is essentially a statement of proportional accuracy, adjusted for chance agreement. However, unlike the traditional
Kappa statistic, VALIDATE breaks the validation down into several components, each with a special form of Kappa or
associated statistic (based on the work of Pontius (2000)):
· Kappa for no information = Kno
· Kappa for location = Klocation
· Kappa for quantity = Kquantity
· Kappa standard = Kstandard
· Value of Perfect Information of Location = VPIL
· Value of Perfect Information of Quantity = VPIQ
With such a breakdown, for example, it is possible to assess the success with which one is able to specify the location of
change versus the quantity of change.
The other validation procedure is the ROC (Relative Operating Characteristic). It is used to compare any statement about
the probability of an occurrence against a Boolean map which shows the actual occurrences. It can be useful, for example,
in validating modifications to the conditional probability maps output from MARKOV. Note that LOGISTICREG incor-
porates ROC directly in its output.
References:
Pontius Jr., R.G., 2000. Quantification error versus location error in comparison of categorical maps. Photogrammetric
Engineering and Remote Sensing. 66(8) pp. 1011-1016.
Introduction
The Land Change Modeler (LCM) for Ecological Sustainability is an integrated software environment for Commissioned
by the Andes Center for Biodiversity Conservation of Conservation International (our inspiration for the Andes Edition
name), LCM is the first extensive vertical application developed by Clark Labs. (IDRISI is a horizontal application – a
software product meant to fulfill many applications. In contrast, a vertical application is directed towards a specific appli-
cation.) The Land Change Modeler for Ecological Sustainability is oriented to the pressing problem of accelerated land
conversion and the very specific analytical needs of biodiversity conservation.
Within each tab, a series of tasks/analytical stages are presented as a series of drop-down panels. You can have as many
drop-down panels open as you wish – they are presented this way simply to accommodate varying screen resolutions.
The first three of the five tabs of LCM are intended for the integrated analysis of landcover change and its projection into
the future. As a result access to almost all panels on these tabs requires the specification of a minimal set of project
parameters located on the first panel of the first tab.
Note that the panels on the first three tabs are generally organized around a sequential set of operations that should be
followed one after the other.
A Recommendation
We strongly recommend that you complete the tutorial exercises for LCM. This is the fastest way of learning the full scope
of the system.
Notes
1. The landcover maps must be byte or integer images with identical legends.
Notes
1. Changing the units on this panel causes the units on the Change Maps panel to also change.
Notes
1. The Ignore Transitions checkbox is very important – please read the entire text of this note. This checkbox is used to fil-
ter out minor transitions that may be the result of map errors or may be considered to be insignificant for the purpose of
the study. This checkbox affects not only the maps produced from this panel, but also the transitions that are automati-
cally included for analysis on the Transition Potentials tab. This is the quickest and most effective way of narrowing down
the transitions to those that are essential for understanding and modeling change.
2. Specifying an output name is optional. If one is not specified, a temporary filename is used.
3. The Map Exchanges option is designed for the examination of exchanges such as those between agriculture and sec-
ondary forest in areas of swidden agriculture.
4. Changing the measurement units on this panel also changes the units on the Change Analysis panel.
5. Note that an All option is provided in the drop-down lists of landcover categories. Thus choosing to map the changes
from All to a specific category maps any change that ended up in the designated category, differentiated by the start cate-
gory.
Notes
1. The intention of this module is to provide a means of generalizing about the pattern of change. The numeric values do
not have any special significance. The surface is created by coding areas of change with 1 and areas of no change with 0
and treating them as if they were quantitative values.
2. The analytical work done by this option is achieved by a call to the TREND module. For details on how it works, please
refer to the on-line Help System for TREND.
Multi-Layer Perceptron
Initially the dialog for the Multi-Layer Perceptron neural network may seem daunting, but most of the parameters pre-
sented do not need to be modified (or in fact understood) to make productive use of this very powerful technique.
As launched by LCM, the Multi-Layer Perceptron starts training on the samples it has been provided of pixels that have
and have not experienced the transitions being modeled. At this point, the MLP is operating in automatic mode whereby it
makes its own decisions about the parameters to be used and how they should be changed to better model the data. In
most cases, you can let it run in this mode until it completes its training. However, you are also free to stop its training
operation, modify parameters and start it training again. Ultimately, after training has been completed, you will need to
click the Classify button to complete the process of transition potential modeling. When it is finished, LCM will display
each of the transition potential models. For more information about LCM’s specific use of the MLP, see Note 3 in the sec-
tion titled How It Works in this chapter. For detailed information on the MLP parameters, see the on-line Help System
for MLP.
Note some specific tips on using MLP in this context:
- The critical factor in the use of the MLP is the learning rate. What you ideally want to achieve is a smooth descent of the
RMS error curve. If it is flat over a large number of iterations (more than 2000), stop the training (by clicking the Stop
button) and halve the start and end learning rates. Continue to do this as necessary until the error curve descends.
- If the RMS error curve has descended and flattens out over a large number of iterations (>1000), stop the training and
proceed to the Classify button. If, however, you experience a slow but progressive increase in accuracy and decrease in the
RMS errors, let the MLP run until the end of its iterations. If it reaches the end of its iterations and it still appears to be
learning (the accuracy is increasing and the RMS is dropping), re-run it with a larger number of iterations (e.g., an addi-
tional 25%).
- In general, manipulating the learning rate alone will yield better than 90-95% of the best solution. In general, we do not
Logistic Regression
In contrast to the MLP, logistic regression can only model a single transition at a time. Thus it launches in a mode that is
ready to model the specific transition indicated in the drop-down combo box in the Transition Sub-Models: Status panel.
1. Dynamic road development in IDRISI was inspired by the pioneering work of the DINAMICA team. See Soares-Filho, B.S.; Assunção, R.M.; Pan-
tuzzo, A. Modeling the spatial transition probabilities of landscape dynamics in an Amazonian colonization frontier. BioScience, v. 51, p.1039-1046, 2001.
Display Options
LCM provides several options for display of the prediction. One is to display the intermediate stage images (as opposed to
only the final prediction). This option should be used with care as Windows display memory can be rapidly exhausted,
putting the entire Windows system into an unstable state. The limits here will depend, in part, on how much RAM is
installed on your system.
A second option for display is to create an AVI video file. In IDRISI, this file can be played in IDRISI’s Media Viewer – a
utility provided under the Display menu. It can also be played with programs such as Microsoft Media Player and can be
inserted into a Microsoft PowerPoint presentation. For long sequences, a frame rate of 0.25 generally works well, but
slower rates may be more appropriate for slower sequences.
For more information about how the hard allocation procedure is undertaken, see the section titled How It Works in this
chapter.
2. The habitat assessment procedure introduced here was inspired by the work of the Bow Corridor Ecosystem Advisory Group (BCEAG) in the devel-
opment of a corridor strategy for the Southern Canmore Region of Alberta, Canada. For more information, please refer to http://www.stratalink.com/
corridors/wildlife_corridors_report.htm.
Normalized Entropy
This measure is Shannon’s Entropy measure normalized by the maximum entropy for the number of landcover classes
involved. Another common term for this measure is Diversity. It is calculated over the local neighborhood of each pixel,
defined as either a 3x3, 5x5 or 7x7 neighborhood. The formula is as follows:
E = -Σ(p*ln(p)) / ln(n)
where p is the proportion of each class within the neighborhood, ln is the natural logarithm3 and n is the number of
classes. The result is an index that ranges from 0-1 where 0 indicates a case where the landcover is uniform within the
neighborhood and 1 indicates maximum diversity possible of landcovers within the neighborhood.
Relative Richness
This is another measure of diversity of cover classes, measured as:
R = n/nmax*100
where n is the number of different classes present in the neighborhood and nmax is maximum number of classes possible.
Edge Density
Edge Density is a simple measure of fragmentation. Edge density is tabulated as the number of adjacent pairs of pixels
within the neighborhood that are different from each other relative to the maximum number of different pairs possible.
Patch Area
Patch Area groups adjacent pixels of similar landcover category into patches, calculates their areas, and outputs an image
where each pixel expresses the area of the patch to which it belongs.
Patch Compactness
Patch Compactness groups adjacent pixels of similar landcover category into patches, calculates their compactness, and
outputs an image where each pixel expresses the compactness of the patch to which it belongs. Compactness is calculated
as:
C = SQRT(Ap/Ac)
where SQRT is the square root function, Ap is the area of the patch being calculated, and Ac is the area of a circle having
the same perimeter as that of the patch being calculated.
Change Process
The Change Process option compares the earlier and later landcover maps and measures the nature of the change under-
way within each landcover class. It does this by using a decision tree procedure that compares the number of landcover
patches present within each class between the two time periods to changes in their areas and perimeters.4 The output is in
3. Log base 2 is more commonly used in communications theory, but the difference is immaterial with this normalized procedure.
General Logic
The underlying principle of the refinement process is to uncover the common environmental logic of the areas delineated
by the range polygon. It does this by creating clusters of environmental conditions according to a set of environmental
variables that the user believes can characterize the niche of the species. It then compares these clusters with the range
polygon to determine the proportional inclusion of clusters within the range polygon. Clusters that fall wholly or largely
within the polygon are assumed to describe essential components of that niche. Those that fall mostly or wholly outside
are assumed to be unlikely components. The polygon is thus refined by removing areas that fall below a designated confi-
dence. In addition, another option is provided to simply output a confidence map that can be used in conjunction with
the original range polygon by the Weighted Mahalanobis Typicality procedure in the Habitat Suitability / Species Distribu-
tion panel. This is the default option and the one we generally recommend.
4. This is an implementation of the procedure outlined in Bogaert, J., Ceulemans, R., and Salvador-Van Eysenrode, D. (2004) “Decision tree algorithm
for detection of spatial processes in landscape transformation.” Environmental Management, 33, 1, 62-73.
Output Options
Four output options are provided:
1. Presence. This is a refined range polygon where areas that are poorly associated with the core environmental character-
istics of the original range polygon are removed.
2. Presence/Pseudo-Absence. The output is the same as the above except that areas that are extremely unlikely to be asso-
ciated with the core environmental characteristics of the original range polygon are treated as absence while only those
that have a close association are considered as presence.
3. Confidence. This is the default option and the one we generally recommend. Each pixel within the original polygon is
assigned a confidence value from 0-1 based on how well it fits the general nature of a coherent pattern of environmental
conditions (as will be explained further below).
4. Thresholded Confidence. This option is the same as the above, except that areas that fall below a minimum specified
confidence are forced to have a confidence of zero.
Thresholds
For all options except the Confidence output, an upper and/or lower threshold will need to be selected to establish areas
of presence or absence. The default thresholds will serve as a general guideline of the values that would be used. In gen-
eral, for presence, you are looking for a value that separates a clear group of clusters that strongly overlap the range poly-
gon, while for absence you want to isolate clusters that have very little or no presence in the polygon. In many instances,
this is very hard to do, which is why we recommend the use of the Confidence option coupled with the Weighted Mahala-
nobis Typicality procedure in the Habitat Suitability / Species Distribution panel. Using this option, no decision needs to
be made.
Background Mask
The background mask option is quite important to the use of this procedure. If you are modeling a land species and are
working in an area with significant ocean areas, you should provide a mask image to remove these from the calculations of
proportional areas. The mask should have 1’s over land areas and 0’s over water areas. For marine species, clearly the
opposite applies.
5. We tested several clustering procedures including K-Means, Fuzzy ARTMAP and SOM. However, the Histogram Peak technique provided by the
CLUSTER module in IDRISI was so much superior to the others that we decided to use it despite the limitation on the number of independent vari-
ables that could be used.
Assign a weight to each factor. Factor weights can be any numeric value that is convenient for expressing the rel-
ative importance of each to the final suitability map. The weights will automatically be normalized to a 0.0-1.0
range before use.
Add any constraints necessary. Contraints are Boolean images which exclude areas from consideration. They
should have 0’s in constrained areas and 1’s otherwise.
Choose an aggregation option (see below) and then click OK to create the suitability map.
Note that factors can also be added or removed directly from the factors grid. However, be sure that any directly added
factors have a 0.0-1.0 numeric range.
Standardization Options
The FUZZY Option
With the FUZZY option, you need to indicate the nature of the relationship between the variable and suitability. The
graph will illustrate each case along with the general positions of the control points for linking the curve to your variable.
The graph will also indicate the nature of the various shape options. For further information, please refer to the on-line
Help System on the FUZZY module.
The ASSIGN Option
With the ASSIGN option, you will be provided with a grid in which you must indicate the identifiers of classes in the left-
hand column and the suitabilities that should be assigned in the right-hand column (on a 0.0-1.0 range). Any classes that
are not included in this grid will automatically be assigned a value of 0.
Aggregation Options
The aggregation options dictate how the factors will be combined to create a single suitability map. The default is
weighted linear combination (WLC) which is appropriate when you wish the factors to trade-off (i.e., to allow poor quali-
ties to be compensated by good qualities). The Minimum operator allows no trade-off and characterizes each location by
its worst quality. This is clearly the most conservative operator. The Maximum operator also allows no trade-off, but char-
acterizes locations by their best quality.
If you find that you have some factors that should trade-off and others that should not, process the group that do trade-
off first. Then combine that result with the others that do not trade-off using either the Minimum or Maximum operator.
Presence Data
Presence data is probably the most common form of training data for species modeling – it records where the species has
been observed, but not where it has been observed to be absent. Two procedures are available for dealing with these data.
Mahalanobis Typicality
The Mahalanobis Typicality option assumes that the underlying species distribution is normal with respect to environ-
mental gradients. However, our tests have shown that it performs reasonably even with mildly skewed data. The output is
in the form of typicality probabilities – an expression of how typical the pixel is of examples it was trained on. Thus a
value of 1.0 would indicate a location that is identical to the mean of environmental conditions that were evident in the
training data. However, be careful about the interpretation of low typicalities. Since typicalities express the full range of
variability, a low typicality may be unusual, but still legitimately a location that is part of the species’ range. If you are look-
ing for a threshold for when to consider an area as being unlikely to be part of its range, it is likely to be a very low value
(e.g., 0.001). As an illustration of this concept, consider the case of a blue lobster. Blue lobsters are very rare, but they are
Abundance Option
With abundance data, the MULTIREG (multiple regression) option is launched with the selected environmental variables
loaded. Please refer to the on-line Help System for MULTIREG regarding this option.
Gamma Diversity: the total number of considered species over a large region.
Beta Diversity: the ratio of Gamma to average Alpha Diversity over a large region, and thus a measure of the
turnover of species. There are many measures of beta diversity that have been proposed. The measure used here
Range Restriction: a continuous measure of vulnerability that can also be interpreted as a measure of endemism.
Input Data
In all cases, the input for this analysis is in the form of species range polygons. Three input formats are supported. The
first is a vector composite polygon where all species polygons are contained within the same vector file. The second is a
vector group file that lists the names of a set of vector files that contain the range polygons for a single species6. The third
is a raster group file that lists a set of raster files that contain the rasterized range polygons of a single species.
With the exception of Alpha Diversity, the data must ultimately be converted to a raster form for analysis. Thus if a vector
group file is supplied, each file is rasterized (using the spatial characteristics of the reference file) and a raster group file is
created with the same name as the vector group file. If a vector composite file is used, it is first broken out into a set of
separate vector files along with a vector group file of the same name as the vector composite file. These vector files are
then in turn rasterized.
An Important Note
Because of the potentially large number of files that may be generated by this analysis, the option is provided to delete
generated intermediate layers. However, if you intend on running further analyses with the same data, it is recommended
that you do not choose this option until the last run and that subsequent analyses be run from the raster group file.
Regional Definition
All measures except Alpha Diversity and the Range Restriction Index require the definition of a region over which the
index is calculated. Three options are provided. The vector and raster region polygon options will yield a mapping where
all pixels within a region (such as an ecoregion) will have the same index value. The focal zone option, however, is quite
different and can produce a different value at each pixel location.
The focal zone option calculates values by comparing the species composition in each pixel to those in a circular zone sur-
rounding it. To use the focal zone option, you must set the focal zone diameter (e.g., 50 km). This focal zone is moved
successively over every pixel in the image. As a consequence, the analysis does take considerable time to complete. Conti-
nental scale analyses at a moderate resolution (e.g., 1 km) are probably best set up at the end of the day so that they can
run overnight.
Alpha Diversity
Alpha Diversity is computed simply as the richness of species at each location – i.e., it is the total number of species found
at each location.
Gamma Diversity
Gamma Diversity is calculated as the richness of species over a region. Thus the value recorded at any pixel represents the
richness within the region to which it belongs and not the richness at that particular spot.
Beta Diversity
Beta Diversity is calculated as gamma diversity divided by the average alpha diversity within each region. This formulation
is the original one developed by Whittaker (1972)7.
6. Vector group files have the same format as raster group files. At this time, vector group files are only used for the biodiversity analyses in LCM.
Range Restriction
The Range Restriction Index is based on a comparison of the area over which the species is found relative to the entire
study region. It is intended for continental or global scale analyses and should include a mask file to mask out water areas
for land species or vice versa for marine species. The index ranges from 0-1 with high values indicating that the majority
of species present at that location have restricted ranges. Note that the index is continuous and does not rely on a thresh-
old area to define range restriction. For more information on the Range Restriction Index, see the section titled How it
Works in this chapter.
Important Note
Use incentives and disincentives with care. Small changes can have huge impacts. In normal use, you will have areas of
absolute constraints (0), areas where normal transition potentials apply (1) and a few areas that may be slightly above or
below 1 (e.g., 0.9 to 1.1).
7. Whittaker, R.H. (1972) “Evolution and measurement of species diversity”, Taxon, 21, 213-251.
How it Works
1: The Evidence Likelihood Transformation Utility
The Evidence Likelihood transformation requires two inputs:
1. A Boolean map of areas that have gone through the transition being modeled (this can easily be created from the
Change Analysis tab).
2. A categorical variable or a continuous variable that has been binned into classes (such as with the STRETCH module).
The procedure looks at the relative frequency of pixels belonging to the different categories of that variable within areas
of change. In effect, it asks the question of each category of the variable, “How likely is it that you would have a value like
this if you were an area that would experience change?”
2
n ⎛ ⎛ range _ area ⎞⎞
∑ ⎜ 1 − ⎜⎜
⎜ ⎟⎟ ⎟⎟
i =1 ⎝ ⎝ total _ area ⎠⎠
RRI =
Alpha _ Diversity
where alpha diversity is expressed as richness (the number of species) and the total area is the total area of the image
minus any masked areas.
References
Eastman, J. R., L. Solorzano and M. Van Fossen. "Transition Potential Modeling for Land-Cover Change." In GIS, Spatial
Analysis and Modeling, edited by David J. Maguire, Michael Batty and Michael F. Goodchild, 357-385. Redlands, CA:
ESRI Press, 2005.
Introduction
Environmental image series provide a critically important resource for understanding both the dynamics and evolution of
environmental phenomena. As a consequence, Earth Trends Modeler (ETM) is focused on the analysis of trends and
dynamic characteristics of these phenomena as evident in time series images. Further, the system is highly interactive, with
the process of exploration largely being an active process. The trends and dynamics emphasized include:
-- Interannual trends – trends that persist over several years or longer that may be indicative of major environmental
changes. ETM provides tools for determining the presence of trends that are both linear and monotonic (non-linear) as
well as their significance. In addition, trend analysis tools are provided that are resistant to the presence of outliers in short
time series.
-- Seasonal trends – trends in the character of the annual seasonal progression. This is a revolutionary new analytical pro-
cedure developed for ETM. The axis of the earth leads to substantial annual variation in solar energy, leading to major
seasonal cycles. However, traditional multi-year trend analysis procedures consider seasonality to be a contaminant and
thus intentionally reject it. This new procedure specifically seeks trends in seasonality and displays it in a dramatic manner.
-- Cyclical components – ETM offers a completely new procedure for examining cyclical components in image time
series. Marrying the elements of Fourier Analysis and Principal Components Analysis, ETM provides a unique form of
Spatial-Temporal Spectral Analysis, called Fourier PCA, that discovers not simply cycles in isolation, but patterns of cycles
that occur together. A final stage of correlation analysis also shows when these patterns were most prevalent.
-- Irregular but recurrent patterns in space/time -- ETM offers a variety of tools for the analysis of recurrent patterns in
space/time including a unique form of Wavelet Analysis and both standardized and unstandardized Principal Compo-
nents Analysis (also known as Empirical Orthogonal Functional Analysis).
-- Teleconnections – ETM offers cutting-edge tools for the analysis of climate teleconnections.
-- Sub-annual variability – Short-term variability and motion over space and time are among the most difficult phenomena
to detect and understand.
A Recommendation
We strongly recommend that you complete the tutorial exercises for ETM. This is the fastest way to learn the full scope of
the system.
1. The troposphere is the portion of the atmosphere in which weather events (e.g., rain storms) occur. It extends to approximately the cruising altitude
of large aircraft.
3. These files contain the three planes of the visualization cube. BSQ = band sequential, containing a single image for each time slice. BIL = band inter-
leaved by line, containing an X/time slice for position in Y. BIP = band interleaved by pixel, containing a Y/time slice for each X position.
4. Use the middle instant stretch option on Composer for series that should be symmetric about 0 (i.e., they contain both negative and positive values)
and the left-most instant stretch option otherwise.
PCA
Principal Components Analysis (also known as Empirical Orthogonal Function analysis) decomposes an image series into
a set of underlying components, ordered by the amount of variance they explain in the original series. For each compo-
nent, a pair of outputs is provided – a component image showing the spatial pattern of the component and a loading
graph that shows the degree to which that pattern is present over time6. Please see the explanatory discussion of PCA in
the Analysis Tab section below.
All series for which a PCA analysis has been run will be listed in the series drop-down selector. Select a series and then the
adjacent drop-down will list specific analyses. After you select the specific analysis, the component loading of the first
component view will appear as a graph. To view the associated component image, click the Display icon at the upper right
of the form. If your component image has both positive and negative values, it is recommended that you use the symmet-
ric instant stretch option on Composer (the middle button) to achieve a proper visual interpretation of the component.
EOT
An Empirical Orthogonal Teleconnection analysis also produces a series of components that consist of a pair of outputs.
The graph is the EOT itself while the image indicates the partial correlation of the series with that EOT with the effects
of other EOTs removed. Please see the explanatory discussion of EOT in the Analysis Tab section below.
All series for which an EOT analysis has been run will be listed in the series drop-down selector. Select a series and then
the adjacent drop-down will list specific analyses. After you select the specific analysis, the first EOT will appear as a
graph. To view the associated partial correlation image, click the Display icon at the upper right of the form. If your com-
ponent image has both positive and negative values (which is very likely), it is recommended that you use the symmetric
instant stretch option on Composer (the middle button) to achieve a proper visual interpretation of the partial correlation
5. Named after the climatologist who is widely attributed to have first proposed the utility of this view.
6. Empirical Orthogonal Function (EOF) analysis produces the exact same output, with the exact same interpretation, as PCA. However in EOF, the
graph is called the component and the image is called the loading map.
Fourier PCA
Fourier PCA also produces a series of components, but in this case with an image and several possibilities of associated
graphs. The default graph is a pseudo-periodogram where the X axis indicates frequencies ranging from the lowest fre-
quency of one sine wave over the whole series to the highest with n/2 sine waves over the whole series (where n is the
total number of images in the series). The Y axis indicates the amplitude loading of the wave. The associated component
image is that which has this combination of waves present (the degree to which waves are present is represented by their
amplitude loading). In the upper-right corner of the pseudo-periodogram is a drop-down list where you can choose to
focus on just the inter-annual frequencies (those wavelengths that are longer than a year) or the sub-annual frequencies. In
addition, there is a temporal loading option. The temporal loading indicates the correlation between the Fourier PCA
component image and each of the original images in the series. Please see the explanatory discussion of Fourier PCA in
the Analysis tab section below.
All series for which a Fourier PCA analysis has been run will be listed in the series drop-down selector. Select a series and
then the adjacent drop-down will list specific analyses. After you select the specific analysis, a periodogram of the first
Fourier component will appear as a graph. To view the associated component image, click the Display icon at the upper
right of the form. If your component image has both positive and negative values, it is recommended that you use the
symmetric instant stretch option on Composer (the middle button) to achieve a proper visual interpretation of the com-
ponent.
Wavelets
The wavelet view requires special explanation. Wavelets are used for multiple purposes. A common application is data
compression. However, the application here is the search for patterns of anomalous events over time. One such pattern
would be a perfect cycle, for which Fourier Analysis is a good detection tool. A perfect cycle is one which oscillates in a
consistent manner over time, such as a sine wave. However, not all oscillations are perfect, nor are they necessarily sine
waves or continuous over time.
A wavelet is a little wave, or perhaps better stated, a briefly appearing wave. Wavelets can be of any form. For example,
one could use a sine wave as a wavelet. In practice, there are a variety of wavelets that are used for special reasons. In
ETM, we have introduced an Inverse Haar wavelet that leads to a very simple form of interpretation in the context of
image time series.
The Inverse Haar wavelet looks like this:
+1
0 0
-1
The graph represents a series of weights over four adjacent samples for a single pixel in time. However, since the weight
of 0 applies to all time samples before the first and after the fourth, it’s probably simpler to think of it as applying to a pair
of adjacent samples over time as follows:
1 month
-1
Applying the wavelet filter to any pixel allows us to determine the degree to which the wavelet is present at any moment in
time. Let’s assume that for a particular pixel, we have the following values over a sequence of seven months:
5 7 15 12 8 9 3
Moving the filter to the first two positions, we multiply the 5 by -1 and the 7 by +1. Then we add the results which yields
+2. Now the filter is slid one time sample to the right. We multiply the 7 by -1 and the 15 by +1 which when added
together yields +8. In essence, we’re saying that the wavelet is present four times as strongly at this second position. Now
move it to the right one more time step. Here we see something different. We multiply the 15 by -1 and the 12 by +1
yielding -3 after summation of the results. This implies that the wavelet is present in its opposite orientation with an
amplitude of 3.
Why is this of interest? If you look at the nature of the mathematical operations undertaken at each time step, it equates to
calculating the rate of change from one point of time to the next. The result of the Inverse Haar filter for the first two
dates indicates that the rate of change is +2, i.e., a gain of two units. The filter thus results in an expression of monthly
gains and losses.
In the next step, we increase the width of the filter to span over four months:
+1
2 months
2 months
-1
Applying this filter, we will first average over the months spanned by each of the two month filter sections. Thus applying
this to the first four months in the series, the 5 and the 7 are averaged to become 6 and the 15 and 12 are averaged to
become 13.5. Then multiplying the 6 by -1 and adding it to 1 times 13.5 yields a result of +7.5. This indicates that there
was a gain of 7.5 units between the first 2 month period and the second 2 month period. Then the filter is slid just one
month. The next calculation is thus on the months with values of 7, 15, 12 and 8, yielding a result of -1. Sliding this filter
only a single month produces what is known as a Maximal Overlap Discrete Wavelet Transform (MODWT). The use of a
MODWT filter has advantages in that it can handle any length of series (and not just those that have lengths that are a
In this example, we see time along the X axis and scale (expressed as the number of months in each half of the Haar filter)
along the Y axis. In ETM, you can move the mouse over the diagram and it will tell you the date, the scale and the value
of the wavelet transform that the color represents.
The importance of using the Inverse Haar filter is that the values it calculates have a very simple interpretation – they are
simply gains in the original units. Thus in the example above based on an analysis of sea surface temperatures in the Lab-
rador sea, the value represents a loss of 0.89 degrees Celsius (because the gain is -0.89). Notice the intense cooling that
took place in 1989. At the 1 month scale, the cooling took place over only a few months (looking horizontally), but look-
ing vertically, we see that its impact lasted for approximately 6 years (72 months). Many of the warmings and coolings are
evident at both the finest and coarsest scales. They thus have a tree-like appearance in the wavelet diagram. However, not
all do. In the middle of the diagram, we see evidence of warming that does not manifest itself at finer scales. This repre-
sents a very gradual warming that is lost in the higher variability that occurs at finer scales.
7. See Chapter 5 on the Maximal Overlap Discrete Wavelet Transform in Wavelet in Percival, D.B., and Walden, A.T., (2000) Wavelet Methods for Time Series
Analysis (Cambridge University Press).
8. The number of different trend elements selected affects the speed with which ETM can calculate the necessary information. Therefore, by default,
only the information necessary for the fitted curves is extracted for the sample region. You can select additional elements by using the checkboxes pro-
vided. The one you will most typically want is the observed curve information.
Linearity
This procedure maps out the coefficient of determination (r2) from a linear regression between the values of each pixel
over time and a perfectly linear series. The result is a mapping of the degree to which a linear trend is present.
Linear Correlation
This maps out the Pearson Product-Moment linear correlation between the values of each pixel over time and a perfectly
linear series. This is a commonly used form of trend analysis, but it is sensitive to noise in short series.
9. See HOAGLIN, D.C., MOSTELLER, F., and TUKEY, J.W., 2000, Understanding Robust and Exploratory Data Analysis, Wiley Classics Library Edition, (New
York: Wiley).
Mann-Kendall Significance
This option produces a pair of images – a significance image expressed as Z scores and a second image that expresses the
probability that the observed trend could have occurred by chance. Strictly speaking, this option is expressing the signifi-
cance of a Mann-Kendall trend. However, it is commonly used as a trend test for the Theil-Sen median slope operator as
well.
10. With a Mann-Kendall statistic, the data series is the dependent variable and time is the independent variable. When the independent variable is some-
thing other than time, the statistic is known as Kendall’s Tau. In that case, one looks at whether the two variables are both increasing or both decreasing
(known as a concordance) or whether one is increasing while the other is decreasing (a discordance) between every pairwise combination of observa-
tions. Tau is then the relative frequency of concordances minus the relative frequency of discordances.
11. See Eastman et al., (in press) “Seasonal Trend Analysis of Image Time Series,” International Journal of Remote Sensing.
After the first component has been calculated, its effects are removed from the pixel vector field12. The new vectors thus
express the residuals after removing the effects of the first component. Then the process is repeated to extract the second
12. Although the actual calculation of components is not done this way, this step is equivalent to calculating the partial correlations between each pair of
pixel vectors, while removing the effects of the component.
13. Van den Dool, H. M., Saha, S., Johansson, A., (2000) Empirical orthogonal teleconnections. Journal of Climate, 13:1421-1435; Van den Dool, H.,
(2007) Empirical Methods in Short-Term Climate Prediction. Oxford University Press, New York. Note that we STRONGLY recommend this book. It is
exceptionally well written and provides many useful insights into image time series analysis.
Harmonic Interpolation
This option is closely based on the procedure known as HANTS (Harmonic Analysis of Time Series) by Roerink, et al.,
(2000)14. The procedure is best for filling missing data in mid-latitude areas. It is not suitable for high-latitude or desert
regions with long periods of uniform response (e.g., snow cover) or for tropical rainforest regions where seasonality is
questionable or extremely subtle.
The basic logic of the procedure is as follows. In mid-latitude areas, the seasonal curve is very well described by the addi-
tive combination of a small set of sine waves. This can be modeled by using either Fourier Analysis or Harmonic Regres-
sion. As in the HANTS procedure, Harmonic Regression is used because it can accommodate missing data and does not
require that the data be at a consistent interval, i.e.,
n=2
⎧ 2πnt 2πnt ⎫
y = α 0 + ∑ ⎨an sin( ) + bn cos( )⎬ + e
n =1 ⎩ T T ⎭
where y is the series value, t is time, T is the length of the series, n is the number of harmonics to be used in the regression,
e is an error term and α0 is the mean of the series. The Julian dates associated with each step in the series cycle (contained
in the .tsf file) provides the time information needed for the regression.
For each missing value in the series the harmonic regression uses a sliding window of one year, centered on the missing
date. The series type will therefore dictate how many data values are used in the regression. The number of harmonics will
14. Roerink, G.J., Menenti, M., and Verhoef, W., (2000) “Reconstructing cloudfree NDVI composites using Fourier analysis of time series”, International
Journal of Remote Sensing, 21, 9, 1911-1917.
Spatial Interpolation
With spatial interpolation, ETM looks at the 3 by 3 neighborhood surrounding a missing pixel and replaces it with the
median value. This will only fill in the pixel if the majority of the neighboring pixels have valid data.
Temporal Filter
The Temporal Filter option smoothes over time. It does so using a symmetric moving filter window. Filter lengths can be
both odd and even numbered except in the case of either a Gaussian filter or a Maximum filter (both of which must be
odd). In the case of even filters, the first and last images have a half weight each in the filtering operation. Sub-options
include:
--Mean filter – the resulting pixel values are the simple averages of all time periods within the moving temporal window.
--Gaussian weighted mean – the resulting pixel values are the weighted averages of all time periods within the moving
temporal window, where the weights follow a Gaussian distribution. This tends to produce a smoother series, but will typ-
ically need a wider filter window than the simple mean filter.
--Maximum value – the resulting pixel values represent the maximum value that occurs over the temporal filter window.
--Cumulative sum – the resulting pixel values represent the sum of the pixel being operated upon and all previous values
that occurs within the length of the temporal filter window.
Inverse PCA
Unstandardized Principal Components Analysis provides a linear transformation of a set of components in which the
early components describe highly prevalent and coherent sources of variability in the series. On the other hand, the later
components will contain minor sources and incoherent sources of variability such as noise. Thus an effective means of
denoising data is to reverse the transformation while leaving out these noisy elements. How many should you leave out?
There is no easy answer here without looking at the components quite carefully. However, most of the coherent variability
in a series is contained within a surprisingly few components – e.g., the first 20 or so. For filtering for the purpose of using
a Cross-EOT, some would recommend even a heavier filtering, using only perhaps the first five components in re-con-
structing the series. Note that in using this option, ETM automatically computes both the forward and inverse transfor-
mations.
Inverse Fourier
The logic here is similar to Inverse PCA. From a forward breakdown of an image series into a set of amplitude and phase
images, it is possible to reconstruct it using an inverse transformation. The higher-numbered harmonics represent increas-
ingly high-frequency elements such as noise. A value that commonly works well is to include all the interannual cycles, the
annual cycle and the semi-annual cycle. The cutoff frequency is thus going to depend of the length of the series. If, for
example you have 25 years of monthly data, interannual cycles would be the harmonics less than 25, the annual cycle
would be harmonic 25 and the semi-annual cycle would be harmonic 50. Therefore the number of harmonics to use
would be 50.
Anomalies
This is the most commonly used form of deseasoning. As a first step, ETM creates what is known as a climatology from the
series as a whole. For example, if the data are monthly, then it creates an average January, an average February, and so on.
It then goes through the series and subtracts the long term average from each month. For example, the January 1998
anomaly image would be equal to January 1998 minus the long term average of all Januaries.
Standardized Anomalies
This is a variant of regular anomalies. In calculation of the climatology (the long-term means), it also calculates the standard
deviation of values for each member of the climatology. Then when the anomaly images are created (by subtracting the
long term mean) the result is divided by the standard deviation to create a standardized anomaly (a z-score). In this new
system, a value of 0 would mean that it has a value equal to the long term mean, a value of +1 would mean that it is 1 stan-
dard deviation of the long term mean, and so on.
Lagged Series
Generates a lagged version of the input series. Specifying a lag other than 0 will start the series with the position indicated,
starting the count from 0. For example, for a monthly series that starts in January, specifying a lag of 3 will have the series
start from April and end with the normal end of the series. To pair an unlagged series of the same length, use the Truncated
Series option and remove from the end of the series the same number of images as specified for the lag. Note – for nega-
tive lags truncate from the end of the series.
Truncated Series
Generates a truncated version of the input series – i.e., a shortened series by removing images either from the end or the
beginning.
Supplemented Series
This is used to add additional data on either the beginning or the end of a series. Note that the series to be added (even if
it contains only one image) must be correctly documented (in its .tsf file) with regards to the start and end dates of the
series, series type, etc.
Aggregate Series
This option aggregates series to a coarser or similar temporal resolution. Depending upon the nature of the series, one or
several conversion options may be available (a small number do not have any conversion options because of peculiarities
in their nature). For cases where a mean is used and values must be taken from more than one input image, a weighted
mean is used.
Detrend (linear)
This option removes the presence of a linear trend in the series. It does this by extracting the residuals from a linear
regression between the series (the dependent variable) and time (the independent variable).
where a is an intercept, b is the trend slope, t is time, Xt is a lag-1 red noise process, ρ is the serial correlation and e is a
white noise error term. The trend preserving prewhitening in ETM uses the procedure described by Wang and Swail
(2001) 15to remove the red noise component yielding a new series (Wt) that can be described by:
15. Wang, X.L., and V.R. Swail, 2001. Changes of extreme wave heights in northern hemisphere oceans and related atmospheric circulation regimes, Jour-
nal of Climate, 14, 2204-2221.
This prewhitened series has the same trend as the original series, but with no serial correlation (Wang and Swail (2001).
Note that this prewhitening method decreases the sample size by one which is recovered using the Prais-Winsten trans-
formation (Kimenta, 2004)16.
Durbin Watson
The Durbin Watson option computes the Durbin Watson statistic of serial correlation over time for each pixel. For index
series, a tabulation is made of the slope and intercept of the best fit linear trend, along with the Durban Watson statistic.
Its value always lies between 0 and 4. A value of 2 indicates no serial autocorrelation. A Durbin-Watson statistic less than
2 indicates evidence of a positive serial correlation and a statistic greater than 2 indicates evidence of a negative serial
autocorrelation. Critical values for the Durbin Watson statistic can be found in standard statistical texts.
Cochrane-Orcutt
The Cochrane-Orcutt transformation17 transforms the dependent variable and each of the independent variables in order
to remove serial correlation. It estimates the first order serial correlation ρ by calculating the correlation between the
residuals and the residuals at lag 1. The transformation of both the dependent and independent variables is calculated as:
As with Prewhiten, the Prais-Winsten transformation (Kmenta, 2004) is used to estimate the initial value of the trans-
formed variables. Note that it is assumed that all variables have been deseasoned such that the expected intercept is 0. In
this case, the transformed intercept is also 0. Therefore the intercept term drops out of the transformation and the user
can use the transformed dependent and independent variables with the Linear Modeling tool to complete the analysis.
17. Cochrane D, Orcutt GH (1949) Application of Least Squares Regression to Relationships Containing Auto-correlated Error Terms. J Amer Statisti-
cal Assoc 44:32-61
Isotropic Costs
Isotropic cost surface modeling is accomplished in IDRISI with the COST module. Given input images of a set of fea-
tures from which cost distances should be calculated and the frictions that affect movement, COST outputs a cost surface
that expresses costs of movement in terms of distance equivalents. Thus, for example, if a cell contains a value of 100, it
simply expresses that the cost of moving from the nearest starting feature (target) to that point is the equivalent of moving
over 100 cells at the base cost. It could equally arise from traveling over 100 cells with a relative friction (i.e., relative to the
friction associated with the base cost) of 1, or 50 cells with frictions of 2, or 1 cell with a relative friction of 100.
Anisotropic Costs
With the COST module, frictions have identical effect in any direction. It doesn't matter how you move through a cell—
its friction will be the same. We can call such a friction isotropic since it is equal in all directions. However, it is not very dif-
ficult to imagine anisotropic frictions—frictional elements that have different effects in different directions. Take, for exam-
ple, the case of slopes. If we imagine the costs of walking (perhaps in calories per hour at normal walking speed), then
slopes will affect that cost differently in different directions. Traveling upslope will cause that friction to act full-force;
traveling perpendicularly across the slope will have no effect at all; and traveling downslope will act as a force that reduces
the cost. Traditional cost analysis cannot accommodate such an effect.
1. Azimuths express directions in degrees, clockwise from north. In IDRISI, it is also permissible to express an azimuth with the value of -1 to indicate
that no direction is defined.
2. To undertake a process similar to RESULTANT, DECOMP is used to decompose all force/friction image pairs acting upon an area into their X and
Y components. These X and Y component images are then added to yield a resulting X and Y pair. The recomposition option of DECOMP is then
used with these to produce a resultant magnitude/direction image pair.
Anisotropic Functions
With force/friction image pairs, one has an indication of both the magnitude and direction with which forces and fric-
tions act. However, what is the interpretation of direction? If a force is said to act at 45° (northeast), does this mean it acts
fully at 45° and not at all at 44°? The answer to this is not easily determined and it ultimately depends upon the applica-
tion. If one takes the earlier example of walking against slopes of varying degrees, the force/friction image describes only
the direction and magnitude of the steepest descending slope. If one faced directly into the slope one would feel the full
force of the friction (i.e., effective friction = stated friction). Facing directly away from the slope (i.e., pointing
downslope), the friction would be transformed into a force to the fullest possible extent (i.e., effective friction = 1/(stated
friction)). Between the two, intermediate values would occur. Moving progressively in a direction farther away from the
maximum friction, the friction would progressively decrease until one reached 90°. At 90°, the effect of the slope would
be neutralized (effective friction = 1). Then as one moves past 90° towards the opposite direction, frictions would
become forces progressively increasing to the extreme at 180°.
This variation in the effective friction/force as a function of direction is here called the anisotropic function. With
VARCOST, the following default function is used:
effective_friction = stated_frictionf
where f = coskα
and k = a user-defined coefficient
and α = difference angle.
The difference angle in this formula measures the angle between the direction being considered and the direction from
which frictions are acting (or equivalently, the direction to which forces are acting). Figure 1 indicates the nature of this func-
tion for various exponents (k) for difference angles from 0 to 90°.
1.0
k=1
function value
k=2
0.5
k=10
k=100
0.0
0 30 60 90
difference angle
Figure 1
You will note in Figure 1 that the exponent k makes the function increasingly direction-specific. At its limit, an extremely
high exponent would have the effect of causing the friction to act fully at 0°, to become a fully acting force at 180°, and to
0 infinity
VARCOST DISPERSE
90 90
-90 -90
0 infinity
Figure 2 Friction Force
Like VARCOST, DISPERSE also allows the entry of a user-defined function. The procedure is identical, allowing for the
reading of a data file containing function values for difference angles from 0-360° in increments of 0.05°. The format for
this file is indicated in the DISPERSE module description in the on-line Help System. Unlike VARCOST, however, the
values of that function represent a multiplier (rather than an exponent) as follows:
effective_friction = stated_friction * f
where f = a user-defined function.
Introduction
In GIS, we often want to combine information from several layers in analyses. If we only know the values of a selection of
points and these sample points do not coincide between the layers, then such analyses would be impossible. Even if the
sample points do coincide, we often want to describe a process for all the locations within a study area, not just for
selected points. In addition, we need full surfaces because many processes modeled in GIS act continuously over a sur-
face, with the value at one location being dependent upon neighboring values.
Any GIS layer, whether raster or vector, that describes all locations in a study area might be called a surface. However, in
surface analysis, we are particularly interested in those surfaces where the attributes are quantitative and vary continuously
over space. A raster Digital Elevation Model (DEM), for instance, is such a surface. Other example surfaces might
describe NDVI, population density, or temperature. In these types of surfaces, each pixel may have a different value than
its neighbors.
A landcover map, however, would not be considered a surface by this definition. The values are qualitative, and they also
do not vary continuously over the map. Another example of an image that does not fit this particular surface definition
would be a population image where the population values are assigned uniformly to census units. In this case, the data are
quantitative, yet they do not vary continuously over space. Indeed, change in values is present only at the borders of the
census units.
No GIS surface layer can match reality at every scale. Thus the term model is often applied to surface images. The use of
this term indicates a distinction between the surface as represented digitally and the actual surface it describes. It also indi-
cates that different models may exist for the same phenomenon. The choice of which model to use depends upon many
things, including the application, accuracy requirements, and availability of data.
It is normally impossible to measure the value of an attribute for every pixel in an image. (An exception is a satellite image,
which measures average reflectance for every pixel.) More often, one needs to fill in the gaps between sample data points
to create a full surface. This process is called interpolation. IDRISI offers several options for interpolation which are dis-
cussed in this chapter. Further technical information about these modules may be found in the on-line Help System.
Surface Interpolation
The choice of interpolation technique depends on what type of surface model you hope to produce and what data are
available. In this section, the techniques available in IDRISI are organized according to input sample data type—points or
lines. A description of the algorithm used and the general characteristics of the techniques are given. For a more theoreti-
cal treatment of the characteristics of surface models produced by particular interpolation techniques, consult the refer-
ences provided at the end of this chapter.
Interpolation techniques may be described as global or local. A global interpolator derives the surface model by consider-
ing all the data points at once. The resulting surface gives a "best fit" for the entire sample data set, but may provide a very
poor fit in particular locations. A local interpolator, on the other hand, calculates new values for unknown pixels by using
the values of known pixels that are nearby. Interpolators may define "nearby" in various ways. Many allow the user to
determine how large an area or how many of the nearest sample data points should be considered in deriving interpolated
values.
Interpolation techniques are also classified as exact or inexact. An exact interpolation technique always retains the original
Potential Model
INTERPOL also offers a second technique called a potential model. It is similar in operation to the distance-weighted
average. The difference is in the function that is employed. The calculation is the same as that described above except that
the sum of weighted attribute values is not divided by the sum of weights. This causes the values at sample points to often
be higher than the original value, especially when sample points are close together. The method is therefore an inexact
interpolator. The surface appears to have spikes at sample points and tends to approach zero away from sample points.
This type of interpolation method is based on the gravity model concept and was developed to model potential interaction
between masses measured at sample points. For example, the amount of interaction (e.g., in terms of commerce) between
the people of two villages is related to the number of people in each village and how close these villages are to each other.
More people who are closer together produce a greater total interaction. The interaction at a location far from any village
would tend to be zero. The potential model method is applied for different purposes than the other methods discussed in
this chapter. It would not be used to develop a surface model from elevation data, for example.
1. Gstat, © Edzer Pebesma, is licensed freeware available from GNU. See the on-line Help System for more details.
2. The line of greatest slope is used to avoid flat lines that result when a line intersects the same isoline on both ends. This is quite common with topo-
graphic maps and would lead to an abundance of flat areas in the interpolated surface.
Introduction
Triangulated Irregular Networks (TINs) are the most commonly-
used structure for modeling continuous surfaces using a vector
data model. They are also important to raster systems because
they may be used to generate raster surface models, such as
DEMs. With triangulation, data points with known attribute val-
ues (e.g., elevation) are used as the vertices (i.e., corner points) of
a generated set of triangles. The result is a triangular tessellation
of the entire area that falls within the outer boundary of the data
points (known as the convex hull). Figure 1 illustrates a triangula-
tion from a set of data points.
There are many different methods of triangulation. The Delaunay
triangulation process is most commonly used in TIN modeling
and is that which is used by IDRISI. A Delaunay triangulation is Figure 1 A set of data points (left) and a
defined by three criteria: 1) a circle passing through the three triangulation of those data points (right).
points of any triangle (i.e., its circumcircle) does not contain any
other data point in its interior, 2) no triangles overlap, and 3) there
are no gaps in the triangulated surface. Figure 2 shows examples
of Delaunay and non-Delaunay triangulations.
A natural result of the Delaunay triangulation process is that the
minimum angle in any triangle is maximized. This property is
used by the IDRISI algorithm in constructing the TIN. The num-
ber of triangles (Nt) that make up a Delaunay TIN is Nt=2(N-1)-
Nh, and the number of edges (Ne) is Ne=3(N-1)-Nh, where N is
the number of data points, and Nh is the number of points in the
convex hull.
IDRISI includes options for using either true point data or vertex
points extracted from isolines1 as input for TIN generation. The
TIN module also offers options to use non-constrained or con-
strained triangulation, to optimize the TIN by removing “tunnel”
and “bridge” edges, and to generate a raster surface from the TIN Figure 2 Delaunay triangulation (left)
by calling the module TINSURF. Modules are also available for and non-Delaunay triangulation (right).
preparing TIN input data. These are all discussed in detail below. The shaded triangle doesn’t meet the
empty circumcircle criterion.
In IDRISI, the TIN file structure consists of a vector line file
(containing the triangle edges) and an associated ASCII TIN file
(containing information indicating which points make up each tri-
angle). File structure details may be found in the on-line Help System.
1. In this chapter, the term isoline refers to any line representing a constant attribute value. Elevation contours are one example of isolines.
Lines
When an isoline file is used as input to TIN, only the vertices2 that make up the lines are used in the triangulation. It may
be useful to examine the density of the vertices in the isoline file prior to generating the TIN. The module GENERAL-
IZATION may be used to extract the vertices of a line file to a vector point file for visualization.
It may be desirable to add points along the lines if points are so far apart they create long straight-line segments that result
in large TIN facets. Point thinning along lines is also sometimes desirable, particularly with isoline data that was digitized
in stream mode. In this case, the number of points in the lines may be much greater than that necessary for the desired
resolution of the TIN, and thus will only serve to slow down the TIN generation process. The module TINPREP per-
forms along-line point addition or thinning. Other line generalization options are also available in the GENERALIZA-
TION module.
If line data are used to create a TIN, both the non-constrained and the constrained triangulation options are available. The
differences between these options are described below. If a raster surface is the desired final output of input isoline data,
the module INTERCON offers an alternative to TIN/TINSURF. However, the latter normally produces a superior
result.
Command Summary
Following is a list and brief description of the modules mentioned in this section.
GENERALIZATION thins or "generalizes" point vector data, extracts the vertices (points) from a line vector to a point
vector file, and generalizes line vector data..
INTERPOL interpolates a raster surface from point data.
IDRISI interfaces to Gstat provide geostatistical tools that can be used to create a raster surface from point data.
TINPREP adds or thins vertices along vector lines.
INTERCON interpolates a raster surface from rasterized isoline data.
TIN creates a TIN from point or line vector data. TINSURF may be automatically called from the TIN dialog if a raster
surface output is desired.
TINSURF creates a raster surface from an extisting TIN.
2. In this context, the term "vertices" refers to all the points that make up a line, including the beginning and ending points.
a b c d
Figure 4 a: Contours at the top of a hill; b: triangulation of highest contour,
with B/T edges identified; c: placement of critical points on B/T edges; d:
re-triangulation.
Tunnel Edges
Bridge Edge
110 m
100 m
500m 600m 600m 500m
a b
Figure 5 a: Contours at a stream; b: contours at a “saddle” feature. Con-
tours are shown with solid lines, constrained triangle edges with dashed
lines. B/T edges are shown in red.
y
2
( x-a) = 2p0(y-b)
y
(a,b)
2
( x-a) = 2p1(y-b)
2
( x-a) = 2p2(y-b)
2 2
( x-a) = 2p1(y-b) ( x-a) = 2p2(y-b)
x
o
(a,b)
2 x
( x-a) = 2p0(y-b) o
Figure 6 Example parabolas and their equations. On the left, p is negative. On the right, p is
positive.
To translate the general parabolic equation to the critical point attribute interpolation problem, we re-label the axes of Fig-
ure 6 from X,Y to S,H where S represents distance from the origin (o) and H represents the attribute value (e.g., elevation)
from the origin (o). (The origin is defined by the location and attribute of the original point as described below.) In the
3. Although the parabolic algorithm is recommended, linear and optimized linear options are also available as critical point interpolation methods in the
module TIN. In the example of the hilltop, a sharp peak would be modeled by the linear method in contrast to the rounded peak of the parabolic
method. The optimized linear method uses a linear interpolation unless slopes in all eight directions (see the discussion of the parabolic interpolation)
are zero, in which case it uses the parabolic.
P0 H
critical
P1 point
Hp
critical point P1 P2
P2
P0
Sp S
a b c
Figure 7 a: Eight-direction search for isoline intersections for one critical point; b: intersection
points for one direction; c: parabola derived from intersection points. Attribute (hp) for the critical
point can be found, given the critical point’s distance (Sp)0 from P.
If three intersections are not found for a particular parabola (e.g., at the edge of a coverage), then it is undefined and the
number of parabolas used to interpolate the attribute value for that critical point will be fewer than eight.
For each defined parabola, the attribute value of any point on the parabola can be found by entering its distance from the
original point into the parabolic equation. The following equation can be used to calculate the attribute of a critical point
for one of its parabolas:5
2 2
( S point – S j )
H = ∑ hi ⋅ ∏ ------------------------------
( Si – Sj )
i=0 j = 0, j ≠ i
Where hi, i = 0, 1, 2 are attribute values of the three intersection points, P0, P1 and P2; Si, Sj, i, j=0,1,2 represent the dis-
tances from the original point to the intersection points, and Spoint represents the distance from the original point to the
critical point. According to the above definitions of the intersection points (Figure 22-6b), we know S0≡0, while S1=P1P0 ,
and S2=P2P0.
4. Any parabola can be defined once three points on it are known. Three equations (one for each point) can be written as below. For each, the distance
(S) from that point to the origin and the attribute (H) are known. The simultaneous equations can then be solved for a, b, and p.
P0 P1 P2 P0 P1 P2 P0 P2
P1
H P2 H P2 H P2
h0+2Δh
P1 P1 P1
h0+Δh
P0 P0 P0
h0 S S S
a b c
h0 h0+Δh h0+Δh
h0 h0 h0+Δh
P0 P0 P1 P2 P0
P1 P2 P1 P2
H H H
h0+2Δh
P2 P1
h0+Δh P1 P2 P1 P2
P0 S P0 P0
h0 S S
d e f
Figure 8 Six examples of parabolic critical point interpolation. The upper part of each exam-
ple illustrates the map view of the isolines (dashed) and the intersection points (P0, P1 and
P2) for one search direction (dotted line) for a critical point. The lower part shows the parab-
ola for that set of intersection points. The attribute for the critical point (which is always
between P1 and P2) can be found by plotting the critical point on the parabolic curve at its
distance (S) from P0.
Invalid Cases
There are two extreme circumstances in which the parabolic interpolation procedure is invalid:
Outputs of TIN
The outputs of the TIN module are a vector line file defining the triangle edges, an ASCII .TIN file containing the topo-
logical information for the triangulation and, if B/T edge removal was used, a point vector file of the critical points that
were added. All these pieces except for the triangle edge vector file, in addition to the original vector data file, are used by
the TINSURF module to create a raster surface from the TIN.
6. The algorithm uses the local contour interval for each critical point, so isoline data with variable contour intervals do not pose a problem.
7. Each pixel center will fall in only one TIN facet, but a single facet may contain several pixel center points.
Further Reading
Lee J., 1991. Comparison of Existing Methods for Building Triangular Irregular Network Models of Terrain From Grid
Digital Elevation Models, International Journal of Geographic Information Systems, 3: 267-285.
Tsai, V. J. D., 1993. Delaunay Triangulations in TIN Creation: an Overview and a Linear-time Algorithm, International Jour-
nal of Geographic Information Systems, 6: 501-512.
Zhu, H., Eastman, J. R., and Schneider, K., 1999. Constrained Delaunay Triangulation and TIN Optimization Using Con-
tour Data, Proceedings of the Thirteenth International Conference on Applied Geologic Remote Sensing, 2: 373-380, Vancouver, British
Columbia, Canada.
Introduction
Geostatistics provides tools for the exploration and statistical characterization of sample point data. It also provides a
number of techniques for interpolating surfaces from such data. Ordinary kriging is the most well-known of these. While
the techniques originated with scientists working in the mining industry, a broader audience has been found in those fields
in which both data values and their locations are considered analytically important.
Several interpolation techniques were introduced in the chapter Surface Interpolation. Geostatistical techniques are dis-
tinct from these in that they provide GIS analysts with the ability to incorporate information about patterns of spatial
continuity into the interpolation model as well as to produce surfaces that include elements of local variation. The meth-
ods allow for a high degree of user flexibility in detecting and defining structures that describe the nature of a data set.
Indeed, a set of structures can be nested, each describing a particular aspect of the data set.
With this flexibility, however, also comes some risk. From the same data set, it is possible to produce many surfaces—all
very different, and all seemingly reasonable representations of reality. The new user is encouraged to enter into geostatis-
tics deliberately and with some caution. An understanding of, and respect for, the underlying assumptions of these tech-
niques is essential if the results are to provide meaningful information to any analysis.
This chapter presents a very brief overview of the geostatistical capabilities offered through IDRISI interfaces to Gstat.1
For more complete and theoretical treatments of geostatistics, consult the references listed at the end of this chapter. The
Tutorial includes an extensive exercise illustrating the use of the geostatistical tools available in IDRISI.
Spatial Continuity
The underlying notion that fuels geostatistical methods is quite simple. For continuously varying phenomena (e.g., eleva-
tion, rainfall), locations that are close together in space are more likely to have similar values than those that are further
apart. This tendency to be most similar to one's nearest neighbors is quantified in geography through measures of spatial
autocorrelation and continuity. In geostatistics, the complement of continuity, variability, is more often the focus of analy-
sis.
The first task in using geostatistical techniques to create surfaces is to describe as completely as possible the nature of the
spatial variability present in the sample data. Spatial variability is assessed in terms of distance and direction. The analysis
is carried out on pairs of sample data points. Every data point is paired with every other data point. Each pair may be char-
acterized by its separation distance (the Euclidean distance between the two points) and its separation direction (the azimuth in
degrees of the direction from one point to the other).2 The sample data point set shown in Figure 1 would produce pairs
characterized as shown in Table 1.
1. IDRISI provides a graphical user interface to Gstat, a program for geostatistical modeling, prediction and simulation written by Edzer J. Pebesma
(Department of Physical Geography, Utrecht University). Gstat is freely available under the GNU General Public License from www.gstat.org. Clark
Labs' modifications of the Gstat code are available from the downloads section of the Clark Labs Web site at www.clarklabs.org.
2. The points in a pair are identified as the from point and the to point. No pair is repeated.
BC 85 m 235
Figure 1
Table 1
The distance measure is typically referred to in units of lags, where the length of a lag (i.e., the lag distance or lag interval)
is set by the user. In specifying a particular lag during the analysis, the user is limiting the pairs under consideration to
those that fall within the range of distances defined by the lag. If the lag were defined as 20 meters, for example, an analy-
sis of data at the third lag would include only those data pairs with separation distances of 40 to 60 meters.
Direction is measured in degrees, clockwise from grid north. As with distance, direction is typically specified as a range
rather than a single azimuth.
The h-scatterplot is used as a visualization technique for exploring the variability in the sample data pairs. In the h-scatter-
plot, the X axis represents the attribute at one point of the pair (the from point) and the Y axis represents that same attri-
bute at the other point of the pair (the to point). The h-scatterplot may be used to plot all of the pairs, but is more often
restricted to a selection of pairs based on a certain lag and/or direction. Figure 2 shows the spatial distribution of 250
rainfall sample points from a 1000 km2 area. These points were paired and data pairs that are within 1 lag (0-1 km) and for
all directions are plotted in the h-scatterplot shown in Figure 3.
Figure 2
Figure 3
Variability
Min
Max
Figure 4
Figure 5
Usually one begins with plotting an omnidirectional semivariogram. From the omnidirectional graph, one may gain
insight into the overall variability of the data. The user then may create several plots, using different directions and lag dis-
tances, to gain a better understanding of the structure of the data set.
3. Note that some geostatistical software plot zero degrees to the right rather than the top of the surface graph.
variance
where an increase in separation distance between pairs no longer increases the
variability between them, i.e., there is no spatial dependence at this and larger dis-
tances. The variance value at which the curve reaches the plateau is called the sill.
The total separation distance from the lowest variance to the sill is known as the
range
range. The range signifies the distance beyond which sample data should not be nugget
considered in the interpolation process when selecting points that define a local distance between pairs
neighborhood.
Figure 6
The nugget refers to the variance at a separation distance of zero, i.e., the Y-inter-
cept of the curve that is fit to the data. In theory, we would expect this to be zero.
However, noise or uncertainty in the sample data may produce variability that is not spatially dependent and this will result
in a non-zero value, or a nugget effect. A nugget structure increases the variability uniformly across the entire graph because
it is not related to distance or direction of separation.
The fourth parameter that defines the structure is the anisotropy of the data set. The transition of spatial continuity may
be equal in all directions, i.e., variation is dependent on the separation distance only. This is known as an isotropic model. A
model fit to any direction is good for all directions. In most environmental data sets, however, variability is not isotropic.
The data used in Figure 2, for example, exhibits a minimum direction of variability in the West-East direction. In any
other direction, variability increases more rapidly at the same separation distance. This type of data requires an anisotropic
model. Anisotropy is described by directional axes of minimum and maximum continuity. To determine the parameters to
be used, the user views directional semivariograms for multiple directions.
In kriging and simulation interpolation processes, structures that describe the pattern of spatial variability represented by
directional semivariograms are used to determine the influence of spatial dependence on neighborhoods of sample points
selected to predict unknown points. The structures influence how their attributes should be weighted when combined to
produce an interpolated value. Semivariograms, however, because they are based on the inherent incompleteness of sam-
ple data, need smoother curves that define the shape of the spatial variability across all separation distances. Using ancil-
lary information and the semivariograms, mathematical functions are combined to delineate a smooth curve of spatial
variability. At this stage, a nugget structure, and sills, ranges, and anisotropies of additional structures are defined for the
smooth curve. The Model Fitting interface offers several mathematical functions that may be used to design a curve for
the spatial variability. Those functions that do not plateau at large separation distances, such as the linear and the power
functions, are termed non-transitional. Those that do reach a plateau, such as the gaussian and exponential functions, are
called transitional functions.
Together, the nugget structure, and the sills, ranges, and anisotropies of additional structures mathematically define a
nested model of spatial variability. This is used when locally deriving weights for the attributes of sample data within the
neighborhood of a location to be interpolated. Using the Spatial Dependence Modeler interface, one unearths a pattern of
spatial variability through the plotting of many variograms until a representative semivariogram can be determined.
Through the Model Fitting interface, the user fits a mathematical curve described by sills, ranges, a nugget, anisotropy and
selected functions to the detected spatial variability. This curve is used to derive the weights applied to locally selected
samples during the interpolation by kriging or conditional simulation.
Semivariograms are statistical measures that assume the input sample data are normally distributed and that local neigh-
borhood means and standard deviations show no trends. Each sample data set must be assessed for conformity to these
assumptions. Transformations of the data, editing of the data set, and the selection of different statistical estimators of
spatial variability are all used to cope with data sets that diverge from the assumptions.
The ability to identify true spatial variability in a data set depends to a great extent on ancillary knowledge of the underly-
ing phenomenon measured. This detection process can also be improved with the inclusion of other attribute data. The
Summary
Geostatistics provides a large collection of tools for exploring and understanding the nature of a data set. Rather than sim-
ply seeking to produce a visually-pleasing interpolated surface, one engages in geostatistical analysis with the foremost
purpose of understanding why various methods produce particular and different results. Interpretation of the information
presented through the various techniques is dependent upon knowledge of other data characteristics and the actual sur-
face. While spatial variability measures themselves are relatively simple descriptive statistics, understanding how they may
be used with data sets that diverge from ideal assumptions requires practice and experience.
ABIDJAN 1987 Cote D’Ivoire (Ivory Coast) Clarke 1880 -124.76 53 466.79
ADINDAN MEAN FOR Ethiopia, Sudan Clarke 1880 -166 -15 204
AMERICAN SAMOA 1962 American Samoa Islands Clarke 1866 -115 118 426
ANNA 1 ASTRO 1965 Cocos Islands Australian National -491 -22 435
ANTIGUA ISLAND ASTRO 1943 Antigua (Leeward Islands) Clarke 1880 -270 13 62
ARC 1950 MEAN FOR Botswana, Lesotho, Malawi, Swa- Clarke 1880 -143 -90 -294
ziland, Zaire, Zambia, Zimbabwe
ARC 1960 MEAN FOR Kenya, Tanzania Clarke 1880 -160 -6 -302
ASTRO DOS 71/4 St. Helena Island International 1924 -320 550 -494
ASTRO TERN ISLAND (FRIG) 1961 Tern Island International 1924 114 -116 -333
ASTRONOMICAL STATION 1952 Marcus Island International 1924 124 -234 -25
AUSTRALIAN GEODETIC 1966 Australia & Tasmania Australian National -133 -48 148
AUSTRALIAN GEODETIC 1984 Australia & Tasmania Australian National -134 -48 149
BELLEVUE (IGN) Efate & Erromango Islands International 1924 -127 -769 472
BUKIT RIMPAH Indonesia (Bangka & Belitung Islands) Bessel 1841 -384 664 -48
CAMP AREA ASTRO Antarctica (McMurdo Camp Area) International 1924 -104 -129 239
CANTON ASTRO 1966 Phoenix Islands International 1924 298 -304 -375
CHATHAM ISLAND ASTRO 1971 New Zealand (Chatham Island) International 1924 175 -38 113
DOS 1968 New Georgia Islands (Gizo Island) International 1924 230 -199 -752
EASTER ISLAND 1967 Easter Island International 1924 211 147 111
EUROPEAN 1950 MEAN FOR Austria, Belgium, Denmark, Fin- International 1924 -87 -98 -121
land, France, West Germany, Gibraltar,
Greece, Italy, Luxembourg, Netherlands, Nor-
way, Portugal, Spain, Sweden, Switzerland
EUROPEAN 1979 MEAN FOR Austria, Finland, Netherlands, International 1924 -86 -98 -119
Norway, Spain, Sweden, Switzerland
Fahud
FORT THOMAS 1955 Nevis, St. Kitts (Leeward Islands) Clarke 1880 -7 215 225
GRACIOSA BASE SW 1948 Azores (Faial, Graciosa, Pico, Sao Jorge, Ter- International 1924 -104 167 -38
ceira)
HONG KONG 1963 Hong Kong International 1924 -156 -271 -189
INDIAN 1960 Vietnam (near 16oN) Everest 1830 198 881 317
ISTS 061 ASTRO 1968 South Georgia Islands International 1924 -794 119 -298
ISTS 073 ASTRO 1969 Diego Garcia International 1924 208 -435 -229
JOHNSTON ISLAND 1961 Johnston Island International 1924 189 -79 -202
KERGUELEN ISLAND 1949 Kerguelen Island International 1924 145 -187 103
KERTAU 1948 West Malaysia & Singapore Everest 1948 -11 851 5
KUSAIE ASTRO 1951 Caroline Islands, Federal States of Micronesia International 1924 647 1777 -1124
MIDWAY ASTRO 1961 Midway Island International 1924 912 -58 1227
MONTSERRAT ISLAND ASTRO 1958 Montserrat (Leeward Islands) Clarke 1880 174 359 365
NAPARIMA BWI Trinidad & Tobago International 1924 -10 375 165
NORTH AMERICAN 1927 MEAN FOR Antigua, Barbados, Barbuda, Clarke 1866 -3 142 183
Caicos Islands, Cuba, Dominican Republic,
Grand Cayman, Jamaica, Turks Islands
Aleutian Islands -2 0 4
Hawaii 1 1 -1
OBSERVATORIO METER- EO 1939 Azores (Corvo & Flores Islands) International 1924 -425 -169 81
OLD HAWAIIAN MEAN FOR Hawaii, Kauai, Maui, Oahu Clarke 1866 61 -285 -181
ORD. SURVEY GREAT BRITIAN 1936 MEAN FOR England, Isle of Man, Scotland, Airy 375 -111 431
Shetland Islands, Wales
PICO DE LAS NIEVES Canary Islands International 1924 -307 -92 127
POINT 58 MEAN FOR Burkina Faso & Niger Clarke 1880 -106 -129 165
PORTO SANTO 1936 Porto Santo, Madeira Islands International 1924 -499 -249 314
PROVISIONAL S. AMERICAN 1956 MEAN FOR Bolivia, Chile, Colombia, Ecuador, International 1924 -288 175 -376
Guyana, Peru, Venezuela
PROVISIONAL S. CHILEAN 1963 Chile (South, Near 53oS) (Hito XVIII) International 1924 16 196 93
SAO BRAZ Azores (Sao Miguel, Santa Maria Islands) International 1924 -203 141 53
SOUTH AMERICAN 1969 MEAN FOR Argentina, Bolivia, Brazil, Chile, South American -57 1 -41
Colombia, Ecuador, Guyana, Paraguay, Peru, 1969
Trinidad & Tobago, Venezuela
TIMBALAI 1948 Brunei, East Malaysia (Sabah, Sarawak) Everest (Sabah & -679 669 -48
Sarawak)
TOKYO MEAN FOR Japan, Okinawa, South Korea Bessel 1841 -148 507 685
TRISTAN ASTRO 1968 Tristan da Cunha International 1924 -632 438 -609
VITI LEVU 1916 Fiji (Viti Levu Island) Clarke 1880 51 391 -36
WAKE ISLAND ASTRO 1952 Wake Atoll International 1924 276 -57 149
Geodetic (Latitude/Longitude)
A single REF file named LATLONG is supplied for geodetic coordinates. This file is based on the WGS84 datum. For
other datums, use the COPY function in IDRISI Explorer to copy this file to another name and then use Metadata in
IDRISI Explorer or Edit along with the data in Appendices 1 and 2 to enter the new datum information.
Michigan SPC27MI1 Current Michigan State Plane Coordinate System Northern Zone L
New Hampshire SPC27NH1 New Hampshire State Plane Coordinate System Zone 1 TM
New Jersey SPC27NJ1 New Jersey State Plane Coordinate System Zone 1 TM
New Mexico SPC27NM1 New Mexico State Plane Coordinate System Eastern Zone TM
New York SPC27NY1 New York State Plane Coordinate System Eastern Zone TM
SPC27NY4 New York State Plane Coordinate System Long Island Zone L
North Carolina SPC27NC1 North Carolina State Plane Coordinate System Zone 1 L
North Dakota SPC27ND1 North Dakota State Plane Coordinate System Northern Zone L
Puerto Rico & Virgin Islands SPC27PR1 Puerto Rico & Virgin Islands State Plane Coordinate System Zone 1 L
SPC27PR2 Puerto Rico & Virgin Islands State Plane Coordinate System Zone 2 (St. Croix) L
Rhode Island SPC27RI1 Rhode Island State Plane Coordinate System Zone 1 TM
South Carolina SPC27SC1 South Carolina State Plane Coordinate System Northern Zone L
South Dakota SPC27SD1 South Dakota State Plane Coordinate System Northern Zone L
West Virginia SPC27WV1 West Virginia State Plane Coordinate System Northern Zone L
Michigan SPC83MI1 current Michigan State Plane Coordinate System Northern Zone L
New Hampshire SPC83NH1 New Hampshire State Plane Coordinate System Zone 1 TM
New Jersey SPC83NJ1 New Jersey State Plane Coordinate System Zone 1 TM
New Mexico SPC83NM1 New Mexico State Plane Coordinate System Eastern Zone TM
New York SPC83NY1 New York State Plane Coordinate System Eastern Zone TM
SPC83NY4 New York State Plane Coordinate System Long Island Zone L
North Carolina SPC83NC1 North Carolina State Plane Coordinate System Zone 1 L
North Dakota SPC83ND1 North Dakota State Plane Coordinate System Northern Zone L
Puerto Rico & Virgin Islands SPC83PR1 Puerto Rico & Virgin Islands State Plane Coordinate System Zone 1 L
SPC83PR2 Puerto Rico & Virgin Islands State Plane Coordinate System Zone 2 (St. Croix) L
Rhode Island SPC83RI1 Rhode Island State Plane Coordinate System Zone 1 TM
South Carolina SPC83SC1 South Carolina State Plane Coordinate System, single zone L
South Dakota SPC83SD1 South Dakota State Plane Coordinate System Northern Zone L
West Virginia SPC83WV1 West Virginia State Plane Coordinate System Northern Zone L
Gauss-Kruger
The Gauss-Kruger reference system is used primarily in the countries of the former Soviet Union and Eastern Bloc. REF
Miscellaneous
Note that IDRISI also includes miscellaneous regional and local reference system files. A list can be found within the
IDRISI Help System, in the Notes section of the PROJECT module.
Arithmetic Operations
In the formulas below, S refers to RMS error. Formulas are presented for each of the arithmetic operations performed by
the OVERLAY and SCALAR modules in IDRISI. In OVERLAY operations, Sx would refer to the RMS error in Map X,
Sy refers to the RMS error in Map Y, and Sz refers to the RMS error in the final map produced, Map Z. In SCALAR oper-
ations, K refers to a user-defined constant. Often, error is computed as a uniform value for the entire resulting map. How-
ever, in some cases, the formula depends upon the values in corresponding cells of the input maps. These are referred to
as X and Y. In these instances, the error would vary over the face of the map and would thus need to be computed sepa-
rately for each cell. Note that these formulas assume that the input maps are uncorrelated with each other.
Overlay Add / Subtract
2 2
(e.g., Z=X+Y or Z=X-Y) S z = Sx + Sy
Logical Operations
For Boolean operations, logical errors may be expressed by the proportion of cells that are expected to be in error (e) in
the category being overlaid. Since a Boolean overlay requires two input maps, the error of the output map will be a func-
tion of the errors in the two input maps and the logic operation performed as follows:
Logical AND
ez=ex+ (1-ex)*ey or equivalently ez=ex+ ey - (ex*ey)
Logical OR
ez=ex*ey