Geostatistics Without Tears
Geostatistics Without Tears
..
..
..
.. Gregg Babish
GIS Specialist
Edition 2006.03
July 5, 2006
Environmen
t Canada
Geostatistics Without
Tears
. . . . . . . . . . . . . . . . . .
Gregg Babish
GIS Specialist
Environment Canada
Regina, Saskatchewan
INTRODUCTION ............................................................................................................9
INTERPOLATION TECHNIQUES...............................................................................9
SPATIAL INTERPOLATION ..............................................................................................10
Deterministic/Stochastic ...........................................................................................10
Global/Local.............................................................................................................10
Exact/Approximate ...................................................................................................1 1
Gradual/Abrupt ........................................................................................................11
CLIMATE DATA .............................................................................................................11
ANUSPLIN................................................................................................................12
Centro Internacional de Agricultura Tropical .........................................................13
Climatologically Aided Interpolation .......................................................................13
Daymet......................................................................................................................14
Gradient plus Inverse Distance Square ....................................................................14
Optimal Interpolation ...............................................................................................15
Parameter-evaluation Regressions on Independent Slopes Model ..........................15
Thin-Plate Spline ......................................................................................................16
DIGITAL ELEVATION MODELING ...................................................................................16
FOURIER SERIES ............................................................................................................1 8
GEOSTATISTICS .............................................................................................................18
INVERSE DISTANCE WEIGHTED .....................................................................................1 9
NEAREST NEIGHBOUR ...................................................................................................20
SPLINES .........................................................................................................................20
TREND SURFACE ANALYSIS...........................................................................................2 1
TRIANGULATION............................................................................................................21
HISTORY OF KRIGING ..............................................................................................23
DATA ANALYSIS..........................................................................................................27
CENSORED DATA...........................................................................................................37
CO-LOCATED SAMPLES..................................................................................................37
DECLUSTERING..............................................................................................................37
EXTREME VALUES.........................................................................................................38
SKEWED DATA WITH ZEROS ..........................................................................................38
SPATIAL OUTLIERS ........................................................................................................39
STRATIFICATION (DATA SET SUBDIVISION) ...................................................................39
TIME ..............................................................................................................................40
TRANSFORMATION ........................................................................................................40
Arcsine Transform ....................................................................................................40
Box-Cox Transform ..................................................................................................40
Lognormal Transform...............................................................................................41
Normal Score Transform..........................................................................................41
Scale Transform........................................................................................................42
Square Root Transform ............................................................................................42
OFFSETS AND BACK-TRANSFORMS ................................................................................42
Offset.........................................................................................................................42
Back-transformation.................................................................................................42
TRENDS .........................................................................................................................43
APPLIED GEOSTATISTICS .......................................................................................45
VARIOGRAMS..............................................................................................................45
CALCULATING THE VARIOGRAM....................................................................................46
CONCLUSION...............................................................................................................8 7
PREFACE
Geostatistics (also known as kriging) was developed for the mining industry
during the 1960s and 1970s to estimate changes in ore grade. The principles of
geostatistics are now applied to many applications that require statistically based
interpolation techniques. Geostatistics provides a data value estimate for
locations that cannot be sampled directly by examining data taken at locations
that can be sampled. If a model of spatial correlation can be established for the
sampled data then this model is used to interpolate data values for those locations
that cannot be sampled directly. The effectiveness of geostatistics depends upon
how well the observation points represent the phenomenon under study as well as
the appropriate selection of the model parameters.
The intent of this paper is to provide the reader with an introduction and
understanding of surface interpolation techniques and geostatistics (kriging).
Hopefully no tears will be shed.
Data to create these surfaces are usually collected through field work. As a rule,
data collection can only be conducted at a restricted number of point locations
due to limited resources and high costs. In order to generate a continuous surface
some type of interpolation method must be used to estimate data values for those
locations where no samples or measurements were taken.
INTERPOLATION TECHNIQUES
Surface interpolation is the process of estimating a data value for unsampled
sites within an area of existing data observations. The best surface interpolation
method to use will depend on the data, the study objectives, and the ease of
creating a surface versus the accuracy of the surface created.
• the sample measurements are accurate and represent the true value at that
location
Deterministic/Stochastic
Deterministic techniques are based on surrounding measurements (mathematical
functions) to calculate the surface. These techniques are based on the measured
values of a parameter at samples near the unmeasured location for which a
prediction is made. For deterministic interpolation there is only one possible
predicted value for a location. Thiessen polygons, IDW and spline interpolation
are deterministic techniques
Global/Local
A global interpolator derives a surface model using all of the available data to
provide predictions for the whole area of interest by applying a single function
across the entire region. The resulting surface gives a best fit for the entire
sample data but may provide a very poor fit at some locations. There are a
number of disadvantages to a global fit procedure. The most obvious is that
global interpolation methods don’t capture short-range variation. A change in one
input value can affect the entire map. Global algorithms tend to produce
smoother surfaces with less abrupt changes.
Gradual/Abrupt
Gradual interpolation produces a surface with gradual changes by applying the
same rules over the entire data source. Gradual interpolation is appropriate for
interpolating data with low local variability.
Climate Data
Collection of meteorological data requires specialized instrumentation collecting
data over many years and is subject to siting, data collection errors and biases.
Climate data is available only at a relatively small number of observed points.
Given a set of meteorological data, researchers are confronted with a variety of
interpolation methods to estimate meteorological variables at unsampled
locations. Depending on the spatial attributes of the data the accuracy of the
interpolated surface can vary widely among different methods (MacEachren and
Davidson, 1987; and Rhind, 1975).
Climate interpolation models provide few estimates of error because these estimates
rely on the same assumptions used in the interpolation process itself, and are
therefore, not independent or reliable (Daly, 2006). Experts are the most useful
source of verification because they can integrate information from disparate sources
to give justifiable feedback on how well the model results reproduce their
knowledge from a variety of perspectives. This feedback can be in the form of
evaluation of the spatial patterns and magnitudes of the mapped climate values, as
well as insight into station data quality issues (Daly and Johnson, 1999).
Kriging has been used for interpolation of precipitation data (Chua and Bras,
1982; Dingman et al., 1988; Phillips et al., 1992). Meteorologists sometimes
consider that a potential drawback of kriging is that kriging implicitly relies on
the data to directly represent the spatial variability of the actual precipitation
field. If the data are not representative (as is often the case in complex terrain,
such as mountainous regions), the accuracy of the resulting interpolated field will
be in question (Daly et al. 1993).
Ordinary kriging has been applied for developing contour maps of design storm
depth using intensity-duration-frequency (IDF) data (Cheng, et al., 2003). The
design storm, a crucial element in urban design and hydrological modeling, is a
hypothetical storm of specific storm duration, and recurrence interval. Variogram
parameters, the sill and range are functions of the recurrence interval and the
storm duration. The sill accounts for the time non-stationarity of the rainfall field,
and design storms with higher total rainfall depths have higher sill values.
ANUSPLIN
ANUSPLIN software developed in the late 1980’s by Michael Hutchinson
(Hutchinson, 1991 and 1994), Australian National University (ANU), Centre for
Resource and Environmental Studies (CRES), Canberra is based on the srcinal
thin-plate surface fitting technique (Wahba, 1979 and 1990) for mapping climate
variables, especially for the interpolation of rainfall amounts (Hutchinson, 1995;
1998a; 1998b). ANUSPLIN fits thin-plate splines (usually second- or third-order
polynomials) through station data in three dimensions: latitude, longitude, and
elevation (Hutchinson, 1995). Booth and Jones (1996) conclude that ANUSPLIN
provides a powerful
used to develop set ofconsistent
globally programs30”
for climate
climaticsurfaces
analysis.covering
ANUSPLINmost has been
areas on
the Earth’s surface (Hijmans et al. 2004 a,b).
uses a simple
between interpolation
the station and thealgorithm based
interpolated on of
point thethe
inverse square
nearest five of the distance
stations (Jones
et al. 1990, Jones, 1995).
The major difference between CIAT and ANUSPLIN is that the CIAT method
uses a standard lapse rate applied over the whole dataset whereas ANUSPLIN
uses a 3-dimensional spline algorithm to determine a local lapse rate from the
data. The CIAT method introduces error when the local lapse rate deviates from
the standard lapse rate function. ANUSPLIN suffers when there are erroneous
data or insufficient data range in the local area, resulting in spurious correction
for elevation (Booth and Jones, 1996). CIAT has the advantage of speed and ease
of use for large data sets where computational capacity is limited. The influence
of a bad data point can be significant and can cause significant circling in the
resultant surface. Because CIAT uses only five data points, it relies less on the
underlying DEM than other methods (Booth and Jones, 1996).
Daymet
Daymet (http://www.daymet.org/) is similar to PRISM in that both methods use
local regression techniques, but Daymet and PRISM are different in their
formulation. Daymet was developed at the University of Montana, Numerical
Terradynamic Simulation Group (NTSG) to fulfill the need for fine resolution, daily
meteorological and climatological data necessary for plant growth model inputs.
Using a digital elevation model and daily observations of minimum/maximum
temperatures and precipitation from ground-based meteorological stations, an 18
year daily data set (1980-1997) of temperature, precipitation, humidity and radiation
have been produced as a continuous surface at a 1 km resolution for the United
States (http://www.daymet.org/dataSelection.jsp
).
Daymet develops local linear regressions between climate and elevation for each
grid cell on a digital elevation model using data from surrounding stations. Each
station is weighted in the regression function by its distance from the target grid
cell. This method takes into account the elevation variation of climate and a
simple station distance weighting algorithm. Daymet does not have the ability to
simulate nonmonotonic relationships between climate and elevation, such as
temperature inversions, and does not explicitly account for terrain-induced
climatic transitions or coastal effects (Daly, 2006).
Kagan, 1974;
al., 1982; Bretherton
Hiller and Käse,et1983;
al., 1976; Bretherton
Swanson, et. al., and Williams,
2001; 1980; Sarmiento et
AWI, 2006.
PRISM develops local climate-elevation regression functions for each DEM grid
cell, but calculates station weights on the basis of an extensive spatial climate
knowledge base that assesses each station’s physiographic similarity to the target
grid cell (Daly et al., 2002, 2003). The knowledge base and resulting station
weighting functions currently account for spatial variations in climate caused by
elevation, terrain orientation, effectiveness of terrain as a barrier to flow, coastal
proximity, moisture availability, a two-layer atmosphere (to handle inversions),
and topographic position (valley, midslope, ridge).
While PRISM accounts for more spatial climate factors than other methods, it also
requires more effort, expertise and supporting data to take advantage of its full
capability (Daly 2006). The effects of other variables, such as slope, aspect, coastal
proximity, the influence of the boundary layer, etc. are controlled by the use of
special algorithms, parameters and weights. Some of these values are set as defaults
based on the general knowledge of physiographic and atmospheric
Thin-Plate Spline
The thin-plate spline (TPS) method refers to a physical analogy involving the
bending of a thin sheet of metal constrained not to move at the sample points and
is free from any external force. The resulting shape of the plate represents the
natural bend of the metal to accommodate all of the fixed sample points. The use
of thin-plate splines has been found useful for a number of applications including
DEM construction, fingerprint modeling, image manipulation, geophysical
applications, medical research, and rubber sheeting. Thin-plate splines have
found application in meteorology (Hutchinson and Bischof, 1983; Hutchinson et
al., 1984; Kalma et al., 1988; Hutchinson, 1990; Mitchell, 1991; Zheng and
Basher, 1995; Chapman et al., 2005).
The key features of the thin-plate smoothing spline analysis are robustness,
simplicity and the advantage of being able to map sparse and noisy data. One of
the main shortcomings of the thin-plate spline method has been the limited range
of practical options for user control of the fit. The thin-plate spline method can be
problematic if sample points are sparse. Small changes in sample position could
produce global changes in the warping transformation (Glasbey and Mardia,
1998). The thin-plate spline is equivalent to kriging with a specific spatial
covariance structure (Kent and Mardia, 1994).
There areNetwork
Irregular two forms of Digital
(TIN) Elevation
with their Model (DEM):
own advantages Grid and Triangular
and disadvantages (Cadell,
2002). A number of algorithms and computer programs have been developed to
create a DEM.
ESRI’s ARCTIN method makes a Triangular Irregular Network (TIN) and then
converts the TIN to a DEM. ANUDEM and ESRIs ArcInfo TOPOGRID are two
common spline-based methods of obtaining a reasonable interpolation between
Triangular models have been used in terrain modeling since the 1970s, and are
sometimes used to create Digital Terrain Models (DTMs). Commercial systems
using TIN began to appear in the 1980's as contouring packages. Due to
limitations of computers and the complexity of TIN data structures, gridded
models have
effectively overshadowed
divided triangular
into triangles models.
with plane Certain
facets. Thistypes of terrain true
is particularly are very
with
fluvially eroded landscapes. However, other landscapes, such as glaciated ones,
are not well represented by flat triangles. Triangles work best in areas with sharp
breaks in slope where TIN edges can be aligned with breaks such as mountainous
terrain, but triangular models aren’t as suitable for representing prairie
landscapes (Goodchild, 1999).
Geostatistics
Interpolation methods such as spline, IDW, triangulation, trend surfaces and
Fourier series are based on mathematical models (deterministic). These methods
assume that the sampled data has no errors, which is often an incorrect
assumption since errors can be minimized but not eliminated. The best
representative map created with these techniques may look like they model the
spatial process but this model may not provide a correct representation of the
spatial process (Krivoruchko, 1998).
ii. interpolate values at those locations which have not been sampled
based on the degree of spatial pattern (spatial autocorrelation).
Kriging is the process of prediction of the unknown values
The results from kriging are generally of higher quality and have a more realistic
look compared to techniques such as IDW and triangulation. The “bull’s eye”
effect is avoided; a measure of error or uncertainty is provided; confidence
placed in the estimates can be calculated; extreme weighting of values caused by
irregular distribution of sample points is reduced; redundant data (clustering) and
anisotropies is accounted for; and kriging can be used to determine where more
Shepard’s Method (Shepard, 1968) is a variation on the inverse power with two
different weighting functions (an inverse distance weighted least squares method,
and the inverse distance to a power interpolator). Shepard’s Method, while
similar to IDW, eliminates or reduces “bull’s eye” patterns, especially when a
smoothing factor is used. The Shepard’s Method can extrapolate values beyond
the data’s range.
For climate studies, both the nearest neighbor and IDW can be significantly
improved by including topographical effects. For example, an environmental
lapse rate (–6.5 °C/km) can be applied to temperature values at the stations
before the interpolation is performed. Then, after the interpolation, the grid
values can be transformed back by re-applying the lapse rate. This normalizing
technique standardizes the temperature data with respect to elevation, resulting in
a more accurate interpolation of ‘like’ data (Tait and Zheng, 2005).
Splines
Splines were srcinally flexible pieces of wood or metal used by draftsmen to fit
curvilinearly smooth shapes when the mathematics and/or tools were not
available to create the shapes directly (airplane wings/fuselages, boat hulls,
railway tracks). Spline interpolation was later developed (early 1960’s) to help in
the computer aided manufacturing of car bodies, for computer graphics and to fit
isopleths (contours) to continuous data.
Splines are polynomials (usually cubics) which are fitted to coordinate (X,Y)
and surface values (Z) and forced to be smooth and continuous at the data joining
points called 'knots' (Smith, et al., 2004). To visualize the spline in action,
imagine a rubber sheet being used to interpolate the surface. The rubber sheet
must be bent and stretched so that it passes through all of the measured values.
The rubber sheet can either be forced to form nice curves (regularized spline
interpolation) or the edges of the rubber sheet can be tightly controlled (tension
spline). The spline curve goes through data points exactly and assumes a curved
shape elsewhere. A spline interpolation demo can be seen at
http://www.math.ucla.edu/~baker/java/hoefer/Spline.htm.
Splines are easy to compute mathematically and are useful in fitting a smooth
line or surface to irregularly spaced data points while still retaining localized
features. Spline functions produce good results for gently varying surfaces, such
as elevation, water table heights or pollution concentrations.
The spline function is inappropriate if there are large changes in the surface
within short horizontal distances and/or when the sample data are prone to error
or uncertainty. With this type of data the spline method can wildly overshoot
estimated values and closely spaced data points can develop embolisms.
Extrapolating beyond the edges of the data domain often yields erratic results
because there is no information beyond the data to constrain the extrapolation
and splines will grow to large values (positive or negative). Splines tend to
emphasize trend rather than anomalies (outliers).
A trend surface can be used to filter out large scale spatial trends in order to focus
on the smaller scale variation (residuals) (Goodman, 1973, 1999, and
Klinkenberg, 2002). Variability in spatial data is generally the product of two
effects: the first the result of broad-scale regional changes in the value of the
phenomena, the second the result of smaller-scale ‘local’ variations (Goodman,
1973, 1999; Davis, 1973). Trend surfaces typically fit one of three
mathematically defined ideal surface models: linear (a constant dip in some
single direction); quadratic (a bowl or dome shape, anticline or syncline); or
cubic (saddle point, perhaps large scale folding).
Triangulation
When a set of coordinate (X,Y) and surface value (Z) data points are connected
in a triangular network of varying size the resulting lattice work is irregular and
is called a Triangular Irregular Network (TIN) or irregular triangular mesh. The
TIN partitions a surface into a set of contiguous, non-overlapping, triangles. The
TIN is a vector data structure and is made from a set of points called mass points.
Triangular structures were invented in the early 1970’s because computers were
limited and grids weren’t as inefficient (Thurston, 2003).
The main disadvantage of triangulation is that the generated surfaces are not
smooth and may have a jagged appearance. The size and shape of the areas
depend on the sample layout, which can lead to irregularly shaped polygons,
especially on the edges. Sparse data sets result in distinct triangular facets.
The Voronoi diagram, sometimes named Dirichlet tessellation breaks an area into
pieces or tiles. Lines that bisect the lines between a center point and its
surrounding points define a Voronoi polygon. The bisecting lines and the
connection lines are perpendicular to each other. Voronoi polygons are created so
that every location within a polygon is closer to the sample point in that polygon
than any other sample point. Interest in the Voronoi diagram srcinates from the
climatologist A.H. Thiessen’s use of Thiessen polygons to define regions that
surround unevenly distributed weather stations.
Given that concentrations of high-grade ore are easier and more profitable to
mine, while regions of low-grade ore should be ignored, the estimation of
recoverable ore is very important to the mining industry because local variability
can make or break a mining venture. The first steps to resolve this problem were
taken in the early 1950’s in South Africa with the work of the mining engineer
Danie Krige (Krige, 1951; Krige et al. 1989) and the statistician Herbert Sichel
(1915-1995) working on the Witwatersrand goldfields.
Krige and Sichel studied the best ways of predicting the value of a block to be
mined using only samples from the faces already worked. Krige assumed that there
was a relationship between neighboring blocks of ground in that blocks close to the
stope (area of a mine from which ore is or has been extracted) would have a
different and stronger relationship than those being estimated that are further away.
Which is the basis for the assumption: “two blocks of ground will have a
covariance which depends on the distance between them and on their relative
locations.” Krige
necessary, To simplify
devised athe calculations
template andbe cut
that could laid down the
over the amount
mined areas.of work
The word “krigeage” appears in Matheron's srcinal work (in French) and the
English translation was kriging. Matheron’s first English paper appeared in 1963
(Matheron, 1963a). The two most common ways to pronounce kriging in English
are “kree-ging” or “kree-jing”. Danie Krige pronounced kriging with a hard 'g'
(as in 'grand' or 'organize'), and not with the guttural Afrikaans 'g' (which is how
he pronounced his name).
From 1964 to 1968, Georges Matheron turned his attention to the mathematical
characterization of geometric shapes and in collaboration with Jean Serra, created
the discipline of "Mathematical Morphology" which has since become an
essential branch of the field of image analysis (Rivoirard, 2000). In 1986 the
Centre de Morphologie Mathématique became two programs, one on
mathematical morphology and one on geostatistics (Centre de Geostatistique,
2000). Two of Matheron’s first students (Journel and David) would start new
centers of teaching and research in the USA and Canada.
In 1993 ISATIS was released which offered in one package all the techniques
previously available in MAGMA. ISATIS (http://www.geovariances.com/) is the
result of 40 years of experience in industrial applications and applied research in
geostatistics. Today, ISATIS is widely used the world over by more than 250
private oil & gas companies, consultant teams, mining corporations and
environmental agencies.
The first step is to verify three data features: dependency, stationarity and
distribution (Krivoruchko, 2002a). If data are independent it makes little sense to
analyze them geostatistically. Autocorrelation assumes stationarity, meaning that the
spatial structure of the variable is consistent over the entire domain of the data set. If
the data are not stationary, they need to be made so, usually by data detrending and
data transformation. Geostatistics works best when input data are Gaussian (normal).
If not, the data need to be made to be close to Gaussian distribution.
SUMMARY DATAANALYSIS
Summary Statistics
Summary statistics includes four groups:
Frequency Distribution
The histogram is the graphical version of a table
Histogram which shows what proportion of cases fall into
each of several or many specified categories. The
y
c
100 100.00% frequency distribution summarizes discrete data
n Frequency
e
u
by counting the number of observations falling
q 50 50.00% Cumulative %
e
r
into each category. The number associated with
F 0 .00% each category is called the frequency and the
20 25 30 35 40 45 50
collection of frequencies over all categories gives
Value the frequency distribution of that variable.
Cumulative Distribution
The cumulative distribution determines the number of observations that lie above
(or below) a particular value in a data set. The normal distribution produces an s-
shaped graph when the frequency value is plotted against cumulative proportion.
Skewness
Skewness measures the degree to which a distribution is asymmetric. Normal
distributions (symmetrical distribution) will have a skewness value near zero. If a
distribution's mean and median do not coincide, the distribution is skewed.
Positive Skew
If the distribution's mean is greater than (to the right of) the median, the
distribution is said to be skewed to the right. The tail of the distribution points to
the right, and the mound of data are on the left. Data found in pollution studies or
geological applications tend to have a lot of values clustered below the mean
value and a long tail into the high values.
Negative Skew
The opposite is referred to as a negative skew. Data found in geological
applications in limestone and sedimentary iron ores tend to have a lot of high
values with a long tail into the lower values.
Mesokurtic
A normal random variable has a kurtosis of 3 irrespective of its mean or standard
deviation. Kurtosis should be near 3 for a normal distribution (normal tails).
Leptokurtic
Kurtosis is greater than 3 for distributions that have thicker than normal tails.
Leptokurtosis is associated with distributions that are simultaneously peaked and
have fat tails. The graph on the right is leptokurtic.
Platykurtic
Kurtosis is less than 3 for distributions with thinner than normal tails.
Platykurtosis is associated with distributions that are simultaneously less peaked
and have thinner tails. Platykurtic distributions are said to have shoulders. The
graph on the left is platykurtic.
Short Tails
The normal probability plot based on data that has short tails shows a
non-linear pattern. In this example, the normal distribution is not a
good model for this data which shows up in two ways. First, the
middle of the data shows an s-like pattern. This is common for both
short and long tails. Second, the first few and the last few points
show a marked departure from the reference fitted line. In comparing
this plot to the long tail example the important difference is the
direction of the departure from the fitted line for the first few and last
few points. For short tails, the first few points show increasing
departure from the fitted line above the line and the last few points show
increasing departure from the fitted line below the line.
Long Tails
The normal probability plot based on data that has long tails shows a
reasonably linear pattern in the centre of the data. However the tails,
particularly the lower tail, show departures from the fitted line. For
data with long tails relative to the normal distribution, the non-
linearity of the normal probability plot can show up in two ways.
First, the middle of the data may show an s-like pattern. This is
common for both short and long tails. In this particular case, the s-
pattern in the middle is fairly mild. Second, the first few and the last
few points show marked departure from the reference fitted line. This
is most noticeable for the first few data points. In comparing this plot to the
short-tail example, the important difference is the direction of the departure from
the fitted line for the first few and the last few points. For long tails, the first few
points show increasing departure from the fitted line below the line and the last
few points show increasing departure from the fitted line above the line.
Quantile-Quantile Plot
A Quantile-Quantile (QQ) plot is a plot of the quantiles of one data set
against the quantiles of a second data set. The QQ plot can determine if
two data sets come from populations with a common distribution. The
QQ plot has a number of advantages: sample sizes do not need to be
equal; many distributional aspects can be simultaneously tested (shifts in
location, shifts in scale, and changes in symmetry); the presence of
outliers can be detected. For example, if the two data sets came from
populations whose distributions differ only by a shift in location, the
points should lie along a straight line that is displaced either up or down
from the 45 degree reference line.
Measures of Location
Mean
The arithmetic mean (commonly called the average) is the sum of all data values
divided by the number of data values. The mean is a good measure of central
tendency for roughly symmetric distributions but can be misleading in skewed
distributions since the mean can be greatly influence by extreme data values.
Median
The median is the middle of a distribution. To calculate the median the data must
be ranked (sorted in ascending order). The median is the number in the middle.
The median value corresponds to a cumulative percentage of 50% (i.e., 50% of
the values are below the median and 50% of the values are above the median).
If the mean and the median are approximately the same value this is evidence
that the data may be normally distributed. A frequency distribution is said to be
skewed when its mean and median are different.
Mode
The mode is the most common (frequent value). The easiest way to look at
modes is on the histogram. There may be no mode if no value appears more than
any other. There may also be two modes (bimodal), three modes (trimodal), or
Quantiles
A quantile is the variable value that corresponds to a fixed cumulative frequency.
Any quantile can be read from the cumulative frequency plot:
A quantile is the percent of points below a given value. That is, the 30% (0.3)
quantile is the point at which 30 percent of the data fall below and 70 percent fall
above that value. Probability intervals can also be read from the cumulative
frequency plot, i.e. the 90% probability interval.
Measures of Spread
Measures of spread describe the variability of a data set. This group includes the
range, variance, standard deviation, interquartile range and the coefficient of
variation.
Range
The range is the largest data value minus the smallest data value. As a general
rule of thumb, the range should be somewhere between 4 and 6 standard
deviations depending on how many samples are present.
Variance
One of the most useful statistics to measure the variation of the sample values is the
variance, which is found by considering the deviation of each sample from the
average value. The variance is calculated by squaring the deviations, then adding
them up and taking the average. Variance is a mean squared deviation and its units
are squared rather than in the srcinal sample units.
Standard Deviation
To get a descriptive statistic the square root of the variance is required to obtain the
standard deviation (root mean squared deviation). The standard deviation can be
interpreted as the difference between any data set member and the average of the
data set. The mean value fixes the centre of the distribution, and the standard
deviation scales the horizontal axis. In a normal distribution, 68% of the
population values lie within one standard deviation of the mean value. Just over
95% of the values lie within two standard deviations. Approximately 99.7% of
the observations are within 3 standard deviation of the mean.
Interquartile Range
The interquartile range is used to describe the spread of a data set. The
interquartile range is the difference between the first quartile (a number for which
25% of
75% of the
the data
data is
is less
less than
that that
that number) anda third
number) of set ofquartile (a interquartile
data. The number for which
range
is the range of the central 50% of the sample values.
Coefficient of Variation
The coefficient of variation is calculated by dividing the mean by the standard
deviation. The general rule of the coefficient of variation says, “if the standard
deviation is relatively large compared to the arithmetic average, the average can’t
be used to make decisions.” As a rule of thumb, when the standard deviation is
smaller than the mean, the data are relatively closely clustered and mean is
considered a reasonably good representation of the full data set. By contrast, if the
standard deviation is greater than the mean, then the data are relatively widely
dispersed and the mean is a rather poor representation of the full data set.
Coefficient of Correlation
Pearson’s coefficient of correlation (r) is the most common measure of
correlation or predictability between two variables. Pearson’s r ranges in value
from -1 to 1. The lager the r (ignoring the sign) the stronger the association and
the more accurate it is to predict one variable from knowledge of the other
variable. If r is 0 there is no correlation (no relationship) between the two
variables.
The sign of the correlation implies the “direction” of the association. A positive
correlation means that relatively high scores on one variable are paired with
relatively high scores on the other variable, and low scores are paired with
relatively low scores. A negative correlation means that relatively high scores on
one variable are paired with relatively low score on the other variable.
A zero correlation does not necessarily mean there is no relationship between two
variables – it means there is no linear relationship. For this reason, scatter plots
are important supplements to statistical measures of association by providing a
graphical picture of where variables may be associated with each other.
Correlation measures the extent of association, but association does not imply
causation. It can happen that two variables are highly correlated, not because one
is related to the other, but because they are both strongly related to a third
variable.
Coefficient of Determination
The coefficient of determination (r2, r2 or R2 – the square of the coefficient of
correlation) is the ratio of the explained variation to the total variation and
provides the proportion of the variation (fluctuation) of one variable that is
predictable from the other variable. The coefficient of determination is a measure
that provides the mechanism to determine how certain one can be in making
predictions from a certain model or graph.
The better the regression line fits the data the larger the r2. If the regression line
passes exactly through every point on the scatter plot, it would be able to explain
all of the variation. The further the line is away from the points the less it is able
to explain. A correlation coefficient of one (perfect correlation) is rarely, if ever,
achieved. The coefficient of determination should be at least 0.6 and preferably
between 0.8 and 1.0 for a reliable surface model.
The standard error of estimate (se, SE, Se or SEE) is a measure of the accuracy of
predictions. The top regression example show the points are closer to the
regression line then they are in the bottom example. The predictions in the top
graph are more accurate than the bottom graph. A regression line is the line that
minimizes the sum of square deviations of prediction (also called the sum of
squares error) . For example, if the standard error of the estimate is 775 then the
prediction will be within +/- 775 units. The standard error of estimate is a
measure of the variation (scatter) of the points about the regression line and is
analogous to the standard deviation. However, the standard deviation measures
the scatter about the arithmetical mean of a sample whereas the standard error of
estimate measures the scatter of the dependent variable about the regression line.
If the standard error of estimate is equal to the standard deviation of the
dependent variable, then the correlation coefficient is equal to zero. The standard
error of estimate provides a measure of the closeness of the point to the curve
relation; it does not show the degree of correlation.
Censored Data
Censored data is data having a <,> or estimated qualifier, as is typical with water
quality data. An accepted practice is to convert not detected (ND) values to one-
half of the detection limit; to convert less than detection (<DL) values to the
detection limit; and to convert estimated (E) values to the value estimated.
Co-located Samples
If the samples represent replicates of the same sample collected at the same time, an
accepted practice is to average the samples and only include the average in the data
set. If the sample represents a different element or a different time of collection an
accepted practice is to change the location coordinates slightly. A problem arises
with exact kriging when multiple values are measured at locations. A method of
dealing with multiple samples was documented in Matheron's srcinal works. With
multiple samples at certain locations, the diagonal entry in the kriging system can
be modified. Instead of γ(0) = 0 use γ(0) = (n-1)/n times nugget effect. This tells the
kriging system you have replicates and it will adjust weights and optimal estimator
accordingly (Isobel Clark, 2001b)
Declustering
In many situations the spatial location of data collection sites are not randomly or
regularly spaced. Data may have been sampled with a higher density in some
places that in others. The mean value will be over-estimated if the samples are
spatially clustered in the high value areas. A declustering method must be
implemented to adjust for preferential sampling. One solution to preferential
sampling is to weight the data, with data in densely sampled areas receiving less
weight and data in sparsely sampled areas receiving greater weight. This can be
accomplished in a couple of ways.
One method is cell declustering. A grid of cells is overlaid the data locations, and
the weight attached to each data point is inversely proportional to the number of
data points in its cell. Some researchers suggest that the cell size can be chosen
corresponding to the minimum weighted mean if the data have been
preferentially sampled in areas of high values. Conversely pick the cell size
corresponding to the maximum weighted mean if the data have been
Another scheme uses a polygon method that defines a polygon around each
spatial data location, such that all locations within that polygon are closer to the
data location than any other data location. The idea is to weight each data point
in proportion to the area that it represents. The problem with this method is that it
is difficult to define weights toward the edge. The edge points can often receive
large weights unless a border encloses the data.
Extreme Values
The first step in dealing with extreme values is to establish why the data contain
extreme highs. Is there a high degree of imprecision in measuring values (the
sample observations are actually inaccurate)? Is the distribution of data values
skewed? Are there two (or more) populations, only one of which gives the high
values? Are there other reasons?
Once the reason for extreme values has been determined it will be possible to
deal with this problem more objectively. You may be able to:
• stratify the data set by grouping extreme values together and treating
them as a separate statistical population. Stratified kriging can be done
by interpolating for different sub-regions separately and then the results
of the interpolation can be combined as a single map.
• extreme values can be accepted and their effect on the distribution can be
mathematically moderated. There are two sets of mathematical methods
to do this. The first is to transform the data set before applying statistical
measures. The other is to apply robust statistical measures that are better
designed mathematically to reduce the effect of outliers or a small
number of samples.
Spatial Outliers
Special attention
since these valuesshould be given to
will seriously abnormal
distort values (spatial
the interpolation outliers)
process. A
good way to detect the spatial outliers is to map or visually display
the data set. Map display provides insights about the data set, how the
samples are spatially distributed (uniform or clustered), their general
pattern, trend and the extreme high or low values associated with the
data set. The spatial location of extreme values is helpful in detecting
erroneous data.
The decision to split the data into more homogeneous subsets should be based on
physical considerations and the density of the available data. Each subset should
have enough data to allow the reliable inference of statistics for each population.
Stratification should be based on information other than the data values
themselves. Don’t stratify by putting high values into one stratum and low value
into another. Secondary information (soil type, vegetation class) can be used
when there is justification for stratification. Alternatively the data could be
divided according to geographic location or elevation. Care should be taken
because there may be no justification for stratification.
Transformation
Data transformations can be used to make data normally distributed. Kriging
relies on the assumption of stationarity, and this assumption requires in part, that
all data values come from distributions that have the same variability. Data
transformations can improve predictions in many but not all cases.
Transformations and trend removal are often applied to data to justify
assumptions of normality and stationarity. The goal of transformations is to
remove the relationship between the data variance and the trend. When data are
composed of counts of events, such as crimes, the data variance is often related to
the data mean. That is, if you have small counts in part of your study area, the
variability in that region will be larger than the variability in a region where the
counts are larger. The histogram and normal QQ plots can be used to see what
transformations, if any, are needed to make a data set more normally distributed.
The same transformation will likely equalize variances as well, helping to satisfy
the stationarity assumption.
Arcsine Transform
For percentages or proportions (data between 0 and 1) the arcsine transform can
be used. When data consists of proportions, the variance is smallest near 0 and 1
and largest near 0.5. The arcsine transformation often yields data that has
constant variance throughout the study area and often makes the data appear
normally distributed (Krivoruchko, 2005). Kriging prediction with a Box-Cox or
arcsine transformation is known as transGaussian kriging (Krivoruchko, 2002a).
To create an output map, the results of transGaussian kriging require back
transformation to the srcinal data, however, it should be noted that this can only
be done approximately (Cressie, 1993).
Box-Cox Transform
Box-Cox (also knownas power transformation) is a useful method to alleviate
heteroscedasticity when the distribution of the dependent variable is not known.
These
causedtransformations
by nonnormality areof
defined
one ofonly
thefor positive an
variables, dataindirect
values.relationship
Heteroscedasticity
betweenis
variables, or the effect of a data transformation.
Sichel (1971, 1973) found that lower frequency distribution drop off can be
brought back into line by adding a constant number to every one of the sample
values before taking logarithms. If the probability plot shows a straight line, the
population is said to be “three parameter lognormal.” According to Clark and
Harper (2000) the additive constant has no physical meaning and is basically, just
an artifact to normalize the logarithms. The size of the additive constant is a
debatable issue with the most common rule of thumb being that it would be a
problem if the additive constant starts to get to the same order of magnitude as
the average value.
Sampling can reflect a mixture of normal or lognormal populations. For example
if measurements are taken from both male and female populations the results may
be comprise of two different normal populations. This structure will be reflected
in the histogram and probability plots.
The goal of the normal score transform is to make all random errors for the whole
population normally distributed. Thus it is important that the cumulative
distribution from the sample reflect the true cumulative distribution of the whole
population (Krivoruchko, 2005). An unfortunate property of the normal score
transformation is that it destroys the interval property of the srcinal data,
because the quantile has a different meaning depending whether it occurs in the
middle or at the tails of a standard normal distribution.
Scale Transform
In some cases it can be useful to scale data to a range of 0 to 1 if the values are
extremely large.
Offset
In the case of a lognormal or square root transform if the Z value spans the range
of less than one to greater than one (<1 to >1) all values should be made greater
than one prior to transformation by adding an offset value.
Back-transformation
When a transformation is chosen, after analysis of the transformed data, the
output data are customarily (but not necessarily) back-transformed to the srcinal
Standard
The standard back-transform is simply the converse of the transformation –
scaled values are rescaled to the srcinal range, log transformed values are raised
to the natural exponent (e) and squared values are raised to the square root (0.5).
Offset values are subtracted from the back-transformed values.
Weighted
The weighted back-transformation is a complex back-transformation that more
closely approximates true population statistics than simple back transformations
(Hann, 1977 and Krige, 1981).
Trends
When there is a trend in the data the first task is to identify the pattern and degree
of the trend and then to decide whether the trend needs to accommodated. Trends
can occur because of the measurement process, sampling design, or can be due to
the physical qualities of the sampled object. Climatic data may follow a strong
directional pattern. Increasing rainfall from west to east leads inevitably to
constantly increasing local means in the same direction.
1. Study the data to establish the predictability of values from place to place
(determine dependency rules) and define the degree of autocorrelation.
The sample semivariance is used to estimate the shape of the variogram (the
curve that represents the semivariance as a function of distance). The variogram
describes the spatial relationship between the data points by showing the spatial
structure and the data associations.
VARIOGRAMS
Geostatistics defines the correlation between any two values separated by a
distance (known as the lag distance) then uses this information to make
predictions at unsampled locations. Semivariance is a measure of the degree of
spatial correlation among sample data points as a function of the distance and
direction between the sample data points. The independent variable is the
distance between pairs of points. The dependent variable is the semivariance of
the differences in the data values for all samples a given distance apart. The
semivariogram (or simply the variogram) controls how kriging weights are
assigned to points during interpolation, and consequently controls the quality of
the results. The semivariance increases with distance until the semivariance
equals the variance around the average value and no longer increases, causing a
flat region to occur on the semivariogram called the sill.
Scale
It is not difficult to calculate and plot the semivariogram.
0 K1m The first step is to define a lag increment h(), which is the
spacing between any two points in the data. For the
20.1 14.6 13.0 15.4 12.9 16.1 16.5
7 Km example data a 1 kilometer lag increment is arbitrarily
defined. For each pair separated by 1 kilometer, calculate
15.7 11.8 20.9 17.1 24.3
6 Km
) the difference, and then square the difference. Sum up all
g the differences and divide by twice the number of pairs.
n i
th 5 Km
28.5 13.6 35.4 27.2
r This gives the measure of similarityγ(h) for the variogram
o
(N 30.2 0.0 33.3 19.6 23.8
for a particular lag increment or distance. The same is done
te 4 Km for other lag distances of 2, 3, 4, 5, and 6 kilometers.
a
in Grouping the pairs by location is referred to as binning.
d
r 29.9 31.3 7.5 33.3 23.3 18.2
That is, all points that are within 0 to 1 kilometres apart are
o 3 Km
o
c
- grouped into the first bin, those that are within 1 to 2
Y
2 Km
43.7 14.4 12.1 16.3 15.5 55.1
kilometres apart are grouped into the second bin, and so on.
N
33.6 29.7 31.9
1 Km
1 Km 2K m 3K m 4K m 5K m 6K m 7K m
γ(h) = (Σ [Z(xi) - Z(xi + h)]
2
) / 2N
i=1
X-coordinate ( Easting )
1
First Lag Distance
Twenty (N = 20) pairs are found for lag distance h = 1. The pairs in the east-west
direction and their calculated differences:
20.1 – 14.6 = 5.5 14.6 - 13 = 1.6 13 - 15.4 = -2.4 15.4 - 12.9 = 2.5
12.9 - 16.1 = -3.2 16.1 - 16.5 = -0.4 15.7 - 11.8 = 3.9 20.9 - 17.1 = 3.8
17.1 - 24.3 = -7.2 28.5 - 13.6 = 14.9 33.6 - 19.6 = 13.7 19.6 - 23.8 = -4.2
29.9 - 31.3 = -1.4 31.3 - 7.5 = 23.8 7.5 - 33.3 = -25.8 33.3 - 23.3 = 10
14.4 - 12.1 = 2.3 12.1 - 16.3 = -4.2 16.3 - 15.5 = 0.8 15.5 -55.1 = -39.6
30.25 + 2.56 + 5.76 + 6.25 + 10.24 + 0.16 + 15.21 + 14.44 + 51.84 + 222.01 + 187.69 +
17.64 + 1.96 + 566.44 + 665.64 + 100.00 + 5.29 + 17.64 + .64 + 1568.2 = 3489.82
1
The example variogram calculations have been adapted from Carr (1995).
20.1 - 12.9 = 7.2 14.6 - 16.1 = -1.5 13.0 - 16.5 = -3.5 15.7 - 17.1 = -1.4
11.8 - 24.3 = -12.5 28.5 - 35.4 = -6.9 30.2 - 33.3 = -3.1 0 - 23.8 = -23.8
29.9 - 23.3 = 6.6 7.5 - 18.2 = -10.7 43.7 - 16.3 = 27.4 14.4 - 55.1 = -40.7
15.7 11.8 20.9 17.1 24.3 15.7 11.8 20.9 17.1 24.3
6 Km 6 Km
) )
g g
in in
h
t 5 Km
28.5 13.6 35.4 27.2 ht 5 Km
28.5 13.6 35.4 27.2
r r
o o
N N
( 30.2 0.0 33.3 19.6 23.8 ( 30.2 0.0 33.3 19.6 23.8
te 4 Km te 4 Km
a a
n i ni
d d
r 29.9 31.3 7.5 33.3 23.3 18.2 r 3 Km 29.9 31.3 7.5 33.3 23.3 18.2
o 3 Km o
o o
c c
-Y -Y
43.7 14.4 12.1 16.3 15.5 55.1 43.7 14.4 12.1 16.3 15.5 55.1
2 Km 2 Km
1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km 1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km
15.7 11.8 20.9 17.1 24.3 15.7 11.8 20.9 17.1 24.3
6 Km 6 Km
) )
g g
in in
th 5 Km th 5 Km
28.5 13.6 35.4 27.2 28.5 13.6 35.4 27.2
r r
o o
N N
( 30.2 0.0 33.3 19.6 23.8 ( 30.2 0.0 33.3 19.6 23.8
e t 4 Km e t 4 Km
a a
n
i n
i
d d
r 29.9 31.3 7.5 33.3 23.3 18.2 r 29.9 31.3 7.5 33.3 23.3 18.2
o 3 Km o 3 Km
o o
c - c -
Y 43.7 14.4 12.1 16.3 15.5 55.1 Y 43.7 14.4 12.1 16.3 15.5 55.1
2 Km 2 Km
1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km 1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km
15.7 11.8 20.9 17.1 24.3 15.7 11.8 20.9 17.1 24.3
6 Km 6 Km
) )
g g
n
i n
i
ht 5 Km
28.5 13.6 35.4 27.2 ht 5 Km
28.5 13.6 35.4 27.2
r r
o o
N N
( 30.2 0.0 33.3 19.6 23.8
( 30.2 0.0 33.3 19.6 23.8
te 4 Km te 4 Km
a a
n
i n
i
d d
r 3 Km 29.9 31.3 7.5 33.3 23.3 18.2 r 3 Km 29.9 31.3 7.5 33.3 23.3 18.2
o o
o o
c
- c
-
Y 2 Km 43.7 14.4 12.1 16.3 15.5 55.1 Y 2 Km 43.7 14.4 12.1 16.3 15.5 55.1
1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km 1 Km 2 Km 3 Km 4 Km 5 Km 6 Km 7 Km
y 200 ar
ti il
r m
si
la ss
u le
im
s
f 150 Spherical model
o sill 140
e
r )
u (C
s e
a
e n
c no further corralation between
ia sample points ( covariance becomes 0)
m 100 r
a
- v
l
) a
r
(h tcu ( C0 + C)
u
e
c rts
n 50
a
ir n u g g et ( C0) 2 0
a r
v a A0
ie
m e ilm
r
o ra ng e of in flu en ce ( ) 3
m si
s 0
0Km 1Km 2Km 3Km 4Km 5Km 6Km
Coordinate System
The geographic latitude/longitude coordinate reference system uses angular
measurements to describe a position on the surface of the earth. Only at the
equator is the distance represented by one degree of longitude equal to the
distance represented by one degree of latitude. Moving towards the poles the
distance between lines of longitude becomes progressively smaller until at the
exact location of the pole all 360° of longitude are represented by a single point.
Because lines of longitude are not uniform, a latitude/longitude is not a suitable
coordinate system for lag distances.
Lag
Lag Distance
The lag distance is the range over which autocorrelation is calculated. The
number of lags and the size of the lags must be specified based on knowledge of
the phenomena being analyzed and the reason for modeling the variogram. The
selection of a lag size has important effects on the semivariogram.
• If the lag distance is too small, there may be too many empty sample and
• The general rule is that lag size times the number of lags is
approximately equal to one-half of the largest distance among all points.
Optimally, a class size should be found where approximately the same number of
pairs result for each lag increment. For some data sets such a class size cannot be
found due to sampling geometry. For these data sets a class size should be sought
yielding an interpretable semivariogram showing a clear sill and range. It is
common that the sill will be reached in less than 20 lags (Myers, 1997).
Use the smallest lag class distance spacing that produce acceptable number of
pairs. Ideally 30 to 50 data pairs are consider a practical minimum to obtain a
reliable estimate of the variogram at any lag (Armstrong, 1984 and Russo, 1984).
In statistics, this approximates the number of pairs needed to produce a normal
distribution if drawing samples from a random distribution. Webster and Oliver
(1992) reported that variograms computed on fewer than 50 sample points are of
little value and that at least 100 data are needed. Their experiments suggest that for
a normally distributed isotropic variable a variogram computed from a sample of
150 data might often be satisfactory, while one derived from 225 data will usually
be reliable. The best lag spacing may differ for an isotopic versus anisotropic
variograms of the same data (Welhan, 2005).
Variable Lag Class Distance
Specifying a variable non-uniform lag class is appropriate when locations are
irregularly spaced or clustered across the interpolation area, or if the phenomenon
being studied is known or suspected to be composed of several components. In
this situation, a suggestion is to use lags with small separation distances at short
distances and use lags with increasingly larger separation distances at longer
distances. Changing the width of each lag may help improve detection of
continuity.
Distance Tolerance
Lag distance tolerance is defined to deal with irregularly spaced data. Typically
lag tolerance is defined to be half the lag increment used. For a lag increment of
one kilometer the lag tolerance would be ± 500 meters. When the variogram is
plotted the lag distance will be the average of all the distances between pairs
falling within the tolerance. The larger the lag tolerance the more pairs that can
be defined and the variogram will look smoother. Larger lag tolerances may also
lead to unstable variograms that mask the continuity structure. The smaller the
tolerance the fewer number of pairs will be defined and the variogram will not be
as smooth.
camera. There will be a certain lag (in each direction) at which the
22.5
angular
o
L
ag
3
le
angular
tolerance
will mask anisotropy that may be present. A tolerance of 90 degrees
g
an
2 ec
t io
n means an omni-directional variogram, regardless of azimuth.
ag d ir
L
1 nce
Directional
examining thetolerance must be chosen carefully and the process of
h -scatter plots at different lag widths or angular
ag ar tolera
L angul
o
tolerances is essential.
90
x-axis
Range
The range (A0) is the distance where the semivariance reaches the sill. The range
of influence defines a neighborhood within which all locations are related to one
another. Samples separated by distances in excess of this range are said to be
spatially independent (not correlated). The range signifies the distance beyond
which sample data should not be considered in the interpolation. A larger range
means more continuous behavior, the variable is very well correlated and
predictions result in fairly smooth maps. The range typically tends to increase as
more and better data become available.
Sill
When the distance between sampling points is zero the value at each point is
compared to itself. Therefore the difference is zero and the calculated
semivariance for γ(0) should be zero. If h is a small distance, the points being
compared tend to be very similar, and the semivariance will be a small value. As
the distance h increases, the points being compared are less and less closely
related to each other and their differences become larger, resulting in larger
values of γ(h) . Spatial variability in most environmental data between sample
pairs increases as the separation distance increases.
At some distance the points being compared are so far apart that they are no
longer related to each other, and their squared differences become equal in
magnitude to the variance around the average value. The semivariance no longer
increases and the semivariogram develops a flat region called the sill. The sill
indicates that there is no further correlation between pairs of samples. The higher
the sill value, the higher the prediction variances.
The semivariogram sill is theoretically equivalent to the variance of the spatial
samples – however the sample variance is often an inappropriate measure of the
population variance (Clark, 1979b, Rossi, et al. 1992; Isaaks and Srivastava
1989). The standard formula for estimating the variance assumes independent
data, which is invalid in most environmental situations (Goovaerts, 2002, Barnes,
1991). In general if a sill is clearly present its value should be used as an estimate
of the population variance, and the sample variance should not be used as an
• the data are evenly distributed across the area of interest; there is no
significant trend in the data across the area of interest; and
• the dimension of the area of interest is more than three times the effective
variogram range.
Nugget
When the semivariogram is plotted back to the zero lag distance the nugget effect
(nugget variance) is found. The nugget effect would be expected to be zero since
two samples from the same point should have the same value. If the intercept is
greater than zero then a random or unstructured component is present. No matter
how close sampling is done the difference between neighboring samples will
never be zero.
The term nugget effect was coined from gold mining. In most gold deposits the
nugget effect tends to be quite large due to the nuggety nature of the
mineralization, so that samples taken close together can potentially have very
different grades. No matter how close the samples get, there will be large
differences in value between the samples because of the nuggety occurrence of
the gold. The value of the nugget effect will be close to zero in those deposits that
have a very uniform grade distribution, such as porphyry coppers (Surpac, 2004).
The nugget effect represents the inherent variability of the data which can be
interpreted as sampling errors, measurement errors, reproducibility issues, short
scale variations or random variation at distances smaller than the sampling
interval (white noise). A nugget structure increases the variability uniformly
across the entire variogram because the variability is not related to distance or
direction of separation. Increasing the nugget effect inflates the prediction error
(kriging variance) and hence increases uncertainty of estimated values (Siska &
Kuai Hung, 2001). Specifying a nugget effect causes kriging to become more of
a smoothing interpolator, implying less confidence in individual data points
versus the overall trend of the data. The higher the nugget affect the smoother the
resulting grid.
Ideally the nugget effect should be as small as possible. A large nugget effect
denotes unreliable data and more effort should be put into getting the data correct
rather than using the unreliable data with large nugget effect (ACE, 2005). A
high nugget variance indicates large point-to-point variation at short distances
and increased sampling often reveals more detail in structure. The nugget effect
tends to increase with lag tolerance and with data scarcity. Typically, the nugget
effect decreases, as more and better data become available.
The way in which the sample variogram behaves at near zero separation
distances is critical in describing the spatial continuity. A variogram that consists
of pure nugget effect indicates complete lack of spatial dependence. It is not
There are two possible ways in which geostatistical software will deal with the
semivariogram model at zero distance:
2. allow the model to hit the vertical axis (γ(0) = nugget effect.
The first option makes kriging an exact interpolator and the sample location will
be honored and the kriging variance will be zero. This is what Matheron
srcinally specified and is described in early test books. The second option means
that kriging will not exactly honor the data but will put most of the weight on the
sample and some weight on the other samples.
If the software only allows the second option the only way to honor the sample
values is to have a zero nugget effect. The nugget effect does not have to be
removed from the model, but another component (perhaps spherical) is added to
the model where the sill equals the real nugget effect and whose range of
influence is just below the closest sampling spacing. If it is not known which
option is implemented in the software, run the interpolation with the nugget
effect and with this alternative. If there is no difference in the results the software
does the first option.
If the sampling
combination canerrors can where
be used be quantified and the
short range second(for
spherical method is implemented
example) replaces thea
smaller scale variability and the nugget effect reflects the ‘true’ replication error.
The choice to filter out the replication error by removing the nugget effect from
the model must be made. It is important to keep in mind that when using the
second method and/or removing the nugget effect when kriging, the calculated
variances will be too low by a factor of 2*nugget effect. If the nugget effect is
divided as suggested, the kriging variance will be too low by a factor of
2*replication error (Clark, 2006b).
Species of North American birds censused on the breeding bird survey vary
considerably in their nugget variances (Villard and Maurer, 1996). Temperate
migrants tend to have higher nugget variances than Neotropical migrants, and
species that use a number of different habitat types have higher nugget variances
than species using a single type of habitat. For estimates of abundances obtained
from censuses like the breeding bird survey, part of the nugget variance could be
due to the inherent variability among observers in counting birds at different
breeding bird survey routes. The second component is due to the discontinuous
nature of the process in space. For birds, this would be related to the continuity of
the habitat in which they were being counted. Since the habitat of a species is
generally discontinuous in space, this is a likely component of the nugget
variance for data like the breeding bird survey.
y-axis
cloud). of As
the pairs
variogramof
locations become farther
apart (moving to the right
on the x-axis of the
variogram cloud), pairs of
locations should become
more dissimilar and have a
higher squared difference
(high on the y-axis of the
semivariogram cloud). The
variance cloud is useful for
discovering outliers.
Scattergram
The major
for the purpose of between
relationship fitting a model
values isatto specified
give to provide a mathematical
distances. formula
Semivariances are
estimated at discrete values of h, whereas the true variogram is continuous. The
estimates are subject to error, and unless a large sample is taken the experimental
variogram will appear erratic. The better correlated the variable the smoother the
interpolated surface. Semivariograms can be a mix of two or more models.
The model selected influences the prediction of the unknown values, particularly
when the shape of the curve near the srcin changes significantly. The steeper the
curve as it nears the srcin, the more influence the closest neighbors will have on
the prediction. As a result, the output surface will be less smooth. To consider the
effect of the model on predictions, think in terms of the shape of the variogram
for early lags
Linear
e
The linear model is the simplest model for a semivariogram and applies when the
c
n spatial variability increases linearly with distance and never levels off. The linear
a
r
a
v
model indicates that the greater the separation of two samples, the greater the
m
e
difference in the two samples. The linear model does not have a plateau and may be
s
considered to be the beginning of the spherical or exponential model (Matheron,
distance 1963a, 1965). The range is defined arbitrarily to be the distance interval for the last
lag class in the variogram. Because the range is an arbitrary value it should not be
compared directly with the ranges of other models. There is no sill, and the sill is
the calculated semivariance for the arbitrarily defined range.
e
c
n The linear/sill model is similar to the linear model except that at some distance
a
r
a (the range), pairs of points will no longer be autocorrelated and the variogram
v
m
e will reach a sill. This model should not be used unless the variation is limited to
s
one dimension (Webster, 1985).
distance
The semivariogram immediately takes its maximum value (represented by a flat
semivariogram) if there is no correlation. In this case the phenomenon is
e
c completely
n
a the scale of random
sampling.andMeasurement
there is no spatial
errorsdependence
or samplingamong thefar
sites too observations at
apart (and as
r
a a result are spatially independent) or both may lead to noisy semivariograms that
v
m appear as pure nugget effect. Choosing a pure nugget effect model is an extreme
se
modeling decision that precludes use of kriging.
distance
of the two spheres, given the distance between their centres (Clark, 2006a)
The spherical model shows a progressive decrease of spatial autocorrelation
(increase of semivariance) of the variable with increasing distance. Dependence
fades away altogether when the model reaches a sill. The spherical model is a
modified quadratic function for which at some distance (the range) pairs of points
will no longer be autocorrelated and the semivariogram reaches a sill (becomes
asymptote). The spherical model is one of the most commonly used models.
Exponential
The exponential model was developed to represent the notion of exponential
decay of “influence” between two samples. The exponential model applies
e
when the spatial dependence decreases exponentially with increasing distance.
c
n
a
This dependence disappears completely at infinite distance. The exponential
r model is similar to the spherical model in that it approaches the sill gradually,
a
v
m but different from the spherical model in the rate at which the sill is
e
s approached and in the fact that a definite sill is never actually achieved. The
distance
exponential model reaches 98% of its sill value at around three times the
range of influence. The exponential model is also a commonly used model.
Gaussian
The Gaussian (hyperbolic) model is similar tothe exponential model but assumes a
e
c
n gradual rise for the y-intercept. The Gaussian semivariogram model is named
a
r because the formula is basically identical to that for the normal (Gaussian)
a
v
m probability distribution. Spatial dependence vanishes only at an infinite distance.
se This model represents phenomena that are extremely continuous or similar at short
distance
distances. The main feature of this model is its parabolic shape at the srcin. It
expresses a rather smooth spatial variation of the variables. The Gaussian model
does not have a true range but rather a parameter used in the model to provide
range (called the effective range). For a Gaussian model the range of influence is
square root of three times the distance scaling parameter. The Gaussian model is
generally unstable in the presence of a nugget effect or of very dense data. This sort
of model occurs in topographic applications or where samples are very large
compared to the spatial continuity of the values being measured.
Nugget Variance
Nugget variance should be viewed as an isotropic parameter because it is not
physically meaningful to say that uncorrelated (random) variability at a point is
different depending on the direction away from that point. If the apparent nugget
in two directional variograms is substantially different the data should be
modeled as an additional anisotropic nested (non-nugget) structure with a
maximum range so large as to add negligibly to the variogram at small lags
(Welhan, 2005).
Geometric Anisotropy
Geometric (or affine) anisotropy is a situation where the variogram exhibits
different ranges in different directions and the sill is constant in all directions. A
longer range in one direction means that the values are more continuous in that
direction than for a direction with a shorter range (Liebhold et al., 1993). For
example, in an Aeolian deposit, permeability might have a larger range in the
wind direction compared to the range perpendicular to the wind direction (Syed,
1997).
The anisotropy ratio is the ratio between the smallest range and biggest range
(these directions are assumed to be approximately perpendicular to each other). A
ratio of one denotes an isotropic variogram (Syed, 1997). The ratio between the
lowest range and highest range, and the difference between their two angles can
be used to transform the model to an isotropic model suitable for kriging (Ricci,
A.K., 1998).
Correlogram
The correlogram shows the correlation among sample points by
plotting autocorrelation against distance. Correlation usually
decreases with distance until it reaches zero. A correlogram mirrors
the variogram with the expectation that the correlogram assumes a
stationary mean (Srivastava, 1996). As a variogram gradually rises to
a plateau the correlogram gradually drops and levels out.
Correlograms can be used to detect and describe regular spatial
patterns between patches (Radeloff, et al. 2000). For a repeating
pattern of patches separated by some distance, the patch size appears
as a peak in autocorrelation for distances up to average patch size; the
distance between patches appears as a second peak.
Covariance
The covariance (like the semivariogram) measures the strength of
statistical correlation as a function of distance – when two locations
are close to each other they are expected to be similar and so their
covariance will be large. As the two locations move further apart,
they become less similar and their covariance eventually becomes
zero. Since covariance decreases with distance, covariance can be
thought of as a similarity function.
Cross-K
The Cross-K (Cressie, 1993) is similar to Ripley’s K. A point from the first data
set is chosen. A radius ( h) is delineated and the points from the second data set
that are within the circle are counted. A line above that designated by the
interaction between two random point sets would imply that the data points are
attracted to each other or tend to exist together in the same place. A line below
shows that the points repel each other or tend to exist in opposing space.
Drift
Drift (also called trend) is calculated as the difference between
values at different sample locations within a particular lag
distance. Drift function is used to assess non-stationarity – does
the mean value change over distance).
Fractal
Mandelbrot introduced the term fractal for temporal or spatial
phenomena that are continuous but not differentiable, and that
The fractal dimension is directly related to the slope of the best-fitting line
produced when the log of the distance between samples is regressed against the
log of the mean-squared difference in the elevations for that distance. A graphical
representation of variations of fractal dimension values as a function of scale has
been called a fractogram (Palmer, 1988). According to Burroughs (1983), the
slope of a variogram model plotted on log-log axes can be translated directly into
Geary’s C
Inverted Covariance
Madogram
The madogram is a graph of mean absolute difference of sample
measurements as a function of distance and direction.
Madograms are less sensitive to extreme data values and can be
useful for inferring range and anisotropy of data sets with outlier
values that may make variogram features difficult to discern
using traditional methods (Sobieraj, Elsenbeer and Cameron,
2003). The madogram should not be used for modeling the
nugget of the semivariograms (Deutsch and Journel, 1992;
Goovaerts, 1997).
No variography or kriging should be carried out for data with a zero Moran’s I
for all lag distances since this does not respect Tobler’s Law. Deeper slopes in
the Moran’s I correlogram indicate a shift from a spherical to an exponential
model while flat profiles with a high spatial autocorrelation reveal a shift from a
spherical to a Gaussian one (Garrott, 2003).
squares).
deviationsWhen modeled
must be and by
multiplied used
thefor kriging values
estimated the relative kriging standard
to be comparable with
kriging standard deviations produced with ordinary variogram models.
Local
The local relative variogram considers the data within each region as a separate
population and the values are proportional to their means. To describe the overall
spatial continuity the local variogram values are scaled by their means. This
method is rarely used because of the difficulty of determining the regions.
General
The general relative variogram weights the semivariance by the
mean value of the data by dividing each semivariance by the
squared mean of all samples used to estimate the semivariance.
General relative variograms are less sensitive to clustering and
data outliers and therefore may provide cleaner estimates of
spatial correlation (Deutsch and Journel, 1997, p.45).
Pairwise
The pairwise relative variogram calculations are adjusted by a
square mean performed for each pair of sample values by dividing
each sample pair’s difference by its mean. Experience has shown
that the general relative and pairwise relative semivariograms are
effective in revealing spatial structure and anisotropy when the
scatter points
relative are sparse
variograms should (Deutsch & on
only be used Journel, 1992).
positively Pairwise
skewed data
sets. Pairwise relative variograms are less sensitive to clustering
and data outliers and therefore may provide cleaner estimates of
spatial correlation (Deutsch and Journel, 1997, p.45).
Standardized
In standardized variogram analysis the variogram is computed on
the natural logarithms of the variates.
Ripley’s K
Ripley’s K (Ripley, 1977; Cressie, 1993) is a spatial statistics function that is
used to analyze spatial autocorrelation (usually clustering, rather than
dispersion). Ripley’s K is also called the reduced second moment measure – that
is it measures second order trends (local clustering as opposed to a general
pattern over the region. The function chooses a point within the data set and
counts all the other points within a distance of that point. Like any statistic
Ripley’s K is prone to biases which include: edge biases, sample size, scale of
interpretation, and the shape of the boundaries (Ned Levine & Associates, 2004).
zij = Σ wijzi j = 1, …, m
i=1
where the weight values Wij can be calculated by solving the following system of
equations:
n
Σ wki = 1 j = 1, …, m
k=1
Ordinary Kriging
Ordinary kriging is done when there is no underlying spatial trend (drift), the
mean of the variable is unknown and the sum of the kriging weights is equal to
one. This method assumes that the data set has a stationary variance but also a
non-stationary mean within the search radius. Ordinary Kriging is highly reliable
and is recommended for most data sets. One of the main issues concerning
ordinary kriging is whether the assumption of a constant mean is reasonable.
Sometimes there are good scientific reasons to reject this assumption.
Simple Kriging
Simple kriging assumes that the data set has no underlying spatial trend, a
stationary variance and a known mean value. Because the mean is assumed
known, it is slightly more powerful than ordinary kriging, but in many situations
the selection of a mean value is not obvious. Simple kriging also assigns a weight
to the population mean, and is in effect making a strong assumption that the
mean value is constant over the site. Simple kriging also requires that the
available data be adequate to provide a good estimate of the mean. Simple
kriging uses the average of the entire data set (a constant known mean) while
ordinary kriging uses a local average (the average of the scatter points in the
kriging subset for a particular interpolation point). Simple kriging can be less
accurate than ordinary kriging, but it generally produces a result that is
"smoother" and more aesthetically pleasing.
Universal Kriging
Universal kriging combines trend surface analysis (drift) with ordinary kriging
(Bonham-Carter, 1994). Universal kriging is more difficult to apply. Once trends
have been accounted for (or assumed not to exist), all other variation is assumed
to be a function of distance. This method represents a true geostatistical approach
to interpolating a trend surface of an area.
1. Fit a trend (local or global) and calculate the residuals. From these
residuals obtain a de-trended (driftless) semivariogram.
The recommended setting for the trend is a first degree polynomial which will
avoid unpredictable behavior at the outer margins of the data set (Loránd, 2005).
Punctual Kriging
Punctual (point) kriging provides an estimate for a given point by estimating the
value of that point from a set of nearby sample values. Punctual kriging can be
used if sampling was done to represent point values in a field or in time.
A fine spatial resolution (small blocks) will require denser sampling than a
coarser resolution (large blocks) to predict block means at a given level of
precision. An optimal sampling scheme will attempt to find the combination of
block size and sampling spacing in order to achieve a given level of precision at
the lowest total cost (Castrignanò and Lopez, 2003).
Block kriging can be used if sampling was done to represent the area around the
actual sample point. For example if soil samples from an area around the
sampling location were composited before analysis then block kriging may be
more appropriate. Block kriging produces smoother maps than punctual kriging
by interpolating average values over blocks with the effect of smoothing local
discontinuities. These effects are particularly desirable when the main interest is
to study regional patterns of variation rather than local details (Castrignanò and
Lopez, 2003). Consequently block estimates appear more reliable than those for
points. In environmental work block kriging is usually more appropriate than
punctual kriging.
By computing block estimates, the nugget effect, which may be due either to
measurement error or very short-range variation, can be avoided. Block
interpolation may be more appropriate than punctual interpolation where average
values of properties are more meaningful than exact single-point values,
especially where spatial or temporal dependence is weak (Burgess and Webster
1980).
1.250 Regression coefficient = .416 If all estimated values were the same as
(SE = .045, r2 = .399, y intercept = 0.222)
the true value the scatter plot would be
a 45-degree straight line. The distance
0.952
Z
each point is from the 45-degree line is
d the estimation error for that sample
et
a 0.655 value. The closer the cloud of points is
ts to this line, the better the estimation.
E The slope of the cross-validation scatter
0.358
plot of predicted versus true values is
usually less than one. It is a property of
0.060
kriging that it tends to under-predict
0.060 0.358 0.665 0.952 1.25 large values and over-predict small
Actual Z values. Systematic deviations from the
line indicate lack-of-fit with the model.
Individual points deviating from the line may be considered outliers.
Cokriging
Cokriging is a multivariate extension of kriging to allow the incorporation of two
or more variables that are correlated with each other by using one set of data to
help explain and improve the description of variability in another (Goovaerts,
1997, pp 203-248, Wackernagel, 1995). Cokriging assumes that the second data
set is highly correlated with the primary data set to be interpolated. For example
altitude is considered as an important additional variable when estimating air
temperature and hence cokriging should provide a better estimation than kriging
because cokriging would take altitude as an additional variable. As another
example, the quantity of vegetation (the biomass) may be related to elevation and
soil moisture, and if this relationship can be determined, it would be possible to
use cokriging to predict biomass based on both elevation and soil moisture.
Disjunctive Kriging
Disjunctive kriging (full indicator cokriging) is a specific nonlinear transformation
of the srcinal data that does not have simple (Gaussian) distributions (Rivoirard,
1994). The first step is to perform a normal score transformation of the data. The
next step is to express the normalized function as an expansion of hermitian
polynomials. Disjunctive kriging assumes all data pairs come from a bivariate
standard normal distribution. The bivariate normal distribution is weaker than the
multi-Gaussian condition. If bivariate normality exists, disjunctive kriging can
produce better results than other geostatistical methods. If not, another type of
kriging should be used. Disjunctive kriging tries to do more than ordinary kriging
and indicator kriging by considering functions of the data and requires making
stronger assumptions. The technique is computational daunting even with the use of
variogram modeling software.
Indicator Kriging
Indicator kriging (Journel, 1993, and Journel and Arik, 1988) is a method
developed for use on data populations with any distribution. The method is easy
to understand. A cutoff value is defined, and all data below the cutoff are
assigned a value of 0, all data above the cutoff are assigned 1. Variogram
modeling and kriging is done on binary (0 or 1) values. An advantage of
indicator kriging is to minimize the influence of outliers, to contend with skewed
distributions, to deal with populations of both high value and values below the
detection limit. Indicator kriging is not recommended for data having a trend.
Although sampling error is not taken into account the indicator transform can
improve estimation of spatial continuity parameters in two ways. By reducing the
scale of the variogram the behavior of the sample variograms frequently improves.
Being an arithmetic average kriging can be drastically affected by even a single
outlier. Deleting outliers is one approach to dealing with unusual values, but this
may not be desirable because data are thrown away. Alternatively an indicator
transform may be applied to the data, because no matter how unusual the data
values are the data are transformed to either 0 or 1. Since the indicators are 0 or 1
the indicator variogram is very well behaved and resistant to outliers. Indicator
variograms have a linear relationship (Lemmar, 1985).
When the threshold level does not correspond well with the continuity structure
the indicator variogram will not produce good result, for example, when a
skewed proportion of the data fall into one level relative to the other. Experience
shows transforming the data into indicators using the median of the distribution
as the threshold can provide a variogram revealing the approximate description
of variability which at times is a reasonable approximation (Myers, 1997).
Lognormal Kriging
To cope with the presence of a few high grades South African mining
geostatisticians developed lognormal kriging, which is kriging applied to
lognormal transforms of data followed by a back-transform of final estimates
(Journe, Huijbregts, 1978). The lognormal hypothesis is very strict in that any
departure can result in completely biased estimates. Lognormal kriging is rarely
used in mining applications today (Vann and Guibal, 2000). Some environmental
parameters are lognormally distributed, and in such cases, lognormal kriging may
be useful. The lognormal transform is referred to as a “sledgehammer” approach
to data manipulation and can significantly reduce or even eliminate true
distribution characteristics of the data (Myers, 1997). It is also believed that log
transforming data often goes too far, especially if data are not srcinally
lognormally distributed (Cressie and Hawkins, 1980).
indicators
recoverableatresource
several estimation
cut-offs. Multiple indicator
which is robust kriging values
to extreme is an and
approach to
practical
to implement (Vann and Guibal, 2000).
The distribution histogram, probability plot, and other characteristics of the data
should be examined before resorting to the complicated multiple indicator
approach. If a decent semivariogram is obtained from ordinary values, indicator
kriging is not needed. Theoretically multiple indicator kriging gives a worse
approximation of the conditional expectation than disjunctive kriging, which can
be shown to approximate a full cokriging of the indicators at all cut-offs, but does
not have the strict stationarity restriction of disjunctive kriging (Vann and
Guibal, 2000).
Uniform Conditioning
Uniform condition is a variation of Gaussian disjunctive kriging more adapted to
situations where stationarity is not very good. In order to ensure that the
estimation is locally well constrained, a preliminary ordinary kriging of relatively
large
value.panels is made,
Uniform and the is
conditioning proportions per panelbut
a robust technique aredoes
conditional
depend to that kriging
heavily on the
quality of the kriging of the panels.
Since cokriging must be done for a discrete number of cutoffs, there are a large
number of variograms required to implement probability kriging. If ten cutoff
values are used, then twenty-one variograms are necessary. The hybrid nature of
this estimate as well as the time-consuming complexity of the structural analysis
makes probability kriging rather unpractical (Vann and Guibal, 2000). Since
probability kriging uses additional information it should theoretically produce
better estimates than indicator kriging (Knudsen, 1987). Probability kriging is not
recommended for data having a trend.
Regression Kriging
Regression kriging (RK) is a spatial interpolation technique that combines a
regression of the dependent variable on auxiliary variable (such as terrain
parameters, remote sensing imagery and thematic maps) with kriging of the
regression residuals. Regression kriging is mathematically equivalent to
interpolations methods such as universal kriging and kriging with external drift
(KED) where auxiliary predictors are used directly to solve the kriging weights
(Hengl, et. al., 2003). Hengl (2006) has a web page that presents step-by-step
instruction for running regression kriging in a statistically sound manner – called the
generic framework for regression-kriginghttp://spatial-analyst.net/regkriging.php
( ).
According to the authors cited above some of the major points for variogram
interpretation are:
the spatial of
variability behavior
the sizeofofthe data –inthe
objects theshape
terrainof the variogram is related to
• the sill gives information on the total variability – the height of the
variogram is influenced by the density of coverage of the objects and the
spectral differences between the objects
• as the spatial resolution become coarser the overall variance of the data is
reduced and the fine scale variation becomes blurred – consequently, the
sill height will reduce, the range will increase, and the nugget will
increase
Each realization
from known datais points
different
andto so
others because there
individually is always
realizations areuncertainty
simulationsaway
not
estimates. Simulation has different objectives to estimation. The point of
simulation is to reproduce the variance of the input data, both in a univariate
sense via the histogram and spatially via the variogram. Thus simulations provide
an opportunity to study any problem relating to variability, for example risk
analysis, in a way that estimates cannot (Vann, et. al., 2002)
Simulations can be used to deal with a range of problems that involve variability,
and for such problems they are far superior than any estimate (including kriging).
One example of an application is to generate confidence intervals for an
estimated block. Other applications include modeling local spatial variability for
blending, stockpiling or mining selectivity studies (Bertoli, 2005).
Gaussian Simulation and Sequential Gaussian Simulation are the most common
conditional simulation methods. Non-parametric simulation techniques such as
Sequential Indicator Simulations and Probability Field Simulations are becoming
more and more popular.
Gaussian Simulation
Variogram modeling is performed on normally distributed data. The variogram
model is used to krig the data at all locations providing a base map. In
simulation, multiple realization maps are created that honor the actual data value
(a conditional map), and approximately the same variogram and distribution. To
generate each realization, an unconditional map is simulated which honors the
variogram model but not the data at the data locations. Then the same
unconditional map is used to create another map by kriging the values at the
same locations as the actual data. At each grid node a simulated error is obtained
– the difference between the kriged and simulated value. This error is added to
the base map at each grid location. This gives the first realization which can then
be back transformed to its srcinal distribution. This is repeated for other
realization using a different random number sequence to generate multiple
realizations of the map (Syed, 1997).
Plurigaussian simulation
Plurigaussian simulation tries to simulate categoricalvariables by the intermediate
simulation of continuous Gaussian variables (Armstrong et al., 2003). A
Plurigaussian simulation is way of simulating the geometry of reservoirs and
deposits. Fluid flow in oil reservoirs depends on their internal architecture - on the
spatial layout of the lithofacies. A similar problem arises in mining when different
facies with different grades or metallurgical recoveries are intermingled in a
complicated way. Plurigaussian simulations make it possible to simulate complex
types of geometry such as double (or multiple) anisotropies (e.g. gold bearing veins
in one dominant direction with conjugate veins in another direction) or ore types
that cannot touch each other. Plurigaussian simulation is an extension of Truncated
Gaussian simulation. Plurigaussian simulation has proved to be very efficient in
providing images reproducing the main features of the geology encountered in
kimberlite crater deposits (Deraisme et al., 2004).
is kriged(Gaussian)
normal which also gives the kriged
distribution that hasvariance. A random
a variance number
equivalent to theiskriged
drawnvariance
from a
and a mean equivalent to the kriged value. This number is then the simulated
number for that grid node. Another grid node is selected at random and the process
is repeated. This procedure defines a random path through all the gird nodes. For
the kriging all previously simulated nodes are included to preserve the spatial
variability as modeled in the variogram. When all the nodes have been simulated a
back transform to the srcinal distribution is performed. This gives the first
realization. This process is repeated for all the other realizations using a different
random number sequence to generate multiple realizations of the map (Syed, 1997).
For Sequential Gaussian Simulation the biggest problem is the search neighborhood
selection. The selection of small neighborhoods can lead to poor condition and poor
replication of the variogram.
Variables are divided into several thresholds (or cut-offs). The data indicators are
then coded (1 or 0) depending on exceedence or not of the thresholds. Sequential
indicator simulation proceeds as follows: a random location is chosen and
indicator kriging is performed for the threshold at that location. After correcting
for possible order relations violations the indicator kriging will define a full
cumulative probability density Function at that location. A simulated value from
this Cumulative Probability Density Function is drawn based on a uniform
random number generator. The sequence simulations then moves on to another
location with this simulated value assumed as “hard” data until all nodes are
Geostatistics is concerned with spatial data and each data value is associated with
a location in space and there is at least an implied connection between the
location and the data value. The semivariogram provides a spatial picture of the
data and choosing the correct variogram model to use is an art. The development
of an appropriate variogram model requires numerous correct decisions based on
a solid understanding of the data and the underlying processes from which the
data are drawn. The following is provided as a summary guide to understanding
the kriging process.
Data Collection
• State the objective of the modeling exercise in clear quantitative term.
• Select a sample size and a sample collection strategy that are consistent
with the objectives.
• Try to obtain a sample size of 100 data points or more – any calculations
using less than 100 samples may give poor results.
•
Mixed data sets – if a number of sampling methods were used, each data
set should be statistically analyzed separately to see if there are any
significant differences.
• If the data are multi-modal, the data may need to be split into different sets.
Semivariogram Visualization
• Check for enough number of data pairs at each lag distance (a
minimum of 30 to 50 pairs).
• Use a large lag tolerance to get more pairs and a smoother variogram.
distance spacing
average and tolerance.
betweenAthe
good starting point for the class size is the
samples.
• Remove trends.
Evaluation of Model
• Correct data errors.
Geostatistics is not a cure-all nor is it useful for all problems. It is not a black box
scheme and must be used with some care. In some cases, and for certain data sets
given some objectives, a user need only have access to a geostatistical software
package. In other cases it may be necessary to seek the advice and assistance of
someone more experienced and with a stronger understanding of the
mathematics/statistics in order to adequately apply geostatistics to a data set.
The information presented in this paper was developed to provide a good base
understanding to the field of geostatistics to either design or to analyze the data
from environmental monitoring surveys. Hopefully no tears were shed during
this learning process. An extensive bibliography is given which goes beyond the
references cited in this paper and is provided to give readers an entry to the
primary literature on geostatistics and surface interpolation.
ACE, 2005. Options for Variogram. Applied Computer Engineering (ACE), Inc. Integrated
Reservoir Studies. Houston Texax 77036.http://www.gridstat.com/ACEweb/example8.htm
Acevedo. M.F., 1997. Geog 5190 Advanced Stats and Quantitative Methods, Lecture
Outlines, Lab Guides and Assignments.
http://www.geog.unt.edu/~acevedo/courses/590/week10/lectur10.htm
Australian National University. 2004. Centre for Resource and Environmental Studies
(CRES)-ANUSPLIN Version 4.3. http://cres.anu.edu.au/outputs/anusplin.php
Arad, N., Dyn, N., Reisfeld, D. and Yeshurun, Y. 1994. Image warping by radial basis
functions: applications to facial expressions. CVGIP: Graphical Models and Image
Processing, 56, 161-172.
Ardini, F., Fioavanti, S. and Giusto, D.D., 1991. Multifractals towards Remote Sensing
Surface Texture Characterization. Proc. Int. Geoscience and Remote Sensing Symp.
(IGARSS’91), 3-6 June, Espoo, Finland, pages 317-320.
Armstrong A., Galli A., Le Loc’h G., Geffroy G. and Eschard R., 2003. Plurigaussian
Simulations in Geosciences, Springer.
Armstrong, M., 1984. Common problems with universal Kriging, Mathematical Geology
16(1): 101-108.
Armstrong, M., 1986. On the Fractal Dimensions of some Transient Soil Properties,
Journal of Soil Science, 37, 641-652.
Armstrong, M., and Alain G., 2001. Matheron's Contributribution to Geostatistics, Centre
de Géostatistique, Fontainebleau, France, 2001 Annual Conference of the International
Association for Mathematical Geology, Cancún, Mexico, September 6-12, 2001.
Ashraf, M., Loftis, J.C., and Hubbard, K.G., 1997.Application of Geostatistics to Evaluate
Partial Weather Station Networks. Agricultural & Forest Meteorology. 84(3-4): 225-271.
Atkinson, P.M. and Lloyd, C.D., 1998. Mapping Precipitation in Switzerland with
Ordinary and Indicator Kriging. Journal of Geographic Information and Decision
Analysis, Vol. 2, No. 2.: 72-86. ftp://ftp.geog.uwo.ca/SIC97/Atkinson/Atkinson.html
Barnes, R.J. 1991. The variogram sill and the sample variance. Mathematical Geology,
Vol. 23, No. 4: 673-678.
Barnes, R. 2005. Variogram Tutorial. Golden Software, Inc. Golden, Colorado 80401-
1866, U.S.A. http://www.goldensoftware.com/variogramTutorial.pdf
Bastin, G., Lorent, B., Doque, C., and Gevers, M., 1984 Optimal estimation of the average areal
rainfall and optimal selection of rain gauge locations, Water Resources Res. 20(4):
463-470.
Bates, D., Lindstrom, M., Wahba, G. and Yandell, B. 1987 GCVPACK - Routines for
Generalized
ComputationCross Validation. Communications in Statistics B - Simulation and
16: 263-297.
,
Berry, J.K, 1999. Extending Spatial Dependency to Maps. Beyond Mapping, Geoworld,
Vol 12 (1): 26-27.
Billings, S.D., Newsman, G.N., and Beatson, R.K. 2006. Fourier Transformation of
Geophysical Data Using Continuous Global Surfaces. University of British Columbia.
http://www.geop.ubc.ca/~sbilling/papers/fourier_transformation_using_cgs.PDF
Binkley, M.R., Carlson, R. and Lee, J., 1997. Spatial Interpolation of Lake Water Quality
Data. Kent State University. http://humboldt.kent.edu/~binkley/poster.html
Bogaert, P., Mahau, P. and Beckers, F., 1995. The Spatial Interpolation of Agro-Climatic
Data – Cokriging Software and Source Code,Use’ Manual, Version 1.0b. Unité de
Biométrie, Faculté des Sciences Agronomiaques, Université Catholique de Louvain,
Louvain-la-Neuve, Belgium. Agrometeorology Series Working Paper, Number 12, FAO
Rome, Italy
Boomer, K.M. and Brazier, R.A., 1997. Geostatistics, Seismic Risk Analysis of Northern
Arizona from Acceleration Contour Maps, Enhancing the Sampling Procedure through a
Geostatistical Analysis.http://www.geo.arizona.edu/geophysics/students/brazier/geo.html
Booth, T.H. and Jones, P.G. 1996. Climatic Databases for use in Agricultural Management
and Research Proceedings of the Arendal II Workshop on UNEP/GRID and CGIAR
cooperation to meet requirements for the use of digital data in agricultural management and
research. Arundal, Norway 9-11 May 1995.
http://www.grida.no/cgiar/htmls/climpres.htm#climpres
M_CONCEPTS/Assim_concepts8.html
Bretherton F.P., R.E. Davis and C.B. Fandry. 1976. A Technique for Objective Analysis and
Design of Oceanic Experiments applied to MODE-73. Deep Sea Research, 23, p. 559-582.
Bretherton F.P. and J.C. Williams. 1980. Estimations from Irregular Arrays. Reviews of
Geophysics and Space Physics, Vol.18, 4, p.789-812.
Briggs, P.R., and Cogley, J.G. 1996. Topographic Bias in Mesoscale Precipitation
Networks. Journal of Climate, 9, 205-218.
Bruce, J.P. and Clark, R.H., 1980. Introduction to Hydrometeorology. Pergamon Press,
Ontario. 324 pages.
Bruno, O.P, and Pohlman, M.M., 2003. High-order Surface Representation. California
Institute of Technology. http://www.acm.caltech.edu/~bruno/afosrposter03.pdf
Burgess, T.M. and Webster, R., 1980. Optimal Interpolation andIsarithmic Mapping of Soil
Properties I. The semivariogram and punctual kriging. Journalof Soil Science 31: 315-331.
Burgess, T.M. and Webster, R., 1980. Optimal Interpolation and Isarithmic Mapping of
Soil Properties II. Block kriging. Journal of Soil Science 31: 333-341.
Burrough, P.A., 1981. Fractal dimensions of landscapes and other environmental data.
Nature 294:240-242.
Burrough, P.A., 1993b. Soil Variability: A Lake 20 th Century View. Soils and Fertilizers,
pages 529-562.
Cadell, W., 2002. Report on the Generation and Analysis of DEMs for Spatial
Modelling. The Macaulay Institute. http://www.macaulay.ac.uk/ladss/documents/DEMs-
for-spatial-modelling.pdf
Calter, P., 1990. Technical Mathematics with Calculus. Prentice Hall, Engelwoods Cliffs,
New Jersey.
Carr, James R., 1995. Numerical Analysis for the Geological Sciences. Prentice Hall,
Englewoods Cliffs, New Jersey.
Carrasco, P., Ibarra, F., Le Loc’h, G., Rojas, R., Seguret, S., 2006. Application of the
Truncated Gaussian Simulation Method to the MM deposit at Codelco Norte, Chile.
http://www.geovariances.com/IMG/pdf/EAGE_Madrid_Leloch.pdf
Carrat, F. and Valleron, A., 1992. Epidemiological Mapping Using the “Kriging”
Method: Application to an Influenza-like Illness Epidemic in France. American Journal
of Epidemiology, 135(11): 1293-1300.
Castrignanò, A. and Lopez, R., 2003. GIS and Geostatistics: An Essential Coupling For
Spatial Analysis. Istituto Sperimentale Agronomico. Italian Interest Group on Spatial
Statistics. Text: http://194.119.196.14/istituti/ieif/spatialstatistics/res/gg.pdf, PPT:
http://194.119.196.14/istituti/ieif/spatialstatistics/res/gg_ppt.pdf
Chapman, A.D., 2003. The Case for a 3-minute Climate Surface for South America.
SpeciesLink, Appendix_e.pdf. http://cres.anu.edu.au/outputs/anudem.php
Chapman, A.D., Muňoz, M.E.S., and Koch, I. 2005. Environmental Information: Placing
Biodiversity Phenomena in an Ecological and Environmental Context. Biodiversity
Informatics, 2, 24-41. http://jbi.nhm.ku.edu/index.php/jbi/article/viewFile/5/3
Chang, T.J. and Teoh, C.B., 1995. Use of the Kriging Method for Studying Characteristics
of Ground Water Droughts. Water Resources Bulletin, American Water Resources
Association, Vol. 31, No. 6: 1001- 1007.
Cheng, K-S, Wei, C., Cheng, Y-B, and Yeh, H-C. 2003. Effect of Spatial Variation
Characteristics on Contouring of Design Storm Depth. Hydrological Processes, 17, 1755-
1769. Published online 28 January 2003 in Wiley InterScience (www.interscience.wiley.com).
DOI: 10.1002/hyp.1209. http://www.rslabntu.net/KSC/Publications/IDF_Contouring.pdf
Chellemi, D.O., Rohrbach, K.G., Yost, R.S. Yost, and Sonoda, R.M., 1988. Analysis of
the Spatial Pattern of Plant Pathogens and Diseased Plants Using Geostatistics. The
American Phytopathological Society. Vol. 78, No. 2: 221-226.
Chilès, J.-P., and Delfiner, P., 1999. Geostatistics: Modeling Spatial Uncertainty, John
Wiley & Sons, New York and Toronto, 695 pages.
Chua, S. H. and Bras, R. L., 1982. Optimal Estimators of Mean Areal Precipitation in
Regions of Orographic Influence. J. Hydrology, 57: 23-48.
Clark, I., 1983. Regression Revisited. Mathematical Geology, Vol. 15, No. 4,
1983.
th
Clark, I., 1986. The Art ofCross Validation in Geostatistical Applications. 19 Application
of Computers and Operations Research in theMineral Industry, R.V. Ramani , Editor.
Sponsored by The Pennsylvania State University, April 14-16, 19 86. Published by Society of
Mining Engineers, Inc., Littleton, Colorado 1986.
Clark, I., 1987. Turning the tables an interactive approach to the traditional estimation of
reserves", J. S. Afr. Inst. Min. Metall., Vol 87, No 10, pp.293-306. Full copy at
http://uk.geocities.com/drisobelclark/resume/Publications.html
Clark, I., 1998. Geostatistical Estimation Applied to Highly Skewed Data. JointStatistical
Meetings, Dallas, Texas, August, 1998.
Clark, I., 2000. Erratic highs -- a perennial problem in resource estimation SME 2000, Salt
Lake City, Utah.
Clark, I., 2001a. AI-GEOSTATS listserver email: Anisotropy with varying Range and Sill.
Clark, I., 2002. AI-GEOSTATS listserver email: Log transformation and zeros.
Clark I. and Harper,W.V., 2000. Practical Geostatistics 2000. Ecosse North America Llc,
Columbus, Ohio, U.S.A.
Clark I. and Harper,W.V., 2000. Practical Geostatistics 2000, Answers to the Exercises.
Cressie, N., 1985. Fitting variogram models by weighted least squares. Mathematical
Geology 17: 563 - 586.
Cressie N., 1990. The Origins of Kriging. Mathematical Geology, Vol. 22, No. 3, 1990: 239 - 252.
Cressie, N.A.C. 1991. Statistics for Spatial Data. John Wiley, New York.
nd
Cressie, N., 1993. Statistics for Spatial Data. 2edition. John Wiley and Sons, Inc, New York.
Cressie, N., 2003. Statistics for Spatial Data. Revised edition. John Wiley and Sons, Inc,
New York.
Cressie, N.A.C. and Hawkins D.M., 1980. Robust Estimation of the Variogram. Journal
International Association Mathematical Geology 12:115-125.
Cressman,
367-374. G., 1959. An Operation Objective Analysis System. Monthly Weather Review, 87,
Creutin, J.D., and Obled. C., 1982. Objective Analysis and Mapping Techniques for Rainfall
Fields: an objective comparison. Water Resource Research 18(2):413-431.
Daly, C. 2006. Guidelines for Assessing theSuitability of Spatial Climate Data Sets.
International Journal of Climatology. 26: 707-721. Published online 27 March 2006 in Wiley
InterScience (www.interscience.wiley.com) DOI: 10.1002/joc.1322.
Daly, C., Gibson, W.P., Taylor, G.H. Johnson, G.L., and Pasteris P. 200
2. A Knowledge-
based Approach to the Statistical Mapping of Climate. Climate Research 22: 99-113.
Daly, C., Gibson, W.P., Doggett, M.,Smith, J and Taylor, G. 2004. A Probabilistic-spati
al
Approach to the Quality Control ofClimate Observations. In Proceedings of the 14th AMS
Conference on Applied Climatology, American Meteorological Society, Seattle,
Washington, January 13-16, 2004. http://www.ocs.orst.edu/pub/prism/docs/appclim04-
probabilistic_spatial_approach-daly.pdf
Daly, C., Hlemer, E.H., Quinones, M. 2003. Mapping the Climate of Puerto Rico,
Vieques,
and Culebra. International Jounal ofClimatology 23: 1359-1381.
Daly, C, and Johnson, G.L. 1999.PRISM Spatial Climate Layers: Their development and
th
use. Short course on topics in applied climatology. 79 Annual Meeting of the American
Meteorological Society, American Meteorological Society: Dallas, Texax, 10-15 January.
http://www.ocs.orst.edu/pub/prism/docs/prisguid.pdf
David, M., 1977. Geostatistical Ore Reserve Estimation, Elsevier, 364 pages.
Davis, J.C., 1973. Statistics and Data Analysis in Geology. JohnWiley and Sons, New York.
nd
Davis, J.C., 1986. Statistics and Data Analysis in Geology, 2 edition, Wiley, New York.
Davis, B.M., and Borgman L.E., 1979. A Test of Hypothesis Concerning a Proposed
Model for the Underlying Variogram, Proceedings 16th APCOM, 163-181.
De Jong, S.M. 1993. An Application of Spatial Filtering Techniques for Land Cover
Mapping Using Tm Images, Geocarto International, 8(1):43-49.
Delfiner, P. and Delhomme, J.P. 1975. Optimum Interpolation y Kriging. Display and
Analysis of Spatial Data, Davis, J.C. and McCullagh, M.J. (Eds.) Wiley and Sons, New
York. pp. 96-114.
Delhomme, J.P., 1979. Kriging in the Hydrosciences, Advance in Water Res. 1(5): 251-266.
Deraisme. J., 1997. Interpolation and Reality. Geovariances News Letter #7, page 3.
http://www.geovariances.fr/publications/n1/7/page3.html
Deutsch, C.V., 1996. Correcting for Negative Weights in Ordinary Kriging. Computers
and Geosciences Vol. 22, No. 7: 765-773.
Deutsch, C.V. and A.G. Journel 1992. GSLIB Geostatistical Software Library. Oxford
University Press, New York. www.gslib.com
Deutsch, C.V. and A.G. Journel 1997/1998. GSLIB Geostatistical Software Library and
Users Guide, 2nd Edition. Oxford University Press, New York.
Dingman, S.L., Seely-Reynolds, D.M., and Reynolds III R.C. 1988. Application of
Kriging to Estimate Mean Annual Precipitation in a Region of Orographic Influence.
Water Resource Bulletin, 24,329-339.
Dorsel, D. and La Breche, T., 1998. Environmental Sampling & Monitoring Primer,
Kriging. http://www.ce.vt.edu/enviro2/smprimer/kriging/kriging.html
Duval, R., Lévy, R., Matheron, G., 1955. Travaux de D. G. Krige sur l’évaluation des
gisements dans les mines d’or sud-africaines. Annales des Mines, Paris, vol. 12, p. 3-49.
Eastman, J., 1999. IDRISI32, Guide to GIS and Image Processing – Volume 2, p 139 and
p 155-161. Clark Labs, Clark University.
Englund, E.J., 1990. A Variance ofGeostatisticians. J. Math. Geology. vol. 22, no. 4, pp.
417-455.
Englund, E.J., 1993. Spatial Simulation: Environmental Applications. A version of this
manuscript has been published in Environrnental Modeling with GIS, M.F. Goodchild et
al., eds., Oxford University Press (1993) 432-437.
http://www.epa/gov/nerlesd1/cmb/research/papers/ee104.pdf
Franke, R. 1985. Thin-plate Splines with Tension. Computer Aided Geometric Design 2,
87-95, North-Holland.
Freek, D., 1997. Classification of High Spectral Resolution Imagery Using an Indicator
Kriging Based Technique. International Institute for Aerospace Survey and Earth Sciences
(ITC), Department of Earth Resources Surveys. 350 Boulevard 1945, P.0. Box 6, 7500 AA,
Enschede, The Netherlands.
Galli, A., Beucher H., Le Loc’h, G., Doligez, B. and Group H., 1994. The pros and cons of
the truncated Gaussian method, in Geostatistical Simulations, pp. 217-233, Kluwer
Academic, Dordrecht.
Gamma Design. 2004. GS+ for Windows is a Windows 95/98/NT version of a
geostatistical mapping program for the PC. Gamma Design Software, LLC, Plainwell,
Michigan 49080 USA. http://www.gammadesign.com
Gambolatic, G. and Volpi, G., 1979a. A Conceptual Deterministic Analysis of the Kriging
Technique in Hydrology. Water Resources Res. 15(3): 625-629.
Gandin, L.S., 1963. Objective Analysis of Meteorological Fields: Gidrometeorolgicheskoe
Izdatel’stvo (GIMIZ), Leningrad (translated by Israel Program for Scientific Translations,
Jerusalem, 1965, 238 pages.
Gandin, L.S., and Kagan, R.L. 1974. Construction of a System for Objective Analysis of
Heterogeneous Data Based on the Method of Optimum Interpolation and Optimum
Agreement. Meteorological I Gidrol., No. 5 (Moscow), 1-11. [Translated by Joint
Publications Research Service.]
Garrott, J., 2003. SAKWeb: (Spatial Autocorrelation and Kriging Web) A W3
Geocomputation Perspective. Universidade Nova De Lisboa.
http://www.isegi.unl.pt/labnt/presentations/2003/031215Garrot.pdf
Geary, R., 1954. The Contiguity Ratio and Statistical Mapping. The Incorporated
Statistician, Volume 5, 115-145.
Genoton,, M.G. and Furrer, R., 1998. Analysis of Rainfall Data by Simple Good Sense: is
Spatial Statistics Worth the Trouble? Journal of Geographic Information and Decision
Analysis, Vol. 2, No. 2: 11-17. ftp://ftp.geog.uwo.ca/SIC97/Genton_1/Genton_1.html
Goovaerts, P., 2000. Geostatistical approaches for incorporating elevation into the spatial
interpolation of rainfall. Journal of Hydrology 228 (2000) 113-129.
http://www.terraseer.com/training/geostats/jhydrol00.pdf
Goovaerts, P., 2001. Geostatistical Modelling of Uncertainty in Soil Science. Geoderma
103:3-26.
Goovaerts, P. 2002. AI-GEOSTATS: Tuning a semivariogram.
Goovaerts, P. 2005. Geostatistical Analysis of Disease Data: Estimation of cancer mortality
risk from empirical frequencies using Poisson kriging. International Journal
of Health
Geographics 4:31.http://www.ij-healthgeographics.com/content/pdf/1476-072X-4-31.pdf
Goovaerts, P. 2006. Geostatistical Analysis of Disease Data: Visualization and
propagation of Spatial Uncertainty in Caner Mortality Risk using Poisson kriging and p-
field simulation. International Journal of Health Geographics 2006, 5:7. http://www.ij-
healthgeographics.com/content/pdf/1476-072X-5-7.pdf
Gottwald, T.R., Avinent, L, Llacer, G., de Mendoza, H. and Cambra, M., 1995. Analysis
of the Spatial Spread of Sharka (Plum Pox Virus) in Apricot and Peach Orchards in
Eastern Spain. Plant Disease. Vol. 79 No. 3: 266-278.
Gottwald, T.R., Cambra, M., Moreno, P, Camarasa, E. and Piquer, J., 1996. Spatial and
Temporal Analyses of Citrus Tristeza Virus in Eastern Spain. The American
Phytopatholgical Society. Vol. 86, No. 1: 45 - 55.
Griffith, D.A., 1987. Spatial Autocorrelation. A Primer. Resource Publication in
Geography, Association of American Geographers, Washington D.C.
Hansen,
No. 10: R.O., 1993. Interpretive Gridding by Anisotropic Kriging. Geophysics, Vol. 58,
1491-1497.
Hengl, T. 2006. Regression-Kriging. Spatial-analyst.net. http://spatial-
analyst.net/regkriging.php
Hengl, T., Heuvelink, G.B.M., and Stein, A. 2003. Comparison of Kriging with External
Drift and Regression-kriging. Technical note, ITC, Available on-line at
http://www.itc.nl/library/Academic_output/
http://spatial-analyst.net/GRK/Hengl_et_al_Comparison_RK_KED.pdf
Henley, S. 2001. Geostatistics – Cracks in the Foundations? Stephen Henley, Resources
Computing International Ltd. Derbyshire, U.K.
http://www.infomine.com/softwaremine/articles/GeostatCracks.asp and
http://www.earthsciswinfo.com/swmine/swarticles/GeostatCracks.html
Herzfeld, U.C, Eriksson, M.G., and Homlund P., 1993. On the Influence of Kriging
Parameters on the Cartographic Output – A study in Mapping Subglacial Topography.
Mathematical Geology, Vol. 25, No. 7: 881-900.
Hevesi, J.A., and Istok, J.D. 1992. Precipitation Estimation of Mountainous Terrain Using
Multivariate Geostatistics. Part I: Structure Analysis. Journal of Applied Meteorology, 31,
661-667.
Hijmans, R.J., Cameron, S., and Para J. 2004a. WorldClim, version 1.2. A Square
Kilometre resolution database of global terrestrial surface climate.
Hijmans, R.J., Cameron, S., and Para J. 2004b. WorldClim, a new high-resolution global
climate database. Abstract: Inter-American Workshop on Environmental Data Access.
Campinas, Brazil.
Hiller W. and R. Käse. 1983. Objective Analysis of Hydrographic Data Sets from
Mesoscale Surveys. Berichte aus dem Institut für Meereskunde an der Christian-
Albrechts-Universität Kiel, 116 78 pages.
,
Holawe, F. and Dutter, R., 1999. Geostatistical Study of Precipitation Series in Austria:
Time and Space. Vienna University of Technology.
Jones, T.A.,
With the Hamilton,
Computer. VanD.E., and Johnson,
Nostrand C.R.,
Reinhold, 1986. England.
London, Contouring Geological Surfaces
Jones, P.G. 1995. Centro Internacional de Agricultura (CIAT) climate database version
3.41, Digital data tape, Cali, Columbia: CIAT.
Jones, P.G., Robison, D.M. and Carter, S.E. 1990. A geographical information approach
for stratifying tropical Latin America to identify research problems and opportunities in
natural resource management for sustainable agriculture in Centro Internacional de
Agricultura Tropical (CIAT). Cali, Colombia: Agroecological Studies Unit., CIAT.
Journel, A.G., 1974 Geostatistics for Conditional Simulation of Ore Bodies. Economic
.
Krivoruchko, K., 2000. Applied Reservoir Characterization Using Geostatistics: The Value
of Spatial Modeling. AAPG Hedberg Conference, Houston, Texas, December 3-6.
Krivoruchko, K., 2002a. Introduction to Modeling Spatial Processes Using Geostatistical
Analyst. ESRI Australia.http://zappa.nku.edu/~longa/geomed/modules/esda/index.html
Krivoruchko, K. 2002b. Bridging the GapBetween GIS and Solid Spatial Statistics. New
Tools for Spatial Data Analsysis: Procceding of the CSISS Specialist Meeting, Santa Barbara,
California, May 10-11, 2002. University of California, Santa Barbara, Center for Spatially
Integrated Social Science.http://www.dpi.inpe.br/gilberto/csiss/papers/krivoruchko.pdf
Krivoruchko, K., 2002. Working on Nonstationarity Problems in Geostatistics Using
Detrending and Transformation Techniques: An Agricultural Case Study. paper was
presented at the Joint Statistical Meetings, New York City, August 2002. Available at
http://www.esri.com/software/arcgis/arcgisxtensions/geostatistical/research_papers.html
and
http://campus.esri.com/campus/library/ConfProc/GeostatisticsTeam/Krivoruchko_2002_
Working.pdf
Krivoruchko, K. 2005. Introduction toModelling Spatial Processes Using Geostatisti
cal
Analyst. ESRI White Papers.http://www.esri.com/library/whitepapers/pdfs/intro-
modeling.pdf
Kufs, C., 1987. Trend-Surface Modeling of Groundwater Data. Management of
Uncontrolled Hazardous Waste Sites. Silver Springs, MD. Hazardous Materials Control
Research Institute. http://members.home.net/terrabyte/t_surf.pdf
Kufs, C., 1998. Interpolation for Fun and Profit.
http://members.home.net/terrabyte/didyou.htm
Lorenc,
Scheme.A., 1981: A
Monthly GlobalReview,
Weather Three-Dimensional Multivariate Statistical Interpolation
109, 701-721.
Lovejoy, S., 1982. Area-Perimeter Relation for Rainand Cloud Areas. Science, 216:185-187.
Lynch, S.D. and Schulze, R.E., 1997. Techniques for Estimating Areal Daily Rainfall.
Department of Agricultural Engineering, University of Natal, Private Bag X01,
Scottsville, 3209, South Africa. http://www.ccwr.ac.za/~lynch2/p241.html
Mitasova,
ApplicationH.toand Hofierka,
Terrain J., 1993.
Modeling and Interpolation by Regularized
Surface Geometry Spline with Tension:
Analysis. Mathematical II.
Geology,
Vol. 25, No. 6: 657-669.
Myers, J., 1997. Geostatistical Error Management: Quantifying Uncertainty for
Environmental Sampling and Mapping, Van Nostrand Reinhold, New York.
Myers, D.E., 1999. What is Geostatistics?http://www.u.arizona.edu/~donaldm/whatis.html
Nalder, I.A. and Wein, R.W. 1998. Spatial interpolation of climatic Normals: test of a new
method in the Canadian boreal forest. Agric. For. Meteorol.92: 211-225.
Ned Levine & Associates, 2004. Chapter 5 – Distance Analysis I and II. CrimeStat III, The
national Institute of Justice, Washington, DC.
http://www.icpsr.umich.edu/NACJD/crimestat.html/and
http://www.icpsr.umich.edu/CRIMESTAT/files/CrimeStatChapter.5.pdf
New, M., and Hulme. M. 1997. Monthly Rainfall and Temperature Surfaces for Africa:
1951-1995 (CD-ROM) information: Production of Climate Surfaces. Climatic Research
Unit, University of East Anglia Norwich NR4 7TJ, UK.
http://www.mara.org.za/climatecd/info.htm
NIST, 2005. Engineering Statistics Handbook. NIST/SEMATECH e-Handbook of Statistical
Methods, Home:http://www.itl.nist.gov/div898/handbook/index.htm , Normal Probability
Plot: http://www.itl.nist.gov/div898/handbook/eda/section3/normprpl.htm
, Quanitile-
Quantile Plot:http://www.itl.nist.gov/div898/handbook/eda/section3/qqplot.htm
Olea, R., Ed., 1991. Geostatistical Glossary and Multilingual Dictionary, New York:
Oxford University Press.
Oliver, M.A. and Webster, R., 1991. How Geostatistics Can Help You. Soil Use and
Management. Volume 7, Number 4, December 1991: 206-217.
Oliver, M., Webster, R. and Gerrard, J., 1989. Geostatistics in Physical Geography. Part I:
theory. Trans. Inst. Br. Geogr. N.S. 14: 259-269 (1989) ISSN:0020-2754.
Oliver, M. Webster, R. and Gerrard, J., 1989. Geostatistics in Physical Geography. Part II:
applications. Trans. Inst. Br. Geogr. N.S. 14: 270 -286 (1989) ISSN:0020-2754.
Omre, H., 1984. The Variogram and its Estimation, in Verly, G., David, M., Journel, A.
G. and Marechal, A. (eds) Geostatistics for Natural Resources Characterization (Reidel,
Dordrecht): 107-125.
Onsoy, Y.S., Harter, T., Ginn, T.R., and Horwath, W.R., 2005. Spatial Variability and
Transport of Nitrate in a Deep Alluvial Vadose Zone. Published in Vadose Zone Journal
4:41-54 (2005), Soil Science Society of America 677 S. Segoe Rd., Madison, WI 53711
USA. http://groundwater.ucdavis.edu/Publications/Onsoy%20et%20al%20-
%20Spatial%20variability%20of%20nitrate%20-%20Vad%20Zone%20J%202005.pdf
Surpac Minex, 2004. Polygon Kriging. Surpac Minex Group, Perth Western Australia
6850. http://www.surpac.com/refman/default/stats/pkrig.htm
Swanson, R.T., Ritz, R. L. , McAtee, M.D. 2001. Operational Mesoscale Data
Assimilation at Air Force Weather Agency Using a Parallelized Multivariate Optimal
Interpolation Scheme. PSU/NCAR Mesoscale Modeling System Users’ Workshop 2001.
http://www.mmm.ucar.edu/mm5/workshop/ws01/swanson.doc
Syed Abdul Rahman Shibli. 1997. Spatial Correlation: General. AI-GEOSTATS.
http://www.ai-geostats.org/Geostats_Faq/Syed/sc_general.html
Taghvakish, S. and Amini, J., 2004. Optimum Weight in Thin-plate Spline for Digital
Surface Model Generation. TS26 Positioning and Measurement Technologies and
Practices II. FIG Working Week, 2004. Athens, Greece, May 22-27, 2004.
http://www.fig.net/pub/athens/papers/ts26/TS26_6_Amini_Taghvakish.pdf
Tait, A., and Zheng, X. 2005. Final Report: Optimal Mapping and Interpretat
ion of Fire
Weather Information, prepared for New Zealand Fire Service. NWA Client Report:
WLG2005-1, June 39, 2005. National Institute of Water andAtmospheric Research (NIWA)
Project: NZF03301 / Year 2.http://www.fire.org.nz/research/reports/reports/Report_49.pdf
Tangmar, B.B., Yost, R.S., and Uehara, G., 1985. Application of Geostatistics to Spatial
Studies of Soil Properties. Advanced Agronomics, 38:45-94.
Taylor, G.H., Daly, C., W.P. Gibson, andSibul-Weisberg, J. 1997. Digital andMap Products
Produces using PRISM. Proceedings 10th AMS Conference on Applied Climatology,
American Meteorological Society 20-24 October 1997,Reno, Nevada pp. 217-218
Ten Berge, H.F.M., Stroosnijder, L., Burrough, P.A., Bregt, A.K., and de Haes, M.J.
1983. Spatial Variability of Soil Properties Influencing the Temperature of the Soil
Surface, Agricultural Water Management, 6:213-226.
Thomas, G.S., 1997. Interactive Analysis and Modeling Semi-Variograms. Snowden
Associates Pty Ltd. Australia.
http://www.metal.ntua.gr/msslab/MineIT97/papers/GT67/GT67.html.
Thornton, P.E., Running, S.W., and White, M.A. 1997. Generating Surfaces of Daily
Meterological Variable over Large Regions of Complex Terrain. Numerical
Terradynamics Simulation Group, School of Forestry, University of Montana, Missoula.
Journal of Hyrdology 190 (1997) 214-251.
http://www.cgd.ucar.edu/tss/aboutus/staff/thornton/pubs/thornton_jh_1997.pdf
Thurston, J. (editor), 2003. Looking Back and Ahead: The Triangulated Irregular Network
(TIN). GeoInformatics, October. http://www.vectorone.info/Publish/tin.pdf
Tobler, W., 1979: A Transformational View of Cartography. The American Cartographer,
6, 101-106.
Todini, E., and Ferrarcsi, M., 1996. Influence of Parameter Estimation Uncertainty in
Kriging. Journal of Hydrology 175: 555-566.
Tolosana-Delgado, R., and Pawlowsky-Glahn. 2006. Kriging Coordinates: What Does that
Mean? Departament d’Informàtica i Matemàtica Aplicada Universitat de Girona, E-17071
,
Girona, Spain.http://ima.udg.es/Activitats/CoDaWork03/paper_Tolosana_Pawlowsky.pdf
Tomczk, M., 1998. Spatial Interpolation and it Uncertainty Using Automated Anisotropic
Inverse Distance Weighting (IDW) - Cross-Validation/Jackknife Approach. Journal of
Geographic Information and Decision Analysis, Vol. 2, No. 2: 18-33.
ftp://ftp.geog.uwo.ca/SIC97/Tomczak/Tomczak.html
Tsai, V.J.D., 1993. Delaunay triangulations in TIN creation: An overview and a linear-
time algorithm. Int. J. Geographic Information. Science, 7~6!, 501-524.
universal...................................43, 70
weights...........................................72 mode...................................................31
model evaluation ................................84
kriging algorithm................................69 Monte Carlo simulation .......................79
kriging and remotely sensed images.. 77 Monte-Carlo.......................................63
kriging quality....................................18 Moran’s I......................................65, 66
kriging with external drift.................. 71 multi-Gaussian kriging........................41
kurtosis ........................................ 27, 29 multi-indicator kriging .......................76
multimodal .........................................32
L
multiple samples .................................37
lag ......................................................51
N
lag class..............................................52
lag class variable................................52 ND......................................................37
lag distance ............................ 45, 50, 51 nearest neighbor .................................20
lag distance tolerance...................52, 53 negative skew.....................................28
lag increment................................ 46, 52 normal.................................................41
lag spacing..........................................52 normal distribution.............................28
lag tolerance.......................................52 normal probability plot.................27, 29
latitude/longitude............................... 51 normal QQ plots .................................40
left skewed...................................28, 31 normally distributed ............................40
leptokurtic.......................................... 29 normally distributed data....................30
less than detection.............................. 37 not detected ........................................37
linear ............................................ 21, 57 nugget effect.......................................54
linear/sill ............................................ 57 nugget effect in different directions.....55
local.................................................... 10 nugget variance ..................................60
local relative variograms.....................67
O
local trend........................................... 70
lognormal kriging .............................. 75 objective analysis .........................15, 24
probability of exceedence..................74
process................................................18 aniistrophic.....................................60
correlogram ....................................63
projection........................................... 51 covariance ......................................63
pronunciation of kriging ..................... 23 cross-K ...........................................63
punctual kriging............................ 70, 71 directional.......................................60
drift.................................................64
Q
exponential...................57, 58, 66, 82
QQ plot .............................................. 31 fractal .............................................64
quadratic ............................................ 21 Gaussian .............................57, 58, 66
quality of geostatistics/kriging...........18 Geary’s C .................................65, 66
quantile .............................................. 32 general relative variograms .............67
quantile-quantile plot ............. 27, 29, 31 hole effect.......................................59
holesin ............................................59
R
hyperbolic.......................................58
r 34, 35 InvCov............................................65
r2 34, 35 inverted covariance ........................65
Radarsat..............................................17 istrophic....................................60, 61
range ......................................32, 53, 57 linear ........................................57, 65
rank order transform ..........................77 linear/sill.........................................57
regionalized variable...........................18 local relative variograms .................67
regression kriging ............................... 77 madogram.......................................65
regression line....................................35 Matheron ........................................58
regularized spline................................ 20 Moran’s I ..................................65, 66
rejection ............................................. 39 omnidirectional ..............................60
relative variograms ............................. 67 pairwise relative variograms............67
remotely sensed images .....................77 relative variograms..........................67
replacement........................................39 Ripley’s K ................................63, 68
skewed distributions
skewed left ..........................31
.........................................31 arcsine.............................................40
back ................................................43
skewed right.......................................31 Box-Cox ...................................40, 42
skewness ............................................ 28 lognormal........................................41
social phenomena ................................9 normal score ...................................41
spatial autocorrelation. 9, 18, 45, 56, 58, offset...............................................42
60, 65, 66, 68, 69 power..............................................40
spatial dependency.............................45 scale................................................42
spatial independence............................9 square root ......................................42
spatial outliers..............................39, 74 standard back ..................................43
spatial pattern .....................................39 weighted back.................................43
sphere of influence.............................58 transGaussian......................................40
spherical....................................... 57, 58 trend............................43, 62, 64, 70, 74
splines ................................................ 20 global..............................................70
square root transform..........................42 local................................................70
standard back-transform ..................... 43 trend removal......................................40
standard deviation........................ 27, 32 trend surface .................................21, 43
standard error of estimate ........ 27, 34, 35 cubic...............................................21
standarized variograms.......................68 linear ..............................................21
stationarity .................................... 27, 40 quadratic.........................................21
stochastic ........................................... 10 trend surface analysis .........................70
stochastic simulation ...........................79 trend versus anomalies .......................39
storm depth ........................................12 triangular irregular network .........16, 21
strata...................................................39 triangular mesh...................................21
stratification.................................38, 39 trimodal ..............................................31
subset ................................................. 39 truncated gaussian simulation ............82
truncated plurigaussian simulation.....82