Copulas in Stochastic Processes Analysis
Copulas in Stochastic Processes Analysis
vorgelegt von
Diplom-Mathematiker
Volker Schmitz
aus Düren
This thesis was written during my time at the Institute of Statistics of Aachen University.
I would like to thank Professor Hans-Hermann Bock for giving me the opportunity and
freedom to choose this topic, his valuable comments, continuous interest and support for
this work. I am very grateful for having been given the opportunity to work in this highly
interesting and motivating field.
Many thanks to Professor B. Rauhut for also taking the effort to referee this thesis.
A special thankyou goes to Dr. E. Pier-Ribbert who drew my attention to the interesting
field of copulas during my stay at Dresdner Bank, Frankfurt.
I would also like to thank my colleagues at the Institute of Statistics of whom many con-
tributed to this thesis in one way or another. In particular, I have to mention Stefan Merx
who provided me with many useful hints and critical remarks and Eric Beutner for his
proof-reading of some of the chapters.
Last, I want to thank my wife Fiona for her patience and perseverance in supporting me
during the past years.
1 Introduction 1
3 Dependence Measures 33
3.1 Comonotonicity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.2 Concordance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.3 Kendall’s τ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
3.4 Spearman’s ρ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.5 Tail Dependencies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
3.6 Quadrant Dependence and Positive Likelihood Ratio Dependence . . . . . 39
Bibliography 103
Index 109
1
Introduction
The modelling of dependence relations between random variables is one of the most widely
studied subjects in probability theory and statistics. Thus, a great variety of concepts for
dependence structures has emerged (see Jogdeo, 1982). However, without specific assump-
tions about the dependence, no meaningful stochastic model can be developed.
The dependence of random variables can basically be classified into spatial and temporal
dependencies where the former means the dependence between a number of variables at the
same time whereas the latter means the intertemporal dependence structure of a process.
The classic approach uses second order moments of the underlying random variables,
known as the covariance. It is well known that only linear dependence relationships can
be captured this way and that it is characterizing only for special classes of distributions,
for example the normal distribution. The question arises if there is a possibility to capture
the whole dependence structure without any disturbing effects coming from the marginal
distributions. This is exactly where copulas provide a beneficial approach.
As the details will be given in the next chapter we will only say that copulas al-
low to separate the effect of dependence from the effects of the marginal distributions,
i. e., if X1 , . . . , Xn are real-valued random variables on some probability space (Ω, A, P)
with joint distribution function H, i. e., H(x1 , . . . , xn ) := P(X1 ≤ x1 , . . . , Xn ≤ xn ),
xi ∈ , 1 ≤ i ≤ n, a copula C of (X1 , . . . , Xn ) is given by the relation H(x1 , . . . , xn ) =
C(F1 (x1 ), . . . , Fn (xn )) for all xi ∈ where the Fi are the univariate marginal distributions
of the Xi , i. e., Fi (x) := P(Xi ≤ x), x ∈ , 1 ≤ i ≤ n.
Simply put, a copula is a multivariate distribution function on the n-dimensional unit
cube with uniform marginal distributions.
The question of existence and uniqueness of such a copula was answered by Sklar (1959)
who also introduced the term “copula” which is now the most widely used term. (“Copula”
is derived from the Latin verb “copulare”, meaning “to join together”.) Also used are the
terms “dependence function” (e. g., Galambos, 1978, Definition 5.2.1) in the context of
2 1. Introduction
multivariate extreme value theory, “uniform representation” (e. g., Kimeldorf and Sampson,
1975) and “standard form” (e. g., Cook and Johnson, 1981).
Although the concept is rather old, the first volume of the Encyclopedia of Statistical
Sciences to contain the entry “copulas” was the Update Volume from 1997 (Fisher, 1997).
This may be due to the fact that from 1959 to 1974 results concerning copulas were ob-
tained in the course of the development of the theory of probabilistic metric spaces and
distributions with given marginals (for a historic account see Schweizer, 1991). Nowadays,
however, interest in copulas has grown especially in mathematical finance; spatial depen-
dence between assets is tried to be captured by copulas instead of the classic models such
as correlation structures. For an example of this approach see Embrechts et al. (1999) or
Schmitz and Pier-Ribbert (2001).
Most of the current research in copulas is done for spatial dependence due to great interest
in practice for new spatial dependence models, but the analysis of temporal dependence is
also possible by the copula approach. The first paper dealing exclusively with copulas and
stochastic processes was presented by Darsow et al. (1992), who established the connection
between copulas and Markov processes. Only few papers have appeared since then, and
most are working papers not yet published in a journal.
This thesis extends the relationship between copulas and stochastic processes into new di-
rections and tries to show that the copula approach can provide new structural insights into
the temporal dependence underlying the processes. It will become clear that many prop-
erties of stochastic processes can be analyzed by splitting conditions up into a “marginal
part” and a “dependence part”, thus providing a better insight into the laws governing the
temporal behaviour of the process.
Many known theorems can be given alternative and in most cases simpler proofs via
copulas.
Although the focus is laid upon univariate stochastic processes, we also will—in a small
excursion—deal with two special and interesting aspects of copulas for spatial dependence
(Chapter 4).
After this general Introduction, Chapter 2 (Copulas and Their Properties) will give
a basic introduction to the definition and theory of copulas. It will be a basic reference for
all other chapters to follow. A special focus should be laid upon Section 2.2 (The Role of
Partial Derivatives) in which the connection between conditional distribution functions and
partial derivatives of copulas will be developed in a mathematical rigorous way. Although
many of the results are well known, it is difficult to find mathematically satisfying proofs
for them.
As the family of Archimedean copulas is the most prominent of all currently used models,
we give a short introduction to them in Section 2.3.
As a preparation for the analysis of symmetric stochastic processes we need to introduce
the survival copula and some symmetry concepts for multivariate random variables in
1. Introduction 3
Section 2.4. However, this section is of its own interest. The most important result is
that radial symmetry of a n-variate distribution function can be equivalently established
by radial symmetry of the underlying copula and (univariate) symmetry of the marginals
(Corollary 2.41).
Chapter 5 deals with the main topic of this thesis: the connection between univariate
stochastic processes and copulas. As we will look at a rather wide field of different concepts,
an introduction to definitions used in the sequel is given in Section 5.1.
Section 5.2 introduces basic relations between copulas and stochastic processes. Concepts
such as strict stationarity, equivalence of stochastic processes and symmetry of processes
are characterized by copulas.
Section 5.3 will specialize on processes in continuous time, e. g., Brownian motion. We
develop sufficient conditions to the univariate marginals and the copula structure of a pro-
cess to ensure stochastic continuity (Section 5.3.1). These conditions are then used to give
an elegant proof of the findings of Cambanis (1991) for EFGM-processes. He showed that
a stochastic process based on multivariate EFGM-distributions cannot be stochastically
continuous. However, he established this result by a number of inequalities whereas we are
able to give a direct proof with the help of Theorem 5.30 in Theorem 5.35.
Section 5.4 (Markov Processes) introduces the results of Darsow, Nguyen, and Olsen
(1992) and extends the findings to a characterization of symmetric Markov processes by
radial symmetry of the bivariate copulas (Section 5.4.2). Interestingly, within the class of
Archimedean copulas, symmetric Markov processes can be characterized by the so-called
Frank copula.
Finally, the copula structure of continuous local martingales is derived in Section 5.5.
As continuous local martingales are not defined by their finite-dimensional distributions,
a different technique has to be used to determine the underlying copulas. We show that
4 1. Introduction
the copula of Brownian motion plays a central role as each continuous local martingale
is equivalent to a time change of Brownian motion, a result due to Dambis (1965) and
Dubins and Schwarz (1965). It is then possible, for example, to derive the copula of the
Ornstein-Uhlenbeck process and the Brownian bridge. These can be used to construct
and/or simulate processes which have the same intertemporal dependence structure but
with other marginal distributions.
In the Appendix we will provide some additional information about basic tools used
within this thesis which are not worth being dealt with in the main part. A special focus
is laid on the simulation of random variables and processes from copulas, since in practice
simulation from copulas is one of the main tools to analyze the underlying models.
The List of Symbols and Abbreviations tries to give the reader a central point
of access for all symbols used within the text. Where appropriate, the page number is
indicated where the corresponding symbol is defined or introduced.
I tried to used notations and symbols which seem to be standard in the field concerned,
the only speciality is the symbol “” which marks the end of an example or remark.
Concerning the Bibliography, two things should be mentioned. Firstly, the numbers at
the end of each entry refer to the page on which the work is cited in this thesis. Secondly,
it is clear that the average reader does not necessarily have access to every source in the
references. However, I decided to refer to the literature wherein the exact formulation of
the needed theorems can be found, but it is possible to read the text with a single reference,
e. g., Rogers and Williams (2000a,b).
Finally, the Index should certainly help the reader find what he or she is looking for. If
page numbers behind entries are emphasized (such as 4711 in contrast to 4711), they refer
to the most important pages concerning these entries. It usually means that the definition
can be found there.
2
Copulas and Their Properties
In this chapter we give a basic introduction to copulas and their properties. As there seems
to be a lack of literature connecting copulas and conditional probabilities in a mathematical
rigorous way, we also shed some light on this aspect of copulas which turns out to be a
very powerful tool when dealing with the modelling of stochastic processes with copulas.
Let us define some notions first. We denote the extended real line by , i. e., :=
n
∪ {−∞} ∪ {∞}, analogously, denotes the extended n-dimensional real space.
n
For two vectors x = (x1 , . . . , xn ), y = (y1 , . . . , yn ) ∈ , we will write x ≤ y if xi ≤ yi ,
1 ≤ i ≤ n, and x < y if xi < yi for all 1 ≤ i ≤ n.
1 All random variables in this thesis will be assumed to be real-valued if not stated otherwise.
6 2. Copulas and Their Properties
2.1 Definition
n
A (halfopen) rectangle or interval in is the Cartesian product of n one-dimensional
intervals of the form
n
Rxy := ×(xi , yi ] = (x1 , y1 ] × · · · × (xn , yn ], x ≤ y,
i=1
n
where x = (x1 , . . . , xn ), y = (y1 , . . . , yn ) ∈ .
n
The set of all rectangles in will be denoted by Rn .
The vertices of a rectangle Rxy ∈ Rn (of which we clearly have 2n ) are the points
vert(Rxy ) := {(u1 , . . . , un ) : ui ∈ {xi , yi }, 1 ≤ i ≤ n}.
The unit cube I n is the product I × · · · × I (n times) where I = [0, 1] is the unit interval.
The domain and range of a function H will be denoted by dom(H) and ran(H), respec-
tively.
As we will use n-dimensional Lebesgue-(Stieltjes-)integration, we need a n-dimensional
analogue of an increasing2 function (i. e., some prototype of a measure generating function).
This can be established via the n-dimensional volume of a rectangle as we will see later.
with
(n) (n−1) (1)
∆ba H(x1 , . . . , xn ) := ∆(an ,bn ) ∆(an−1 ,bn−1 ) · · · ∆(a1 ,b1 ) H(x1 , . . . , xn ) (2.2)
where
(i)
∆(ai ,bi ) H(x1 , . . . , xn ) := H(x1 , . . . , xi−1 , bi , xi+1 , . . . , xn ) − H(x1 , . . . , xi−1 , ai , xi+1 , . . . , xn )
(2.3)
is the first order difference operator. ∆ba H(x1 , . . . , xn ) is therefore the n-th order difference
of H with regard to a and b.
2.3 Remark
The above notation is used to clarify on which coordinate the operator ∆ operates. As this
is notationally inconvenient, we will use the notation
(i)
∆baii ≡ ∆(ai ,bi )
where the coordinate is clear from the context. Where confusion could arise, we will use
the full notation.
2 Note that the terms “increasing” and “decreasing” mean f (x) ≤ f (y), x ≤ y, for a function f and f (x) ≥ f (y), x ≤ y,
respectively. If a strict inequality holds, we will use the term “strictly in-/decreasing”.
2.1 Basic Properties 7
The next definition is intuitive in the sense that the notion of groundedness is a natural
requirement for a multivariate distribution function on n .
for all xj ∈ Sj , 1 ≤ j ≤ n, j 6= i, ai ∈ Si .
We can now introduce the definition of a copula:
and:
If all coordinates of u ∈ I n are 1 except uk , then C(u) = uk . (2.6)
VC (Rab ) ≥ 0. (2.7)
We will now present some simple but basic properties of copulas. All these properties
hold for the general case n ≥ 2, although formulated and proved for the case
n = 2. The general case can be proved analogously, but a rigorous treatment would be
notationally complex.
2.7 Lemma
Let C ∈ C2 and 0 ≤ x1 ≤ x2 ≤ 1, then
is increasing on I.
Similarly, for 0 ≤ y1 ≤ y2 ≤ 1, the function
is increasing on I.
Proof. We only consider the former case as the latter one follows analogously. Let 0 ≤ y 1 ≤
y2 ≤ 1 and 0 ≤ x1 ≤ x2 ≤ 1. As C is 2-increasing (see Def. 2.4), we have (with x = (x1 , x2 ),
y = (y1 , y2 ))
2.8 Corollary
Copulas are Lipschitz-continuous (hence continuous), i. e., for any C ∈ C2 and 0 ≤ x1 ≤
x2 ≤ 1, 0 ≤ y1 ≤ y2 ≤ 1, and x, y ∈ I, it holds that
2.9 Corollary
Let C ∈ C2 , then y 7→ C(x, y) is increasing for all x ∈ I and analogously, x 7→ C(x, y) is
increasing for all y ∈ I.
Proof. Take x1 = 0 and y1 = 0, respectively, in Lemma 2.7.
2.1 Basic Properties 9
2.10 Remark
The set Cn of all n-copulas is a compact and convex subset of the space of all continuous
real valued functions defined on I n under the topology of uniform convergence. It follows
that—in Cn —pointwise convergence implies uniform convergence.
The convexity follows directly by the definiton of a copula. The compactness follows by
noticing that the set of all continuous real-valued functions on I n is a compact metric space
and the subset of all n-copulas is closed.
Two functions, of which only one is a copula for all dimensions (see Remark 2.12), are
special in that they provide bounds for all copulas:
The functions W and M are called lower and upper Fréchet bounds.
2.12 Remark
The multivariate version of Theorem 2.11 is given by
n
!
X
W (u) := max ui − n + 1, 0 ≤ C(u) ≤ min (ui ) =: M (u) (2.12)
1≤i≤n
i=1
for u = (u1 , . . . , un ) ∈ I n , n ≥ 2, C ∈ Cn .
Although the lower bound W is never a copula4 for n ≥ 3, the left-hand inequality in
(2.12) is “best-possible” in the sense that for any n ≥ 3 and any u ∈ I n , there is a n-copula
C such that C(u) = W (u) (see Nelsen, 1999, Theorem 2.10.12).
Qn
By Π(u) := i=1 ui , u ∈ I n , we denote the so-called independence copula.
For a stochastic interpretation of the functions M , W and Π we need the following classic
result of Sklar (1959, 1973) which shows the relationship between copulas and random
variables and their distribution functions, respectively. This will explain the interest in the
study of copulas which also underlies this thesis.
4 This can be seen from the fact that V ((1/2, 1]) = 1 − n/2 with (1/2, 1] := (1/2, 1] × · · · × (1/2, 1] so that W is not
W
grounded.
10 2. Copulas and Their Properties
W 1 1 1
M
0 0 0
1 1 1
v
v
1 1 1
00 u 00 u 00 u
Sfrag replacements
0.5 0.5 0.5
FIGURE 2.1. Lower Fréchet bound, independence copula and upper Fréchet bound; perspective
and related contour plots for the case n = 2.
In Figure 2.1, plots of the Fréchet bounds and the independence copula are given. The
geometric interpretation of Theorem 2.11 is that the surface of any copula of two random
variables must lie between the surfaces of the lower and the upper Fréchet bound. Every
kind of dependence between two random variables has a corresponding copula with a surface
which can be identified in the unit cube.
2.14 Remark (cf. Nelsen, 1999, Section 2.5)
With the theorem of Sklar, the copulas W , Π and M have the following stochastic inter-
pretations:
Two random variables U, V distributed uniformly on (0, 1) have the joint distribution
function
2.1 Basic Properties 11
• M iff (if and only if) U is almost surely (a. s.) an increasing function of V :
U = f (V ) for some monotone, increasing function f : → ,
We will now see that property (2.7) guarantees that C induces a Lebesgue-Stieltjes-
measure on (I n , B n ∩ I n ), where B n denotes the Borel-σ-algebra.
2.16 Theorem
Let C be a n-copula. Then C induces a unique probability measure PC on (I n , B n ∩ I n ).
Proof. As C is a distribution function on I n , the fact follows easily from Shiryaev (1996,
Theorem II.3.2, p. 160) or Billingsley (1995, Example 1.1, p. 9).
However, it is not true that every probability measure P on I n is induced by a copula
C.5 For a measure to be induced by a copula, it must spread mass in a manner consistent
with the boundary conditions of a copula. Therefore, any probability measure on I n where
at least one one-dimensional margin is not uniformly distributed cannot be induced by a
copula.
In detail, we have:
5 This means that for a rectangle Ru being spanned by 0 and u (both in I n ), there is always a n-copula C such that
0
P(R0u ) = C(u) holds for all u ∈ I n .
12 2. Copulas and Their Properties
2.17 Corollary
A probability measure P on (I n , B n ∩ I n ) is induced by a copula C ∈ Cn if and only if
P (0, xi ] = xi , for all xi ∈ I, 1 ≤ i ≤ n,
Proof. If P is induced by a copula C we must have P (0, xi ] = C(1, . . . , 1, xi , 1, . . . , 1) = xi
due to (2.6) for all xi ∈ I. On the other hand, if we have P (0, xi ] = xi , we can define the
copula C by C(u) := P((0, u]), u ∈ I n , and easily check the definition.
The following theorem concerning the denseness of the subset of absolutely continuous
copulas in the space of all copulas was firstly stated for the case n = 2 in Darsow et al.
(1992). For a proof, the authors refer to an oral presentation given at their institute. To
make things more rigorous, we state the theorem for the general case and give a detailed
proof.
As copulas are uniformly continuous on I n , the topology of uniform convergence and
its induced supremum metric dsup provide a natural framework for measuring distances
between copulas.
For notational convenience, λn will be used for the usual Lebesgue measure on ( n , B n )
and the Lebesgue measure restricted on I n , i. e., (I n , B n ∩ I n ). The meaning depends on
the context; where confusion could arise, additional remarks will be made.
2.18 Theorem
Let Cn denote the set of all n-copulas and Cλn ⊂ Cn the subset of all copulas which are
absolutely continuous with respect to (w. r. t.) the Lebesgue measure λn restricted on I n .
Then Cλn is dense in Cn w. r. t. the supremum metric dsup , i. e., the metric generating the
uniform topology on I n .
i1 − 1 i1 in − 1 in
Si1 ,...,in := , × ··· × , , 1 ≤ ij ≤ N, ij ∈ , 1 ≤ j ≤ n, (2.14)
N N N N
2.1 Basic Properties 13
and let X
RiN1 ,...,in := Sj1 ,...,jn (2.16)
1≤jk ≤ik
1≤k≤n
be the n-dimensional rectangle defined by the points (0, . . . , 0) and (i1 /N, . . . , in /N ).
N is chosen such that N > n/ε.
Now, for 1 ≤ i1 , . . . , in ≤ N , let C in a first step be defined on the points (i1 /N, . . . , in /N )
by
i1 in N ∗ i1 in
C ,..., := PC ∗ (Ri1 ,...,in ) = C ,..., . (2.17)
N N N N
Note that the C-volume of any rectangle of the form R = (0, i1 /N ] × · · · × (0, in /N ] can
be determined by (2.16) from these values, i.e.,
X
VC (RiN1 ,...,in ) = VC (Sj1 ,...,jn ).
1≤jk ≤ik
1≤k≤n
n
Recall that using the notation for the n-th order difference operator ∆ba for a, b ∈ we
have
(i /N,...,in /N )
VC (Si1 ,...,in ) = ∆((i11 −1)/N,...,(i n −1)/N )
C(u1 , . . . , un ),
and this is well defined by (2.17).
In a second step, let us now define C by a λn -density in the following way which guar-
antees that the mass of C is spread uniformly on the cubes Si1 ,...,in :
dC X
n
(x , . . . , x ) ≡ c(x , . . . , x ) := N Si1 ,...,in (x) · VC (Si1 ,...,in ), (2.18)
1 n 1 n
dλn 1≤i ,...,i ≤N
1 n
and we have
N N N N
C(ui ) = C , . . . , , ui , , . . . ,
N N N N
Z1 Zui Z1 X
(2.18)
= ... ... Nn Si1 ,...,in (x) · VC (Si1 ,...,in ) dλn (x)
0 0 0 1≤i1 ,...,in ≤N
ji −1
X hZ1 ZN Z1
n n
= N VC (Si1 ,...,in ) ... ... Si1 ,...,in (x) dλ (x)
1≤i1 ,...,in ≤N 0 0 0
Z1 Zui Z1 i
n
+ ... ... Si1 ,...,in (x) dλ (x)
0 ji −1 0
N
X n
n 1
= N VC (Sk1 ,...,kn ) ·
1≤k1 ,...,kn ≤N
N
ki ≤ji −1
X n−1
n ji − 1 1
+ N VC (Sk1 ,...,kn ) ui −
1≤k1 ,...,kn ≤N
N N
ki =ji
X X
ji − 1
= VC (Sk1 ,...,kn ) + N ui − VC (Sk1 ,...,kn )
1≤k1 ,...,kn ≤N
N 1≤k ,...,k 1 n ≤N
ki ≤ji −1 ki =ji
(2.17) ∗ ji − 1 ji − 1
= C (1, . . . , 1, , 1, . . . , 1) + N ui −
N N
∗ ji ∗ ji − 1
· C (1, . . . , 1, , 1, . . . , 1) − C (1, . . . , 1, , 1, . . . , 1)
N N
C ∗ is copula ji − 1 ji − 1 ji ji − 1
= + N ui − − = ui .
N N N N
2.1 Basic Properties 15
Therefore, C is a copula.
Therefore,
(2.10) (2.19)
≤ 2kx − v ∗ k1 < ε,
thus
kC − C ∗ k < ε,
which completes the proof.
2.19 Remark
The same construction idea is used for the bivariate case by Li et al. (1998) and Kulpa
(1999), who call it a “checkerboard approximation”.
16 2. Copulas and Their Properties
with ∆(h) := FY (y+h)−FY (y) wherever the derivative exists. This is the basic idea behind
the interpretation of partial derivatives which will be put in a more rigorous framework
hereafter.
Analogously, by Lemma 2.7 and Corollary 2.8, the functions in Lemma 2.7 have deriva-
tives almost everywhere, and
Dk C(u1 , . . . , un )u =y − Dk C(u1 , . . . , un )u =x ≥ 0 (2.22)
k k
if 0 ≤ x ≤ y ≤ 1, ui ∈ (0, 1).
{X≤x}
2.2 The Role of Partial Derivatives 17
2.20 Notation
As we will need the operator Dk quite often, we will use the notation
∂
Dk C(u1 , . . . , uk−1 , x, uk+1 , . . . , un ) := C(u1 , . . . , un ) . (2.23)
∂uk uk =x
It follows that the functions x → Dk C(u1 , . . . , uk−1 , x, uk+1 , . . . , un ) are defined and
increasing almost everywhere.
Before formulating a lemma which relates the partial derivatives to conditional expecta-
tions, we need the following result from Elstrodt (1996) which—in connection with (2.21)
and the above remarks—gives us the integrability of the partial derivatives.
2.21 Theorem (Elstrodt, 1996, Satz 6.1, p. 150)
Let K ∈ { , } and f : [a, b] → K (a, b ∈ p , a < b, p ∈ ) be a bounded function.
Then f is Riemann-integrable if and only if the set of its points of discontinuity has λp -
measure 0. Then, the Riemann-integral and the Lebesgue-integral are identical.
We can now formulate the basic result connecting partial derivatives of copulas with
conditional distribution functions.
2.22 Lemma
Let X and Y be two real-valued random variables on the same probability space (Ω, A, P)
with corresponding copula C and continuous marginals FX , FY . Define the conditional
probabilities
{X≤x} {Y ≤y}
Proof. (Cf. Darsow et al., 1992) We will only consider the first case, as the second one is
completely analogous. Let σ(Y ) denote the σ-Algebra generated by Y and consider the
Y
setting Ω, σ(Y ) −→ ( 1 , B 1 ). We have to show
Z Z Z
!
P(X ≤ x | Y ) dP = {X≤x} dP = D2 C FX (x), FY (Y ) dP (2.26)
A A A
R
R all A ∈ σ(Y ), x ∈ . With the definitions Qx (A) := A P(X ≤ x | Y ) dP and Q̃x (A) :=
for
A
D2 C FX (x), FY (Y ) dP for A ∈ σ(Y ) we therefore must show
7 See Billingsley, 1995, Sections 34 and 35, or Rogers and Williams, 2000a, II.40–42.
18 2. Copulas and Their Properties
Note that Qx is well defined due to the definition of conditional expectation and Q̃x is
well defined due to (2.21) and Theorem 2.21. It is elementary to show that Qx and Q̃x are
σ-finite measures on σ(Y ).
It is known that E := {(−∞, a] : a ∈ } is a σ-finite generator of B 1 w. r. t. the
Lebesgue-measure λ1 which is closed under intersections. From Bauer (1992, Satz 5.4) or
Rogers and Williams (2000a, Corollary II.4.7) it follows that we only need to show (2.27)
for a σ-finite generator of σ(Y ) which is closed under intersections. We now have
(∗) (∗∗)
σ(Y ) = Y −1 (B 1 ) = Y −1 σ(E) = σ Y −1 (E) ,
where (∗) is valid due to Billingsley (1995, Theorem 20.1, p. 255) and (∗∗) due to Elstrodt
(1996, Satz 4.4, p. 19).
(∗) is due to Kamke (1960, Satz 1, p. 164) with η = FY (ξ) and the continuity of FY in
combination with Lemma A.2.
This is exactly (2.27) from which the assertion follows.
2.23 Remark
In view of Lemma A.3 we see that FX (X), FY (Y ) ∼ U(0, 1). Thus, (2.25) is equivalent to:
D2 C FX (x), V is a version of P(X ≤ x | Y ) and
(2.28)
D1 C U, FY (y) is a version of P(Y ≤ y | X),
where U, V ∼ U(0, 1) have copula C and are defined on the same probability space as X
and Y . This means that for P(X ≤ x | Y ), say, we only need the dependence information
via C and the marginal FX .
A statement analogous to Lemma 2.22 holds for the general case of n random variables
and can be formulated as follows:
2.2 The Role of Partial Derivatives 19
2.24 Corollary
Let X1 , . . . , Xn be real-valued random variables on the same probability space (Ω, A, P)
with corresponding copula C and continuous marginals F1 , . . . , Fn .8 Then, for any k ∈
{1, . . . , n},
Dk C F1 (x1 ), . . . , Fk−1 (xk−1 ), Fk (Xk ), Fk+1 (xk+1 ), . . . , Fn (xn )
is a version of (2.29)
P Xi ≤ xi , 1 ≤ i ≤ n, i 6= k | Xk
for all x1 , . . . , xn ∈ .
The next goal should be to connect partial derivatives of n-copulas to expressions such
as P(Xn ≤ xn | X1 , . . . , Xn−1 ). This would be particularly helpful when dealing with
modelling conditionally specified time series by copulas. Indeed we can show more than
Lemma 2.22: If the higher partial derivatives of a n-copula exist and are continuous, they
are related to conditional probabilities in an analogous way as exemplified in Lemma 2.22.
Before stating the corresponding lemma let us introduce a notation for higher-order partial
derivatives of n-copulas as a generalization of (2.20):
2.25 Notation
Let C be a n-copula. For k ∈ {1, . . . , n}, we define
∂k
D1,...,k C(x1 , . . . , xn ) := C(u1 , . . . , un ) (2.30)
∂u1 · · · ∂uk (x1 ,...,xn )
for all (x1 , . . . , xn ), (u1 , . . . , un ) ∈ (0, 1)n where the derivative exists. For k = n, we write
as this is the probability density function of (U1 , . . . , Un ) where the Ui ∼ U(0, 1) are con-
nected by the copula C.
More generally, let J ⊂ {1, . . . , n} be such that J = {j1 , . . . , jk }, ji 6= jl for i 6= l. Then,
∂k
DJ C(x1 , . . . , xn ) := C(u1 , . . . , un ) . (2.31)
∂uj1 ∂uj2 · · · ∂ujk (x1 ,...,xn )
2.26 Remark
Due to our convention, expressions such as D1,2 C(0.2, 0.2, 0.4) make sense and are well
defined.
8 Note that for a specific k we only need the continuity of the marginal Fk for the statement to hold.
20 2. Copulas and Their Properties
We can formulate the following theorem for uniformly distributed random variables. The
general expression will follow as a corollary.
2.27 Theorem
Let U = (U1 , . . . , Un ) be n real-valued random variables with uniform marginal distri-
butions U(0, 1), n ∈ , and let C1,...,n = ChU i, C1,...,n−1 = ChU1 , . . . , Un−1 i where Ch · i
denotes the corresponding copula. Assume that C1,...,n is (n − 1) times partially differen-
tiable w. r. t. the first (n − 1) arguments and the partial derivatives are continuous in these
arguments. Then,
Proof. The proof is essentially analogous to the one of Lemma 2.22. However, care has
to be taken of the fact that, e. g., D1,...,n−2 C1,...,n (0, u2 , . . . , un ) is not necessarily equal to
0.9 As we will see, we do not need this either. We will follow the outline of the proof of
Lemma 2.22.
Let σ(U1 , . . . , Un−1 ) denote the σ-algebra generated by U1 , . . . , Un−1 . We need to show
that
Z Z
D1,...,n−1 C1,...,n (U1 , . . . , Un−1 , un )
P(Un ≤ un | U1 , . . . , Un−1 ) dP = dP
D1,...,n−1 C1,...,n−1 (U1 , . . . , Un−1 )
A A
for all un ∈ (0, 1), A ∈ σ(U1 , . . . , Un−1 ). As En−1 := {(0, a] : a = (a1 , . . . , an−1 ) ∈ (0, 1)n−1 }
is a σ-finite generator of B n−1 ∩ (0, 1)n−1 which is closed under intersections, we can assume
A to be of the form
A = U1−1 (0, a1 ] ∩ · · · ∩ Un−1
−1
(0, an−1 ]
9 This
Qn
can be easily seen by Π(u1 , . . . , un ) = i=1 ui : We have D1,...,n−2 Π(0, u2 , . . . , un ) = un un−1 which is greater than
0 for u1 , u2 ∈ (0, 1).
2.2 The Role of Partial Derivatives 21
Z
D1,...,n−1 C1,...,n (U1 , . . . , Un−1 , un )
dP
D1,...,n−1 C1,...,n−1 (U1 , . . . , Un−1 )
A
Z
D1,...,n−1 C1,...,n (u1 , . . . , un−1 , un )
= dC1,...,n−1 (u1 , . . . , un−1 )
D1,...,n−1 C1,...,n−1 (u1 , . . . , un−1 )
(0,a]
Z
(∗)
= D1,...,n−1 C1,...,n (u1 , . . . , un−1 , un ) du1 · · · dun−1
(0,a]
Z
Fubini
= D1,...,n−2 C1,...,n (u1 , . . . , un−2 , an−1 , un )
(0,a1 ]×···×(0,an−2 ]
− D1,...,n−2 C1,...,n (u1 , . . . , un−2 , 0, un ) dun−2 · · · du1
Z
= D1,...,n−3 C1,...,n (u1 , . . . , un−3 , an−2 , an−1 , un )
(0,a1 ]×···×(0,an−3 ]
Pn−1
where δ(b1 , . . . , bn−1 ) := i=1 {0} (bi ) counts the number of zeros.
The following corollary is proven completely analogously so that we will omit its proof.
2.28 Corollary
Let X1 , . . . , Xn , n ∈ , be real-valued, continuous random variables with distribution
functions F1 , . . . , Fn and copula C1,...,n . For k ∈ {1, . . . , n}, assume that C1,...,n is k-times
partially differentiable w. r. t. the first k arguments and the partial derivative is continuous
in these arguments, i. e., D1,...,k C1,...,n (u1 , . . . , un ) exists and is continuous in u1 , . . . , uk .
Further, let C1,...,k be the copula of (X1 , . . . , Xk ) and c1,...,k the associated density func-
tion, i. e., c1,...,k (u1 , . . . , uk ) = D1,...,k C1,...,k (u1 , . . . , uk ). The marginal densities fi (x) :=
22 2. Copulas and Their Properties
d
F (x)
dx i
are assumed to exist for 1 ≤ i ≤ k. Then,
D1,...,k C1,...,n F1 (X1 ), . . . , Fk (Xk ), Fk+1 (xk+1 ), . . . , Fn (xn )
Qk
c1,...,k F1 (X1 ), . . . , Fk (Xk ) i=1 fi (Xi )
(2.33)
is a version of
P(Xk+1 ≤ xk+1 , . . . , Xn ≤ xn | X1 , . . . , Xk )
for all xk+1 , . . . , xn ∈ .
2.29 Remark
It would be possible to formulate an even more general form by allowing mixed partials.
I. e., for I ⊂ {1, . . . , n} we could express P(Xi ≤ xi , i ∈ I | Xj , j ∈ {1, . . . , n} \ I) with
partials of copulas (cf. equation (2.31) on page 19). As this is notationally inconvenient,
we will omit this.
10 Note that ϕ will also stand for the density of a standard normal random variable. As it is common to denote the generator
by ϕ we will stick to that convention as the meaning is always clear from the context.
2.3 Archimedean Copulas 23
2.31 Remark
a) The original proof of this theorem was given by Alsina et al. (2001). As this mono-
graph is not yet available, we refer the reader to the proof in Nelsen (1999).
b) If ϕ(0) = ∞, ϕ is called a strict generator and ϕ[−1] = ϕ−1 , i. e., the usual inverse.
The induced copula is also called strict.
c) The generator ϕ of an Archimedean copula is not unique, as cϕ, c > 0, yields the
same copula as ϕ.
2.32 Example
a) For ϕ(t) = − ln(t), t ∈ (0, 1], we have ϕ(0) = limt&0 ϕ(t) = ∞, therefore ϕ is strict
with ϕ−1 (t) = exp(−t). The induced copula is given by
C(u, v) = exp(−[− ln u − ln v]) = uv = Π(u, v).
Thus, the independence copula is a strict Archimedean copula.
b) The lower Fréchet bound W (u, v) = max(0, u + v − 1), u, v ∈ I, is Archimedean:
Choosing ϕ(t) = 1 − t, 0 ≤ t ≤ 1, yields W .
However, the upper Fréchet bound M is not Archimedean. This is due to the result
that the diagonal δC (u) := C(u, u) of any Archimedean copula C meets the condition
δC (u) < u for all u ∈ (0, 1) (which in combination with associativity11 is characteriz-
ing for Archimedean copulas, see Ling, 1965). As δM (u) = u, this condition cannot
be met by M .
c) Setting ϕϑ (t) = − ln exp(−ϑt)−1
exp(−ϑ)−1
, ϑ ∈ \ {0}, yields ϕ(0) = ∞ so that a strict
copula family given by
1 (exp(−ϑu) − 1)(exp(−ϑv) − 1)
Cϑ (u, v) = − ln 1 + , ϑ ∈ \ {0}, (2.36)
ϑ exp(−ϑ) − 1
is generated. This is the Frank family which will prove to be important in the context
of symmetric Markov processes (see Section 5.4.2).
Many classes of Archimedean copulas and some of their properties are described in Nelsen
(1999, pp. 94–97).
[−1] −1 −1
For 0 ≤ t ≤ ϕ(0) we have
ϕ (t) = ϕ (t), so that C(u, v) = ϕ ϕ(u) + ϕ(v) which
is equivalent to ϕ C(u, v) = ϕ(u) + ϕ(v) for ϕ(u) + ϕ(v) ≤ ϕ(0). So we get
∂
ϕ C(u, v) = ϕ0 C(u, v) · D1 C(u, v) = ϕ0 (u) and
∂u
∂
ϕ C(u, v) = ϕ0 C(u, v) · D2 C(u, v) = ϕ0 (v),
∂v
11 A 2-copula C is associative if C(C(u, v), w) = C(u, C(v, w)) holds for all u, v, w ∈ I.
24 2. Copulas and Their Properties
thus
ϕ0 (u)
D1 C(u, v) = ,
ϕ0 ϕ−1 (ϕ(u) + ϕ(v))
ϕ0 (v)
D2 C(u, v) = 0 −1
ϕ ϕ (ϕ(u) + ϕ(v))
The level curve of a copula is the set Lt ≡ L(t) := {(u, v) ∈ I 2 : C(u, v) = t}, t ∈ I. For
an Archimedean copula, L(t) = {(u, v) ∈ I 2 : ϕ(t) = ϕ(u) + ϕ(v)} ⊂ I 2 which connects
the points (1, t) and (t, 1) (as ϕ(1) = 0). We can write the level curve as Lt (u) = v since
solving for v as a function of u yields
v = Lt (u) = ϕ[−1] ϕ(t) − ϕ(u) = ϕ−1 ϕ(t) − ϕ(u)
where the last step is justified because ϕ(t) − ϕ(u) is in the interval [0, ϕ(0)).
It can be shown (Nelsen, 1999, Theorem 4.3.2, p. 39) that the level curves Lt (u) are
convex for all t ∈ [0, 1).
An important theorem concerning applications such as estimation (cf. Genest and Rivest,
1993) and random variate generation (see Section B.2) can be obtained by finding the C-
measure of the region in I 2 lying on or below and to the left of each level curve.
ϕ(t)
Kϕ (t) = t − . (2.37)
ϕ0 (t+)
The following corollary presents a probabilistic interpretation of the above theorem which
will be useful in random variate generation (it is also useful for dealing with Kendall’s τ
for Archimedean copulas).
2.4 Survival Copula and Symmetry Concepts 25
The next theorem which is due to Genest and Rivest (1993) is an extension of Corol-
lary 2.34; an application is the algorithm for generating random variates from distributions
with Archimedean copulas (see Section B.2).
2.35 Theorem
Under the hypotheses of Corollary 2.34, the joint distribution function H(s, t) of
S = ϕ(U )/ ϕ(U ) + ϕ(V ) and T = C(U, V ) is given by s · KC (t) for all s, t ∈ I. Hence S
and T are independent, and S is uniformly distributed on (0, 1).
2.36 Remark
a) The function Kϕ (z) = P(C(U, V ) ≤ z), z ∈ I, can be estimated by means of its em-
(n)
pirical distribution function Kϕ (z): Given a sample {(x1 , y1 ), . . . , (xn , yn )} of size
n from
P (X, Y ), construct the empirical bivariate distribution function Hn (x, y) :=
n−1 ni=1 (−∞,x] (xi ) · (−∞,y] (yi ), x, y ∈ , of the sample. Then, calculate Hn (xi , yi ),
(n)
b) Kϕ (z) can be fitted by the distribution function Kϑ (z) of any family of Archimedean
copulas, where the parameter ϑ is estimated in such a manner that the fitted distri-
bution has a coefficient of concordance (Kendall’s τ ) equal to the empirical coefficient
(see Genest and Rivest, 1993).
But what does it mean to say that (X, Y ) is “symmetric” about some point (a, b) ∈ 2 or
more generally that (X1 , . . . , Xn ) is “symmetric” about (a1 , . . . , an ) ∈ n ? In the multivari-
ate setting there is a number of possibilities in which sense random vectors are symmetric.
We will shortly introduce three of them for the bivariate case and then generalize the most
important one (radial symmetry) to n dimensions which will be a key tool for characterizing
symmetric processes.
c) (X, Y ) is jointly symmetric about (a, b) if the following four pairs of random vectors
have a common joint distribution:
(X − a, Y − b), (X − a, b − Y ), (a − X, Y − b), (a − X, b − Y ).
2.38 Remark
a) For X and Y we can express radial symmetry in terms of the joint distribution and
the univariate survival functions of X and Y in a manner analogous to (2.39) (see
below).
b) In the above definition we have c)⇒b)⇒a). It can be shown that jointly symmet-
ric random variables must be uncorrelated when the second order moments exist
(Randles and Wolfe, 1979). Thus, the concept is too strong for many applications
and we will focus on radial symmetry.
c) Marginal symmetry does not imply an intuitive bivariate symmetry as can be seen
from Figure 2.2. Although the margins are symmetric N(0, 1) and t3 -variables, respec-
tively, the bivariate density does not seem to be symmetric due to the unsymmetric
dependence structure.
2.4 Survival Copula and Symmetry Concepts 27
3 −3
2 −2
1 −1
0.10 0 0.10 0
x
0.05 −1 0.05 1
0.00
−3 −2 −2 0.00
−3 −2 2
−1 0 −1 0
x 1 2 −3 y 1 2 3
3 3
−3 3
−2 2
−1 1
0.10 0 0.10 0
y
x
0.05 1 0.05 −1
0.00
3 2 0.00
3 −2
2 1 2 1
0 −1 −2 3 0 −1 −2 −3
x −3 y −3
1.0
0.8
0.6
0.4
0.2
0.0
FIGURE 2.2. Bivariate distribution based on the min-max copula with parameter 3 and N(0, 1)-
and t3 -distributed marginals (perspective and contour plots).
28 2. Copulas and Their Properties
H(a − x, b − y)
(a − x, a − y)
(a, b)
PSfrag replacements
(a + x, b + y)
H(a + x, b + y)
FIGURE 2.3. Regions of equal probability for radially symmetric random variables.
Note that the marginals of H are the univariate survival functions, i. e., H(x, −∞) = F (x)
and H(−∞, y) = G(y).
If we define
b v) := u + v − 1 + C(1 − u, 1 − v),
C(u, u, v ∈ I, (2.43)
we have
b F (x), G(y) .
H(x, y) = C
b is called the survival copula of C (it is easy to check that C
C b is indeed a copula).
Note that C b is not the joint survival function C of two U(0, 1) distributed random
variables with copula C as we have C(u, v) := P(U > u, V > v) = 1 − u − v + C(u, v) =
b − u, 1 − v).
C(1
Now we want to define the n-variate survival copula C b of (X1 , . . . , Xn ) such that
b F 1 (x1 ), . . . , F n (xn ) .
H(x1 , . . . , xn ) = P(X1 > x1 , . . . , Xn > xn ) = C (2.44)
We can derive it with the help of the inclusion-exclusion formula (Billingsley, 1995, p. 24):
Let (Ω, A, P) be a probability space and A1 , . . . , An ∈ A. Then,
n
! n
[ X X X
P Ak = P(Ak ) − P(Ai ∩ Aj ) + P(Ai ∩ Aj ∩ Ak ) + · · ·
k=1 k=1 1≤i<j≤n 1≤i<j<k≤n
An equivalent representation due to the fact that all marginals Ci1 ,...,ik are derivable from
C = C1,...,n is given in Georges et al. (2001, Theorem 2), which states that
b 1 , . . . , un ) = C(1 − u1 , . . . , 1 − un )
C(u (2.47)
with
n
X X
C(u1 , . . . , un ) = (−1)k C(v1 , . . . , vn )
k=0 v∈Z(n−k,n,1;u)
X Pn
= i=1 εi
(−1) C ε1 u1 + (1 − ε1 ), . . . , εn un + (1 − εn ) .
(ε1 ,...,εn )∈{0,1}n
Pn
where Z(m, n, ε; u) := {v ∈ I n : vi ∈ {ui , ε}, 1 ≤ i ≤ n, {ε} (vk ) = m}.
k=1
b 1 , u2 , u3 ) (2.46)
C(u = u1 + u2 + u3 − 2 + C1,2 (1 − u1 , 1 − u2 ) + C1,3 (1 − u1 , 1 − u3 )
+ C2,3 (1 − u2 , 1 − u3 ) − C1,2,3 (1 − u1 , 1 − u2 , 1 − u3 )
(2.47)
= C(1, 1, 1) − C(1 − u1 , 1, 1) − C(1, 1 − u2 , 1) − C(1, 1, 1 − u3 )
+ C(1 − u1 , 1 − u2 , 1) + C(1, 1 − u2 , 1 − u3 ) + C(1 − u1 , 1, 1 − u3 )
− C(1 − u1 , 1 − u2 , 1 − u3 ).
We are now ready to derive the following corollary from Theorem 2.40.
2.41 Corollary
A random vector (X1 , . . . , Xn ) with copula C and univariate survival functions F i , 1 ≤ i ≤
n, is radially symmetric about (a1 , . . . , an ) ∈ n iff
a) F i (ai − xi ) = Fi (ai + xi ) for all xi ∈ , 1 ≤ i ≤ n, i. e., all the Xi are symmetric
about ai , and
b 1 , . . . , un ) for all (u1 , . . . , un ) ∈ ran(F1 ) × · · · × ran(Fn ).
b) C(u1 , . . . , un ) = C(u
Proof. By Theorem 2.40, radial symmetry is equivalent to
H(a1 + x1 , . . . , an + xn ) = H(a1 − x1 , . . . , an − xn ) ⇐⇒
b F 1 (a1 − x1 ), . . . , F n (an − xn ) .
C F1 (a1 + x1 ), . . . , Fn (an + xn ) = C
Thus, the sufficiency of a) and b) is clear. For the necessity, we observe that in the case of
radial symmetry
Fi (ai + xi ) = H(∞, . . . , ∞, ai + xi , ∞, . . . , ∞)
radial symm.
= H(−∞, . . . , −∞, ai − xi , −∞, . . . , −∞) = F i (ai − xi ) (2.48)
2.4 Survival Copula and Symmetry Concepts 31
for all xi ∈ .
Now, let (u1 , . . . , un ) ∈ ran(F1 ) × · · · × ran(Fn ). Then there exist xi ∈ such that
ui = Fi (ai + xi ), 1 ≤ i ≤ n. Thus,
C(u1 , . . . , un ) = C F1 (a1 + x1 ), . . . , Fn (an + xn ) = H(a1 + x1 , . . . , an + xn )
radial symm. (2.44)
= H(a1 − x1 , . . . , an − xn ) = C b F 1 (a1 − x1 ), . . . , F n (an − xn )
(2.48)
= C b F1 (a1 + x1 ), . . . , Fn (an + xn ) = C(u
b 1 , . . . , un )
In this chapter we will give a short introduction to some dependence measures commonly
used in multivariate settings. Basic references are Nelsen (1999) and Joe (1997). It is obvious
that we are interested in dependence measures which are closely related to copulas.
3.1 Comonotonicity
Two random variables X and Y are said to be comonotonic if a random variable Z exists
such that
X = f (Z), Y = g(Z) a. s.,
where f and g are two increasing real-valued functions.
Comonotonicity is an extension of the concept of (perfect) positive correlation to vari-
ables with arbitrary distributions, i. e., while correlation of two random variables is often
understood as linear dependence, we may here have any monotone relation between the
variables.
We note that the upper Fréchet bound for copulas is reached if X and Y are comono-
tonic, the lower bound is reached if X and Y are countermonotonic, i.e., if X and −Y
are comonotonic (see Remark 2.14 and Embrechts et al., 1999). If X and Y are perfectly
positively correlated, then X and Y are comonotonic; the converse is generally not true.
3.2 Concordance
In the following we will need the notion of concordance. Informally, a pair of random
variables X and Y is concordant if “large” values of one tend to be associated to “large”
values of the other variable, and analogously for “small” values. To be precise, let (X, Y )
be a vector of two random variables and (x1 , y1 ), (x2 , y2 ) two samples from (X, Y ). We
will say that (x1 , y1 ) and (x2 , y2 ) are concordant if (x1 − x2 )(y1 − y2 ) > 0, the samples are
discordant if (x1 − x2 )(y1 − y2 ) < 0. It is clear that we could define concordance of two
34 3. Dependence Measures
vectors (x1 , y1 ) and (x2 , y2 ) without using random variables. However, in our context there
is no need for such a generalization.
Note that P(X1 = X2 ) = 0 if X1 and X2 are continuous. Therefore, if X and Y above are
continuous, the regions of concordance and discordance split the sample space as a subset
of 2 into two non-intersecting regions whose P-measure is 1.
3.3 Kendall’s τ
3.1 Definition (Kendall’s τ )
Let (X1 , Y1 ) and (X2 , Y2 ) be two independent and identically distributed random vectors on
some probability space (Ω, A, P). Kendall’s τ is defined as
τ ≡ τ (X1 , Y1 ) ≡ τX1 ,Y1 := P (X1 − X2 )(Y1 − Y2 ) > 0 − P (X1 − X2 )(Y1 − Y2 ) < 0 . (3.1)
We will also call this the population version of Kendall’s τ .
Thus, Kendall’s τ is just the difference between the probability of concordance and the
probability of discordance.
As one may suppose, there is also a“sample version”of this dependence measure, based on
a sample from (X, Y ) ∼ H. Let {(x1 , y1 ), . . . , (xn , yn )}, n ≥ 2, denote a sample of n obser-
vations from a vector (X, Y ) of continuous random variables. Each pair {(xi , yi ), (xj , yj )},
i, j ∈ {1, . . . , n}, i 6= j, is either discordant or concordant
as there are no ties in the sample
with probability one. There are obviously n2 distinct pairs of observations in the sam-
ple. Let c denote the number of concordant pairs, d the number of discordant pairs. Then
Kendall’s τ for the sample (the above mentioned “sample version”) is defined to be
c−d c−d
τ̂ ≡ t := = n . (3.2)
c+d 2
If we choose a pair
of observations (xi , yi ) and (xj , yj ) randomly from the sample (for
n
which there are 2 possibilities which are equally likely) τ̂ can be interpreted as the
probability that we get a concordant pair minus the probability that we get a discordant
pair.
We can express concordance and discordance with indicator functions A (x) for Borel
(xi , yi ) and (xj , yj ) are discordant ⇔ (−∞,0) (xi − xj )(yi − yj ) = 1.
(0,∞)
for discordance. Therefore, the number of concordant values c (x, y) in a sample (x, y) =
{(x1 , y1 ), . . . , (xn , yn )} is given by
n
1 X
c (x, y) = (xi − xj )(yi − yj ) .
(0,∞)
2 1≤i<j≤n
The next theorem is the basis for connecting copulas with Kendall’s τ .
3.4 Spearman’s ρ 35
3.2 Theorem
Let (X, Y ) be a random vector with continuous marginals, i. e., X is a continuous random
variable with distribution function F and Y is continuous with distribution function G.
Let (X1 , Y1 ) and (X2 , Y2 ) be two random vectors with distribution function H1 and H2 ,
respectively, where Xi ∼ F and Yi ∼ G, i = 1, 2. Assume that all random variables are
defined on a probability space (Ω, A,P).
Further, H1 (x, y) = C1 F (x), G(y) and H2 (x, y) = C2 F (x), G(y) , C1 and C2 denoting
the underlying copulas. If Q denotes the difference between the probabilities of concordance
and discordance of (X1 , Y1 ) and (X2 , Y2 ),
Q := P (X1 − X2 )(Y1 − Y2 ) > 0 − P (X1 − X2 )(Y1 − Y2 ) < 0 ,
we have: ZZ
Q = Q(C1 , C2 ) = 4 C2 (u, v) dC1 (u, v) − 1. (3.3)
I2
where U and V are U(0, 1) random variables with copula C and EC denotes the expectation
w. r. t. the measure PC induced by the copula.
Proof. Nelsen (1999, p. 129).
3.4 Spearman’s ρ
Spearman’s ρ is another nonparametric measure of dependence defined in terms of concor-
dance and discordance.
3.4 Definition (Spearman’s ρ)
Let (X1 , Y1 ), X2 , and Y2 be independent random variables and vector, respectively, where
Xi ∼ F , Yi ∼ G, i = 1, 2, and (X1 , Y1 ) has joint distribution function H. All random
variables are defined on (Ω, A, P).
The population version of Spearman’s ρ is defined to be proportional to the probability of
concordance minus the probability of discordance for the two vectors (X 1 , Y1 ) and (X2 , Y2 ):
ρ ≡ ρ(X1 , Y1 ) ≡ ρsX1 ,Y1 := 3 · P (X1 − X2 )(Y1 − Y2 ) > 0 − P (X1 − X2 )(Y1 − Y2 ) < 0 .
(3.5)
36 3. Dependence Measures
We can then formulate the following relationship between the underlying copula C and
Spearman’s ρ. For a proof see Nelsen (1999, pp. 135).
3.5 Theorem
Let X and Y be continuous random variables with copula C. Then
ZZ ZZ
s s
ρX,Y ≡ ρC = 12· uv dC(u, v)−3 = 12 C(u, v) du dv −3 = 12·E[C(U, V )]−3, (3.6)
I2 I2
The following concept was firstly established in the works of Lehmann (1966) and
Esary et al. (1967).
P(X ≤ x | Y ≤ y)
3.7 Remark
Note that we have the following equivalent relations. They directly follow from elementary
calculations:
We can give a sufficient condition for ρ(X, Y ) ≥ τ (X, Y ) ≥ 0, which is a feature often
observed for positively dependent random variables.
3.5 Tail Dependencies 37
It is now obvious to ask for a translation of these definitions and results into the “copula-
language”. We must observe that these definitions are not necessarily independent of the
marginals which can be seen from the form P(X ≤ x | Y ≤ y) = C(F (x), G(y))/G(y).
However, we are able to give sufficient conditions for two random variables X and Y to
be RTI and LTD based only on their copula(s)1 by using the next lemma and showing the
relations a) or b) for all u, v ∈ (0, 1).
3.9 Lemma
Let X and Y be two (not necessarily continuous) random variables on some probability
space (Ω, A, P) with marginals F and G, respectively, and copula C.2 Let IF := (0, 1) ∩
ran(F ) and IG := (0, 1) ∩ ran(G). Then we have the following relations:
Before we provide the proof, let us remark that we do not need to include the points
u = 0 or u = 1 as the copula is either 0 or v which results in constant and insofar increasing
functions.
Proof. We will only show the first part as the second one can be shown analogously. For
the “if”-part, assume C(u, v)/v to be a decreasing function in v ∈ IG for all u ∈ IF . Take
x ∈ , y 1 , y2 ∈ with y1 < y2 . Define u := F (x), v1 := G(y1 ) and v2 := G(y2 ). Then
v2 ≥ v1 due to the monotonicity of G. By assumption, we have
C(u, v1 ) C(u, v2 ) C F (x), G(y1 ) C F (x), G(y2 )
≥ ⇔ ≥
v1 v2 G(y1 ) G(y2 )
⇔ P(X ≤ x | Y ≤ y1 ) ≥ P(X ≤ x | Y ≤ y2 ).
3.10 Remark
a) If the marginals F and G are continuous, we have ran(F ) = ran(G) = I and therefore
a unique copula C. This is due to the fact that in the proof of Sklar’s theorem, the
copula is defined on the points ran(F ) × ran(G) and interpolated for all other points.
However, the interpolation is not unique. Thus, if ran(F ) × ran(G) = I 2 , the copula
can be uniquely defined.
The next dependence concept relates to the amount of dependence in the upper quadrant
or lower quadrant tail of a bivariate distribution. It is particularly relevant to dependence
in extreme values and for the modeling of consecutive extremal events. In contrast to τ
and ρ, upper and lower tail dependence is a local dependence measure in the sense of being
defined for some interesting areas (the upper and lower quadrants) instead of being defined
globally.
Like Kendall’s τ and Spearman’s ρ tail dependence is a copula property. The amount of
tail dependence is therefore invariant under strictly increasing transformations of X and
Y.
3.11 Definition (upper and lower tail dependence)
Let X and Y be random variables with cumulative distribution functions (cdf ) F X and FY ,
respectively. The coefficient of upper tail dependence is then defined as
lim P Y > FY−1 (u) | X > FX−1 (u) =: λU (3.8)
u%1
if λU ∈ [0, 1] exists. If λU ∈ (0, 1], X and Y are said to be asymptotically dependent in the
upper tail. If C is the copula associated to X and Y we obtain
C(u, u) 1 − 2u + C(u, u)
λU = lim = lim , (3.9)
u%1 1 − u u%1 1−u
3.6 Quadrant Dependence and Positive Likelihood Ratio Dependence 39
where C(u, v) = 1 − u − v + C(u, v) is the joint survival function. The coefficient of lower
tail dependence is analogously defined as
C(u, u)
λL := lim P Y ≤ FY−1 (u) | X ≤ FX−1 (u) = lim (3.10)
u&0 u&0 u
where the limit exists.
λ{L,U } is extensively used in extreme value theory for describing the property that one
variable is extreme given the other one is extreme. Note that λU = limu%1 P(U1 > u | U2 >
u) = limu%1 P(U2 > u | U1 > u).
In this chapter we will give two examples of the use of copulas for the analysis of spatial
dependence, i. e., of (at least) two random variables at the same time.
The first application deals with the depencence between the minimum and the maximum
of n iid random variables. We will derive the corresponding copula and calculate Kendall’s τ
and Spearman’s ρ for (X(1) , X(n) ). We can finally show that the relation 3τn ≥ ρn ≥ τn > 0
holds in this situation.
As a second application, we will take a closer look at the joint distribution of Brownian
motion (Bt )t∈[0,∞) (see Definition 5.22) and its supremum process (St )t∈[0,∞) where St :=
sup0≤s≤t Bt . By deriving the copula we see that the dependence structure of Bt and St does
not depend on time, i. e., although we have to start with the derivation of Ct as a function
of time, the copula will prove to be time-independent.
where Fs (x) := P(X(s) ≤ x) is the distribution function of the s-th order statistic:
X n
n n−i
Fs (x) = F i (x) 1 − F (x) . (4.2)
i=s
i
42 4. Two Examples: Analysis of Spatial Dependence by Using Copulas
Note that the conditions for the domain of the copula stem from solving F1 (x) > Fn (y).
We thus found the family of copulas describing the dependence structure of the minimum
and maximum of n independent random variables. To get a visual impression we provide
some plots of these copulas for different values of n in Figure 4.1.
4.1 The Min-Max Copula 43
n=3 n = 10 n = 50 n=∞
1.0
1.0
1.0
1.0
0.8
0.8
0.8
0.8
0.6
0.6
0.6
0.6
0.4
0.4
0.4
0.4
0.2
0.2
0.2
0.2
0.0
0.0
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
5 5 5 5
1.0 1.0 1.0 1.0
y
y
c(x,y)
c(x,y)
c(x,y)
c(x,y)
0.5 0.5 0.5 0.5
1.0
1.0
1.0
0.8
0.8
0.8
0.8
0.6
0.6
0.6
0.6
0.4
0.4
0.4
0.4
0.2
0.2
0.2
0.2
0.0
0.0
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
FIGURE 4.1. First row: contourplots of Cn (dotted line: contours of independence copula; main
diagonal added to provide better orientation), second row: corresponding densities; third and
fourth row: surface- and contourplots of (Cn − W )/(M − W ).
44 4. Two Examples: Analysis of Spatial Dependence by Using Copulas
We also note the interesting fact that this copula is closely related to a well known copula
family. From Theorem 2.15 we know that for a strictly decreasing function f and a strictly
increasing function g, the copula of (f (X(1) ), g(X(n) )) is given by
However, this is a member of the family of copulas discussed by Kimeldorf and Sampson
(1975), Clayton (1978), Cook and Johnson (1981, 1986), Oakes (1982, 1986), and implicitly
already by Mardia (1962) which is usually called Clayton or Cook and Johnson copula. For
ϑ ∈ [−1, ∞) \ {0} it is defined as
CϑCl (u, v) := max (u−ϑ + v −ϑ − 1)−1/ϑ , 0 , u, v ∈ (0, 1). (4.4)
Note that it is an Archimedean copula with generator ϕϑ (t) = (t−ϑ − 1)/ϑ (see Nelsen,
1999, pp. 94).
This observation can also be interpreted stochastically: Let X1 , . . . , Xn be iid random
variables and define Yi = (Y1,i , Y2,i ) := (−Xi , Xi ), 1 ≤ i ≤ n, and
Then
X(n) = max (−Xi ), max Xi = −X(1) , X(n) .
1≤i≤n 1≤i≤n
Note that X(n) = f (X(1) ), g(X(n) ) for f (x) = −x, g(x) = x. In view of the above remarks
the copula of the distribution of X(n) is given by
For further investigations, set An := {(u, v) ∈ (0, 1)2 : 1 − (1 − u)1/n < v 1/n } and
Bn := {(u, v) ∈ (0, 1)2 : 1 − (1 − u)1/n > v 1/n }. On these open sets, Cn is partially
differentiable twice, and elementary calculation yields
∂ (v 1/n + (1 − u)1/n − 1)n−1 · (1 − u)(1/n)−1 , (u, v) ∈ An ,
D1 Cn (u, v) := Cn (u, v) =
∂u 0, (u, v) ∈ Bn ,
(4.6)
∂ 1 − (v 1/n + (1 − u)1/n − 1)n−1 · v (1/n)−1 , (u, v) ∈ An ,
D2 Cn (u, v) := Cn (u, v) =
∂v 1, (u, v) ∈ B ,
n
(4.7)
4.1 The Min-Max Copula 45
and
∂2
D12 Cn (u, v) = Cn (u, v) (4.8)
∂u∂v
n−1 (v(1 − u))(1/n)−1 (v 1/n + (1 − u)1/n − 1)n−2 , (u, v) ∈ An ,
n
= (4.9)
0, (u, v) ∈ Bn .
The next result has been known for a long time (Walsh, 1969, or David, 1981, p. 267),
but it is interesting to see its derivation with the help of copulas. Additionally, its proof is
simpler as, e. g., the ones in Reiss (1989, Section 4.2) or Galambos (1978, Theorem 2.9.1).
From an intuitive point of view, it is clear that the minimum and maximum of n iid random
variables should be asymptotically independent, i. e., we expect the limiting copula to be
the independence copula.
4.1 Lemma
Let Cn be as in (4.3). Then
and
lim Cn (u, v) = v − v(1 − u) = uv = Π(u, v) for (u, v) ∈ (0, 1)2 .
n→∞
4.2 Remark
The above proof is already short and non-technical in the sense that no other results
and only elementary calculus are used. However, it is known for the Clayton copula that
limϑ→0 CϑCl (u, v) = uv for all u, v ∈ (0, 1) (e. g., Cook and Johnson, 1981). As mentioned
Cl
above, v − Cn (1 − u, v) = C−1/n (u, v) so that
Cl
lim Cn (u, v) = v − lim C−1/n (1 − u, v) = v − lim CϑCl (1 − u, v)
n→∞ n→∞ ϑ→0
= v − (1 − u)v = uv.
Based on these findings we can start investigating some of the properties of the above de-
fined min-max copula related to dependence. We will start with the calculation of Kendall’s
τ (see page 34) with the help of Theorem 3.3.
4.3 Theorem
Let X1 , . . . , Xn , n ∈ , be iid random variables. Then Kendall’s τ for X(1) and X(n) is
given by
1
τn (X(1) , X(n) ) ≡ τn = . (4.11)
2n − 1
Proof. From (4.3) we know that the corresponding copula for (X(1) , X(n) ) is given by Cn .
Thus, with Theorem 3.3 we know that
ZZ
τn = 4 · Cn (u, v) dCn (u, v) − 1.
I2
= 4 I1 (n) − I2 (n) − 1
with
ZZ
n − 1 1/n n−2
I1 (n) := · v (1 − u)(1/n)−1 v 1/n + (1 − u)1/n − 1 · (u, v) du dv
An
n
I2
4.1 The Min-Max Copula 47
and
ZZ
n − 1 (1/n)−1 2n−2
I2 (n) := ·v (1 − u)(1/n)−1 v 1/n + (1 − u)1/n − 1 · (u, v) du dv.
An
n
I2
Z1 Z1
n−2
= (n − 1)s(1 − u)(1/n)−1 s + (1 − u)1/n − 1 sn−1 ds du
0
1−(1−u)1/n
Z1 Z1
= (n − 1)nsn (s + t − 1)n−2 tn−1 t1−n ds dt
0 1−t
Z1 Z1 Z1 Z1
(∗)
= (n − 1)n sn (s + t − 1)n−2 ds dt = (n − 1)n sn (s + t − 1)n−2 dt ds
0 1−t 0 1−s
Z1 Z1
n
t=1 1
=n s (s + t − 1)n−1 t=1−s ds = n s2n−1 ds = .
2
0 0
(∗) is valid by Fubini’s theorem (e. g., Rogers and Williams, 2000a, II. 12).
With the same substitutions, I2 (n) can be evaluated to
n−1
I2 (n) =
2(2n − 1)
so that
1 n−1 1
τn = 4 − −1= .
2 2(2n − 1) 2n − 1
Analogously, we can derive the following result for Spearman’s ρ (see p. 35) based on
Theorem 3.5.
4.4 Theorem
Let X1 , . . . , Xn , n ∈ , be iid random variables. Then Spearman’s ρ for X(1) and X(n) is
given by
n
12n X (−1)k 2n (n!)3
ρn (X(1) , X(n) ) ≡ ρn = 3 − 2n + 12 (−1)n . (4.12)
n k=0
2n − k n + k (3n)!
48 4. Two Examples: Analysis of Spatial Dependence by Using Copulas
Proof. With Theorem 3.5 and the sets An and Bn (see p. 44) we have
ZZ
ρn = 12 Cn (u, v) du dv − 3
2
I Z Z ZZ
1/n 1/n
n
= 12 v− v + (1 − u) −1 du dv + v du dv −3
A Bn
ZnZ ZZ
1/n 1/n
n
= 12 v du dv − v + (1 − u) −1 du dv − 3.
An ∪Bn A
|n {z }
=:I(n)
It is easy to see that the first integral yields 1/2. With the substitutions s = (1 − u)1/n and
t = v 1/n , the integral I(n) can be calculated as follows:
1−(1−v 1/n )n
Z1 Z Z1 Z1
n n
I(n) = v 1/n + (1 − u)1/n − 1 du dv = n v 1/n + s − 1 sn−1 ds dv
0 0 0 1−v 1/n
Z1 Z1 Z1 Z1
= n2 (t + s − 1)n sn−1 tn−1 ds dt = n2 tn−1 (t + s − 1)n sn−1 ds dt
0 1−t 0 1−t
n Z
X
1 Z1
2 n k n−1
=n (t − 1) t sn−k sn−1 ds dt
k=0
k
0 1−t
n Z
X
1
2 n k n−1 1 2n−k
=n (t − 1) t 1 − (1 − t) dt
k=0
k 2n − k
0
n
X k Z1 Z1
n (−1)
= n2 tn−1 (1 − t)k dt − tn−1 (1 − t)2n dt
k=0
k 2n − k
0 0
Xn
2 n (−1)k
=n [B(n, k + 1) − B(n, 2n + 1)] ,
k=0
k 2n − k
R1
where B denotes the Beta function B(x, y) := 0 tx−1 (1 − t)y−1 dt. With Gould (1972,
formula (1.43)) the second sum is reduced to
Xn n
2 n (−1)k (n − 1)! (2n)! 2 X n (−1)k n (2n)! n! (−1)n
B(n, 2n + 1) n = n =
k=0
k 2n − k (3n)! k=0
k 2n − k (3n)! n 2n n
(n!)3
= (−1)n .
(3n)!
Using B(n, k + 1) = (n − 1)! k!/(n + k)! for the first sum proves the result.
Sfrag replacements
4.1 The Min-Max Copula 49
1.0
0.8
Kendall’s τ
Spearman’s ρ n ρn τn ρn /τn
2 0.46 0.3 1.4
0.6
τ /ρ
0 10 20 30 40 50
Figure 4.2 and Table 4.1 reveal the fact that τn is smaller than ρn which is not surprising
in virtue of the results from Capéraà and Genest (1993) who showed that Spearman’s
ρ is generally larger than Kendall’s τ for positively dependent random variables in the
sense of left-tail decrease and right-tail increase (see Theorem 3.8). To be precise, they
showed that ρ ≥ τ ≥ 0 for two random variables X and Y whenever one of X or Y
is simultaneously left-tail decreasing (LTD) and right tail-increasing (RTI) in the other
variable (cf. Definition 3.6). Additionally, ρn /τn seems to converge to 3/2; however, this
has not yet been proven due to the unwieldy expression for ρn in equation (4.12). The
convergence to the value 3/2 of ρn /τn must be seen in a more general context: Nelsen (1991)
showed this relation for some copula families and posed the (yet unresolved) question, if
it is valid for all families of copulas {Cϑ (u, v)} indexed by a (possibly multidimensional)
parameter ϑ such that Cϑ0 = Π and Cϑ is a continuous function of ϑ at ϑ0 .
As it is not easy to see that ρn ≥ τn in the case of (X(1) , X(n) ) by the above results, we
will make use of Definition 3.14 and Theorem 3.15 to show that X(1) and X(n) are positively
likelihood ratio dependent so that this part of the the assertion follows.
4.5 Theorem
Let τn and ρn denote Kendall’s τ and Spearman’s ρ for (X(1) , X(n) ) in the above situation.
Then
3τn ≥ ρn ≥ τn > 0.
Proof. τn > 0 follows trivially from Theorem 4.3. If we show that PLR(X, Y ), the other
inequalities follow from Theorem 3.8, Theorem 3.13 and Theorem 3.15. In virtue of Def-
inition 3.14, let u1 , u2 , v1 , v2 ∈ (0, 1) and u1 ≤ u2 , v1 ≤ v2 . It is clear from (4.10) that
cn (x, y) = 0 for (x, y) ∈ \ {(0, 1)2 }. We have to show
with An = {(u, v) ∈ (0, 1)2 : 1 − (1 − u)1/n < v 1/n }. We may therefore assume (u1 , v2 ),
(u2 , v1 ) ∈ An , for if this is not the case, the right hand side of (4.13) is zero and the
inequality trivially true. As u1 ≤ u2 and v1 ≤ v2 , we must also have (u1 , v1 ), (u2 , v2 ) ∈ An .
But then (4.13) is equivalent to
1/n−1 h 1/n 1/n in
v1 (1 − u1 )v2 (1 − u2 ) v1 + (1 − u1 )1/n − 1 v2 + (1 − u2 )1/n − 1
1/n−1 h 1/n 1/n
1/n 1/n
in
≥ v1 (1 − u2 )v2 (1 − u1 ) v1 + (1 − u2 ) − 1 v2 + (1 − u1 ) − 1
1/n 1/n
⇔ v1 + (1 − u1 )1/n − 1 v2 + (1 − u2 )1/n − 1
1/n 1/n
≥ v1 + (1 − u2 )1/n − 1 v2 + (1 − u1 )1/n − 1 .
1/n 1/n
Setting x1 := u1 , x2 := u2 , y1 := 1 − (1 − v1 )1/n and y2 := 1 − (1 − v2 )1/n so that x1 ≤ x2
and y1 ≤ y2 , this inequality is equivalent to
4.6 Remark
The question may arise if it is not possible to derive similar results for (X(r) , X(s) ), (r, s) 6=
(1, n). However, for the derivation of Cn it was of fundamental importance to be able to
determine the generalized inverses of F1 and Fn which could be done via (4.2). It does
not seem to be possible to determine the generalized inverses of Fs for 1 < s < n analyti-
cally. This is an unsolved problem and prevents the copula approach from being used for
(X(r) , X(s) ).
4.2 The Copula of Brownian Motion and Its Supremum 51
We will be interested in the derivation of the copula CtBSup of (Bt , St ), t > 0. At this
stage we do not have any clues about its form; particularly, we do not have reason not to
expect the dependence structure to depend √ on the time-parameter−1t. √
As Bt ∼ N(0, t), we have FBt (x) = Φ(x/ t), x ∈ , such that FBt (u) = t · Φ−1 (u), 0 <
u < 1, where Φ denotes the cumulative distribution function of a standard normal random
variable. For the distribution function FSt of St , we have (e. g., Rogers and Williams, 2000a,
Corollary I.13.3 and the remark thereafter):
√
FSt (y) = P(St ≤ y) = 1 − 2 · P(Bt ≥ y) = 2 · Φ(y/ t) − 1, y ≥ 0, (4.14)
such that
√ v+1
FS−1 (v) = t · Φ−1 , v ∈ (0, 1). (4.15)
t
2
This yields
t · Φ−1 (v + 1)/2 ):
BSup 1 √ −1
Ct (u, v) = Φ √ · t · Φ (u)
t
1 √ −1
√ −1
(4.19)
−Φ √ t · Φ (u) − 2 · t · Φ (u + 1)/2
t
= u − Φ Φ−1 (u) − 2 · Φ−1 (v + 1)/2 ,
We see that neither of these expressions depends on the point of time t. We can thus say
that the dependence between Brownian motion and its supremum is entirely determined
by the marginal distributions and the time-independent copula
(
−1 −1
u − Φ Φ (u) − 2 · Φ (v + 1)/2 , u ≤ (v + 1)/2,
C BSup (u, v) = (4.21)
v, u > (v + 1)/2.
What could be the use of these findings? One possible application could be found in
mathematical finance where geometric Brownian motion is a stochastic process (still) used
to model stock prices. Now, if this model were adequate, we could calculate the supremum
process of the stock price process and transform both processes by the probability integral
transform. For the resulting two uniform processes, we could develop a test for homogeneity
w. r. t. time. Under the null hypothesis that the underlying model is geometric Brownian
motion, we should not be able to detect dependencies on time; conservative tests could be
developed by testing for a trend in the data (as a trend is a special case of time dependence).
We will leave these aspects to future research.
5
Stochastic Processes And Copulas
In this chapter we will investigate the relationship between copulas and univariate stoch-
astic processes. Before we explore these relationships we first give the basic definitions and
properties used in the following sections concerning stochastic processes. For a detailed
introduction, please refer to the “standard literature”, e. g., Ross (1996) or Todorovic (1992)
for a general account to stochastic processes, Karatzas and Shreve (2000) or Revuz and Yor
(1991) for an emphasis on Brownian motion and stochastic calculus; Rogers and Williams
(2000a,b) may deal as a general account to diffusions, Markov processes and martingales.
The basic idea is that many properties of stochastic processes can be characterized by
their finite-dimensional distributions so that copulas can be used for analysis. However,
many concepts in the theory of stochastic processes are stronger than the finite-dimensional
distribution approach; examples are martingales and stochastic differential equations which
rely heavily on pathwise structures. Nevertheless, it is possible to derive the copula structure
in some special cases.
From a stochastic point of view we are often only interested in the finite-dimensional
distributions of processes so that we declare two processes with the same finite-dimensional
distributions to be equivalent:
It is important to realize that this is only one version of a number of equivalence concepts
for stochastic processes. Others are modification and indistinguishability which both are
stronger than the finite-dimensional distribution concept (cf. Karatzas and Shreve, 2000,
p. 2). However, as copulas are distribution functions, we will focus on the finite-dimensional
distributions.
A valuable source for the relationship between these concepts and sufficient conditions
for the existence of certain modified versions of a given process with desired properties
(such as continuity or differentiability) is Cramér and Leadbetter (1967).
A certain flow of information is revealed by the concept of filtrations and adapted pro-
cesses:
(Ω, F , , P).
We will need right-continuity of a filtration for the “usual conditions” which will be used
for technical reasons in Section 5.5.
Stopping times play a prominent role in the technique of localizing a stochastic process
(see Revuz and Yor, 1991, pp. 117). I. e., many theorems in the context of stochastic anal-
ysis are valid if some conditions are met locally by the process, meaning essentially that
the stopped process has the desired properties.
A class of widely studied processes is the class of martingales. They may be thought of
as representing “fair games” as the expected value in the future based on “the present’s
point of view” is exactly the current value of the process.
lim Ft1 ,...,tn (x1 , . . . , xn ) = Ft1 ,...,tk−1 ,tk+1 ,...,tn (x1 , . . . , xk−1 , xk+1 , . . . , xn ) for all 1 ≤ k ≤ n,
xk %∞
(5.5)
for all n ∈ , x1 , . . . , xn ∈ , t1 , . . . , tn ∈ T . Then there exist a filtered probability space
lim Ct1 ,...,tn (u1 , . . . , un ) = Ct1 ,...,tk−1 ,tk+1 ,...,tn (u1 , . . . , uk−1 , uk+1 , . . . , un ) (5.7)
uk %1
P(Xt1 ≤ x1 , . . . , Xtn ≤ xn ) = Ct1 ,...,tn Ft1 (x1 ), . . . , Ftn (xn ) (5.8)
For a given process X = (Xt )t∈T , we will denote the associated family of copulas by
ChXi. In the same sense, ChXt1 , . . . , Xtn i ≡ CtX1 ,...,tn denotes a copula of (Xt1 , . . . , Xtn ).
Proof. The assertion follows directly from Theorem 5.10 and Theorem 2.13.
5.2 General Properties 57
5.13 Remark
There is little overlap between the usual concepts of stochastic processes (Markov pro-
cesses, martingales, time series analysis or diffusion processes) and the concept of mod-
elling processes via copulas. For example, time series analysis heavily relies on the con-
cept of autocorrelations whereas it is of great importance to see that the copula ap-
proach tries to capture the full intertemporal dependence structure in the data. However,
Darsow, Nguyen, and Olsen (1992) managed to relate Markov processes to copulas and
even to derive necessary and sufficient conditions for a process being Markov based on its
copula family. We will come to this interesting point later.
58 5. Stochastic Processes And Copulas
{t1 + h, . . . , tn + h} ⊂ T , we have
P(Xt1 +h ≤ x1 , . . . , Xtn +h ≤ xn ) = P(Xt1 ≤ x1 , . . . , Xtn ≤ xn ) (5.9)
for all x1 , . . . , xn ∈ .
5.15 Remark
It is clear (set n = 1 in the above definition) that all marginal distributions must be the
same, i. e., Xt ∼ F for all t ∈ T for some one-dimensional distribution function F .
The processes in Example 5.12 clearly yield strictly stationary processes for Xt ∼ F for
some arbitrary distribution function F .
5.17 Remark
If (S, S) = ( 1 , B 1 ), (5.10) is equivalent to
for all n ∈ , xi ∈ , 1 ≤ i ≤ n.
A fortiori, Xt and Yt must have the same distribution function Ft for all t ∈ T .
If CtX1 ,...,tn and CtY1 ,...,tn denote the copulas of (Xt1 , . . . , Xtn ) and (Yt1 , . . . , Ytn ), respectively,
we can give an equivalent characterization of stochastically equivalent processes:
5.18 Theorem
Let (Xt ) and (Yt ) be real-valued stochastic processes on (Ω, F , P) with Xt ∼ Ft , Yt ∼ Gt ,
t ∈ T , where Ft and Gt denote the distribution functions of Xt and Yt . For t1 < · · · < tn ,
n ∈ , set CtX1 ,...,tn = ChXt1 , . . . , Xtn i and CtY1 ,...,tn = ChYt1 , . . . , Ytn i.
a) Ft = Gt for all t ∈ T ,
Proof. For the “only if”-part, Remark 5.17 yields Ft = Gt for all t ∈ T . Further, assume
n
(u1 , . . . , un ) ∈ ran(Ft1 ) × · · · × ran(F tn ) so that there exists (x1 , . . . , xn ) ∈ such that
(u1 , . . . , un ) = Ft1 (x1 ), . . . , Ftn (xn ) . Thus,
CtX1 ,...,tn (u1 , . . . , un ) = CtX1 ,...,tn Ft1 (x1 ), . . . , Ftn (xn )
equivalence
= P(Xt1 ≤ x1 , . . . , Xtn ≤ xn ) = P(Yt1 ≤ x1 , . . . , Ytn ≤ xn )
Ft =Gt Y
= CtY1 ,...,tn Gt1 (x1 ), . . . , Gtn (xn ) = Ct1 ,...,tn Ft1 (x1 ), . . . , Ftn (xn )
= CtY1 ,...,tn (u1 , . . . , un ).
It should be clear that stronger concepts of equivalence do not relate to copulas in the
same fashion: For example, (Yt ) is a modification of (Xt ) if P(Xt = Yt ) = 1 for all t ∈ T , but
there is no connection to copulas for this relation. The same is true for indistinguishability,
i. e., P(Xt = Yt for all t ∈ T ) = 1, which is the strongest form of equivalence (almost all
sample paths coincide). For details, cf. Todorovic (1992, pp. 5).
60 5. Stochastic Processes And Copulas
d
Proof. We must have (Xt ) = (−Xt ). Take t1 < · · · < tn arbitrarily. Then the proof is
completely analogous to the one of Corollary 2.41.
1 Recall that a process (Xt ) is adapted to a filtration = (Ft )t∈T if each Xt is Ft -measurable for all t.
5.3 Continuous Time Stochastic Processes 61
a) X0 = 0 almost surely,
It follows from Karatzas and Shreve (2000, Theorem 5.12, p. 75) that this process is a
Markov process. We can now extract the underlying bivariate copula family ChXi.
It is now easy to derive the corresponding partial derivatives and the copula density. Ele-
mentary calculus yields
√ −1 √
B tΦ (v) − sΦ−1 (u)
D1 Cs,t (u, v) = Φ √ , (u, v) ∈ (0, 1)2 , (5.16)
t−s
r Zu √ √
B 1 t tΦ−1 (v) − sΦ−1 (w)
D2 Cs,t (u, v) = ϕ √ dw, (u, v) ∈ (0, 1)2 .
ϕ Φ−1 (v) t−s t−s
0
(5.17)
62 5. Stochastic Processes And Copulas
For the last result, an interchange of integration and differentiation is necessary which
holds from the “differentiation lemma” (e. g., Bauer, 1992, Lemma 16.2, p. 102).
Differentiation of (5.16) yields the density
r √ √ √
B t ϕ ( tΦ−1 (v) − sΦ−1 (u))/ t − s
cs,t (u, v) = , (u, v) ∈ (0, 1)2 , (5.18)
t−s ϕ Φ−1 (v)
for the Brownian copula.
After having derived the copula we are free to generate a new stochastic process (again
of Markovian nature as the Markov structure is characterized by the bivariate copulas, see
Theorem 5.44) with the same intertemporal dependence structure as Brownian motion but
with different marginals. An empirical example for this is given in Figure 5.1. Every path
is generated the following way:
• Choose a time grid (t1 , . . . , tn ) where ti > 0, ti < ti+1 .
• A realization (u1 , . . . , un ) from the Brownian copula process is generated on the time
grid (cf. Section B.3).
√
• For the path of standard Brownian motion, set bn := tn Φ−1 (un ) such that bn can
be interpreted as a realization of a normal random variable with µ = 0 and σ 2 = tn .
• For
q the path of the generalized Brownian motion with scaled tν -marginals, set dn :=
tn ν−2
ν
· Fν−1 (un ) where Fν−1 is the quantile function of a tν -distributed random
variable (cf. Section A.2, p. 94) and take ν = 3. The dn can then be interpreted as
realizations from a rescaled t3 distributed random variable Dn such that E(Dn ) = 0,
Var(Dn ) = tn .
Note that it is not possible to distinguish the standard Brownian motion from the general-
ized one only considering first and second moments of the marginals and the dependence
structure. This is a widely unrecognized danger in practice.
5.24 Remark
a) It is now straightforward to see that the copula of any strictly monotone increasing
transformation of a Brownian motion is exactly the Brownian copula (see Theo-
rem 2.15).
For example, a geometric Brownian motion X = (Xt )t∈T satisfies the stochastic
integral equation
Zt Zt
Xt = X0 + µ Xs ds + ν Xs dBs (5.19)
0 0
(Karatzas and Shreve, 2000, pp. 349) where µ ∈ , ν > 0, X0 > 0. Thus, Xt is just
a strictly increasing transformation of Bt by f (z) := X0 · exp (µ − 12 ν 2 )t + νz . The
copula of geometric Brownian motion is therefore given by equation (5.15).
However, this is only a special case of a related question: How can we derive the
family of copulas for general transformations of diffusions or—more general—local
(semi-)martingales? We will deal with this interesting point in Section 5.5.
2A Gaussian process is a process whose finite-dimensional distributions are all multivariate normal distributions.
64 5. Stochastic Processes And Copulas
1.0
4
0.8
3
0.6
2
B(t)
Ct
0.4
1
0
0.2
−1
0.0
0 2 4 6 8 10 0 2 4 6 8 10
t t
Generalized Brownian motion with t3 -marginals Difference between standard and t3 Brownian motion
0.5
4
3
0.0
2
1
−0.5
0
g replacements
−1
−1.0
0 2 4 6 8 10 0 2 4 6 8 10
t t
FIGURE 5.1. Standard Brownian motion and underlying copula process (upper row); “generalized
Brownian motion” with scaled marginals such that E(Xt ) = 0, Var(Xt ) = t as in the Brownian
case.
5.3 Continuous Time Stochastic Processes 65
We will make the general assumption that the marginal distributions Ft are
continuous and strictly increasing for all t ∈ T .
If (5.21) holds for all t ∈ T , we say that the process is stochastically continuous on T .
Following Todorovic (1992, Remark 1.9.1), we note that “stochastic continuity is a reg-
ularity condition on bivariate marginal distributions of the process”. Having in mind the
goal of constructing a stochastically continuous process, we will examine the relationship
between (5.21) and copulas.
First, note that
(5.22)
Z Zx+ε
=1− dPXt |Xt0 =x (y) dPXt0 (x),
x−ε
and
ZZ ZZ
P( |Xt − Xt0 | > ε) = Ht0 ,t ( dx, dy) = Ct0 ,t Ft0 ( dx), Ft ( dy) .
|x−y|>ε |x−y|>ε
66 5. Stochastic Processes And Copulas
Z Zx+ε
P( |Xt − Xt0 | > ε) = 1 − dFt|t0 (y | x) dFt0 (x)
x−ε
Z
=1− Ft|t0 (x + ε | x) − Ft|t0 (x − ε | x) dFt0 (x)
Z
(5.23)
=1− D1 Ct0 ,t Ft0 (x), Ft (x + ε) − D1 Ct0 ,t Ft0 (x), Ft (x − ε) dFt0 (x)
Z
=1− D1 Ct0 ,t u, Ft (Ft−1
0
(u) + ε) − D1 Ct0 ,t u, Ft (Ft−1
0
(u) − ε) du.
(0,1)
For the process to be stochastically continuous in t0 , this expression must tend to 0 for
t → t0 , which is equivalent to
Z
lim D1 Ct0 ,t u, Ft (Ft−1
0
(u) + ε) − D1 Ct0 ,t u, Ft (Ft−1
0
(u) − ε) du = 1.
t→t0 | {z }
(0,1) =:f (u;t0 ,t,ε)
Now, from (2.22) we have 0 ≤ f (u; t0 , t, ε) ≤ 1 for all u ∈ (0, 1). Further, assume that
the limit limt→t0 f (u; t0 , t, ε) exists for almost all u ∈ (0, 1) and is a measurable function.
With Lebesgue’s dominated convergence theorem (e. g., Billingsley, 1995, Theorem 16.4)
we obtain the condition
Z Z
lim f (u; t0 , t, ε) du = lim f (u; t0 , t, ε) du = 1.
t→t0 t→t0
(0,1) (0,1)
which follows from Bauer (1992, Theorem 13.2). Moreover, we see that the following is
equivalent to stochastic continuity in t0 :
lim D1 Ct0 ,t u, Ft (Ft−1
0
(u) + ε) = 1 [λ], and
t→t0
(5.24)
lim D1 Ct0 ,t u, Ft (Ft−1
0
(u) − ε) = 0 [λ]
t→t0
for any ε > 0. This is due to the fact that 0 ≤ D1 C(u, v) ≤ 1 for any copula (see (2.21))
and Ft Ft−1
0
(u) + ε ≥ Ft Ft−10
(u) − ε for all ε > 0.
The next goal should be to provide sufficient conditions for (5.24) to hold.
Let us remark that for the upper Fréchet bound M (u, v) = min(u, v), u, v ∈ (0, 1), we
have D1 M (u, v) = (0,v) (u) for v ∈ (0, 1). If we compare this with (5.24), we see that the
limit should yield strictly monotone dependence. This will be formalized in the following
Lemma.
5.3 Continuous Time Stochastic Processes 67
5.26 Lemma
Let (Cn )n∈ ⊂ C2 be a sequence of 2-copulas and M (u, v) = min(u, v), u, v ∈ (0, 1), the
n→∞ 0, u > v,
(5.25)
for all v ∈ (0, 1).
Then limn→∞ Cn (u, v) = M (u, v) for all u, v ∈ (0, 1).
Proof. First, assume u < v. Then,
(2.11)
|Cn (u, v) − M (u, v)| = M (u, v) − Cn (u, v) = min(u, v) − Cn (u, v) = u − Cn (u, v)
(∗)
= u − [Cn (u, v) − Cn (0, v)] = u − D1 Cn (ξn , v) · u
| {z }
=0
for some ξn ∈ (0, u). (∗) is valid due to the existence of the partial derivative and the
continuity of Cn (mean value theorem of differential calculus, e. g., Walter, 1992, Theorem
10.10). As ξn < u < v for all n ∈ , the assumption yields limn→∞ D1 Cn (ξn , v) = 1 so that
for all 0 < ε < u so that limn→∞ Cn (u, u) = M (u, u) for all u ∈ (0, 1).
limn→∞ D1 Cn (u, v) exists for all u ∈ (0, 1) \ {v} and limn→∞ D1 Cn (u, v) is continuous
in u for all v, then, for any v ∈ (0, 1),
(
1, u < v,
lim D1 Cn (u, v) = D1 M (u, v) = for u ∈ (0, 1) \ {v}.
n→∞ 0, u > v,
68 5. Stochastic Processes And Copulas
Due to the continuity of the limit, there exists an ε > 0 such that limn→∞ D1 Cn (x, v) < 1
for all x ∈ (u − ε, u + ε) ∩ (0, 1). Choose δ = min{ε, 1 − u, u, v − u} so that (u − δ, u +
δ) ⊂ (0, v). Then also limn→∞ D1 Cn (x, v) < 1 for all x ∈ (u − δ, u + δ). Now, observe
Cn (u + δ, v) − Cn (u, v). First,
Cn (u + δ, v) − Cn (u, v) = D1 Cn (ξn , v) · δ
which is clearly a contradiction to the above equation (∗). The assertion follows.
From these two Lemmas, we can immediately derive the following corollary.
5.28 Corollary
Let (Cn )n∈ ⊂ C2 be a sequence of 2-copulas and M (u, v) = min(u, v). For any v ∈ (0, 1),
let limn→∞ D1 Cn (u, v) exist for all u ∈ (0, 1) \ {v} and be continuous in u. Then the
following properties are equivalent:
n→∞
Proof. a)⇒b) follows from Lemma 5.27, b)⇒a) from Lemma 5.26.
5.29 Remark
Lemmas 5.26 and 5.27 and Corollary 5.28 could equivalently have been formulated for a
family {Cs,t : s, t ∈ T } of bivariate copulas where limt→s Cs,t (u, v) must be considered. This
is an important aspect when dealing with continuous time stochastic processes.
With these preparations, we are able to give sufficient conditions for the stochastic con-
tinuity of a process generated by a family of copulas:
5.3 Continuous Time Stochastic Processes 69
5.30 Theorem
Let (Xt )t∈T be a stochastic process, Ft the cumulative distribution function of Xt , t ∈ T ,
and Cs,t = ChXs , Xt i the copula of (Xs , Xt ), s, t ∈ T .
If
a) lim Ft−1 (u) = Fs−1 (u) for almost all u ∈ (0, 1), and
t→s
b) lim D1 Cs,t (u, v) = (0,v) (u) exists for all u ∈ (0, 1) \ {v},
t→s
= 0 and = 1.
lim D1 Cs,t (u, v) lim D1 Cs,t (u, v)
t→s u,Ft (Fs−1 (u)−ε) t→s u,Ft (Fs−1 (u)+ε)
5.31 Remark
a) We cannot expect to enhance the above result w. r. t. stronger continuity concepts as
they are usually based on other concepts than finite-dimensional distributions.
b) The finite-dimensional distributions alone do not give all the information regarding
the continuity properties of a process. For example, let (Ω, F , P) = [0, ∞), B, µ
where B denotes the Borel σ-algebra on [0, ∞), and µ is a probability measure on
[0, ∞) with no mass on single points. It is easy to see that Xt (ω) = {t} (ω) and
Yt (ω) = 0 for all (t, ω) ∈ [0, ∞)×[0, ∞) have the same finite-dimensional distributions
and that (Xt ) is a version of (Yt ),3 but t → Yt (ω) is continuous for all ω while
t → Xt (ω) is discontinuous for all ω.
Let us give an example for the use of the theorem. We have already determined the copula
of a Brownian motion (see Example 5.23). It is known that Brownian motion almost surely
has continuous sample paths, from which the stochastic continuity on + follows. We
should be able to check the latter property via the copula approach:
√
Ft (x) = Φ(x/ t), x ∈ , which is equivalent to Ft−1 (y) = tΦ−1 (y), y ∈ (0, 1). Then
√ √ t→s
Ft−1 (y) − Fs−1 (y) = ( t − s)Φ−1 (y) −−→ 0
B
for y ∈ (0, 1). From Example 5.23 we know that the copula Cs,t is given by
Zu √ √
B tΦ−1 (v) − sΦ−1 (w)
Cs,t (u, v) = Φ √ dw,
t−s
0
i. e., √ −1 √
B tΦ (v) − sΦ−1 (u)
D1 Cs,t (u, v)
=Φ √ , t > s.
t−s
√ √ √
(Note that the copula would be Φ ( sΦ−1 (u) − tΦ−1 (v))/ s − t for s > t.) Clearly,
B
D1 Cs,t (u, v) is continuous in u. Now,
√ √ −1 t→s √
tΦ−1 (v) − sΦ (u) −−→ s Φ−1 (v) − Φ−1 (u)
√ t→s
and t − s −−→ 0; therefore
√ −1 √ (
tΦ (v) − sΦ−1 (u) ∞, u < v,
lim √ =
t→s t−s −∞, u > v,
B
so that limt→s D1 Cs,t (u, v) = (0,v) (u). The stochastic continuity now follows from Theo-
rem 5.30.
We will show in the next section that Theorem 5.30 can be used advantageously to
shorten existing proofs.
continuous-parameter stationary EFGM processes. Cambanis (1991) showed that these pro-
cesses are not stochastically continuous and even not measurable. To be precise, he showed
this for strictly stationary EFGM processes. He did not use an approach via the underlying
copulas but considered the complete parametric form (including the marginals), showing
the discontinuity of any such process by using a set of inequalities. These results were
obtained under a weak symmetry-type condition on the support of the marginal distribu-
tion4 . Due to this approach, Cambanis was not able to derive results for the non-stationary
case. (Note: “non-stationarity” can mean time-varying marginals, different copulas C s,t and
Cs+h,t+h , or both!)
When both marginals are absolutely continuous, αmin = −1, αmax = 1, but generally,
αmin = − min (M1 M2 )−1 , ((1 − m1 )(1 − m2 ))−1 ,
(5.28)
αmax = min (M1 (1 − m2 ))−1 , ((1 − m1 )M2 )−1 ,
where
mk = inf ({Fk (x) : x ∈ } \ {0, 1}), Mk = sup ({Fk (x) : x ∈ } \ {0, 1}), k = 1, 2.
where X
1+ εk εj αk,j ≥ 0 for all εk ∈ {−Mk , 1 − mk } (5.30)
1≤k<j≤n
must be valid.
Based on this multivariate distribution, one can now introduce EFGM random processes
(Xt )t∈T by requiring that all finite-dimensional distributions are multivariate EFGM, i. e.,
for the distribution function of (Xt1 , . . . , Xtn ), t1 < · · · < tn , ti ∈ T , n ∈ , we have
n
!
Y X
Ht1 ,...,tn (x1 , . . . , xn ) = Fti (xi ) 1 + α(tk , tj ) 1 − Ftk (xk ) 1 − Ftj (xj ) ,
i=1 1≤k<j≤n
(5.31)
where Ft is the cdf of Xt .5 The coefficient function α(s, t), s 6=P
t, is assumed to be symmetric
(α(s, t) = α(t, s)) and the admissible values are given by 1 + 1≤k<j≤n εk εj α(tk , tj ) ≥ 0 for
4 Note that we deliberately write “marginal distribution” instead of “marginal distributions” as Cambanis considers strictly
stationary processes.
5 The consistency condition (5.5) can be easily checked.
72 5. Stochastic Processes And Copulas
5.33 Theorem
Let (Xt )t∈T be a real-valued, strictly stationary EFGM process with nondegenerate mar-
ginal distribution F and dependence function α. If M (F ) = 1 − m(F ), where m(F ) :=
inf({F (x) : x ∈ } \ {0, 1}), M (F ) := sup({F (x) : x ∈ } \ {0, 1}), then the process is
not stochastically continuous.6
From Section 5.3 we know that stochastic continuity is a matter of bivariate distributions.
It is therefore clear and advantageous for the analysis of stochastic continuity that we only
need the bivariate distributions of the process, i. e., a stochastic process form of (5.27):
Hs,t (xs , xt ) = Fs (xs )Ft (xt ) 1 + α(s, t) 1 − Fs (xs ) 1 − Ft (xt ) , xs , xt ∈ , (5.32)
where Hs,t is the distribution function of (Xs , Xt ), s < t. We can easily identify the copula
as
Cs,t (u, v) = uv 1 + α(s, t)(1 − u)(1 − v) , u, v ∈ [0, 1]. (5.33)
Note that the conditions for α(s, t) yield that Cs,t is 2-increasing for all s < t; generally,
only α(s, t) ∈ [−1, 1] is required for Cs,t to be a copula. Now,
5.34 Lemma
A uniform process (Ut )t∈T generated from the EFGM family is not continuous in probability
in any point t ∈ T .
Proof. From (5.24) and (5.34), we have (Ft (u) = Ft−1 (u) = u for all u ∈ [0, 1]) for all ε > 0:
This can clearly not be the case as α(s, t) is independent from u. Thus, the process cannot
be continuous in t.
5.35 Theorem
Let (Xt )t∈T be a real-valued, strictly stationary EFGM process with continuous marginal(s)
F . Then (Xt ) is not stochastically continuous in any t ∈ T .
Proof. For the process to be continuous in t, (5.24) must hold for all ε > 0 and almost all
u ∈ (0, 1). Taking the second equation and (5.34), we must particularly have
t→s
D1 Cs,t u, F (F −1 (u)−ε) = F F −1 (u)−ε +α(s, t) 1−F F −1 (u)−ε (1−2u) −−→ 0 [λ1 ].
Without loss of generality, assume u 6= 1/2, as {1/2} is a λ1 null set. Let ε > 0 be given.
The case F (F −1 (u) − ε) = 1 cannot occur as this is equivalent to F −1 (u) ≥ ω(F ) + ε.7
As u ∈ (0, 1), this is impossible. For the same reason, g(u, ε) := F (F −1 (u) − ε) < 1 for all
u ∈ (0, 1). So the above equation is equivalent to
s→t g(u, ε) 1
α(s, t) −−→ · [λ1 ],
g(u, ε) − 1 1 − 2u
which evidently cannot be the case, as any Borel set with positive measure consists of
uncountably many points. (The left side is independent of the right side.) Therefore, (Xt )
is not stochastically continuous in t.
+
We assume T ⊂ .
on ), we have
5.37 Remark
(5.35) is equivalent to P(Xtn ≤ x | Xt1 , . . . , Xtn−1 ) = P(Xtn ≤ x | Xtn−1 ) [P] for all x ∈
as {(−∞, x] : x ∈ } is a generator of B closed under intersections.
We will use the previously introduced notation for the conditional distribution functions:
This can be interpreted as a continuous version of the law of total probability, modified by
the Markov property.
We will now give an outline of the work of Darsow, Nguyen, and Olsen (1992) who
were the first to relate copulas to Markov processes. The key to the understanding of this
relationship was the introduction of a product for 2-copulas which essentially captures the
dependence model of the Chapman-Kolmogorov equations.
but D2 C1 (0, z) = limh&0 (C1 (0, z + h) − C1 (0, z))/h = 0 as copulas are grounded so that
(C1 ∗ C2 )(0, v) = 0 for v ∈ I; the case (C1 ∗ C2 )(u, 0) = 0 for all u ∈ I follows analogously.
Further,
Z Z
(C1 ∗ C2 )(1, v) = D2 C1 (1, z) · D1 C2 (z, v) dz = D1 C2 (z, v) dz = C2 (1, v) − C2 (0, v)
(0,1) (0,1)
=v−0=v
for v ∈ I, as D2 C1 (1, z) = limh&0 (C1 (1, z + h) − C1 (1, z))/h = limh&0 (z + h − z)/h = 1.
The case (C1 ∗ C2 )(u, 1) = u for all u ∈ I is analogous.
It remains to show that C1 ∗ C2 is 2-increasing, i. e., the (C1 ∗ C2 )-volume of any rectangle
(u, v] (u = (u1 , u2 ), v = (v1 , v2 ), ui , vi ∈ I, u1 ≤ u2 , v1 ≤ v2 ) must be non-negative:
VC1 ∗C2 (u, v]
= (C1 ∗ C2 )(u2 , v2 ) − (C1 ∗ C2 )(u1 , v2 ) − (C1 ∗ C2 )(u2 , v1 ) + (C1 ∗ C2 )(u1 , v1 )
Z
= D2 C1 (u2 , z)D1 C2 (z, v2 ) − D2 C1 (u1 , z)D1 C2 (z, v2 )
(0,1)
− D2 C1 (u2 , z)D1 C2 (z, v1 ) + D2 C1 (u1 , z)D1 C2 (z, v1 ) dz
Z
= [D2 C1 (u2 , z) − D2 C1 (u1 , z)] [D1 C2 (z, v2 ) − D1 C2 (z, v1 )] dz ≥ 0.
| {z }| {z }
(0,1) ≥0, see (2.22) ≥0, see (2.22)
Thus, C1 ∗ C2 is a copula.
hold:
n→∞ n→∞
a) An ∗ B −−−→ A ∗ B and B ∗ An −−−→ B ∗ A.
where Ct1 ,...,tn = ChXt1 , . . . , Xtn i and Ctk ,tk+1 = ChXtk , Xtk+1 i.
5.4 Markov Processes 77
These findings create a new approach to the theory of Markov processes and yield a new
technique for constructing such processes. We cite Darsow et al. (1992):
(
1−t1 /t2 1/t 1/t
u2 , 0 < u1 1 < u2 2 < 1,
D1 Ct1 ,t2 (u1 , u2 ) = 1/t 1/t (5.43)
0, 0 < u2 2 < u1 1 < 1,
8 This is a known fact (see Resnick, 1987, Proposition 4.1); we will derive this result with the help of the new tools based
and
(
t1 −t1 /t2 1/t1 1/t2
u1 1 − t2
u2 , 0 < u1 < u2 < 1,
D2 Ct1 ,t2 (u1 , u2 ) = 1/t 1/t
(5.44)
1, 0< u2 2 < u1 1 < 1.
Note that “∧” and “∨” denote the minimum and maximum, respectively. For t1 = t2 we
have (clearly) C(u, v) = M (u, v) = min(u, v), u, v ∈ I, so this case is consistent with the
findings. For t2 → ∞, we have Ct1 ,t2 (u, v) → Π(u, v), u, v ∈ I, as can be easily seen. Mk
and Ml are asymptotically independent.
To derive the finite-dimensional distributions, i. e. for (Mt1 , . . . , Mtk ), we note that for
t1 < t 2 < · · · < t k ,
k
! k
!
^ ^
Ht1 ,...,tk (x1 , . . . , xk ) = F t1 xi · F t2 −t1 xi · · · F tk −tk−1 (xk )
i=1 i=2
Z
(Ct1 ,t2 ∗Ct2 ,t3 )(u1 , u3 ) := D2 Ct1 ,t2 (u1 , v) · D1 Ct2 ,t3 (v, u3 ) dv
(0,1)
t /t
u12 1
Z Z1
= D1 Ct2 ,t3 (v, u3 ) dv + u1 (1 − t1 /t2 )v −t1 /t2 D1 Ct2 ,t3 (v, u3 ) dv,
| {z }
0 t /t1
u12
t /t3
=0 for v>u32
t /t3
2
uZ
3
t /t1 1−t2 /t3
= Ct2 ,t3 (u12 , u3 ) + u1 (1 − t1 /t2 ) · v −t1 /t2 · u3 dv
t /t1 t /t
u12 ∧u32 3
5.4 Markov Processes 79
h iu3t2 /t3
t /t1 1−t2 /t3
= Ct2 ,t3 (u12 , u3 ) + u 1 u3 v 1−t1 /t2 t /t1 t /t
u12 ∧u32 3
t 2 t3 −t2 h i
1/t 1/t 1/t 1−t /t t /t −t /t 1/t 1/t t −t
= u1 1 ∧ u 3 3 u3 3 + u1 u3 2 3 u32 3 1 3 − u1 1 ∧ u3 3 2 1
1/t1 1/t3 t2 1−t2 /t3 1−t1 /t3 1−t2 /t3 1/t1 1/t3 t2 −t1
= u1 ∧ u3 u3 + u 1 u3 − u 1 u3 u1 ∧ u3
ut12 /t1 u1−t
3
2 /t3 1−t /t 1−t /t (t −t )/t 1−t /t 1/t 1/t
+ u 1 u3 1 3 − u 1 u3 2 3 u1 2 1 1 = u 1 u3 1 3 , u1 1 < u 3 3 ,
=
u + u u1−t1 /t3 − u u1−t1 /t3 = u , u1
1/t1 1/t
≥ u3 3 .
3 1 3 1 3 3
As we had expected, this yields Ct1 ,t2 ∗ Ct2 ,t3 = Ct1 ,t3 when comparing with equation (5.42).
We can now prove the above mentioned theorem within the copula framework:
5.45 Theorem
Let the situation be as above. Then {XL(n) : n ∈ } is a Markov process.
Proof. We will use the characterization provided by Theorem 5.44.
For t1 < t2 < · · · < tn , ti ∈ , n ≥ 3, we have to show
2 t /t3
u2 ∧u
Z3 h iz=u2 ∧ut32 /t3
1−t /t 1−t /t
= D2 Ct1 ,t2 (u1 , z) · u3 2 3 dz = u3 2 3 · Ct1 ,t2 (u1 , z)
z=0
0
t 1 t2 −t1
1−t /t 1/t t /t 1/t2 t /t 1/t2
= u3 2 3 u1 1 ∧ u2 ∧ u22 3 u2 ∧ u32 1
1−t2 /t3 1/t1 1/t2 1/t3 t1 1/t2 1/t3 t2 −t1
= u3 u1 ∧ u2 ∧ u3 u2 ∧ u3
(5.45)
= Ct1 ,t2 ,t3 (u1 , u2 , u3 ).
80 5. Stochastic Processes And Copulas
Assuming the validity for some n ∈ , n ≥ 3, we get the validity for the case n + 1:
Ct1 ,t2 ? · · · ? Ctn−1 ,tn ? Ctn ,tn+1 (u1 , . . . , un , un+1 )
“?” associative
= Ct1 ,t2 ? · · · ? Ctn−1 ,tn (u1 , . . . , un ) ? Ctn ,tn+1 (un , un+1 )
assumption
= Ct1 ,...,tn (u1 , . . . , un ) ? Ctn ,tn+1 (un , un+1 )
Z
Def.
= Dn Ct1 ,...,tn (u1 , . . . , un−1 , z) · D1 Ctn ,tn+1 (z, un+1 ) dz
(0,un )
tn /t
un ∧un+1 n+1
Z
1−t /tn+1
= un+1n Dn Ct1 ,...,tn (u1 , . . . , un−1 , z) dz
0
h iz=un ∧un+1
tn /tn+1
1−t /t
= un+1n n+1 Ct1 ,...,tn (u1 , . . . , un−1 , z)
z=0
n−1
^ 1/tn t1
1−t /tn+1 1/ti t /t
= un+1n ui n n+1
∧ un ∧ un+1
i=1
n−1
^ 1/tn t2 −t1 1−tn−1 /tn
1/ti n n+1 t /t t /t
n n+1
· ui ∧ un ∧ un+1 · · · un ∧ un+1
i=2
n+1
^ n+1
^ n+1
^
1−t /t 1/t t 1/t t −t 1/ti tn −tn−1
= un+1n n+1 ui i 1 ui i 2 1 ··· ui
i=1 i=2 i=n
= Ct1 ,...,tn+1 (u1 , . . . , un+1 ).
This finishes the proof.
5.46 Remark
This result easily yields Proposition 4.1 in Resnick (1987, p. 165) which says that the
stationary transition probabilities of {XL(n) } are given by
(
1−F (y)
, y > x,
Π x, (y, ∞) := P(XL(n+1) > y | XL(n) = x) = 1−F (x) (5.46)
1, y ≤ x,
as Π x, (y, ∞) = 1 − D1 CL(n),L(n+1) F L(n) (x), F L(n+1) (y) . F denotes the distribution
function of the Xi as usual.
However, for absolutely continuous cdf F we have
r−1 j
X 1 1
FXL(r) (x) = 1 − 1 − F (x) log
j=0
j! 1 − F (x)
(Kamps, 1995, p. 32). Thus, an analytic derivation of the generalized inverse of FXL(r) and
therefore the copula of (XL(r) , XL(s) ) does not seem to be possible.
5.4 Markov Processes 81
5.4.2 Symmetry
d
A well-known property of standard Brownian motion is its symmetry in the sense of (Bt ) =
(−Bt ). We can now ask under which condition(s) a “generalized Brownian motion”, which
is a process (Xt ) with Ch(Xt )i = Ch(Bt )i (copulas coincide but possibly other marginals)
will be symmetric. Recalling that Brownian motion is a Markov process we can generalize
this question:
Which condition(s) must be fulfilled by the copulas and the marginals of the process for
Xt to be symmetric? In Section 5.2.3 a characterization for symmetry of general processes
was given so that we can make use of it in the current context.
We will tackle the problem by considering a general Markov process (Xt )t∈T where T =
[0, ∞) without loss of generality.
To this end, the Markov property yields
P(Xt1 ≤ x1 , . . . ,Xtn ≤ xn )
Yn
= P(Xti ≤ xi | Xt1 ≤ x1 , . . . , Xti−1 ≤ xi−1 ) · P(Xt1 ≤ x1 )
i=2
n
Y
Markov property (5.47)
= P(Xti ≤ xi | Xti−1 ≤ xi−1 ) · P(Xt1 ≤ x1 )
i=2
Qn
i=2 Cti−1 ,ti Fti−1 (xi−1 ), Fti (xi )
= Qn−1
i=2 Fti (xi )
5.47 Example
For the EFGM copula (see Section 5.3.2)
It is well known that Brownian motion is symmetric about 0 (Karatzas and Shreve,
2000, Lemma 9.4, p. 104). In the following example (which ends on page 84) we will give
an alternative proof for the symmetry with the help of Theorem 5.21. Note that we only
need to check condition (5.50) for the bivariate copulas due to the above remarks.
5.49 Theorem
Let G ⊂ p be an open connected set and f = (f1 , . . . , fq ) : G → q , p, q ∈ . If the
5.50 Corollary
Let G ⊂ p be an open connected set and p, q ∈ . If f = (f1 , . . . , fq ), g = (g1 , . . . , gq ) :
G → q and all partial derivatives are the same, i. e., ∂x∂ i fj (x1 , . . . , xp ) = ∂x∂ i gj (x1 , . . . , xp )
for all 1 ≤ i ≤ p, 1 ≤ j ≤ q, (x1 , . . . , xp ) ∈ G, and there exists x0 ∈ G with f (x0 ) = g(x0 ),
then f ≡ g on G.
5.51 Remark
If f and g are continuous on G (the closure of G), then f (x0 ) = g(x0 ) for some x0 ∈
∂G := G \ int G (the boundary, “int” denotes the interior of G) will be sufficient for the
above corollary to hold.
5.4 Markov Processes 83
Due to Corollary 5.50 and the following Remark 5.51 it is sufficient for proving symmetry
B
to check if the partial derivatives of Cs,t b B coincide on (0, 1)2 where C B is the copula
and C s,t s,t
of Brownian motion (see Example 5.23).9
Let (u, v) ∈ (0, 1)2 , 0 < s < t < ∞. It was shown in (5.16) and (5.17) that
√ √
B tΦ−1 (v) − sΦ−1 (u)
D1 Cs,t (u, v) =Φ √
t−s
and
r Zu √ √
B t 1 tΦ−1 (v) − sΦ−1 (w)
D2 Cs,t (u, v) = · ϕ √ dw.
t − s ϕ Φ−1 (v) t−s
0
bB (u, v) = 1 − D2 C B (1 − u, 1 − v)
D2 C s,t s,t
r Z1−u √ −1 √
t 1 tΦ (1 − v) − sΦ−1 (w)
=1− · ϕ √ dw
t − s ϕ Φ−1 (1 − v) t−s
0
r Z1−u √ −1 √
Φ−1 (1−v)=−Φ−1 (v)
and ϕ symmetric t 1 − tΦ (v) + sΦ−1 (1 − w)
= 1− · ϕ √ dw
t − s ϕ Φ (v)
−1 t−s
0
r Z1−u √ √ −1
t 1 tΦ−1 (v) − sΦ (1 − w)
=1− · ϕ √ dw
t − s ϕ Φ (v)
−1 t−s
0
r Z1 √ √
1−w=x t 1 tΦ−1 (v) − sΦ−1 (x)
= 1− · ϕ √ dx.
t − s ϕ Φ (v)
−1 t−s
u
9 Note that both C B and C b B are copulas so that they are continuous on I 2 and all values C B (u, v) and C
b B (u, v) are
s,t s,t s,t s,t
the same for u ∈ {0, 1} or v ∈ {0, 1}.
84 5. Stochastic Processes And Copulas
Thus,
r Z1 √ √
B bB (u, v) t 1 tΦ−1 (v) − sΦ−1 (x)
D2 Cs,t (u, v) − D2 C s,t = · ϕ √ dx − 1
t − s ϕ Φ−1 (v) t−s
0
Z1
(5.18)
= cB
s,t (x, v) dx − 1 = (0,1) (v) −1=0
0
(5.51)
The question arises if there is a characterization of those classes of bivariate copulas, i. e.,
of a set C sym ⊂ C2 , which generate symmetric Markov processes by solving the functional
equation C = C. b
For a wide class of copulas, the Archimedean family (see Section 2.3), we can give a
positive answer due to the work of Frank (1979). The answer leads to a special class
of Archimedean copulas, the Frank family which is the only Archimedean family with
b v) = C(u, v) for all u, v ∈ I. This is exactly the desired property.
C(u,
Then this family of copulas is called Frank family, each member is called a Frank copula.
Please refer to Figure 5.2 for a visual impression of the density when the margins are
N(0, 1).
5.53 Remark
e−ϑt −1
a) The characteristic generator of Cϑ for ϑ ∈ \ {0} is ϕϑ (t) = − ln e−ϑ −1
(see
Section 2.3).
b) The definition of C−∞ , C0 and C∞ is motivated by the fact that these copulas are the
limits of Cϑ , e. g., limϑ→0 Cϑ (u, v) = Π(u, v). They have been added for completeness
as they also generate symmetric Markov processes.
c) Some of the statistical properties of this family were discussed in Nelsen (1986) and
Genest (1987).
5.4 Markov Processes 85
θ = −8 θ = −2
1 2 3
1 2 3
−1
−1
−3
−3
−3 −1 1 2 3 −3 −1 1 2 3
θ = −0.5 θ = 0.5
1 2 3
1 2 3
−1
−1
−3
−3 −1 1 2 3 −3 −3 −1 1 2 3
θ=2 θ=8
1 2 3
1 2 3
−1
−1
−3
−3
−3 −1 1 2 3 −3 −1 1 2 3
FIGURE 5.2. Contour plots of the Frank copula density with standard normal margins.
d) Cϑ is radially symmetric for all ϑ ∈ . This could be checked tediously, but we get
this from Theorem 5.21 and the following Theorem 5.54.
Thus, this result yields a characterization of symmetric Markov processes whose copulas
are in the Archimedean class.
86 5. Stochastic Processes And Copulas
5.59 Corollary
t→∞
If M is a (Ft , P)-local martingale vanishing at 0 and such that h(t) := hM it −−−→ ∞, the
bivariate copulas of the process are given by
Zu p p !
h(t)Φ−1 (v) − h(s)Φ−1 (w)
Cs,t (u, v) = Φ p dw, 0 ≤ s < t. (5.56)
h(t) − h(s)
0
Let us give some examples for the application of Corollary 5.59. Bt always denotes
standard Brownian motion.
(see, e. g., Karatzas and Shreve, 2000, p. 358, or Øksendal, 1998, p. 74). The solution is
given by
Zt
Xt = X0 · exp(−αt) + σ exp(−α(t − s)) dBs , 0 ≤ t < ∞. (5.58)
0
88 5. Stochastic Processes And Copulas
However, (Xt ) is not a martingale if X0 6= 0 on a set A ∈ F with P(A) > 0. This is due to
Zt
E(Xt | Fs ) = X0 · exp(−αt) + σ · E exp(−α(t − u)) dBu | Fs
0
Zs
= X0 · exp(−αt) + σ exp(−α(t − u)) dBu
0
= Xs + X0 · exp(−α(t − s)), 0 ≤ s < t,
Rt
where the second last equality is valid because 0 exp(−α(t−u)) dBu is a martingale w. r. t.
(Ft ) (see Øksendal, 1998, Corollary 3.2.6 or Karatzas and Shreve, 2000, Proposition 3.2.10,
p. 139). If we set
Zt
Yt := exp(αt)Xt = X0 + σ exp(αs) dBs (5.59)
0
Mt := Yt − X0 , t ≥ 0,
i. e., Xs and Xt are asymptotically independent. Note that Lebesgue’s dominated conver-
gence theorem must be used to interchange integration and the taking of the limit.
The copula is also independent of σ. This means that α alone determines the dependence
between Xs and Xt .
For the parameter α, we have the following limiting copulas:
α→∞:
We have
e2αt − 1 1 − e−2αt α→∞
= −−−→ 1
e2αt − e2αs 1 − e−2α(t−s)
and
α→0:
1 − e−2αt 2te−2αt t
lim −2α(t−s)
= lim −2α(t−s)
=
α→0 1 − e α→0 2(t − s)e t−s
and
1 − e−2αs 2se−2αs s
lim 2α(t−s)
= lim −2α(t−s)
=
α→0 e − 1 α→0 2(t − s)e t−s
90 5. Stochastic Processes And Copulas
so that
OU B
lim Cs,t = Cs,t
α→0
B
where Cs,t is the Brownian copula. This result is not surprising in view of the defining SDE
(5.57) and its solution (5.58): Letting α → 0 in (5.58), we get
Zt
Xt = X 0 + σ dBs = X0 + σBt .
0
This again is only a strictly monotone increasing transformation of the Brownian motion
(Bt ), so the copulas must be the same. The interchange of the integral and limit is justified
by monotone convergence; it is standard to show it for elementary functions and then
generalize it by monotone class theorems.
b − Xt
dXt = dt + dBt , 0 ≤ t < T, X0 = a. (5.62)
T −t
It can be shown (Karatzas and Shreve, 2000, pp. 358 or Øksendal, 1998, p. 75) that the
solution is given by
Zt
t t dBs
Xt = a 1 − + b + (T − t) , 0 ≤ t < T. (5.63)
T T T −s
0
The process can be thought of to be a “linear bridge” between the points (0, a) and (T, b)
which is disturbed R t by sa time-scaled Brownian motion.
Setting Yt := 0 TdB −s
, we see that, again, Xt = f (Yt ) for f (x) ≡ f (x; a, b, T, t) = a(1 −
t t
T
) + b( T ) + (T − t)x so that the copulas must coincide. (Yt )0≤t<T is a continuous martingale
with Y0 = 0. In analogy to Example 5.60 we calculate the quadratic variation process to
be
Zt t
1 −1 1 1
hY it = 2
ds = (T − s) = − . (5.64)
(T − s) 0 T −t T
0
t→∞
One problem remains: We do not have hY it −−−→ ∞, but we do not need this requirement.
Taking a look at the proof of Theorem 5.58 in Karatzas and Shreve (2000, pp. 174) we see
that in the case of a bounded time interval [0, T ), we need the condition limt→T hY it = ∞
which is met by (5.64). For h(t) = T 1−t − T1 , 0 ≤ t < T , we get
s s s s
h(t) t(T − s) h(s) s(T − t)
= and = .
h(t) − h(s) T (t − s) h(t) − h(s) T (t − s)
5.5 Continuous Local Martingales 91
5.62 Remark
Note that h(t) = hM, M it is usually a stochastic process, i. e., one should write h(t, ω).
However, in the above examples, h(t) turned out to be a deterministic function which is
due to the fact that the quadratic variation of Brownian motion is only a function of time
which is a consequence of the Itô-isometry.
It is this fact which makes the theory of time transformation applicable to the derivation
of copulas, for, if the time transformation were stochastic, we would end up with a copula
with stochastic time. With reference to equation (5.56) we had to write Cs,t (u, v; ω) and
h(t, ω), yielding a random copula. This would be of only theoretical interest.
Nonetheless, from a theoretical point of view, the central result is that of Monroe (1978)
who showed that a process is equivalent to a time change of Brownian motion if and only if
it is a local semimartingale. This is the most general result in that direction. Note that—for
example—Lévy processes fall into this class.
It is well-known that stochastic integrals are Markov processes, and therefore, the above
examples also yield bivariate copulas generating Markov processes. In accordance with the
remarks on page 62 we are now free to construct new processes with the intertemporal
dependence given by the above copulas but with arbitrary marginals.
92 5. Stochastic Processes And Copulas
Appendix A
General Results from Probability Theory
(with inf(∅) := ∞, sup(∅) := −∞) denote the right and the left endpoint of the support
of F .
Uniform Distribution
Symbol: U(a, b), a, b ∈ ,a<b
Density: f (x) = (b − a)−1 · (a,b) (x)
Normal Distribution
Symbol: N(µ, σ 2 ), µ ∈ , σ > 0
√ 2
Density: f (x) = ( 2πσ)−1 exp − (x−µ)
2σ 2
t-Distribution
Symbol: tν (µ, σ 2 ); tν for µ = 0, σ = 1; µ ∈ , σ > 0, ν ∈
− ν+1
Γ( ν+1 ) (x−µ)2 2
Density: f (x) = νπΓ( ν )σ 1 + νσ2
√ 2
2
In most cases of the application of copulas no closed form solutions exist (e. g., portfolio
analysis and option pricing in mathematical finance) so that there is a great interest in
simulation and Monte Carlo studies.
We will therefore give a short introduction to the generation of random samples from
a given joint distribution. We assume that procedures for generating independent uniform
variates are known to the reader (e. g., Lange, 1998, pp. 269, and Devroye, 1986, for al-
ternative methods). From these uniform samples we can generate samples x of a random
variable X with distribution function F by the following two steps:
1. Generate u from U(0, 1).
2. Set x = F −1 (u) where F −1 is the generalized inverse of F (see page 93).
Here we will focus on generating random samples from a copula. No proofs will be given
as they can be found in Nelsen (1999) or follow directly as in the case of the generation of
time-series from bivariate copulas.
By virtue of Sklar’s theorem, we only need to generate a pair (u, v) of U(0, 1) variables
(U, V ) with copula C (which is also the copula of X := F −1 (U ) and Y := G−1 (V )) and
then transform this pair by the quantile transformation.
Simulating uk from Ck (uk | u1 , . . . , uk−1 ) can be done by simulating q from U(0, 1) from
(−1)
which uk = Ck (q | u1 , . . . , uk−1 ) can be obtained by solving q = Ck (uk | u1 , . . . , uk−1 )
w. r. t. uk 1 if the inverse Ck−1 of Ck does not exist in closed form.
Because of the potential numeric root finding, this general algorithm is often not suit-
able. However, other and more efficient algorithms exist only in special cases such as the
Archimedean family, as we will see in the next section.
2. Set t = λ−1 −1 0
ϕ (q) where λϕ is the generalized inverse of λϕ (v) = ϕ(v)/ϕ (v+), 0 < v ≤
1.
3. Set u = ϕ−1 sϕ(t) , v = ϕ−1 (1 − s)ϕ(t)
simulate a uniform process (Ut ).2 Let Cti ,ti+1 (ui , ui+1 ) be a bivariate copula of (Uti , Uti+1 )
and D1 Cti ,ti+1 the partial derivative w. r. t. the first argument.
Then the following algorithm yields a time-series (u0 , u1 , . . . , un ) generated from
(Ut0 , Ut1 , . . . , Utn ) with starting value ut0 (which may be a realization of an U(0, 1)-variable
or simply Ft−1 0
(x0 ) for some starting value x0 ∈ ) with bivariate copulas Cti ,ti+1 :
2. Set qu (v; s, t) := D1 Cs,t (u, v) and let qu−1 (z; s, t) denote the generalized inverse, z ∈
(0, 1)
The vector (u0 , u1 , . . . , un ) contains the sample including the starting value.
A library for the computer language R is available from the author on request (email:
[Link]@[Link]). R is an integrated suite of software facilities for data
manipulation, calculation and graphical display not unlike STM (see Ihaka and Gentleman,
1996) and freely available under the GNU General Public License for many platforms from
[Link]
The copula-library contains a set of fully documented functions for dealing with copulas
and can be easily installed within a working R-environment by the command
R CMD INSTALL copula*.tgz (Linux)
or within the graphical user interface. Please consult the documentation.
2 Recall that we get a general process from this by applying the quantile transformation Ft−1
i
to the sample.
98 Appendix B. Simulation/Random Variate Generation
List of Symbols and Abbreviations
× cartesian product
<, ≤ less than (or equal to) componentwise
hXi quadratic variation process of X, page 86
d
= equality of finite-dimensional distributions
∼ “is distributed as”
∗ product of copulas, see equation (5.37), page 74
? generalized product of copulas, see equation (5.39), page 76
, A (x) indicator function (of a set A): A (x) = 1 if x ∈ A and otherwise 0
a. s. “almost surely”
α(F ), ω(F ) left and right endpoint of the support of F
Bn n-dimensional Borel-σ-algebra
Bt Brownian motion
cdf “cumulative distribution function”
C, Ci , Ct , Ct1 ,...,tn , etc. copulas
Ch · i copula (family) of the argument
Cn set of all n-copulas
b
C survival copula of C, see equation (2.43)
∆ba H(x1 , . . . , xn ) n-th difference of a function H w. r. t. (a, b), see equation (2.2),
page 6
δC (u) := C(u, u) diagonal of the copula C
(i)
∆(ai ,bi ) H(x1 , . . . , xn ) first difference of the function H in the i-th component w. r. t.
(ai , bi ), see equation (2.3), page 6
100 List of Symbols and Abbreviations
Numbers at the end of each entry refer to the page(s) on which the work is cited.
Alsina, Claudi, Maurice J. Frank, and Berthold Schweizer (2001). Associative functions on
intervals: A primer of triangular norms. Monograph, to appear in 2001. 23
Bauer, Heinz (1992). Maß- und Integrationstheorie. de Gruyter, Berlin, second edition.
18, 62, 66
Billingsley, Patrick (1995). Probability and Measure. Wiley, New York, third edition. 11,
17, 18, 29, 66
Capéraà, Philippe and Christian Genest (1993). Spearman’s ρ is larger than Kendall’s τ
for positively dependent random variables. Nonparametric Statistics, 2, 183–194. 37, 49
Clayton, David G. (1978). A model for association in bivariate life tables and its application
in epidemiological studies of familial tendency in chronic disease incidence. Biometrika,
65(1), 141–151. 44
104 Bibliography
Cook, R. Dennis and Mark E. Johnson (1981). A family of distributions for modelling
non-elliptically symmetric multivariate data. Journal of the Royal Statistical Society B,
43(2), 210–218. 2, 44, 46
Cramér, Harald and M. Ross Leadbetter (1967). Stationary and Related Stochastic Pro-
cesses: Sample Function Properties and Their Applications. Wiley, New York. 54
Darsow, William F., Bao Nguyen, and Elwood T. Olsen (1992). Copulas and Markov
processes. Illinois Journal of Mathematics, 36(4), 600–642. 2, 3, 12, 17, 57, 74, 75, 76,
77
David, Herbert Aron (1981). Order Statistics. Wiley, New York. 41, 45
Devroye, Luc (1986). Non-Uniform Random Variate Generation. Springer, New York. 95
Dubins, Lester E. and Gideon Schwarz (1965). On continuous martingales. Proc. Nat.
Acad. Sciences U.S.A., 53, 913–916. 4, 86, 87
Embrechts, Paul, Alexander McNeil, and Daniel Straumann (1999). Correlation and
dependence in risk management: Properties and pitfalls. Working Paper. Risklab:
[Link] 2, 33
Esary, James D., Frank Proschan, and David W. Walkup (1967). Association of random
variables, with applications. The Annals of Mathematical Statistics, 38(2), 1466–1474.
36
Eyraud, Henri (1938). Les principes de la mesure des corrélations. Annales de l’Université
de Lyon, Series A, 1, 30–47. 70
Farlie, D. J. G. (1960). The performance of some correlation coefficients for a general bi-
variate distribution. Biometrika, 47, 307–323. 70
Fisher, Nicholas I. (1997). Copulas. In: Samuel Kotz, Campbell B. Read, and David L.
Banks (editors), Encyclopedia of Statistical Sciences, Update Volume 1. Wiley, New York,
pp. 159–163. 2
Frank, Maurice J. (1979). On the simultaneous associativity of f (x, y) and x + y − f (x, y).
Aequationes Mathematicae, 19, 194–226. 84, 85
Bibliography 105
Fréchet, Maurice (1957). Les tableauxs de corrélations dont les marges sont données.
Annales de l’Université de Lyon, Sciences Mathématiques et Astronomie, Serie A(4),
13–31. 11
Galambos, János (1978). The Asymptotic Theory of Extreme Order Statistics. Wiley,
New York. 1, 45
Genest, Christian (1987). Frank’s family of bivariate distributions. Biometrika, 74(3),
549–555. 84
Genest, Christian and Louis-Paul Rivest (1993). Statistical inference procedures for bi-
variate Archimedean copulas. Journal of the American Statistical Association, 88(423),
1034–1043. 24, 25, 96
Georges, P., A.-G. Lamy, E. Nicolas, G. Quibel, and Thierry Roncalli
(2001). Multivariate survival modelling: A unified approach with copu-
las. working paper, Groupe de Recherche Opérationelle Crédit Lyonnais,
[Link] 30
Gould, Henry W. (1972). Combinatorial Identities. A Standardized Set Of Tables Listing
500 Binomial Coefficient Summations. Rev. Ed. Morgantown, W. Va. 48
Gumbel, Emile J. (1958). Distributions à plusieurs variables dont les marges sont données.
C. R. Acad. Sci. Paris, 246, 2717–2719. 70
Heuser, Harro (1988). Lehrbuch der Analysis, Teil 2. Teubner, Stuttgart, fourth edition.
82
Hoeffding, Wassily (1940). Maßstabinvariante Korrelationstheorie. Schriften des Mathe-
matischen Seminars und des Instituts für Angewandte Mathematik der Universität
Berlin, 5(3), 179–233. 11
Ihaka, Ross and Robert Gentleman (1996). R: A language for data analysis and graphics.
Journal of Computational and Graphical Statistics, 5(3), 299–314. 97
Joe, Harry (1997). Multivariate Models and Dependence Concepts. Chapman & Hall,
London. 33, 36
Jogdeo, Kumar (1982). Dependence, concepts of. In: Samuel Kotz, Norman L. Johnson,
and Campbell B. Read (editors), Encyclopedia of Statistical Sciences, volume 2. Wiley,
New York, pp. 324–334. 1
Johnson, Norman L. and Samuel Kotz (1975). On some generalized Farlie-Gumbel-
Morgenstern distributions. Communications in Statistics, 4(5), 415–427. 71
Kamke, Erich (1960). Das Lebesgue-Stieltjes-Integral. B. G. Teubner, Leipzig, second
edition. 18, 21
Kamps, Udo (1995). A Concept of Generalized Order Statistics. Teubner Skripten zur
Mathematischen Stochastik. Teubner, Stuttgart. 80
106 Bibliography
Karatzas, Ioannis and Steven E. Shreve (2000). Brownian Motion and Stochastic Calculus.
Graduate Texts in Mathematics. Springer, New York, second edition. 51, 53, 54, 60, 61,
63, 82, 86, 87, 88, 90
Kimeldorf, George and Allan Sampson (1975). Uniform representations of bivariate distri-
butions. Communications in Statistics, 4(7), 617–627. 2, 44
Lange, Kenneth (1998). Numerical Analysis for Statisticians. Springer, New York. 95
Li, Xin, Piotr Mikusiński, and Michael D. Taylor (1998). Strong approximation of copulas.
Journal of Mathematical Analysis and Applications, 225, 608–623. 15
Mardia, Kanti V. (1962). Multivariate Pareto distributions. Annals Math. Statist., 33,
1008–1015. 44
Meyn, Sean P. and Richard L. Tweedie (1996). Markov Chains and Stochastic Stability.
Springer, London, third edition. 74
Monroe, Itrel (1978). Processes that can be embedded in Brownian motion. The Annals
of Probability, 6(1), 42–56. 91
Nelsen, Roger B. (1991). Copulas and association. In: Giorgio Dall’Aglio, Samuel Kotz, and
Gabriella Salinetti (editors), Advances in Probability Distributions with Given Marginals:
Beyond the Copulas, volume 67 of Mathematics and Its Applications. Kluwer Academic
Publishers, Dordrecht – Boston – London. 49
Oakes, David (1982). A model for association in bivariate survival data. Journal of the
Royal Statistical Society B, 44(3), 414–422. 44
Øksendal, Bernt (1998). Stochastic Differential Equations. Springer, New York, fifth edi-
tion. 86, 87, 88, 90
Randles, Ronald H. and Douglas A. Wolfe (1979). Introduction to the Theory of Nonpara-
metric Statistics. Wiley, New York. 26
Resnick, Sidney I. (1987). Extreme Values, Regular Variation, and Point Processes.
Springer, New York. 77, 78, 80
Revuz, Daniel and Marc Yor (1991). Continuous Martingales and Brownian Motion.
Springer, Berlin. 53, 55, 86, 87
Rogers, L. Chris G. and David Williams (2000a). Diffusions, Markov Processes and Martin-
gales, Volume 1: Foundations. Cambridge University Press, Cambridge, second edition.
4, 17, 18, 47, 51, 53, 86
Rogers, L. Chris G. and David Williams (2000b). Diffusions, Markov Processes and Martin-
gales, Volume 2: Itô Calculus. Cambridge University Press, Cambridge, second edition.
4, 53, 86
Ross, Sheldon M. (1996). Stochastic Processes. Wiley, New York, second edition. 53
Schmitz, Volker and Erwin Pier-Ribbert (2001). Copulas in finance. derivatives week,
10(6), 6–7. 2
Schweizer, Berthold (1991). Thirty years of copulas. In: G. Dall’Aglio, S. Kotz, and
G. Salinetti (editors), Advances in Probability Distributions with Given Marginals, Math-
ematics and Its Applications. Kluwer Academic Publishers, Dordrecht. 2, 22
Shiryaev, Albert N. (1996). Probability. Springer, New York, second edition. 11, 56, 57
Sklar, Abe (1959). Fonctions de répartition à n dimensions et leurs marges. Publ. Inst.
Statist. Univ. Paris, 8, 229–231. 1, 9, 10
Sklar, Abe (1973). Random variables, joint distribution functions, and copulas. Kyber-
netika, 9(6), 449–460. 9
Walsh, John E. (1969). Asymptotic independence between largest and smallest of a set of
independent observations. Ann. Inst. Statist. Math., 21, 287–289. 45
Quadratic variation, 86
Quasi-monotone, see Increasing, n-increasing
14. Mai 1973 In Düren geboren als Sohn von Irene und Klaus Schmitz