The Analytic Hierarchy Process A Mathematical Model For Decision
The Analytic Hierarchy Process A Mathematical Model For Decision
Open Works
Senior Independent Study Theses
2014
Recommended Citation
Nguyen, Giang Huong, "The Analytic Hierarchy Process: A Mathematical Model for Decision Making Problems" (2014). Senior
Independent Study Theses. Paper 6054.
https://openworks.wooster.edu/independentstudy/6054
This Senior Independent Study Thesis Exemplar is brought to you by Open Works, a service of The College of Wooster Libraries. It has been accepted
for inclusion in Senior Independent Study Theses by an authorized administrator of Open Works. For more information, please contact
[email protected].
by
Giang Huong Nguyen
The College of Wooster
2014
Advised by:
Dr. John Ramsay
Abstract
The ability to make the right decision is an asset in many areas and lines of
profession including social work, business, national economics, and
international security. However, decision makers often have difficulty
choosing the best option since they might not have a full understanding of
their preferences, or lack a systematic approach to solve the decision making
problems at hand. The Analytic Hierarchy Process (AHP) provides a
mathematical model that helps the decision makers arrive at the most logical
choice, based on their preferences. We investigate the theory of positive,
reciprocal matrices, which provides the theoretical justification of the method
of the AHP. At its heart, the AHP relies on three principles: Decomposition,
Measurement of preferences, and Synthesis. Throughout the first five chapters
of this thesis, we use a simple example to illustrate these principles. The last
chapter presents a more sophisticated application of the AHP, which in turn
illustrates the Analytic Network Process, a generalization of the AHP to
systems with dependence and feedback.
iii
Acknowledgements
I would like to thank Professor John Ramsay, my advisor, for his tremendous
support with this thesis. His knowledge of the subject, passion for
mathematics, and relentless encouragement were the main source of
motivation that got me through a research project that at times seemed
impossible. Being his student has been the most extraordinary experience in
my sixteen years of schooling. I am most grateful to complete an Independent
Study project under his guidance.
My sincere thanks goes to all of the professors in the Department of
Mathematics and Computer Science. Each of them has offered advice that
helped shape the outcome of this thesis.
I would also like to thank my friends, Dung Nguyen’14 and Chris and
Christy Ranney, for their support. Without them, this thesis would not have
been possible.
Last but not least, thank you, Mom and Dad, for having faith in my
abilities and judgment.
v
Contents
Abstract iii
Acknowledgements v
1 Introduction 1
1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
3.3 Finding the Weight Vector for the Inconsistent Alice Example . . 41
vii
4 Finding the Score of an Alternative on an Objective 45
4.1 Finding the Score of Each Alternative in the Alice Example . . . . 46
4.2 The Hierarchy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
4.3 Extension of the Hierarchy in the Alice Example . . . . . . . . . . 53
5 Metrics 59
5.1 Measure of Consistency . . . . . . . . . . . . . . . . . . . . . . . . 60
5.1.1 Consistency of a Pairwise Reciprocal Matrix . . . . . . . . 60
5.1.2 Consistency Index . . . . . . . . . . . . . . . . . . . . . . . 68
5.1.3 Finding the Consistency Ratio for the Alice Example . . . 70
5.2 Measure of Pairwise Preferences . . . . . . . . . . . . . . . . . . . 72
7 Conclusion 113
List of Figures
ix
x LIST OF FIGURES
List of Tables
xi
Chapter 1
Introduction
1.1 Motivation
We make decisions every day. High school graduates decide which college to
attend. Moms make up their minds as to how much time they should let their
children watch TV each day. Chefs choose the spices to use in dishes. The
Federal Reserve makes judgments about its monetary policy. Decision making
is a part of everyday life, whether it be a personal choice, a decision with
global impact, or anything in between. Being able to make right decisions can
have a tremendously positive impact on individuals’ lives.
However, making the right decision is often easier said than done. For
starters, assuming that the decision makers are aware of their options, they
need to identify their goals in choosing among these options. (As a simple
example, consider a buyer, who is trying to decide between two kinds of
tomatoes. He or she needs to know what qualities in tomatoes he or she
values: the color, the shape, or the price.) Then, the decision makers need to
1
2 CHAPTER 1. INTRODUCTION
determine how well each option satisfies their goals. Even when the decision
makers have clear ideas of their goals and the qualities of their options, their
judgments can still exhibit inconsistency, or can vary according to time.
Variations can also occur due to the environment that surrounds the decision
making process.
Alice is in her last year of high school. She has been accepted to three major
universities and is trying to decide which one to attend. Those universities are
City University, Suburb University, and Town University. In evaluating a
university, Alice considers three factors: the academic quality of the university,
the financial aid package that the university offers her, and the quality of the
town or city that surrounds the university. In our discussion of the AHP, we
will refer to the universities that extend their acceptance to Alice as
alternatives and the three factors as objectives.
CHAPTER 1. INTRODUCTION 3
However, what will happen if Alice has ten, instead of three, objectives?
Then scoring the universities will be a more complicated and error-prone task.
How about the possibility that each objective has a different level of
importance to Alice? Then we have to take into account not only the score of
each university on each objective, but also the weight of each objective to
Alice’s decision. The problem quickly escalates in terms of complexity if Alice
also has different potential majors in mind, and her choice of university affects
her choice of major. For example, one of the universities might not have a
Pre-Health program, but it has a strong Music department (assuming that
Alice is considering studying Pre-Health and Music). Another university has a
prestigious Pre-Health curriculum, but it does not have a Music major. The
third university has both Pre-Health and Music, allowing Alice more
opportunities to explore her interest, but both of the programs are only
average.
to the importance of the other objectives. This relative importance is called the
weight of the objective. For ease of computing, we require that the measured
weights sum to 1 [4, p. 30]. For this example, suppose that the AHP has found
the weights for academic quality, financial aid, and quality of location to be
0.6, 0.3, and 0.1, respectively. According to these weights, Alice considers the
quality of a university’s education the most important factor in making her
decision. Following in importance are the financial aid package that she is
offered, and then the quality of the university’s location.
In the next step, the AHP measures how well each of the decision maker’s
alternatives satisfies each of the objectives. The extent to which each
alternative meets the decision maker’s expectation in each objective is
measured as a numerical value. This value is referred to as the score of the
alternative on that objective. In this example, suppose that the AHP has found
the scores for one of the universities, City University, on academic quality,
financial aid, and quality of location, to be 5, 7.5, and 10, respectively. These
scores indicate that Alice loves the setting of City University, while the
financial aid that she is offered is moderately good, and the academic quality is
only average. We noted earlier that a decision maker’s preferences can exhibit
inconsistency. The scores found by the AHP are not simply the decision
maker’s assessment of his or her own preferences. The AHP achieves these
scores using a systematic approach that will be discussed in Chapter 4, helping
the decision maker make the most logical choice given his or her preferences.
The reader might have guessed that in order for Alice to make a decision,
both the weights of her objectives and the score of each alternative on those
objectives must be taken into account. Indeed, we utilize both kinds of
CHAPTER 1. INTRODUCTION 5
information to compute the total score for each alternative. For the first
alternative, City University, we first take the products of each of City
University’s scores on an objective and the weight of that objective. Taking the
sum of those products will give us the total score of City University. Using the
values in this example, the total score of City University is computed as
follows:
5 × 0.6 + 7.5 × 0.3 + 10 × 0.1 = 6.25.
Given that we also know the score of the other alternatives on Alice’s
objectives, we can compute the total score of those alternatives using the same
method explained above. The alternative with the highest total score should
be chosen.
In subsequent chapters, it will be clear that the AHP involves more than
just finding the weights of the objectives and the scores of the alternatives. In
fact, the two steps outlined above are merely part of a hierarchy that is used to
model the decision makers’ preferences. The same idea of a hierarchy that is
used to solve Alice’s problem could be used to answer questions about
promotion and tenure in higher education [10, p. 162], the optimum choice of
coal plants [10, p. 156], and measuring the world influence of nations such as
the U.S., China, France, the U.K., Japan, . . . [10, p. 40]. For the purpose of
understanding the foundations of the AHP, the next three chapters will
investigate the two outlined tasks: finding the weights of the objectives and
the scores of the alternatives. The discussion of the hierarchy will soon follow.
the first step of the AHP, for the ith objective, the AHP generates a weight wi . In
the second step, for the ith objective and the kth alternative, the AHP obtains a
score sik of the kth alternative on the ith objective. The total score of the kth
alternative is then computed by the following formula:
n
X
wi sik . (1.1)
i=1
After the total scores of all alternatives have been calculated, the decision
maker should choose the alternative that has the highest total score.
This thesis focuses on the method and mathematical reasoning of the AHP.
The next two chapters cover the task of finding the weights of the objectives.
Specifically, in Chapter 2, we will develop Alice’s problem, introducing the
basic terms and concepts of the AHP. As will be clear in this chapter, an
essential concept of the AHP is the consistency of the decision maker’s
preferences. We will include the definition of consistency, and then focus on
the consistent decision maker. In Chapter 3, we investigate the case of an
inconsistent decision maker. We turn our attention to finding the scores of the
alternatives in Chapter 4. The discussion of the scores naturally leads to the
idea of a hierarchy. Having discussed the hierarchy, we apply it to extend our
Alice’s problem. In Chapter 5, we examine some measurement issues of the
AHP. Specifically, we will provide an explanation for the measure of the
decision maker’s consistency, and comment on the scale that is used to
CHAPTER 1. INTRODUCTION 7
In this chapter, we tackle the first task of the AHP: finding the weight of each
objective of the decision maker. We continue with the Alice example presented
in the last chapter, introducing terms and concepts that are fundamental to the
AHP. One such concept is what is called the pairwise comparison matrix. This
concept will naturally lead to the idea of the decision maker’s consistency in
preferences. The AHP problem will then be divided into two sub-problems:
one for the consistent and the other for the inconsistent decision maker. In the
rest of the chapter, we present the method of finding the weight vector for the
consistent case. This method involves the eigenvalue problem, a concept
central to the AHP. Finally, we provide an illustration of the eigenvalue
problem for the Alice example.
It is the writer’s intention to start the discussion of the AHP with a simple,
instructive example such as the problem of choosing university for Alice.
9
10 CHAPTER 2. THE CONSISTENT DECISION MAKER
Returning to the Alice example, the AHP assumes that Alice knows the
pairwise comparisons of her objectives. Suppose she values the quality of a
university’s education twice as much as the financial aid that the university
offers, and six times as much as the quality of the university’s location.
Similarly, financial aid is three times as important to Alice as the quality of
CHAPTER 2. THE CONSISTENT DECISION MAKER 11
1 2 6
A = 21 1 3 .
1 1
6 3
1
The entry in the ith row and the jth column of matrix A is the importance of
objective i compared to that of objective j. For example, a13 = 6 means that the
first objective, academic quality, is six times as important as the third objective,
location. It follows that the entries in the diagonal of matrix A are equal to 1.
Moreover, if objective i is twice as important as objective j, then objective j
must be half as important as objective i. In other words, a ji = 1
aij
. We also note
that if Alice is consistent in her preferences, then ai j a jk = aik . That is, if she
prefers academic quality twice as much as financial aid, and financial aid three
times as much as location, then she must prefer academic quality six times as
much as location to be considered a consistent decision maker. According to
this definition, Alice in our example is consistent. However, a decision maker
in real life is rarely consistent in her preferences. In Chapter 3, we will discuss
the process of obtaining the weights of the objectives in the inconsistent case.
It is in this inconsistent case that the mathematical reasoning becomes more
complicated. For now, we turn our attention to the consistent case.
h i
matrix for a decision maker with n objectives is an n × n matrix A = ai j such
that:
(ii) a ji = 1
aij
for i, j = 1, . . . , n.
w1 w1 w1
w1 w2
··· wn
w2 w2 w2
h i w1 w2
··· wn
A = ai j = .. ,
.. .. ..
. . . .
wn wn wn
w1 w2
··· wn
Pn
where wi > 0 and i=1 wi = 1.
Definition 2.3 leads to the formal definition of the weight vector of the
decision maker:
T
w = [wi ] = w1 w2 · · · wn ,
Pn
where wi > 0 and i=1 wi = 1. The weight vector w is also referred to as the
priority vector of the decision maker.
Theorem 2.1. Suppose a decision maker is consistent and has n objectives. Let A be
the corresponding pairwise comparison matrix, and w the weight vector. Then w is an
eigenvector of A with corresponding eigenvalue λ = n.
14 CHAPTER 2. THE CONSISTENT DECISION MAKER
w1 w1 w1 w1 w1 w1
w1 w2
··· wn
w
1 w
w1 1
+ w
w2 2
+ ··· + w
wn n
w2 w2 w2 w2 w2
+ + · · · + wwn2 wn
w
w1 w2
··· wn 2 w
w1 1
w
w2 2
Aw = .. .. =
.. .. .. ..
. . . . . .
wn
w1
wn
w2
··· wn
wn
w
n
wn
w
w1 1
+ wn
w
w2 2
+ · · · + wwnn wn
nw w
1 1
nw2 w2
= = n . = nw.
.. .
. .
nw w
n n
From Theorem 2.1, we have a way to obtain the weight for each objective,
given that we know the number of objectives n and the pairwise comparison
matrix A. We know this by observing that:
Aw = nw
⇔ Aw − nw = 0
⇔ Aw − nIw = 0
⇔ (A − nI)w = 0.
The equation
Aw = nw (2.1)
The purpose of this section has been to present the eigenvalue problem in
Equation 2.1. We close the section by the next theorem, which provides a quick
and easy way to identify the weight vector when the decision maker is
consistent.
Theorem 2.2. The normalized form of any column of the matrix A = wwij is a
T
solution to the eigenvalue problem Aw = nw, where w = w1 w2 · · · wn is the
weight vector solution and n is the number of objectives.
T
w1
wj
w2
wj
··· wn
wj
,
In the next section, we apply both Theorem 2.1 and Theorem 2.2 to find the
16 CHAPTER 2. THE CONSISTENT DECISION MAKER
weights of Alice’s objectives. Our expectation is that the methods from both
theorems give the same weight vectors.
Example
In this section, we illustrate the process of finding the weight vectors for the
Alice example, using both Theorem 2.1 and Theorem 2.2. By our definition of
consistent pairwise comparison matrices, the matrix A in the Alice example,
which was given in the last section, is consistent. This guarantees that we can
apply both theorems to obtain the weight vectors to solve Alice’s problem.
1 2 6 3 0 0 −2 2 6
A − nI = A − 3I = 21 1 3 = .
1
− 0 3 0
2 −2 3
1 1 1 1
1 0 0 3 −2
6 3 6 3
1 0 −6
The reduced row echelon form of A − 3I is: 0 1 −3 . Using
0 0 0
Gauss-Jordan elimination [8, p. 78], the eigenvectors of A with corresponding
T
eigenvalue 3 have the form 6s 3s s (s ∈ R).
T
eigenvalue 3 is 6 3 1 . Since the weights of all of the objectives sum to 1,
the weight vector is
T
w= 6
10
3
10
1
10
.
19
20 CHAPTER 3. THE INCONSISTENT DECISION MAKER
inconsistent pairwise comparison matrix will not deviate much from the
consistent matrix. We can find the weight vector for an inconsistent decision
maker based on the weight vector in the consistent case. Thus, we want to
further explore the eigenvalues and corresponding eigenvectors of a consistent
pairwise matrix. The next theorem serves that purpose.
Theorem 3.1. Suppose a decision maker is consistent and has n objectives. Then the
pairwise comparison matrix has a unique largest eigenvalue n. All of the other
eigenvalues are zero.
Aw = nw.
Consider:
w w1 w1 w1 w
1
w1 w2
··· wn 1
w1 − w1
0
w
1
w2 w2 w2
0
w1 w2
··· wn
0
w2 − w2
0
0
..
= .. .. ..
..
=
..
= ..
= 0
..
.
A . . . ··· . . . . .
wn−1 wn−1 wn−1
0 ··· 0 wn−1 − wn−1 0 0
w1 w2 wn
wn wn wn −w
−wn w1 w2
··· wn n wn − wn 0 −wn
CHAPTER 3. THE INCONSISTENT DECISION MAKER 21
w
1
0
.
.
. is an eigenvector of A with corresponding eigenvalue
Thus, the vector
0
−wn
λ = 0. Using the same approach, we can prove that the following vectors are
the eigenvectors corresponding to the eigenvalue λ = 0 for A:
w 0 0 0
1
0 w2 0 0
0 0 w3 0
, , , . . . , .
.. .. .. ..
.
.
.
.
0 0 0 wn−1
−wn −wn −wn −wn
| {z }
n-1 vectors
It can be easily seen that these n − 1 vectors are linearly independent, since
none of them could be written as a linear combination of the others. So the
basis of the eigenspace associated with the eigenvalue λ = 0 has at least n − 1
vectors. In other words, the geometric multiplicity of the eigenvalue λ = 0,
which is the dimension of the eigenspace associated with that eigenvalue, is at
least n − 1. Since the algebraic multiplicity of the eigenvalue λ = 0, which is
the number of times its factor occurs in the characteristic polynomial, is
always bigger than or equal to its geometric multiplicity, the factor of the
eigenvalue 0 occurs at least n − 1 times in the characteristic polynomial.
Thus, the matrix A has a unique largest eigenvalue λ = n and all of the
other eigenvalues equal zero.
Our problem for the inconsistent case becomes: given a decision maker
that is inconsistent in her preferences, find the weight vector w0 that satisfies
where λmax is the unique largest eigenvalue for A. In Section 3.2, we will
discuss Perron’s theorem, which guarantees that Equation 3.1 always has a
unique solution.
In this section, we discuss Perron’s theorem for positive matrices. The proof of
this theorem provides the theoretical foundation for the method of finding the
CHAPTER 3. THE INCONSISTENT DECISION MAKER 23
Definition 3.1. A real matrix A is non-negative (or positive) if all of the entries
of A are non-negative (or positive). We write A ≥ 0 (or A > 0).
lim Mk = ev,
k→∞
where v = v1 v2 · · · vn is a positive row vector, ni=1 vi = 1, and
P
T
e = 1 1 ··· 1 .
24 CHAPTER 3. THE INCONSISTENT DECISION MAKER
Proof. Theorem 3.2 states that for large k, Mk approaches a matrix with
identical rows. In order to prove this theorem, we prove that each column of
Mk approaches a column vector with identical entries. Let
T
y0 = y0 y0 · · · y0
1 2 n be an arbitrary column vector in Rn . Define
yk = Mk y0 , with k = 0, 1, 2, . . .. Let ak and bk be the maximum and minimum
components of yk , respectively. Further, let α be the minimum entry in M.
The outline of the proof is as follows: (i) first, we demonstrate that the
sequences (ak ) and (bk ) are monotone, and (ii) bounded. As a result, (ak ) and
(bk ) converge. (iii) Next, we prove that the difference between (ak ) and (bk )
approaches 0 as k approaches infinity. Therefore, (ak ) and (bk ) tend to a
common limit, and all of the components in yk also approach this limit. So yk
approaches a column vector with identical entries. (iv) Lastly, we choose y0 so
that for each y0 , yk represent a column of Mk . Putting everything together,
each column of Mk approaches a column vector whose entries are the same,
and Mk approaches a matrix whose rows are identical.
n
X
j
yik+1 = mi, j yk = mi,1 y1k + mi,2 y2k + . . . + mi,n ynk . (3.2)
j=1
The last line was achieved because the entries in each row of M sum to
unity. Further, since α is the minimum entry in M, α ≤ mi,1 and
1 − α ≥ 1 − mi,1 . Therefore
≥ αak + (1 − α)bk .
≤ αbk + (1 − α)ak .
which is equivalent to
ak+1 − ak ≤ α(bk − ak ).
Because bk ≤ ak for k = 0, 1, 2, . . .
ak+1 − ak ≤ α(bk − ak ) ≤ 0
Similarly, the bounds in Equation 3.3 hold for the smallest component of
CHAPTER 3. THE INCONSISTENT DECISION MAKER 27
yk+1 . As a result
bk+1 ≥ αak + (1 − α)bk , (3.5)
which is equivalent to
bk+1 − bk ≥ α(ak − bk ) ≥ 0,
(ii) The sequences (ak ) and (bk ) are bounded and convergent:
which is equivalent to
a1 − b1 ≤ (1 − 2α)1 (a0 − b0 ),
By induction,
ak − bk ≤ (1 − 2α)k (a0 − b0 ). (3.7)
We showed that (ak ) and (bk ) converge. Therefore, the Algebraic Limit
Theorem [1, p. 45] implies that (ak − bk ) also converges. Next, we
compute the limit of (ak − bk ).
CHAPTER 3. THE INCONSISTENT DECISION MAKER 29
Since M is a positive matrix whose row sums are ones, 0 < α < 1. Thus
0 ≤ ak − bk ≤ (1 − 2α)k (a0 − b0 ).
We showed that lim(ak ) and lim(bk ) exist. Again, by the Algebraic Limit
Theorem
lim (ak − bk ) = 0 = lim (ak ) − lim (bk ).
k→∞ k→∞ k→∞
Thus, (ak ) and (bk ) approach a common limit. Since ak and bk are the
maximum and minimum components of yk , respectively, all of the
components in yk also approach this limit. In other words, as k
approaches infinity, the vector yk approaches a column vector whose
entries are all the same. Since the decreasing sequence (ak ) is bounded
below by b0 and the increasing sequence (bk ) is bounded above by a0 ,
limk→∞ yk = Ce,
30 CHAPTER 3. THE INCONSISTENT DECISION MAKER
T
where b0 ≤ C ≤ a0 and e = 1 1 · · · 1 .
c1 c2 · · · cn 1 1
c1 c2 · · · cn 1 1
lim Mk = .. =
c c · · · c = v v · · · v = ev.
.. .. .. 1 2 .. 1 2
n n
k→∞
. . . . .
c c ··· c
1 2 n 1 1
Since ci is positive, v is a positive row vector. The last item in the proof is
to show that the sum of all of the components of v is 1.
T
Since the rows of M all sum to 1, Me = e, with e = 1 1 · · · 1 . We
prove that Mk is also a stochastic matrix by induction.
Pn
unity, and therefore j=1 v j = 1.
Ak
lim = wv,
k→∞ λk
Proof. Let S be the set of all non-negative, column n-vectors such that the
entries of each of those vectors sum to 1. We denote
n
X
S = {x|x = [xi ]n×1 , xi ≥ 0, xi = 1, i = 1, 2, . . . , n}.
i=1
n
X
f (y) = yi ,
i=1
1
T(y) = Ay,
f (Ay)
The outline of the proof is as follows: (i) first, we demonstrate that the
transformation T is a continuous function, and T(S) ⊂ S. That is for an
arbitrary x ∈ S, T(x) ∈ S. (ii) Next, using the Brouwer Fixed Point Theorem
with the function T and the space S, we are able to find a positive fixed point
w. (iii) Finally, we apply Theorem 3.2 to prove the proposed limit.
n n
X X 1
(T(x))i = ( (Ax))i
i=1 i=1
f (Ax)
1
= f( Ax)
f (Ax)
(Ax)1 (Ax)2 (Ax)n
= + + ... +
f (Ax) f (Ax) f (Ax)
n
1 X 1
= (Ax)i = f (Ax)
f (Ax) i=1 f (Ax)
= 1.
By the Brouwer Fixed Point Theorem [14, p. 277], there exists a fixed
point w ∈ S such that
T(w) = w.
1
Aw = w, (3.8)
f (Aw)
which is equivalent to
Aw = f (Aw)w.
Aw = λw, (3.9)
We observe that
1 1 1
(D−1 ( )AD)e = (D−1 ( )A)De = D−1 ( )Aw = D−1 w = e.
λ λ λ
Thus, D−1 ( λ1 )AD is a row stochastic matrix. Further, since D−1 and D are
non-negative n × n matrices, A is a positive n × n matrix, and λ is
positive, D−1 ( λ1 )AD is a positive n × n matrix. Using Theorem 3.2,
1
lim (D−1 ( )AD)k = ev∗ ,
k→∞ λ
h i Pn
where v∗ = v∗i > 0, and i=1 v∗i = 1.
1×n
1 1 1 1
limk→∞ (D−1 ( )AD)k = lim (D−1 ( )AD)(D−1 ( )AD) . . . (D−1 ( )AD))
λ k→∞ λ λ λ
| {z }
kterms
1
= lim D−1 ( A)k D
k→∞ λ
1
= D−1 (lim k Ak )D.
k→∞ λ
CHAPTER 3. THE INCONSISTENT DECISION MAKER 35
Therefore,
1 k
D−1 (lim A )D = ev∗
k→∞ λk
Ak
⇔ lim = Dev∗ D−1 = wv∗ D−1 .
k→∞ λk
Ak
lim = wv,
k→∞ λk
We are now ready to state and prove Perron’s theorem, which justifies the
existence of a unique largest eigenvalue and a corresponding eigenvector w0
that satisfies Equation 3.1.
1. A has a real positive simple (i.e., not multiple) eigenvalue λmax , whose modulus
is larger than the modulus of any other eigenvalues.
2. Each of the right and left eigenvectors of A corresponding to the eigenvalue λmax
has positive components, and is essentially (to within multiplication by a
constant) unique.
36 CHAPTER 3. THE INCONSISTENT DECISION MAKER
3. The number λmax (also called the Perron root of A) is bounded above by the
maximum row (or column) sum of A, and bounded below by the minimum row
(or column) sum of A.
Proof. We claim that for a matrix A > 0, the real number λ and the vectors w
and v constructed in Lemma 3.1 satisfy Perron’s theorem. We present the
proof in the following steps: (i) We first prove that λ is a eigenvalue of A with
corresponding right eigenvector w and left eigenvector v, with λ, w, and v
constructed in Lemma 3.1. (ii) Next, we prove that λ is the unique, largest
eigenvalue of A, which is the first item in Perron’s theorem. (iii) The proof that
w and v are unique will be presented next. (iv) Lastly, we prove that λ is
bounded above by the maximum row (or column) sum, and bounded below
by the minimum row (or column) sum of A.
1 1 Ak Ak+1
wvA = lim k A = lim k+1 = wv.
λ λ k→∞ λ k→∞ λ
Therefore
wvA = λwv,
which is equivalent to
ewvA = λewv,
CHAPTER 3. THE INCONSISTENT DECISION MAKER 37
in which e = 1 1 · · · 1 is a row n-vector. Since ew is a constant, we
have
vA = λv, (3.10)
We just showed that w and v are the right and left eigenvectors of A,
respectively, with eigenvalue λ. The vectors w and v, as well as λ, are
constructed in Lemma 3.1. In Lemma 3.1, we proved that λ is a positive
real number, w is a positive column vector and v is a positive row vector.
Taking the limit of both sides of the above equality and using the result
of Lemma 3.1, we have: wvu = limk→∞ ( λh )k u.
Since wvu is an column n-vector, the limit on the right-hand side of the
above equality has to exist, which means that limk→∞ ( λh )k has to exist.
Therefore, | λh | < 1, which is equivalent to
Taking the limit of both sides of the above equality and using the result
of Lemma 3.1, we have: limk→∞ ( λ1k Ak u) = limk→∞ u ⇔ wvu = u. Since vu
is a constant, we can set a = vu. Thus
aw = u.
Taking the limit of both sides of the above equation and using the result
Ak
of Lemma 3.1, we have y limk→∞ λk
= limk→∞ y ⇔ ywv = y. Since yw is a
constant, we can set c = yw. Thus
cv = y.
(iv) The eigenvalue λ is bounded above by the maximum row (or column)
sum of A and bounded below by the minimum row (or column) sum
of A:
T
Let e = 1 1 · · · 1 be a column n-vector. The row sums of A are
given by the components of Ae. Let M be the maximum row sum and m
be the minimum row sum of A. Then
me ≤ Ae ≤ Me. (3.11)
Therefore
vme ≤ λve ≤ vMe. (3.12)
Since ve is a positive real number (we proved that v > 0), we can divide
3.12 by ve, yielding:
m ≤ λ ≤ M.
Using Equation 3.9, we have eAw = eλw. Therefore, enw ≤ eλw ≤ eNw,
which is equivalent to
n ≤ λ ≤ N.
Theorem 3.4.
Ak e
lim = w1 ,
k→∞ eT Ak e
For the proof of this theorem, the reader is referred to [10, p. 176]. Theorem
3.4 states that in order to compute the weight vector of an inconsistent
pairwise comparison matrix, we raise the matrix to an arbitrarily large power,
and then divide the sum of each row by the sum of the entries in the matrix .
In the next section, we illustrate this method of finding the weight vector
for an inconsistent decision maker.
CHAPTER 3. THE INCONSISTENT DECISION MAKER 41
Alice Example
In this section, we illustrate the use of Theorem 3.4 with the inconsistent
version of our Alice example. The consistent pairwise comparison matrix A is
modified so that Alice is inconsistent in her preferences:
1 2 5
A = 21 1 3 .
1 1
1
5 3
1 2 5 1 8
9
A1 e = Ae = 12 1 3 2 ,
1 =
1 1 23
5 3
1 1 15
1 2 5 1
421
eT A1 e = 1 1 1 21 1 3 1 = .
30
1 1
5 3
1 1
Applying Theorem 3.4 and approximating the result to five decimal places,
the eigenvector is
8 0.57007
1
A e 30
w = T 1 =
1 9 = 0.32066 .
e A e 421 2
23
0.10926
15
42 CHAPTER 3. THE INCONSISTENT DECISION MAKER
0.58176
2
A e
w2 = T 2 = 0.30896
e A e
0.10928
0.58157
3
A e
w = T 3 = 0.30898
3
e A e
0.10945
0.58155
4
A e
w4 = T 4 = 0.30900
e A e
0.10945
0.58155
5
Ae
w = T 5 = 0.30900 .
5
e A e
0.10945
where > 0 is predetermined. The weights of Alice’s objectives are the entries
of w5 : the weight of the first objective is the first entry of w5 , the weight of the
second objective is the second entry, and so on. From w5 , we can see that the
weights of academic quality, financial aid, and location are 0.58155, 0.30900,
and 0.10945, respectively. These weights are close to the weights in the
CHAPTER 3. THE INCONSISTENT DECISION MAKER 43
consistent case, which were found in Chapter 2 to be 0.6, 0.3 and 0.1. This
indicates that Alice is not too inconsistent in her preferences. The method to
measure the degree of inconsistency will be presented in Chapter 5.
Up until now, we have explained the methods used to find the weight
vectors in both the consistent and the inconsistent sub-problems of the AHP. If
a decision maker is consistent, Equation 2.1 states that the desired weight
vector w is an eigenvector of A with corresponding eigenvalue n. By Theorem
2.2, we are assured that w is simply the normalized form of any column in A.
If the decision maker is inconsistent, we find the largest eigenvector of A
corresponding to the principal eigenvalue λmax . (In other words, we find a
solution to Equation 3.1). Theorem 3.3 (Perron’s theorem) guarantees the
existence of a unique solution to Equation 3.1. In order to find this unique
principal eigenvector, we apply Theorem 3.4: raising A to an arbitrarily large
power, and then dividing each row sum by the sum of the entries in the
matrix. The iterations are stopped when the difference between two resulting
vectors is less than a prescribed value. Since the goal is for the resulting
vectors to converge, a quick way to obtain the weight vector is to raise A to the
2 · i power at the i iteration, for i = 1, . . . , n [10, p. 179].
In the next chapter, we turn our attention to the second task of the AHP:
finding how well each alternative satisfies each objective.
44 CHAPTER 3. THE INCONSISTENT DECISION MAKER
Chapter 4
45
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
46 OBJECTIVE
4.1 Finding the Score of Each Alternative in the
Alice Example
Recall that Alice has been accepted into three universities: City University,
Suburb University, and Town University. Suppose further that we know how
well each university satisfies each objective, compared to how well each other
university satisfies the same objective. The extent to which alternative i
satisfies an objective, compared to the extent to which alternative j satisfies
that same objective, is measured on an integer-valued, 1–9 scale.1 These
1
This scale is the same scale that was used to measure the pairwise comparisons of objectives
in Chapter 2. This scale will be further discussed in Chapter 5.
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
OBJECTIVE 47
pairwise comparative scores are placed into a pairwise comparison matrix for
each objective. For example, for the first objective, academic quality, the
pairwise comparison matrix comparing each pair of the three universities is
1 4 8
B1 = 41 1 3 .
1 1
1
8 3
0.71788
2
B e
s21 = T 1 2 = 0.20459
e B1 e
0.07753
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
48 OBJECTIVE
0.71677
3
B 1
e
s1 = T 3 = 0.20496
3
e B1 e
0.07826
0.71664
4
B 1
e
s41 = T 4 = 0.20509
e B1 e
0.07826
0.71665
5
B 1
e
s51 = T 5 = 0.20509
e B1 e
0.07826
0.71665
6
B1
e
s1 = T 6 = 0.20509 .
6
e B1 e
0.07826
The values of s1 are within our convergence tolerance = 0.00001 after the
sixth iteration. The first entry of this vector is the score of the first alternative
on the objective academic quality, the second entry corresponds to the score of
the second alternative on that objective, and so on. Based on s61 , we know that
the scores of City University, Suburb University, and Town University on the
objective academic quality are 0.71665, 0.20509, and 0.07826, respectively. Note
that just like the weights of the objectives, these scores sum to 1.
We can use the same process to compute the score vectors of the three
universities for the other two objectives. Suppose the pairwise comparison
matrix comparing each pair of the three universities on the objective financial
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
OBJECTIVE 49
aid is
1 1 1
7 4
B2 = 7 1 2 .
4 1 1
2
0.08234
5
B 2
e
s2 = T 5 = 0.60263 .
5
e B2 e
0.31503
Based on this vector, the scores of City University, Suburb University, and
Town University on financial aid are 0.08234, 0.60263, and 0.31503, respectively.
Finally, the pairwise comparison matrix comparing each pair of the three
universities on location is
1 1 4
2
B3 = 2 1 9 .
1 1
1
4 9
The score vector s3 , whose values are within our convergence tolerance
= 0.00001 after five iterations of Theorem 3.4, is
0.30116
5
B e
s53 = T 3 5 = 0.62644 .
e B3 e
0.07239
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
50 OBJECTIVE
Now that we have calculated the weights of the objectives, as well as the
scores of the alternatives on the objectives, we can compute the priorities of
the three alternatives. Applying Equation 1.1, we find the total score of City
University:
0.71665 0.08234 0.30116 0.58155 0.475172829500000
Sw = 0.20509 0.60263 0.62644 0.30900 = 0.374046617500000 .
0.07826 0.31503 0.07239
0.10945 0.150779458500000
One way to think of the computation Sw is that we are weighing the scores
of each university on the objectives by the importance of those objectives.
From the obtained priority matrix, the priorities (or total scores) of City
University, Suburb University, and Town University, are 0.48, 0.37, and 0.15,
respectively. These scores are the same as the scores calculated separately for
each alternative. Based on these scores, Alice should choose to go to City
University, as it has the highest total score.
The similarity of the methods used to compute the weights of the objectives
and the scores of the alternatives might have led the reader to guess that there
is a connection between the roles of the objectives and the alternatives in the
AHP. Indeed, this similarity results from the fundamental idea of the AHP: the
construction of a hierarchy, which consists of different levels, in order to
reflect the various layers of factors that affect the decision making process.
the goal of the decision making process, choosing a university. The second
level consists of the objectives used when choosing a university, academic
quality, financial aid, and location. The third level specifies the alternatives
available to Alice: City University, Suburb University, and Town University.
In Figure 4.1, each line segment represents the influence (or impact) that
an element of a higher level has on an element of a lower level. Through the
intermediary level (which is Objectives in this case), the goal has an impact on
the alternatives. In our Alice example, the influence of the second level on the
lowest level is how well each alternative satisfies each objective. This influence
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
OBJECTIVE 53
The hierarchy for the Alice example can be easily extended to solve a slightly
different problem. Suppose Alice is trying to decide between studying
Engineering and studying Medicine in college. All of the three universities
that she considers offer these two programs, and her choice of major will affect
her choice of university. The two majors add a fourth layer to the hierarchy
called the Sub-alternatives. Figure 4.2 shows this extended hierarchy.
objectives in the second level and the alternatives in the third level.
1 2 1 4 1 1
3
C1 = C2 = C3 = .
1 1 1 1 3 1
2 4
We observe that all of the three matrices are consistent. This is in fact the
case for every 2 × 2 positive, reciprocal matrix, since it always holds that
aik = ai j a jk . The corresponding eigenvectors of these matrices are:
2 4 1
3 5 4
s1 = s2 = s3 = .
1 1 3
3 5 4
In order to compute the priority vector of the two majors, we first put s1 ,
s2 , and s3 into a 2 × 3 matrix S0 . Right-multiplying S0 by Sw, we obtain the
desired priority vector.
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
56 OBJECTIVE
0.71665 0.08234 0.30116 0.58155
2 4 1 0.65371
3 5 4
S0 Sw = 0.30900 = .
0.20509 0.60263 0.62644
1 1 3
3 5 4
0.34628
0.07826 0.31503 0.07239 0.10945
Further, since we have already determined that City University is the most
logical choice for Alice, among the three universities, and Engineering is twice
as strong as Medicine at City University (the entry in the first row and the
second column of C1 is 2), we can easily conclude that Alice should study
Engineering at City University, without using the pairwise comparison
matrices to find S0 . A reasonable argument would be that if the strength of
CHAPTER 4. FINDING THE SCORE OF AN ALTERNATIVE ON AN
OBJECTIVE 57
based on theories in mathematics and are therefore beyond the scope of this
thesis. For more information, the interested reader can refer to [10, p. 14–16].
At this stage, we have investigated most of the foundations of the AHP,
including the idea of hierarchy, and the justification and computation of the
weights of the elements in a hierarchy. However, we still need to fill some
remaining gaps in the theory of the AHP. These are two of the metrics utilized
in the AHP: measure of a decision maker’s consistency, and the integer-valued
1–9 scale that was used to quantify the pairwise comparisons. We will spend
more time developing both of these metrics in the next chapter.
Chapter 5
Metrics
In Chapter 3, we presented the argument made by Saaty [10, p. 179] that when
the entries of a positive reciprocal matrix change by small amounts, then the
eigenvalues change by small amounts. This argument provides the foundation
necessary for us to move from an eigenvalue problem in the consistent case to
the same problem in the inconsistent case. In this chapter, we shall provide the
theoretical justification for our argument. The central content of this
justification is the first metric utilized in the AHP to measure a decision
maker’s consistency.
59
60 CHAPTER 5. METRICS
A0 w0 = nw0 ,
Aw = λmax w. (5.1)
[10, p. 192].
n
X
ai j w j = λmax wi .
j=1
Therefore
n
X wj
λmax = ai j
j=1
wi
w1 w2 wi wn
= ai1 + ai2 + . . . + aii + . . . + ain
wi wi wi wi
X wj
= ai j + 1,
j,i
wi
for each i = 1, 2, . . . , n.
The last line was achieved because by construction, the entries on the
diagonal of the pairwise matrix A are ones. Taking the sum of λmax over n rows
yield:
X wj X wj X wj
nλmax = a1j +1+ a2 j + 1 + ... + an j + 1.
j,1
w1 j,2
w2 j,n
wn
62 CHAPTER 5. METRICS
X wj X wj
nλmax = ai j + ai j + n. (5.2)
1≤i<j≤n
wi 1≤j<i≤n wi
The first quantity on the right-hand side of Equation 5.2 corresponds to the
entries above the diagonal of A, and the second quantity corresponds to the
entries below the diagonal. Combining these two quantities, we have
X wj wi
nλmax = (ai j + a ji ) + n,
1≤i<j≤n
wi wj
which is equivalent to
X wj wi
nλmax − n = (ai j + a ji ). (5.3)
1≤i<j≤n
wi wj
Define
λmax − n
µ= . (5.4)
n−1
We have
λmax − 1 + 1 − n n(λmax − 1) 1 − n
µ= = + ,
n−1 n(n − 1) n−1
CHAPTER 5. METRICS 63
1 X wj wi
µ = −1 + (ai j + a ji ). (5.5)
n(n − 1) 1≤i<j≤n wi wj
wi
ai j = i j , (5.6)
wj
we have
wj 1
a ji = . (5.7)
wi i j
Substituting Equation 5.6 and Equation 5.7 into Equation 5.5, we have
1 X w wj wj 1 w
i i
µ = −1 + ( i j + ).
n(n − 1) 1≤i<j≤n w j wi wi i j w j
Thus
1 X 1
µ = −1 + (i j + ). (5.8)
n(n − 1) 1≤i<j≤n i j
64 CHAPTER 5. METRICS
We note that
1 X 1
lim µ = −1 + lim (i j + )
ij →1 n(n − 1) 1≤i<j≤n ij →1 i j
1 X
= −1 + 2
n(n − 1) 1≤i<j≤n
1 2n(n − 1)
= −1 +
n(n − 1) 2
= 0.
The second to last line was achieved because the number of entries that are
n(n−1)
above the diagonal of A is n−1
i=1 i =
P
2
.
The above limit states that as consistency is approached (i j approaches 1),
µ approaches 0. This suggests that we choose µ as a measure of the decision
maker’s inconsistency. We will show that the general form of µ in Equation 5.4
suffices as a measure of consistency. In order to accomplish this goal, we need
to introduce the next three theorems.
First, we define
i j − 1 i j ≥ 1
δi j =
ji − 1 i j < 1.
δi j ≥ 0,
1 X 1
µ= (i j + ) − 1
n(n − 1) 1≤i<j≤n i j
1 X 1 2n(n − 1)
= ( (i j + ) − ).
n(n − 1) 1≤i<j≤n i j 2
2n(n−1)
In computing the limit of µ, we found that =
P
2 1≤i<j≤n 2. Therefore we
can write µ as
1 X 1
µ= (i j + − 2).
n(n − 1) 1≤i<j≤n i j
If i j > 1, then i j + 1
ij
= 1 + δi j + 1
1+δij
.
Therefore
1 X 1
µ= (1 + δi j + − 2).
n(n − 1) 1≤i<j≤n 1 + δi j
1 X δ2i j
µ= . (5.9)
n(n − 1) 1≤i<j≤n 1 + δi j
With this new notion of µ, we can proceed to prove the next theorems,
which will be helpful to our discussion of the measure of consistency.
λmax − n 1 X δ2i j
=µ= ,
n−1 n(n − 1) 1≤i<j≤n 1 + δi j
Thus,
λmax ≥ n. (5.10)
Then
n−1 2
λmax − n < δ.
2
CHAPTER 5. METRICS 67
1 X δ2i j
λmax − n =
n 1≤i<j≤n 1 + δi j
1 X 2
< δ
n 1≤i<j≤n i j
1 X 2
≤ δ
n 1≤i<j≤n
1 n(n − 1) 2
= δ
n 2
n−1 2
= δ.
2
To arrive at the conclusion above, we have used the fact that δi j > 0 for some i, j
(by construction of δi j ), assuming that A is inconsistent. Consequently,
δ2ij
1 + δi j ≥ 1 for all i,j and 1 + δi j > 1 for at least one i, j. Therefore, 1+δij
< δ2i j . This
justifies the second step of our argument. To arrive at the fourth step, we used
the fact that the number of the entries above the diagonal of a matrix is
Pn−1 n(n−1)
i=1 i = 2
.
Theorem 5.3 states that the deviation of λmax from n depends on the
maximum perturbation δ of the entries in A and the number of objectives n.
Small perturbations in the entries of A would cause λmax to deviate from n by a
small amount. Therefore, we can find the weight vector of the inconsistent
case using the same eigenvalue problem in the consistent case, which is
A0 w0 = nw0 . Theorem 5.3 thus provides the theoretical reasoning for our
method in Chapter 3.
inconsistent. In this case, the entries of A will deviate from the entries of a
consistent matrix by large amounts. The weight vector that results from our
approach will be meaningless. As such, it is necessary that we develop a
measure of consistency. This measure will provide some insight into the
accuracy of the weight vector obtained from our eigenvalue approach. In the
next subsection, we will present this measure of consistency.
λmax − n
µ=
n−1
λmax − n δ2
µ= < .
n−1 2
δ2
Therefore, 2
provides an upper bound for our measure of the consistency
δ2
index. However, we note that if n is large, then µ < 2
even if λmax is far away
from n. Therefore, for a large number of objectives, µ might not provide a
meaningful measure of consistency. For n small, µ provides a reasonable
CHAPTER 5. METRICS 69
measure of how far λmax is from n, and, consequently, how far the decision
maker is from consistency. Saaty suggests that n should be less than 10 [10,
p. 181].
In order to check for consistency, Saaty uses both the consistency index
and another measure called the random index. A random sample of 500
pairwise reciprocal matrices is constructed. Each matrix is generated
randomly, and its entries are subject to the constraints that
(ii) ai j = 1 if i = j, and
(iii) ai j = 1
a ji
for i, j = 1, 2, . . . , n.
The average value of the consistency indexes of these 500 matrices is called
the random index (R.I.). Table 5.1 gives the random indexes and the
corresponding matrix orders [10, p. 21]. Since Saaty suggests using the AHP
when the number of objectives is less than 10, this table only lists the R.I. for
matrices up to order 10.
n 1 2 3 4 5 6 7 8 9 10
R.I. 0.00 0.00 0.58 0.90 1.12 1.24 1.32 1.41 1.45 1.49
The consistency ratio (C.R.) of a matrix is the ratio of the C.I. of that matrix
to the R.I. for the same matrix order. Thus, for a specific decision maker, the
consistency ratio takes into account both the measure of consistency for that
70 CHAPTER 5. METRICS
specific decision maker (through the consistency index), and the measure of
consistency of a random sample of 500 other decision makers (through the
random index). Therefore, the consistency ratio provides a reasonable
measure of the consistency of the decision maker in question.
If the consistency ratio is 0.10 or less, the decision maker is not too
inconsistent and the result obtained by the AHP is acceptable. However, if the
C.R. is larger than 0.10, more serious inconsistency exists and the priority
vector might not provide an accurate solution to the decision making process
[17, p. 788]. We illustrate the use of the consistency ratio with our Alice
example.
where
1 2 5
A = 21 1 3 .
1 1
5 3
1
0.58155
A5 e
w0 = T 5 = 0.30900 .
e A e
0.10945
CHAPTER 5. METRICS 71
From Equation 5.11 and the rule of matrix multiplication, we know that
3
X wj
λmax = ai j ,
j=1
wi
h i
for i = 1, 2, 3, A = ai j , and w0 = [wi ]. Thus, λmax can be computed as the ratio
of any component in Aw0 to the corresponding component in w0 . Computing
these ratios for all pairs of components in Aw0 and w0 yield
3
X wj 1(0.58155) + 2(0.30900) + 5(0.10945)
a1j = = 3.003694708,
j=1
w1 0.58155
3
X wj 1
(0.58155) + 1(0.30900) + 3(0.10945)
a2j = 2 = 3.003694450,
j=1
w2 0.30900
3
X wj 1
5
(0.58155) + 31 (0.30900) + 1(0.10945)
a3j = = 3.003694636.
j=1
w3 0.10945
λmax − n 3.003694 − 3
= = 0.001847.
n−1 3−1
For a matrix of order 3, the random index (as in Table 5.1) is 0.58. Therefore,
the consistency ratio is
0.001847
= 0.003184 < 0.1.
0.58
In this section, we have investigated the theories that allow us to apply the
same eigenvalue approach in the consistent case to the inconsistent case. In
λmax −n
the process, we also derived the consistency index µ = n−1
, which is used
together with the random index to measure the decision maker’s inconsistency.
In the next section, we will discuss the integer-valued scale 1–9 that was used
to represent the decision maker’s preferences for each pair of objectives.
Value of ai j Interpretations
An Application in Medical
Diagnosis
In the previous chapters, we have used the simple Alice example to illustrate
the method of the AHP. The objective of this chapter is to present a more
sophisticated application of the AHP in medical diagnosis. We first introduce a
few concepts that are useful to the execution of this application. Such concepts,
e.g. the supermatrix, are not fundamental elements of the AHP and therefore
are presented in this chapter as extensions of the AHP. The second half of the
chapter is dedicated to the application of the AHP in medical diagnosis.
75
76 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
We make use of this section to develop several key ideas for the medical
application of the AHP. In Chapter 4, we mentioned that the hierarchy for the
Alice example was a linear system: we start with the highest level, and then
extend downward from one level to the next. In terms of the relationship
between two consecutive levels, the higher has an influence on the lower, as
was pointed out in the extended Alice example (where the sub-alternatives
were added). However, in a real-world problem, it is possible that the lower
level of a hierarchy has an influence on a higher level as well, or the elements
in a level have dependent relationships. In fact, the hierarchy for the medical
application that will be presented in the next section exhibits both of those
characteristics. A hierarchic structure with such nonlinear relationship
between layers, or between elements in a layer, is called a system with
feedback. We will soon demonstrate how a linear hierarchy, such as that for
the Alice example, is modified to display the feedback relationship. In order to
do that, we first represent the hierarchy in the form of a network, which shows
clearly the relationship between levels in the hierarchy.
Figure 6.1 shows the network for the Alice example. The network has
three nodes, which correspond to the three levels in the hierarchy given in
Figure 4.1. Each node in turn includes one or more elements. For example, the
node Goal consists of only one element: the goal of choosing a university,
while the node Objectives has three elements: academic quality, financial aid,
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 77
and location. Each arrow in the network represents an influence that the
elements in the higher node has on the elements in a lower node. In this
example, the influence of the elements in the node Objectives on the elements
in the node Alternatives is represented by the scores of the alternatives on the
objectives. Similarly, the influence of the element in the node Goal on the
elements in the node Objectives corresponds to the weights of the objectives
on the goal of choosing a university. In Chapter 4, we expressed these scores
and weights as the score matrix S and the weight vector w, respectively.
Equation 6.1 shows the matrix W for a network with three nodes. The
construction of W is as follows. Let the first, second, and third node of the
network correspond to the first, second, and third row (column) of W,
respectively. Then the ij component in W reflects the influence of the elements
in the j node on the elements in the i node of the network. By influence, we
mean the priorities of the elements in the i node, with respect to the elements in
the j node of the network.
0 0 0
W = w 0 0 .
(6.2)
0 S I
Since w is a weight vector, the entries in w are positive and sum to 1. Each
of the columns in S is itself a weight vector, so S is a column stochastic matrix.
By putting the identity matrix in the 3,3 position of W, we have made W a
column stochastic matrix. Saaty contends that the powers of W will eventually
reach a stable stage, denoted W k , and that by raising W to powers, we will
obtain the desired priorities [13, p. 494]. Since the purpose of this chapter is to
illustrate a sophisticated application of the AHP, we will not investigate the
theoretical justification of this claim. However, for more theoretical discussion,
we refer the reader to [10, p. 206–214] and [11].
The target priority vector can be found as a column in W k or as a
component in W k , depending on the form of W. Thus, we need several
definitions about certain types of matrix that will help us categorize W.
the network, with respect to the elements in the j node, is given by the
component at the n, j position of W k , where n is the number of nodes in the
hierarchy. Thus, the priorities of the elements in the lowest node, with respect
to the elements in the remaining nodes, can always be read from the last row
of components in W. We note that by component of W k , we are also referring
to a matrix, since W k is the result of raising the supermatrix W to powers.
0 0 0 0 0 0 0
0 0 0 0 0 0 0
0 0 0 0 0 0 0
W 3 = .
0 0 0 0 0 0 0
0.47517282950 0.71665 0.08234 0.30116 1 0
0
0.37404661750 0.20509 0.60263 0.62644 0 1 0
0.15077945850 0.07826 0.31503 0.07239 0 0 1
82 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
The component in the 3,1 position of W 3 (in bold) gives the priorities of the
elements in the last node (Alternatives) of the network, with respect to the
element in the first node (Goal). These priorities are precisely what we found
by multiplying Sw in Chapter 4. The component in the 3,2 position of W 3
gives the priorities of the alternatives with respective to the objectives. This
component is exactly the score matrix S that we found in Chapter 4. For the
Alice example, raising the supermatrix W to powers yields the same result as
successively weighing the elements in a hierarchy.
The purpose of our discussion thus far has been to introduce the network
and the supermatrix, so that we can utilize them to solve a decision making
problem in a more complex system. The next subsection will present the
supermatrix approach in light of a system with feedback.
We open this section with a brief discussion of the usefulness of systems with
feedback, which itself necessitates the supermatrix approach to obtain the
desired priority vector. Saaty contends that many problems in the social
sciences have such complex situations that the linear form of a hierarchy fails
to capture. An example is the various forms of organizations that cannot be
put in a hierarchic structure [10, p. 199].
In both the initial and the extended versions of the Alice example, the
desired priority vector was obtained by matrix multiplication. In other words,
the priorities of the elements in the lowest level, with respect to the element in
the highest level of the hierarchy, were obtained by successively weighing the
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 83
priorities of the elements in each level with respect to the elements in the level
immediately above. In the initial Alice example, the priorities of the three
alternatives on the goal were obtained by weighing the priorities of the
alternatives on the objectives, and then the priorities of the objectives on the
goal. In terms of matrix multiplication, this successive weighing process is
represented by Sw. However, this approach is impossible to execute in a
system with feedback, since we no longer have a hierarchy extending from
one level down to the next. We now have a network, in which it is no longer
clear which node is lower and which is higher. Furthermore, there might be
more than one interaction between the elements in two nodes. The
supermatrix comes to our rescue. A supermatrix enables us to express and
compute two-way interactions between elements in different nodes, as well as
the relationship among elements in the same node.
Figure 6.2 illustrates the network for a system with feedback. As in the
network for the Alice example, each arrow represents the influence that the
elements in the node at the starting point of the arrow has on the elements in
the node at the ending point of the arrow. In this network, the elements in the
node Objectives exhibit dependent relationship. The matrix multiplication
approach used in Chapter 4 has no procedure to take into account this
relationship while deriving the priorities of the alternatives. The supermatrix
approach, however, offers a solution to this problem. Let us take a look at the
corresponding supermatrix for this system with feedback:
84 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
This rule for construction of the supermatrix extends to the general case, as
given in the next definition.
N1 N2 ··· Nn
N1 W11 W12 ··· W1n
N2 W21 W22 ··· W2n
W = . . .. .. .. ,
.. .. . . .
N W Wn2 ··· W
n n1 nn
In this section, we apply the ideas developed in the first half of the chapter to
solve a problem in medical diagnosis. The case study in this section, as well as
the data on the pairwise comparison matrices and the supermatrix, is drawn
from an article by Saaty [13].
88 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
• Anemia (An),
At this point the physicians considered four possible diseases that could
cause the symptoms:
• Lupus,
For more information on these medical terms, the reader can consult the
encyclopedic reference [3].
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 89
Given the condition of the patient, the physicians need to make a decision
between two alternatives: to terminate the pregnancy, or to treat the patient
for her symptoms and let her proceed with the pregnancy. The network for
this case study is given in Figure 6.3.
example, [5], [15], and [9].) However, the use of Bayes’ theorem in diagnosing
diseases requires an assumption that is seldom satisfied in real life: the
symptoms are independent of one another [6].
Saaty suggests that the reason for this assumption is the lack of
information on the relationship of symptoms, as well as the daunting task of
conducting a sufficient number of experiments to obtain the necessary
statistical data to apply Bayes’ theorem [10, p. 492]. Using the AHP, on the
other hand, allows the incorporation of physicians’ judgment in order to take
into account the relationship among symptoms.
From Figure 6.3, the interactions among the elements in the nodes of the
network for the case study are described as follows.
(i) Each symptom observed in the patient has an influence on the possible
diseases. This influence is interpreted as the likelihood of the diseases,
given the symptom;
(ii) Each possible disease has an influence on the observed symptoms. This
influence means the extent to which the symptoms are characteristic of
the disease;
(iv) Each possible disease has an influence on the alternative treatments. This
influence is precisely the priority of the alternative treatments, given the
possible diseases.
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 91
0
and α2 , we will make the block W21 in W column stochastic.
W31
where the ij entry in M is the weight of the i node with respect to Diseases,
compared to the weight of the j node with respect to Diseases. In particular,
the entry m21 reflects the importance of knowing about the alternatives,
compared to the importance of knowing about the symptoms, in diagnosing the
T
disease(s). Then the eigenvector u = u1 u2 of M such that u1 + u2 = 1
gives the weights of the nodes Symptoms and Alternatives on the node
Diseases. By our construction of M, we have that u1 = α1 and u2 = α2 .
The reader must have realized that this process of finding the weight
vector for the nodes Symptoms and Alternatives, with respect to Diseases,
resembles the process of finding the priority vector that we discussed in
Chapter 2. Indeed, these two procedures are the same. In Chapter 2 and
Chapter 3, we discussed the method to find the weights of the objectives on
the goal, and of the alternatives on an objective. In light of our recent
discussion of network, what we did in those chapters was finding the weights
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 93
0
β1 W12 0
T = α1 W21 β2 W22 0
α2 W31
0 I
is column stochastic. Each of W12 , W21 , W22 , and W31 in T reflects the weights
of the elements in a node, with respect to the elements in another node. Each
of the coefficients α1 , α2 , β1 , and β2 reflects the weight of a node, with respect to
another node. Therefore, T represents all of the interactions of the factors in
this case study. Raising T to powers would give us the target priorities of the
alternatives.
Second, we note that the question asked in finding α1 and α2 does not
make much sense. Certainly, knowing about the alternatives, which can only
be "terminate" or "not terminate the pregnancy", does not have any
contribution whatsoever to the diagnosis of a disease. Thus, α2 must be 0,
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 95
Figure 6.4: Reduced Network for the Case Study in Medical Diagnosis
Diseases Symptoms
Diseases 0 β1 W12
Q=
,
(6.3)
β2 W22
Symptoms W21
Figure 6.5: Linear Network for the Case Study in Medical Diagnosis
linear network as in Figure 6.5, provided that we have obtained the priorities of the
possible diseases.
The case study becomes a problem that can be solved in three steps:
(a) Finding the priorities of the observed symptoms with respect to the
possible diseases. This is W21 in Q.
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 97
(b) Finding the priorities of the possible diseases with respect to the
observed symptoms. This is W12 in Q.
(c) Finding the priorities of the observed symptoms with respect to one
another, which is W22 in Q.
rest of this chapter will be devoted to finding the priorities of the alternatives
using the presented approach.
In this subsection, we explain how Saaty found the likelihood of the possible
diseases. The method is to apply the supermatrix approach to the network in
Figure 6.4. Recall that the supermatrix associated with this network is Q given
in Equation 6.3. Each of the nonzero components in Q is a matrix of priorities,
comprised of principal eigenvectors of pairwise comparison matrices. We
illustrate the process to find these components below.
For (the given) disease and for two symptoms, which symptom is more
characteristic of the disease, and how much more is it?
method presented in Chapter 2 and Chapter 3 are used to obtain the principal
eigenvector of the pairwise comparison matrix, which gives the priorities of
symptoms with respect to the disease.
The process outlined above follows closely what we have described in the
first five chapters. In the rest of this chapter, we will encounter more
opportunities in which our knowledge of the pairwise comparison matrix and
the procedure to find the priority vector is utilized. In such situations, our
discussion in Chapter 2 through Chapter 5 applies.
For the disease Lupus, the pairwise comparison matrix A that provides
comparisons on the pairs of symptoms is
For example, the entry in the fifth row and the sixth column in A is 9, so
ANA-H is absolutely more characteristic of Lupus, compared to ACA-H. The
symptom ABL was not included in this matrix because ABL is not
characteristic of the disease Lupus. The priority of ABL with respect to Lupus
is therefore 0. The weight vector w, which is the principal eigenvector of A,
100 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
gives the priorities of all of the symptoms, except for ABL, with respect to
Lupus.
T
w1 = 0.156 0.050 0.046 0.046 0.630 0.072 .
For the other diseases, the original article does not provide the pairwise
comparison matrices, but information about the corresponding eigenvectors is
available. Next, the priority vectors are placed into Table 6.1. The entries of w1
are placed in the first column of the table.
In Table 6.1, the entries in the j column give the priorities of the symptoms,
with respect to the j disease. In terms of our case study, these entries represent
the relative probabilities that the symptoms are observed, given the j disease.
By relative probabilities, we mean, for example, the probability that the
symptom An, out of all of the other symptoms, is observed in the patient if she
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 101
has Lupus. According to the first column of the table, the probability that An
is observed in the patient if she has Lupus is 0.156. Thus, by looking at all of
the entries in a column, we can identify the symptom that is most prevalent for
a given disease. From the first column of the table, given that the disease is
Lupus, ANA-H is the most likely symptom to be exhibited, with a probability
of 0.63.
We note that Table 6.1 gives the entries in W21 . Since the table is obtained
from the eigenvectors of the pairwise comparison matrices, such that the
entries of each vector sum to 1, each column sum of the table is 1, and W21 is
column stochastic.
Before moving on to the next step, we note that the probabilities in Table
6.1 are not for a general, hypothetical pair of disease-symptom, but for the pair
of disease-symptom pertinent to the patient in this case study. This means that
the physicians whose judgments allowed the derivation of these probabilities
gave their answers based on their knowledge both of the disease-symptom
pair, and of the patient’s medical history and current condition. This subtle
difference also applies to the probabilities found in the next two steps.
For (the given) symptom and two diseases, which disease is more likely to exhibit
this symptom, and how much more likely is it?
For the symptom An, the pairwise comparison matrix B that provides
paired comparisons of all the diseases, with respect to An, is
For example, the entry in the 2,1 position of B is 5. This means that TTP is
strongly more likely than Lupus to exhibit An as a symptom. The principal
eigenvector of B gives the priorities of all of the diseases, with respect to An:
T
w2 = 0.052 0.299 0.350 0.299 .
The pairwise comparison matrices for the other symptoms are not
available in the original article. The priority vectors are available and were
placed into Table 6.2. The entries of w2 are placed in the first column of the
table.
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 103
In Table 6.2, the j column gives the priorities of the possible diseases with
respect to the j symptom. In other words, the entries in the j column
represents the relative probabilities that the possible diseases cause the j
symptom. Relative probabilities mean, for example, the probability that the
disease HELLP, among all of the other possible diseases, causes the symptom
LP. From the table, the probabilities that HELLP causes LP is 0.231, which
corresponds to the entry in the 3,2 position in the table. Therefore, by looking
at all of the entries in the j column, we can identify the disease that is most
likely to cause the symptom j. For example, by looking at the second column
in the table, we know that the disease that is most likely to cause LP is TTP,
with a probability of 0.461.
Table 6.2 gives the entries in W12 . Since each column in Table 6.2 is
obtained from a priority vector, the sum of the entries in each column is 1, and
W12 is column stochastic.
The entry in the ij position of this matrix reflects the likelihood that the i
symptom occurs at the same time as the given symptom, compared to the
likelihood that the j symptom occurs jointly with the given symptom.
Likelihood means the chance that a symptom occurs at the same time as the
given symptom, as judged by the physicians. The physicians answered the
following question in order to help derive the pairwise comparison matrix:
Given a symptom, e.g., ANA-H, and two other symptoms that may be related to
it, e.g., An and LP, which of the two latter symptoms is more likely to be associated
with, or occur jointly with, the given symptom? How much more likely is it?
ANA-H An LP BC ACA-H
An 1
1 4 4
LP 1 1 1 1
C= .
1
BC
4 1 1 1
ACA-H 14 1 1 1
For example, the entry in the first row and the second column in C is 1,
which means that An is as equally likely as LP to occur jointly with ANA-H.
Note that ABL and APTT-H were omitted from C because these symptoms are
not related to ANA-H, as judged by the physicians. ANA-H is also skipped,
since the question asks about a symptom and two other symptoms that are
different from the original one. The priorities of ABL, APTT-H, and ANA-H,
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 105
T
w3 = 0.455 0.235 0.155 0.155 .
The pairwise comparison matrices for the other symptoms are not
available from the original articles. The priority vectors are available and
placed into Table 6.3. The entries of w3 are placed in the sixth column of this
table.
The j column in Table 6.3 gives the priorities of the observed symptoms
with respect to the j symptom. These priorities mean the relative probabilities
that given symptom j, the other symptoms that are observed in the patient are
also present. In particular, the entry in the i row and the j column of the table
represents the relative probability that the i symptom occurs jointly with the j
106 CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS
Recall that β1 is the weight of the node Diseases on the node Symptoms,
and β2 is the weight of the node Symptoms on the node Symptoms. To
estimate β1 and β2 , we ask: Is knowledge of a symptom and its usefulness in making
the diagnosis more a direct result of knowing about the diseases, or of knowing about
the other symptoms, and how much more so?
The original article solves the case study problem with β1 = β2 = 0.5, which
implies that knowledge of the diseases and knowledge of the observed
symptoms both contribute equally to knowledge of a symptom and its
usefulness in making the diagnosis. The article then solves the problem with
β1 = 0.99, and β2 = 0.01, which means that knowledge of the dependent
relationships among symptoms contributes only 1 percent to the final
diagnosis. The results in both cases vary by an insignificant amount. (The
priority for one of the alternative treatment decreases by 0.03 if β2 decreases
from 0.5 to 0.01.) Since our purpose is solely to illustrate an application of the
extensions of the AHP, we choose to pursue only the case when β1 = β2 = 0.5.
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 107
T
w= 0.073 0.068 0.087 0.106 0.169 0.144 0.063 0.067 0.066 0.088 0.070 .
T
w0 = 0.218 0.203 0.262 0.317 .
The first, second, third, and fourth entry in w0 represents the relative
likelihood of Lupus, TTP, HELLP, and ACA Syn, respectively. This is the target
result for the first step in the three-step solution to the case study. The second
step is explained in the next subsection.
the Diseases
The j column of Table 6.4 gives the priorities of the alternatives with
respect to the j disease. These priorities are interpreted as the extent to which
the alternative is appropriate for the given disease. For example, from the first
CHAPTER 6. AN APPLICATION IN MEDICAL DIAGNOSIS 111
column of the table, the appropriate treatment for Lupus is to not terminate
the pregnancy (NT), with priority 0.8.
We note that the entries in Table 6.4 give the entries in W31 . Even though
W31 was not entered into the supermatrix Q, it is the priority matrix desired in
step 2 of the solution to the case study. In the next subsection, we illustrate the
final step of the solution.
This is the last step in the solution to this case study. It uses the results from
the first two steps. The method is to apply the matrix multiplication approach
presented in Chapter 4 to the linear hierarchy in Figure 6.5.
Recall that to obtain the priorities of the elements in the lowest level of a
hierarchy, with respect to the element in the highest level, we successively
weigh the priorities of the elements in each level with respect to the elements
in the level immediately above. In the initial Alice example in Chapter 4, we
weighed the priorities of the alternatives on the objectives, and then the
priorities of the objectives on the goal. This successive weighing process is
represented by right-multiplying the score matrix S by the weight vector w.
The result was the priorities of the alternatives with respect to the goal.
0.218
0.200 0.800 0.800 0.833 0.203 0.68
W31 w0 = = .
0.800 0.200 0.200 0.167 0.262 0.32
0.317
The priority of T is 0.68, and the priority of NT is 0.32. Thus, the result of
the AHP recommends that the pregnancy be terminated. Saaty noted that this
recommendation was in agreement with the decision made by a doctor of the
patient [13, p. 500]. Even though this fact does not guarantee the infallibility of
the AHP, it gives some insight into its validity as a mathematical model for
solving decision making problems.
Chapter 7
Conclusion
The case study in the last chapter is in fact an illustration of the Analytic
Network Process (ANP). The ANP is a generalization of the AHP to the case
in which there exists dependence and feedback among factors in decision
making problems. In particular, the supermatrix approach is the
generalization of the hierarchy approach. Regardless, both the AHP and the
ANP rely on these three steps: Decomposition, Measurement of preferences,
and Synthesis [p. 492][10].
113
114 CHAPTER 7. CONCLUSION
The last step, Synthesis, occurs after we have obtained the priority
matrices for all of the interactions in the hierarchy (or network). The
supermatrix approach is the synthesizing step for the ANP, while for the AHP,
we successively weigh the priorities of the levels, from the top to the bottom of
the hierarchy.
Recall that the AHP yields meaningful results only when the number of
objectives is less than 10. Thus, the AHP is not particularly advantageous
when the decision making problem involves a very large number of objectives.
Moreover, the priority vector provides a meaningful result only when the
decision maker is not too inconsistent. Specifically, we found in Chapter 5 that
the consistency ratio needs to be 0.10 or less in order for the result of the AHP
to be acceptable.
The strength of the AHP, as well as the ANP, lies precisely in the three
CHAPTER 7. CONCLUSION 115
Due to time constraints, this thesis has not investigated in depth the
theoretical justification of the ANP. Researchers who wish to pursue further
work on the AHP might find it worthwhile to further explore the theory
behind the ANP. So far, two issues have emerged as worthy of future research.
The first issue is the explanation of the stable stage of the supermatrix. An
interesting fact, which is perhaps also useful for future research, is that the
method of the ANP parallels that of Markov chains, as Saaty himself notes in
[10, p. 206]. Specifically, the nodes in a network correspond to the states in
Markov chains, and the influence of the i node on the j node at time k
corresponds to the transition from state i to state j at time k. Further, the
concepts "priority" in the ANP and "probability" in Markov Chains coincide.
For a more comprehensive list of the correspondence of the terminologies
between the two systems, the reader is referred to the cited source.
Second, recall that the case study was solved by replacing the original
network (Figure 6.3) with a linear hierarchic structure (Figure 6.5), through the
reduction of the original network to a simpler one (Figure 6.4). The author of
this thesis based this approach on the discussion of the AHP in the thesis. This
can be viewed as an alternative explanation to the approach used by Saaty,
which is the same as the three-step solution presented in the last chapter.
116 CHAPTER 7. CONCLUSION
[3] Robert W. Colman. Hemostasis and thrombosis: basic principles and clinical
practice. Lippincott Williams & Wilkins, Philadelphia, PA, 2006.
[4] Bruce L. Golden, Edward A. Wasil, and Patrick T. (Eds.) Harker. The
Analytic hierarchy process: applications and studies. Springer-Verlag, Berlin,
1989.
117
118 BIBLIOGRAPHY
[8] David Poole. Linear algebra: a modern introduction. Available 2011 Titles
Enhanced Web Assign Series. Cengage Learning, Boston, MA, 2010.
[9] Martin J. Prince. Predicting the onset of alzheimer’s disease using bayes’
theorem. American Journal of Epidemiology, 143(3):301–308, 1996.
[10] Thomas L. Saaty. The analytic hierarchy process: planing, priority setting,
resource allocation. Analytic Hierarchy Process Series. RWS Publ.,
Pittsburgh, PA, 1990.
[11] Thomas L Saaty. The analytic network process: decision making with
dependence and feedback. RWS Publ., Pittsburgh, PA, 2001.
[12] Thomas L. Saaty and David S. Turner. Prediction of the 1996 super bowl:
An application of the ahp with feedback (the supermatrix approach). In
Fourth International Symposium on The Analytic Hierarchy Process, pages
12–15, 1996.
[15] Steven J. Skates, Feng-Ji Xu, Yin-Hua Yu, Kerstin Sjövall, Nina Einhorn,
Yuchiao Chang, Robert C. Bast, and Robert C. Knapp. Toward an optimal
algorithm for ovarian cancer screening with longitudinal tumor markers.
Cancer, 76(S10):2004–2010, 1995.