Game Theory For Control of Optical Network PDF
Game Theory For Control of Optical Network PDF
Series Editor
Tamer Başar, University of Illinois, Urbana-Champaign, IL, USA
Game theory has recently been enjoying increased popularity in the research com-
munity, as it provides a new perspective on optimization, networking, and dis-
tributed control problems. It incorporates paradigms such as Nash equilibrium and
incentive compatibility, it can help quantify individual preferences of decision-
makers, and it has an inherently distributed nature. Consequently, game theoretic
models have been applied to a variety of problem domains ranging from economics
to computer science, networking, and security. Game theoretic models provide not
only a basis for analysis but also for design of network protocols and decentralized
control schemes. This makes it attractive as a theoretical framework for the design
of networks and in particular communication networks. Applications range from
power control in wireless communications to congestion control for the Internet, as
well as sensor or ad-hoc networks.
As the backbone of the Internet, optical networks provide huge bandwidth and in-
terconnect countries and continents. Unlike conventional networks, which are well-
established and even standardized, the optical networks field is much younger, and
in fact still evolving. Designing and setting up networking applications within the
optical context is inherently more difficult than in conventional wired or wireless
networks, in part due to the more complex physical-layer effects, and in part due to
the lack of automatic methodologies developed for optical networks. But in these
respects, the optical networking field has seen a tremendous growth in recent years.
From a game theory and control perspective, there are a multitude of problems to
be tackled in the optical networks area and the field is still in its infancy. Aside from
the fact that the area is much younger, research work requires an unusual blend of
interdisciplinary expertise. In this spirit, the present monograph draws the author’s
research background in control theory, practical industrial experience in optical net-
works, and more than five years of research in game theory and control in optical
communications.
The book is focused on mathematical models, methodologies, and game theory
for optical networks from a control perspective. The general setup is that, given an
optical communication network and some performance measures to be optimized
among many networks units/players/channels, one must design an algorithm that
v
vi Preface
achieves as good a performance as possible for each channel. The algorithm should
be decentralized and provably convergent, and should use localized and decentral-
ized feedback. By regarding channels in the network as players in a game this multi-
objective optimization problem fits within the setup of game theory.
The first theoretical question of interest is how to formulate meaningful and
tractable game theoretical problems that can be used as a basis for developing such
algorithms, taking into account the various mathematical characteristics arising nat-
urally in optical networks. These mathematical characteristics are imposed by phys-
ical constraints and by network specific features and topologies. They translate into
constraints on the game formulation (separable versus coupled action space), con-
straints on player’s interaction (localized versus global interaction, i.e., one shot
game versus stage or partitioned game), and constraints on players’ actions (global
constraints versus propagated/modified constraints).
Thus, one of the characteristics of optical networks is that game theoretic for-
mulations and results cannot be transferred directly from other application domains
(e.g., wireless networks or congestion control). Due to inherent physical complex-
ities, the optical area uncovers new theoretical problems, such as games with cou-
pled utilities and coupled constraints. Moreover, in the networking setup these con-
straints are modified across the network links and a framework for dealing with
network games has to be developed. This makes game theory in optical networks
an excellent starting point that could open new research problems and that could be
generalized to other classes of games in networks.
This monograph has a two-fold aim. Its first goal is to provide researchers in
the control and game theoretic community with background on the rich problems
and the initial results in this area. There is a broad scope for fundamental control
and game theoretical research; the hope is that the book will provide background
material such that non-specialists in optical networks can approach these research
problems within the optical networking domain. The book’s second goal is to pro-
vide researchers in the networking and optical community with game theoretical
methodologies that could be used to solve optical networking problems.
The following topics are covered. In the first part of the book, two chapters
present non-cooperative game theory background and some new mathematical re-
sults for Nash equilibria computation in games with coupled constraints. In the sec-
ond part, background and mathematical models for optical networks are presented,
followed by game theory formulations developed for various topologies. The basic
game considered is a power control game in the class of games with continuous
action spaces, coupled utilities in normal form. The book in fact introduces in a sys-
tematic, gradual way the different types of game theoretical problems: first games
with no coupled constraints in normal form, then games with coupled constraints;
all-to-all interaction versus localized player interaction, which leads to games in
ladder-nested form; multi-link (single-sink) and finally mesh topologies and how to
deal with games in such scenarios, while building on the simpler cases and results.
The third part considers issues such as robustness and time-delay effects, as well
as other types of game problems in optical networks, including routing and path
coloring games.
Preface vii
Much of the work in game theory applications has been within the framework of
networking or computer science fields. Essentially, this book uses a mathematical
approach developed from a control theoretical perspective. This is different from
a networking approach, which is typically application-specific and focused on par-
ticular protocols, and unlike algorithmic-only approaches as typically used in com-
puter science. Game theoretical problems are mathematically formulated herein in
a systemic manner, with analytic conditions derived for existence and uniqueness of
Nash equilibrium. Based on these, iterative algorithms with provable convergence
properties to Nash equilibria are developed. The control theoretic approach to games
as taken here allows for the treatment of important issues such as stability and time-
delay effects in a dynamic system context. The abstract mathematical models and
results could be applied to other application domains.
I wish to acknowledge and express my thanks to the many people who influ-
enced and shaped my work over the past several years and thus indirectly helped me
in writing this book: Tamer Başar, Tansu Alpcan, Peter Caines, Roland Malhamé,
and Eitan Altman from the game theoretical and control community; and Stewart
Aitchison, Li Qian, and Dan Kilper from the optical communications community.
I am indebted to my graduate students, Yan Pan, Nem Stefanovic, and Quanyan
Zhu, with whom many of the research work was done and without whom this book
would not have been possible.
Toronto Lacra Pavel
Contents
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Game Theory in Networks . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Optical Networks . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Scope of This Monograph . . . . . . . . . . . . . . . . . . . . . . 5
ix
x Contents
xiii
Chapter 1
Introduction
strategy games [14]. Applications include power control problems in different mul-
tiuser environments [8, 41, 61, 72, 118, 133], routing [12, 99] or congestion con-
trol [4, 11, 13, 21, 76, 129, 139, 160], extending the system-based optimization
approaches [68, 80–82, 87, 153]. The many “players” interact within the network
and make (sequential) decisions, i.e., play a game. For example, in a noncooper-
ative (Nash) game framework the natural players/agents can be the Internet ser-
vice providers (ISP) or domain operators [79, 83], the routers [12, 21, 99, 162], or
even the users themselves in an access network application with dedicated wave-
lengths [112]. As another example in wired networks, there could be two sets of
players: telecom firms/ISP and end users. Both sets of players have different objec-
tives and non-negligible interaction across players exists. In wireless networks there
could be wireless LANs where users/players communicate with a fixed access point,
or wireless ad-hoc networks where users/players communicate with each other in
the absence of any fixed infrastructure support.
In a noncooperative (Nash) game [20, 96, 128] each player pursues the maxi-
mization of its own utility, or equivalently the minimization of its own cost func-
tion, in response to the actions of all other players. The stable outcomes of the
interactions of noncooperative selfish agents correspond to Nash equilibria. On the
other hand, in a cooperative game framework the natural players/agents can be the
network nodes, routers or switches (as software agents), or users [9]. These play-
ers/agents cooperate to redistribute the network resources (bandwidth, wavelength
capacity, power).
Why game theory? Consider the case of a multiple access network problem; most
optimization-based approaches find the optimal multiple access control (MAC) and
routing parameters that optimize network throughput, lifetime, delay etc. and as-
sume all nodes in the network use these parameters. But there is no reason to believe
that nodes will adhere to the actions that optimize network performance. Cheaters
may deviate in order to increase their payoffs which in turn affects other users. In
effect any scenario where there is some strategic interaction among self-interested
players is best modeled via a game-theoretic model. Game theory helps to capture
this interaction, the effect of actions of rational players on the performance of the
network. Although the selfish behavior of players causes system performance loss
in a Nash game [1, 38, 64, 73, 122], it has been shown in [64, 134] that proper
selection of network pricing mechanism can help preventing the degradation of net-
work system performance. In the context of evolution, a game captures the essential
features where strategic interactions occur. The basic idea is that actions which are
more “fit”, given the current distribution of actions, tend over time to displace less
fit actions [47, 49, 50].
to be processed outside the optical domain. Optical line terminals (OLT) are net-
work elements that interface with the optical network clients (SONET, IP, ATM)
and are comprised of transponders (Tx/Rx pairs), as well as optical multiplex-
ers/demultiplexers. Individual powers of the channels can be adjusted at the signal
sources (Tx) as well as at some intermediary nodes. Optical amplifiers (OAs) are
used to amplify the channel powers every 80 to 100 kilometers in an optical link.
Recent years have witnessed the evolution from point-to-point WDM fiber links to
switched optical networks (also referred to as wavelength-routed optical networks).
Channels can be rerouted dynamically through different links within the network
using optical cross-connects (OXCs) or optical add/drop multiplexers (OADMs).
See the List of acronyms for commonly used ones. By using direct photonic links
and optical-domain connectivity unnecessary and expensive optical to electrical to
optical (o/e/o) conversion is avoided.
In static optical networks, lightpaths are statically provisioned up front. The op-
erational expense of reconfiguring the network is quite high and such reconfigura-
tion is time-consuming in the sense that any change in the virtual topology, i.e., the
topology needed for a set of transmission requests, needs manual reconfiguration
and additional equipments. Evolution of emerging tunable network element (such
as reconfigurable OADMs, OXCs, and tunable lasers) make it possible to realize
reconfigurable optical networks [131, 152, 156, 166]. Specific for these is that net-
work reconfiguration can be quickly performed under software control for arbitrary
virtual topologies (control plane layer, see Fig. 1.2). As important is the fact that
the length of fiber links and the number of optical amplifiers deployed along a link
can be changing. Different channels may travel on different fiber links and arbitrary
virtual topologies can be formed over the transport layer (see Fig. 1.2).
Then the interesting question is how to deploy reconfigurable optical networks in
which quick reconfiguration is realized under software control and at the same time,
the network stability and channel transmission performance are maintained. Typical
optical network topologies are mesh, consisting of interconnection of multiple links,
which complicates network performance analysis and network control. This book is
aimed at understanding how such control algorithms can be designed for optical
networks from a game-theoretic perspective.
1.3 Scope of This Monograph 5
From a game theory and control perspective, there are a multitude of problems to be
tackled in optical networks and the field is still in its infancy. An important problem
is how to control the channel transmission performance in optical networks while
incorporating physical-layer impairments. More precisely, transmitted signals de-
grade in quality due to physical-layer impairments (attenuation, noise, dispersion,
nonlinear effects). Channel transmission performance at the physical layer (trans-
port layer) is characterized through the bit-error rate (BER), which is defined as
the average probability of incorrect bit identification. Performance degradation is
associated with an increase in BER, which in turn depends on optical signal-to-
noise ratio (OSNR), dispersion, and various nonlinear effects [2]. While a channel
traverses a number of optical elements (e.g., OAs and OXCs) through the network
(see Fig. 1.3), it may degrade in quality due to these impairments. Moreover, signals
of new established channels affect the quality of existing channels on their shared
fiber links. An optical network is considered as a dynamic multiuser environment, if
the signal of a channel is regarded as interfering for all others on the same fiber link.
Power control is a key issue in designing an interference-limited multiuser commu-
nication network system. A major portion of this monograph will be focused on this
first type of problem by using game-theoretic approaches. Another part of the book
will discuss game-theoretic approaches for other challenging problems such as rout-
ing and wavelength-assignment in optical networks [99, 125, 165] in the context of
reconfigurable optical networks.
For simplicity, we use “optical networks” to generally refer “reconfigurable
switched optical networks”. We also use “channel (input) power” to refer to
“channel (input) signal power”. Our approach is primarily based on game theory.
Throughout the book the focus is on analysis of dynamic systems arising from game
formulations with non-separable player utilities and with coupled constraints as well
as propagated (modified) constraints, in the class of continuous strategy games. It
turns out this is the case for a class of games in optical networks, where the utility
of each player is affected by decisions of other players, and action sets are no longer
orthogonal (Chaps. 7 and 8).
The monograph is organized as follows. Chapter 2 provides an overview of basic
concepts in game theory, mostly focused on noncooperative (Nash) game theory.
Chapter 3 focuses on matrix games, while Chap. 4 on games with continuous action
sets and cost functions. Chapter 5 presents some relatively new theoretical results for
Nash games with coupled constraints, i.e., coupled action sets. Chapter 6 provides
an overview of basic background on transmission in optical networks and on gen-
eral topologies to be studied. The concepts of optical signal-to-noise ratio (OSNR)
and coupled power constraints in an optical link are introduced, and the various
models are presented. In Chap. 7 the basic formulation of a game framework to-
wards maximizing channel OSNR is presented. We restrict our attention to single
point-to-point optical links, as the simplest network topology. A Nash game played
among channels is set up towards maximizing OSNR, in this first instance without
coupled power constraints. Thereafter, two approaches are considered in order to
6 1 Introduction
theoretic formulations. Stability properties of game and pricing dynamics are in-
vestigated by using Lyapunov theory and a singular perturbation approach, under
the assumption of time-scale separation and considering time-delay effects. Other
game formulations for wavelength-routing and wavelength-assignment (routing and
path-coloring) fall into the class of finite strategy games and are typically treated
as extension of integer-linear programming (ILP) heuristics. These are covered in
Chap. 12.
Techniques developed herein and game-theoretic formulations as exemplified for
optical networks could be well extended to other cases of more general classes of
games with propagated constraints. There is a lot of scope for fundamental control
and game-theoretic research. The hope is that this book will provide background
material such that non-specialists in optical networks can approach such research
problems within the optical networking domain. Its second goal is to provide re-
searchers in the networking and optical community with game-theoretic method-
ologies that could be used to solve optical networking problems.
Part I
Game Theory Essentials
Chapter 2
Basics of Game Theory
Abstract This chapter provides a brief overview of basic concepts in game theory.
These include game formulations and classifications, games in extensive vs. in nor-
mal form, games with continuous action (strategy) sets vs. finite strategy sets, mixed
vs. pure strategies, and games with uncoupled (orthogonal) vs. coupled action sets.
The next section reviews basic solution concepts, among them Nash equilibria being
of most relevance. The chapter is concluded with some remarks on the rationality
assumption and learning in classical games. The following chapters will introduce
these concepts formally.
2.1 Introduction
Game theory is a branch of applied mathematics concerned with the study of situ-
ations involving conflicting interests. Specifically game theory aims to mathemat-
ically capture behavior in strategic situations, in which an individual’s success in
making choices depends on the choices of others. The field was born with the work
of John von Neumann and Oskar Morgenstern [159], although the theory was de-
veloped extensively in the 1950s by many among whom John Nash [95, 96].
In this chapter we shall introduce the game-theoretic notions in simplest terms.
Our goal will be later on to study and formalize mathematically various game prob-
lems, by which we understand problems of conflict with common strategic features.
While initially developed to analyze competitions in which one individual does bet-
ter at another’s expense (zero-sum games), it has been expanded to treat a wide
class of interactions, which are classified according to several criteria, one of these
being cooperative versus noncooperative interactions. Typical classical games are
used to model and predict the outcome of a wide variety of scenarios involving a
finite number of players (or agents) that seek to optimize some individual objective.
Noncooperative game theory studies the strategic interaction among self-interested
players.
Historically, game theory developments were motivated by studies in economics,
but many interesting game theory applications have emerged in fields as diverse as
biology [141], computer science [54], social science and engineering [74]. In engi-
neering, the interest in noncooperative game theory is motivated by the possibility
of designing large scale systems that globally regulate their performance in a dis-
The extensive form of a game amounts to a translation of all the rules into technical
terms of a formal system designed to describe all games.
Extensive-form games generally involve several acts or stages, and each player
chooses a strategy at each stage. The game’s information structure, i.e., how much
information is revealed to which players concerning the game’s outcomes and their
opponents’ actions in the previous stages, significantly affects the analysis of such
games. Extensive-form games are generally represented using a tree graph. Each
14 2 Basics of Game Theory
node (called a decision node) represents every possible state of play of the game
as it is played [20]. Play begins at a unique initial node, and flows through the tree
along a path determined by the players until a terminal node is reached, where play
ends and cost s are assigned to all players. Each non-terminal node belongs to a
player; that player chooses among the possible moves at that node, each possible
move is an edge leading from that node to another node. Their analysis becomes
difficult with increasing numbers of players and game stages.
A formal definition is as follows.
Example 2.2 In the game of Matching Pennies (see Fig. 2.1) player 1 chooses
“heads” (H) or “tails” (T), player 2, not knowing this choice, also chooses between
H or T. If the two choose alike (matching) than player 2 wins 1 cent from player 1
(hence +1 for player 2 and −1 for player 1); else player 1 wins 1 cent from player
2 (reverse case). The game tree is shown below with vectors at the terminal vertices
indicating the cost function, while the number near vertices denote the player to
2.3 Games in Extensive Form 15
whom the move corresponds. The dotted (shaded) area indicates moves in the same
information set.
The next two figures show other two zero-sum game examples which differ by the
information available to player 2 at the time of its play (information set), denoted
by the shaded area (dotted). In the first case, Fig. 2.2, the two possible nodes of
player 2 are in the same information set, implying that even though player 1 acts
before player 2 does, player 2 does not have access to it s opponent decision. This
means that at the time of its play, player 2 does not know at which node (vertex)
he is. This is as saying that both players act simultaneously. The extensive form in
the second case Fig. 2.3, admits a different matrix game in normal form. In this
case each node of player 2 is included in a separate information set, i.e., has perfect
information as to which branch of the tree player 1 has chosen.
16 2 Basics of Game Theory
One can transform any game in extensive form into an equivalent game in normal
form, so we shall restrict most of our theoretical development to games in normal
form only.
In a matrix game with m players, each player i has a finite number of discrete
options to choose from, i.e., there are ni possible actions, so that the set of its actions
is simply identified with a set of indices Mi = {1, . . . , ni } corresponding to these
possible actions. Then one considers the action ui ∈ Ωi with the action sets being
defined as Ωi = {e1 , . . . , eni }, and ej being the j th unit vector in Rni . Given the
action u ∈ Ω chosen by all players, player i has a cost matrix Ai ∈ Rn1 ×···×nm ,
∀i ∈ M, that defines his cost by
Ji (u) = [Ai ]u1 ,...,um ∈ R, i∈M
This is easiest seen in the case of a matrix game when we can explicitly write the
cost functions as
J1 (u) = uT1 A u2 , J2 (u) = uT1 B u2 (2.1)
Such as two-player matrix game is the Matching Pennies game in Example 2.3.
A symmetric game is a game where the payoffs for playing a particular strategy
depend only on the other strategies employed, not on who is playing them. If the
identities of the players can be changed without changing the payoff to the strate-
gies, then a game is symmetric and this corresponds to B = AT . Many of the com-
monly studied 2 × 2 games are symmetric. The standard representations of the
18 2 Basics of Game Theory
Chicken game, Prisoner’s Dilemma game, and the Stag Hunt game are all symmet-
ric games [20].
Unlike matrix games, where players have a finite set of actions, in a continuous
game each player can choose its action from a continuum of (possibly vector-valued)
alternatives, that is, Ωi ⊂ Rni . We shall review results for both matrix games and
continuous games in the next two chapters, but most of the games we shall consider
afterwards are continuous games.
A cooperative game is one in which there can be cooperation between the players
and/or they have the same cost (also called team games). A noncooperative game
is one where an element of competition exists and among these we can mention
coordination games, constant-sum games, and games of conflicting interests. We
give below a few such examples.
In coordination games, what is good for one player is good for all players. An ex-
ample coordination game in normal form is described by
(−3, −3) (0, 0)
M=
(0, 0) (−4, −4)
2.5 Game Features 19
In this game, players try to coordinate their actions. The joint action (j, k) = (2, 2)
is the most desirable (least cost), but the joint action (j, k) = (1, 1) also produces
negative costs to the players. This particular game is called a pure coordination game
since the players always receive the same payoff.
Other coordination games move more toward the domain of games of conflicting
interest. For example, consider the Stag Hunt game: stag hare (we shall come back
to this example)
(−4, −4) (0, −1)
M=
(−1, 0) (−1, −1)
In this game, each player can choose to hunt stag (first row or first column) or hare
(second row or second column). In order to catch a stag (the biggest animal, hence
the bigger payoff or lowest cost of −4), both players must choose to hunt the stag.
However, a hunter does not need help to catch a hare, which yields a cost of −1.
Thus, in general, it is best for the hunters to coordinate their efforts to hunt stag, but
there is considerable risk in doing so (if the other player decides to hunt hare). In
this game, the costs (payoffs) are the same for both players when they coordinate
their actions, but their costs are not equal when they do no coordinate their actions.
Constant-sum games are games in which the sum of the players’ payoffs sum to the
same number. These games are games of pure competition of the type “my gain is
your loss”. Zero-sum games are particular example of these games, which in terms
of the two-players cost matrices can be described by B = −A. An example of such
game is the Rock, Paper, and Scissors game with the matrix form
⎡ ⎤
(0, 0) (1, −1) (−1, 1)
M = ⎣ (−1, 1) (0, 0) (1, −1) ⎦
(1, −1) (−1, 1) (0, 0)
These fall in between constant-sum games and coordination games and cover a large
class, whereby the players have somewhat opposing interests, but all players can
benefit from making certain compromises. One can say that people (and learning
algorithms) are often tempted to play competitively in these games (both in the real
world and in games), though they can often hurt themselves by doing so. However,
on the other hand, taking an extreme cooperative approach (same actions) can lead
to similarly bad (or worse) payoffs (high costs). One of the most celebrated games
of this type is the Prisoners’ Dilemma game, with the choices of to “Confess” (co-
operate) or “Don’t Confess” (defect) as the actions of two prisoners (players) put in
separate cells. If they both confess each they each receive 3 years in prison. If only
one confesses, he will be freed, used as witness, and the other will be convicted
20 2 Basics of Game Theory
and receive 8 years in prison. If neither confesses they will be convicted of a minor
offense and receive each only 1 year. The normal form (strategic form) of this game
is described by the matrix
(−3, −3) (0, −8)
M=
(−8, 0) (−1, −1)
or, as pair of cost matrices (A, B) for the two players,
−3 0 −3 −8
A= , B=
−8 −1 0 −1
2.5.4 Repetition
Any of the previously mentioned kinds of game can be played any number of times
between the same players, and the game can be the same at each play or can be
state-dependent.
In one-shot games, players interact for only a single round (or stage). Thus, in these
situations there is no possible way for players to reciprocate (by inflicting punish-
ment or rewards) thereafter.
In repeated games, players interact with each other for multiple rounds (playing the
same game). In such situations, players have opportunities to adapt to each others’
behaviors (i.e., “learn”) in order to try to become more successful. There can be
finite-horizon repeated games where the same game is repeated a fixed number of
times by the same players, or infinite-horizon games in which the play is repeated
indefinitely.
The case where the game changes when players interact repeatedly is what can be
called a repeated dynamic game, characterized by a state. These are also called dif-
ferential games. Unlike a repeated game where the agents play the same game every
time, in a dynamic game the state of the game influences the play and the outcome.
Important in this class are the so called stochastic games, which are extensions of
Markov decision processes to the scenario with m multiple players, where prob-
abilistic transitions are modeled. We shall not cover these types of game in this
monograph.
2.6 Solution Concepts 21
Depending on the amount of information a player has different plays and outcomes
may be possible. For example, does an player know the costs (or preference order-
ings) of other players? Does the player know its own cost (payoff) matrix? Can he
view the actions and costs of other players? All of these (and other related) questions
are important as they can help determine how the player should learn and act. The-
oretically, the more information an player has about the game, the better he should
be able to do. In short, the information an player has about the game can vary along
the following dimensions: knowledge of the player’s own actions; knowledge of the
player’s own costs; knowledge of the existence of other players; knowledge of the
other players’ actions; knowledge of the other players’ costs and in case learning is
used, knowledge of the other players’ learning algorithms.
In a game with complete information each player has knowledge of the payoffs
and possible strategies of other players. Thus, incomplete information refers to situ-
ations in which the payoffs and strategies of other players are not completely known.
The term perfect information refers to situations in which the actual actions taken by
associates are fully observable. Thus, imperfect information implies that the exact
actions taken by associates are not fully known.
One of the most basic properties of every game is the minimax solution (or minimax
strategy), also called security strategy. The minimax solution is the strategy that
minimizes a player’s maximum expected loss (cost). There is an alternate set of
terminology we can use (often used in the literature as we mentioned before). Rather
than speak of minimizing our maximum expected loss, we can talk of maximizing
our minimum expected payoff. This is known as the maximin solution. Thus, the
terms minimax and maximin can be used interchangeably. The minimax solution is
an essential concept for zero-sum games.
22 2 Basics of Game Theory
Let us look at the Prisoner’s Dilemma matrix game above. In the prisoner’s
dilemma, both players are faced with the choice of cooperating or defecting. If both
players cooperate, they both receive a relatively low cost (which is −3 in this case).
However, if one of the players cooperates and the other defects, the defector receives
a very low cost (−8 in this case) (called the temptation cost), and the cooperator re-
ceives a relatively high cost (0 in this case). If both players defect, then both receive
a higher cost (which is −1 in this case). So what should you do in this game? Well,
there are a lot of ways to look at it, but if you want to play conservatively, you might
want to invoke the minimax solution concept, which follows from the following rea-
soning. If you play cooperate, the worst you can do is get a cost of 0 (thus, we say
that the security of cooperating is 0). Likewise, if you play defect, the worst you
can do is get a cost of −1 (security of defecting is −1). Alternately, we can form a
mixed strategy over the two actions. However, it just so happens in this game that
no mixed strategy has higher security than defecting, so the minimax strategy in this
game is to defect. This means that the minimax value (which is the maximum cost
one can incur when plays the minimax strategy) is −1.
However, even though the minimax value is the lowest cost you can guarantee
yourself without the cooperation of your associates, you might be able to do much
better on average than the minimax strategy if you can either outsmart your asso-
ciates or get them to cooperate or compromise with you (in a game that is not fully
competitive). So we need other solution concepts as well.
Another basic solution concept in multi-player games is to play the strategy that
gives you the lowest cost given your opponents’ strategies. That is exactly what
the notion of the best response suggests. Suppose that you are player i, and your
opponents’ play u−i . Then the your best response in terms of pure strategies is u∗i
such that
Ji u−i , u∗i ≤ Ji (u−i , ui ), ∀ui ∈ Ωi
In the case of mixed strategies, assuming your opponents’ play the strategy x−i ,
your best response is the strategy xi∗ such that
Ji x−i , xi∗ ≤ Ji (x−i , xi ), ∀xi ∈ Δi
where Δi is the probability simplex. The best-response idea has had a huge impact
on learning algorithms. If you know what your other players are going to do, why not
get the lowest cost (highest payoff) you can get (i.e., why not play a best response)?
Taking this one step further, you might reason that if you think you know what other
players are going to do, why not play a best response to that belief? While this
obviously is not an unreasonable idea, it has two problems. The first problem is that
your belief may be wrong, which might expose you to terrible risks. Secondly, this
“best-response” approach can be quite unproductive in a repeated game when other
players are also learning/adapting [48].
2.6 Solution Concepts 23
We now introduce briefly a most celebrated solution concept for a N -player nonco-
operative game G. John Nash’s identification of the Nash equilibrium concept has
had perhaps the single biggest impact on game theory. Simply put, in a Nash equi-
librium, no player has an incentive to unilaterally deviate from its current strategy.
Put another way, if each player plays a best response to the strategies of all other
players, we have a Nash equilibrium.
We will discuss the extent to which this concept is satisfying by looking at a few
examples later on.
While all the Nash equilibria we have identified so far for these two games are
pure strategy Nash equilibrium, they need not be so. In fact, there is also a third
Nash equilibrium in the above coordination game in which both players play mixed-
strategies. The next chapter we shall formally review this extension.
Here are a couple more observations about the Nash equilibrium as a solution
concept:
• In constant-sum games, the minimax solution is a Nash equilibrium of the game.
In fact, it is the unique Nash equilibrium of constant-sum games as long is there
is not more than one minimax solution (which occurs only when two strategies
have the same security level).
• Since a game can have multiple Nash equilibrium, this concept does not tell us
how to play a game (or how we would guess others would play the game). This
poses another question: Given multiple Nash equilibria, which one should (or
will) be played? This leads to considering refinements of Nash equilibria.
Strategic dominance is another solution concept that can be used in many games.
Loosely, an action is strategically dominated if it never produces lower costs (higher
payoffs) and (at least) sometimes gives higher costs (lower payoffs) than some other
action. An action is strategically dominant if it strategically dominates all other ac-
tions. We shall formally define this later on. For example, in the Prisoner’s Dilemma
(PD) game, the action defect strategically dominates cooperate in the one-shot game.
This concept of strategic dominance (or just dominance, as we will sometimes call
it) can be used in some games (called iterative dominance solvable games) to com-
pute a Nash equilibrium.
One of the features of a Nash equilibrium (NE) is that in general it does not cor-
respond to a socially optimal outcome. That is, for a given game it is possible for
all the players to improve their costs (payoffs) by collectively agreeing to choose a
strategy different from the NE. The reason for this is that a posteriori some players
may choose to deviate from such a cooperatively agreed-upon strategy in order to
improve their payoffs further at the group’s expense. A Pareto optimal equilibrium
describes a social optimum in the sense that no individual player can improve his
payoff (or lower his cost) without making at least one other player worse off. Pareto
optimality is not a solution concept, but it can be an important attribute in deter-
mining what solution the players should play (or learn to play). Loosely, a Pareto
optimal (also called Pareto efficient) solution is a solution for which there exists no
other solution that gives every player in the game a higher payoff (lower cost). A PE
solution is formally defined as follows.
Definition 2.6 A solution u∗ is strictly Pareto dominated if there exists a joint ac-
tion u ∈ Ω for which Ji (u) < Ji (u∗ ) for all i, and weakly Pareto dominated if there
exists a joint action u = u∗ ∈ Ω for which Ji (u) ≤ Ji (u∗ ) for all i.
2.7 The Rationality Assumption 25
Definition 2.7 A solution u∗ is weakly Pareto efficient (PE) if it is not strictly Pareto
dominated and strictly Pareto efficient (PE) if it is not weakly Pareto dominated.
Often, a Nash equilibrium (NE) is not Pareto efficient (optimal). Then one speaks
of a loss of efficiency, which is also referred to as the Price of Anarchy. An interest-
ing problem is how to design games with improved Nash efficiency, and pricing or
mechanism design is concerned with such issues.
In addition to these solution concepts other important ones include the Stackel-
berg equilibrium [20], which is relevant in games where the information structure
plays an important role, and correlated equilibria [48, 98], which case is relevant
in games where the randomization used to translate players’ mixed strategies into
actions are correlated.
The important problem to study is not only to classify games for which equilibria
exist and have favorable properties such as uniqueness, but also, in conjunction, to
classify the strategy update laws that yield convergence to these equilibria under
repeated play. In the terminology of [158], a game with learning can be said to have
an “inner game” (i.e. the underlying classical game G) and an “outer game” (i.e. the
dynamics of the strategy update laws).
One of earliest strategy update laws to be studied is fictitious play (FP) [31]. In
FP, each player keeps a running average, known as an empirical frequency of his op-
ponent’s actions, and chooses his next move as a best response to this average. The
term “fictitious play” comes from the consideration of playing under the unrealistic
assumption that the opponent is playing a constant strategy, hence he is a fictitious
player. It has been proved that under a number of different sufficient conditions FP
is guaranteed to converge to one. One of the earlier cases studied was the class of fi-
nite, two-player, zero sum games, for which convergence of FP was proved in [127],
followed by other results [75].
A continuous time version of fictitious play can be derived by considering the
infinitesimal motions of the empirical frequencies [58]. It can be shown that if both
players are updating their strategies so that these are the best response to the other
one, we may write
u̇i = −ui + ri (u−i )
This equation known as the best-response (BR) dynamic, and clearly displays the
very convenient feature that u̇ = 0 if and only if u has reached a fixed point of the
reaction function r(u). By the characterization of NE given above (see (4.5)), we
conclude that the set of equilibria of the best-response dynamic coincides precisely
with NE(G). Most of these studies are done within the framework of evolutionary
game theory [132, 142], while this monograph is concerned only with the setup of
classical games.
2.9 Notes
Abstract This chapter considers normal-form games with finite action sets, hence
matrix games. Two-player bimatrix cases are treated first, followed by m-player ma-
trix games, both introducing pure- and mixed-strategy Nash equilibrium concepts.
The concepts of dominance and best replies are reviewed, as well as Nash equilibria
theorem and Nash equilibria refinements. Results are mostly adapted from (Basar
and Olsder in Dynamic Noncooperative Game Theory. SIAM Series Classics in
Applied Mathematics, 1999; Owen in Game Theory, Academic Press, San Diego,
1995).
3.1 Introduction
In this chapter we consider the class of m-player matrix games, where each player i
has a finite number of discrete options to choose from, i.e., there are ni possible ac-
tions, so that the set of its actions is simply identified with a set of indices {1, . . . , ni }
corresponding to these possible actions. On the other hand, in a continuous game
each player can choose its action from a continuum of (possibly vector-valued) al-
ternatives, that is, Ωi ⊂ Rni . This will be the focus of the next chapter.
We consider first the two-player matrix game case, or bimatrix games, and gen-
eralize then to m-players matrix games. We discuss pure and mixed- strategy game
formulations, review concepts of dominance and best replies, and then prove the
important Nash equilibria theorem, followed by a brief review of Nash equilibria
refinements.
Consider a two-player matrix game, where player 1 and 2 each have a finite number
n1 and n2 of discrete options or pure strategies to choose from. Then the set of their
actions Ω1 and Ω2 can be simply identified with the set of indices M1 := {1, . . . , n1 }
and M2 := {1, . . . , n2 } corresponding to these possible actions.
Let the action of player 1 be denoted by u1 ∈ Ω1 ; the j th action can be identified
with the index j ∈ M1 . Similarly for player 2, let its action be denoted by u2 ∈ Ω2 ;
the kth action can be identified with the index k ∈ M2 . Each of the two players have
cost matrix A and B, respectively. Then we can write for the game outcome denoted
by J1 when strategy pair (j, k) is used
T
aj k = e1j Ae2k := J1 e1j , e2k
where e1j ∈ R n1 and e2k ∈ R n2 are the j th unit vector in R n1 and the kth unit vector
in R n2 , respectively. Thus player 1’s cost when strategy pair (u1 , u2 ) is set to be the
(j, k) pair is
T
J1 (u1 , u2 ) = J1 e1j , e2k = e1j Ae2k
Then for player 1 we can write u1 ∈ Ω1 := {e11 , . . . , e1j , . . . , e1n1 } and correspond-
ingly for player 2, u2 ∈ Ω2 := {e21 , . . . , e2k , . . . , e2n2 }, and its cost when (u1 , u2 ) is set
to be the (j, k) pair is
T
J2 (u1 , u2 ) = J2 e1j , e2k = e1j Be2k
Each of the players aims to minimize its own cost, J1 and J2 , respectively, described
by its own cost matrix A and B, respectively. We call this a (A, B) bimatrix game.
In the case when B = −A the game is called a zero-sum game.
The definition of a Nash equilibrium in pure strategies for a bimatrix game is
given next.
Definition 3.1 (Pure-strategy Nash equilibrium (NE)) For a given (A, B) bimatrix
game let (j ∗ , k ∗ ) be pair of pure strategies chosen by the two players. Then if for
any other pure strategies (j, j ) both
aj ∗ k ∗ ≤ aj k ∗ , ∀j = 1, . . . , n1
bj ∗ k ∗ ≤ bj ∗ k , ∀k = 1, . . . , n2
hold, then (j ∗ , k ∗ ) is a pure-strategy Nash equilibrium (NE) solution for the bima-
trix game, and (aj ∗ k ∗ , bj ∗ k ∗ ) is a NE equilibrium outcome of the game.
In the case when more than a single NE solution exist, we shall denote this set
by NE(G). The game outcome can be different for different NE solutions and the
question of ordering the elements in NE(G) arises. Since we cannot have total or-
dering between pairs of numbers one can resort to partial ordering in order to arrive
at some preferential choice.
It can be shown that the set of NE equilibrium points is invariant to positive affine
transformations on the cost functions.
Definition 3.3 Two bimatrix games (A, B) and (Ã, B̃) are said to be strategically
equivalent games if there exist positive constants α1 , α2 > 0 and scalar β1 , β2 such
that
ãj k ∗ = α1 aj k ∗ + β1 , α1 > 0, β1 ∈ R, ∀j
b̃j ∗ k = α2 bj ∗ k + β2 , α2 > 0, β2 ∈ R, ∀k
Proposition 3.4 All strategically equivalent bimatrix games have the same NE equi-
libria.
Example 3.5 (Prisoner’s Dilemma) Recall the Prisoner’s Dilemma (PD) game re-
viewed in Chap. 2, where cost matrices (A, B) for the two players are
−3 0 −3 −8
A= , B=
−8 −1 0 −1
for the choices of to “Confess” (first strategy) or to “Not Confess” (second strategy)
available for the two prisoners. This is a game in which there are gains from coop-
eration between players; the best outcome is for the players to not confess (second
strategy), but each player has an incentive to be a “free rider”. Whatever one player
does, the other prefers “Confess” so that game has a unique pure Nash equilibrium
(Confess, Confess) or pair (1, 1). Indeed player 1’s first pure strategy (“Confess”)
gives a smaller loss (higher payoff) than its second pure strategy, irrespective of what
strategy player 2 uses. Similarly, player 2’s first pure strategy (defect) gives always
a smaller loss (higher payoff) than its second pure strategy: each entry in first col-
umn of B matrix is less than the corresponding entry in the second column. Hence
individual rationality of minimizing its cost or loss leads to each player to select
the first strategy, hence both “Confess”. The dilemma arises because both players
would have lower loss if they were to select their second strategy (“Not Confess”),
but this would require some coordination.
Unfortunately not every game has a pure equilibrium. For example the game of
Matching Pennies in Example 2.3 in Chap. 2 does not have a (pure) Nash equilib-
rium. This leads to the mixed-strategy extension of a game.
30 3 Matrix Games
where χ denotes random variable of choosing one of the pure strategies in set Ω1 .
Since Ω1 is finite this probability distribution is a vector of probabilities associ-
ated with the pure action. A mixed strategy is identified as the vector composed of
the probabilities associated with available actions, i.e., x = [xj ], j ∈ {1, . . . , n1 },
3.2 Bimatrix Games 31
x ∈ Δ1 , where
n1
Δ1 := x ∈ R n1 xj = 1, xj ≥ 0, ∀j = 1, . . . , n1 (3.1)
j =1
is the unit simplex of (n1 − 1) dimension. Note that pure strategies are just extreme
case of mixed strategies (vertices of the simplex Δ1 ). Similarly, for player 2 we
shall identify his mixed strategy with y = [yk ], k ∈ {1, . . . , n}, y ∈ Δ2 , where yk
denotes the probability that player 2 will choose action k from his n2 available
(pure) alternatives in Ω2 and y ∈ Δ2 ,
n2
Δ2 := y ∈ Rn2 yk = 1, yk ≥ 0, ∀k = 1, . . . , n2
k=1
n1
n2
= J1 e1j , e2k xj yk (3.2)
j =1 k=1
and we denote this by J 1 (x, y) or J 1 (x), as the mixed-strategy cost function, when
x ∈ Δ1 , y ∈ Δ2 . Using J1 (e1j , e2k ) = aj k into (3.2) this gives
n1
n2
J 1 (x) = J 1 (x, y) = aj k xj yk = x T Ay
j =1 k=1
n1
n2
J 2 (x) = J 2 (x, y) = bj k xj yk = x T By
j =1 k=1
Remark 3.6 In the case when B = −A, the game is a two-player zero-sum matrix
game
J 1 (x, y) + J 2 (y, x) = 0, ∀x, ∀y
So in this game each player minimizes its cost J 1 and J 2 , but with respect to J
player 1 is the minimizer, while player 2 is regarded as the maximizer of J .
32 3 Matrix Games
Definition 3.7 (Mixed-strategy Nash equilibrium (NE)) For a given (A, B) bimatrix
game let x∗ = (x ∗ , y ∗ ) be pair of mixed strategies chosen by the two players. Then
if for any other mixed strategies (x, y) both
x ∗T Ay ∗ ≤ x T Ay ∗ , ∀x ∈ Δ1
x ∗T By ∗ ≤ x ∗T By, ∀y ∈ Δ2
denotes the unit simplex in R ni space, which has dimension ni − 1. The vertices of
Δi are the unit vectors eij . Every mixed strategy xi ∈ Δi is some convex combination
of the unit vectors (pure strategies) eij ∈ Ωi ,
ni
xi = xi,j eij
j =1
Definition 3.8 For a mixed strategy xi ∈ Δi , we define its support or carrier as the
set of pure strategies that have assigned positive probabilities,
The subset
int(Δi ) = {xi ∈ Δi | xi,j > 0, ∀j ∈ Mi }
is called the interior of Δi . Hence, mixed strategies in the interior xi ∈ int(Δi ) are
called completely mixed or interior, in that they assign positive probabilities to all
player’s pure strategies, hence have full support, i.e., supp(xi ) = Mi .
m
x(u) = xi (ui )
i=1
m
m
km =
x e1k1 , . . . , em xi eiki = xi,ki , k i ∈ Mi
i=1 i=1
In terms of pure strategies the cost for player i when the pure-strategy m-tuple u
is used is denoted by Ji (u) ∈ R. If mixed strategies xi , i ∈ M are used according
34 3 Matrix Games
to the distribution x(u) then the cost will be the statistical expectation of Ji with
respect to the probability distribution x(u), namely
E Ji (u) = Ji (u)x(u) := J i (x)
u∈Ω
m
J i (x) = J i (x1 , . . . , xm ) = ··· Ji (u1 , . . . , um ) xl (ul ) (3.4)
u1 ∈Ω1 um ∈Ωm l=1
m
J i (x−i , ui ) = ··· ··· Ji (u) xl (ul ) (3.5)
u1 ∈Ω1 ui−1 ∈Ωi−1 ui+1 ∈Ωi+1 um ∈Ωm l=1,l =i
We also write (x−i , zi ) ∈ ΔX for the strategy profile in which player i plays mixed
strategy zi ∈ Δi while all other players use strategies according to x ∈ ΔX . This
notation is particularly useful when a single player considers “deviations” zi ∈ Δi
from a given profile x ∈ ΔX .
The definition of an NE equilibrium point in mixed strategies (MSNE) is given
next.
where x∗ = (x∗−i , xi∗ ) and x∗−i denotes x∗ of all mixed strategies except the ith one,
i ∈ M.
3.4 Dominance and Best Replies 35
Since there can be more than a single NE we denote this set as NE(G), and x∗ is
an element of this set, x∗ ∈ NE(G). Then, based on Definition 3.9,
NE(G) = x∗ ∈ ΔX | J i x∗ ≤ J i x∗−i , wi , ∀i ∈ M, ∀wi ∈ Δi (3.7)
Now, based on the above identification of Ω1 with {e11 , . . . , e1k1 , . . . , e1M1 }, we can
replace u1 ∈ Ω1 by e1k1 for k1 ∈ M1 and we can write
m
J i x1∗ , . . . , xm
∗
= ··· Ji (u1 , . . . , um ) ∗
xl,kl
u1 ∈Ω1 um ∈Ωm l=1
1
m
∗
= ··· m
Ji ek1 , . . . , ekm xl,kl
k1 ∈M1 km ∈Mm l=1
with the second form explicitly showing the finite matrix form. Using this form and
km ) = Ak1 ,...,km (element in the cost matrix of player i), we
identifying Ji (e1k1 , . . . , em i
see that (3.6) can be re-written as a set of m inequalities, and the ith one, i ∈ M,
∀wi = [xi,j ] ∈ Δi , j ∈ Mi
m
m
∗ ∗
··· Aik1 ,...,km xl,kl
≤ ··· Aik1 ,...,km xl,k x
l i,j
(3.8)
k1 ∈M1 km ∈Mm l=1 k1 ∈M1 km ∈Mm l=1,l =i
One could define partial ordering of a player’s (pure or mixed) strategy set as given
by the consequences on the outcome for that player. We shall define everything
herein in terms of mixed strategies since pure strategies are special cases of these,
so we work on the mixed-strategy simplex Δi of each player i. A strategy weakly
dominates another strategy if it never results in a worse outcome (higher loss) than
the second and sometimes results in lower loss. A strategy is undominated if there is
no strategy that weakly dominates it. A strategy strictly dominates another strategy
if it always results in lower loss.
One of the basic assumptions in non-cooperative game theory is that players are
rational and do not ever use strictly dominated strategies. This is the reason why
pure dominated pure strategies can be deleted without affecting the game outcome.
for any given x−i ∈ Δ−i , also called the optimal response set (or rational reac-
tion set) [20]. The overall mixed-strategy best response of the m-players to strategy
profile x ∈ ΔX is Φ : ΔX → ΔX
Φ(x) = Φ1 (x−1 ), . . . , Φi (x−i ), . . . , Φm (x−m ) ⊂ ΔX
which can be multi-valued, hence a correspondence. Now since every mixed strat-
egy is a convex combination of pure strategies and J i (x−i , zi ) is linear in zi , it
follows that no mixed strategy zi ∈ Δi can give a lower cost against x−i than one
of the pure best replies to x−i of player i. The best-response set Φi (x−i ) is a face
of Δi , i.e., a convex hull of some pure strategies (vertices) in Δi ) and is always
non-empty, closed, and convex.
Based on the best-reply correspondence Φ, (3.9) for any mixed-strategy Nash profile
(equilibrium point) x∗ = (x∗−i , xi∗ ) ∈ NE(G), we can write
xi∗ ∈ Φi x∗−i , ∀i ∈ M
response to all his opponents’ actions, and the same is true for all the players i ∈ M.
Hence
NE(G) = x∗ ∈ ΔX | xi∗ ∈ Φi x∗−i , ∀i ∈ M = x∗ ∈ ΔX | x∗ ∈ Φ x∗ (3.10)
Proof (Necessity) If x∗ is an mixed NE point, (3.6) holds for any wi . Since any pure
strategy is a particular case of a mixed strategy, we can take wi = eij which shows
that (3.11) holds.
(Sufficiency) Choose an arbitrary mixed strategy wi ∈ Δi , hence
wi = eij αji , for some αji ≥ 0, αji = 1
j ∈Mi j ∈Mi
38 3 Matrix Games
This yields
i
J i x∗−i , xi∗ αj ≤ J i x∗−i , wi
j ∈Mi
Since j ∈Mi αji = 1, by (3.9) this implies that
xi∗ ∈ Φi x∗−i
Proof
We prove the statement by contradiction. Consider any xi ∈ Δi , xi =
j ∈Mi xi,j ej =
i i
k∈supp(xi ) xi,k ek as a convex combination of pure strategies,
where k∈supp(xi ) xi,k = 1, and supp(xi ) = 0 since for at least one eik we have
xi,k > 0. Assume by contradiction that for all pure strategies k ∈ supp(xi ), i.e., with
xi,k > 0, we have J i (x−i , xi ) > J i (x−i , eik ). Hence for all such strategies,
J i (x)xi,k > J i x−i , eik xi,k
Then factoring J i (x) on the LHS and using the representation of xi gives
hence a contradiction.
Proof If player i has ni pure strategies, then the set of his mixed strategies can be
represented geometrically by an (ni − 1)-dimensional simplex, Δi . Each Δi is a
convex, closed, and bounded (hence compact) subset of the R ni Euclidean space.
3.5 Nash Equilibria Theorem 39
and we see that Cji is a function that measures the decrease in the cost for player i
when in the mixed-strategy m-tuple (profile) x, player i replaces strategy xi by its
pure strategy ui = eij . We also see that Cji (x) ≥ 0. Any possible increase in cost is
not indicated since in this case the function vanishes.
Now for each i ∈ M, j ∈ Mi we consider the transformation
where xi (eij ) = xi,j . Each of these is non-negative also. Note that, since xi ∈ Δi ,
xi eij = xi,j = 1
j ∈Mi j ∈Mi
T (x̃) = x̃
The theorem is proved if we can show that every such fixed point is necessarily a
Nash equilibrium and vice versa that every Nash equilibrium is a fixed point of this
mapping. Since x̃ = (x̃ 1 , . . . , x̃ m ) is a fixed point for T it follows that for all i ∈ M
and for all l ∈ Mi
x̃ i (eil ) + Cli (x̃)
x̃ i eil := (3.12)
1 + j ∈Mi Cji (x̃)
40 3 Matrix Games
For this x̃ ∈ ΔX , by Lemma 3.14 and the definition of Cji it follows that for any
player i there exist a k and a pure strategy eik such that x̃ i (eik ) = x̃i,k > 0
x̃ i eik > 0 and J i (x̃) − J i x̃−i , eik ≤ 0
hence
x̃ i eik > 0 and Cki (x̃) = 0
For this particular strategy (3.12) becomes
x̃ i (ei )
x̃ i eik := k i
1 + j ∈Mi Cj (x̃)
so that
i
x̃ i eik Cj (x̃) = 0
j ∈Mi
which since x̃ i (eik ) > 0 implies j ∈Mi Cji (x̃) = 0. Since all Cji ≥ 0 the foregoing
implies
Cji (x̃) = 0, ∀i ∈ M, ∀j ∈ Mi
and from the definition of Cji this means that for each player i we have
J i (x̃) ≤ J i x̃−i , eij , ∀i ∈ M, ∀j ∈ Mi
Proof (based on Kakutani’s fixed-point theorem) Recall that Φi (x−i ) is the set of
tuples of mixed strategies chosen by each player in a game, and Φ(x) is the overall
best response. Since there may be a number of responses which are equally good,
Φi (x−i ) is set-valued rather than single-valued. Then the Nash equilibrium of the
game is defined as a fixed point of Φ.
Note that ΔX is a non-empty, compact, and convex set. The best-reply correspon-
dence (mapping) Φ : ΔX → ΔX is upper hemi-continuous by the continuity of the
payoff function J i , which in itself follows form Berge’s maximum Theorem A.17
3.6 Nash Equilibria Refinements 41
(see Appendix A). For every x ∈ ΔX , the image Φ(x) ⊂ ΔX is a non-empty, closed,
and convex set. Hence, by Kakutani’s fixed-point Theorem A.19, Appendix A, Φ
has at least one fixed point in ΔX , i.e., there exists x∗ ∈ ΔX such that x∗ ∈ Φ(x∗ ).
This refinement is due to Selten [136], best known as the “trembling hand perfec-
tion”. This intuitively means that NEs that are not robust to “trembles” in players’
strategies are discarded.
Consider a game G(M, Δi , J i ) and let μ be an error function such that a number
μi,k ∈ (0, 1) is assigned for each player i ∈ M, and pure strategy k ∈ Mi which
defines the probability
that this k pure strategy will be played by mistake (trembling
hand), with k ∈ Mi μi,k < 1. Note that such a small probability > 0 is assigned to
all pure strategies.
This means that for each player i ∈ M the error function μ defines a subset of
mixed strategies
that player i can use. Define a perturbed game G(μ) = G(M, Δi (μ), J i ), where
ΔX (μ) = i∈M Δi (μ) ⊂ int(ΔX ). By a similar argument based on Kakutani’s
frixed-point theorem, every such perturbed game has an non-empty set NE(G(μ))
of NEs. The smaller all error probabilities are, the larger ΔX (μ) is and as μ → 0
the perturbed game approaches the original game.
Definition 3.16 An NE, x∗ ∈ NE(G) is a perfect equilibrium if, for some sequence
{G(μt )}μt →0 of perturbed games, there exist profiles xt ∈ NE(G(μt )) such that
xt → x ∗ .
Let us call PE(G) the set of perfect equilibria. Note that every interior (com-
pletely mixed) or inner NE is perfect. This is seen because if x∗ ∈ int(ΔX ),
then for sufficiently small μi,k , x∗ ∈ int(ΔX (μ)). If in addition x∗ ∈ NE(G) then
42 3 Matrix Games
x∗ ∈ NE(G(μ)). The definition above requires only that the NE be robust with re-
spect to some trembles (error function), so existence can be established even if there
are no interior NE.
Proof For any sequence {G(μt )}μt →0 let xt ∈ NE(G(μt )) for each t. Since {xt },
t = 1, . . . , ∞ is a sequence in a compact set ΔX it has a convergent subsequence,
{xts }, s = 1, . . . , ∞ with limit x∗ ∈ ΔX . For each s, G(μts ) is the associated per-
turbed game. Then since xts ∈ NE(G(μts )) it follows that x∗ ∈ NE(G) by continuity
arguments, and x∗ is perfect since xts → x∗ and xts ∈ NE(G(μts )) for all s.
We give without proof the following useful result (for a proof see [157]).
Definition 3.19 An NE, x∗ ∈ NE(G) is a strictly perfect equilibrium if, for every
sequence {G(μt )}μt →0 of perturbed games, there exist profiles xt ∈ NE(G(μt )) such
that xt → x∗ .
Note that any interior (inner) NE equilibrium x ∈ NE(G) is strictly proper, since
one could take xt = x∗ for all t sufficiently large such that x∗ ∈ NE(G(μt )). In
fact a perfect NE equilibrium that is not strictly perfect is vulnerable against some
sequence of trembles.
-proper if
J i x−i , eik > J i x−i , eij =⇒ xi,k ≤
xi,j
3.7 Notes 43
3.7 Notes
In this chapter we reviewed basic concepts and results for matrix games, i.e., games
with finite action sets. In the next chapter we consider games with infinite (continu-
ous) action sets.
Chapter 4
Games with Continuous Action Spaces
Abstract This chapter focuses on Nash games with continuous kernel, i.e., games
with continuous action spaces and cost functions. Basic concepts and results are
reviewed focused on game formulations, Nash equilibria, reaction curves, and exis-
tence results. These are mostly adapted from (Basar and Olsder in Dynamic Nonco-
operative Game Theory. SIAM Series Classics in Applied Mathematics, 2nd edn.,
1999). For more in-depth treatment the reader is referred to this and other more
extensive references on the subject.
4.1 Introduction
In this chapter we consider games with infinite action spaces (hence continuous)
and with players’ payoff (cost) functions as continuous in the actions. We mostly
treat pure strategies (actions). We introduce basic NE concepts and existence re-
sults for the case in which no coupling exists between those action spaces, results
mostly adapted from [20, 98]. For more in-depth treatment the reader is referred to
those and other more extensive references on the subject. A review of the associated
optimization results is presented in one of the appendices.
The chapter is organized as follows. We present basic concepts and game formu-
lation for games with continuous action spaces and uncoupled constraints, and then
we present fundamental existence and uniqueness results for the Nash equilibrium.
A game consists of a set of players, an action set action set (also referred to as a set
of strategies) available to those players and an individual objective function for each
player. In a game, each player individually takes an optimal action which optimizes
its own objective function and each player’s success in making decisions depends
on the decisions of the others.
Let us consider games with continuous action spaces and pure strategies. Specif-
ically, as the focus of this chapter we consider games with continuous action spaces
and uncoupled constraints. The following chapter will treat such games, but with
Ω = Ω1 × · · · × Ωm (4.1)
Similarly to the concept of a mixed strategy for matrix games, this concept can be
defined for continuous (or infinite) games by employing the appropriate distribution
function. We only consider for simplicity the N = 2 case. Let the two players’ pure
strategies be denoted u1 ∈ Ω1 , u2 ∈ Ω2 , where this time Ωi are infinite set such as
for example the interval [0, 1].
A mixed strategy for player 1 is a cumulative distribution function σ1 defined
on Ω1 : for every u1 ∈ Ω1 , σ1 (u1 ) is a random process for choosing a number not
greater than u1 . That is,
σ1 (u1 ) = Pr (ξ ≤ u1 )
where Pr denote the probability and ξ the random variable of the process.
(i) σ1 (u1 ) ≥ 0, ∀u1 ∈ Ω1 , (ii) σ1 (0) = 0, σ (1) = 0, (iii) σ1 (u1 ) is non-decreasing,
(iv) σ1 (u1 ) is right-continuous in the open (0, 1). Similarly σ2 (u2 ) is the mixed
4.4 Nash Equilibria and Reaction Curves 47
Now if player 1 and 2 use mixed strategies σ1 (u1 ) and σ2 (u2 ), then joint probability
distributions have to be considered, and assuming them independent, the expected
cost for player 1 is
J 1 (σ1 , σ2 ) = J (u1 , u2 ) dσ1 (u1 ) dσ2 (u2 )
Ω1 Ω2
and similarly the expected cost for player 2. Then definition for a Nash equilibrium
can be given in the mixed-strategy case using the set of distribution functions in
terms of these expected costs.
Return to the basic case of games with continuous action spaces in the pure-strategy
case, G(I, Ωi , Ji ). Let us define an NE solution for this case. In such a game an
individual cost function Ji : Ω → R is assigned to each player i, Ji (u), and each
player aims to minimize its own cost function Ji (ui , u−i ), in the presence of all
other players.
Then a pure-strategy Nash equilibrium (NE), or simply a Nash equilibrium (NE)
of this noncooperative game is the defined as a point u∗ ∈ Ω that satisfies
Ji u∗ ≤ Ji u∗−i , ui , ∀ui ∈ Ωi , ∀i ∈ M
or, alternatively given the actions of all other players, u−i ∈ Ω−i , each player i
independently minimizes its own cost:
min Ji (u−i , ui )
(4.2)
subject to ui ∈ Ωi
which is called the optimal reaction set of player i ∈ M, i.e., the set of all optimal re-
sponses by player i to any fixed u−i . The concepts of best response correspondence
and the NE concept, defined here for pure-strategy game G(M, Ωi , Ji ), correspond
to those reviewed in Chap. 3 for a mixed-strategy matrix game G(M, Δi , J i ).
Note that u∗ ∈ Ω is an NE solution if u∗i is a solution to the optimization problem
(4.2), given all other players taking equilibrium actions, u∗−i , i.e., u∗i is player i’s
best response to all his opponents’ actions, and the same is true for all the players
i ∈ M. Based on this we define next the reaction function of a player.
Definition 4.2 (Reaction function) If the optimal reaction set Ri (u−i ) (4.4) is a
singleton (i.e., a set with exactly one element) for every given u−i ∈ Ω−i , then
Ri (u−i ) is called the reaction function of player i, and is specifically denoted by
ri : Ω−i → Ωi .
Proof Under Assumption 4.3, the cost function Ji (u−i , ui ) is strictly convex in ui .
By Proposition A.2, there exists a unique minimum ui , for any given u−i , such that
By Definition 4.2, this implies that a unique mapping, i.e., a reaction function ri :
Ω−i → Ωi exists for each player i. The continuity can be obtained directly from
Berge’s Maximum Theorem (Theorem A.17).
Proof By strict convexity of the cost function and Proposition 4.4, for every given
u−i ∈ Ω−i , there exists a unique reaction function ri of player i ∈ M such that
ui = ri (u−i ) ∈ Ωi , ∀i ∈ M.
the other hand, r(u) maps a compact and convex set Ω into itself. From (Theo-
rem A.18), there exists a u∗ such that u∗ = r(u∗ ). By Definition 4.1, it follows that
u∗ is an NE solution of G(M, Ωi , Ji ).
∂Ji
(u−i , ui ) = 0, ∀i ∈ M (4.6)
∂ui
Now let us define an associated optimization problem from the cost functions Ji ,
i.e.,
m
min Jj (u)
j =1
subject to u∈Ω
Example 4.7 Consider a two-player Nash game with the cost functions
min J0 (u1 , u2 )
subject to 0 ≤ ui ≤ 8, i = 1, 2,
From the individual cost functions Ji (u−i , ui ) let us define now an augmented
system-like two-argument cost function J˜ : Ω × Ω → R as in [20]
m
J˜(u; x) := Ji (u−i , xi ), (4.7)
i=1
where each Ji (u−i , ui ) satisfies Assumption 4.3. We shall call this two-argument
function J˜ the Nash game (NG) cost function of the game. Based on it another
52 4 Games with Continuous Action Spaces
Remark 4.9 The two definitions, Definitions 4.1 and 4.8 are equivalent. Indeed (4.8)
can be equivalently rewritten as for every given u∗−i ,
m
m
Ji u∗−i , u∗i ≤ Ji u∗−i , xi , ∀x ∈ Ω
i=1 i=1
where ũ := (u∗−i , ũi ) and ũi = u∗i . This contradicts the hypothesis (4.8).
Based on the concept of NG cost function, (4.7), we give next a proof of Propo-
sition 4.5. The proof essentially follows arguments similar to those used in The-
orem 4.4 in [20]. We present it here to give readers a better interpretation of the
two-argument NG cost function, because this will prove a very useful analytical
vehicle in later chapters for games with coupled constraints (Chaps. 5 and 7). The
proof helps reformulate the necessary conditions (4.6) with respect to the NG cost
function.
Proof of Proposition 4.5 From (4.7), the two-argument NG cost function J˜(u; x)
is separable in the second argument x for every given u, i.e., each component cost
function in J˜(u; x) is decoupled in x for every given u. Therefore by using (4.7), for
every given u, the gradient of J˜(u; x) with respect to x is written as
⎡ ⎤
∂J1
∂x1 (u−1 , x1 )
⎢ ⎥
∇x J˜(u; x) := ⎢
⎣
..
.
⎥
⎦ (4.9)
∂Jm
∂xm (u−m , xm )
4.5 Existence and Uniqueness Results 53
or, equivalently,
where minimization on the right-hand side is done with respect to the second argu-
ment x in J˜(u; x). Recall that Ω is compact. By the continuity and convexity prop-
erty of J˜, it follows from Theorem A.17 (Berge’s Maximum Theorem) that R is an
upper-semi-continuous mapping that maps each point u in Ω into a compact and
convex subset of Ω. Then by the Kakutani Fixed-Point Theorem (Theorem A.19),
there exists a point u∗ such that u∗ ∈ R(u∗ ), i.e., u∗ satisfies (4.8). By Definition 4.8,
u∗ is an NE solution of G(M, Ωi , Ji ).
where the notation “|x=u ” denotes finding a fixed-point solution. By using (4.9), we
get the component-wise form of (4.11):
∂Ji
(u−i , xi ) = 0, ∀i ∈ M, (4.12)
∂xi xi =ui
which are equivalent to the necessary conditions (4.6) with respect to Ji presented
in Proposition 4.6.
Based on this one can summarize the procedure to find an inner NE solution with
respect to J˜ as follows. As a first step solve ∇x J˜(u; x) = 0 for every given u (i.e.,
x is the only variable), which gives x as a function of u. Then look for a fixed-
point solution for these equations, this fixed-point solution being denoted by x = u.
Solving the resulting set of m equations, (4.12), leads to an inner NE solution. We
give next an illustration of this procedure.
54 4 Games with Continuous Action Spaces
Example 4.10 (Consider the two-player Nash game in Example 4.7 and its NG cost
function)
Assume there exists an inner NE solution. Then given every u, in the first step we
solve ∇x J˜(u; x) = 0, i.e.,
∂J1
(u2 , x1 ) = 4x1 − 2 − u2 = 0
∂x1
∂J2 1
(u1 , x2 ) = 2x2 − − u1 = 0
∂x2 2
This leads to x as a function of u, i.e.,
1
1
0 4 2
x= u+
1 1
2 0 4
In the second step we solve for a fixed-point solution, by setting x = u in the above,
which leads to (u∗1 , u∗2 ) = ( 14
9 4
, 7 ).
4.6 Notes
This chapter provided an overview of game with continuous action spaces and cost
functions (continuous kernels) and uncoupled constraints. A review of optimization
results is presented in the appendix.
Chapter 5
Computational Results for Games
with Coupled Constraints
Abstract This chapter provides some results for Nash games with coupled con-
straints, i.e., coupled action sets. Work on games with coupled action spaces has
been going on for more than 50 years. These are also called generalized Nash games,
games with coupled constraints, or social equilibria. Game theoretical formulations
of problems and computational approaches towards solving coupled or generalized
Nash games have been areas of much recent interest. We present some new results
mainly based on the Lagrangian approach extension proposed in Pavel (Automatica
43(2):226–237, 2007). We review a relaxation via an augmented optimization, the
Lagrangian extension in a game setup, followed by duality and hierarchical decom-
position in a game setup.
5.1 Introduction
As seen in the previous chapter, in Nash games with uncoupled constraints, the
action space is the Cartesian product of the individual action sets and players can
affect only the cost functions of the other players but not their feasible action sets.
On the other hand, in Nash games with coupled constraints, each player’s action
affects the feasible action sets of the other players. In Example 4.7, the action sets
are Ωi = [0, 8], i = 1, 2, such that the action space Ω is rectangular in R2 .
Now consider the following example with a modified action space.
Example 5.1 Consider the two-player Nash game in Example 4.7 with an action
space, Ω := Ω1 × Ω2 and ui ∈ Ωi = [0, 8], i = 1, 2. An additional constraint is
considered: u1 + u2 ≤ 8. Now the action space is modified to be
Ω = {u ∈ Ω | u1 + u2 − 8 ≤ 0}.
u1 ∈ [0, 8 − u2 ] and u2 ∈ [0, 8 − u1 ]. That is the action space Ω is coupled and this
game is called a two-player Nash game with coupled constraints.
Starting from this example, we present in this chapter some theoretically results
for Nash games with coupled constraints, i.e., coupled action sets. We call such
a game a coupled Nash game. Coupled Nash games are also called generalized
Nash games (games with non-disjoint strategy sets) [56], games with coupled con-
straints [128], social equilibria games or pseudo-Nash equilibria games [16, 37].
Game theoretical formulations of problems and computational approaches towards
solving coupled or generalized Nash games have been areas of much recent interest
[15, 40, 90, 91, 120, 135, 138]. The work in [40] shows that a generalized Nash
equilibrium can be calculated by solving a variational inequality and the results
express conditions in terms of variational inequality problem and Karush–Kuhn–
Tucker (KKT) conditions for the pseudo-gradient. Another related work is [90]
where the authors present a scheme that associates to a generalized variational in-
equality, a dual problem and KKT conditions, thus allowing to solve primal and dual
problems in the spirit of classical Lagrangian duality for constrained optimization
problems, using set theoretic concepts and set-valued operators. Extensions based
on this approach have been developed for non-compact and non-smooth settings in
[138].
The treatment is this chapter mostly follows the Lagrangian approach extension
proposed in [120]. This is a procedural method for computing a Nash equilibrium,
based on an extension of duality to a game theoretical framework [32]. The setting
of the construction uses the two-argument NG cost function (see Chap. 4) and re-
laxes also the constraints into a two-argument form. Thus the problem is enlarged
into a constrained optimization problem in a space of twice the dimension followed
by projection back into a one dimension (with a fixed-point solution). Moreover,
for convex constraints, duality leads to hierarchical decomposition into a lower-
level game with no coupled constraints and an optimization problem for Lagrangian
prices.
The chapter is organized as follows. In Sect. 5.2 some results on Nash equilibria
existence are reviewed and relaxation via an augmented optimization problem is
considered. This is followed by results for Lagrangian extension in a game setup in
Sect. 5.3. Sections 5.4 and 5.5 present results for duality extension and hierarchical
decomposition in a game setup.
5.2 Nash Equilibria and Relaxation via an Augmented Optimization 57
gr (u) ≤ 0, r = 1, . . . , R (5.1)
for u ∈ Ω. The coupled constraint set associated with each gr (u) is denoted by
Ω r = u ∈ Ω gr (u) ≤ 0 (5.3)
R
Ω= Ω r = u ∈ Ω g(u) ≤ 0 (5.4)
r=1
For each i ∈ M, this projection action set Ωi (u−i ) is the feasible action set under
the given u−i . A vector u = (u−i , ui ) is called feasible if u ∈ Ω. The resulting
coupled Nash game is denoted by G(M, Ω i , Ji ), and an NE solution is defined as
follows.
Let us look again at Example 5.1. Note that (u∗1 , u∗2 ) = ( 14 , 7 ) ∈ Ω (see Fig. 5.2),
9 4
∗ ∗
where (u1 , u2 ) is the unique NE solution obtained in Example 4.7. Then obviously
(u∗1 , u∗2 ) is the NE solution of the associated Nash game with coupled constraints.
The following proposition (adapted from Theorem 4.4 in [20]) gives sufficient
conditions for existence of an NE solution.
58 5 Computational Results for Games with Coupled Constraints
Proposition 5.3 (Theorem 4.4, [20]) Let the action space Ω be a compact and
i , Ji ) admits an NE solution.
convex subset of Rm . Under Assumption 4.3, G(M, Ω
Next let us use the concept of NG cost function presented in Chap. 4 to charac-
terize an NE solution of G(M, Ω i , Ji ).
Recall the two-argument NG cost function J˜(u; x) defined in (4.7),
m
J˜(u; x) := Ji (u−i , xi )
i=1
and note that J˜(u; x) is in its second argument x. Then with respect to J˜(u; x), an
NE solution u∗ of G(M, Ω i , Ji ) satisfies
J˜ u∗ ; u∗ ≤ J˜ u∗ ; x , ∀x ∈ Ω, with g u∗−i , xi ≤ 0, ∀i ∈ M (5.7)
This result can be obtained by using Definition 5.2, the definition of J˜(u; x), (4.7),
i (u−i ), (5.5).
and the projection action set Ω
Now, let us also augment the coupled constraints g(u) in (5.2) into an equivalent
two-argument form, g̃,
m
g̃(u; x) = g(u−i , xi ), (5.8)
i=1
m
g̃r (u; x) = gr (u−i , xi ), ∀r = 1, . . . , R
i=1
5.2 Nash Equilibria and Relaxation via an Augmented Optimization 59
with g̃(u∗ ; u∗ ) ≤ 0.
Specifically one arrives at the following result.
This result has been proved by contradiction in [120]. We briefly review the argu-
ment here. By contradiction assume that u∗ is not an NE solution of G(M, Ω i , Ji ).
∗
It follows that for some i ∈ M, there exists an x̃i ∈ Ωi with g(u−i , x̃i ) ≤ 0, such
that
Ji u∗−i , x̃i < Ji u∗−i , u∗i
By adding a term j ∈M,j =i Jj (u∗−j , u∗j ) to both sides, the following inequality
holds:
J˜ u∗ ; x̃ < J˜ u∗ ; u∗ ,
with x̃ := (u∗−i , x̃i ) ∈ Ω. From g̃(u∗ ; u∗ ) ≤ 0 and (5.9), it follows that
g(u∗−j , u∗j ) ≤ 0, ∀j ∈ M, such that
g̃ u∗ ; x̃ := g u∗−j , u∗j + g u∗−i , x̃i ≤ 0
j ∈M,j =i
60 5 Computational Results for Games with Coupled Constraints
Proposition 5.5 Consider G(M, Ωi , Ji ). Let the action space Ω be a compact and
convex subset of R and Assumption 4.3 is satisfied.
m
μ∗ T g(u) = 0, (5.13)
where the notation “|x=u ” is defined in (4.11) and denotes finding a fixed-point
solution.
(b) (Sufficiency): Let u∗ be a feasible vector together with a vector μ =
[μ1 , . . . , μR ]T , such that μ ≥ 0 and μT g(u∗ ) = 0. Assume that u∗ minimizes
the Lagrangian function, L̃ (5.11), over x ∈ Ω, as a fixed-point solution, i.e.,
u = u∗ satisfies
u = arg min L̃(u; x; μ), (5.14)
x∈Ω
5.3 Lagrangian Extension in a Game Setup 61
Remark 5.6 The Lagrangian optimality condition in Proposition 5.5 shows that u∗
is obtained by first minimizing the augmented Lagrangian function L̃(u; x; μ) with
respect to the second argument x, which gives x = φ(u) for every given u. The next
step involves finding a fixed-point solution u∗ of φ by setting x = u, i.e., solving
u = φ(u). This u∗ , a fixed-point solution to the minimization of L̃ over x ∈ Ω, the
following holds: L̃(u∗ ; u∗ ; μ) ≤ L̃(u∗ ; x; μ), ∀x ∈ Ω. Note that u∗ thus obtained
depends on μ, u∗ (μ). An optimal μ∗ is achieved by solving
μ∗ T g u∗ μ∗ = 0 and μ∗ ≥ 0 (5.15)
Example 5.7 Consider the two-player Nash game presented in Example 5.1 with a
coupled constraint. The corresponding J˜(u; x) is
2 1
˜
J (u; x) = 2x1 − 2x1 − x1 u2 + x2 − x2 − u1 x2
2
2
Note that as Ji and gr are differentiable convex functions and Ω = Rm , the La-
grangian function L̃(u; x; μ) is convex with respect to x, so the Lagrangian mini-
mization is equivalent to the first order necessary condition. Thus in the presence of
convexity the first order optimality conditions are also sufficient.
The primal and dual optimal solution pairs are characterized by the following
result (Theorem 2 in [120]).
so that
L̃ u∗ ; u∗ , μ∗ ≤ L̃ u∗ ; x; μ∗ , ∀x ∈ Ω
Therefore (5.10) holds and u∗ is an optimal NE game solution with J˜∗ = J˜(u∗ ; u∗ ).
Using (5.17), evaluated at μ∗ , and the foregoing relations yields
D μ∗ = min L̃ u; x; μ∗ x=u = J˜ u∗ ; u∗
x∈Ω
The separability in the second argument of both NG cost function and constraints
ensures that D(μ) in (5.17) can be decomposed. Such a decomposition result for
the minimization of L̃(u; x; μ) (Theorem 3, [120]) is presented next. The result
shows that the minimum of L̃(u; x; μ) with respect to x ∈ Ω can be obtained by
minimizing a set of one-argument Lagrangian functions. Thus the fact that both the
NG-game cost and the constraints are separable in the second argument is exploited
to show that the dual NG cost function D(μ) can be decomposed and, equivalently,
found by solving a modified Nash game with no coupled constraints.
The decomposition result is given next.
Proposition 5.9 (Theorem 3, [120]) Consider G(M, Ω i , Ji ). Let the action space
Ω be a compact and convex subset of R and Assumption 4.3 is satisfied. The
m
m
D(μ) = Li u∗−i (μ), u∗i (μ), μ , (5.19)
i=1
where
Li (u−i , xi , μ) = Ji (u−i , xi ) + μT g(u−i , xi ) (5.20)
and u∗ (μ) = [u∗i (μ)] ∈ Ωminimizes a set of Li defined in (5.20) over xi ∈ Ωi as a
fixed-point solution, ∀i ∈ M. In other words, ui = u∗i (μ) satisfies
Proof By Proposition 5.5, the necessary conditions for NE optimality with respect
to the Lagrangian L̃, (5.11), require one to solve
m
L̃(u; x; μ) = Li (u−i , xi , μ) (5.22)
i=1
5.5 Hierarchical Decomposition in a Game Setup 65
m
∂
Lj (u−j , xj ; μ) = 0, i = 1, m
∂xi
j =1
∂
Li (u−i , xi ; μ) = 0, i = 1, m (5.23)
∂xi
Therefore, component-wise (5.21) is the same as (5.23). To either (5.21) or (5.23)
one needs to find a fixed-point solution. Now (5.23) are the first order necessary
conditions for minimizing Li (5.20) with respect to xi . Since Ji and gr are convex,
they are also sufficient. For each given u−i , from (5.23) one obtains xi∗ = xi (u−i ),
so that
xi (u−i ) = arg min Li (u−i , xi , μ) , i = 1, m (5.24)
xi ∈Ωi
xi (u−i ) = ui , ∀i = 1, . . . , m
for a fixed-point vector denoted u∗ = [u∗i ] and x = [u∗i ], which depends on μ. With
this u∗ let us return now to the value functional in (5.22). The first step taken in
order to obtain u∗ was minimization with respect to x, so that from (5.22) one has
m
min L̃(u; x; μ) = min Li (u−i , xi , μ), x∈Ω
x∈Ω x∈Ω
i=1
m
min L̃(u; x; μ) = min Li (u−i , xi , μ) (5.25)
x∈Ω xi ∈Ωi
i=1
for any given u. Now evaluating (5.25) at the fixed-point u∗ = [u∗i ], x = [u∗i ] ob-
tained as above, one can write
m
min L̃(u; x; μ) = min Li (u−i , xi , μ)
x∈Ω u=u∗ ,x=u∗ xi ∈Ωi ui =u∗i ,xi =u∗i
i=1
m
= Li u∗−i (μ), u∗i (μ), μ
i=1
66 5 Computational Results for Games with Coupled Constraints
m
min L̃(u; x; μ) = min Li (u−i , xi , μ)
x∈Ω x=u xi ∈Ωi xi =ui
i=1
The proof is completed by using (5.17) and recalling that u∗ is a fixed-point solution
to the set of m optimizations (5.24), i.e., equivalently u∗ is an NE solution to the
Nash game with cost functions Li , (5.20).
5.6 Notes
6.1 Introduction
This book is focused on methodologies and game theory for optical networks
from a control perspective. The general setup is that given an optical communica-
tion network, some performance measures to be optimized among many networks
units/players/channels, one must design an algorithm that achieves as good a per-
formance as possible for each channel. In the following two chapters we consider
the problem of developing such algorithms from a game theoretical perspective.
Channels seek to re-adjust parameters such that they maintain an optimal quality of
service (QoS) (OSNR). Because of optical networks’ characteristics, game theoretic
formulations and results cannot be ported directly from other application domains
(e.g. wireless networks or congestion control). These characteristics are either phys-
ical constraints or network specific topologies that translate into various mathemat-
ical constraints in the game formulation: coupled as opposed to disjoint action sets,
constraints on player’s interaction (global versus localized interaction, i.e., one shot
end-to-end games versus stage or partitioned games), constraints on players’ ac-
tions (global constraints versus propagated/modified constraints). In this chapter we
present the background and mathematical modeling on optical networks needed to
develop these game theoretically inspired algorithms.
Optical wavelength-division multiplexed (WDM) communication networks are
evolving beyond statically designed point-to-point links. That is, one goal is how
to realize reconfigurable networks with arbitrary topologies, while at the same time
maintaining network stability, optimal channel transmission performance and qual-
ity of service (QoS), [92].
At the physical transmission level, one parameter that directly determines chan-
nel performance and QoS is the bit-error rate (BER). BER in turn, depends on op-
tical signal-to-noise ratio (OSNR), dispersion, and nonlinear effects, [2]. Typically
in link optimization, OSNR is considered as the dominant performance parameter,
with dispersion and nonlinearity effects being kept low by proper link design. This
is called the noise-limited regime versus the dispersion-limited regime as these two
regimes can be treated separately. Detailed OSNR computations are needed to de-
termine whether the channel OSNR on a given optical device or route is acceptable
[152]. Therefore, OSNR optimization can be directly translated to QoS optimiza-
tion.
The dominant impairment in OSNR is given by noise accumulation in chains of
optical amplifiers and its effect on OSNR, [45]. A traditional approach uses a static
budget of fiber and device impairments along a fiber link with sufficient tolerance
margins added [126]. Therefore at least a desired OSNR is achieved for each chan-
nel. The process is repeated manually after reconfiguration. An optimum map of the
amplifier gain distribution has been proposed in [85]. The map not only minimized
the effects of the nonlinearity, but also gave rise to the minimum OSNR for a given
average power and the minimum average power for a given OSNR. In fact, power
control has been used to solve OSNR equalization problems since the 1990s. The
primary objective of power control is to obtain a satisfactory channel OSNR by reg-
ulating channel input power. In [34], a heuristic algorithm was proposed for on-line
OSNR equalization in single point-to-point links, while an analysis of the perfor-
mance limits associated with end-to-end equalization was performed in [154, 155].
The algorithm in [34] is centralized, in that each channel needs to know the OSNR of
all other channels. OSNR equalization has been formulated as a static optimization
problem in a point-to-point link in [45]. Optimal channel optical power vector at the
transmitter is found by solving an eigenvalue problem for the system transmission
matrix, composed of link gains of all channels. The proposed algorithm, developed
only for a single point-to-point link, requires centralized, global link information.
Current approaches that are developed for single links are not appropriate for
reconfigurable optical networks, where the length of the links and the number of
devices in a link are changing. Moreover, in these networks arbitrary topologies
can be formed, different channels can travel via different optical paths, and also
different channels can have different levels of QoS (OSNR) requirements. There-
fore one seeks to adjust network parameters (optical power, gains) in an optimal
way, based on on-line network feedback (from receiver to transmitter and various
nodes). This adjustment can result in increased network flexibility and capacity,
so that the needed quality of service (QoS) is ensured whenever a new channel is
added/dropped.
These observations justify the need for on-line network control algorithms to-
wards OSNR optimization that have provable convergence properties for general
network configurations. Moreover, particularly useful are decentralized algorithms,
such that channel power at the transmitter (Tx), can be adjusted based on feedback
from the corresponding receiver (Rx) only, plus other channel measurements. Fig-
ure 6.1 depicts one such feedback system (encapsulating the mesh network above)
6.1 Introduction 73
with a power control algorithm. We take channel signal power at the sources (chan-
nel power) as input to the optical network system and channel OSNR output and
other network measurements as feedback signals.
This problem on how to design such algorithms can be approached from an opti-
mization perspective or from a game theoretical perspective. A game theoretical one
is suited particularly because it naturally enables one to capture user preferences and
player incentives and leads to distributed algorithm. This is the approach we take in
this book, although from time to time we refer to and compare it to an optimization
approach.
Such power control algorithms can be implemented in an iterative feedback man-
ner. Each channel OSNR is measured by the optical spectrum analyzer (OSA) at the
optical switches and the destination. Feedback network measurements are trans-
mitted via optical service channels (OSCs) in optical networks. Each source has
a computation unit which performs an algorithmic calculation and determines the
corresponding channel input power.
This problem is similar to power control in wireless communication systems, a
topic which has been explored extensively, via either centralized approaches [164],
or decentralized, noncooperative game approaches [8, 41, 61]. There are several
differences that make this problem more challenging in optical networks. In wireless
networks, channels are characterized by static loss/gain only, with single Tx to Rx
links, or multiple links with no gains (ATM networks). Optical networks bring a
new combination of challenges: amplified spans, multiple links, accumulation, and
self-generation of optical (ASE) noise, as well as cross-talk generated at the routing
elements.
In the following chapters we consider this problem of developing algorithms
from a game theoretical perspective for re-adjusting network parameters such that
channels maintain an optimal QoS (OSNR), while at the same time taking into ac-
count constraints imposed by dispersion and nonlinearity. The material is based
on mathematical modeling, presented in this chapter, organized as follows. In the
next two sections (Sects. 6.2 and 6.3) we provide basic background on transmis-
sion basics in optical networks and on general topologies to be studied. In Sect. 6.4
the analytical mathematical model for OSNR is presented for point-to-point link
(Sect. 6.4.1) and for network topology (Sect. 6.4.2), both as recursive relations and
end-to-end relations. Link power capacity constraint is introduced in Sect. 6.4.3.
74 6 Optical Networks: Background and Modeling
The collection of lightpaths is called the virtual topology in the optical layer.
The optical switch may be equipped with a wavelength converter. Then some chan-
nels may be converted from one wavelength to another as well along their route by
wavelength conversion.
Consider an optical network that is defined by a set of optical fiber links L =
{1, . . . , L} connecting optical switch nodes. A link l ∈ L is composed of Nl optical
amplified spans. Each span includes an optical fiber followed by an optical amplifier.
These nodes allow channels in the network to be added, dropped or routed and
can also provide the flexibility of channel power adjustments [2].
Optical switch nodes could be optical cross-connects (OXCs) or optical add/drop
multiplexers (OADMs). An OXC deals with multiple wavelengths and selectively
drops some of these wavelengths locally or adds selected wavelengths while letting
others pass through. An OADM provides a similar function but at a much smaller
size.
There are various topologies that an optical network can be configured into, the
simplest one being the point-to-point link topology (Fig. 6.2).
Another simple topology is a single-sink topology where channels are added at
different links, but dropped off at the output of the same link (e.g. Fig. 6.3), multi-
link topology (Fig. 6.4), moving to the most complex, mesh topology (Fig. 6.5(a)).
A set of channels, M = {1, . . . , m}, corresponding to a set of wavelengths,
are transmitted across the network by intensity modulation and wavelength-
multiplexing. We denote by Ml the set of channels transmitted over link l ∈ L.
Also, we denote by Ri , i ∈ M, the set of links from the associated Tx to the corre-
sponding Rx that channel i uses in its optical route. For a link l ∈ Ri , we sometimes
use the notation Ri,l to denote the set of links previous to link l on the path Ri , i.e.,
for ∀l ∈ Ri ,
Ri,l := l
∈ L | l
∈ Ri , l
precedent to l
We denote by ui , n0i , pi and ni the channel signal power at Tx, the chan-
nel noise power at Tx, the channel signal power at Rx and the channel power
at Rx, respectively, for channel i ∈ M (illustrated in Fig. 6.5(a)). We let u =
6.3 Topologies and Setup 77
Fig. 6.5 A wavelength-routed optical network configuration: mesh, link, and span
Let channel signal and noise power at the output of the link (at Rx) be denoted by
pi , nout out
i , respectively, or in vector form p and n . These are taken to be the same
as the channel signal and noise power at the output of span N , i.e.,
pi = pi,N , p = pN (6.2a)
i = ni,N ,
nout nout = nN (6.2b)
6.4 Power Control and OSNR Model 79
The optical signal-to-noise ratio (OSNR) at Rx, for the ith channel, i ∈ M, denoted
by yi is defined by
pi
yi := out or y = p · /nout (6.3)
ni
in vector form.
The following basic propagation relations will be used. An optical amplified
span s, s = 1, . . . , N on each link is composed of an optical fiber with attenuation
(loss) coefficient, Ls , which is wavelength independent, and an optical amplifier. An
optical amplifier simultaneously increases the optical power of all channels i ∈ M
multiplexed (transmitted) over the link, with a wavelength-dependent gain (spectral
gain shape).
Let us consider an amplifier on span s and let the gain it provides to channel i be
denoted by gi,s . Then the span transmission of span s for channel i is defined as
s
Hi,s = hi,q (6.5)
q=1
Assumption 6.1 ASE noise does not contribute to the amplifier gain saturation.
ASE noise power generated within the bandwidth of the ith channel by the sth
amplifier is denoted by ASEi,s and is given as [2]
where nsp > 1 is a constant factor called amplifier excess noise factor, h is the Planck
constant, Bo is the optical bandwidth, and νi is the optical frequency corresponding
to wavelength λi , where λi = c/νi , with c the speed of light. Typical values are λi
in the 1550 nm (infrared) range and νi in the THz range. In vector form we denote
ASEs = [ASE1,s . . . ASEi,s . . . ASEm,s ]T .
Noise is generated additively at each amplifier. Moreover in cascaded amplifiers,
each OA amplifies simultaneously signal and noise power, since ASE is in-band
80 6 Optical Networks: Background and Modeling
noise (same wavelength or optical frequency as the signal channel). Thus, channel
signal and noise power at the output of span s on the link, denoted by pi,s , ni,s are
given for ∀s = 1, . . . , N as
where Diag(v) denotes the diagonal matrix with the elements of vector v as diagonal
entries.
In the following we present the OSNR model, i.e., its relation to the input optical
signal and noise power. We consider only forward propagation of signal and noise in
steady-state, [45], i.e., we do not consider amplifier gain dynamics. This is justified
for any OSNR game or optimization problem where power adjustments need to be
made at steady-state, after updating network topology.
We can consider separately two cases depending on the amplifier operation
mode: gain control mode (Lemma 6.1), and power control mode (Lemma 6.2) in
[119]. In the gain control mode, the gain of an amplifier is kept constant, while in
the power control mode the total output power is kept constant. The case of gain
control mode leads to a simpler mathematical model (given in Lemma 6.1): OSNR
on a particular channel does not depend on the power of the rest of the channels.
However, the typical operation mode is the automatic power control mode because
this ensures that a constant power is launched into the next span of a link, which
leads to uniform total power distribution across a link and thus limits the nonlinear
effects [2, 85]. The model for this case is given in Lemma 6.2.
The following lemma gives the model for gain control mode.
Lemma 6.1 Let u, n0 be the optical signal and noise power at the input (Tx) of a
point-to-point optical link with N spans and amplifiers in gain control mode. Under
Assumption 6.1 the optical signal power p, noise power nout and the OSNR at Rx y
(6.3) are given in vector form as
p = Diag(HN )u (6.9a)
N
−1
n out
= Diag(HN ) n +0
Diag(Hr ) ASEr (6.9b)
r=1
and
!
N
−1
y = u. n0 + Diag(Hr ) ASEr (6.10)
r=1
6.4 Power Control and OSNR Model 81
Proof The proof follows by developing propagation relations for both signal and
noise. Using (6.8a), (6.8b) yields
Using these recursively after s together with (6.1a), (6.1b), yields for ∀s = 1, N
s
ps = Diag(hq )u,
q=1
s
s
s−2
ns = Diag(hq )n0 + Diag(hv ) ASEs−r−1 + ASEs ,
q=1 r=0 v=s−r
or
s
ps = Diag(hq )u,
q=1
s
ns = Diag(hq )n0
q=1
s s−r−1 −1
s−2
+ Diag(hv ) Diag(hv ) ASEs−r−1 + ASEs
r=0 v=1 v=1
"s
After some manipulation, factoring out v=1 Diag(hv ), denoting s − r − 1 by new
index r we obtain
s
ps = Diag(hq )u,
q=1
r −1
s
s
ns = Diag(hq ) n +0
Diag(hv ) ASEr
q=1 r=1 v=1
s
Diag(Hs ) = Diag(hs ) Diag(hs−1 ) · · · Diag(h1 ) = Diag(hq )
q=1
82 6 Optical Networks: Background and Modeling
ps = Diag(Hs )u,
s
−1
ns = Diag(Hs ) n + 0
Diag(Hr ) ASEr
r=1
Using these for s = N with p = pN , nout = nN (see (6.2a), (6.2b)) gives (6.9a),
(6.9b). Then (6.10) follows immediately by using these into (6.3).
Next one can use this result and develop a model for the case when amplifiers are
operated in automatic power control mode, which is the typical operation mode for
optical amplifiers when the network is in steady-state, [2]. We will see that this case
leads to a more complex mathematical model; the inherent scaling on the total power
translates into coupling between all channels’ powers, and OSNR on a particular
channel does depend on all other channels’ powers.
The following assumption is used for optical amplifiers.
Assumption 6.2 All spans on each link have equal length and all optical amplifiers
are operated in automatic power control mode with the same total power target P
and have the same gain spectral shape [2, 85].
This assumption of same gain spectral shape and power target is a reasonable
assumption if all spans in a link have the same length, i.e., for uniformly designed
optical links, [2]. Then gi,s = gi for s = 1, . . . , N (same spectral shape), and by (6.6)
the ASE noise generated is the same at each amplifier on the link. In vector form
let g = [g1 . . . gi . . . gm ]T or g = [gi ] denote the vector spectral gain shape of any
amplifier on the link. Similarly for each s = 1, . . . , N let gs = [g1s . . . gis . . . gm
s ]T or
g = [gi ].
s s
Also by Assumption 6.2 at the output of each span s = 1, . . . , N one can write
,
1Tm ps = P (6.11)
having the loss coefficient ηs adjusted to achieve the constant total power target P
[45]. Thus under Assumption 6.2, the effective gain of an optical amplifier can be
written for s = 1, . . . , N , i ∈ M, as
gi,s = gi ηs , (6.13)
where gi is the corresponding spectral gain value for channel i, and ηs the filter loss.
As we will see next the scaling via ηs introduces coupling between channel powers,
and channel OSNR is no longer independent of other channels’ powers.
The following result gives the OSNR model for a point-to-point WDM link de-
veloped based on (6.2a), (6.2b), (6.11), (6.20), and (6.7a), (6.7b). An earlier similar
simple model can be found in [45].
Lemma 6.2 Let u, n0 be the signal and noise power at the input (Tx) of a point-
to-point optical link with N spans and amplifiers in power control mode. Under
Assumptions 6.1 and 6.2, the optical signal power p and the OSNR at Rx y (6.3) are
given in vector form as
p = Diag(HN )u (6.14)
where HN = P
[gN ]T u
gN and
y = u./ n0 + Γ u (6.15)
where Γ = [Γi,j ] is the (m × m) link system matrix
N
−1 ASE
r T
Γ = Diag gr g
P
r=1
and Diag(v) denotes the diagonal matrix with the elements of vector v as diagonal
entries.
where
N gs
j ASE i
Γi,j = (6.17)
gis P
s=1
and HN = [Hi,N ], i ∈ M
giN P
Hi,N = N
j ∈M gj uj
84 6 Optical Networks: Background and Modeling
Proof From Lemma 6.1, (6.9a), (6.9b), one can write for ∀s = 1, . . . , N
where ASEr = ASE, for r = 1, . . . , N by Assumption 6.2. Using (6.18a) into (6.11)
gives
= 1Tm Diag(Hs )u = HTs u
P (6.19)
Recall that Hs = [Hi,s ], i ∈ M where Hi,s is defined in (6.5) as
s
Hi,s = hi,q
q=1
hi,q = gi Lq ηq = gi δq , ∀q = 1, . . . , N, (6.20)
hq = gδq , ∀q = 1, . . . , N
where hs = [h1,s . . . hi,s . . . hm,s ]T = [hi,s ], g = [gi ]. Substituting (6.20) into the
expression above for Hi,s yields
s
s
Hi,s = gi δq = gis δq , ∀s = 1, . . . , N,
q=1 q=1
"s
or, by denoting ρs = q=1 δq ,
Hi,s = gis ρs , ∀s = 1, . . . , N
H s = gs ρ s , ∀s = 1, . . . , N (6.21)
so that
P
ρs = , ∀s = 1, . . . , N
[gs ]T u
6.4 Power Control and OSNR Model 85
Hence
P
Hs = gs , (6.22)
[gs ]T u
which when used in (6.18a) for s = N yields (6.14). Note that component-wise one
can write
gs P
Hi,s = m i s
j =1 gj uj
Next, for the second part, observe that using (6.22) one can write for the last term
in (6.18b)
−1 1
−1
T
Diag(Hr ) ASE = Diag gr ASE gr u
P
where the fact that (gr )T u is a scalar was used. Thus replacing this into (6.18b)
yields
ns = Diag(Hs ) n0 + Γs u , (6.23)
where Γs is defined by
s
1
−1
T
Γs = Diag gr ASE gr
P
r=1
Thus, for OSNR, y = p./nout , (6.3), using p in (6.14) and nout = nN in (6.23) for
s = N , yields
y = p./n = u./ n0 + Γ u
with Γ = ΓN , or component-wise
pi ui
yi = OSNRi = = 0
ni ni + j ∈M Γi,j uj
Remark 6.3 This is the OSNR model typically used in optical links. Mathematically
the OSNR model in (6.16) is similar to the wireless signal-to-interference ratio (SIR)
model [8]. However, it has a richer system structure: Γ with Γi,j defined in (6.17)
is a full matrix, with cross-coupling terms, non-zero diagonal elements and all el-
ements dependent on network parameters (e.g., the gain and ASE noise of optical
amplifiers). We can rewrite (6.16) as
ui
yi := OSNRi = , (6.24)
X−i + Γi,i ui
86 6 Optical Networks: Background and Modeling
where
X−i = n0i + Γi,j uj . (6.25)
j ∈M,j =i
Hence OSNR is no longer a linear function of channel input power like SIR as in [8].
Furthermore, since optical amplifiers are often cascaded along the link, ASE noise
accumulates over many amplifiers and degrades channel OSNR as the number of
amplifiers increases, which is reflected by (6.16).
Let us look now at the network model. The same assumptions are used and similar
notations. Thus consider an optical network with a set L = {1, . . . , L} of links. A set
M = {1, . . . , m} of channels transmitted over the network. Let Ml = {1, . . . , ml }
be the set of channels transmitted over link l ∈ L. Each link l is composed of Nl
cascaded optical amplifiers and optical fiber spans and has P̂l as its constant total
power target. We denote by gi,l the gain experienced by channel i due to the spectral
gain shape (wavelength dependency) of the amplifier on link l. By Ri we denote the
route of channel i from Tx to Rx. Let the optical fiber from each Tx to the optical
switch that is connected be defined as a virtual optical link (VOL) (see Fig. 6.7).
Thus the set of VOLs, Lv , is equivalent to M. Therefore, the set of all optical links
and VOLs is denoted by L
= L ∪ Lv = {1, . . . , L
}, where L
= L + m.
We define two network connection matrices. First, let A = [Ai,l ]m×L
denote the
channel transmission matrix, where
#
1, channel i uses link or virtual link l;
Ai,l =
0, otherwise
Similar notations are used for the output powers: pl denotes the m-dimensional
output power vector at each link l (all channels), and pi denotes the L
-dimensional
output power vector for each channel (all links), respectively.
Based on the network connection matrices, the following result gives intercon-
nection relation for all the channel powers at link l, i.e., how the input channel
powers ul are obtained from the other links or from the virtual optical links (VOL).
The first case we treat is the end-to-end case where we assume that channel pow-
ers are adjustable only at the transmitter sites. Later on in this section we treat the
recursive case where channel powers are individually adjustable also at the input of
each link by some adjustment factors γ , hence called γ -link. In this first case the
channel signal power at the input of link l, denoted by ui,l is identical to the signal
power at the output of the preceding link l
of link l on route Ri , i.e.,
Lemma 6.4 The optical signal powers at the input of link l, ul , l ∈ L, are given in
vector form as
ul = VlT p, (6.27)
where p = [p1 . . . pi . . . pm ]T and pi = [pi,1 . . . pi,k . . . pi,L
]T , with
#
pi,k , k∈L
pi,k =
ui , k ∈ Lv
where ui is the
' signal power of channel i at Tx, and the matrix Vl is the direct sum
of Vi,l , Vl = mi=1 Vi,l , where ∀i = 1, . . . , M, l ∈ L,
Proof For channel i, i ∈ M its signal power launched into link l, l ∈ L is trans-
mitted either from one of previous links of link l or one of virtual links (Txs). Let
k
∈ L
be such a link. Then Bk
,l = 1, Ai,k
= 1 and
Notice that there is one and only one k
such that both Bk
,l and Ai,k
are non-zero.
So the above equation can be rewritten as
ui,l = Ai,l Bk
,l Ai,k
pi,k
+ Bk,l Ai,k pi,k = Ai,l γi,l Bk,l Ai,k pi,k
k∈L
,k =k
k∈L
where
#
pi,k , k ∈ L,
pi,k =
ui , k ∈ Lv ,
and ui is the signal power of channel i at Tx. Since L
= {1, . . . , L
}, ui,l can be
expressed as
⎡ ⎤
pi,1
⎢ ⎥
ui,l = [Ai,l B1,l Ai,1 , . . . , Ai,l BL
,l Ai,L
] ⎣ ... ⎦ = γi,l · Vi,l · pi (6.28)
pi,L
'm
Recalling that Vl = i=1 Vi,l = Diag(Vi,l ), the vector form of the foregoing
yields (6.27).
indicating not only the connections between links but also channel routing condi-
tions. From the system control point of view, for the network system with channels
transmitted over it, input of this system is the signal power at the input side of each
link and the output is the output signal power of each link.
Remark 6.6 Since noise is propagated in the same manner via the optical switch (no
O/E conversion), it can be immediately seen that, as in (6.26),
i,l = ni,l
.
nin out
where
#
nout
i,k , k ∈ L,
nout =
i,k n0i , k ∈ Lv ,
6.4 Power Control and OSNR Model 89
l = Vl n.
nin T
or in vector notation
l ,
eT pl = P ∀l ∈ L,
yl := pl ./nout
l
and
yl = ul ./ nin
l + Γl ul (6.31)
90 6 Optical Networks: Background and Modeling
Nl
−1 ASEl
r T
Γl = Diag grl gl
l
P
r=1
Component-wise
pi,l ui,l
yi,l := = in (6.32)
ni,l ni,l + j ∈Ml Γli, j uj,l
where
Nl s
gj,l ASEi,l
Γli,j = (6.33)
s
gi,l Pl
s=1
Based on (6.31) a recursive OSNR model for the network can be obtained.
Lemma 6.8 Consider a link l and a channel i, i ∈ Ml , and let l
be the link prece-
dent to link l for channel i along path Ri . Let yi,l , and yi,l
denote the channel
OSNR at the output of link l and l
, respectively. Then the following OSNR recursive
relation holds for ∀i ∈ Ml :
1 1 uj,l
= + Aj,l Γli,j (6.34)
yi,l yi,l
ui,l
j ∈M
Proof Let us write (6.32) with the help of the connection matrix A, such that for
∀i ∈ Ml
ui,l
yi,l =
nin
i,l + j ∈M Aj,l Γli,j uj,l
6.4 Power Control and OSNR Model 91
so that
1 nin
i,l
uj,l
= + Aj,l Γli,j
yi,l ui,l ui,l
j ∈M
A similar model as in Lemma 6.2 can be developed for the network end-to-end
OSNR albeit with a richer network system matrix Γ (Fig. 6.9). The model can be
derived in a similar way based on recursive use of propagation relations. The reader
is referred to [119] for the full proof of this result as well as for extension to the case
when cross-talk at the optical nodes (OXC or OADM) is included (Lemmas 2 and 3
[119]).
Proposition 6.9 (Lemma 2, [119]) Under Assumptions 6.1 and 6.2, the OSNR of
channel i at Rx, yi is given as
ui
yi = , (6.36)
n0i + j ∈M Γi,j uj
where Γli,j and Hi,Nq are defined in Corollary 6.7 and Rl,i denotes the set of links
on the path Ri before link l.
The network OSNR model (6.36) has the same form as the link OSNR model
(6.2), but with a more complex system matrix Γ = [Γi,j ], which we refer to as net-
work system matrix. The OSNR model shows that as input power of one channel
92 6 Optical Networks: Background and Modeling
increases, thereby increasing its OSNR, the noise in the other channels increases,
thus decreasing their OSNRs. Based on the OSNR model, we take the channel in-
put power as input to the optical network system and OSNR output and other net-
work measurements as feedback signals to design power control algorithms such
that channel OSNRs can be optimized.
Now we consider the second case, whereby in an optical network, channel pow-
ers are adjustable not only at Txs but also at optical switches [2]. Thus channel
powers can be individually adjusted in the beginning of each optical link via an
adjustment parameter for channel i on link l is denoted by γi,l , γi,l ∈ [γmin , γmax ]
(Fig. 6.10). This is the case called γ -link topology. Then
Lemma 6.10 The optical signal powers at the input of link l, ul , l ∈ L, are given
in vector form as
ul = Diag(γ l )VlT p, (6.37)
where γ l = [γ1,l , . . . , γm,l ]T , and Vl , p are defined as in Lemma 6.4.
Now, following Remark 6.6, since the adjustment factors affect both input signal
and noise power simultaneously, one can write
That is, for noise power a relation similar to (6.37) holds, i.e.,
l = Diag(γ l )Vl n
nin T
nin
i,l γi,l nout
i,l
1
= =
ui,l γi,l pi,l
yi,l
Thus one can see that in this second case with γi,l adjustment factors (γ -link), the
same recursive OSNR model as in (6.35) and Lemma 6.8 is valid on a per-link basis.
Recall that in Sect. 6.2 we mentioned that nonlinear effects on optical fiber along
links limit the total input power. After each span of optical fiber, the total power is
influenced by the gain distribution of optical amplifiers [33]. Under Assumption 6.2,
the same total power is launched into each span of a link. This leads to a uniform
94 6 Optical Networks: Background and Modeling
total power distribution along the link, which minimizes the effects of optical non-
linearity [85].
However, this optimal gain distribution of optical amplifiers is not applicable
to limit nonlinear effects on the optical fiber at the beginning of each link, i.e.,
the fiber segment between Txs (or optical switches) and the first optical amplifier
along the link. Since the effects of the nonlinearity can be reduced by lowering
the total launched power, the following condition is imposed as the coupled power
constraint, also called the link capacity constraint in optical networks:
ui,l ≤ P̂l , ∀l ∈ L (6.38)
i∈Ml
where ui,l is the signal power of channel i at the input of link l. When channel i is
transmitted from its associated Tx directly to link l, ui,l = ui , where ui is the signal
power of channel i at Tx. Otherwise, due to the power propagation along links and
among networks, ui,l is a function of ui , or u where u is defined as the signal power
vector of all channels at Tx. This is similar to capacity constraints in flow control
[13]. However, capacity constraints in flow control are unchanged along links, while
link coupled power constraints in optical networks are propagated along links. The
foregoing (6.38) can be rewritten as
gl (u) = ui,l − P̂l ≤ 0, ∀l ∈ L (6.39)
i∈Ml
Example 6.11 In a single point-to-point WDM fiber link (Fig. 6.6), channel i is
transmitted from its associated Tx directly to the link. Then one can drop the link
index and (6.39) is reduced to
g(u) = ≤ 0.
ui − P (6.40)
i∈M
In optical networks, the coupled power constraints are propagated along links,
and due to this, convexity of (6.39) is not automatically ensured, as shown in the
next example.
Example 6.12 Consider the network shown in Fig. 6.12 with three links (L = 3)
and three channels (m = 3) transmitted over links. Each link l has Nl spans and a
l . Let us study the convexity of the constraint on the third link,
total power target P
g3 (u):
3
g3 (u) = u3,1 + u3,2 − P
6.5 Notes 95
By using (6.26), i.e., ui,l is identical to pi,l
with l
being the preceding link of l on
route Ri , we have
3
g3 (u) = p2,1 + p2,2 − P
Then by using the system constraint (6.29) on link 2, we get
3 = −p2,3 + (P
g3 (u) = p2,1 + p2,2 − P 2 − P
3 )
As can be easily seen g3 (u) is a more intricate function of u. Moreover, when com-
2 2
puting the Hessian matrix of g3 (u), the sign of ∂ g3 2(u) and ∂ g3 2(u) may be different.
∂u1 ∂u2
Thus the Hessian matrix ∇ 2 g3 (u) may not be positive semidefinite, that is, g3 (u)
may not remain convex independently of link parameters.
6.5 Notes
Abstract This chapter provides the basic formulation of a game framework to-
wards solving the OSNR optimization problem in optical networks. We restrict
the analysis to single point-to-point optical links, as the simplest network topol-
ogy. A Nash game played among channels is employed towards maximizing OSNR
firstly without coupled link capacity constraint. Then for incorporating the coupled
power constraint, two approaches are considered—an indirect and a direct one,
based on Lagrangian pricing and duality extension. Sufficient conditions are de-
rived for the existence and uniqueness of an NE solution for both approaches. Two
convergent iterative algorithms are developed towards finding the NE solution.
a single point-to-point optical link. A Nash game played among channels is em-
ployed towards maximizing OSNR firstly without coupled link capacity constraint
(Sect. 7.2). Channel optical signal-to-noise ratio (OSNR) optimization is formulated
as an m-player noncooperative game. Conditions for the existence and uniqueness of
the game equilibrium solution are given. We discuss an iterative algorithm for power
control, which uses only channel feedback measurements and which is shown to
converge to the equilibrium solution [115, 118]. Then for incorporating this link ca-
pacity coupled constraint, two approaches are considered: an indirect penalty-based
approach (Sect. 7.3) and a direct Lagrangian pricing approach (Sect. 7.4). Sufficient
conditions are derived for the existence and uniqueness of an NE solution for both
approaches. Two convergent iterative algorithms are developed towards finding the
NE solution.
The OSNR game problem is similar to power control via noncooperative game ap-
proaches in wireless networks, a topic which has been explored extensively in works
such as [8, 133]. The problem needs to be formulated and addressed separately for
optical networks, as there are several differences when compared to wireless net-
works [8]. In wireless networks, channels are characterized by uncoupled loss/gain
with single Tx to Rx links. Optical networks have cascaded amplified spans, accu-
mulation and self-generation of optical noise, cross-talk and possible coupling, and
saturation [2].
Considering these differences and physical features of optical networks, relevant
questions that we address in this section are: what is a natural way of formulating
a tractable OSNR optimization game in optical links and then networks of general
topology? How is this game compared to the SIR optimization game in wireless
networks? Is there a natural decomposition that leads to iterative algorithms with
provable convergence, decentralized with respect to channels?
We concentrate on a basic game formulation for an end-to-end optical link, i.e., a
point-to-point link topology. Conditions for uniqueness of the Nash equilibrium as
well as proof techniques and conditions for selecting pricing parameters are different
than in the wireless case. This chapter focuses on a point-to-point link topology and
it provides a starting point for the next chapter where optical network topologies are
considered.
We study a point-to-point WDM fiber link shown in Fig. 6.6 in Chap. 6. We
assume that reconfiguration is finished, i.e., channels will not be added or dropped
while performing adjustments. A set M = {1, . . . , m} of channels are transmitted
over the link. The link consists of N cascaded spans of optical fiber each followed
by an OA (Fig. 7.1). All OAs have the same gain spectral shape and the gain value
for channel i is gi . We denote ui and n0i the signal power and noise power of channel
7.2 Games Without Coupled Constraints 99
i ∈ M at Tx, respectively. Similarly, we denote pi and ni the signal power and noise
power of channel i ∈ M at Rx, respectively. Let u = [u1 , . . . , um ]T = [ui ] denote
the vector form of the signal power at Tx. Equivalently, we write u = (u−i , ui )
with u−i = [u1 , . . . , ui−1 , ui+1 , . . . , um ]T in some context to represent the same
vector u. The signal power at Tx is typically limited for every channel. That is, ui
is in a bounded set Ωi = [0, umax ] where umax is a positive constant. We use Ω to
represent the Cartesian product of Ωi , i ∈ M, Ω = Ω1 × · · · × Ωm . We also let
Ω−i = Ω1 × · · · × Ωi−1 × Ωi+1 × · · · × Ωm . Thus ui ∈ Ωi and u ∈ Ω.
The OSNR of channel i at Rx denoted by yi , yi = pnii , is given as in (6.16), i.e.,
ui
yi = , (7.1)
n0i + j ∈M Γi,j uj
where X−i denotes the total interference on channel i due to other channels’ power.
The OSNR model reflects that the signal power of one channel can be regarded
as an interfering noise for others, which leads to OSNR degradation. Regulating the
optical powers at Tx, i.e., allocating optical power as a resource among channels,
aims to achieve a satisfactory OSNR for each channel at Rx.
Let us now formulate a game-theoretic approach to solve an OSNR optimiza-
tion problem from the user optimization point of view, based on this OSNR model.
100 7 Games in Point-to-Point Topologies
for any given u∗−i . Recall that NE optimality means that no player has an incentive
to change its action, since no further individual improvement in its cost is possible.
Let us consider that each individual cost function Ji is defined as the difference
between a pricing function Pi and a utility function Ui
The utility Ui is related to channel’s performance criteria, while the pricing term Pi
is used to penalize a channel for using too large an action (power). Minimizing the
cost function is equivalent to maximizing the net utility, i.e., the difference between
utility and pricing.
In general a pricing mechanism is known to improve the NE efficiency, and a
linear pricing is the simplest one [133]. For such a simple linear pricing term each
channel minimizes the cost function Ji :
(A.ii.2) ui = 0, ui = umax are not solutions to the minimization of the cost func-
tion Ji .
We construct a utility function that satisfies (A.ii.1), (A.ii.2). Note that OSNRi ,
i.e., yi in (7.2), is a strictly increasing function with respect to ui , and tends to 1/Γi,i ,
for infinite channel power. It is apparent that relation (7.2) bears a striking similarity
with the wireless SIR model, [8], even though a different physical mechanism is
present (ASE noise accumulation). The system matrix in the SIR model [8] has a
special structure with equal rows, which is instrumental in obtaining uniqueness
results. In contrast, for optical networks, Γ is a full general structure matrix, with
coupling due to all channels and all spans. Moreover, OSNRi is no longer a linear
function of ui and a direct logarithmic utility function of associated SIR as in the
wireless case cannot be applied. For the general full matrix Γ , one can define a
more general utility function Ui (u) here chosen to be a logarithmic function of the
associated channel’s OSNR, yi (u),
yi (u)
Ui (u) = βi ln 1 + ai , ∀i ∈ M (7.5)
1 − Γi,i yi (u)
where X−i is given as in (7.2), as a function of the full system-matrix Γ . This de-
pendence will be instrumental in the following results. Therefore, the cost function
102 7 Games in Point-to-Point Topologies
Herein αi is the price set by the network/link and βi is set by each channel, re-
spectively. These act as weighting factors, quantifying the trade-off between pricing
(penalizing a channel for using large power) and the desire to maximize its utility.
From (7.6) it follows immediately that Ui satisfies (A.ii.1), i.e., the utility func-
tion Ui (u−i , ui ) is twice continuously differentiable in its arguments, monotonically
increasing and strictly concave in ui . Using (7.7), it can be shown that there exists a
non-empty interval from which to select βi /αi , such that (A.ii.2) holds. Thus αi , βi
are selected such that an NE solution is an interior point of the action set, and hence
is an inner NE (see remark after Definition 4.1 in Chap. 4). This is what we assume
in the following.
One can note that this is only a particular utility function that has certain nice
properties being logarithmic, in that it allows for closed-form expression of the
Nash equilibrium solution. Other examples can be given such as a linear function in
OSNR. A logarithmic function is analytically useful and moreover is widely used
as a utility function in flow control [7, 67, 139] and power control [3, 6, 8] for
general communication networks. In some cases, the logarithmic utility function is
intimately associated with the concept of proportional fairness [67].
Let us recall from Chap. 4 that a point u∗ is an NE when u∗i solves individ-
ual optimization problem Ji , given all channels on its path have equilibrium power
levels, u∗−i . Existence of an NE solution depends on existence of a common in-
tersection point for all players’ reaction curves, while uniqueness depends on the
particular problem. The following result characterizes the Nash equilibrium (NE)
solution.
Theorem 7.1 Consider G(M, Ωi , Ji ) with individual cost functions Ji , (7.7). This
game admits a unique NE solution u∗ if ai are selected such that
Γi,j < ai , ∀i ∈ M (7.8)
j =i
u∗ = Γ(−1 b̃ (7.9)
This result shows that while similar in some respects to wireless case, [8], for
optical networks a more general uniqueness condition is needed for uniqueness of
the NE solution.
2
Proof From (7.4, 7.7) and (A.ii.1) it follows directly that ∂ J2i > 0. Since the cost
∂ui
function Ji is strictly convex in ui , there exists a minimizing u∗i , for any given u−i ,
such that
J u−i , u∗i < J (u−i , ui ), ∀ui = u∗i
on the closed and bounded (compact) set [0, umax ]. Furthermore by (A.ii.2) u∗i is
inner.
To find u∗i we solve the necessary conditions ∂u
∂Ji
i
= 0. From (7.7) one obtains
ai βi
ai u∗i + X−i
∗
= , ∀i (7.10)
αi
rewrite (7.10) as
ai βi
ai u∗i + Γi,j u∗j = − n0,i , ∀i
αi
j =i
Γ(u∗ = b̃ (7.11)
where matrix Γ( and vector b̃ are defined as Γ( = [Γ(i,j ] and b̃ = [b̃i ] with
#
a, j =i ai βi
Γ(i,j = i b̃i = − n0,i
Γi,j , j = i αi
Remark 7.2 For a given Γ , the ai factors can be selected such that the diagonal
dominance condition (7.8) holds on Γ(. If Γ is itself diagonal dominant, then a pos-
sible choice is ai = Γi,i . Once Γ changes (e.g. upon reconfiguration) these factors
can be adjusted to satisfy (7.8).
104 7 Games in Point-to-Point Topologies
βi X−i (n)
ui (n + 1) = − , ∀i (7.12)
αi ai
as recursive relation for updating transmitter power level, based on (7.10). Thus
(7.12) corresponds to a parallel adjustment scheme (PUA) based on best response
(BR) (see Chap. 4 and [20]), whereby each player responds optimally to the pre-
viously selected action of the other players. Relation (7.12) requires the total inter-
ference factor X−i , which from (7.2) depends on all channel powers, uj , and all
channel gains, i.e., centralized information. However, using (7.2) one can express
(7.12) in terms of yi , i.e,
βi 1 1
ui (n + 1) = − − Γi,i ui (n) (7.13)
αi ai yi (n)
Lemma 7.3 If (7.8) holds, then algorithm (7.13) converges to the NE solution.
Proof Let ei (n) = ui (n) − u∗i where u∗ = [u∗i ] is the NE solution. Using (7.10, 7.12,
7.13) yields
1
ei (n + 1) = − Γi,j ej (n) (7.14)
ai
j =i
so that
) ) 1
)e(n + 1))
= max ei (n + 1) ≤ max
Γi,j ej (n)
∞ i i ai
j =i
(a) Proportional Pricing This is the case when pricing parameters αi are selected
to be proportional to the system matrix entry for channel i, Γi,i , i.e., αi = Γi,i ki , and
all βi are equal to 1. It can be shown that the scaling factors ki can be selected such
that all channels achieve some desired OSNR level, γi∗ . Using (7.2) and (7.10) for
the NE solution, one can write
1 ∗ ai 1
u = (Γi,i − ai )u∗i + , ∀i
γi∗ i Γi,i ki
In matrix vector form this yields
where εi = 1
γi∗ + ai − Γi,i , and vi = ai 1
Γi,i ki . Using the NE solution, (7.11), u∗ =
Γ˜ −1 b̃, and b̃ = v − n0 , yields after some manipulation
I − Γ˜ Σ −1 v = n0 (7.16)
assuming that εi = 0. Then if ρ(Γ˜ Σ −1 ) < 1, where ρ is the matrix spectral radius,
equation (7.16) has a unique solution v. Moreover, for εi = ε = 0, using (7.15), one
can rewrite this condition as an upper bound condition on the desired OSNR level
γi∗ , i.e.,
1
γi∗ < , ∀i (7.17)
ρ(Γ˜ ) + (Γi,i − ai )
Therefore, if (7.17) holds v can be uniquely found as in (7.16), and hence the ki
factors can be found such that the desired OSNR level γi∗ is achieved. This corre-
sponds to a centralized pricing strategy, and as in [8], shows the trade-off between
“gain” of the system matrix, in terms of the spectral radius, and the level of OSNR
achieved. Unlike [8], for the general model Γ these parameters and the upper bound
for the desired OSNR level are different for each channel.
(b) Decentralized Pricing For the case when βi can be adjusted individually, one
can show that if βi satisfies the lower and upper bounds
1 + (ai − Γi,i )γi∗ αi αi
βi > X −i , β < u a + Γ + n 0 , ∀i
1 − Γi,i γi∗
i max i i,j
ai ai
j =i
then each channel will achieve at least γi∗ , i.e., yi > γi∗ , with ui ≤ umax at each
iteration.
106 7 Games in Point-to-Point Topologies
(c) Minimum OSNR Level Finally, consider the case when all ui = umax and
yi ≥ γmin . Using the OSNR model in Chap. 6, it can be shown that a sufficient
condition for this is
1
γmin ≤
maxi∈M ( j ∈M Γi,j )
which shows the trade-off between minimum achievable OSNR, γmin , and the norm
(gain) of system matrix Γ . Unlike the wireless case, [8], the system matrix Γ plays
an essential role, i.e., a condition independent of the system matrix cannot be ob-
tained.
As long as coupling constraints are not considered these point-to-point link re-
sults are directly extendable to end-to-end games in optical network topologies (see
Chap. 8). In the next section we show how distributed links can be treated.
This Ω is coupled in the sense that one player’s action affects the feasible action
sets of the other players. The feasible action set for each channel i is the projection
set
# *
i (u−i ) = ξ ∈ Ωi
Ω u + ξ −
P ≤ 0 (7.19)
j
j ∈M,j =i
The first approach we consider in order to treat such a game is based on incor-
porating the coupled constraint into each cost function, thus in effect an indirect
approach. Part of the results in this section are based mainly on [105, 108, 110].
7.3 Games with Coupled Constraints: Indirect Penalty Approach 107
Let us consider then that for each channel i ∈ M an individual cost function
Ji : Ω → R is assigned which is defined as the difference between a pricing function
Pi : Ω → R and the same utility function as before , Ui : Ω → R:
i − Ui
Ji = P (7.20)
Specifically, the utility function Ui is the same as in (7.5), while the new pricing
function consists of two terms: a linear pricing term (as in Sect. 7.2.1) and a new
regulation (penalty) term, i.e.,
i (u−i , ui ) = αi ui + 1
P , ∀i ∈ M, (7.21)
−
P j ∈M uj
βmin
βi < Γj,i
, (7.23)
j =i,j ∈M aj
+
,
, βi Γj,i
αi >αmax - , (7.24)
βj aj
j =i,j ∈M
Proof (Existence) The action space Ω is a compact and convex set with a non-
empty interior. Each cost function Ji (u−i , ui ) is continuous and bounded and the
first and second partial derivatives of J
i (u−i , ui ) with respect to ui are well defined
on Ω except the hyperplane {u ∈ Rm | j ∈M uj = P }, given as
∂ Ji (u) 1 βi ai
= αi + − , ∀i ∈ M (7.25)
∂ui
(P − j ∈M uj )2 X−i + ai ui
∂ 2 Ji (u)
It follows that is positive and therefore Ji (u−i , ui ) is strictly convex in ui .
∂u2i
Since each point on the hyperplane {u ∈ Rm | j ∈M uj = P } is not an NE solution,
by Proposition 5.3, the game with the action space Ω admits an NE solution, which
is the inner NE solution of G(M, Ω i , Ji ).
Since an NE solution is inner, it follows from Proposition 4.6 that (4.6),
∂ Ji (u)
∂ui = 0, ∀i ∈ M, holds. The vector form is ∇ J (u) = 0, where ∇ J (u) is defined
as in (A.3).
(Uniqueness) The uniqueness part is proved by contradiction and involves rather
long manipulations. The reader is referred to [105, 110] for the detailed steps.
Based on the results in Theorem 7.4, let us briefly discuss parameter selection
strategies. The link sets fixed channel prices and each channel decides its willing-
ness βi to obtain a satisfactory OSNR, yi , denoted by βi (yi ). From the necessary
i (u)
condition, ∂ J∂u i
= 0, using (7.25) one obtains βi as a function of yi to be
αi
X−i + ai ui (yi )
βi (yi ) = X−i + ai ui (yi ) + , (7.27)
ai
ai (P − j ∈M,j =i uj − ui (yi ))2
7.3 Games with Coupled Constraints: Indirect Penalty Approach 109
where
X−i
ui (yi ) =
1/yi − Γi,i
For a given lower OSNR bound
γi , one can show that if βi is adjusted to satisfy the
lower bound
αi 1 + (ai − Γi,i )
γi
βi > X−i
ai 1 − Γi,i
γi
(1 − Γi,i γi )(1 + (ai − Γi,i )
γi )
+ X−i ,
−
ai (P
j ∈M,j =i uj − (P − j ∈M,j =i uj Γi,i + X−i )
γi )2
(7.28)
1 βi ai
αi + = , ∀i ∈ M (7.29)
−
(P j ∈M uj )
2 X−i + ai ui
Then it follows that for every given u−i , ri (u−i ) is single-valued. Furthermore, by
Berge’s Maximum Theorem (Theorem A.17), ri (u−i ) is continuous. Thus, there is
a unique point, ui = ri (u−i ) for every i ∈ M. In vector form,
u = r(u), (7.31)
or u∗ = r(u∗ ). Unlike the case with no coupling constraints in Sect. 7.2.1 where a
closed-form expression was found, here the reaction function ri is highly nonlinear
and the NE solution is analytically intractable.
110 7 Games in Point-to-Point Topologies
The reaction function is in fact the implicit solution of (7.29). Notice that over
Ω, the left-hand side of (7.29) is a monotonically increasing function with respect
to ui , while the right-hand side of (7.29) is monotonically decreasing. Therefore,
for every i ∈ M, there exists a unique intersection between the left-hand side and
the right-hand side of (7.29), which is the NE solution u∗ . One can get insights into
the NE solution’s characteristics by studying the reaction function.
Lemma 7.5 (See [110]) The reaction function r(u) has the following properties:
– r(u) ≥ 0 (non-negativity);
– If u > u
, then r(u) < r(u
) (monotonicity).
It follows from Lemma 7.5 that the reaction function r(u) is not standard. As
defined in [163], a function I (p) is standard if for all p ≥ 0, the following properties
are satisfied: I (p) > 0; if p ≥ p
, then I (p) ≥ I (p
); for all α > 1, αI (p) > I (αp).
Such a property typically helps in developing iterative algorithms.
In this section we present two iterative algorithms towards finding the unique inner
NE solution: a parallel update algorithm (PUA) and a gradient algorithm (GA). PUA
is developed using the reaction functions and GA is developed based on the gradient
descent method.
In the parallel update algorithm each channel i updates its input power ui based
on its associated reaction function ri (u−i ). Let
u(n) := u(0), u(1), . . . , u(n), . . .
denote the sequence of channel input power vectors, where u(0) is the initial chan-
nel input power vector. Then at each iteration time (n + 1), channel input power
ui (n + 1) is updated by
(PUA) ui (n + 1) = ri u−i (n) , ∀i ∈ M, (7.33)
If the reaction function r(u) were standard, PUA would be the standard power
control algorithm [163] and the convergence results in the synchronous case could
be directly applied to PUA. However, this is not the case here and convergence is
really difficult to prove. Results have been obtained only for the case of 2-players
(m = 2) and extension to m-player case, m > 2 is an open problem [110].
7.3 Games with Coupled Constraints: Indirect Penalty Approach 111
Remark 7.6 PUA may not converge when more than two channels (m > 2) exist.
The intuitive reason is that each channel updates its optical power only based on
instant costs and parameters, ignoring future implications of its action. Therefore,
at some iterations the total optical power of all other channels will exceed the target
power P when m > 2. To overcome these power fluctuations, one can use a relaxed
PUA, in which a relaxation parameter is used to determine the step size that each
channel takes towards finding the NE solution at each iteration step. For example,
(7.33) can be modified as
ui (n + 1) = (1 − μi )ui (n) + μi ri u−i (n) ,
where the coefficient 0 < μi < 1 and it can be shown that the relaxed PUA converges
if μi is restricted.
PUA is developed by using the reaction function r(u) but convergence can be
shown only for m = 2. As alternative a gradient algorithm (GA) can be designed
based on the gradient descent method [24]. Convergence of this algorithm can be
proved even when m > 2. For convenience, particularly for the simplification of
convergence proof, we follow the approach in [107, 109] and develop the algorithm
in the continuous-time domain.
We consider a model where each channel uses a gradient algorithm (GA) to up-
date its power, given as
dui ∂ Ji (u−i , ui )
u̇i (t) = = −μ , ∀i ∈ M, (7.35)
dt ∂ui (t)
where t is the continuous-time variable and the coefficient μ > 0.
Using (7.3) and (7.25), one can rewrite (7.35) as
1 βi ai
(GA) u̇i (t) = −μ αi + −
− j ∈M uj (t))2 m Γ˜i,k uk (t)
(7.36)
(P k=1
where
#
Γi,k , k = i
Γ˜i,k =
ai , k=i
Recalling the definition of OSNR, one can rewrite (7.36) as
1 βi ai
u̇i (t) = −μ αi + − (7.37)
− j ∈M uj (t))2 ( 1 + ai − Γi,i )ui (t)
(P yi (t)
Hence, at each iteration, individual channels need only the sum of power of all
channels and local measurements, namely, its own power and current OSNR level.
Thus the iterative algorithm (7.37) is distributed.
Next let us define a set Ω δ , which is a subset of the action space Ω. We slightly
modify the bounded set Ωi = [0, umax ] to be
(i = [umin , umax ],
Ω ∀i ∈ M
112 7 Games in Point-to-Point Topologies
We set umin to be sufficiently small such that Theorem 7.4 holds, i.e., u∗i > umin for
all i ∈ M and
∂ Ji
(u−i , ui ) < 0, ∀u−i , ∀i ∈ M (7.38)
∂u i ui =umin
Notice that because of the capacity constraint, the following statement is always
true:
∂ Ji
(u−i , ui )
> 0, ∀u−i , ∀i ∈ M (7.39)
∂ui − j ∈M,j =i uj
ui =P
Thus the set Ω δ is invariant under the algorithm (7.36). That is, the trajectory
remains inside Ω δ . It follows that the trajectory lies in Ω if the initial state is
in Ω δ . Moreover, the equilibrium of (7.36) in Ω δ is the unique NE solution u∗
of G(M, Ω i , Ji ).
The following result proves the convergence of algorithm (7.36).
Theorem 7.7 For G(M, Ω i , Ji ), let the initial condition u(0) ∈ Ω δ . Then the up-
date scheme (7.35) converges to the NE solution u∗ if
2
2P
amin > max Γj,i , (7.40)
umin i∈M
j ∈M, =i
Proof The proof is based on a Lyapunov approach. Let φi (u) := u̇i (t), where u̇i (t)
is defined in (7.35) or (7.36), and define a candidate Lyapunov function,
1 2
V (u) := φi (u) (7.43)
2
i∈M
Note that V (u) is restricted to the set, Ω δ . Because of the uniqueness of the NE
solution u∗ , φi (u) = 0, ∀i ∈ M, if and only if u = u∗ . Therefore, V (u) is strictly
positive for all u = u∗ .
7.3 Games with Coupled Constraints: Indirect Penalty Approach 113
From (7.36), taking the second derivative of ui with respect to time t, yields
2 βi ai Γ˜i,j
üi (t) = −μ + φj (u) (7.44)
− j ∈M uj (t))3 (
(P ˜ 2
j k∈M Γi,k uk (t))
where vector φ = [φi ] and matrix Θ(t) = [θij (t)]m×m are appropriately defined. If
Θ(t) is uniformly positive definite then V̇ (u) in (7.46) is negative definite, and the
system is asymptotically stable by Lyapunov’s stability theorem [69]. This is the
main idea of the proof and the reader is referred to [107] for the detailed steps. Then
φi (u(t)) = u̇i (t) → 0, ∀i ∈ M which implies that ui (t) converges to the unique NE
solution point u∗ .
A gradient algorithm has been studied for stability in a congestion control game
[4, 5] where a special structure matrix is automatically uniformly positive definite in
the proof. In contrast, herein the corresponding matrix Θ(t) is more general. System
dependent conditions are required for Θ to be positive definite.
Compared the conditions (7.40) and (7.41) for the stability of GA with the suf-
ficient conditions (7.22)–(7.24) for the existence of a unique NE solution, it read-
ily follows that βi is not only upper-bounded by (7.23), but also lower-bounded
by (7.41).
114 7 Games in Point-to-Point Topologies
In this section we consider Nash games with coupled constraints and their solv-
ability by a direct approach based on Lagrangian extension. Specifically we apply
directly the procedure developed in Chap. 5. We note that this direct approach has
advantages compared to the indirect one in the previous section. As we saw the indi-
rect, penalty-based approach does not allow for an analytically tractable NE solution
and developing decentralized algorithms is not immediate.
As in Sect. 7.2, each channel i ∈ M minimizes the original individual cost func-
tion Ji : Ω → R, with Ji defined in (7.4), not the penalized cost Ji as used the
previous section. The cost Ji is composed of a linear pricing and a channel utility
Ui related to OSNR maximization defined as before. Unlike Sect. 7.2, where cou-
pled constraints were not considered, here the action set is coupled as in (7.18),
i.e.,
Ω = u ∈ Ω | 1T u ≤ P (7.47)
and the action set of channel i ∈ M is the projection of Ω on channel i’s direction,
namely as in (7.19):
m
i (u−i ) = xi ∈ Ωi
Ω uj + xi ≤ P (7.48)
j =i
We denote the game by G(M, Ω i , Ji ), which is in the class of m-player Nash games
with coupled utilities and coupled constraints. It follows from Sect. 7.2 that Ji is
continuously differentiable in its arguments and convex in ui . Note that the over-
all coupled action space Ω is compact and convex as well. Then from Proposi-
tion 5.3, G(M, Ω i , Ji ) admits an NE solution. Moreover, due to the coupled con-
straint (7.47), solving directly for an NE solution of this game requires coordination
among possibly all channels and is impractical as seen in the previous section.
In the following, we use Lagrangian extension and decomposition results in
Chap. 5 as a natural way to obtain a hierarchical decomposition and compute an NE
7.4 Games with Coupled Constraints: Lagrangian Pricing Approach 115
solution of G(M, Ω i , Ji ). For this game, using the notations as defined in Chaps. 4
and 5, consider the separable NG cost function J((u; x), (4.7),
m
J((u; x) = Ji (u−i , xi ),
i=1
m
(
g (u; x) = g(u−i , xi ),
i=1
( x; μ), (5.11),
together with the augmented Lagrangian function L(u;
( x; μ) = J((u; x) + μT (
L(u; g (u; x)
( x; μ). In a fixed-point
where u∗ is such that u = u∗ satisfies u = arg minx∈Ω L(u;
notation,
D(μ) = min L(u;( x; μ) ,
x∈Ω (
arg min L=u
x∈Ω
or
( x; μ)
D(μ) := min L(u;
x∈Ω x=u
Proposition 7.8 Consider the coupled OSNR Nash game G(M, Ω i , Ji ) with cost
functions Ji (u−i , ui ), (7.4), subject to the linear constraint (7.47), i.e., over Ωi .
116 7 Games in Point-to-Point Topologies
m
∗ m
D(μ) = ∗
Li u−i , ui (μ), μ +
μ eT u∗−i − P (7.50)
i=1 i=1
where u∗ (μ) = [u∗i (μ)] is an NE solution to the Nash game G(M, Ωi , Li ) with cost
functions Li (7.51) and no coupled constraints.
g(u−i , xi ) = eT u−i + xi − P (7.52)
where e = [1, . . . , 1]T is the (m − 1) × 1 all ones vector. Then by using Proposi-
tion 5.9, D(μ) can be decomposed as
m
D(μ) = Li u∗−i (μ), u∗i (μ), μ
i=1
Recall that in D(μ) in Proposition 5.9 one has to minimize first with respect to
xi on the right-hand side, and then solve for a fixed-point solution. From (7.53)
it can be seen that only the first two terms depend on xi . Hence, substituting for
Li (u−i , xi , μ), (7.53), on the right-hand side of D(μ) and isolating the terms that
are independent of xi , yields
m
m
D(μ) = min Li (u−i , xi , μ)|xi =ui +
μ eT u−i − P
xi ∈Ωi
i=1 i=1
i.e., Li is the same as Ji , (7.7) for αi replaced by αi + μ, ∀i. Therefore, for each
given μ, the NE solution u∗ (μ) to the lower-level game G(M, Ωi , Li ) with cost Li
is unique and can be obtained as in Theorem 7.1 as
u∗ (μ) = Γ(−1 Diag 1./(α + μ) b0 − n0 (7.54)
u∗ (μ) = Γ(−1(
b(μ),
Proof The result can be proved by applying Proposition 4.5. We rewrite Li (7.51)
as
ai ui
Li (u−i , ui , μ) = (αi + μ)ui − βi ln 1 + 0 , ∀i ∈ M (7.57)
ni + j =i Γi,j uj
It can be seen that for any given μ ≥ 0, Li is jointly continuous in all its arguments
and
∂ 2 Li
> 0, ∀ui ∈ Ωi
∂u2i
∂Li
An inner NE solution can be found by solving the necessary conditions, ∂ui = 0.
From (7.57), one can obtain
ai βi
ai u∗i (μ) + Γi,j u∗j (μ) = − n0i , ∀i ∈ M
αi + μ
j =i
Γ( · u∗ (μ) = (
b(μ), (7.58)
where matrix Γ( and vector ( b(μ) are defined in (7.56). Therefore a unique solution
of (7.58) exists if the matrix Γ( is invertible. Notice that Γ( is a positive-entry matrix.
If (7.55) holds, then Γ( is strictly diagonal dominant. From Gershgorin’s Theorem
(Theorem A.7), it follows that Γ( is invertible and a unique solution of (7.58) exists,
u∗ (μ) = Γ(−1 · (
b(μ),
Next, based on the explicit solution (7.54) and on pricing coordination at the
higher level, a recursive hierarchical algorithm is discussed. By Theorem 5.8 applied
to the coupled OSNR game G(M, Ω i , Ji ) with costs Ji and coupled constraints
∗ ∗
(7.47), (u , μ ) is an optimal NE solution–Lagrange multiplier pair if and only if
u∗ is NG-feasible,
m
,
u∗i (μ) ≤ P u∗i ∈ Ωi , i ∈ M (7.59)
i=1
μ∗ ≥ 0, μ∗ ( m ∗
i=1 ui − P ) = 0 (complementary slackness condition) and the La-
grangian optimality condition
( u; x; μ∗
u∗ = arg min L (7.60)
x∈Ω x=u
holds. By Proposition 7.8 and (7.50), note that u∗ (μ) solving (7.60) can be found as
an NE solution to the modified Nash game G(M, Ωi , Li ) with costs Li (7.51), with
no coupled constraints. For every given price μ, this NE solution u∗ (μ) is unique
as in (7.54). Furthermore, from (7.54) it is seen that all components of u∗ (μ) de-
crease with μ. One can exploit the linear constraint and adjust the price μ to satisfy
the slackness condition. Instead of maximizing D(μ), the optimal price μ∗ can be
obtained such that the slackness condition holds, i.e.,as the point of interception
between the curve representing total power, u∗T (μ) = m ∗
i=1 ui (μ), with the level P
(Fig. 7.3). This method has the interpretation of a coordination mechanism. The link
as the coordinator sets the price at the optimal value μ∗ . The channels respond by
adjusting their power levels to u∗i (μ∗ ), which minimizes their own cost.
7.4 Games with Coupled Constraints: Lagrangian Pricing Approach 119
Link Algorithm Every K iterations of the channel algorithm, the new link price
is computed based on the received total power for all channels in the link uT (K) =
μ
j =1 uj (K) as
μ(k̄ + 1) = μ(k̄) + η +
uT (K) − P (7.61)
where k̄ is the link iteration number, η is the step-size and [z]+ = max{z, 0}.
This simple price update based on Fig. 7.3 requires only measurement of total
power. Moreover it corresponds to a gradient descent technique if link price is ad-
justed slower than channel powers. At the higher level, μ(k̄) acts as a coordination
signal that aligns individual optimality with the system constraint, (7.47) or (7.59).
This is the new price given to the channels, who repeat K iterations of the following
algorithm.
Channel Algorithm Based on a pricing μ(k̄) from the link, the optimal channel
power vector u∗ (μ(k̄)) can be found explicitly as in (7.54). This requires global
centralized information. However, the following iterative update algorithm can be
used:
βi 1 1
ui (n + 1) = − − Γi,i ui (n) (7.62)
αi + μ(k̄) ai yi (n)
where n is the channel iteration number. This is again a decentralized algorithm,
since the only information feedback is the individual channel yi , which can be mea-
sured in real-time, and the channel “gain”, Γi,i . For fixed μ this algorithm converges
to the optimal NE solution (7.54). Convergence of the combined algorithms will be
proved later on in the context of network topologies (see Chap. 8). Even if the op-
timal solution is coupled as in (7.54), it can be iteratively computed by using the
decentralized algorithm (7.62). Thus individual channels do not have to coordinate
with other channels at the lower-level game.
120 7 Games in Point-to-Point Topologies
7.5 Notes
Abstract This chapter provides approaches on how to deal with games in more
complicated network topologies, starting from the basic games in single point-to-
point WDM fiber links studied in Chap. 7. The multi-link topologies studied are
representative for selected paths extracted from a mesh configuration in which no
closed loops are being formed by channel optical paths. In network configurations,
coupled constraints are propagated along fiber links and constraint functions be-
come complicated from the end-to-end point of view. The non-convexity introduces
additional complexities for analysis. In this chapter, we present a partition approach.
More precisely, we partition the general multi-link structure into stages each stage
being a single link. Then we formulate a partitioned Nash game for general multi-
link topologies composed of ladder-nested stage Nash games. We also show that
convexity is ensured in single-sink topologies, so that a partition approach could be
based on single-sink stages.
8.1 Introduction
We have seen that in their most complete form games in optical networks belong
to the class of m-player games with coupled utilities and constraints. For single
point-to-point link topologies these were studied in Chap. 7. The purpose of this
chapter is to treat multi-link and mesh topologies. In fact end-to-end games without
coupled constraints are direct extensions of the point-to-point games in Chap. 7,
with the appropriate Γ network system matrix (as in Chap. 6) replacing the link
system matrix. Because of this we shall not consider them here. Instead we shall
focus on how to deal with games with coupled constraints in network topologies.
When coupled constraints are considered in network configurations, constraint
functions become complicated from an end-to-end point of view. This is due to
power propagation and constraint propagation along fiber links as shown in Exam-
ple 6.12 in Chap. 6. The approach presented in Sect. 7.3 of Chap. 7 cannot easily be
extended to such configurations. One reason is the difficulty in building an appro-
priate cost function that needs a complicated penalty pricing function term for the
constraints. On the other hand the Lagrangian extension and decomposition results
as developed in Sect. 7.4 of Chap. 7 provide a good path to follow. One thing to keep
in mind is that the approach can be applied under convex constraints. Since convex-
ity may not be preserved across the network (see Example 6.12), this is unlike ca-
pacity constraints in flow control [21, 68]. This non-convexity introduces additional
difficulties in directly applying the approach to network topologies from an end-to-
end point of view. A possible approach is a partitioned game formulation and this
is the approach we present. The main idea is that a general multi-link structure can
be partitioned into stages whereby each stage is a γ -link or a single-sink structure,
for which convexity is satisfied. A partitioned Nash game can be formulated for a
general multi-link topology as being composed of o series of ladder-nested stage
Nash games. Stage Nash games are played sequentially and solutions are intercon-
nected. Based on Lagrangian extension and decomposition results, a hierarchical
iterative algorithm towards computing a solution of the partitioned Nash game can
be developed.
The remainder of this chapter is organized as follows. As a building block for the
partitioned game approach, in Sect. 8.2 focuses on games in a γ -link topology. This
is a simple network topology where power is adjustable not only at the transmitter
sites but at all nodes (see Sect. 6.4.2). For simplicity we consider games with no cou-
pled constraints in a γ -link so that the partitioned game approach is shown better.
Then in Sect. 8.4 a partitioned game is presented for the case with constraints and
general mesh and multi-link topologies. This follows the approach in Sect. 8.2 for
the partitioned game where each stage game is a link game. Coupled capacity con-
straints are treated as in Sect. 7.4 for each stage game. Alternatively in Sect. 8.3 we
show that for single-sink topologies convexity is maintained and thus approach in
Sect. 7.4 can be directly applied to an end-to-end game. The material in this chapter
is mostly based on [106, 112].
As a stepping stone for solving games with coupled constraints in general network
topologies in this section we consider a relatively simple topology of interconnec-
tion of L point-to-point links linked via network elements such as OXC or dynamic
gain equalizers [126]. Thus parameters are adjustable at all nodes that have these
elements (see Sect. 6.4.2). This allows us to develop a partitioned game approach
into stages that will be used later on in multi-link and mesh topologies. For sim-
plicity we treat only the case with no coupled constraints. We will consider coupled
constraints from the following sections onwards (Sects. 8.3 and 8.4).
In this section we study γ -link topologies with a set of links L = {1, . . . , L} where
channel powers are adjustable not only at Txs but also at some K bifurcation nodes
via the adjustment factors γ (see Fig. 6.11, Sect. 6.4.2 and Fig. 8.1 below).
8.2 Games in γ -Link Topologies 123
with the vector ul,−i obtained by deleting the ith element from ul . Let γl = [γi,l ]
denote the vector of channel adjustments at stage l such that
where stage l
is the stage precedent to stage l for channel i, i.e., l
= l − 1 (Fig. 8.2).
The adjustment parameter γi,l is bounded within [γmin , γmax ]. Each stage is com-
posed of a link. Thus the OSNR for channel i, i ∈ M, at the output of stage l, l ∈ L,
is defined as
pi,l
yi,l = out .
ni,l
Based on this partition, the OSNR of channel i at the output of stage (link) l, can be
obtained from Proposition 6.7 as
ui,l
yi,l = , (8.1)
nin
i,l + j ∈M Γli,j uj,l
where Γl = [Γli,j ] is the link (stage) system matrix. Based on (8.1), we define
1 1 1 uj,l
:= − = Γli,j (8.2)
δQi,l yi,l yi,l
ui,l
j ∈M
where stage l
is the link precedent to stage l for channel i. In this case of γ -
link topologies this is simply l
= l − 1. It follows that δQ1i,l measures the OSNR
degradation of channel i from stage l
to stage l. Thus, instead of maximization of
OSNR from Tx to Rx, one can consider minimization of OSNR degradation be-
tween stages.
We formulate a partitioned Nash game, composed of lower-level ladder-nested
stage Nash games. On stage l, a Nash game is played with each channel attempting
to minimize its individual cost over m channels. Games on stages are played in
sequence and solutions are interconnected as will be shown below.
We first specify channel individual cost function Ji,l in the lower-level stage Nash
game. Channel individual cost function Ji,l has a similar form as Ji (7.3) (Chap. 7),
defined as a difference between a pricing function αi,l ui,l and a utility function.
8.2 Games in γ -Link Topologies 125
Unlike the games in Chap. 7, herein the utility function reflects the associated
channel’s OSNR degradation. Based on the OSNR model (8.1) and the new measure
variable δQ1i,l (8.2) for stage l, let us define the cost function Ji,l as
That is, the utility function Ui,l reflects a preference for lower OSNR degradation of
channel i on stage l. Substituting (8.2) into (8.3) yields
ui,l
Ji,l (ul,−i , ui,l ) = αi,l ui,l − βi,l ln 1 + ai,l , (8.5)
X̃l,−i
where
X̃l,−i = Γli,j uj,l (8.6)
j ∈M,j =i
It follows that Ji,l (ul,−i , ui,l ) is continuously differentiable in its arguments and
convex in ui,l . Let us denote the lower-level stage Nash game by G(M, Ωi , Ji,l ).
Recall that the overall action space Ω is convex, then from Proposition 4.5
G(M, Ωi , Ji,l ) admits an NE solution.
We exploit the partitioned Nash game composed of L lower-level stage Nash
games, given by G(M, Ωi , Ji,l ). Each G(M, Ωi , Ji,l ) is played by m channels such
that δQ1i,l is minimized for each channel i on stage l. The lower-level stage Nash
games are played sequentially
(in a precedence order) with the interpretation that
across all L stages, l∈R i Ji,l is related to the overall OSNR degradation for chan-
nel i. Solutions of all G(M, Ωi , Ji,l ) are interconnected, explained as follows.
Let u∗l = [u∗i,l ] be an NE solution of G(M, Ωi , Ji,l ). Recall that channel powers
are adjustable at bifurcation nodes and γi,l is the adjustable parameter for channel i
on stage l. Given the precedent actions of channel i, i.e., u∗i,l
and p∗i,l
, the adjust-
ment for channel i on stage l is obtained as
∗
u∗i,l
γi,l =
p∗i,l
u∗l = diag γl∗ · p∗−l ,
where p−l consists of the corresponding channel signal powers from the output of
other stages for all i ∈ M. Recall that each adjustment parameter is bounded within
126 8 Games in Network Topologies
∗ ∈ [γ
[γmin , γmax ]. The partitioned Nash game admits a solution if each γi,l min , γmax ].
∗
Alternatively, γl can be written as
γl∗ = Fl u∗l , p∗−l ,
where Fl is a one-to-one mapping. The reader is referred to [117, 121] for a detailed
proof (Theorem 1).
Remark 8.1 Note that using a system-like cost interpretation the overall L×m game
between γ -links and channels has a cost function
L
m
Jt = Ji,l (8.7)
l=1 i=1
This could be interpreted as the sum of net utilities over all γ -links and over chan-
nels. After changing the summation order this gives
m
L
Jt = J˜i , with J˜i = Ji,l (8.8)
i=1 l=1
Thus Jt can be interpreted as the sum of net utilities over channels, each channel
with cost J˜i
L
ai,l
J˜i = αi,l ui,l − βi,l ln 1 +
l=1
1
δQi,l − Γli,i
An interesting parallel can be made between J˜i and the cost Ji (7.3) (Chap. 7), used
in the end-to-end game. This relation is more evident if one considers the case of
identical γ -links. In this case the parameters αi,l , βi,l can be taken to be the same
for all γ -links, αi,l = αi , βi,l = βi , ai,l = ai and J˜i can be rewritten as
L
L
ai
J˜i = αi ui,l − βi ln 1 + (8.9)
l=1 l=1
1
δQi,l − Γi,i
Comparing this with Ji (7.3) (Chap. 7), it can be seen that J˜i , (8.9), is similarly
expressed as the difference between a linear pricing term and a logarithmic util-
ity term. In fact J˜i , (8.9), generalizes Ji since it captures in pricing and utility the
contribution of each γ -span from Tx to Rx.
Remark 8.2 We started the formulation from a game with action space in R mL that
was further decomposed. This formulation has a computational advantage when
compared to an overall link m-player game with action space in R L , and a possi-
ble cost J˜i . For the latter game formulation one cannot find an explicit, analytically
tractable NE solution, due to the coupling between channel powers at one γ -link and
8.2 Games in γ -Link Topologies 127
all previous spans. Alternatively, the game formulation thus developed corresponds
to a ladder-nested structure: in the game between the γ -spans the lth player decision
is taken after the (l − 1)th player’s action. The overall NE solution is still coupled
but it has decoupled necessary existence conditions. Moreover, the specific triangu-
lar structure of these conditions enables a recursive computation of the unique NE
solution. For more details the reader is referred to [121], where a similar partitioned
Nash game formulation was proposed for the number of stages K < L. As in [121]
herein the number of channels at each γ -link (stage) was identical for simplicity.
In the following two sections this formulation is extended to multi-link and mesh
topologies which are more complex. Firstly, at each stage, the number of channels at
different stages (links) can be variable and, secondly, more importantly the coupled
link capacity constraint has to be considered.
Theorem 8.3 If for all i ∈ M, ai,l are selected such that (8.11)
Γli,j < ai,l , ∀i ∈ M (8.11)
j =i
Proof By Theorem 7.1 (point-to-point link) applied to link l, if on stage (link) l, ai,l
is selected satisfying (8.11), then an inner unique NE solution u∗l exists given as in
(7.9). Component-wise, (7.9) is written for i ∈ M as
βi,l 1
u∗i,l = − Γli,j u∗j,l (8.12)
αi,l ai,l
j ∈M,j =i
The rest of the proof follows directly as the proof of Lemma 4 in [121], stated here
for completeness. Let
ei,l (n) := ui,l (n) − u∗i,l (μl )
128 8 Games in Network Topologies
The corresponding vector form is el (n) = [. . . , ei,l (n), . . .]T . Let us also define
) )
)el (n)) := maxei,l (n)
∞ i∈M
1
ei,l (n + 1) = − Γli,j ej,l (n)
ai,l
j ∈M,j =i
Using this it can be shown that under condition (8.11), el (n + 1)∞ ≤ C0 el (n)∞ ,
where 0 ≤ C0 < 1 and el (n)∞ ≤ C0n el (0)∞ , such that the sequence {el (n)}
converges to 0. Therefore channel algorithm (8.10) converges to the inner NE solu-
tion u∗l .
Note that for a given γ -link matrix Γl , the selected ai,l factors can be set such
that the diagonal dominance condition holds. If Γl is itself diagonal dominant, then
a possible choice is ai,l = Γli,i . If all γ -links are identical then the same ai,l = ai
can be used.
In the simple network topology in the previous section, i.e., γ -link topology, we did
not consider coupled constraints. Towards treating the full case of network topolo-
gies with coupled constraints, we now turn to games with coupled constraints but
we restrict the network topology to another simple case, called single-sink topology.
As we will see this particular topology has the feature of maintaining convexity of
the coupled constraints.
In a single-sink topology with a set L = {1, . . . , L} of links and a set M =
{1, . . . , m} of channels, channels are added at different links, but dropped off at
the output of the same link (e.g. Fig. 8.3).
Compared to the point-to-point topology here we can have different number
channels over different links. We order the m channels in the following way: a set
Ma1 = {1, . . . , m1 } of m1 channels are added into the network at the first link, a set
l−1
l
Mal = mj + 1, . . . , mj
j =1 j =1
L
ml = m
l=1
8.3 Games in Single-Sink Topologies 129
Example 8.4 Consider an example shown in Fig. 8.3, where three channels (m = 3)
are transmitted over a single-sink three-link network (L = 3).
Similar to Example 6.12 in Chap. 6, the link capacity constraints gl (u) can be
written as
2
g1 (u) = u1,i − P10 = u1 + u2 − P10 ≤ 0
i=1
4
2
g2 (u) = u2,i − P20 = p1,i + u3 + u4 − P20 = u3 + u4 − P20 − P10 ≤ 0
i=1 i=1
6
4
g3 (u) = u3,i − P30 = p2,i + u5 + u6 − P30 = u5 + u6 − P30 − P20 ≤ 0,
i=1 i=1
where the system constraint (6.29) was used twice. Thus the .coupled constraint set
for link l ∈ L, Ω l = {u ∈ Ω | gl (u) ≤ 0} is convex and Ω = l∈L Ω l is also convex.
This result can be generalized for single-sink multi-link topologies with m chan-
nels, presented in the following lemma.
130 8 Games in Network Topologies
Lemma 8.5 Consider a single-sink topology with a set of links L. For any link
l ∈ L,.the coupled constraint set Ω l is convex. The overall coupled action space
Ω = l∈L Ω l is a convex set.
l−1 l
Proof Let us denote m̄l−1 = j =1 mj and m̄l = j =1 mj . Then
ml
gl (u) = l ≤ 0,
um̄l−1 +j − P (8.13)
j =1
Remark 8.6 From (8.14), one can see that in the inequality constraint for link l,
gl (u) ≤ 0, only channels in the set Mal are coupled, i.e., channels that are added
into the network at link l are coupled. This is advantageous for a hierarchical de-
composition, as will be shown later.
Based on the convexity result in Lemma 8.5, let us consider now an m-player Nash
game in a single-sink topology where there are L coupled constraints. The action
space is
Ω= u ∈ Ω | gl (u) ≤ 0 ,
l∈L
8.3 Games in Single-Sink Topologies 131
As before in the Nash game formulation, each channel i ∈ M minimizes its indi-
vidual cost function Ji : Ω → R,
Ji (u−i , ui ) = αi ui − Ui (u−i , ui ), (8.17)
with a utility Ui indicating preference for better OSNR:
ai ai ui
Ui (u−i , ui ) = βi ln 1 + 1 = ln 1 + 0 ,
yi − Γi,i
ni + j =i Γi,j uj
where αi , βi > 0 are weighting factors, ai > 0 is a channel specific parameter and
Γ = [Γi,j ] is the network system matrix being defined in (6.36).
We denote the game by G(M, Ω i , Ji ), which is in the class of m-player Nash
games with coupled utilities and coupled constraints. It follows from (8.17) that
Ji is continuously differentiable in its arguments and convex in ui . Note that the
overall coupled action space Ω is compact. By Lemma 8.5, Ω is convex as well.
Then from Proposition 5.3, G(M, Ω i , Ji ) admits an NE solution. In the following,
we use Lagrangian extension and decomposition results presented in Chap. 5 to
compute an NE solution.
An NE solution of G(M, Ω i , Ji ) can be computed by using the Lagrangian ex-
tension to the game-theoretic framework. We continue using the notations defined
in Chap. 5 for the separable NG cost function J((u; x), (4.7),
m
J((u; x) = Ji (u−i , xi ),
i=1
the separable augmented constraint (
g (u; x), (5.8),
m
(
g (u; x) = g(u−i , xi ),
i=1
( x; μ), (5.11),
the augmented Lagrangian function L(u;
( x; μ) = J((u; x) + μT (
L(u; g (u; x)
and the dual cost function D(μ), (5.17),
( u∗ ; u∗ ; μ ,
D(μ) = L (8.18)
( x; μ). In a fixed-point
where u∗ is such that u = u∗ satisfies u = arg minx∈Ω L(u;
notation,
D(μ) = min L(u;( x; μ) ,
x∈Ω (
arg minx∈Ω L=u
where (
g (u; u) ≤ 0.
132 8 Games in Network Topologies
m
m
L
D(μ) = Li u∗−i (μ), u∗i (μ), μr(i) + l (8.19)
μl eTml ,−i u∗−i (μ) − P
i=1 i=1 l=1
where r(i) refers to link r where channel i is added, eml ,−i is an (m − 1) vector
obtained by deleting the ith element of the m vector eml , and u∗ (μ) is an NE solution
to G(M, Ωi , Li ) where
This is similar to the point-to-point case (see Chap. 7) but herein each of the L
constraints is accounted for.
Proof Let us order the L links and m channels in the same way as in the proof of
Lemma 8.5. Then the constraint function (8.15) can be written in a two-argument
form
T
em ,−i u−i + xi − P l , l = r(i)
T
gl (u−i , xi ) = eml · u − Pl = T l (8.21)
eml ,−i u−i − Pl , l = r(i)
m
D(μ) = Li u∗−i (μ), u∗i (μ), μ
i=1
8.3 Games in Single-Sink Topologies 133
where
Li (u−i , xi , μ) = Ji (u−i , xi ) + μT g(u−i , xi )
Using (8.21) one finds that, ∀i ∈ M, Li is given as
Li = Ji (u−i , xi ) + μr(i) gr(i) (u−i , xi ) + μl gl (u−i , xi )
l =r(i)
= Ji (u−i , xi ) + μr(i) xi + l
μl eTml ,−i u−i − P (8.22)
l
The following arguments are similar to those in Corollary 1 in [120], outlined here.
Let us minimize RHS of (8.22) with respect to xi and then solve for a fixed-point
solution xi = ui . Since only the first two terms on RHS of (8.22) depend on xi ,
isolating the terms without xi yields
m
D(μ) = min Li (u−i , xi , μr(i) )
xi ∈Ωi arg minxi ∈Ωi Li =ui
i=1
m
L
+ l ,
μl eTml ,−i u−i − P (8.23)
i=1 l=1
xi∗ = xi (u−i )
u∗ (μ) = Γ(−1 · (
b(μ),
The proof follows similar to the proof of Proposition 7.9 for the point-to-point
topology except that it involves the network matrix and μr(i) .
Based on this single-sink case and the method in Sect. 8.2, a partitioning method
for general multi-link topologies was proposed in [101, 112]. Such a topology is
partitioned into stages where each stage has a single-sink structure and algorithms
developed based on the extragradient method [70, 71]. However, this partition does
not fully exploit the power adjustment flexibility, is not completely scalable and
cannot be easily extended to treat mesh topologies. In the next section we present
an approach that addresses this most general case, of how to treat mesh topologies.
After these preliminary simple network topologies (γ -link in Sect. 8.2 and single-
sink topologies in Sect. 8.3), we now consider games with coupled constraints in
general mesh and multi-link network topologies. A possible case with three channel
optical paths is shown in Fig. 8.4, where each channel is represented by a different
color (wavelength).
8.4 Games with Coupled Constraints: Mesh Topologies 135
uses the flexibility of power adjustment at each node. Thus the network is parti-
tioned as in Sect. 8.2 into stages composed only of single optical links. By breaking
the loop and selecting one stage as the start, stages can be placed sequentially in a
ladder-nested form. One benefit of such a partition is that the convexity of coupled
constraints propagated along links on each stage is automatically satisfied.
In the partitioned Nash game, each link Nash game is played towards minimizing
channel OSNR degradation. Based on such a partition, the hierarchical decomposi-
tion is applicable to each link game, leading to a lower-level uncoupled game for
channels on each link and a higher-level problem for link pricing. Computation of
equilibria is based on a three-level hierarchical algorithm. Such a partition simplifies
the structure of each stage, makes it regular and scalable, and benefits the develop-
ment of a link pricing algorithm. This approach can be thus used for general mesh
topologies.
We can apply Lemma 6.8 and use the recursive OSNR model for the network.
Consider link l in Fig. 6.10. The OSNR of channel i at the output of link l is given
as
1 1 uj,l
= + Aj,l Γli,j , i ∈ Ml , (8.26)
yi,l yi,l
ui,l
j ∈M
where link l is the link precedent to link l for channel i. Using (8.26), let us define
1 1 1 uj,l
= − = Aj,l Γli,j (8.27)
δQi,l yi,l yi,l
ui,l
j ∈M
It follows that δQ1i,l measures the OSNR degradation of channel i from link l
to
link l. Thus instead of maximization of yi from Tx to Rx, consider minimization of
each δQ1i,l between links, i.e., minimization of individual OSNR degradation. In the
next section, by using the similar methodology as in Sect. 8.2, a partitioned Nash
game composed of stage Nash games is presented. Each stage is a link and the stage
Nash game is formulated to minimize the OSNR degradation, and moreover the
coupled constraints are considered.
8.4 Games with Coupled Constraints: Mesh Topologies 137
In this section we discuss a partitioned game framework with link capacity con-
straints in mesh topologies. As mentioned above the network is partitioned into
stages composed of single links. By partitioning, each stage game is defined for a
link, thus a link game, as in Sect. 8.2. Each link Nash game is played towards min-
imizing channel OSNR degradation. Then a partitioned Nash game composed of
link games is defined. The set of OSNR degradation minimizations on stages/links
is related to OSNR maximization from Tx to Rx (see Remark 8.1 in Sect. 8.2). This
setup is described below.
On each link l ∈ L, the set of channels is Ml = {1, . . . , ml }. Let us consider the
link capacity constraint on link l:
gl (ul ) = l ≤ 0, ∀l ∈ L,
Ai,l ui,l − P (8.28)
i∈M
where Ω denotes the Cartesian product of Ωi . The action set of individual channel
i ∈ Ml is defined as the projection of Ω l on channel i’s direction, namely,
Ωi,l (ul,−i ) = ξ ∈ Ωi | gl (ul,−i , ξ ) ≤ 0 ,
where ul,−i is obtained by deleting ui,l from vector ul . It can be seen that both the
action space Ω l and the individual action set Ω i,l (ul,−i ) are compact and convex.
The link Nash game is played with each channel attempting to minimize its indi-
vidual cost with respect to its OSNR degradation. We specify a channel cost func-
tion Ji,l for channel i on each link l. Similar to the channel cost function defined in
Sect. 8.2, Ji,l is defined as a difference between a pricing function Ai,l αi,l ui,l and
a utility function Ui,l which reflects the associated channel’s OSNR degradation,
namely,
Ji,l = Ai,l αi,l ui,l − Ui,l , ∀i ∈ M, (8.30)
with
Ai,l ai,l
Ui,l = βi,l ln 1 + , (8.31)
1
δQi,l − Γli,i
where βi,l > 0 indicates the strength of the channel’s desire to minimize its OSNR
degradation and ai,l > 0 is for scalability. Substituting (8.27) into (8.30) yields
Ai,l ui,l
Ji,l = Ai,l αi,l ui,l − βi,l ln 1 + ai,l , ∀i ∈ M, (8.32)
Xl,−i
138 8 Games in Network Topologies
where Xl,−i = j =i,j ∈M Aj,l Γli,j uj,l . It follows that Ji,l is continuously differen-
tiable in its arguments and convex in ui,l .
We denote such a link Nash game by G(M, Ω i,l , Ji,l ). Note that the individ-
ual cost function Ji,l is generally defined for each channel i ∈ M. If channel i is
not transmitted on link l, i.e., i ∈ / Ml , then Ai,l = 0 and Ji,l = 0, which means
that the decision of channel i, i ∈ / Ml , does not affect the decisions made by
other channels j ∈ Ml . Thus G(M, Ω i,l , Ji,l ) is equivalently to a reduced Nash
game played among ml channels. Furthermore, the existence of an NE solution of
G(M, Ω i,l , Ji,l ) is guaranteed by Proposition 5.3.
We exploit the partitioned Nash game. By partitioning and selecting one stage
as the start,
stages can be sorted sequentially with the interpretation that across all
stages, l∈Ri Ji,l is related to the overall OSNR degradation for channel i. So-
lutions of all G(M, Ω i,l , Ji,l ), l ∈ L are interconnected similar to Sect. 8.2. The
explanation is given as follows.
Recall that channel powers are adjustable at optical switches and γi,l is the ad-
justable parameter for channel i on stage l. The vector form is γl = [γi,l ]. Let
u∗l = [u∗i,l ] be an NE solution of G(M, Ω i,l , Ji,l ). The corresponding signal power
vector at the output of link l is p∗l = [pi,l ∗ ] and the corresponding augmented output
power vector defined in Lemma 6.10 is p∗ . Note that for those channel i ∈ / Ml , val-
ues of u∗i,l and pi,l ∗ are randomly set. By using (6.37) in Lemma 6.10, optimal γ
i,l
can be obtained by solving the corresponding component-wise equation in
T ∗
u∗l = diag γl∗ · diag Vi,l ·p .
∗ ∈ [γ
Finally, the partitioned Nash game admits a solution if each γi,l min , γmax ].
Compared to Sect. 8.2 the interconnection matrix diag(Vi,l ) for the full network
topology appears as well as the coupled constraints.
where Ji,l is defined in (8.32) with Ai,l = 1. The two-argument constraints are given
as
gl (ūl ; x̄l ) =
( gl (ūl,−i , x̄i,l )
i∈Ml
where the scalar μl is the Lagrange multiplier. The associated dual cost function
Dl (μl ) is defined as
(l ū∗l ; ū∗l ; μl ,
Dl (μl ) = L (8.34)
where ū∗l is such that ūl = ū∗l satisfies
( ūl ; x̄l ; μl )
ūl = arg min L(
x̄l ∈Ω
where Ji,l is defined in (8.32) and an uncoupled action set Ωi . We denote this game
by G(Ml , Ωi , Li,l ). Notice that Li,l (8.35) has a similar form with Li (7.51) in
Chap. 7. By using Proposition 7.9 we obtain the following result characterizing the
NE solution of G(Ml , Ωi , Li,l ).
Corollary 8.9 For each given μl ≥ 0, G(Ml , Ωi , Li,l ) admits an NE solution if ai,l
in Li,l (8.35) satisfies
Γli,j < ai,l , ∀i ∈ Ml (8.36)
j =i,j ∈Ml
Based on these, an iterative hierarchical algorithm can be developed for both link
pricing and channel power adjustment.
8.4 Games with Coupled Constraints: Mesh Topologies 141
where link l
is the precedent of link l. Note that both yi,l
and μl (k̄) are invariable
during the channel iteration on link l. Each channel i on link l updates its signal
power ui,l based on the feedback information, i.e., its OSNR at the output of link l,
yi,l , and fixed parameters, μl (t) and yi,l
. Note that compared to (8.10) in Sect. 8.2
the term μl appears due to coupled constraint, while compared to (7.62) in Chap. 7)
the term yi,l
appears due to partitioning in link stages and interconnection between
link l
and link l. Then applying Theorem 8.3 in Sect. 8.2 one can obtain the follow-
ing convergence result.
Corollary 8.10 If for all i ∈ Ml , ai,l in Li,l (8.35) are selected such that (8.36)
is satisfied. Then for each given μl ≥ 0, channel algorithm (8.39) converges to the
inner NE solution ū∗l (μl ).
Link Algorithm The link algorithm is a gradient projection algorithm [24, 81],
developed based on (8.38). On each link l, after every K iterations of the channel al-
gorithm (8.39), the new link price μl is generated at each iteration time k̄, according
to the following link algorithm:
+
μl (k̄ + 1) = μl (k̄) + η l
ui,l μl (k̄) − P (8.40)
i∈Ml
where η > 0 is a step-size and [z]+ = max{z, 0}. Practically, K is sufficiently large
such that each channel power converges to its solution. Thus given the total power
of all channels on the link l, the link algorithm (8.40) is completely distributed and
can be implemented by individual links using only local information. A proof for
the convergence of this link algorithm is provided in [111].
The link algorithm above is a simple projection algorithm. Theorem 1 in [111]
provides an explicit upper-bound for the step-size η related to the NE solution. It
can be mentioned that an extragradient method can been used on each stage for
the link pricing [112]. The extragradient method is a modified projection algorithm,
and sufficient conditions for its convergence depend on a Lipschitz constant which
is intractable.
Remark 8.11 For any given μl ≥ 0, Theorem 8.3 shows that the channel algorithm
converges to the NE solution which is unique in the sense of inner-ness. Though
uniqueness of a solution is not guaranteed, μ∗l , (u∗l , μ∗l ) is an optimal NE solution–
Lagrange multiplier pair and the algorithm converges to one such pair.
142 8 Games in Network Topologies
8.5 Notes
For mesh and multi-link network topologies we treated several cases in increasing
order of complexity. First we treated simple γ -link topologies and we formulated
a partitioned Nash game composed of link Nash games, where for simplicity no
coupled constraints were considered. In the partitioned Nash game, each link Nash
game is played towards minimizing channel OSNR degradation. Based on such a
partition, the hierarchical decomposition is applicable to each link Nash game.
Next we considered games with coupled utility and constraints in multi-link
topologies, extending the approach in Chap. 7 where only single point-to-point fiber
links were treated. In multi-link topologies constraints propagate along link and
convexity may not be preserved. The general multi-link topology was dealt with by
formulating a partitioned Nash game composed of stage Nash games, where each
stage has a single-link structure, i.e., a γ -link structure. Connections between stages
are realized by exploiting the flexibility of extra adjustments at bifurcation points.
By selecting a starting link, the game on each link can be played sequentially and
the hierarchical decomposition leads to a lower-level Nash game for channels with
no coupled constraints and a higher-level problem for link pricing. Computation of
equilibria is based on a hierarchical algorithm. An alternative partitioned game could
be based on single-sink structure (which has convexity ensured), but the single-link
partition has a more regular and scalable structure. A possible implementation is
described later on in Chap. 10.
In Chaps. 7 and 8 game-theoretic approaches have been used to solve channel
OSNR optimization with link capacity constraints in optical networks. In particu-
lar, while we formulated a game-theoretic framework towards optimizing channel
OSNR, however, we have not taken desired channel OSNR targets into account yet.
In a Nash game in the presence of player selfishness, Nash equilibria may not op-
timize overall system performance. These efficiency issues as well as robustness of
network control algorithms derived from game-theoretic formulations in the pres-
ence of time-delay will be addressed in the following two chapters.
Chapter 9
Nash Equilibria Efficiency and Numerical
Studies
9.1 Introduction
In previous chapters we showed how a game-theoretic framework can be formulated
towards optimizing channel OSNR, but we have not taken desired channel OSNR
targets into account yet. In a Nash game in the presence of player selfishness, Nash
equilibria may not optimize overall system performance. These efficiency issues
will be addressed next, focusing on a single point-to-point link game. The material
in this chapter is mostly from [102–104]. A combination approach can be found
in [168].
We first consider a system optimization framework towards achieving a target
OSNR level for each channel, while minimizing the interference and hence im-
proving the overall performance. Link capacity constraint is also considered. This
framework is used to investigate the effects of parameters in a game-theoretic ap-
proach.
Efficiency of Nash equilibria has been studied extensively [1, 38, 64, 122, 129,
134]. The efficiency loss as compared to an socially optimal system performance
is quantified by a ratio called “Price of Anarchy” [73]. Work on how to quantify
the efficiency loss under separable and under non-separable costs can be found in
[64] and [122], respectively. Results suggest that, provided a pricing mechanism is
chosen properly, the selfish behavior of players in a Nash game may not degrade
the system performance arbitrarily [64, 134]. Such approaches will be reviewed in
Chap. 12 for routing and path coloring via game approaches in optical networks. In
this chapter, instead of looking from the view point of degree of efficiency (Price
of Anarchy), we study the efficiency of an OSNR game by investigating the effects
of parameters in individual game cost functions. In particular, we use the system
where P 0 is the link total power target, and the target OSNR constraint
OSNRi ≥
γi , ∀i ∈ M, (9.1)
where γi is the target OSNR for channel i. As before the signal optical power at Tx
ui is bounded in Ωi = [0, umax ] with constant umax > P 0 , for all i ∈ M.
Consider an alternative OSNR optimization setup from the perspective of system
performance, as opposed to a game-theoretic approach as developed in previous
chapters. Specifically consider the following cost:
min C(u)
subject to ui ∈ Ωi , ∀i ∈ M,
i ≥
OSNR γi , ∀i ∈ M,
uj ≤ P 0 ,
j ∈M
where C(u) is the system cost function, defined as the sum of all individual costs,
Ci (ui ),
C(u) = Ci (ui ),
i∈M
hence it is separable. Each individual cost function Ci (ui ) is a generic cost function
that satisfies the following assumption:
This cost function can be defined similarly to the form of (7.3), Ci = Pi −Ui . The
pricing function Pi is a linear function of ui . The utility function Ui is a logarithmic
function of ui , which quantifies approximately the link’s demand or the channel’s
willingness to pay for a certain level of OSNRi based on the relationship between
OSNRi and ui . This relationship is illustrated approximately in Fig. 9.1.
By using the OSNR model (6.16), (9.1) can be rewritten as
ui
≥
γi ,
n0i + j ∈M Γi,j uj
or
ui + γi Γi,j )uj ≥ n0i
(− γi ,
j ∈M
or in vector form
T u ≥ b, (9.3)
where
⎡ ⎤ ⎡ ⎤
1−
γ1 Γ1,1 −
γ1 Γ1,2 ··· −
γ1 Γ1,m n01
γ1
⎢ − ⎥ ⎢ 0 ⎥
⎢ γ2 Γ2,1 1−
γ2 Γ2,2 ··· −
γ2 Γ2,m ⎥ ⎢ n2 γ2 ⎥
T =⎢ .. .. .. .. ⎥, b=⎢
⎢ .. ⎥
⎥
⎣ . . . . ⎦ ⎣ . ⎦
−
γm Γm,1 −
γm Γm,2 ··· 1−
γm Γm,m n0m
γm
All T ’s off-diagonal entries −
γi Γi,j are less than zero. By Definition A.9, T is a
Z-matrix.
146 9 Nash Equilibria Efficiency and Numerical Studies
where 1 is the m × 1 all ones vector. Therefore, the constraint set of the system
optimization problem is Ω := {u ∈ Rm | T u ≥ b, 1T u ≤ P 0 and u ≥ 0}. This con-
strained system optimization problem can be formulated as
min C(u)
(9.4)
subject to u ∈ Ω
and denote this problem by OPT(Ω, C). Condition (9.2) in Assumption 9.1 ensures
that the solution to OPT(Ω, C) does not hit ui = 0, ∀i ∈ M. Let
γ = [
γ1 , . . . ,
γm ]T .
The following result characterizes the unique solution of OPT(Ω, C).
1T · T((
γ ) · b(
γ ) ≤ P 0, (9.6)
γ ) = [n01
with b( γm ]T and T((
γ1 , . . . , n0m γ ) = T −1 (
γ ), then the constraint set Ω is
non-empty and OPT(Ω, C) has a unique positive solution uopt .
Proof We first show that the constraint set Ω is non-empty. Note that in the link
OSNR model, the system matrix Γ (6.17) is a positive matrix, so if (9.5) is satisfied,
we have
1− γi Γi,i >
γi Γi,j > 0, ∀i ∈ M,
j ∈M,j =i
or equivalently,
1−
γi Γi,i > |−
γi Γi,j |, ∀i ∈ M,
j ∈M,j =i
which implies that the Z-matrix T has positive diagonal entries and by Definition
A.6, T is strictly diagonally dominant. According to Theorem A.7, each eigenvalue
of T has a positive real part. Then it follows from Theorem A.10 that T is an M-
matrix. So it has the following properties: T u ≥ b > 0 implies u ≥ 0, and T −1 is
non-negative. Thus
u ≥ T −1 b := T(b (9.7)
and then we have 1T u ≥ 1T · T( · b. Note that both T( and b depend on
γ , i.e.,
T( = T((
γ ) and b = b(
γ ). So
1T · u ≥ 1T · T((
γ ) · b(
γ)
9.2 A System Optimization Formulation 147
u = T(b. Then T (
Let ( u = 1T · T(b. By (9.6), 1T · (
u = b. Also we have 1T · ( u ≤ P 0 . It
follows that
(u ∈ u ∈ Rm | T u ≥ b, 1T u ≤ P 0
Thus the above set is non-empty if both (9.5) and (9.6) are satisfied. Since T u ≥
b > 0 implies u ≥ 0, we have proved that if γ is selected such that (9.5) and (9.6)
are satisfied, the constraint set Ω is non-empty.
Moreover, the constraint set Ω is convex and we have 0 ≤ ui ≤ P0 , ∀i ∈ M.
So Ω is bounded. In addition, it is also closed since it consists of the intersection of
half-spaces. Thus this system optimization problem is a strictly convex optimization
problem on a convex compact constraint set [30], which, following Proposition A.2,
always admits a unique globe minimum, uopt .
Example 9.2 To illustrate the constraint set Ω and the conditions (9.5) and (9.6)
consider a simple example in Fig. 9.2 where m = 2. From (9.7), we have
u1 row1 (T() · b
≥
u2 row2 (T() · b
Remark 9.3 Recall that n0 denotes the input noise power at Tx and may include
external noise, such as thermal noise. If the input noise is neglected, n0 includes
only external noise, which is negligible [45]. So b = [n01 γ1 , . . . , n0m
γm ]T ≈ 0 and
P 0 ≥ 1T · T( · b ≈ 0. This means that the constraint set is non-empty under the first
148 9 Nash Equilibria Efficiency and Numerical Studies
Remark 9.4 Let us take a close look at the second condition (9.6). Recall that
T = I − diag(γ )Γ , where I is an identity matrix. We know from the proof of The-
orem 9.1 that T is an M-matrix. By Theorem A.11, ρ(diag( γ )Γ ) < 1 and
∞
−1
k k
T −1 = I − diag(
γ )Γ = diag
γ Γ
k=0
If
γi increases (given
γj , j = i), LHS of (9.8) will increase. We can find a maximum
OSNR target γmax by solving the following equation:
γmax Γ )−1 · n0 = P 0
γmax · 1T · (I − (9.9)
Based on the link OSNR model, we know that the performance for each channel
is interference limited. In addition, (9.9) shows that the OSNR targets significantly
affect the capacity of a link: each link decides the OSNR threshold γmax by us-
ing (9.9). Any new channel with a required OSNR target no more than γmax will be
admitted to transfer over the link. This idea can be used for links to develop channel
admission control schemes. Similar algorithms exist for wireless networks [65, 66].
min C(u)
(9.10)
subject to Tu ≥
b,
9.3 Distributed Algorithm 149
where
T b
T = and
b=
−1T −P 0
Now consider a barrier function λi : R → R with the following properties:
(P.1) ∀i ∈ M, λi (x) is non-increasing, continuous and
yi (u)
lim λi (x) dx → −∞ (9.11)
ui →∞
bi
where
yi (u) := rowi (T)u. (9.12)
(P.2) λi (x) attains the value 0 if x >
bi , where
bi is defined in (9.10).
Such an example is shown in Fig. 9.3.
With this barrier function λi , which has properties (P.1) and (P.2), construct a
function
yi (u)
Vp (u) = Ci (ui ) − λi (x) dx (9.13)
i∈M i∈M bi
Based on (P.1), i.e., the non-increasing property of the barrier function, and from
Assumption 9.1, i.e., that Ci (ui ) is strictly convex, it follows that Vp (u) is strictly
convex and hence has a unique internal minimum value such that
∂Vp (u)
= Ci
(ui ) − rowi TT λ y(u) = 0, ∀i ∈ M
∂ui
150 9 Nash Equilibria Efficiency and Numerical Studies
Thus solving the set of above equations we obtain the unique solution ūopt of (9.14):
ūi = Ci
−1 rowi TT λ y ūopt , ∀i ∈ M
opt
The barrier function λi (·) can be selected such that the unique solution of (9.14) may
arbitrarily closely approximate the optimal solution of OPT(Ω, C). For example, a
barrier function can be defined as [68]
[
bi − x + ε]+
λi (x) = ,
ε2
where [x]+ = max{x, 0}.
For the relaxed system problem a primal algorithm can be defined as
∂Vp (u)
u̇i (t) = gi (ui , si ) = −ki = −ki Ci
ui (t) − si (t) , ∀i ∈ M, (9.15)
∂ui
where the coefficient ki > 0 and si (t) is defined as
si (t) = rowi TT λ y u(t) , (9.16)
and λ(·) = [λ1 (·), . . . , λm+1 (·)]T , with each λi (·) satisfying (P.1) and (P.2).
Algorithm (9.15) is a gradient algorithm and can be implemented in a distributed
way. Each channel varies its input power ui gradually as in (9.15), while the link
(network system) calculates the vector s(t) = [s1 (t), . . . , sm (t)]T based on the re-
ceived input powers, OSNR preference and link constraint, and then feeds this up-
dated information back to each channel. This algorithm is represented in Fig. 9.4.
The following theorem states that the unique equilibrium of the algorithm (9.15)
corresponds to the unique solution of (9.14), ūopt . Moreover, the solution is globally
asymptotically stable.
Theorem 9.5 (See [102]) The unique solution ūopt to the relaxed system optimiza-
tion problem (9.14) is globally asymptotically stable for the system (9.15).
∂V (u)
Proof Since ūopt is the unique solution to the equations ∂u p
i
= 0, ∀i ∈ M, it
follows that it is the unique equilibrium point of the system (9.15). Next, from Vp (u)
9.4 Numerical Study of Efficiency in Nash Game 151
being strictly convex it follows that ūopt is the global minimum point of the function
Vp (u). Let C = Vp (ūopt ). Then Vp (u) > C for all u = ūopt , and we construct a
Lyapunov function for the system (9.15):
V (u) = Vp (u) − C
It follows that V (u) = 0 when u = ūopt , and that V (u) > 0 when u = ūopt . That is,
the function V (u) is positive definite with respect to the equilibrium point u = ūopt .
Taking the derivative of V (u) along the trajectory of the system gives
∂
∂
2
V̇ (u) = Vp (u) · u̇i = ki Vp (u)
∂ui ∂ui
i∈M i∈M
Since V̇ (u) = 0 when u = ūopt , and V̇ (u) < 0 when u = ūopt , it follows that V̇ (u)
is negative definite with respect to the equilibrium point u = ūopt . The conclusion
follows by Lyapunov stability theory.
Note that the unique solution of the relaxed system problem (9.14) may arbitrar-
ily closely approximate the optimal solution of the original system problem (9.10)
with an approximate selection of the barrier function. For details the reader s re-
ferred to [101].
Note that while the individual costs are convex in each player’s action ui , in gen-
eral the aggregate game cost function is not necessarily convex in u. The following
simple example shows this. For simplicity we omit the penalty term and noise in the
OSNR model.
152 9 Nash Equilibria Efficiency and Numerical Studies
Example 9.6 Consider a Nash game with three players (m = 3) with individual
costs,
ui
Ji (u) = ui − ln , i = 1, 2, 3
j =i uj
It follows that
∂ 2 J1 u2 + u3
= >0
∂u21 u1
∂ 2 J2 1
=− <0
∂u1 2 (u1 + u3 )2
∂ 2 J3 1
=− <0
∂u1 2 (u1 + u2 )2
Therefore,
∂ 2J u2 + u3 1 1
= − −
∂u12 u1 (u1 + u3 )2 (u1 + u2 )2
∂2J
Note that has not a definite sign, so that J (u) is not always convex with respect
∂u21
to u1 , even though J1 (u) is strictly convex with respect to u1 .
Ci (ui ) = αi ui − βi ln ui , (9.17)
where αi > 0 and βi > 0. This cost function Ci (ui ) is obviously strictly convex and
continuously differentiable. Moreover Ci (ui ) → +∞ as ui → 0. Comparing to Ji
in the above example it can be seen that Ci (ui ) in OPT(Ω, C) has an interpretation
similar to the one of the cost function Ji (u) in G(M, Ω̂i , Ji ), and moreover Ci is
uncoupled in u for a given set of other power u−i . Based on the relation between
these two formulations, i.e., OPT(Ω, C) and G(M, Ω̂i , Ji ), we can use the central
cost function in OPT(Ω, C) as a welfare function of G(M, Ω̂i , Ji ).
This is shown next, based on the system optimization being used to measure
the efficiency of the NE solution numerically. We compare by simulation the two
approaches: system optimization approach and the game-theoretic approach, re-
spectively. Consider a link with six channels (m = 6) and total power target is
9.4 Numerical Study of Efficiency in Nash Game 153
Within the set of six channels, there are two levels of OSNR target, a 26 dB
level desired on the first three channels and a 22 dB OSNR level on the next
three channels. Conditions (9.5) and (9.6) on the target OSNR are satisfied. So
the feasible constraint set is non-empty. The coefficients in (9.17) are selected as
αi = 1, i = 1, . . . , 6, and β = [0.5, 0.51, 0.52, 0.3, 0.31, 0.32]. Recalling the rela-
tionship between channel OSNR and channel input power shown in Fig. 9.1, the
values of βi implicitly indicate channel OSNR preferences. The coefficient ki is
fixed for each channel with ki = 0.01, i = 1, . . . , 6 and we select a barrier function
6
λi (xi ) = 1000 max{0, b̂i − xi } , (9.18)
where xi (u) = rowi (T)u. Notice that λi (ui ) is zero when the constraints are satis-
fied. So there is a penalty with any violation of the constraints.
Let the NE solution of G(M, Ω̂i , Ji ) be denoted by u∗ . Algorithm GA 7.36 (see
Chap. 7) is used, i.e.,
1 βi ai /ui (t)
u̇i (t) = −μ αi + 0 − ,
(P − j ∈M uj (t))2 ( OSNR 1
(t) + a i − Γi,i )
i
Remark 9.7 Theorem 7.7 states that GA converges to the NE solution if (7.40) and
(7.41) are satisfied, where umin > 0 is a positive lower bound on each ui . A lower
bound on βi is set by (7.41). Since each channel attempts to select larger βi for
the purpose of higher OSNR, the lower bound does not affect the results of the
efficiency study.
In OPT(Ω, C), instead of using a generic Ci (ui ), (9.17) is used as the cost func-
tion for channel i, based on the primal algorithm (9.15) and the barrier function as
in (9.18). Thus the equilibrium point of (9.15) closely approximates the solution of
OPT(Ω, C). Initial channel powers are selected for both cases as
u(0) = [0.216 0.221 0.226 0.231 0.236 0.241] (mW)
We first present three cases in which the parameter selection strategy is not used
as a guideline (thus it is possible that the game settles down at an NE solution where
channels do not reach their OSNR targets). In all cases, the user-defined parameters
βi in G(M, Ω̂i , Ji ) are chosen as same as βi in OPT(Ω, C). A fixed αi is used
of 0.001, 1, and 20, respectively. With these pricing mechanisms, the total power
(uT ) vs. iteration and channel OSNR vs. channel number are shown in Figs. 9.5, 9.6
and 9.7 for three cases.
Thus without proper pricing mechanism, OSNR targets may not be achieved for
some channels or all channels, while the link capacity constraint is satisfied in all
cases. Furthermore, we notice that the penalty term
1
P0 − j ∈M uj
in G(M, Ω̂i , Ji ) plays a key role with small αi . In other words, with larger αi (say,
αi = 20 in the third case), total power is smaller than the link capacity constraint.
While with smaller αi in the first two cases, total power approaches the constraint
and higher channel OSNR is possibly achieved.
Channel powers, uopt and u∗ , inthree games versus channel number are shown
in Fig. 9.8. System cost C(u) = i∈M Ci (ui ) is evaluated via uopt and u∗ , re-
spectively and is shown in Table 9.1. Results imply that larger αi degrades system
performance and even violates the system constraints.
Next three other cases are shown in which proper pricing mechanisms are chosen
such that OSNR targets for all channels are achieved. In the game-theoretic frame-
9.4 Numerical Study of Efficiency in Nash Game 155
work G(M, Ω̂i , Ji ), the parameter selection strategy (7.28) is used such that proper
pricing mechanisms are chosen and OSNR targets for all channels are achieved.
Although the parameter selection strategy acts as a guideline for the selection of
156 9 Nash Equilibria Efficiency and Numerical Studies
each βi , it is practically intractable. Parameters αi are set at 1 for all cases and βi
are selected as in Table 9.2 such that different pricing mechanisms are chosen for
G(M, Ω̂i , Ji ).
Since we do not use Monte Carlo method [44] to simulate, we select βi in three
games by using the following rules. Firstly, βi increases for each channel, i.e., Game
[c] has the largest βi compared to Game [a] and Game [b]. Secondly, Game [b] has
the largest ratio of βi to βmin .
The efficiency of these two solutions u∗ and uopt , is compared by evaluating the
system cost C(u). The corresponding system cost values are obtained and shown in
Table 9.3.
9.4 Numerical Study of Efficiency in Nash Game 157
The results in Table 9.3 (compared with Table 9.1) show that the efficiency in
the solution of the Nash game (user optimization) can be improved by a proper
pricing mechanism. It is a well-known that no full efficiency can be obtained from
results in economics [38], transportation [129] and network resource allocation [64].
However, the Nash game solution gets very close to the optimal solution for system
optimization (see Table 9.3).
Figure 9.9 shows total power versus iteration. Channel power and channel OSNR
versus channel number are shown Figs. 9.10 and 9.11, respectively. Both constraints
(link capacity constraint and channel OSNR target) are satisfied in all cases. From
Fig. 9.11 it is seen that among the three cases, channel final OSNR values in Game
[c] get closest to the optimal solution of OPT(Ω, C).
We recall that the parameters βi in the Nash game are upper-bounded by the
condition (7.23) in Theorem 7.4, or in other words, the ratio of βi to βmin is upper
bounded. This condition restricts each channel asking unilaterally for a much higher
OSNR target than others. This phenomenon is also reflected in the selections of βi
in three cases. We take Game [b] for an example. In this case, β3 = 9.4 which is
greatly larger than other βi , indicating that channel 3 asks for a highest OSNR level.
158 9 Nash Equilibria Efficiency and Numerical Studies
Meanwhile, channel 4 has the smallest β4 = 3.9. Thus the largest ratio of βi to βmin
in Game [b] is around 2.41, which is the largest ratio among the three cases (2.23,
2.41, 1.81, in Game [a], [b], [c], respectively). Recall that Game [b] results in the
9.5 Notes 159
largest system cost value C(u∗ ), 4.6216, which implies a loss of efficiency of the
Nash equilibrium in the case when βi deviates from the average.
9.5 Notes
In this chapter we discussed and alternative constrained OSNR optimization from
the perspective of system performance. This system optimization framework can
be used to measure the efficiency of Nash equilibria of the Nash game presented
in Chap. 7. A numerical approach can be used to study the effects of parameters
in individual game cost functions. Results show that the OSNR target in the game-
theoretic framework can be achieved and the efficiency can be possibly improved
by appropriate selection of parameters.
Chapter 10
Simulations and Experimental Studies
Abstract This chapter provides simulation and experimental results for various al-
gorithms studied in previous chapters. The first section describes the physical setup.
This is followed by simulations and experimental results based on implementing
iterative algorithms in Chap. 7 for a Nash game with two, three, and five channels,
respectively, in a point-to-point link topology. The last section presents results for
partitioned Nash game framework in a multi-link topology and a quasi-ring topol-
ogy, based on implementing the hierarchal algorithms studied in Chap. 8. Full use of
the flexibility of channel power adjustment at each optical switch is assumed and the
multi-link is partitioned into stages with single links. Simulation and experimental
results are given for each type of network topology.
10.1 Introduction
In the following we present simulation and experimental results for the various algo-
rithms studied in some of the previous chapters. Simulations are done in MATLAB
while the experimental results are performed on a physical setup described below.
An optical network test system (ONTS) composed of the following optical
devices has been used: stabilized light sources (LS), variable optical attenuators
(VOA), optical spectrum analyzer (OSA), optical Erbium-Doped Fiber Amplifier
(OA), tunable bandpass fiber optic filters, fiber optical couplers and an ASE broad-
band source. Parts of the ONTS are shown in Fig. 10.1.
Any transmitter (Tx) is composed of a light source (of a particular wavelength)
and a variable optical attenuator. Each channel input power is adjustable by setting
the value of the corresponding VOA. An OSA provides accurate and comprehensive
measurement capabilities for spectral analysis, and can be used in ONTS to mea-
sure channel optical powers and OSNRs. An OA amplifies optical signals and noise
simultaneously. Tunable filters are used to adjust the center wavelength of a narrow
passband so that it separates optical signals with different wavelengths. Fiber optical
couplers are used to combine or distribute optical powers from single (or multiple)
input to single (or multiple) output. Couplers are designed bi-directionally and thus
can be used as a coupler or a splitter. Input optical noise in ONTS is obtained by
using the ASE broadband source.
Results have been implemented in LabVIEW1 is used for the purpose of com-
munication and control. Communication between instruments and computers is re-
alized via GPIB (General Purpose Interface Bus, also referred to as IEEE-488).
LabVIEW first communicates with the light sources to initialize them (e.g., wave-
length selection and light source power setting). This is followed by communication
with VOAs to get and set the attenuation configurations and then with OSAs to
measure the OSNR level and output power of each channel. Distributed optimiza-
tion algorithms are embedded in the control block by using the MathScript Node in
LabVIEW.
The ONTS with a single optical link is setup shown in Fig 10.2.
as a platform and development environment for a visual programming language from National
Instruments® .
10.2 Point-to-Point Link Topologies 163
Consider multiple channels transmitted over a single optical link with a total power
capacity constraint P 0 = 2.5 mW (3.98 dBm). The associated system matrix Γ for
m = 3 is obtained as
⎡ ⎤
0.7463 0.7378 0.7293
Γ = ⎣ 0.7451 0.7365 0.7281 ⎦ × 10−4
0.7438 0.7353 0.7269
For the iterative algorithms, the channel parameters in cost functions are se-
lected satisfying both the sufficient conditions for the existence of a unique NE
solution, (7.22)–(7.24), and the conditions for stability of the update scheme,
(7.40)–(7.41). Simulations are repeated with the following selected parameters:
α = [0.001, 0.001, 0.001], β = [1, 3, 2] and a = [1, 1, 1].
Two distributed algorithms are implemented in simulation. At each iteration time
(n + 1), ui (n + 1) is obtained via PUA, i.e., via solving (7.33),
1 βi ai
αi + =
(P 0 − j ∈M,j =i uj (n) − ui (n + 1))
2
( OSNRi (n) − Γi,i + ai )ui (n + 1)
1
where μ = 0.01. In addition, the initial channel power value in GA are set as u(0) =
[0.116 0.121 0.126] (mW).
A Nash game with two channels is considered first. Two channels compete for
the power resources and the game settles down at the NE solution via PUA and GA.
Figures 10.3, 10.4 and 10.5 show the evolutions of channel input power, total power
and OSNR. It can be observed that wide fluctuations are obtained when running
PUA. These fluctuations are largely avoided when using GA. Moreover, during the
iterative process, the total power constraint is violated when using PUA, but not
when using GA.
164 10 Simulations and Experimental Studies
making full use of input power resource. Simulation results showed above prove the
performance of the algorithms and validate the analytic results.
Consider the ONTS with a single optical link as shown in Fig. 10.2, with one OA.
Multiple channels are transmitted with the following wavelengths: 1533.47 nm,
166 10 Simulations and Experimental Studies
1535.04 nm, 1537.40 nm, 1555.75 nm and 1558.17 nm. We continue using the no-
tations defined in Chap. 7. Since only one OA is used, i.e., N = 1 in (6.17), diagonal
elements of the system matrix Γ are derived from (6.17), i.e.,
ASEi
Γi,i = ,
P0
10.2 Point-to-Point Link Topologies 167
where P 0 = 1.5 mW is the constant total power target of this link and ASEi is the
ASE noise, defined as
ASEi = 2nsp (Gi − 1)hνi B,
where
nsp amplifier excess noise factor, nsp = 1
h Planck’s constant, h = 6.626
B optical bandwidth, B = 1010
νi optical frequency of channel i
Gi gain of OA at the wavelength of channel i
The two algorithms, PUA and GA, are applied. In PUA, each ui (n + 1) is ob-
tained by solving the following equation:
1 βi ai
αi + =
(P 0 − j ∈M,j =i uj (n) − ui (n + 1))
2
( OSNRi (n) − Γi,i + ai )ui (n + 1)
1
In GA,
1 βi ai
ui (n+1) = ui (n)−μ αi + − ,
(P 0 − j ∈M uj (n))
2 1
( OSNR + ai − Γi,i )ui (n)
i (n)
Since all channels are under same conditions, for example, same noise profile
and same network price αi , they compete for the power resource via the value of
βi : the larger the βi is, the higher the OSNR and input power values are obtained.
Snapshots of initial states and final results on OSA are shown in Figs. 10.14 and
10.15, respectively.
herein we also fully use the flexibility of channel power adjustment at each optical
switch and partition the multi-link structure into stages with single links. A mesh
network topology is studied next. Simulation results are given for each type of net-
work topology.
Some assumptions and parameters defined in simulation for all types of network
topology are presented first. As before, in all topologies, we assume that each link
is with same number of amplified spans and all optical amplifiers deployed along
links have the same gain spectral shape. The dynamic adjustment parameter γl,i is
bounded within [γmin , γmax ] = [0, 10] for all l ∈ L and for all i ∈ M. In partitioned
170 10 Simulations and Experimental Studies
Nash games, the individual channel cost function Jl,i is with the following parame-
ter selection:
αl,i = 10 × Γli,i ,
βl,i = 1 + 0.1 × i,
al,i = 50 × l × Γli,i , i ∈ Ml , l ∈ L,
10.3 Multi-link and Mesh Topologies 171
where Γli,i is the diagonal element in the link system matrix Γl . Values of Γli,i
are obtained in each individual network topology. Note that the condition (8.36)
is satisfied.
used in which each link is a stage. This partition simplifies the partitioned structure
and the convexity condition is naturally satisfied, as we discussed before.
We present MATLAB simulation results for the topology shown in Fig. 10.16 by
applying the iterative hierarchical algorithm (8.39), (8.40). Link total power targets
are P10 = 1.5 mW, P20 = 2.5 mW and P30 = 2.0 mW. Diagonal elements of each
link system matrix Γl , l = 1, 2, 3, are obtained as
For channels added directly from Tx, we set the γ value to 0, since channel power is
adjusted at Tx. Thus γ ∗ is feasible with respect to the predefined range [γmin , γmax ].
10.3 Multi-link and Mesh Topologies 173
Let us take a closer look at the simple quasi-ring topology in Fig. 8.6(a) and three
channels whose optical paths are shown in Fig. 8.6(b). We break the closed loop and
select link 3 as the starting link. The unfolded configuration is shown in Fig. 10.23.
The overall recursive process is such that stage Nash games on links are played
sequentially: l3 → l1 → l2 . On link 3, the adjustable parameters for channels 1 and
2 are initially set as γ3,11 and γ 1 , respectively, where the superscript 1 indicates
3,2
the number of iteration of the game among links. The game on link 3 settles down
at u∗3 (μ∗3 ) with the corresponding channel output power p3∗ (μ∗3 ). Sequentially, the
game on link 1 is played with an NE solution, u∗1 (μ∗1 ). Then channel output power
is p1∗ (μ∗1 ). Given p3∗ (μ∗3 ), and the adjustable parameter on link 1, γ1,3∗ , is deter-
mined. The game on link 2 is played after that. The NE solution of this game is
u∗2 (μ∗2 ) and the channel output power is p2∗ (μ∗2 ). Then the adjustable parameters on
10.3 Multi-link and Mesh Topologies 175
Fig. 10.23 Unfolded quasi-ring topology with the starting link, link 3
A step-size of η = 0.1 is used in the link algorithm. The partitioned Nash game is
played as described above. For every 20 iterations, the link adjusts its price via the
link algorithm and then channels readjust their powers. Evolutions in time of chan-
nel input power, total power and link price on each link l are shown in Figs. 10.24–
10.29, respectively.
176 10 Simulations and Experimental Studies
The adjustable parameters for three links and three channels are obtained as in
the following 3 × 3 matrix:
⎡ ⎤
0 0 0.7208
γ ∗ = ⎣ 7.5983 1.4868 0 ⎦
1.5277 0.6042 0
The overall game settles down since γ ∗ is feasible. Note that a different starting
link can be selected, say, link 1, such that games on links are played sequentially:
l1 → l2 → l3 . Typically we select the starting link where channels are added directly
from Txs.
10.3 Multi-link and Mesh Topologies 177
We study a mesh network topology as shown in Fig. 10.30(a), where eight channels
are transmitted over six links. The channel routes are shown in Fig. 10.30(b).
It can be seen from Fig. 10.30(b) that there exists a closed loop among links 1,
4, and 6, which is formed by the optical paths of channels 3, 4, 5, and 6. We break
the closed loop and select link 1 as the starting link. This unfolded configuration is
shown in Fig. 10.31.
The overall recursive play process is described as follows. Games on links 1, 6,
and 4 are played in a precedence order: l1 → l6 → l4 . Since the closed loop among
links 1, 4, and 6 is unfolded, on link 1 the adjustable parameters for channels 5
178 10 Simulations and Experimental Studies
5.409 × 10−4 , Γ32,2 = 8.564 × 10−4 , Γ37,7 = 8.483 × 10−4 , Γ38,8 = 5.34 × 10−4 ,
Γ41,1 = 4.327 × 10−4 , Γ42,2 = 6.852 × 10−4 , Γ43,3 = 0.0011, Γ44,4 = 0.0022, Γ45,5 =
0.0022, Γ46,6 = 0.0011, Γ57,7 = 0.0011, Γ58,8 = 7.12 × 10−4 , Γ63,3 = 0.0018 and
Γ64,4 = 0.0036. We show the evolutions of channel input power, total power and
link price on links 1, 2, and 4 in Figs. 10.32–10.37, respectively. The final adjustable
parameter values for six links and eight channels are obtained as in the 6 × 8 matrix
below:
⎡ ⎤
0 0 0 0 0.7067 1.3548 0 0
⎢ 0 0 0 0 0 0 0 0 ⎥
⎢ ⎥
⎢ 0.6893 0.4308 0 0 0 0 0.6535 1.0415 ⎥
γ∗ =⎢ ⎢ 1.1210 0.7271 0.8486 0.4335
⎥
⎢ 0 0 0 0 ⎥ ⎥
⎣ 0 0 0 0 0 0 0 0 ⎦
0 0 2.5839 1.1778 0 0 0 0
180 10 Simulations and Experimental Studies
Next experimental results are presented on the same ONTS where this time three
links are setup as shown in Fig. 10.38. Each link is composed of an OA. Total
power targets are P10 = 1.5 mW, P20 = 2.5 mW and P30 = 1.5 mW, respectively.
Channels 1 and 2 are added on link 1 and channel 3 is added on link 2. We use
following wavelengths for each channel: 1535.04 nm on Light source 1 (LS 1),
1537.40 nm on Light source 2 (LS 2) and 1533.47 nm on Light source 3 (LS 3).
Channel 3 is dropped after link 2. This is realized by using a filter which is used to
perform wavelength selection. Channels 1 and 2 are transmitted through two filters,
respectively.
10.3 Multi-link and Mesh Topologies 181
We partition this three-link game into two stage games (K = 2): Stage 1 is com-
posed of links 1 and 2 and stage 2 is link 3. Diagonal elements of each stage system
matrix Γk are derived from stage OSNR mode. Particularly in this configuration,
ASE1,i ASE2,i
Γ1i,i = + , i = 1, 2,
P10 P20
ASE2,3
Γ13,3 = ,
P20
ASE3,i
Γ2i,i = , i = 1, 2,
P30
182 10 Simulations and Experimental Studies
where
ASEl,i = 2nsp (Gl,i − 1)hνi B,
with Gl,i the gain of OA at the wavelength of channel i on link l.
The value of the dynamic adjustment parameter γk,i is in fact the attenuation
value of channel i on stage k, which is physically bounded.
The iterative hierarchical algorithm developed in Chap. 8 is implemented, which
is composed of a channel algorithm and a link algorithm. At each stage k, the fol-
lowing channel algorithm is used:
βk,i 1 1 uk,i (n)
uk,i (n + 1) = − − − Γki,i ,
αk,i + μk,r(i) (t) OSNRk,i (n) OSNRk
,i ak,i
10.3 Multi-link and Mesh Topologies 183
where stage k
is the precedent of stage k and OSNRk
,i and μk,r(i) (t) are invariable
during the channel iteration in stage k. Then after every Nk iterations of the channel
algorithm, the new link price is generated according to the following link algorithm:
+
μk (t + 1) = μk (t) − ηk Pk0 − Ek uk μ̄k (t) ,
In the experiment, the game on Stage 1 is played first. Figure 10.39 shows the evo-
lution in iteration time of channel input power on Stage 1.
For every N1 = 10 iteration, link 1 and link 2 adjust their prices simultaneously
(link algorithm) and then channels readjust their powers. Total power and prices
evolutions are shown in Fig. 10.40. After Stage 1 settles down, the game on Stage 2
starts to play. For every N2 = 7 iteration, link 3 adjusts its price. The evolutions are
shown in Figs. 10.41 and 10.42. Snapshots of the values of OSNR and attenuation
184 10 Simulations and Experimental Studies
Fig. 10.43 Snapshot: partial evolution of OSNR and attenuation value on stage 2
during the iteration on Stage 2 are shown in Fig. 10.43. The final values of adjustable
parameters, i.e., the attenuation values are
∗ 0 0 0
γ =
4.454 4.073 0
Part III
Robustness, Delay Effects, and Other
Problems
Chapter 11
Robustness and Delay Effects on Network
Games
11.1 Introduction
Lemma 11.1 Consider the ith channel in an arbitrary multi-link optical network,
i ∈ M. Its OSNR value, fed back from Rx to Tx, evaluated at time t is denoted by
11.2 System Model 191
ui (t − τi,i )
yi (t) = (11.1)
n0,i + j ∈M Γi,j uj (t − τi,j )
where Γi,j , elements of the full (n × n) system matrix Γ , are defined as (6.36).
Note that (11.1) is the OSNR signal that is measured at Rx and then fed back
to Tx. Thus both the forward and the backward time delays are present, since the
OSNR value is fed back to the transmitter.
For a single-link case, as in Lemma 6.2, time-delay notation can be simplified
such that τ = τ f + τ b , where τ is the round-trip time delay of the link, τ f is the
forward propagation delay, and τ b is the backward propagation delay. Thus, from
Lemma 6.2, we obtain the following lemma.
Lemma 11.2 Consider the ith channel in a point-to-point optical link, i ∈ M. Its
OSNR value, fed back from Rx to Tx, evaluated at time t is denoted by OSNRi or yi
and is given as
ui (t − τ )
yi (t) = , i∈M (11.2)
n0,i + j ∈M Γi,j uj (t − τ )
Next let us revisit the iterative algorithm for channel power. From Chap. 7, recall
that the algorithm (7.62) ensures that the closed-loop system converges to the NE
equilibrium when no time delays are considered. In order to find out under what
conditions this holds in the presence of time delay let us introduce tunable control
parameters, ρi , for each channel i, into the control algorithm (7.62) and convert into
it to continuous-time. Thus, for each channel i we subtract ui (n) from each side
of (7.62). The n time index is normalized with respect to the time period T0 , i.e. a
time period of T0 elapses between n and n + 1 iterations, i.e., tn = nT0 , where tn
is the total elapsed time at the nth algorithm update. We also introduce parameter
ρi , 0 < ρi ≤ 1, multiplying the right-hand side of (7.62), and we approximate the
left-hand side by dudti (t) = ui (n + 1) − ui (n), to obtain
# *
dui (t) βi 1 1
= ρi − − Γi,i + ai ui (t) (11.3)
dt αi ai yi (t)
(11.1) within the discrete-time domain. However, as the majority of time-delay sta-
bility results in literature in continuous-time, this is the approach we take also.
Consider further that (11.3) is modified by substituting ui (t) for ui (t − τi,i ),
which yields
# *
dui (t) βi 1 1 Γi,i
= ρi − ui (t − τi,i ) + − 1 ui (t) (11.4)
dt αi ai yi (t) ai
This is used so as to simplify the closed-loop system’s structure. We explicitly write
this as the following assumption.
Assumption 11.1 Each channel, i, has knowledge of its own round-trip time de-
lay, τi,i .
Assumption 11.1 is realistic in practice since round-trip time delays, which are
due to signal propagation times, are measurable. Furthermore, the decentralized na-
ture of the control algorithm (11.3) is also preserved since this added information
is local. Because the control algorithm is implemented in software, this substitution
as in (11.4) is allowed. Furthermore, the channel algorithm running at the channel
sources may store the signal powers, u(t), over time. Thus, with τi,i known, it is
used to reference the appropriate past channel power u(t − τi,i ). Hence, (11.4) is a
modified channel algorithm of (7.62) to handle time delays more effectively. Note
that while the control algorithm (11.4) can be modified, the OSNR model (11.1),
which is the physical model of the optical network, cannot be modified.
or, after shifting it around the equilibrium point u∗ , (7.11) (see Chap. 7), the equiv-
alent form is
dui ρi
=− (ai − Γi,i )ui (t) + Γi,j uj (t − τi,j ) (11.6)
dt ai
j ∈M
This (11.6) is the closed loop that is analyzed next. Specifically, we give sufficient
conditions for stability of the time-delay closed-loop system (11.6). These condi-
tions are given in terms of tunable parameters ρi and ai . The approach is based on
11.3 Delay Effects in Network Games Without Constraints 193
an analysis as in [100], but unlike [100], the system does not have any symmetry in
the system matrix Γ that can be exploited.
First, let us map (11.6) into the form (A.33) (see the appendix) in order to study
its asymptotic stability as in Definition A.32. For convenience, we restate (A.33)
using the state variables and system parameters from (11.6) as
2
m
u̇(t) = A0 u(t) + Ak u(t − τk ), τk ≥ 0 (11.7)
k=1
for m channels in the system. Starting from (11.6), assume ai > Γi,i for all i, let
εi = ρaii (ai − Γi,i ), and define
A0 = diag(−εi )
In order to define Ak one can proceed as follows. Notice in (11.6), for channel i,
there may be up to m delay terms with the form − ρaii Γi,j u(t − τi,j ) for j = 1, . . . , m.
Each time delay, τi,j , may be unique for a total of m2 time delays for all i, j =
1, . . . , m. Thus, to rewrite (11.6) into the form (11.7), each coefficient − ρaii Γi,j ,
associated with the time-delayed state u(t −τi,j ), occupies one element of the m×m
matrix Ak for the associated time-delayed state u(t − τk ). Let k = (i − 1) × m + j
for i, j = 1, . . . , m and let Ak be an (m × m) matrix of zeros except for the (i, j )th
element corresponding to the k index which is equal to − ρaii Γi,j . Furthermore, let
τk = τi,j . With these A0 , Ak , and τk (11.6) is mapped directly to (11.7).
Then by applying Laplace transform to (11.7) we obtain
2 m2
m 0
−sτk −st
sI − A0 − Ak e U (s) = u(0) + Ak e dt e−sτk (11.8)
k=1 k=1 −τk
or
1
Δ(s) = (sI − A0 ) I + diag F (s)
s + εi
194 11 Robustness and Delay Effects on Network Games
m2 −sτk .
where F (s) = k=1 −Ak e Finally, we write
Δ(s) = (sI − A0 ) I + L(s) (11.10)
Γi,j e−τi,j s
Li,j (s) = ρi Γi,i
(11.11)
ai s + ρi (1 −
ai )
We now state and prove a stability result (Theorem 11.6) for the closed-loop
system resulting from games with no coupled constraints; part (i) gives a time-delay
dependent stability condition based on the tuning parameters ρi ; part (ii) provides a
conservative stability condition independent of time delay. Before that we give two
useful lemmas.
Γi,i e−τi,i j ω
Li,i (j ω) = ρi Γi,i
(11.12)
ai j ω + ρi (1 −
ai )
Γi,i μ̄ρi
1− 0 ≤ μ̄Li,i j ω∗ − (−1) (11.14)
ai ω2 + ρ 2 (1 − Γ /a )2
i i,i i
11.3 Delay Effects in Network Games Without Constraints 195
Proof of Lemma 11.4 is immediate, based on the triangle inequality and the fact
that the magnitude of Li,i (j ω) is monotone with respect to ω.
0.68244
≤ ρi
(1 − Γ i,i
ai )τi,i
Theorem 11.6 (i) If the system design parameters ai are selected such that
Γi,j < ai < Γi,i + Γi,j , ∀i (11.15)
j =i j
then the closed-loop time-delay system, (11.4), (11.1) is stable if the tuning param-
eters ρi are selected such that
0.68244
≤ ρi , ∀i (11.17)
(1 − Γ i,i
ai )τi,i
π
0 < ρi < 0 , ∀i (11.18)
Γ
2τi,i (1 − ai,ii ) Γi,i
2 /(a −
i j Γi,j )2−1
(ii) The closed-loop time-delayed system, (11.4), (11.1) is stable for any time delays
if the parameters ai are selected such that
Γi,i + Γi,j < ai , ∀i (11.19)
j
μ̄ρi Γi,j
0 < μ̄Li,i (j ω) + 1 (11.20)
ω2 + ρi2 (1 − Γi,i /ai )2 ai
j =i
196 11 Robustness and Delay Effects on Network Games
∀i, ∀ω ≥ 0 where Li,i (j ω) is defined in (11.12). The RHS of (11.20) is the distance
from the critical point, −1, to the Nyquist curve of μ̄Li,i (j ω). It can be shown that
if
1 Γi,j
Γi,i
< μ̄Li,i (j ω) + 1, ∀i, ∀ω ≥ 0 (11.21)
(1 − )
ai
ai
j =i
Thus by (11.23) and (11.14) it follows that (11.22) and (11.21) hold, and hence
−1 ∈/ eig(μ̄L(j ω)). To show (11.23), note that μ̄ = 1 gives the smallest value
on the RHS of (11.23). After some manipulation, this can be rewritten as
Γi,i 2 Γ2 π2
ρi2 1 − i,i − 1 < (11.24)
ai (ai − j Γi,j )2 2
4τi,i
By (11.15) the third term on the LHS of (11.24) is positive and (11.24) can be
rearranged as in (11.18). Then (11.24) holds along with (11.23) and therefore
−1 ∈/ eig(μ̄Lii (j ω)). By Proposition 11.3, the τ -dependent stability is proved.
(ii) We return to (11.21). On the RHS, we have |μ̄Li,i (j ω) + 1| ≥ 1 − |Li,i (0)|
based on 0 < μ̄ ≤ 1 and the monotonicity of |Li,i (j ω)| with respect to ω. Thus,
|μ̄Li,i (j ω) + 1| ≥ 1 − Γi,i /(ai − Γi,i ) and from condition (11.19) it follows that
(11.21) holds, which implies (11.20) holds, and −1 ∈ / eig(μ̄L(j ω)).
Corollary 11.7 Both (11.17) and (11.18) can only be satisfied simultaneously in
the range
Γi,i
+ Γi,j ≤ ai (11.25)
δ
j
0
where δ
= ( 2(0.68244)
π
)2 + 1 = 2.5096.
11.4 Delay Effects in Network Games with Constraints 197
In this section, we derive the closed-loop expression for the channel-price or primal–
dual algorithms (7.62) and (7.61) applied to the time-delayed OSNR model (11.2).
We then apply a coordinate shift that simplifies the foregoing stability analysis in
the following section.
Let us first briefly review the primal–dual algorithms developed in Chaps. 7
and 8. Recall that the channel algorithm (7.62) is located at the transmitters (Tx)
and the price algorithm or link algorithm (7.61) is located at the receivers (Rx) at
the output of the link. The control algorithm of the ith channel adjusts the channel
powers, ui , towards OSNR optimization using its OSNR value and channel price
μ as feedback. The link algorithm measures the total sum of the channel powers
and computes a channel price, μ, based on the threshold for nonlinear effects. The
channel price is fed back to the channel algorithms at Tx.
Let us consider the closed-loop system obtained by using the channel and price
algorithms and the OSNR model (11.2). Recall, from Chaps. 7 and 8 that the channel
algorithm operates on a faster time-scale reflected by the fact that its index, n, is
related to the index of the link algorithm, k̄, via n = K k̄, where K is large, e.g.,
K = 100. In continuous-time, we denote the “fast” time variable by t and the “slow”
time variable by tˆ. The two time-scales are related by tˆ = εt, where ε = K1 is small,
198 11 Robustness and Delay Effects on Network Games
and we will see later that the stability of the closed-loop system is ensured if ε is
sufficiently small. The time delays on the two time-scales are related similarly by
τ̂ = ετ . A smaller ε value means a more effective time-scale decoupling between
the channel and the price (link) algorithm.
Let us begin with the continuous-time channel algorithm (11.3) derived in
Chap. 7, where αi = μ(tˆ ) = μ(εt) is the channel price, rewritten below as (11.26)
# *
dui (t) βi 1 1
= ρi − − Γi,i + ai ui (t) (11.26)
dt μ(εt) ai yi (t)
Now, substituting the time-delayed model (11.2) into (11.26) and including the
backward propagation delay, τ̂ b = ετ b , into the channel price, μ(tˆ ) = μ(εt), yields
#
dui (t) βi
= ρi
dt μ(ε(t − τ b ))
*
1 n0,i + j ∈M Γi,j uj (t − τ )
− − Γi,i + ai ui (t) (11.27)
ai ui (t − τ )
Note that (11.27) and (11.26) are on the “fast” time-scale, t. For the single-link case
here, τ = τi,i , since there is only one round-trip time delay. The price is adjusted on
the “slow” time-scale; let us rewrite (11.27) in this slow time-scale, tˆ, so that both
the channel and link/price dynamics are on the same time-scale. Using tˆ = ε t, from
(11.27) we obtain
#
d ûi (tˆ ) βi
ε = ρi
d tˆ μ(tˆ − τ̂ b )
*
1 n0,i + j Γi,j ûj (tˆ − τ̂ )
− − Γi,i + ai ûi (tˆ ) (11.28)
ai ûi (tˆ − τ̂ )
where we used ddttˆ = ε and û denotes u on the slow time-scale tˆ. This form is used
to apply standard singular perturbation theory (see Sect. A.9, (A.19)). If we divide
both sides of (11.28) by ε, then the 1ε term on the RHS is a large gain. Thus, the ε
term embodies the fast dynamics of the channel algorithm in the slow time-scale.
Let us move now to the price or link algorithm (7.61) and convert it to a
tˆ )
continuous-time version. Using the same approximation dμ( d tˆ
= μ(k̄ + 1) − μ(k̄),
we obtain in the slow time-scale, tˆ
μ̇(tˆ ) = η ûj (tˆ ) − P0 (11.29)
j =1
Finally let us incorporate the forward time delay that acts on the input powers to get
μ̇(tˆ ) = η ûj tˆ − τ̂ f − P0 (11.30)
j =1
As before one can modify (11.28) to eliminate ûi (tˆ − τ̂ ) in the denominator by
design (see also (11.4) for a similar technique). If we keep a record of past power
11.4 Delay Effects in Network Games with Constraints 199
input, ûi (tˆ − τ̂ ), we can then appropriately replace ûi (tˆ ) with ûi (tˆ − τ̂ ) to obtain
# *
βi 1
ε û˙ i (tˆ ) = ρi − n 0,i + Γ˜i,j ûj (tˆ − τ̂ ) (11.31)
μ(tˆ − τ̂ b ) ai j
where z and x denote in vector form the shifted variables, zi = ûi − u∗i and x =
Γ˜
μ − μ∗ . In the foregoing Γ˜a is a matrix with elements ai,ji , β is a column matrix with
elements βi , 1row is a row vector with all elements equal to 1, ρ = diag(ρi ). Note
that Γ˜a is invertible because of the diagonal dominance condition (see Chap. 7).
Now we can see that (11.32) and (11.33) is written in a standard singular pertur-
bation form. Let
−x
˜
h(x) = Γa β−1
(11.34)
μ∗ (x + μ∗ )
be the isolated root of (11.33) and use the coordinate shift
ẑ(tˆ ) = z tˆ − τ̂ f − h x(tˆ ) (11.35)
Next, taking the derivative of (11.35) with respect to time, tˆ, yields
˙ tˆ ) = ż tˆ − τ̂ f − dh(x) ẋ(tˆ )
ẑ(
dx
and using (11.33) it follows that
# *
x(tˆ − τ̂ )
dh(x)
˙ tˆ ) = ρ −β
ε ẑ( − Γ˜a z tˆ − τ̂ − τ̂ f − ε ẋ(tˆ )
μ∗ (x(tˆ − τ̂ ) + μ∗ ) dx
On the right-hand side, using (11.34) and then again (11.35) yields
dh(x)
ε ẑ˙ = −ρ Γ˜a ẑ(tˆ − τ̂ ) − ε ẋ(tˆ ) (11.37)
dx
200 11 Robustness and Delay Effects on Network Games
Thus, the coordinate shifted system to be analyzed is given by (11.38) and (11.39),
where x is on the fast and ẑ is on the slow time-scale. Let us introduce the reduced
system and the boundary-layer system (see the appendix). These are simpler to an-
alyze and can be used to study the full system (11.38) and (11.39).
To obtain the reduced system, we set ẑ = 0 in (11.38), that is, we assume that the
fast dynamics converge instantly to steady-state value. This yields
˙ tˆ ) = f h x̂(tˆ )
x̂( (11.42)
where we denoted its solution by x̂(tˆ ), which is distinct from x(tˆ ), solution of
(11.38). Let us use the same notation
ẑ tˆ(t) = (ẑ ◦ tˆ )(t) := z̃(t) (11.43)
d z̃(t)
dh(x(εt))
= g z̃(t − τ ) − ε f z̃(t) + h x(εt) (11.44)
dt dx
Finally, we set ε = 0, that is we assume that tˆ and x(tˆ ) are “frozen” (since they
evolve on the slow time-scale). Denoting the solution by ž(t), we get
d ž(t)
= g ž(t − τ ) (11.45)
dt
We can write (11.42) and (11.45) in their explicit forms by using (11.40), (11.41)
˙x̂(tˆ ) = η1row Γ˜ −1 β −x̂
(11.46)
a
μ∗ (x̂ + μ∗ )
d ž
= −ρ Γ˜a ž(t − τ ) (11.47)
dt
11.4 Delay Effects in Network Games with Constraints 201
Thus we can see that the reduced system (11.46) is a nonlinear, scalar system with
no time delay. On the other hand, the boundary-layer system (11.47) is a linear time-
delay system. Since the two systems are decoupled, they are much simpler to study
than the full system (11.38) and (11.39). The full system (11.38), (11.39) or (11.32),
(11.33), is studied based on these simplified subsystems and a composite Lyapunov
function.
In this section, we present conditions for stability of the closed-loop system (11.32),
(11.33) in the presence of time delays, based on the reduced and boundary-layer
systems (11.46) and (11.47). We apply a Lyapunov–Razumikhin stability analysis
(see Sect. A.11.2). This Lyapunov–Razumikhin stability analysis does not rely on
functionals and the stability conditions are simple in form, but relatively conserva-
tive.
The following lemmas are used to prove the main result, Theorem 11.10.
σ (ρ Γ˜a + Γ˜aT ρ)
τ< 0 (11.48)
2 σ̄ ((ρ Γ˜a )2 ((ρ Γ˜a )2 )T )
d ž
= A1 ž(t − τ ) (11.49)
dt
we ensure asymptotic stability if we satisfy the LMI (A.40), in fact ensuring expo-
nential stability.
For the boundary-layer system, (11.47), A1 = −ρ Γ˜a . The LMI (A.40) then be-
comes
⎛ 1 ⎞
− τ [Pρ Γ˜a + Γ˜aT ρP ] + (α̂0 + α̂1 )P 0 −Pρ Γ˜a ρ Γ˜a
⎝ 0 −α̂0 P 0 ⎠<0 (11.50)
˜ T ˜
−Γa ρ Γa ρP
T 0 −α̂1 P
We rewrite (11.50) using the Schur complement, we set Q = P −1 , and after some
manipulation one can obtain
1 1
T
− ρ Γ˜a Q + QΓ˜aT ρ + (α̂0 + α̂1 )Q + (ρ Γ˜a )2 Q (ρ Γ˜a )2 < 0 (11.51)
τ α̂1
202 11 Robustness and Delay Effects on Network Games
Notice that the second and third terms of (11.51) are positive definite. In addition,
notice that there always exists a Q to satisfy (11.51) given a sufficiently small τ .
Thus, we know that a solution exists, the only question remaining is the bound on τ
that gives such a result.
We can simplify (11.51) by a priori choosing Q = vI , where v is a constant
multiplier, and I is the identity matrix. Rearranging (11.51), and setting α̂0 = 0
yields
1
˜ ˜
τ α̂1 I + (ρ Γa ) (ρ Γa )
2 2 T
< ρ Γ˜a + Γ˜aT ρ (11.52)
α̂1
Following rather lengthy manipulations leads to (11.48) as a sufficient condition for
(11.52).
Theorem 11.10 For the singularly perturbed system (11.32) and (11.33), there ex-
ists ε ∗ > 0 such that, for 0 < ε < ε ∗ , the origin is exponentially stable if (11.48) and
(11.53) are satisfied.
for x̂ ≤ r1 , where r1 > −μ∗ , and k∗ > 0 is a constant. By (11.53), 1row Γ˜a−1 β > 0.
Note that we can make r1 arbitrarily large, but (11.54) is satisfied locally for x̂.
Consider the boundary-layer system (11.47) and select the Lyapunov function
W (ž) = žT ž. By (11.48) in Lemma 11.8, this is exponential stable and it can be
shown that
∂W
) )2
g ž(t − τ ) = 2žT −ρ Γ˜a ž(t − τ ) ≤ −k∗2 )ž(t))2
∂ ž
) ) (11.55)
) ∂W )
) ) = 2ž2
) ∂ ž )
2
for x ≥ r2 , where r2 > −μ∗ , and k2 > 0 and k3 > 0 are constants. By (11.53),
1row Γ˜a−1 β > 0. Thus, we have the overall range, r2 ≤ x ≤ r1 . Note that r2 can be
selected arbitrarily close to −μ∗ . Thus, the inequalities (11.56)–(11.59) are valid
locally over x.
We analyze the stability of the full closed-loop system (11.38), (11.39) via a com-
posite Lyapunov function obtained from these two above. This composite Lyapunov
function is defined as χ(x, ẑ) = V (x) + W (ẑ) = 12 x 2 + ẑT ẑ, hence is quadratic.
Thus, we immediately satisfy the first condition in Theorem A.33.
Next we use the composite Lyapunov function χ(x, ẑ) for (11.38) and (11.39),
and exploit the Lyapunov inequalities (11.54), (11.55) and Lipschitz properties
(11.56)–(11.58). The rest of the proof follows as in Theorem 11.4 [69]. We take
the time-derivative of χ(x, ẑ) along the trajectory of the system (11.38) and (11.39),
∂V
∂W ∂h
1 ∂W
χ̇ = f ẑ + h(x) − f ẑ + h(x) + g ẑ(tˆ − τ̂ ) (11.60)
∂x ∂ ẑ ∂x ε ∂ ẑ
where the general functions f , g and h, are defined in (11.40), (11.41), and (11.34),
respectively. Next, we use the norm inequalities (11.54), (11.55), and (11.56)–
(11.58), applied to each term on the RHS of (11.60). For the first term of (11.60),
we apply (11.56), (11.57), and (11.54) to get
∂V
f ẑ + h(x) ≤ L3 x2 ẑ2 − c3 x22 (11.61)
∂x
where L3 and c3 are constants. Using (11.55) the third term of (11.60) can be rewrit-
ten as
1 ∂W
−1
g ẑ(tˆ − τ̂ ) ≤ k∗2 ẑ22 (11.62)
ε ∂ ẑ ε
∂W ∂h
f ẑ + h(x) ≤ b4 ẑ22 + L4 x2 ẑ2 (11.63)
∂ ẑ ∂x
204 11 Robustness and Delay Effects on Network Games
where d̃1 = c3 , d̃3 = k, d̃4 = b4 , and d̃5 = L3 + L4 . The di matrix on the RHS of
(11.64) is positive definite if
4d̃1 d̃3
ε< = ε∗
4d̃1 d̃4 + d̃52
Thus, (11.64) satisfies the second condition in Theorem A.33. The system (11.38)
and (11.39) is exponentially stable.
The time-delay bound (11.48) is very simple in form and offers a clear intu-
ition for selecting the design parameters ai and ρi . This time-delay bound (11.48)
is inversely proportional to the control gain, ρi and is proportional to the design pa-
rameters ai . The trade-off for stability is a slower convergence time for the control
algorithms.
Until now the OSNR models have assumed that the system gains, Γi,j , are constant
(see Lemmas 11.1 and 11.2). In fact due to slow parameter drift over long periods of
time and changes in amplified spontaneous emission (ASE), they are time-varying.
The slow parameter drift occurs due to the aging of network hardware. Uncertainties
f
exist in time delays τi,j and τib for i, j ∈ M due to external environmental effects
on the fibers and system reconfigurations. Fiber lengths may expand or contract due
to environmental effects, or the network may be reconfigured without remodeling
the network and redesigning all of the control algorithms. Finally, uncertainty can
exist in the transmitter noise at the sources due to slow parameter drift. These un-
certainties are captured in the following OSNR model.
11.5 Robustness and Delay Effects Combined 205
Slow parameter drift in optical networks over long periods of time may be mod-
eled as additive uncertainty in the Γ matrix. Additive uncertainty accurately models
the independent changes that occur to each Γi,j element over time. Let the uncer-
tainty set for the perturbed system ΓΔ ∈ ΩΓ be defined as
ΩΓ = Γ + ΔΓ | ΔΓi,row 2 ≤ q̂i , ∀i (11.65)
where q̄j is the independent uncertainty for input channel j . Note q̄j < 1.
The transmitter noise may drift over time as the hardware ages. Assume that
the signal powers change proportionally, modeled via a proportional additive uncer-
tainty and let us define the uncertain transmitter noise, mΔ,0,i ∈ Ωni , where
Ωni = n0,i + δni n0,i | δni 2 ≤ qni , ∀i (11.67)
with qni being the independent uncertainty bound on the transmitter noise for chan-
nel i. Note that qni < 1 since the noise power m0,i is positive.
Then the OSNR model that captures multiple time delays and uncertainties
(11.65), (11.66), and (11.67) follows immediately from Lemma 11.1. This perturbed
OSNR model for the ith channel in a multi-link network is given as
uΔi (t − τi,i )
yΔi (t) = (11.68)
nΔ,0,i + j ∈M ΓΔi,j uΔj (t − τi,j )
i
where Γ is a matrix with elements Γi,j defined in (6.15), ΔΓ is a matrix with ele-
ments ΔΓi,j , and q̂ is an independent uncertainty bound, distinct from q̂i defined in
(11.65). The uncertainty set (11.69) is similar to the uncertainty set (11.65), except
the uncertainty bounds in (11.65) apply to each channel output separately, while
(11.65) has one independent bound for all output.
206 11 Robustness and Delay Effects on Network Games
Consider the continuous-time control algorithms developed based on the link algo-
rithm (7.61) and the channel algorithm (7.62). The following analysis applies to a
single link with time delays and uncertainties as modeled in (11.68). Conditions that
ensure the stability of the closed-loop system are presented.
Using a similar procedure to rewrite (7.61) and (7.62) in their continuous-time
forms yields
μ̇(tˆ ) = η uin,j − P0 (11.70)
j =1
# *
d ûi (tˆ ) βi 1 1
ε = ρi − − Γi,i + ai ûi (tˆ − τ̂ ) (11.71)
dt ˆ μin,i ai yin,i (tˆ )
where uin,j is the signal power transmitted to the link algorithm, μin,i is the channel
price transmitted to the source i, and yin,i is the OSNR signal transmitted back to the
source i. The “in” subscript is used to abstract various signals as input to the control
algorithms without the details of the time delays or uncertainties. This allows us to
substitute for uin,j , μin,i and yin,i appropriately later in the analysis. We use this
notation throughout the remainder of this chapter.
Consider (11.68) for the single-link case, i.e.,
ûΔi (tˆ − τ̂ )
yΔi (tˆ ) = (11.72)
nΔ,0,i + ˆ
j ∈M ΓΔi,j ûΔj (t − τ̂ )
i
where ûΔj ∈ Ωj , mΔ,0,i ∈ Ωni , and ΓΔ ∈ ΩS , where Ωj , Ωni and ΩS are defined
in (11.66), (11.67), and (11.69), respectively. Note that we use ΓΔ ∈ ΩS for the
single-link case rather than ΓΔ ∈ ΩΓ , where ΩΓ is defined in (11.65).
Substitute (11.72) into (11.71), where yin,i = yΔi (tˆ ) and μin,i = μ(tˆ − τ̂ b ) to
obtain the closed-loop uncertain system with time delays,
#
βi Γi,i
ε û˙ i (tˆ ) = ρi + − 1 ûi (tˆ − τ̂ )
μ(tˆ − τ̂ b ) ai
*
1 ûi (tˆ − τ̂ )
− nΔ,0,i + ΓΔi,j ûΔj (tˆ − τ̂ ) (11.73)
ai ûΔi (tˆ − τ̂ )
j ∈Mi
Thus, (11.73) and (11.74) represent the uncertain closed-loop system with time de-
lay for a single link. Next, we use the fact that ûΔj = ûj + δj ûj = (1 + δj )ûj since
11.5 Robustness and Delay Effects Combined 207
ûj is scalar and make the following substitution, which we state as the following
assumption.
Assumption 11.2 In the denominator of the last term of (11.73), we replace the
uncertainty (1 + δi ) in ûΔi with the worst case situation (1 − q̄i ).
Rewrite (11.75) and (11.74) using x = μ − μ∗ and zi = ûi − u∗i , where μ∗ and u∗i
are the equilibrium points of (11.75) and (11.74)
ẋ = η1row I + diag(δj ) z tˆ − τ̂ f (11.77)
# *
−x(tˆ − τ̂ b )
ε ż = ρ β ∗ − Γ¯Δ I + diag(δj ) z(tˆ − τ̂ ) (11.78)
μ (x(tˆ − τ̂ b ) + μ∗ )
where the function h1 (x(tˆ )) is the isolated root of the RHS of (11.78),
−1 −x(tˆ )
h1 x(tˆ ) = I + diag(δi ) Γ¯Δ−1 β (11.80)
μ∗ (x(tˆ ) + μ∗ )
208 11 Robustness and Delay Effects on Network Games
and I + diag(δi ) is invertible, Γ¯Δ = Γ¯ + diag( (1−1q̄i )ai )ΔΓ , Γ¯ is a matrix with
elements defined in (11.76). For now, we assume that Γ¯Δ is invertible. In fact, we
impose this condition later in Lemma 11.9. Proceeding as in the previous section
we can obtain the reduced and the boundary-layers systems as
−x̂1 (tˆ )
x̂˙1 (tˆ ) = η1row Γ¯Δ−1 β (11.81)
1 (tˆ ) + μ )
μ∗ (x̂ ∗
d ž1
= −ρ Γ¯Δ I + diag(δi ) ž1 (t − τ ) (11.82)
dt
The reduced system (11.81) is scalar and nonlinear with no time delay. Notice that
the input multiplicative uncertainty terms have canceled out. The boundary-layer
system (11.82) is linear with multiple time delays. The two systems are decoupled
from each other.
The following lemma presents the LMI that ensures exponential stability for the
boundary-layer system (11.82). A detailed proof can be found in [147].
where λ̄0 > 0, ki > 0, k1 > 0 with k̄i < k1 for all i,
Mn2 −P A20
H̄1 =
−(AT0 )2 P −S1 + λ̄0 (G∗ )T G∗
P E∗ 1
τ (G )
∗ T −P A0 E ∗
H̄2 =
0 −AT0 (G∗ )T λ̄0 (G∗ )T D ∗
⎛ ⎞
−I (D ∗ )T 0
⎜ ∗ ⎟
H̄3 = ⎝ D −I −G∗ E ∗ ⎠
0 −(E ∗ )T (G∗ )T −λ̄0 (I − (D ∗ )T D ∗ )
1
7 7
Mn2 = P A0 + AT0 P + S1 , E ∗ = q̂E2 k1 A0 diag( q̄i k̄i )
τ
7 √ √
q̂ q̄i q̂ √
∗
G = I diag , D ∗ = 0 k1 diag( q̄i )
k1 k̄i 0 0
One can remark that if we set q̂ = 0 and q̄i = 0 (no uncertainty), and we select
S1 = α̂P , where α̂ is a scalar variable, we can recover the LMI used to derive the
time-delay bound (11.48).
11.6 Notes 209
The following lemma ensures the stability of the reduced system (11.81) via the
adjustment of the channel parameter ai .
One can note that the bound on the channel uncertainty term (1 − q̄i ) is not
present. Only the bound on the system matrix uncertainty q̂ is present. If we elimi-
nate the uncertainty, i.e. q̂ = 0, we exactly recover (11.53).
We now state the main stability theorem for the single-link case based on Lem-
mas 11.11 and 11.12.
Theorem 11.13 Consider the uncertain, singularly perturbed system (11.77) and
(11.78), with uncertainty sets defined in (11.65), (11.66), and (11.67). There exists
an ε ∗ > 0 such that for 0 < ε < ε ∗ the origin is asymptotically stable if (11.83) and
(11.84) hold.
The proof of asymptotic stability can be done using a composite Lyapunov func-
tional based on the reduced and boundary-layer system functionals. For details the
reader is referred to [147] and [144].
11.6 Notes
In this chapter we presented stability conditions for the closed-loop system in the
presence of time delays based on frequency domain and on Lyapunov–Razumikhin
analysis. The time-delay bound (11.48) is very simple in form. A Lyapunov–
Krasovskii analysis which utilizes Lyapunov functionals could also be applied. Lya-
punov functionals significantly complicate the system analysis, and they produce
more complicated stability conditions. However, the Lypaunov–Krasovskii analy-
sis produces less conservative stability conditions than the Lyapunov–Razumikhin
analysis. The reader is referred to [150] for such an approach.
Chapter 12
Games for Routing and Path Coloring
Abstract This chapter provides and overview of routing and path coloring prob-
lems in all-optical networks as noncooperative games. We focus on oblivious pay-
ment functions, that is, functions that charge a player according to its own strategy
only. We review results on the relation between such games and online routing and
path coloring. In particular, these results show that the Price of Anarchy of such
games is lower-bounded by, and in several cases precisely equal to, the competitive
ratio of appropriate modifications of the First Fit algorithm.
12.1 Introduction
In this chapter we present other types of game formulations in optical networks. One
of the problems that can be considered in a game-theory framework has to do with
the problem of routing a number of communication requests in WDM all-optical
networks. Specifically, we focus on routing and wavelength assignment as a class of
problems. Recall that communication requests are carried out by assigning a path in
the network (routing) as well as a transmission wavelength (or color). Wavelength
division multiplexing (WDM) allows several requests to be routed through the same
link(s) of the network, and carry them out simultaneously by assigning a different
wavelength to each request [165, 167].
Given an optical network topology and a set of communication requests another
set of questions arise in addition to optimizing channel performance (e.g. OSNR) as
treated until now. In case when the specific routing of the requests is not given, the
routing and path coloring Routing and Path Coloring (RPC) problem asks for both a
routing and a wavelength (color) assignment minimizing the number of colors. See
Fig. 12.1 for a request routed over two different paths. This RPC problem is also
called the RWA assignment problem [125, 126]. On the other hand, if the routing
of the requests is given, the Path Coloring (PC) or wavelength assignment (WA)
problem asks for the minimum number of colors (wavelengths) required such that
requests sharing a common link are assigned different colors. Other optimization
questions can be stated by introducing additional parameters and constraints [165].
A large number of results have been concentrated on the complexity and approx-
imability questions for these optimization problems [39, 51, 99, 166]. A survey of
early results can be found in [52] and references therein.
denoted by (G, P ), and in this last case the players only have to choose a color for
their paths.
Consider a selfish or noncooperative RPC (PC) game on G where each player i
issues a request ri (a path, respectively), hence we identify a player with a request.
A strategy σi for player i is a pair σi = (Pi , λi ), where Pi is a simple path connect-
ing the endpoints of ri and λi is a color assigned to Pi . In the case of PC game this
strategy is just λi . Let Si denote all possible strategies of player i. Note that for each
player these possible strategies depend on the topology of graph G and the number
of colors allowed.
Under the assumption that the number of colors is restricted to be no more than
|R|, this means that there is a finite number of strategies for each player. Thus a
game with |R| players is defined by a finite set of strategies Si , i = 1, . . . , |R|.
Towards formulating the game, for each player i we define a payoff function
(utility function) Ui , Ui : S1 × · · · × S|R| → N , function that herein maps S1 × · · · ×
S|R| to the integers. Such a payoff (utility) is typically associated with a cost or
payment function Ji , Ji : S1 × · · · × S|R| → N , and here we consider that this done
via the relationship Ui = −Ji , i = 1, . . . , |R|.
The elements of S1 × · · · × S|R| are called action combinations or states [28].
Note that unlike the setup in the previous chapter, e.g., Chap. 7, herein the strategies
are finite as in Chap. 3 and mapping is to the integers (unlike a continuous action
set Ωi for games in Chap. 7).
From now on, for simplicity we consider games where all players have the same
cost function or payment function, J . Also, herein the graph plays an important role
so instead of the set of players we use the graph G in the game notation (see below).
216 12 Games for Routing and Path Coloring
Let S-RPC and S-PC denote the class of Selfish-RPC and class of Selfish-PC
games (pre-determined routing), respectively. Then by G(G, R, J ) we denote a
game in S-RPC with input graph G, set of requests R, and payment function J .
Similarly, by G(G, P , J ) we denote a game in S-PC with input graph G, set of
routed requests P , and payment function J .
For a game G(G, R, J ) (and similarly for a game G(G, P , J )), let a pure-strategy
profile, or simply strategy profile, be defined as σ = {σ1 , . . . , σ|R| }, one strategy for
each player. Let S denote the set of all possible strategy profiles.
A game may not have pure-strategy Nash equilibria but if we extend the game
to include as mixed-strategies all possible distributions on Si , then a mixed-strategy
Nash equilibrium (NE) is always guaranteed to exist [95] (see Theorem 3.15 in
Chap. 3). In order to look at the efficiency (performance) of an NE, let NE(G) denote
the set of all NEs, NE(G) ⊂ S.
Definition 12.3 The Price of Anarchy (PoA) of a game G(G, R, J ) (or G(G, P , J ))
is defined as the worst-case number of colors used in a NE (social cost) divided by
Js∗ , that is,
maxσ ∗ ∈NE(G ) Js (σ ∗ )
Price of Anarchy (PoA) =
Js∗
The Price of Stability (PoS) of a game G(G, R, J ) (or G(G, P , J )) is defined as the
best-case number of colors used in a NE (social cost) over Js∗ , that is,
minσ ∗ ∈NE(G ) Js (σ ∗ )
Price of Stability (PoS) =
Js∗
Definition 12.4 Consider a (routing and) path coloring game G(G, R, J ) (or
G(G, P , J )). A cost or payment function J is called oblivious collision-free if:
(a) it guarantees that in a Nash Equilibrium (NE) no color collisions occur, and
(b) it charges a player (who does not collide with other players) according to the
player’s own strategy only.
Note that condition (a) can be satisfied if one charges a very large amount to play-
ers that use the same color and share links of the network. For simplicity the condi-
tion that guarantees collision-free Nash Equilibria is omitted from the descriptions
of the payment functions, since functions we consider here are oblivious collision-
free.
Let us observe that for any instance of S-RPC with oblivious collision-free pay-
ment function, if σ ∗ is a NE, then the following holds:
Js σ ∗ ≤ |R|
Similarly for any instance of S-PC with oblivious collision-free payment function,
if σ ∗ is a NE, then the following hold:
Js σ ∗ ≤ |P |
Thus
|R|
PoA ≤
Js∗
and, respectively,
|P |
PoA ≤
Js∗
in the S-PC case.
Theorem 12.6 (See [28]) The game G(G, R, J1 ) converges to a Nash equilibrium
(NE) in at most |R|2 moves.
This result shows that a NE is reached in polynomial time. However, the payment
function J1 is unable in general to guarantee good performance. As shown by the
following theorem, there exist graphs G for which the Price of Anarchy can be very
high.
Theorem 12.7 (See [28]) For any pricing function f , there exist G and R such that
the Price of Anarchy of game G(G, R, J1 ) is PoA = |R|.
While negative, this result can be justified as follows. Given a routing, finding a
minimum assignment of wavelengths to satisfy the paths in the routing is equiva-
lent to finding a minimum coloring on the graph induced by the routing. It has been
shown that the problem of coloring a graph G = (Ng , Eg ) cannot be approximated
within |Ng |1/7 [23], and since an NE achieved in polynomial time is an approxi-
mated solution to this graph coloring, this implies that the Price of Anarchy cannot
be better than |R|1/7 .
For the second type of cost function J2 the following two results hold.
Theorem 12.8 (See [28]) The game G(G, R, J2 ) converges to a Nash equilibrium
(NE) in a finite number of moves.
Theorem 12.9 (See [28]) For any pricing function f , there exist G and R such that
the Price of Anarchy of game G(G, R, J2 ) is PoA = |R|.
For the last two types of cost function, J3 and J4 , it can be shown that the asso-
ciated games do not even converge in general [28].
Thus the next logical step is to restrict the graph topology, by considering chains
(point-to-point topologies), trees, or rings, or alternatively restricting the communi-
cation patterns. Indeed the following improved results can be shown.
Theorem 12.10 (See [42]) There exist cost or payment functions inducing conver-
gent games with a Price of Anarchy (PoA) of 25.72 in chains, 51.44 in rings and
O(log(|R|)) in trees, all converging in |R|2 steps.
Finally, it is possible to improve the above result for rings and chains by forcing
the agents to simulate the behavior of the online algorithm proposed by Slusarek
[140]. In particular, the following theorem holds.
Theorem 12.11 (See [42]) There exist cost or payment functions inducing conver-
gent games with a Price of Anarchy (PoA) of 6 in rings and 3 in chains.
220 12 Games for Routing and Path Coloring
Another aspect that can be considered is such a RPC game formulation is related
to the different information levels of local knowledge that players may have for
computing their payments [28]. In the complete level of information each player
knows all other players’ routing and coloring strategies. In the intermediate level
of information each player only knows which colors are used on any edge of the
network, while in the minimal level of information each player knows which colors
are used only on edges along paths that the player can choose. For the complete level
it can be shown that the PoA is 1 in chains and 2 in rings, under cost or payment
functions specifically constructed according to the corresponding algorithms [28].
The existence of Nash equilibria and the complexity of recognizing and comput-
ing a Nash equilibrium (NE) for selfish RPC under several payment functions are
considered also in [53]. These results indicate that recognizing a Nash equilibrium
can be done in polynomial time, when each player pays for its own color, when it
pays for the maximum color used by any other overlapping player and when it pays
for the most loaded edge that it uses. On the other hand, when the player pays for
all the different colors appearing along its path, recognizing a Nash equilibrium is
NP-complete.
each request ri the path Pi and color λi that minimize J (ri , P, λ), assuming that
ties are broken arbitrarily. Notice that here we have added the two arguments P, λ
in addition to r in the notation of the payment function to indicate this dependency.
For example, the standard First-Fit for RPC described above can be seen as
FF(J ), where J (ri , P, λ) = λ if P does not overlap with any path of color λ, oth-
erwise J (r, P, λ) = ∞. A similar generalization of First-Fit can be defined for PC
[86]. Note that in the above descriptions, formally for each payment function J ,
the path-color (or just wavelength (color)) assignment for the previous requests rj ,
j < i, should also appear as argument of function J .
The relation between selfish routing and coloring (S-RPC) games and the corre-
sponding online (centralized) RPC problems is shown in the following two results.
Lemma 12.12 (See [86]) Consider a game G(G, R, J ) in S-RPC (S-PC) where J is
an oblivious collision-free payment function. For any ordering R of R, an execution
of FF(J ) algorithm on (G, R) gives a strategy profile which is a Nash Equilibrium
(NE) for G(G, R, J ).
Lemma 12.13 (See [28]) Consider a game G(G, R, J ) in S-RPC (S-PC) where J
is collision-free and non-decreasing on the players’ color (hence also oblivious).
For every strategy profile σ ∗ that is a Nash Equilibrium for G(G, R, J ), there is
an ordering R of R such that there is an execution of FF(J ) algorithm on (G, R)
yielding the same path-color assignment to R as σ ∗ .
This result gives the explicit relationship of an NE to an online solution via FF.
Moreover, any (routing and) coloring solution of RPC (PC) can be converted to a
Nash Equilibrium (NE) for the corresponding game in S-RPC (S-PC, respectively),
with at most the same number of colors, via the so called Nash conversion algorithm
[86]. This leads to the following result.
function of the players’ color, is equal to the competitive ratio of First-Fit for
RPC (PC, respectively).
3. The Price of Stability (PoS) for any game G(G, R, J ) (G(G, P , J ), respectively)
in S-RPC (S-PC), where J is oblivious collision-free and is a non-decreasing
function of the players’ color, is equal to 1.
For the full proof, see [86]. Here we give a short summary.
Proof For 1, by Lemma 12.12, each execution of FF(J ) leads to a path-color (wave-
length) assignment which is a Nash equilibrium (NE) for a game G(G, R, J ). Hence
the social cost Js of that NE is equal to the number of colors used by FF(J ). Di-
viding by Js∗ we get the claim. For 2, let σ ∗ be an NE of the highest social cost.
By Lemma 12.13, it follows that there is an execution of FF(J ) on the correspond-
ing RPC (PC) instance that requires the same number of colors as σ ∗ . Dividing by
Js∗ we find that the competitive ratio of FF(J ) is at least as large as the Price of
Anarchy. Claim 2 follows by combining this with 1. Part 3 can be shown by consid-
ering the optimal coloring and converting it to an NE by using the Nash conversion
algorithm.
Some specific cost functions can be defined based on which some refinements of
the PoA results be obtained [86].
Specifically consider the cost or payment function J (r, P, λ) = λ · N +
length(P), where N is the number of nodes in the ring. Under this function a player
r always selects the smallest possible color even if it requires to follow the longest
one of its two possible alternative paths.
Theorem 12.15 (See [86]) The cost function J (r, P, λ) = λ · N + length(P) used
|R|
for S-RPC games in rings leads to a Price of Anarchy (PoA) equal to 2J ∗ + 1, where
s
R is the given set of requests.
Even if this function does not achieve a low PoA, its PoA is half of that of any
payment function that charges according to the value of the color only (see results
above from [26, 28]).
Next, consider the cost or payment function J (r, P, λ) = length(P) · |R| + λ,
where R is the given set of requests. Under this function a player r always selects
the shortest one of its two possible alternative paths even if it requires to take a larger
color. For this case the following result holds.
Theorem 12.16 (See [86]) The cost function J (r, P, λ) = length(P) · |R| + λ used
for S-RPC games in rings leads to a Price of Anarchy (PoA) such that FF chain + 1 ≤
P oA ≤ 5.06 log N + 10, where N is the number of nodes in the ring.
As seen above, this second payment function (which favors shortest path)
achieves a PoA which does not depend on the number of requests but only on the
number of nodes of the ring (logarithmically). It is still an open question whether
12.5 Notes 223
this upper bound can be further improved. Also worth mentioning is the fact that all
these functions require only local color information, namely to know which colors
are used along edges that can be used by a player (minimal level of information
according to the classification in [28] and [27]).
12.5 Notes
We presented an overview of a class of games for routing and path coloring prob-
lems in all-optical networks. We focused on games where each routing request (pair
of source-destination nodes) is a player, and a strategy of a player consists of a path
from the source to the destination and a wavelength/frequency (color), hence in the
class of finite strategy games. We reviewed results on the relation between such
games and online routing and path coloring. Other approaches are those in which
the ISPs act as brokers or players between the TSP and the end user [162].
Chapter 13
Summary and Conclusions
The goal of this monograph has been to study optical networks as examples of com-
munication networks and to understand how control algorithms can be designed
for optical networks from a game theoretic perspective. From a game-theory and
control perspective, there are a multitude of problems to be tackled in optical net-
works and the field is still in its infancy. An important problem is how to control
channel transmission performance in optical networks while incorporating physical
layer impairments. Of particular interest have been the OSNR optimization prob-
lems with link capacity constraints and/or channel OSNR targets. Game-theoretic
approaches have been introduced to study such problems.
An optical network can be considered as a dynamic multiuser environment, in
which signal of a channel is regarded as interfering for all others on the same fiber
link. Power control is a key issue in designing an interference-limited multiuser
communication network system. A major portion of this monograph has been fo-
cused on this first type of problems by using game-theoretic approaches. Another
part of the book has discussed game-theoretic approaches for other challenging
problems such as routing and wavelength assignment (path coloring) (RWA) or
(RPC) in optical networks [99, 125, 165].
Throughout the book our focus has been on analysis of dynamic systems arising
from game formulations with non-separable player utilities and with coupled as well
as propagated (modified) constraints, in the class of continuous strategy games. It
turns out that this is the case for a large class of games in optical networks, where
the utility of each player is affected by decisions of other players, and action sets
are no longer orthogonal (Chaps. 7 and 8).
The monograph has been organized as follows. Chapter 2 has provided an
overview of basic concepts in game theory, mostly focused on noncooperative
(Nash) game theory. Chapter 3 focused on matrix games, while Chap. 4 on games
with continuous action sets and cost functions. Chapter 5 presented some relatively
new theoretical results for continuous Nash games with coupled constraints, i.e.,
coupled action sets. Chapter 6 presented an overview of basic background on trans-
mission in optical networks and on general topologies. The concepts of OSNR and
link power capacity constraint in optical networks have been introduced. In Chap. 7
of a fiber link composed of cascaded dynamical optical elements makes the opti-
cal network dynamics an important aspect both at the data layer and the physical
layer. Such a problem has been addressed in [114]. In mesh network topologies,
channel routes can be changed by optical switches and closed loop structures can
be formed, for example, quasi-ring structures in Chap. 8. Strictly speaking, time
delay and asynchronism are aspects of dynamics. The material presented in this
monograph considers the network after reconfiguration. In other words, system pa-
rameters are stationary between any updates. From the dynamical point of view,
study of the scalability of algorithms is an open direction.
Finally I would like to end on the same note mentioned in the introduction: I be-
lieve there is a lot of scope for fundamental research in optical networks from a
control- and game-theoretic perspective. Hopefully this book will provide the moti-
vation for many other developments and much research in this area.
Appendix A
Supplementary Material
A.1 Notations
We present below some notations used throughout the monograph.
Let a = [ai ] and b = [bi ] be n-dimensional vectors. We write a ≥ b if all ai ≥ bi
and a > b if all ai > bi . We use diag(a) to denote a diagonal matrix whose diag-
onal entries are elements of vector a = [ai ]. A superscript T denotes the transpose
operation.
For a twice continuously differentiable function f : Rm → R, its gradient at u,
denoted by ∇f (u), is defined as
⎡ ⎤
∂f (u)
⎢ ∂u1 ⎥
∇f (u) = ⎣ · · · ⎦ , (A.1)
∂f (u)
∂um
with elements from the diagonal of the Jacobian matrix of f. The Jacobian of ∇f(u)
2
with respect to u is denoted by ∇ f(u),
⎡ ∂ 2 f (u) ∂ 2 f (u) ∂ 2 f1 (u)
⎤
1
2 ∂u
1
∂u · · · ∂u ∂u
⎢ 2∂u1 1 2 1 m
⎥
⎢ ∂ f2 (u) ∂ 2 f2 (u) ∂ 2 f2 (u) ⎥
⎢ · · · ∂u2 ∂um ⎥
∇ f(u) := ⎢ ⎥
2 ∂u2 ∂u1 ∂u22
⎢ . (A.4)
⎢ .. .. .. .. ⎥ ⎥
⎣ . . . ⎦
∂ 2 fm (u) ∂ 2 fm (u) ∂ 2 fm (u)
∂um ∂u1 ∂um ∂u2 ··· ∂u2m
where ∇u f (u; x) and ∇x f (u; x) are the gradients with respect to the first argument
u = [u1 , . . . , um ] and the second argument x = [x1 , . . . , xm ], respectively, as defined
in (A.1).
min J0 (u)
(A.6)
subject to u ∈ Ω,
and uopt is called strict if the inequality above is strict for u = uopt .
The following is a standard result (combining Proposition 1.1.1 and Proposi-
tion 1.1.2 in [24]) regarding the solution of (A.6).
A.2 Standard Optimization Review 231
min J0 (u)
(A.9)
subject to gr (u) ≤ 0, r = 1, . . . , R,
If r ∈
/ A(u), the constraint gr (u) is inactive at u. A feasible vector u is said to be
regular if the active inequality constraint gradients ∇gr (u), r ∈ A(u), are linear
independent.
The Lagrangian function L : Rm+R → R for problem (A.9) is defined as
R
L(u, μ) = J0 (u) + μr gr (u), (A.11)
r=1
min J0 (u)
subject to gr (u) ≤ 0, r = 1, . . . , R, (A.15)
u ∈ U,
Proposition A.4 (General Sufficiency Condition) Consider the problem (A.15). Let
uopt be a feasible vector, together with a vector μ∗ = [μ∗1 , . . . , μ∗R ]T that satisfies
μ∗r ≥ 0, ∀r = 1, . . . , R
μ∗r = 0, ∀r ∈
/ A uopt
and assume uopt minimizes the Lagrangian function L(u, μ∗ ) (A.11) over u ∈ U ,
denoted as
uopt ∈ arg min L u, μ∗ (A.16)
u∈U
m
|aii | ≥ |aij |, ∀i = 1, . . . , m
j =1,j =i
m
|aii | > |aij |, ∀i = 1, . . . , m
j =1,j =i
Some useful results are shown in the following theorem, adapted from [59].
Theorem A.7 (See [59], p. 349) Let the m × m matrix A = [aij ] be strictly diago-
nally dominant. Then A is invertible and
(a) If all main diagonal entries of A are positive, then all the eigenvalues of A have
positive real part.
(b) If A is Hermitian and all main diagonal entries of A are positive, then all the
eigenvalues of A are real and positive.
Lemma A.8 Let A be an m × m real matrix with all main diagonal entries positive.
Then A is positive definite if A and AT are both strictly diagonally dominant.
Proof If A and AT are both strictly diagonally dominant, then it follows that
m
m
aii > |aij | and aii > |aj i |
j =1,j =i j =1,j =i
Thus As is also strictly diagonally dominant. From Theorem A.7, it follows that As
is positive definite. Therefore A is positive definite in the sense that the symmetric
part, 12 As , is positive definite.
Definition A.9 A square matrix is called a Z-matrix if all off-diagonal entries are
less than or equal to zero.
Definition A.10 (See [43], Theorem 5.1) Let the m × m matrix A = [aij ] be a Z-
matrix. A is called an M-matrix if it satisfies any one of the following conditions.
234 A Supplementary Material
Proof The proof for the first part can be found in [43] where Theorem
5.1kis proved.
In the case when λ = 1, ρ(B) < 1 leads to the convergence of ∞ k=0 B . Further-
more,
∞
B k (I − B) = I + B + B 2 + · · · (I − B) = I
k=0
∞
Thus it follows that (I − B)−1 = k=0 B
k.
Definition A.12 Let S be a (nonempty) subset of Rn . The set of all nonempty sub-
sets of S is called the power set of S, denoted by P(S).
is a continuous function;
2. Φ ∗ : X → Y with
Φ ∗ := arg max f (x, y) : y ∈ Φ(x) = y ∈ Φ(x) : f (x, y) = f ∗ (x)
Theorem A.19 (See [20], Kakutani Fixed-point Theorem) Let S be a compact and
convex subset of Rn , and let f be an upper-semi-continuous correspondence which
assigns to each x ∈ S a closed and convex subset of S. Then there exists some x ∈ S
such that x ∈ f (x).
236 A Supplementary Material
Theorem A.20 (See [25], Projection Theorem) Let X be a nonempty, closed, and
convex subset of Rn .
1. For every x ∈ Rn , there exists a unique z ∈ X that minimizes z − x over all
z ∈ X and is denoted by [x]+ .
2. Given some x ∈ Rn , a vector x ∗ ∈ X is equal to [x]+ if and only if
T
z − x ∗ x − x ∗ ≤ 0, ∀z ∈ X
Proposition A.23 (Proposition 5.1, [25], p. 267) Let γ be a positive scalar and let
G be a symmetric positive definite matrix. A vector x ∗ is a solution of VI(X, f ) if
and only if
∗
+
x − γ G−1 f x ∗ G = x ∗ ,
where [·]+
G is the projection on X with respect to norm xG = (x Gx)
T 1/2 .
Proposition A.24 (Proposition 5.2, [25], p. 268) Suppose that X is compact and
that f : Rn → Rn is continuous. Then there exists a solution to VI(X, f ).
ˆ
ν̇ = dν(
d tˆ
t)
. The system (A.18) is defined on two different time-scales, one fast, and
one slow.
We describe (A.18) in further detail as follows. We denote by tˆ the “slow” time-
scale, and we denote by t the “fast” time-scale. The two time-scales are related to
each other by tˆ = εt. A nonlinear system that is in singular perturbation form, as
presented in (A.18), may be time-decoupled into two subsystems that are easier to
analyze than the full nonlinear system. One subsystem is defined on the slow time-
scale, and the other subsystem is defined on the fast time-scale. The system on the
slow time-scale is called the reduced system, while the system on the fast time-scale
is called the boundary-layer system. A Lyapunov stability analysis may be applied
to the reduced and boundary-layer systems separately. The Lyapunov functions of
the reduced and boundary-layer systems may be combined into a composite Lya-
punov function that applies to the full, singularly perturbed system. This composite
Lyapunov function can be used to prove the stability of the full system.
Next we show how to obtain the reduced system and the boundary-layer system
of (A.18) as in [69]. Assume there is at least one isolated root of 0 = g(x, z) and
let z = h(x) be one such root that vanishes at x = 0. For convenience, we define the
coordinate shift ẑ = z − h(x). Thus, the coordinate shifted system becomes
ẋ = f x, ẑ + h(x) (A.19)
∂h
ε ẑ˙ = g x, ẑ + h(x) − ε f x, ẑ + h(x) (A.20)
∂x
where x and ẑ are functions of the slow time variable tˆ. Note that the ε term appears
in the second equation because it is written with respect to the slow time variable, tˆ.
If we rewrite this equation as a function of the fast time variable t, the ε term disap-
pears.
The reduced model is derived from (A.19) by setting ẑ = 0, which is equivalent
˙ converge instantly. This produces the following
to assuming the fast dynamics, ε ẑ,
reduced system:
ẋ = f x, h(x)
Notice that this system is on the slow time-scale, tˆ.
The boundary-layer system is derived from (A.20). We rewrite (A.20) with re-
spect to the fast time-scale, t, as follows. Let t = εtˆ . Note that setting ε = 0 freezes
the tˆ and x(tˆ) variables with respect to the fast time-scale t so that tˆ and x appear as
constants. Alternatively, we can think of the variables on the t time-scale converg-
ing instantaneously, since t = εtˆ . We introduce the notation tˆ(t) := εt for the scaling
by ε. Thus, we define z̃(t) = ẑ(tˆ(t)) = (ẑ ◦ tˆ)(t), where ◦ denotes composition and
y = ẑ ◦ tˆ. Thus, when we set ε = 0 in (A.20), and rewrite it with respect to the t
time-scale, we obtain
d z̃(t)
= g x(tˆ), z̃(t) + h x(tˆ)
dt
A.9 Singular Perturbation Theory 239
where d z̃
dt = ε ddẑtˆ . Thus, we define the reduced and boundary-layer systems as fol-
lows.
Definition A.25 Consider the system (A.18). Let the reduced system be defined as
ẋ = f x(tˆ), h x(tˆ)
d z̃(t)
= g x(tˆ), z̃(t) + h x(tˆ)
dt
where h(x) is the isolated root of the RHS of (A.18) that vanishes at x = 0, t and
tˆ are the fast and slow time-scales in the system, respectively, z̃ = ẑ ◦ tˆ and ẑ =
z − h(x) is the coordinate shift.
v̇ = f (t, v) (A.21)
and globally exponentially stable if (A.22) is satisfied for any initial state v(t0 ).
We now introduce the formal singular perturbation theorem for analyzing the
stability of (A.18). The following result is Theorem 11.4 simplified from [69] and
restated as Theorem A.27. Assume that (A.18) satisfies all necessary assumptions
outlined below.
The high level proof of Theorem A.27 proceeds as follows. First, a Lyapunov
function for each of the reduced and boundary-layer systems is found. Next, the
two functions are summed together to produce a composite Lyapunov function that
is applied to the full system, (A.23)–(A.24). Using the properties of the Lyapunov
functions, as well as bounds on f , g, and h around the origin, exponential stability
is proved. The details of the proof of Theorem A.27 are found in [69].
A.11.1 Definitions
We define C([−τ, 0], n ) as the set of all continuous functions mapping [−τ, 0] to
n . Let C = C([−τ, 0], n ). The general form of a retarded functional differential
equation is defined as
dx
= f (t, xt ) (A.25)
dt
where xt ∈ C is the function mapping [t − τ, t] to n , and f : × C → n .
The following norm is used in the definition of stability for time-delay systems.
A.11 Time-Delay Stability Theory 241
Definition A.30 The origin of (A.25) is exponentially stable if there exist positive
real numbers h, a, and b such that for every initial condition φ ∈ Ch the solution
xt (φ) of (A.25) exists for all t ≥ 0 and furthermore satisfies
) )
)xt (φ)) ≤ ae−bt φc
c
where x(t), τ, a0 , a1 ∈ . The initial condition for (A.26) is defined as the function
φ : [−τ, 0] → such that
The Laplace transform of (A.26) with initial conditions φ exists, and may be written
as follows
# 0 *
1
X(s) = φ(0) + a1 e−s(v+τ ) φ(v) dv (A.28)
Δ(s) −τ
where
Δ(s) = s − a0 − a1 e−τ s (A.29)
242 A Supplementary Material
is the characteristic quasipolynomial of (A.26). Let Φ(t) denote the inverse Laplace
transform of 1/Δ(s). We call Φ(t) the fundamental solution of (A.26). Notice that
with the initial condition φ(0) = 1, and φ(v) = 0 for all v ∈ [−τ, 0), then Φ(t) is
the solution of (A.26). The growth of the general solution of (A.26) is related to the
exponential growth of the fundamental solution Φ(t), which is determined by the
poles of the system (A.28), i.e., the solutions of the characteristic equation
Δ(s) = 0 (A.30)
In general, (A.30) has an infinite number of solutions. Denote by Re[s] the real com-
ponent of the complex value s. The following proposition from [55] gives conditions
for stability of (A.26).
Proposition A.31 For any α ∈ , there are a finite number of poles with real parts
greater than α. Let si for i = 1, 2, . . . be the poles of the system (A.26), i.e., the
solutions of (A.30), and let
α0 = max Re[si ] (A.31)
i
Then for any α > α0 , there exists a L > 0 such that the solution of (A.26) with the
initial condition (A.27) satisfies the inequality
x(t) ≤ Leαt φc (A.32)
K
ẋ(t) = A0 x(t) + Ak x(t − τk ), τk ≥ 0 (A.33)
k=1
where
K
Δ(s) = sI − A0 − Ak e−τk s
k=1
We say (A.33) is asymptotically stable if and only if p(s; e−τ1 s , . . . , e−τK s ) has
no root in the closed RHP, or C̄+ . This definition of stability is stated formally as
follows [55].
A.11 Time-Delay Stability Theory 243
Definition A.32 The system (A.33) is said to be asymptotically stable if and only
if its characteristic quasipolynomial (A.34) satisfies
p s; e−τ1 s , . . . , e−τK s = 0, ∀s ∈ C̄+ (A.35)
dx(t)
= A0 x(t) + A1 x(t − τ ) (A.36)
dt
where A0 and A1 are (n × n) real matrices.
The following theorem, known as the restricted Lyapunov–Razumikhin theorem
[55] (Proposition 5.1, p. 149) gives conditions for asymptotic stability.
Theorem A.33 The time-delay system (A.36) with maximum time-delay τ is asymp-
totically stable if there exists a quadratic Lyapunov function V : n → such that
for some ψ1 > 0, ψ2 > 0 it satisfies
dV (x)
where the time derivative along the system trajectory dt satisfies
dV (x) ) )2
≤ −ψ1 )x(t))2 (A.37)
dt
if ∀p > 1, ∃t, ∀ξ ∈ [−τ, 0] such that
V x(t + ξ ) ≤ pV x(t)
Notice that Theorem A.33 closely resembles the Lyapunov stability theo-
rem [69]. The first condition is the same as in the Lyapunov stability theorem.
The second condition requires that the time derivative of the Lyapunov function
be non-positive which is also similar to the Lyapunov stability theorem. The third
requirement is an extra condition that only appears for time-delayed systems. We
244 A Supplementary Material
explain this extra condition as follows. The negation of the third condition of Theo-
rem A.33 is
∃p > 1, ∀t, ∃ξ ∈ [−τ, 0] V x(t + ξ ) > pV x(t) (A.38)
Thus, for some p > 1, at every instant of time, t, the value V (x(t)) is less than a
past instant. Intuitively, this means the state trajectory is approaching its equilibrium
point, and hence, the system is stable. Thus, we only need to ensure the first two
conditions of Theorem A.33 are satisfied if the third condition holds. If the third
condition does not hold, then the system is implicitly stable. An interesting note is
that the Lyapunov function and its time derivative are compared to the square of the
norm of the state, which implies exponential stability.
Theorem A.33 presents a general set of stability conditions that do not exploit the
structure of (A.36), i.e., A0 and A1 do not appear in the stability criteria. In addition,
it is not obvious how to pick a Lyapunov function to satisfy the stability criteria in
Theorem A.33. We show that in the foregoing work, the Lyapunov function V =
x T P x may be used.
The remainder of this section is dedicated to delay-independent and delay-
dependent stability criteria specific to the structure of the linear system (A.36). We
first outline time-delay-independent stability conditions, which provide the simplest
and most conservative stability conditions (Proposition 5.3 in [55]). These are pro-
vided below as Proposition A.34.
Proposition A.34 The system (A.36) is asymptotically stable if there exist a scalar
α̂ > 0 and a real symmetric matrix P such that
P A0 + AT0 P + α̂P P A1
<0 (A.39)
AT1 P −α̂P
If A0 = 0 then Proposition A.34 cannot be directly applied since this does not
produces a feasible solution. Thus, more general time-delay-dependent conditions
need to be applied. Time-delay-dependent stability conditions produce less con-
servative stability criteria. The following theorem presents a set of time-delay-
dependent conditions to ensure asymptotic stability of (A.36) (Corollary 5.8 in [55]).
Theorem A.35 The system (A.36) is asymptotically stable if there exist a real sym-
metric matrix P and real scalars α̂0 > 0, α̂1 > 0 such that
⎛ ⎞
M −P A1 A0 −P A21
⎜ ⎟
⎜ −AT AT P −α̂0 P 0 ⎟
⎝ 0 1 ⎠<0 (A.40)
−(A21 )T P 0 −α̂1 P
where
1
M= P (A0 + A1 ) + (A0 + A1 )T P + (α̂0 + α̂1 )P (A.41)
τ
A.11 Time-Delay Stability Theory 245
1. Acemoglu, D., Ozaglar, A.: Costs of competition in general networks. In: Proceedings of the
44th IEEE Conference on Decision and Control, December, pp. 5324–5329 (2005)
2. Agrawal, G.P.: Fiber-Optic Communication Systems, 3rd edn. Wiley, New York (2002)
3. Alpcan, T.: Noncooperative games for control of networked systems. Ph.D. thesis, University
of Illinois at Urbana–Champaign, Illinois, USA (2006)
4. Alpcan, T., Basar, T.: A game-theoretic framework for congestion control in general topology
networks. In: Proceedings of the 41st IEEE Conference on Decision and Control, December,
pp. 1218–1224 (2002)
5. Alpcan, T., Basar, T.: Global stability analysis of an end-to-end congestion control. In: Pro-
ceedings of the 42nd IEEE Conference on Decision and Control, December, pp. 1092–1097
(2003)
6. Alpcan, T., Basar, T.: A hybrid systems model for power control in multicell wireless data
networks. Perform. Eval. 57(4), 477–495 (2004)
7. Alpcan, T., Basar, T.: Distributed algorithms for Nash equilibria of flow control games. Adv.
Dyn. Games: Appl. Econ. Finance Optim. Stoch. Control Ann. Dynam. Games 7, 473–498
(2005)
8. Alpcan, T., Basar, T., Srikant, R., Altman, E.: CDMA uplink power control as a noncoopera-
tive game. In: Proceedings of the 40th IEEE Conference on Decision and Control, December,
pp. 197–202 (2001)
9. Alpcan, T., Fan, X., Basar, T., Arcak, M., Wen, J.T.: Power control for multicell CDMA
wireless networks: a team optimization approach. Wirel. Netw. 14(5), 647–657 (2008)
10. Altman, E., Altman, Z.: S-modular games and power control in wireless networks. IEEE
Trans. Autom. Control 48(5), 839–842 (2003)
11. Altman, E., Basar, T.: Multiuser rate-based flow control. IEEE Trans. Commun. 46(7), 940–
949 (1998)
12. Altman, E., Basar, T., Jimenez, T., Shimkin, N.: Routing into two parallel links: game-
theoretic distributed algorithms. J. Parallel Distrib. Comput. 61(9), 1367–1381 (2001)
13. Altman, E., Basar, T., Srikant, R.: Nash equilibria for combined flow control and routing
in networks: asymptotic behavior for a large number of users. IEEE Trans. Autom. Control
47(6), 917–930 (2002)
14. Altman, E., Boulogne, T., El-Azouziand, R., Jimenez, T., Wynter, L.: A survey on networking
games in telecommunications. Comput. Oper. Res. 33(2), 286–311 (2006)
15. Altman, E., Solan, E.: Constrained games: the impact of the attitude to adversary’s con-
straints. IEEE Trans. Autom. Control 54(10), 2435–2440 (2009)
16. Arrow, K.J., Debreu, G.: Existence of an equilibrium for a competitive economy. Economet-
rica 22(3), 265–290 (1954)
17. EI Azouzi, R., Altman, E.: Constrained traffic equilibrium in routing. IEEE Transactions on
Automatic Control 48(9), 1656–1660 (2003)
18. Bala, K., Brackett, C.: Cycles in wavelength routed optical networks. IEEE J. Lightwave
Technol. 14(7), 1585–1594 (1996)
19. Banerjee, D., Mukherjee, B.: A practical approach for routing and wavelength assignment
in large wavelength-routed optical networks. IEEE J. Sel. Areas Commun. 14(5), 903–908
(1996)
20. Basar, T., Olsder, G.J.: In: Dynamic Noncooperative Game Theory. SIAM Series Classics in
Applied Mathematics, 2nd edn. (1999)
21. Basar, T., Srikant, R.: Revenue-maximizing pricing and capacity expansion in a many-users
regime. In: Proceedings of the 21st IEEE Conference on Computer Communications (INFO-
COM), June 2002, pp. 294–301 (2002)
22. Bazarra, M.S., Sherali, H.D., Shetty, C.M.: In: Nonlinear programming: theory and algo-
rithms, 3rd edn. Wiley-Interscience, New York (2006)
23. Bellare, M., Goldreich, O., Sudan, M.: Free bits, PCPs and non-approximability—towards
tight results. SIAM J. Comput. 27, 804–915 (1998)
24. Bertsekas, D.P.: Nonlinear Programming, 2nd edn. Athena Scientific, Nashua (1999)
25. Bertsekas, D.P., Tsitsiklis, J.N.: Parallel and Distributed Computation: Numerical Methods.
Prentice Hall, New York (1989)
26. Bilo, V., Flammini, M., Moscardelli, L.: On Nash equilibria in non-cooperative all-optical
networks. In: Diekert, V., Durand, B. (eds.) STACS 2005. LNCS, vol. 3404, pp. 448–459.
Springer, Berlin (2005)
27. Bilo, V., Flammini, M., Moscardelli, L.: On Nash Equilibria in Non-cooperative All-Optical
Networks. Optical Switching and Networking (2010)
28. Bilo, V., Moscardelli, L.: The price of anarchy in all-optical networks. In: S’ykora, R.O. (ed.)
SIROCCO 2004. LNCS, vol. 3104, pp. 13–22. Springer, Heidelberg (2004)
29. Borella, M., Jue, J., Banerjee, D., Ramamurthy, B., Mukherjee, B.: Optical components for
WDM lightwave networks. Proc. IEEE 85(8), 1274–1307 (1997)
30. Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge
(2004)
31. Brown, G.W.: Activity Analysis of Production and Allocation. Wiley, New York (1951)
32. Charnes, A., Eisner, M., Kortanek, K.O.: On weakly balanced games and duality theory. Cah.
Cent. Etud. Rech. Oper. 12, 7–21 (1970)
33. Chraplyvy, A.: Limitations on lightwave communications imposed by optical-fiber nonlin-
earities. IEEE/OSA J. Lightwave Technol. 8(10), 1548–1557 (1990)
34. Chraplyvy, A., Nagel, J., Tkach, R.: Equalization in amplified WDM lightwave transmission
systems. IEEE Photonics Technol. Lett. 4(8), 920–922 (1992)
35. de Meyer, B.: Repeated games, duality and the central limit theorem. Math. Oper. Res. 21(1),
237–251 (1996)
36. de Meyer, B., Marino, A.: Duality and optimal strategies in the finitely repeated zero-sum
games with incomplete information on both sides. In: Cahiers de la Maison des Sciences
Economiques, pp. 1–7 (2005)
37. Debreu, G.: A social equilibrium existence theorem. Proc. Natl. Acad. Sci. USA 38(10),
886–893 (1952)
38. Dubey, P.: Inefficiency of Nash equilibria. Math. Oper. Res. 11(1), 1–8 (1986)
39. Erlebach, T., Jansen, K., Kaklamanis, C., Persiano, P.: Optimal wavelength routing on di-
rected fiber trees. Theor. Comput. Sci. 221(1–2), 119–137 (1999)
40. Facchinei, F., Fischer, A., Piccialli, V.: On generalized Nash games and variational inequali-
ties. Oper. Res. Lett. 35(2), 159–164 (2007)
41. Falomari, D., Mandayam, N., Goodman, D.: A new framework for power control in wireless
data networks: games utility and pricing. In: Proceedings of the Allerton Conference on
Communication, Control, and Computing, pp. 546–555 (1998)
42. Fanelli, A., et al.: Game theoretical issues in optical networks. In: Proceedings of ÊICTON
2006, Sep. (2006)
References 253
43. Fiedler, M.: Special Matrices and Their Applications in Numerical Mathematics. Martinus
Nijhoff Publishers, Leiden (1986)
44. Fishman, G.S.: Monte Carlo: Concepts, Algorithms, and Applications. Springer, Berlin
(1996)
45. Forghieri, F., Tkach, R.W., Favin, D.L.: Simple model of optical amplifier chains to evaluate
penalties in WDM systems. IEEE/OSA J. Lightwave Technol. 16(9), 1570–1576 (1998)
46. Foschini, G., Miljanic, Z.: A simple distributed autonomous power control algorithm and its
convergence. IEEE Trans. Veh. Technol. 42(4), 641–646 (1993)
47. Friedman, D.: Evolutionary games in economics. Econometrica 59, 637–666 (1991)
48. Fudenberg, D., Kreps, D.M.: Learning mixed equilibria. Games Econ. Behav. 5, 320–367
(1993)
49. Fudenberg, D., Levine, D.K.: Theory of Learning in Games. MIT Press, Cambridge (1998)
50. Fudenberg, D., Tirole, J.: Game Theory, 2nd edn. MIT Press, Cambridge (1991)
51. Garey, M.R., Johnson, D.S., Miller, G.L., Papadimitriou, C.H.: The complexity of coloring
circular arcs and chords. SIAM J. Algebr. Discrete Methods 1(2), 216–227 (1980)
52. Gargano, L., Vaccaro, U.: Routing in all-optical networks: algorithmic and graph theoretic
problems. In: Numbers, Information and Complexity, pp. 555–578. Kluwer Academic, Dor-
drecht (2000)
53. Georgakopoulos, G.F., Kavvadias, D.J., Sioutis, L.G.: Nash equilibria in all-optical networks.
Discrete Math. 309(13), 4332–4342 (2009)
54. Greenwald, A., Friedman, E.J., Shenker, S.: Learning in network contexts: experimental re-
sults from simulations. Games Econ. Behav. 35, 80–123 (2001)
55. Gu, K., Kharitonov, V.L., Chen, J.: Stability of Time-Delay Systems. Birkhäuser, Boston
(2003)
56. Harker, P.T.: Generalized Nash games and quasi-variational inequalities. Eur. J. Oper. Res.
54(1), 81–94 (1991)
57. Harsanyi, J.C.: Games with randomly disturbed payoffs: a new rationale for mixed-strategy
equilibrium points. Int. J. Game Theory 2, 1–23 (1973)
58. Hofbauer, J., Sigmund, K.: Evolutionary Games and Population Dynamics. Cambridge Uni-
versity Press, Cambridge (1998)
59. Horn, R.A., Johnson, C.R.: Matrix Analysis. Cambridge University Press, Cambridge (1999)
60. Huang, M., Malhame, R.P., Caines, P.E.: Nash certainty equivalence in large population
stochastic dynamic games: connections with the physics of interacting particle systems. In:
Proceedings of the 45th IEEE Conference on Decision and Control, December, pp. 4921–
4926 (2006)
61. Ji, H., Huang, C.: Non-cooperative uplink power control in cellular radio systems. Wirel.
Netw. 4(3), 233–240 (1998)
62. Jia, X., Du, D., Hu, X., Lee, M., Gu, J.: Optimization of wavelength assignment for QoS
multicast in WDM networks. IEEE Trans. Commun. 49(2), 341–350 (2001)
63. Johari, R., Mannor, S., Tsitsiklis, J.N.: Efficiency loss in a resource allocation game: a single
link in elastic supply. In: Proceedings of the 43rd IEEE Conference on Decision and Control,
December, pp. 4679–4683 (2004)
64. Johari, R., Tsitsiklis, J.: Efficiency loss in a network resource allocation game. Math. Oper.
Res. 29(3), 407–435 (2004)
65. Kandukuri, S., Boyd, S.: Simultaneous rate and power control in multirate multimedia
CDMA systems. In: Proceedings of the 6th IEEE International Symposium on Spread Spec-
trum Techniques and Applications, September, pp. 570–574 (2000)
66. Kandukuri, S., Boyd, S.: Optimal power control in interference-limited fading wireless chan-
nels with outage-probability specifications. IEEE Trans. Wirel. Commun. 1(1), 46–55 (2002)
67. Kelly, F.: Charging and rate control for elastic traffic. Eur. Trans. Telecommun. 8, 33–37
(1997)
68. Kelly, F.P., Maulloo, A.K., Tan, D.: Rate control for communication networks: shadow
prices, proportional fairness and stability. J. Oper. Res. Soc. 49(3), 237–252 (1998)
69. Khalil, H.K.: Nonlinear Systems, 3rd edn. Prentice Hall, New York (2002)
254 References
70. Kinderlehrer, D.: An Introduction to Variational Inequalities and Their Applications. Aca-
demic Press, San Diego (1980)
71. Korpelevich, G.M.: The extragradient method for finding saddle points and other problems.
Èkon. Mat. Metody 12, 747–756 (1976) [translated as Matecon.]
72. Koskie, S., Gajic, Z.: A Nash game algorithm for SIR-based power control in 3G wireless
CDMA networks. IEEE/ACM Trans. Netw. 13(5), 1017–1026 (2005)
73. Koutsoupias, E., Papadimitriou, C.H.: Worst-case equilibria. In: Proceedings of the 16th An-
nual Symposium STACS, pp. 404–413 (1999)
74. Krasovskii, N.N., Subbotin, A.I.: Game-Theoretical Control Problems. Springer, Berlin
(1988)
75. Krishna, V., Sjostrom, T.: On the convergence of fictitious play. Math. Oper. Res. 23, 479–
511 (1998)
76. Kunniyur, S., Srikant, R.: End-to-end congestion control schemes: utility functions, random
losses and ECN marks. IEEE/ACM Trans. Netw. 11(5), 689–702 (2003)
77. Laplante, P.A.: Comprehensive Dictionary of Electrical Engineering. Springer, Berlin (1999)
78. Libman, L., Orda, A.: The designer’s perspective to atomic noncooperative networks.
IEEE/ACM Trans. Netw. 7(6), 875–884 (1999)
79. Loja, A., et al.: Inter-domain routing in multiprovider optical networks: game theory and
simulations. In: Proceedings of Next Gen. Internet Networks, May, pp. 157–164 (2005)
80. Low, S., Paganini, F., Doyle, J.: Internet congestion control. IEEE Control Syst. Mag. 22(1),
28–43 (2002)
81. Low, S.H., Lapsley, D.E.: Optimization flow control-I: basic algorithm and convergence.
IEEE/ACM Trans. Netw. 7(6), 861–874 (1999)
82. Luo, Z., Pang, J.: Analysis of iterative waterfilling algorithm for multiuser power control in
digital subscriber lines. EURASIP J. Appl. Signal Process. 2006, 1–10 (2006)
83. Marbach, P., Berry, R.: Downlink resource allocation and pricing for wireless networks. In:
Proceedings of the 21st IEEE Conference on Computer Communications (INFOCOM), June,
pp. 1470–1479 (2002)
84. Marden, J.R., Arslan, G., Shamma, J.S.: Joint strategy fictitious play with inertia for potential
games. In: Proceedings of the 44th IEEE Conference on Decision and Control, December,
pp. 6692–6697 (2005)
85. Mecozzi, A.: On the optimization of the gain distribution of transmission lines with unequal
amplifier spacing. IEEE Photonics Technol. Lett. 10(7), 1033–1035 (1998)
86. Milis, I., Pagourtzis, A., Potika, K.: Selfish routing and path coloring in all-optical networks.
In: Proceedings of CAAN2007, pp. 71–84 (2007)
87. Mitra, D.: An asynchronous distributed algorithm for power control in cellular radio systems.
In: Proceedings of the 4th WINLAB Workshop on Third Generation Wireless Information
Networks, November, pp. 249–257 (1993)
88. Mokhtar, A., Azizoglu, M.: Adaptive wavelength routing in all-optical networks. IEEE/ACM
Trans. Netw. 6(2), 197–206 (1998)
89. Monderer, D., Shapley, L.S.: Potential games. Games Econ. Behav. 14(1), 124–143 (1996)
90. Morgan, J., Romaniello, M.: Generalized quasi-variational inequalities and duality. J. In-
equal. Pure Appl. Math. 4(2), Article 28, 1–7 (2003)
91. Morgan, J., Romaniello, M.: Generalized quasi-variational inequalities: duality under pertur-
bations. J. Math. Anal. Appl. 324(2), 773–784 (2006)
92. Mukherjee, B.: WDM optical communication networks: progress and challenges. IEEE J.
Sel. Areas Commun. 18(10), 1810–1824 (2000)
93. Myerson, R.: Game Theory: Analysis of Conflict. Harvard University Press, Cambridge
(1991)
94. Mync, H.: Nonnegative Matrices. Wiley, New York (1988)
95. Nash, J.: Equilibrium points in n-person games. Proc. Natl. Acad. Sci. USA 36(1), 48–49
(1950)
96. Nash, J.: Non-cooperative games. Ann. Math. 54(2), 286–295 (1951)
References 255
97. Nisan, N., Roughgarden, T., Tardos, E., Vazirani, V. (eds.): Algorithmic Game Theory. Cam-
bridge University Press, Cambridge (2007)
98. Owen, G.: Game Theory, 3rd edn. Academic Press, San Diego (1995)
99. Ozdaglar, A., Bertsekas, D.: Routing and wavelength assignment in optical networks.
IEEE/ACM Trans. Netw. 11(2), 259–272 (2003)
100. Paganini, F., Wang, Z., Doyle, J.C., Low, S.H.: Congestion control for high performance,
stability, and fairness in general networks. IEEE/ACM Trans. Netw. 13(1), 43–55 (2005)
101. Pan, Y.: A game theoretical approach to constrained OSNR optimization problems in optical
networks. Ph.D. Thesis, University of Toronto, Toronto, Canada (2009)
102. Pan, Y., Alpcan, T., Pavel, L.: A distributed optimization approach to constrained OSNR
problem. In: Proceedings of the 17th IFAC World Congress, July (2008)
103. Pan, Y., Alpcan, T., Pavel, L.: Effects of parameters on Nash games with OSNR target. In:
Proceedings of the 2nd International Workshop on Game Theory in Communication Net-
works (GameComm), October (2008)
104. Pan, Y., Alpcan, T., Pavel, L.: A system performance approach to OSNR optimization in
optical networks. IEEE Trans. Commun. 58(4), 1193–1200 (2010)
105. Pan, Y., Pavel, L.: OSNR optimization in optical networks: extension for capacity constraints.
In: Proceedings of the 2005 American Control Conference, June, pp. 2379–2385 (2005)
106. Pan, Y., Pavel, L.: Games with coupled propagated constraints in optical networks: the multi-
link case. In: Proceedings of the 46th IEEE Conference on Decision and Control, December,
pp. 3443–3449 (2007)
107. Pan, Y., Pavel, L.: Global convergence of an iterative gradient algorithm for the Nash equilib-
rium in an extended OSNR game. In: Proceedings of the 26th IEEE Conference on Computer
Communications (INFOCOM), May, pp. 206–212 (2007)
108. Pan, Y., Pavel, L.: Iterative algorithms for Nash equilibrium of an extended OSNR optimiza-
tion game. In: Proceedings of the 6th International Conference on Networking (ICN), April
(2007)
109. Pan, Y., Pavel, L.: OSNR optimization with link capacity constraints in WDM networks:
a cross layer game approach. In: Proceedings of the 4th IEEE Conference on Broadband
Communications, Networks and Systems (BroadNets), September (2007)
110. Pan, Y., Pavel, L.: A Nash game approach for OSNR optimization with capacity constraint
in optical links. IEEE Trans. Commun. 56(11), 1919–1928 (2008)
111. Pan, Y., Pavel, L.: Games with coupled propagated constraints in general topology optical
networks. In: Proceedings of the 1st International Conference on Game Theory in Networks
(GameNets), May, pp. 549–558 (2009)
112. Pan, Y., Pavel, L.: Games with coupled propagated constraints in optical networks with multi-
link topologies. Automatica 45(4), 871–880 (2009)
113. Parrilo, P.A.: Polynomial games and sum of squares optimization. In: Proceedings of the 45th
IEEE Conference on Decision and Control, December, pp. 2855–2860 (2006)
114. Pavel, L.: Dynamics and stability in optical communication networks: a system theory frame-
work. Automatica 40(8), 1361–1370 (2004)
115. Pavel, L.: Power control for OSNR optimization in optical networks: a noncooperative game
approach. In: Proceedings of the 43rd IEEE Conference on Decision and Control, December,
pp. 3033–3038 (2004)
116. Pavel, L.: An extension of duality and hierarchical decomposition to a game-theoretic frame-
work. In: Proceedings of the 44th IEEE Conference on Decision and Control, December,
pp. 5317–5323 (2005)
117. Pavel, L.: A nested noncooperative game formulation for OSNR optimization in distributed
optical links. In: Proceedings of the 44th IEEE Conference on Decision and Control, Decem-
ber, pp. 6958–6965 (2005)
118. Pavel, L.: A noncooperative game approach to OSNR optimization in optical networks. IEEE
Trans. Autom. Control 51(5), 848–852 (2006)
119. Pavel, L.: OSNR optimization in optical networks: modeling and distributed algorithms via
a central cost approach. IEEE J. Sel. Areas Commun. 24(4), 54–65 (2006)
256 References
120. Pavel, L.: An extension of duality to a game-theoretic framework. Automatica 43(2), 226–
237 (2007)
121. Pavel, L.: A nested noncooperative OSNR game in distributed WDM optical links. IEEE
Trans. Commun. 55(6), 1220–1230 (2007)
122. Perakis, G.: The price of anarchy when costs are non-separable and asymmetric. In: Proceed-
ings of the 10th International Integer Programming and Combinatorial Optimization (IPCO)
Conference, June, pp. 46–58 (2004)
123. Perakis, G.: The “price of anarchy” under nonlinear and asymmetric costs. Math. Oper. Res.
32(3), 614–628 (2007)
124. Ramamurthy, B., Datta, D., Feng, H., Heritage, J., Mukherjee, B.: Impact of transmission
impairments on the teletraffic performance of wavelength-routed optical networks. IEEE J.
Lightwave Technol. 17(10), 1713–1723 (1999)
125. Ramaswami, R., Sivarajan, K.: Routing and wavelength assignment in all-optical networks.
IEEE/ACM Trans. Netw. 3(5), 489–500 (1995)
126. Ramaswami, R., Sivarajan, K.N.: Optical Networks: A Practical Perspective, 2nd edn. Aca-
demic Press, San Diego (2002)
127. Robinson, J.: An iterative method of solving a game. Ann. Math. 54, 296–301 (1951)
128. Rosen, J.B.: Existence and uniqueness of equilibrium points for concave n-person games.
Econometrica 33(3), 520–534 (1965)
129. Roughgarden, T.: Selfish Routing and the Price of Anarchy. MIT Press, Cambridge (2005)
130. Roughgarden, T., Tardos, E.: How bad is selfish routing? J. ACM 49(2), 236–259 (2002)
131. Rouskas, G., Perros, H.: A tutorial on optical networks. In: Advanced Lectures on Network-
ing, vol. 2497, pp. 155–193. Springer, Berlin (2002)
132. Sandholm, W.H.: Population Games and Evolutionary Dynamics. Cambridge University
Press, Cambridge (2009)
133. Saraydar, C., Mandayam, N.B., Goodman, D.J.: Pricing and power control in a multicell
wireless data network. IEEE J. Sel. Areas Commun. 19(10), 1883–1892 (2001)
134. Saraydar, C., Mandayam, N.B., Goodman, D.J.: Efficient power control via pricing in wire-
less data networks. IEEE Trans. Commun. 50(2), 291–303 (2002)
135. Scutari, G., Barbarossa, S., Palomar, D.P.: Potential games: a framework for vector power
control problems with coupled constraints. In: Proceedings of IEEE International Conference
on Acoustics, Speech and Signal Processing (ICASSP) (2006)
136. Selten, R.: Re-examination of the perfectness concept for equilibrium points in extensive
games. Int. J. Game Theory 4, 25–55 (1975)
137. Shamma, J.S., Arslan, G.: Dynamic fictitious play, dynamic gradient play and distributed
convergence to Nash equilibria. IEEE Trans. Autom. Control 50(1), 312–326 (2005)
138. Shandbag, U., et al.: Nash equilibrium problems with congestion costs and shared con-
straints. In: Proceedings of the Conference on Decision and Control, December (2009)
139. Shen, H., Basar, T.: Differentiated Internet pricing using a hierarchical network game model.
In: Proceedings of the 2004 American Control Conference, June, pp. 2322–2327 (2004)
140. Slusarek, M.: Optimal online colouring of circular arc graphs. Inform. Theor. Appl. 29(5),
423–429 (1995)
141. Smith, J.M.: The theory of games and the evolution of animal conflicts. J. Theor. Biol. 7,
209–221 (1974)
142. Smith, J.M., Price, G.R.: The logic of animal conflict. Nature 246, 15–18 (1973)
143. Srikant, R.: The Mathematics of Internet Congestion Control. Systems & Control: Founda-
tions & Applications. Birkhäuser, Boston (2004)
144. Stefanovic, N.: Robust power control of optical networks with time-delays. Ph.D. thesis,
University of Toronto, Toronto, Canada (2010)
145. Stefanovic, N., Pavel, L.: Link level power control of optical networks with time-delay. In:
Proceedings of the 46th IEEE Conference on Decision and Control, December, pp. 5198–
5203 (2007)
146. Stefanovic, N., Pavel, L.: A Lyapunov–Krasovskii stability analysis for game-theoretic based
power control in optical networks. In: Proceedings of GameComm 2008, October (2008)
References 257
147. Stefanovic, N., Pavel, L.: Robust power control of optical links with time-delay. In: Proceed-
ings of European Control Conference, Budapest, Hungary, August (2009)
148. Stefanovic, N., Pavel, L.: Robust power control of single sink optical networks with time-
delays. In: Proceedings of the 48th IEEE Conference on Decision and Control, Shanghai,
China, December, pp. 2034–2039 (2009)
149. Stefanovic, N., Pavel, L.: A stability analysis with time-delay of primal-dual power control
in optical links. Automatica 45(5), 1319–1325 (2009)
150. Stefanovic, N., Pavel, L.: A Lyapunov–Krasovskii stability analysis for game-theoretic based
power control in optical links. J. Telecommun. Syst. 47(1), 19–33 (2011)
151. Stein, N.D., Ozdaglar, A., Parrilo, P.A.: Separable and low-rank continuous games. In: Pro-
ceedings of the 45th IEEE Conference on Decision and Control, December, pp. 2849–2854
(2006)
152. Strand, J., Chiu, A., Tkach, R.: Issues for routing in the optical layer. IEEE Commun. Mag.
39(2), 81–87 (2001)
153. Tan, C., Palomar, D., Chiang, M.: Solving nonconvex power control problems in wireless net-
works: low SIR regime and distributed algorithms. In: Proceedings of the 48th IEEE Global
Telecommunications Conference (GLOBECOM), November (2005)
154. Tonguz, O., Flood, F.: Gain equalization of EDFA cascades. IEEE J. Lightwave Technol.
15(10), 1832–1841 (1997)
155. Tonguz, O., Flood, F.: EDFA-based DWDM lightwave transmission systems with end-to-end
power and SNR equalization. IEEE Trans. Commun. 50(8), 1282–1292 (2002)
156. Turkcu, O., Subramaniam, S.: Blocking in reconfigurable optical networks. In: Proceedings
of the 26th IEEE Conference on Computer Communications (INFOCOM), May, pp. 188–
196 (2007)
157. van Damme, E.: Stability and Perfection of Nash Equilibria, 2nd edn. Springer, Berlin (1987)
158. Vincent, T.L., Brown, J.S.: Evolutionary Game Theory, Natural Selection, and Darwinian
Dynamics. Cambridge University Press, Cambridge (2005)
159. von Neumann, J., Morgenstern, O.: Theory of Games and Economic Behavior. Princeton
University Press, Princeton (1944)
160. Yaiche, H., Mazumdar, R., Rosenberg, C.: A game theoretic framework for bandwidth allo-
cation and pricing in broadband networks. IEEE/ACM Trans. Netw. 8(5), 667–678 (2000)
161. Yariv, A.: Signal-to-noise considerations in fiber links with periodic or distributed optical
amplification. Opt. Lett. 15(19), 1064–1066 (1990)
162. Yassine, A., et al.: Competitive game theoretic optimal routing in optical networks. In: Pro-
ceedings SPIE, May, pp. 27–36 (2002)
163. Yates, R.: A framework for uplink power control in cellular radio systems. IEEE J. Sel. Areas
Commun. 13(7), 1341–1347 (1995)
164. Zander, J.: Performance of optimum transmitter power control in cellular radio systems.
IEEE Trans. Veh. Technol. 41(1), 57–62 (1992)
165. Zang, H., Jue, J., Mukherjee, B.: A review of routing and wavelength assignment approaches
for wavelength-routed optical WDM networks. Opt. Netw. Mag. 1(1), 47–60 (2000)
166. Zhu, H., Mukherjee, B.: Online connection provisioning in metro optical WDM networks
using reconfigurable OADMs. IEEE/OSA J. Lightwave Technol. 23(10), 2893–2901 (2005)
167. Zhu, K., Mukherjee, B.: Traffic grooming in an optical WDM mesh network. IEEE J. Sel.
Areas Commun. 20(1), 122–133 (2002)
168. Zhu, Q., Pavel, L.: Enabling differentiated services using generalized power control model
in optical networks. IEEE Trans. Commun. 57(9), 2570–2575 (2009)
Subject Index
Partitioned Nash game, 121, 122, 124, 125, Spectral gain shape, 79, 82, 86
127, 136, 137 Spectral radius, 234
Path, 214 Spectral shape, 75
Payment function, 215, 218 Stackelberg equilibrium, 25
Payoff function, 13, 215 Stackelberg game, 66
Penalty based approach, 114 Stage Nash game, 122–125, 136
Photodetector, 74 Stochastic games, 20
Physical layer, 5 Strategic form, 16
Physical layer impairment, 5 Strategic interaction, 11, 12
Player, 11, 12 Strategically dominant, 24
Point-to-point link topology, 76, 98 Strategy, 13, 16, 18, 211, 212, 215, 217
Point-to-point WDM link, 78, 83 Strategy space, 12, 13, 16, 18
Positive definite, 51, 113, 233 Strictly convex, 48, 51, 53
Positive semidefinite, 95 Strictly Pareto dominated, 24
Power allocation, 12 Symmetric game, 17
Power control, 73, 80, 82 System connection matrix, 86
Price algorithm, 197 System cost function, 144
Price of Anarchy, 25, 143, 213, 216, 222 System matrix, 89, 166
Price of Stability, 216
Pricing, 2, 7, 100, 117 T
Pricing function, 99, 107, 121, 124, 219 Time-delay, 189–192, 205
Primal-dual algorithm, 197 Time-delay system, 192, 240, 241
Probability, 18, 32 Time-scale, 197, 237
Projection set, 106 Time-scale decoupling, 198
Projection Theorem, 236 Transmitter, 74, 75, 87, 161
Pure strategy, 12, 32 Transport layer, 3, 5
Pure-strategy, 216 Two-argument, 51, 56, 58–60, 230
Pure-strategy Nash equilibria, 28 Two-player, 17, 50
Pure-strategy Nash Equilibrium, 216
U
R Uncoupled constraint, 45
Randomization, 13, 18, 32 Upper-semi-continuous, 53, 234, 235
Rational player, 25 Utility function, 13, 100–102, 124, 125, 137,
Reaction function, 48–50, 109 215, 218
Reaction set, 48, 53
Receiver, 74 V
Reconfigurable lightpath, 3 Variable optical attenuator, 161
Recursive OSNR model, 90 Variable optical filter, 82
Reduced system, 200, 238 Virtual optical link, 87
Repeated game, 25 Virtual topology, 4, 76
Repeated games, 20
Routing and Path Coloring, 211 W
Routing and wavelength assignment, 211 Wavelength, 74–76
Wavelength assignment, 211
S Wavelength converter, 76
Schur complement, 201 Wavelength division multiplexing, 3
Selfish-RPC, 216 Wavelength-dependent gain, 79
Separable, 46, 52, 58, 59, 65, 115 Wavelength-division multiplexed, 71
Single-sink topology, 76, 128–130 Wavelength-routed optical network, 4
Singleton, 48, 49, 54 Wavelength-routed optical networks, 75
Singular perturbation, 197, 238 Weakly Pareto dominated, 24
Slow time-scale, 238
Social cost, 213, 216, 217 Z
Socially optimal, 24, 143, 151, 213, 216 Zero-sum games, 19