Constraint Theory Multidimensional Mathematical Model Management
Constraint Theory Multidimensional Mathematical Model Management
George J. Friedman
Phan Phan
Constraint
Theory
Multidimensional Mathematical Model
Management
Second Edition
CONSTRAINT THEORY
MULTIDIMENSIONAL
MATHEMATICAL MODEL
MANAGEMENT
International Federation for Systems Research International
Series on Systems Science and Engineering
Editor-in-Chief: George J. Klir
Binghampton State University
Editorial Board
Gerrit Broekstra Ivan M. Havel
Erasmus University, Rotterdam, Charles University, Prague,
The Netherlands Czech Republic
John L. Castii Klaus Kornwachs
Sante Fe Institute, New Mexico Technical University of Cottbus,
Brian Gaines Germany
University of Calgary, Canada Franz Pichler
University of Linz Austria
This is the last volume in the IFSR series with the above stated editorial board. A reconstituted editorial
board with George Mobus as Editor-in-Chief will continue the series.
IFSR was established “to stimulate all activities associated with the scientific study of systems and to coordinate
such activities at international level.” The aim of this series is to stimulate publication of high-quality
monographs and textbooks on various topics of systems science and engineering. This series complements
the Federations other publications.
A Continuation Order Plan is available for this series. A continuation order will bring delivery of each
new volume immediately upon publication. Volumes are billed only upon actual shipment. For further
information please contact the publisher.
Second Edition
ISSN 1574-0463
IFSR International Series on Systems Science and Engineering
ISBN 978-3-319-54791-6 ISBN 978-3-319-54792-3 (eBook)
DOI 10.1007/978-3-319-54792-3
A single circuit cluster with over 1,000 independent simple circuits (thus
there are 1000 more edges than vertices)
This book is dedicated to our
wives, who have helped us find
the time and priority for our
mathematical dreams.
Preface
Over the past several years, Constraint Theory has been a substantial part
of a special graduate course in the Systems Architecture and Engineering
program at the University of Southern California, where the authors are
adjunct faculty members. The feedback from this group of bright graduate
students was invaluable. Special thanks are given to these graduate students
who performed research studies – summarized in Appendix A – on
Constraint Theory: Leila Habibibabadi, Kathya Zamora-Diaz and Elliott
Morgan [1]. Extra special thanks are given to Mr. Gary L. Friedman who
provided the extremely valuable service of intellectual reflector, editor and
digital format manager for this book.
Introduction
Many thousands of papers have been written about the accelerating pace
of increased complexity and interactivity in virtually every walk of life in the
developed world. Domains which previously could have been studied and
managed separately – such as energy, the environment and economics – must
now be dealt with as intimately intertwined disciplines. With its multitude of
additional capabilities, complex systems also provide a treacherous array of
fragile failure modes, and both the development and operation of new
systems are an increasing challenge to the systems engineer. Advanced
technology is the primary driving force behind the increasing complexity and
the enthusiastic pushing of immature technologies is behind most of the early
failures in the development phases.
Perhaps the most significant advanced technology employed in new
complex systems is the computer science family with its ancillary disciplines
of communications and software. Fortunately, computer science also repre-
sents a major opportunity to control the design and operation of complex
systems because of its ability to perform multi-dimensional modeling to any
level of detail desired. Math models have been used in support of every phase
of systems engineering, including requirements management, preliminary
design, interface and integration, validation and test, risk management,
manufacturing, reliability and maintainability, training, configuration man-
agement and developing virtual universes to measure customer preferences
prior to the implementation of the design. Properly used, the enormous power
of modern computers can even furnish researchers with a synthetic world
where theories can be tried and tested on validated models, thus avoiding far
more expensive tests in the real world. A wide variety of questions – or
“tradeoffs’ – can be asked of the models and, at least in theory, the analyst
has a free choice as to which computations he wishes to observe and which
variables he desires to be independent. Philosophically, it can even be argued
that the math model employed in this fashion provides the technologist a
virtual extension of the scientific method itself.
Those who have actually wrestled with large-scale models will complain
that the above description is far too rosy. Submodels which are developed by
xiv Constraint theory
CHAPTER 1 – MOTIVATIONS 1
WHAT IS CONSTRAINT THEORY AND WHY IS IT IMPORTANT?
1.1 TRENDS AND PROBLEMS IN SYSTEM TECHNOLOGIES
1.2 AN EXAMPLE OF LOW DIMENSION
1.3 THE MANAGER AND ANALYST CONTINUE THEIR DIALOGUE
1.4 PRELIMINARY CONCLUSIONS
1.5 A LITTLE WINDOW INTO FUTURE CHAPTERS
1.6 PROBLEMS FOR THE CURIOUS READER
6.7 SUMMARY
6.8 PROBLEMS FOR THE DISCRETE STUDENT
APPENDICES
APPENDIX C: THE LOGIC OF “IF” AND “IF AND ONLY IF” 201
REFERENCES 211
INDEX 215
Chapter 1 MOTIVATIONS
Gone forever are the simple days! Virtually every identifiable trend is
driving humanity’s enterprises into more intimate interaction and conflict.
Increased population, accelerated exploitation of resources, and expanded
transportation have brought the previously decoupled worlds of economics,
energy and the environment into direct conflict. With the greater efficiency
of travel and communication, the emergence of global marketplaces and the
revolution in military strategies, the international world is incredibly more
interactive, multidimensional and complex than even a decade ago. Locally,
we observe ever tighter coupling between emerging problems in crime,
poverty, education, health and drug misuse. All these issues have been
aggravated by an explosion of new technology and – especially in the United
States – a compulsion to force these new technologies into early and often
simultaneous application. The most vigorous of these advancing
technologies – digital computation – brings with it an unexpected
complexity challenge: software and the management of highly complex and
multidimensional mathematical models.
Fortunately, this most rapidly advancing technology of computer science
not only adds to the complexity of designed systems, it also contributes
enormously to designing these systems themselves. A host of new “computer
assisted” software packages are published each year, running the gamut from
Computer Assisted Design (CAD), Computer Assisted Engineering (CAE),
Computer Assisted Systems Engineering (CASE), Computer Assisted
Manufacturing (CAM), Computer Assisted Instruction (CAI), and eventually
© Springer International Publishing AG 2017
G.J. Friedman, P. Phan, Constraint Theory, IFSR International Series on Systems
Science and Engineering, DOI 10.1007/978-3-319-54792-3_1
2 Constraint theory
T = PE/C (1)
where:
P was the political index of acceptability by the board of directors,
E was the system effectiveness, and
C was the life cycle cost of the system.
where D, the development cost, was to be limited to 30% of the total cost.
S = X + 0.5D (3)
where:
S is the total support cost
X is the cost of ops and support if there were no new technology
D is the development cost for the system, including new technology.
C=D+S (4)
S = K2E/(1-A) (5)
where:
K2 is a constant,
A is the probability that the system is ready when called upon.
E = MA(D/Dmax)1/2 (6)
where:
M is the mission success probability, given the equipment is available
D is the amount spent on development
Dmax is the budget requested by the developers
All these inputs from the specialists were reviewed for reasonableness by
the systems analyst and integrated into the “model” shown in Table 1-1.
1) T = PE/C
2) D = k1C where k1 = 0.3
3) S = x + 0.5D
4) C=D+S
5) S = k2E/(1-A)
6) E = MA(D/Dmax)1/2
where:
T = Top-level systems criterion
P = Political index of acceptability
C = Life cycle cost of a system
D = Development costs of system
S = Support and operations cost
E = Effectiveness of system
M = Mission success probability (working)
A = Availability of system
6 Constraint theory
graphs, there are two distinct types of junctions: squares represent the
relations, and circles represent the variables. The arcs, or “graph edges”
connect each relation to the variables that are relevant to it. The “degree,” d,
of each junction is defined as the number of edges which intersect it.”
Figure 1-1. A right-brain view of relation 5, introducing the concept of representing relations
by squares and variables by circles, as well as demonstrating that computation can flow
through the relations in many directions.
their relevant relations – and thus additional squares are appended to the
bipartite graph.
Figure 1-2. The perspective is expanded to include both relations 5 and 6, which share
variables A and E.
Figure 1-3. The Bipartite Graph: A Metamodel displaying consistency and computability.
Figure 1-4. Symbology and rules of propagation for bipartite model graph.
now apply constraint to their neighbors and the edges take on a directionality
which is determined by the request. Essentially, the computation or
constraint “flows” across the graph.”
“For treelike graphs, the rules to map the sequential propagation of
information – or computation, or constraint – are simple in the extreme: for
d edges intersecting a square, information will propagate if there are (d-1)
inputs and one output; for d edges intersecting a circle, information will
propagate if there is one input and (d-1) outputs.” (See Figures 1-1 and 1-4.)
Note that a square with but a single arc intersecting it will automatically
transmit constraint to the circle it is connected with, since in this case, d=1
and (d-1)=0, thereby requiring no inputs to generate its output.
“For graphs which contain circuits, mapping propagation is a little more
complex. Once we have gone as far as we can with the sequential rules
above, we may require the rule of simultaneous propagation in the vicinity of
a circuit: if a connected subgraph exists which contains an equal number of
unpropagated squares and circles, then all its variables may be computed as
if they were within a set of simultaneous equations.”
“Now I can show you how easy it was to determine the computability of
your computational requests. Looking at Figure 1-5, we can easily see that
relations 2, 3 and 4 form a submodel with an equal number of squares and
circles. This denotes three simultaneous equations covering three unknown
variables and we should expect to be able to solve for the three variables,
converting them from unknown variables to fixed parameters. Thus, your
request, T = f7(S,P) is unallowable since it assumes S is variable rather than
a fixed parameter. For the same reason, C and D cannot be independent
variables either. This type of constraint imposed on the model is called
intrinsic because it existed even before you made any computational
request.”
“Now look at Figure 1-6, which shows constraint propagating from left to
right along variables C, D and S for the above explained reason. When you
requested the computation T = f8(M,A) you established the two independent
variables as a source of extrinsic constraint which propagates into the model
and hopefully gives us a computation of T. Let’s see what happens when we
employ the sequential propagation rules for constraint flow. Since M and A
are extrinsic sources of constraint and D is an intrinsic source, we can satisfy
the (d-1) inputs and one output rule for equation 6, thus producing a
computation for variable E. With A as an extrinsic source and S as an
intrinsic source, we can satisfy the (d-1) inputs and one output rule for
equation 5, thus producing another computation for variable E. This is a case
of local overconstraint, making the requested computation unallowable.”
12 Constraint theory
“Now the term, ‘tradeoffs, with everything else held equal’ is fraught
with ambiguity and most often used with insufficient rigor. Figure 1-10
displays how holding E at some constant value effectively decouples M and
P into different, non-communicating subgraphs, rendering this tradeoff
request unallowable.
If any combination of C, D and S were to be held constant, two types of
problem emerge. First of all, the value to which they were held constant
might not agree with the values computed from the 2, 3 and 4 simultaneous
equations. Even if they did agree, then propagating constraint across the
graph would yield an underconstraint at equation 1 – there would be an
insufficient number of inputs to provide the desired computation of P. The
same underconstraint situation occurs if the variable held constant is A. In
fact, the only variable that can be held constant in order to provide the M vs
P tradeoff is T.”
1. Motivations 15
The manager absorbed all these inputs soberly and after reviewing the
results of his requested computations as well as others which were allowable
on the original model, he complained, “I certainly can’t argue with your
mathematical rigor. But I’m still disappointed that I didn’t get more insight
out of this model for my preliminary design phase – I expected more,
somehow. It seems that whether a computation is allowable or not is like the
flip of a coin.”
“It’s worse than a coin flip; much worse. You have just observed a very
common and generally unappreciated feature of most math models,” the
analyst responded. “Indeed, the vast majority of all possible computational
requests on almost all models are unallowable. Some of the author’s
graduate students performed an exhaustive analysis of the likelihood of
computational allowability on 6- and 8-dimensional connected models of a
variety of topologies. The results are presented in detail in Appendix A, but
the general allowability likelihoods were surprisingly low. Of the 150
1. Motivations 17
analysis on this size model without using computer aids. This was possible
because the rules, and later, the theorems were developed using
comprehensible models of low dimension, and then extending them
rigorously to higher dimensionality. For example, the model depicted in
Figure 1-13 – again about another S times larger than Figure 1-12 – would
undoubtedly boggle the mind of even the most focused analyst unless
computer aids were available. In order to communicate with the computer, a
new construct – called the ‘constraint matrix’, which has all the information
inherent in the bipartite graph – will be employed.”
Figure 1-12. A sequential computational flow in a model about S times larger than the
example.
Figure 1-13. A model about another S times larger certainly invites computer assistance.
Inspired by Edgar Palmer’s [4] delightful title page for his book, Graphical
1
For the allowable requests, draw the directed bipartite graph which
depicts the computational flow. For the unallowable requests,
discuss the reason(s) for the unallowability.
Computational Requests:
M=f(A), M=F(E), E=f(A,M), P=f(T,E), P=f(T,A)
deliberately, could they try to fit their pieces of Truth together – to make
some sense of things.
Mankind’s yearning to understand the world over the eons has been aided
by the development of mathematical models. Groups of researchers,
sometimes spanning centuries contribute their little fragments of data or
understanding and eventually a general theory emerges. In many cases, the
consequences of the new theory are unexpected by the original contributors,
but such is the trust given to mathematics, the unexpected, nonintuitive
results are accepted given they are mathematically sound. Examples:
• In the 16th century, Tycho Brahe organized and extended the
astronomical observations of Copernicus and others into the world’s
finest set of data on stellar and planetary objects. Johann Kepler
took this data and formulated his famous three laws of planetary
motion. Despite his disappointment that planetary orbits were
elliptical – rather than the circles the Greeks maintained were
necessary for “celestial perfection” – he convinced himself and the
scientific world that the ellipse was the correct mathematical form
for all the orbits in Tycho’s data base.
• Decades later, Isaac Newton, with his greater mathematical
understanding, was able to generalize Kepler’s laws into his law of
universal gravitation – a gigantic intellectual feat which unified the
laws of the heavens and earth.
• Centuries later, Albert Einstein provided a refinement of Newton’s
theory of universal gravitation with his general theory of relativity.
Alexander Friedmann solved Einstein’s equations and concluded
that the universe began in a monstrous big bang. This was so against
Einstein’s instincts that he added a cosmological constant to his
equations of relativity to remove the possibility of an expanding
universe or the big bang. However, the rationality of mathematics,
as well as new data by Hubble and others have established
Friedmann correct and Einstein has referred to the cosmological
constant as his greatest blunder.
So in Man’s quest to understand, mathematical modeling has taken an
increasingly central role in building theories, and indeed in the scientific
method itself. The jagged shards of data, incomplete observations and
subdimensional theories are pieced together rationally – often resulting in
unexpected conclusions and a deeper view of the world. With the advent of
modern computer technology, this central importance promises to increase
far more.
“You certainly won’t get arguments from most practitioners of science
and technology about the importance of computers,” remarked the manager,
2. The Four-Fold Way 27
Figure 2-1. It all begins with the simple concept of sets, subsets, and the null set.
Figure 2-2. The sets of variables and their allowable values have enormous flexibility.
Figure 2-3. The Model Hyperspace formed by the orthogonal axes of the variables is a useful
abstraction, although in general it is impossible to be perceived.
Definition 4: The product set of a set of variables is the set containing all
possible combinations of the allowable values of all the variables. In the case
where all the variables are continuous over an infinite range, the product set
is merely every point within the hyperspace defined by the set of variables.
(Figure 2-4)
2. The Four-Fold Way 31
Figure 2-4. The Product Set contains all possible combinations of the allowable sets of the
variables’ values.
Figure 2-5. Wiener suggested that a “relation” between variables can be defined as the subset
within the product set of these variables which satisfies it. This not only provides rigor, but
permits a tremendous variety of relation types.
Figure 2-6. Relations, as well as variables held constant, constrain the product set into a much
smaller subset.
2. The Four-Fold Way 33
Figure 2-7. Only four operations from set theory are employed for Constraint Theory.
Definitions 7: The union of sets A and B is the set of all points which are
either in set A or in set B or both. Symbolically:
x ңAыB if: x ң A or x ң B
34 Constraint theory
The intersection of sets A and B is the set of all points which are in
both A and B. Symbolically:
The extension of set A into dimension y is the set of all points within set
A plus all possible values of the dimension y. For example, if set A is the
point (2,4) in the xy plane, then Exy(2,4) is the line x=2 where y varies over
all its possible values. Extension is a dimension increasing operation.
Symbolically:
ExyA = (x=2) in xy-space (line)
The set theoretic definition of relation was chosen to provide the firmest
and broadest mathematical foundation for the work to follow. Unfortunately,
it cannot also be claimed that this viewpoint is a practical way to describe
the relation. There are some occasions, such as tabulated or plotted
functions, when it is necessary to list every point within the relation subset
exhaustively. In these cases, the relation subset is merely the union of all the
listed points within the hyperspace of the model. However, in the vast
majority of mathematical models, far more efficient means are used to define
the usually infinite number of points comprising the relation subset.
These efficient means almost invariably involve the concept of
describing the total model as the intersection or union (or both) of a set of
submodels or algorithms. (Figure 2-9) This is necessary for at least two
reasons: First, and more obvious, a practical way of specifying infinite sets is
required. Second, and deeper, model builders cannot conceive of the entire
model with their limited perceptual dimensionality and thus attempt to
construct higher dimensional models by aggregating in some fashion a series
of lower dimensional submodels. The rules of aggregation employ the union,
intersection, projection and extension operators defined previously.
Frequently, a function is specified in a piecewise fashion; for example:
x = 0 when t<0
x = t2 when t>0
In cases like this, the meaning is that the contribution of these two sets to
the total model is the union of the sets.
More frequently, a collection of “simultaneous equations” attempt to
define the model; for example:
x + y + z = 13
x - y= 8
In cases like this, the meaning is that the contribution of these two sets to
the total model is the intersection of the sets.
In general, the dimensionality of the total model is far greater than any of
the contributing submodels. Thus, the contributing submodels specify only a
subset of the total model and, in order for them to be able to intersect in the
total dimensional space, they must be extended into all the unspecified
directions. For example, let the total model space be xyz and let the relation
subset for f1(x,y)=0 be A1 and the relation subset for f2(x,z)=0 be A2. Thus,
36 Constraint theory
Figure 2-8. Relevancy of a variable with respect to a relation can be defined in terms of the
projection and extension operations.
AV = Prxy(A6)
2. The Four-Fold Way 37
Figure 2-9. Total model relations are generated by families of submodels which are combined
by the union, intersection, extension and projection operations.
Figure 2-10. After each generating relation is developed, it is extended into the full
hyperdimensional space of the total model, forming the relation AȈ, which in turn is projected
onto the subspace of the computational request, where it can be “viewed” by managers,
analysts and others interested in learning from the model.
“You started off simply enough,” commented the manager. “What can be
easier than the definition of sets and their operations? Also, your extension
of Cartesian coordinates into hyperspace can be grasped by extrapolating
from what we know of one, two and three dimensional spaces.
As a teenager, I was inspired by Abbot’s Flatland [11] and Burger’s
Sphereland [12]. These extensions are interesting philosophically, but do
they really represent the real world and should they be the basis for applied
mathematics? I’ve heard professors argue that Descartes himself only was
thinking of our familiar three dimensional space, not even the four
dimensions for relativity theory, the eleven dimensions for string theory, and
certainly not the hundreds of dimensions we need for a modern mathematical
model.”
2. The Four-Fold Way 41
Figure 2-11. The bipartite graph is a metamodel of the full model which illuminates the
model’s structural and computational properties.
42 Constraint theory
Figure 2-12. The Constraint Matrix is the companion to the bipartite graph and displays
exactly the same information. It is also a metamodel which contains only that information
relating to the model’s structural and computational properties.
2. The Four-Fold Way 43
Figure 2-13. Evolutionary transition from the bipartite graph to the constraint matrix.
Figure 2-14. In order for a problem to be well-posed, the mathematical model must be
consistent and the computation must be allowable. Consistency requires that the hyper-
dimensional relation not be the null set. Allowability requires that all the variables of the
requested computation be relevant to the projection of the total relation onto the
computational subspace.
46 Constraint theory
Figure 2-15. The four representations of a mathematical model. The first two are full models
and the latter two are metamodels.
48 Constraint theory
5. Draw the constraint matrix for Figure 3 of Chapter One. Can you
suggest how the term, “Basic Nodal Square” was developed?
For the allowable requests, draw the directed bipartite graph which
depicts the computational flow. For the unallowable requests,
discuss the reason(s) for the unallowability.
Computational Requests:
E=f(T,M), A=f(T,E), A=f(P,M)
Chapter 3 GENERAL RESULTS
Even the formulas and rules which I applied were not always provided with
the necessary assumptions to clarify their domains of applicability.”
“A major advantage that math has over language is its generizability into
domains which were previously incomprehensible,” the analyst pontificated.
As was mentioned in Chapter 1, we cannot really perceive numbers over
seven or so – the way that we (and many animals) can perceive 1, 2 and 3.
But we so trust the algorithms of arithmetic that we at least have a feeling or
control of understanding that which we need to know about numbers into the
millions or billions: ‘which number is greater and by how much?’ ‘is number
a slightly greater than b or is it an order of magnitude greater?’ These useful
answers can be obtained without fully perceiving large numbers. Similarly,
the objective of constraint theory is to provide the manager of large models a
trustworthy method to obtain answers to certain important properties of
models: ‘Is the model consistent?’, ‘Is the computation a=f(b,z) allowable on
this model?’ These answers are useful whether or not the manager can
actually perceive the very high dimensions involved.”
“OK, thanks for the sermon,” said the manager, feeling that the
explanations were somewhat long. “Let’s see how you can bring me into
Devlin’s “math house” and what trustworthy results you have to provide.”
“I couldn’t fail to disagree with you less,” commented the manager, self-
referentially.
Theorem 2: If any submodel of a total model is inconsistent, then the
entire model is inconsistent.
Proof: The total model relation set is the intersection of the relation sets
of all its submodels. Since the intersection of any set with the null set is the
null set, if any submodel relation is the null set, then the total model will also
become the null set and thus, by Definition 13, will be inconsistent. QED.
“These are a good start it seems,” commented the manager, but aren’t
they rather intuitive? And thus perhaps not so useful?”
“Well I hope all the remaining theorems will be as intuitive, or at least as
plausible as these,” the analyst responded. “As far as useful goes, it tells us
to check the consistency of a model before we attempt computations on it. In
my experience, most of the time computations fail, the model is checked
only after much wasted effort. The utility goes further: even if a part of the
model that is not used in the computation is inconsistent, then the
computation is not allowable. Look at Figure 3-1. At first glance, it would
appear that the computational request d=f(a,c) is allowable. However,
observe that on the left side of the model, we have over three relations
constraining the two variables e and f; this is a serious case of overconstraint
– that is, the relation set for the submodel containing e and f is the null set.
Therefore, no computations, including d=f(a,c), are allowable.”
Figure 3-1. Inconsistency in one part of the model poisoning the whole model.
“OK, that seems nontrivial,” admitted the manager. “I would have treated
the d=f(a,c) computation as a work in progress and suppressed the remainder
of the model as requiring repair. I note that even if the overconstrained
submodel were reduced to just one basic nodal square, a would be
intrinsically constrained, thereby not permitting a to be an independent
variable.”
“Excellent observation,” complimented the analyst, “but you rarely know
in advance all the computational requests you wish to make on a model and
3. General Results 53
what portions of the model will be required for the request’s computational
paths. Thus, any submodel with a null set relation set can poison the entire
model. An even more severe case is shown in Figure 3-2. Here the
overconstrained submodel is in a completely separated component – there
being no possible computational paths between it and the computational
request – and the total model is still inconsistent, rendering all computational
requests unallowable.”
Figure 3-2. Inconsistency even in a disconnected component poisoning the whole model.
Figure 3-3. A simple example of the consistency of relations without common relevant
variables.
of knots. Therefore the number of subsets in the set of K knots equals 2K.
QED. Refer to Figure 3-4 for demonstrations and simple examples.
Figure 3-4. Three demonstrations that the power set of the set of N elements contains 2N
subsets.
“If it weren’t for Theorem 6,” commented the manager, “the number of
possible computational requests predicted by Theorem 7 would be far larger,
I presume.”
“Certainly,” agreed the analyst, “for each of the 2K subsets of dimension
d, one could choose d dependent variables, enormously increasing the
possible computational requests. Of course, one could still choose different
ways to plot the results of the computational requests, but that is really
56 Constraint theory
Thus far, all the results and discussions of consistency and computability
have been on the basis of any conceivable type of general relation. As was
discussed in Chapter 2, relations can take on an extremely wide variety of
properties. In order to progress towards effective tools for the management
of multidimensional math models, it will be necessary to define three
important classes of relations.
3. Generral Res
R sultss 57
F
Figurre 3-5. Exam
E mplees of
o treees, circuits, adjjaceent circuuits and
a circuuit clust
c ters..
These three relation classes will be treated in the next two chapters. We
will discuss the continuum relation class first – in Chapter 4 – because it can
be argued that it is the most important for math model building and also
because constraint theory happens to be most useful for this class. Chapter 6
will treat discrete and interval relations.
3. General Results 59
Figure 3-6A. Three relation classes can be defined – at least locally – by the number of points
in their intersections with a line and the number of points in their projections onto a line.
“I’m beginning to see what you mean by ‘living in a math house,’” the
manager complained mildly. “Although all the definitions and Theorem
proofs are imbedded in natural language, they appear far ‘tighter and
rigorous’ than the ambiguities of normal conversation. It certainly doesn’t
make for fast reading. Especially if one wants to understand the proofs.”
“I’m trying to make it as plausible and painless as I can,” responded the
analyst. “I encourage you to read the proofs and get in the spirit with the
mathematical rhythm of this theory. Otherwise, you won’t gain as much
confidence in the trustworthiness of extending these results into the high
dimensions that we need in order to manage modern models. So far, you’ve
just entered the foyer of the ‘math house’; I want to show you at least three
more rooms.”
60 Constraint theory
“I must admit that the foyer of the math house was fascinating,” said the
manager, “and the rigorous structure based on the previous definitions was
quite impressive. With the theorems and their proofs, I guess I’ve progressed
from protomath to full math. However, it was a little like the ten
commandments: after an admonition on how much I should respect and
revere this central philosophy, all I got was a series of negative statements:
can’t kill, can’t commit adultery, can’t compute if inconsistent... I’m still
looking for useful rules which will permit me to manage, as this book’s title
promises.”
“This Chapter will present many rules and procedures by which you can
more effectively manage large math models,” assured the analyst. “In fact, it
will end with a Constraint Theory Toolkit which summarizes the most useful
of the theorems, rules and procedures.”
“Central to many of the rules will be the treatment of circuits and loops
within extremely tangled bipartite graphs. Many researchers appear to have a
basic phobia about circuits. Logicians dislike self-referential loops because
of the potential for paradoxes. Early in the field of control systems, feedback
was thought to be illogical. Even von Neumann classified intransitive loops
of preferences as ‘irrational.’ (See Figure 4-1) Devlin [12] noted that the
common structure of language worldwide was treelike and hypothesized that
‘our cognitive wiring favors trees.’ So I’m asking you to keep an open mind
on the concept and value of circuits and loops.”
Figure 4-1. The concept of circuits has initially appeared to be contrary to rational thought,
but their careful management has led to many advances.
“I’ll try, but this is the briefest philosophical introduction you’ve given to
any chapter so far,” chided the manager. “I’ve attempted to enjoy your
pontifications on cognitive science, language, origins of mathematics and the
universe. Are we properly warmed up for the meat of this chapter?”
4. Regular Relations 63
“Yes, I hope so,” responded the analyst. “We have more material in this
chapter than any other. It is the heart of Constraint Theory so far in its
development. I’m anxious to get going.”
Definition 20: A pair of relations are locally universal if the ranges and
domains of their relevant variables are mutually compatible. In other words,
any output from one relation is an acceptable input to the other. For example
the circle x2+y2=1 is locally universal with x=0, but not with x=3.
Postulate 1: Model builders inherently wish their relations to be locally
universal.
Although an important assumption for constraint theory analysis, the
actual testing of universality can only be accomplished on the full model; the
constraint theory metamodels do not have sufficient information.
Definition 21: A set of regular relations are continuum relations which
are locally universal with all their interacting relations.
Definitions 22: The constraint potential of a graph G is defined as p(G)
and equals the excess of nodes over knots: p(G)=N-K. It is the negative of
the “degrees of freedom” notion used in Chapter 1. It will be useful to define
two circumstances where constraint potential is insightful:
Definition 23: The degree of a vertex, d(v), equals the number of arcs
that intersect that vertex. For the node vertex, d(n), it is the number of
variables that are relevant to the relation; for the knot vertex, d(k), it is the
number of relations which employ that variable. Refer to Figure 4-2 for
examples.
Thus, for example, the constraint potential of a node with degree d(n) is
merely 1-d(v) since it represents one node attached to d(n) knots. In general,
the average degree of all of a graph’s vertices is a strong indication of the
graph’s connectivity.
Theorem 10: For a model graph of regular relations, with a tree-like
topological structure, the computational rules are:
64 Constraint theory
Figure 4-2. The degree of a vertex is merely the number of edges which intersect it.
Figure 4-3. Computational (constraint) flow in trees requires only the simple rules: “d(n)-1 in,
1 out for nodes; and 1 in, d(k)-1 out for knots.”
Fundamentally, nodal squares and basic nodal squares have the property
that, in their local submodel, the number of variables equal the number of
relations. Examples are shown in Figure 4-4. Recall that submodels are
formed from the total model by grouping subsets of the nodes (or rows of the
constraint matrix); thus the term, “nodal” squares if all the elements of the
constraint matrix are captured in a “square.” In this context, groupings of the
66 Constraint theory
Figure 4-4. A Nodal Square, NS, is a grouping of the rows of a constraint matrix, [C], such
that all the relevant elements are captured within a square. That is, the submodel has an equal
number of relations and variables. A Basic Nodal Square, BNS, is an NS which does not have
a smaller NS within it.
Figure 4-5. Regular relations have the property that each application of a new relation reduces
the dimensionality of the total relation by one.
Figure 4-6. The Nodal Square (NS) is not the kernel of constraint; only the Basic Nodal
Square (BNS) is. In NS-type a, constraint flows from the BNS to fill out the NS. In NS-type
b, multiple BNSs overconstrain some of the knots and constraint doesn’t even reach the other
knots.
Figure 4-7. Even before a computational request is made, intrinsic constraint sources –
formed on the BNS kernels – may exist in consistent models. These sources flow their
constraint outwards into resultant constraint domains. If these domains overlap, overconstraint
of the variables within the overlap is likely, rendering the overall model inconsistent.
how messy it can get. There is an important theorem in set theory by Hall
[14] which we modify into bipartite graph theory as:
Theorem 12: In order that a distinct output variable be associated with
each of m relations it is sufficient that, for each K=1,2,3..m, any selection of
K of the relations shall contain between them at least K relevant variables.
The computer aided method is the only one feasible and we need the
constraint matrix for communication with the computer. However, if we are
to address realistic dimensions of modern models, we must avoid the trap of
exhaustively searching for the BNS culprits through power sets with their
attendant exponential explosions.
Figure 4-8. A bipartite graph so large that gravity forces it into a spherical shape.
Our strategy will be to understand how the properties of the BNS fits
within the easily computed properties of the bipartite graph. If we are alert
72 Constraint theory
and understand what we’re doing, we can improve the computation time to
find the BNS foxes in the dense forests of the bipartite graph by factors of
trillions. But first, in Section 4.3 we must summarize some properties of the
bipartite graph, before we develop practical rules to locate the BNSs in
section 4.4.
The most obvious properties of any graph are its connectedness, its tree-
ness and its circuit-ness. Each of these has important consistency and
computability consequences and we will treat them in this order.
In order to focus on the structure of graphs, let us repeat these three
important definitions from Chapter 3:
Definitions 15: If a pair of relations lie in subgraphs such that there
exists at least one path of edges connecting them, they lie in a connected
component. If the pair of relations lie in subgraphs with no path connecting
them, they lie in disconnected components.
Definition 16: A tree is that structure within a connected component of a
graph such that there is exactly one path connecting every pair of vertices.
Definition 17: A circuit cluster is that structure within a connected
component such that there are two or more independent paths connecting
every pair of vertices. Independent paths have no vertices in common except
their end points.
Examples of definitions 15, 16 and 17 are provided in Figure 3-5. Let us
first address an automatic algorithm to determine the connected components
of a graph.
Definition 26: A node and knot are adjacent when there is an edge
connecting them. A vertex is a separating vertex if its removal disconnects
the graph.
From the viewpoint of the constraint matrix, the existence of a relevant
element in the xth row and yth column denotes the existence of the xy edge
and thus the adjacency of the x node and y knot.
Definition 27: The connectedness algorithm, by repeatedly applying the
adjacency definition, will partition any graph into disjoint connected
components.
Definition 28: The separating vertex algorithm, by employing trial
eliminations of vertices and determining whether the remaining graph is still
connected, will locate the graph’s separating vertices.
These algorithms are exhaustive and will always work to determine the
connected components and separating vertices of any graph, including
bipartite graphs. However, it may sometimes be useful to employ more
4. Regular Relations 73
global results which are a function of only the total number of vertices and
edges and provide faster conclusions on some occasions. These results,
showing when the graph must be connected, may be connected, and cannot
be connected are presented in Figure 4-9.
Figure 4-9. Domains of connectedness for bipartite graphs as a function of the number of
edges, E, and the number of vertices, V. (Recall that V=K+N.)
more edge; thus V-E=1 is still true no matter how many more fragments are
added. QED1.
Since a tree provides only a single path between any vertex pair,
removing any vertex will interrupt that path and disconnect the tree into
separate components. QED2.
Theorem 13 provides us with a simple and powerful tool for the analysis
of trees within graphs, no matter how highly dimensional and complex they
become. Figure 4-10 presents examples of how it can be applied.
Figure 4-10. Theorem 13 is unerring in its ability to identify trees. Initial layouts of bipartite
graphs may frequently disguise the fundamental “tree-ness” of the structure, especially when
the dimension is large.
Definition 29: The tree algorithm for graph G with V vertices and E
edges: If V-E=1, then G is a tree; if V-E>1, then G is disconnected, if V-
E<1, then G has at least one circuit.
4. Regular Relations 75
For the purpose of determining the ei, the edges of G are arbitrarily
assigned directions.
Figure 4-10a. Both the circuit rank and the constraint potential of a bipartite graph will be
useful to locate the kernels of constraint in a complex model.
78 Constraint theory
Figure 4-11. A bipartite graph with 13 simple circuits. However, a simple computation of the
circuit rank, c(G)=E-V+1 reveals that there are only four independent circuits.
Now that we’re somewhat familiar with some aspects of “the math house
of graphs” we can extend the previous section’s results on general graph
theory to bipartite graphs and corner the BNSs which may be lurking deep
within the tangled web of models like Figure 4-8.
80 Constraint theory
Figure 4-12. Classification chart and Venn diagram for the structures of a bipartite graph.
Categories 1, 2, 3.1, 3.2.1, 3.2.2.1, 3.2.2.2.1 and 3.2.2.2.2 are mutually exclusive and
exhaustive.
First of all, let us treat the trivial case of trees which terminate with nodes
of degree = 1. These are really 1x1 BNSs and do not support the spirit of
building multidimensional models. In every case, they represent a relation
with a single variable (because the nodal degree = 1) and this variable can be
solved for prior to its incorporation into the total multidimensional model.
4. Regular Relations 81
In other words, these “nodal twigs” are single dimensional models and serve
only to clutter the total model. They can be easily identified since they are
the rows of the constraint matrix which contain only one element. Once that
single relevant variable has been determined, its value can be “absorbed”
into all its other relevant nodes as a constant, thus eliminated from the
constraint matrix and bipartite graph.
Next, we will prove a simple theorem about integers which will help
speed many of the proofs in this section:
n
Theorem 16: Let p1, p2, p3.... pi be a set of integers. If ¦p i > n then
1
there exists at least one pi such that pi>-1.
Proof: Consider the trivial case of n=1: obviously if p1>-1, then p1>-1.
Next, consider n=2, then as can be seen in the
diagram on the right, if p1+p2>-2, then either
p1>-1, or p2>-1. Assume that the theorem is
true for n=1, and we add a (n+1)th term equal
to -1 (the most stress-full case) to both sides:
thus we obtain: p1+p2+... pn -1> -n-1, which is
n
identical to ¦p i > n . Thus we have shown
1
that the theorem is true for n=1 and n=2, and
furthermore, if it is true for n, then it is also
true for n+1. QED.
“I believe that this theorem is the least interesting one I have ever seen,”
sulked the manager.
“I would tend to agree with you,” said the analyst. “That’s why I
frequently refer to it as the ‘trivial’ theorem and it’s not counted among the
‘Ten Easy Pieces.’ However, as you will soon see, it helps prove other
theorems more compactly which will enable us to find BNSs trillions of
times faster than the method you suggested earlier (the ‘merely’ method).
“Trillions?!” mumbled the manager under his breath.
We are now prepared to relate the properties of a general BNS to its
constraint potential and topological graph structures defined in the previous
section by way of these ten compact results:
Theorem 22: No BNS can lie across circuit clusters with a separating
vertex.
Proof: Assume G is a BNS which lies across circuit
clusters G1and G2 with either a node or a knot as a
separating vertex. If the vertex is a knot, then
p(G1)+p(G2)=-1; if it’s a node, p(G1)+p(G2)=+1. In either
case, by Theorem 16, one or the other of the subgraphs
has constraint potential equal to or greater than zero.
4. Regular Relations 83
Thus, at least one BNS lies within the subgraphs and no BNS can lie across
both. QED.
and thus there must be a BNS within one of the circuit clusters. QED.
Theorem 25: Every BNS is the union of adjacent circuits within a circuit
cluster. (The “inside out” BNS location theorem.)
Proof: Recall the taxonomy of Definition 35 and Theorem 15 which
listed an exhaustive and mutually exclusive set of graph structures. BNSs
cannot lie across connected components, can’t be a tree or have tree
appendages, or be linked by trees, and can’t be within circuit clusters linked
by separating vertices. The only remaining structure is the union of circuits
within a single circuit cluster. QED.
Refer to Figure 4-13 for examples of BNSs in graph structures. Note that,
although every BNS must lie across adjacent circuits, not every circuit or
union of circuits necessarily contains a BNS. This is another example of
84 Constraint theory
the non-symmetry of certain theorems which are not “if and only if”
theorems. This misunderstanding has been the largest cause of confusion
among students.
Figure 4-13. Example of Theorem 25, the “inside out” theorem. All BNSs are the union of
adjacent circuits and lie completely within a circuit cluster.
Figure 4-14. Examples of Definition 36. Sets of independent BNSs can have no members
which are linear combinations2 of other members.
2
From the standpoint of their linear algebra representations.
4. Regular Relations 87
Table 4-1. Ten Easy Pieces Summary; Cornering the BNS within the Bipartite Graph
88 Constraint theory
The analyst continued: “Now that we have cornered the culprit BNSs to
be inside circuit clusters, we must resort to both topological and constraint
properties of the bipartite graph to locate them precisely.”
“I presume that the constraint potential will be much more valuable
here,” the manager ventured.
“True,” agreed the analyst, “but not nearly as valuable as we would like.
For example, one would hope that if p(cc)<0, we could be assured that there
are no BNSs within that cc, and if p(cc)=0, then there are exactly p(cc)+1
BNSs within that cc. Unfortunately, we cannot reach either of these two
conclusions. Referring to Figure 4-15a, we see that a cc with p(cc)<0 can
still have BNSs and a cc with p(cc)>0 can have any number of BNSs equal
or greater than p(cc)+1. Look at the outer ring of the cc; because it is a
bipartite graph, p(outer ring)=0. We can form Y inner loops, each of which
increases p(cc) by 1, while at the same time we can form Z other inner loops
each of which decreases p(cc) by 1. Thus, Y additional BNSs are formed but
the p(cc) = Y-Z. Rearranging, we see that Y=p(cc)+Z and that the number of
BNSs=Y+1=p(cc)+Z+1. Since Z=0, we conclude: the number of
BNSs=p(cc)+1. (This is the proof of T26 promised above.)”
“No more use than that?” complained the manager.
“Well, we can squeeze out a little more: From the viewpoint of the
constraint matrix, if a cc has p(cc)>0 then it has N-K more rows than
columns and thus N!/(N-K)!K! nodal squares can be found within this
rectangle (Figure 4.15b). That’s a help, but beware; just because every NS
must contain at least one BNS does not allow us to conclude that there are at
least N!/(N-K)!K! BNSs in the cc. In Figure 4.15c, we see that a single BNS
may be common to two or more NSs, and in Figure 4.15d, we see there can
be more than one BNS in an NS.”
“So we must also resort to the topology of the cc. From T25, we know
that every BNS lies across the union of circuits and from T14, we know that
the number of independent circuits in a cc is its circuit rank, c(cc). This
allows us to establish the maximum number of BNSs in a cc as the power set
of the cc’s circuit rank:
Theorem 27: The maximum number of BNSs in a cc equals the power set
of the circuit rank of the cc: Max # of BNSs = 2c(cc)
“Thus, T26 sets the lower limit on the number of BNSs as a function of
the cc’s constraint potential and T27 sets the upper limit as a function of the
cc’s circuit rank,” the analyst concluded.
4. Regular Relations 89
Figure 4-15a. Demonstration of T26: the number of BNSs in a circuit cluster is equal or
greater than p(cc)+1.
“So what do you suggest a practical guy to do?” queried the manager.
“Constraint theory provides at least two complementary robust
procedures for locating the BNSs within the cc’s:
Figure 4-15 b-d. In a circuit cluster with p(cc)>0, it is easy to determine the number of nodal
squares within it, but this does not lead directly to the number of basic nodal squares.
The brute force procedure develops the power set of the N relations –
each one being a submodel – and tests for p(submodel)=0 and that there are
no submodels with p(submodel)=0 within it. There will be 2N such
submodels to test and the testing of each one is simple in the extreme. True,
we seem to be trapped again in an exponential number of trials but in general
the number of nodes in even the largest of circuit clusters will be far smaller
4. Regular Relations 91
than the number of nodes in the total model. For example, if the total model
had 100 relations, it would be very unusual to have a circuit cluster as large
as 30 relations. The computational advantage of performing brute force
testing within a cc rather than within the total model equals 2100/230 ~ 1021 –
truly enormous! Assuming that each test can be accomplished in a
nanosecond, the brute force examination of each of the cc’s submodels
would require only a second.
The T25 procedure offers greater efficiency as the number of relations
within the cc grows much larger than 30. By definition, it makes use of the
fact that every BNS is the union of adjacent circuits. The first step in this
procedure is to develop the set of independent circuits within the cc, which
by T14 is equal to the cc’s circuit rank, c(cc). Then examine the submodels
which reside on these circuits – including the twig knots which may be
attached – one at time, then two at a time with adjacent circuits, then three at
a time with adjacent circuits, etc. The disadvantage of this procedure is that
we must find the independent set of circuits and form the potential sets of
combinations of these circuits, before we test for zero constraint potential of
any of the submodels. The advantage of this procedure is that, instead of
searching through the power set of the N relations of the cc, we are searching
through the power set of the cc’s independent circuits. Assuming again that
N=30 and that c(cc)=6 (an average of 5 relations per circuit), then the
computational leverage is about 230/26, a factor of several million. Of course
this must be weighed against the additional “overhead” of forming the
independent set of circuits and its power set.
Generally speaking, it is judged the brute force procedure would be more
practical for cc’s with N<30 if it could be accomplished on the order of a
second. As N grows to 50 and beyond for the cc, it would become
increasingly attractive to employ the T25 procedure. (Refer to Figure 4.16).
Let us now apply the definitions and theorems of the previous sections to
the challenge of finding BNSs within a model graph.
Step a) trims down the model by eliminating all nodes of degree one. These
twig nodes represent single-variable equations which can be solved
separately by employing the full model. Without this key topological
property, the cornerstone theorems (“ten easy pieces”) of constraint theory
could not have been proven. As a subset of the “ten easy pieces”, T-21
through T-25 are specifically employed in the D-37 process. As such, no
terminal nodes are allowed before the rules of constraint theory can be
applied. Not only are the terminal nodes trimmed from the model graph but
their incident edges, which would become “dangling”, must also be cleaned
up explicitly.
Once the relevant single variables (knots) of the above terminal nodes have
been determined, analytically or numerically, their values can be
incorporated into other relevant relations as coefficients. In effect, these are
1 × 1 BNS which only serve to clutter a multi-dimensional model and can be
temporarily eliminated. The solution time for this process is linear as each
and every node will be examined once, and only once, for checking of d(n) =
1 and removal as such. It should be clarified that the terminal nodes are only
temporarily eliminated, or hidden, for the purpose of topological
simplification while we are searching for other higher-order BNS in the
subsequent steps of D-37. Afterwards, these terminal nodes will still need be
re-integrated back into the original, complete model to check for possible
over-constraint among all BNS (Phan, pp. 101-104).
4. Regular Relations 93
Figure 4-16. DFS algorithm used to decompose the BPG into two connected components.
94 Constraint theory
Figure 4-17. Retain only the lower connected component with c(G) > 0.
4. Regular Relations 95
Step d) uses D-30 to trim repeatedly each spanning tree of every external
twigs, i.e. vertex with dC(v) = 1, since no BNS can exist in tree twigs by
T21. The symbol dC(v) is meant to emphasize the degree of a vertex within
the context of its parent component coming out of Step (c), and not that of
the spanning tree which is just a subgraph of the component. Even though
this step only operates on spanning trees, it is important to keep in mind that
the relative complement of each spanning tree must always be carried and
fully accounted for, as an imperative part of every component, to be
simplified and decomposed through every sequentially-related step within
the D-37 process (Phan, pp. 115-117).
Figure 4-18. D-30 used to trim external trees from a connected component with c(G) > 0.
96 Constraint theory
Step e) can employ one of several methods to identify and remove all
internal trees from every spanning tree of each component coming out of
Step (d) since no BNS can exist within internal trees by T23 and T24. A
spanning tree will be decomposed into smaller spanning trees, corresponding
to newly-formed respective sub-components. The relative complement of
each resultant smaller spanning tree will also be automatically identified as
by-products. In an edge-centric method, a spanning tree needs to be grown
only once for a component to save solution time (Phan, pp. 117-122). Other
methods to determine internal trees include those by Tarjan, Cormen et al.
(pp.558-559) and Gross & Yellen (pp. 182-184).
Figure 4-19. All internal trees to be removed from a connected component with c(G) > 0.
4. Regular Relations 97
Step f) identifies separating vertices of every BPG coming out of Step (e)
and partitions its kissing circuits, or circuit clusters, at such points. D-28
stipulates a trial-elimination procedure with polynomial runtime of 4(V·E),
where V and E are the numbers of vertices and edges, respectively (Shirey,
1969). However, a more efficient DFS-based algorithm with linear runtime
can be used to separate kissing circuits and circuit clusters. The notion of a
separating vertex in Constraint Theory (D-26) is identical to that of a cut-
vertex, cut-point or articulation point in Graph Theory. And the structure of
a circuit cluster in Constraint Theory is the same as that of a biconnected
component in Graph Theory (Phan, pp. 123-143). For each partitioned
circuit cluster, if p(cc)>0 then the cc has at least p(cc)+1 BNS within it.
Figure 4-20. Kissing circuits and circuit clusters to be separated at articulation points.
98 Constraint theory
Step g) can follow one of several methods to search for intrinsic BNS within
each cc coming out of Step (e). For smaller cc with N 14, the brute-force
approach is to examine its nodal power set, which would result in an
exponential solution-time of 2N (Phan, 2011, p. 381). For larger cc, a more
efficient method for locating potential BNS is to examine only the fruitful
unions of overlapping nodes, either directly or transitively, i.e. avoiding
combinations of non-overlapping nodes. The search for BNS should also be
systematically implemented in a bottom-up approach, smaller nodal unions
before larger ones. As such, smaller nodal unions identified as BNS, or BNS
containers, can be tagged to not be re-used as components in the construction
of any larger union. Such a larger union can never be a BNS and should not
be unnecessarily constructed and examined (Phan, pp. 144-231).
Figure 4-21. Search for BNS among simple circuits and within circuit clusters.
4. Regular Relations 99
Step h) validates the potential BNSs located by step (g) in simplified circuit
clusters, which are isolated sub-graphs of the overall original model graph.
These potential BNS must still be validated, as complete sub-models, within
the context of a connected component of the model graph identified in step
(b). The validation process is implemented by re-integration of tree twigs
pruned in step (d), internal trees removed in step (e) and circuit clusters
decomposed at separating vertices in step (f) (Phan, pp. 323-325).
“Surely, for those not familiar with D37, there must be other ways of
managing high dimension models on computers”, insisted the manager.
“Yes, there are ways,” admitted the analyst. “But they require far more
attention by the people developing the model and programming it than one
would like in this world of highly computer-automated analysis that we like
to believe we’re in. A pragmatic approach would be to start with models
whose constraint potential is enormously negative – which have perhaps a
100 or so fewer equations than variables. Then, once the computational
request is made known to the model manager, the potential computational
paths – which started out under-constrained – are examined to see where
variables can be set at constant values so that the constraint potential is
brought up to zero along the entire path. This would be very labor intensive
and fraught with the danger that the programmers would make these
‘constraint enhancing’ decisions without properly consulting those with the
responsibility of model fidelity. I have a very sad story in my past – among
many others I’m certain – wherein a programmer decided to add a key
constraint to the model without informing management, and as a result my
company lost a crucial contract we had worked on for years. Typically, the
math model results were presented to management with a minimum of
visibility.”
“Getting off the war stories for a moment,” said the manager changing
the subject, “doesn’t step f) of D37 leave us hanging a bit? It says that if
p(cc)>1 then the cc has multiple BNSs. What happens then?”
“Thanks for the question,” the analyst said. “We have spent most of this
chapter on the issue of the location of BNSs but, as crucial as that is, it is still
only part of the job of math model management. It’s time to broaden our
sights to the whole problem.”
4. Provide a shorter proof than that given in the text for Theorem 21,
using only the constraint matrix.
Let I(N) = d(N) 1; we see that this drives pr(N) to zero, and we have
just derived the “(d 1) in / 1 out” rule for nodes.
Figure 5-1. The Goldilocks rules for constraint flow through a network. In order for a
computational request to be allowable, the resultant constraint must equal zero at every vertex
and circuit along the computational paths from independent variables and variables held
constant to the dependent variable.
5. Consistency and Allowability 103
Next, examine the knots of Figure 5-1. The intrinsic constraint potential
of a knot is by definition: Pi(K) = N K = d(K) 1. If this knot produces
O(K) outputs, then the resultant constraint potential of the knot is decreased
by O(K): pr(K) = pi(K) – O(K). Recognizing that O(K) + I(K) = d(K), we see
that setting O(K) = d(K) 1 will drive pr(K) to zero. Thus we have just
derived the “1 in / (d 1) out” rule for knots.
Finally, examine the circuits on the Figure 5-1 computational paths. As
we have seen above, if p(circuits) > 0, we will have multiple BNS and
therefore over-constraint on their common variables. On the other hand, if
p(circuits) < 0, then we will experience under-constraint. If p(circuits) = 0,
then a single BNS will provide constraint at all its variables, permitting the
flow of constraint though these circuits.
In summary, the “Goldilocks” rule stating the necessity of zero constraint
all along the computational path is a unifying concept for computational
allowability.
the graph completely (the {Zo} constants will be “absorbed” into their
relevant nodes as parameters in their equations rather than variables). This
action will tend to simplify the model and sometimes even disconnect it,
making computations across disconnected components impossible. See
Figure 5-2.
Figure 5-2. Applying extrinsic constraint by holding some variables constant tends to simplify
the bipartite graph. Sometimes the graph is disconnected into separate components, rendering
computational requests involving variables in different components unallowable.
5. Consistency and Allowability 105
Next, start the flow of extrinsic constraint from each of the {Y}
independent variables into the remainder of the bipartite graph. Several
circumstances may occur, as is shown in Figure 5-3.
“So, I was referring to the nature of the world as well as the nature of
model builders. Since the natural world itself is fundamentally consistent,
over-constraint is invariably contributed by the fragmented understanding of
the human model builders, either by inadvertently applying excessive
relations to the description of a phenomenon or, more frequently, adding too
5. Consistency and Allowability 107
Steps (a) – (f) address the issue of model consistency, and steps (g) – (k)
that of computational allowability. The following sections of this chapter
will examine the above general procedure in more details, step by step, with
the goal of realizing the utility of Constraint Theory. D-37 and D-38 will be
further extended, refined and improved into a set of more effective and
efficient algorithms, ready for implementation. As important as these issues
are to discuss, they do not represent exponential explosions of computational
time.
Step (a) of D-38 states that: “Locate the BNS by employing D-37. If no
BNS found, including 1 × 1 BNS (a twig node and its relevant knot) and
c(G) = 0, the model is inherently consistent. Skip steps (b) – (f), go to step
(g).”
Per T-25, all BNS exist within simple circuits or across clusters of
adjacent circuits. If a model graph G with circuit rank c(G) = 0, it has no
circuits within. As such, G contains only tree structures. By T-9, any set of
universal relations whose BPG has a tree structure is consistent. As a global
criterion, and a quick check, to distinguish between trees and circuit clusters,
T-13 asserts that “a connected model graph having V – E = 1 is a tree”. This
assertion comes directly from the definition of circuit rank (Phan, pp. 31-32).
Otherwise, after all the intrinsic BNS have been identified and validated
within the context of a connected component in step (h) of D-37, they need
be compared against one another for any direct over-lapping. To check for
over-lapping, a set of definitions and vectorial operations will be developed
and illustrated herein.
Definition 39: Two BNS are said to directly overlap if they share at least
one variable in common, i.e. the intersection of their knot sets is non-null. In
other words, the common variable(s) are said to be over-constrained (or
over-specified), which causes the parent model graph to be inherently
inconsistent.
Since these intrinsic BNS overlap one another, i.e. sharing and thus over-
constraining at least one common variable, G is inherently inconsistent and
no computational requests made on G are allowable.
Figure 5-5. Inherently inconsistent model graph with multiple intrinsic BNS
overlapping one another.
| N BNS |
charvec(BNS) = *i =1
charvec(ni ) (5 – 1)
5. Consistency and Allowability 111
charvec(n6) = ( 0, 0, 0, 0, 1, 1 )
charvec(BNS1) = charvec(n6) U charvec(n7) = (0, 0, 0, 0, 1, 1 )
charvec(n7) = (0, 0, 0, 0, 1, 1 )
charvec(n1) = ( 1, 1, 0, 0, 0, 0 )
charvec(n2) = ( 1, 0, 1, 0, 1, 0 )
charvec(n3) = ( 0, 0, 1, 1, 0, 0 )
charvec(BNS2) = charvec(n1) U . . . U charvec(n6)
charvec(n4) = ( 0, 1, 0, 1, 0, 0 )
= ( 1, 1, 1, 1, 1, 1 )
charvec(n5) = ( 0, 0, 0, 1, 0, 1 )
charvec(n6) = ( 0, 0, 0, 0, 1, 1 )
charvec(n1) = ( 1, 1, 0, 0, 0, 0 )
charvec(n2) = ( 1, 0, 1, 0, 1, 0 )
charvec(n3) = ( 0, 0, 1, 1, 0, 0 )
charvec(BNS3) = charvec(n1) U . . . U charvec(n5) U charvec(n7)
charvec(n4) = ( 0, 1, 0, 1, 0, 0 )
= ( 1, 1, 1, 1, 1, 1 )
charvec(n5) = ( 0, 0, 0, 1, 0, 1 )
charvec(n7) = ( 0, 0, 0, 0, 1, 1 )
Definition 41: The BNS matrix of a bipartite graph G, denoted :G, is the
rectangular array of binary numbers whose rows correspond to the BNS and
columns correspond to the knots in G. In :G, the entry ei, j = 1 if knot kj
KG has been specified by BNSi, and 0 otherwise.
Figure 5-6 presents the BNS matrix :G for the example BPG in Figure 5-
5. Note that each row of :G represents the characteristic vector of the
corresponding BNS.
112 Constraint theory
Knots
a b c d e f
BNS1 0 0 0 0 1 1
BNS2 1 1 1 1 1 1
BNS3 1 1 1 1 1 1
Definition 42: The overlapping factor between BNSi and BNSj, denoted
as Zi, j, is defined as the dot product between their characteristic vectors.
Symbolically,
Theorem 28: BNSi and BNSj are overlapping if Zi, j 1. They are not
overlapping if Zi, j = 0.
Proof: If two BNS share any common knot in ki then the corresponding
ith elements in both of their respective characteristic vectors are equal to one.
Therefore, the dot product between their characteristic vectors must be equal
to one, or larger, if there are one, or more, common knot between the BNS.
QED
Proof: The matrix :G has R rows representing all BNS and K columns
representing knots (variables) in G. There are (R – 1) + (R – 2) + . . . + 2 + 1
= R2 – R/2 pairwise comparisons of BNS. For each pairwise comparison of
BNS, all K columns must be examined to look for commonly shared, and
thus over-constrained, variable(s) between the two corresponding BNS.
QED.
1
Re-integrated terminal node
b BNS2 c 3
BNS1
At this point in the process, the model builder needs to examine the set of
over-constrained variables output by the algorithm in Definition 43. Any
inconsistency, i.e. directly overlapping intrinsic BNS, inherent within the
model graph must be resolved before proceeding with the D-38 process. This
requirement for reconciliation has been established in steps (b) and (d) of D-
38.
Steps (b) and (d) of D-38 aim to relieve over-constraint inherent within a
model graph by eliminating nodes in overlapping BNS. In their current form,
the procedures for these steps state:
1 3
(!)
Intrinsic Intrinsic
a b 5 c d
BNS #1 BNS #2
2 4
(a) Inherently inconsistent BPG with two intrinsic BNS over-constraining node 5.
1 e
3
a b 5 c d
Resultant
2 1 × 1 BNS 4
(b) Addition of knot e relieves over-constraint and renders BPG inherently consistent
1 e
3
Intrinsic Intrinsic
a b 5 c d
BNS #1 BNS #2
2 4
f
Figure 5-8. Adding knots to BPG relieves over-constraint and allows for
computational requests
5. Consistency and Allowability 117
Figure 5-9. Model graph with intrinsic BNS and resultant constraint
domains
a. For each knot discovered along the constraint flow path, add it to
Kcd.
b. For each vertex along the path, including the source BNS, remove it
from G via a vertex-deletion operation (Phan, pp. 101-102). This
may fragment the remainder graph into several disjoint connected
sub-graphs as evidenced from Figure 5-9. As a quick check for this
fragmentation, or disconnectivity, of the remainder graph, Friedman
[1967, p. 91] has proven that “a model graph with V vertices and
more than (V 2 4 ) – V + 2 edges is connected”.
c. If any over-constraint encountered during propagation, then G is
inherently inconsistent. Employ technique(s) outlined in Section 5.5
to relieve local over-constraints.
d. Re-process G through steps (a) – (h) of D-37 to identify a new set of
intrinsic BNS since the model topology may be unpredictably
altered as a result of step (c). Also repeat steps (1a) – (1c) after the
new set of intrinsic BNS has been validated.
2) Per step (e) of D-38, locate potential resultant BNS, and possibly
continue expanding the constraint domain further, by re-processing the
remainder subgraph(s) through steps (a) – (h) of D-37.
a. To facilitate computation, a vertex-deletion subgraph can be
represented by a remainder matrix (see Definition 46 below).
b. Re-apply the algorithm outlined in Definition 43 to detect
overlapping among newly located BNS and to assist the model
builder to resolve any inherent over-constraint among them.
c. Similarly to steps (1c) and (1d) above, if G is altered to relieve any
local over-constraint among the newly located BNS, it must be re-
processed through D-37.
d. Resume propagation of constraint from resultant BNS by re-iterating
steps (1) – (2c) until no more BNS can be located. As such, Kcd has
reached its maximum and cannot expand any farther.
3) Repeat steps (1) – (2) for each and every intrinsic BNS within G to
define its respective constraint domain.
4) Per step (d) of D-38, re-apply the algorithm outlined in Definition 43 to
detect overlapping among the constraint domains defined in step (3).
a. If any overlapping detected, then G is inherently inconsistent.
Employ technique(s) outlined in Section 5.5 to relieve over-
constraints.
120 Constraint theory
3 4
1 7
Intrinsic Intrinsic
a b f d g
BNS1 BNS2
2 8
6 5
c
Original model graph G with intrinsic BNS1 and BNS2.
3 Resultant 4
BNS3
Figure 5-10. Removal of intrinsic BNS and repeating D-37 locate resultant
BNS
• The flow of constraint is started from knot b of the intrinsic BNS, as the
source of internal constraint, into node n5. The vertices of this source
BNS are removed from the model graph via a vertex-deletion operation
(Phan, pp. 101-102) to simplify the graph during propagation.
• By the T-10 rule of d(n) – 1 in and 1 out for nodes, constraint continues
to flow from n5 into knot c, after which n5 is removed to further reduce
the graph.
• By the T-10 rule of 1 in and d(k) – 1 out for knots, constraint can
simultaneously flow from c into n3 and n4. However, as a matter of
practical implementation, the branches of propagation needs be executed
one of a time, in a single thread of execution. Parallel threads of
execution for multiple branches may result in various issues with
concurrency, e.g. unpredictable racing condition. And it would be
122 Constraint theory
• Again, by the T-10 rule of d(n) – 1 in and 1 out for nodes, constraint
continues to flow from n3 into knot d, after which n3 is removed to
reduce the graph even further.
• Again, by the T-10 rule of 1 in and d(k) – 1 out for knots, constraint can
propagate from d to n4.
• Without any outlet from n4, the flow of constraint stalls here. Remember
that the edge between n4 and c has already been deleted above, and no
longer exists in the remainder matrix. This violates the T-10 rule of d(n)
– 1 in and 1 out for nodes. Therefore, the original model can be
concluded as inherently inconsistent, and no computational requests can
be made on it.
1 3
Source
a b 5 c d
BNS
1 3
Source
a b 5 c d
BNS
1 3
Source
a b 5 c d
BNS
1 3
Source
a b 5 c d
BNS
Figure 5-11. Violation of the d(n) – 1 in and 1 out rule for nodes exposes an
inherently inconsistent model
124 Constraint theory
i) If the computational path does not reach the dependent variable then
examine the residue for BNS, using D-37, and continue with T-10,
followed by D-37 as necessary until the computational path can go no
further.
2 3
Model graph G
d a m
c s 5 6
e
4 p
Intrinsic 3 × 3BNS
a m
5 6
e
p
Vertex-deletion sub-graph H
Steps (h) and (j) of D-38 have pointed out two possible scenarios in
which a computational request becomes unallowable due to over-constraint
of a knot along the computational path. The following sub-sections will
discuss other classes of unallowable computational requests.
Figure 5-1 graphically illustrates this situation. If the two constants in the
bottom half of the graph are propagated first, then knots d and g will be
discovered as over-constrained as stipulated by steps (h) and (j) of D-38. At
this time, the procedure can be halted and the computational request declared
unallowable. However, if the two independent variables in the top half of the
graph are applied first as external constraints, then node a will be discovered
as over-constrained. Accordingly, the same computational request can also
be rendered unallowable and the procedure immediately halted.
Step (k) of D-38 states that: “If the path reaches the dependent variable
by employing all independent variables, without either over- or under-
constraint along the way, then the computational request is allowable. It is
acceptable to have local under-constraint elsewhere in the model. If the
dependent variable can be reached without having to propagate the
constraint externally imposed on one or more independent variables, then the
computation is unallowable”.
It has been previously illuminated and asserted (pp. 40, 45) that: “If the
projection has variables that are not relevant, these variables take on all their
possible values, and are therefore under-constrained. Allowability requires
that all the variables of the requested computation be relevant to the
projection of the total relation onto the computational sub-space.”
130 Constraint theory
a m
Independent variable
Resultant
5 6
BNS
e
p
Dependent variable
Independent variable
1
Per T-2, over-constraint must not exist anywhere in the model, for a
computational request to be allowable, since any sub-model inconsistency
would “poison” the entire model. However, it is acceptable to have locally
under-constrained vertices outside of the computational path. For example,
in Figure 5-14, the computational request m = f2(a) is allowable even though
variables p and t are locally under-constrained elsewhere.
5. Consistency and Allowability 131
Independent variable
a m Dependent variable
5 6
e
p
constraint potential of that circuit cluster p(cc) is equal or greater than zero,
then that circuit cluster contains p(cc) + 1 BNSs. See Figure 5-15.
Once the BNSs are located, the model consistency and computational
allowability are easily determined, as is summarized in Tables 5-1 and 5-2,
the Constraint Theory Toolkit.
Figure 5-16 demonstrates a typical scenario for the flow of constraint
across a small bipartite graph model.
The many definitions and theorems of Chapters 4 and 5 may appear
onerous but they are necessary to establish the rules of D-37 and D-38 as
applicable to any mathematical model of any size. This is a demonstration of
the power of generalizability of mathematics. We start in tiny domains of
low dimension which we can comprehend, and then extend our
understanding and tools to dimensions of any size. As large as the numbers
2N and 2K are, the number of possible topologies of a bipartite graph is
unimaginably larger: 2NK.
5. Consistency and Allowability 133
Add the sources of extrinsic constraint – independent variables and variables held
constant – to the sources of intrinsic constraint and propagate constraint through the
BPG using the propagation rules above. If overconstraint occurs at any vertex, the
computation is NOT ALLOWABLE. If the constraint does not propagate to the
dependent variable, using the above procedure, search for BNSs which can continue
the propagation. If the constraint flow still does not reach the dependent variable, the
computation is NOT ALLOWABLE due to underconstraint. If neither over- or
underconstraint occurs, the computation is ALLOWABLE.
defined by the logical equation, these forbidden domains take priority. This
convention permits the simplification of logical equations which is important
in the logical design of digital computers. In accordance with the
hyperspatial viewpoints discussed in Chapter 2, when the total model is
defined by a set of simultaneous Boolean equations, the intersection of the
Boolean equations and the union of the forbidden areas are taken.
Both the truth table and Veitch diagram representations suffer from
implications of independent/variables and forbidden zone ambiguity.
Figure 6-2d displays the most complete representation of any discrete
function: the identification of the set of points within the hyperspace of
relevant variables. For any binary Boolean equation, this set can reside only
at the corners of the hypercube formed by the points 0 and 1 along the axis
of every relevant variable. That is, these are the values of the variables that
are allowable by the defined relation. In the spirit of the four fold way
presented in Chapter 2, call the set of points that satisfy the equation “A.”
Let us now examine how the computational rule: (d-1)in/1 out that was
developed for regular relations applies to Boolean relations.
Figure 6-3 displays all possible computational paths through the relation
A=BC. It shows that for B and C taken as input variables, every one of the
four possible cases, the (d-1)in/1 out rule works. (This should not be
surprising , since the format of the Boolean equation, truth table and Veitch
diagram were predisposed towards this computational flow.) However, in the
cases of A,B or A,C as input variables, it is seen that only half of the cases
follow the (d-1)in/1 out rule, while the other half result in either multiple
answers or represent forbidden inputs. Noting that multiple results also can
occur in regular relations and can still be carried forward to other
downstream computations, and that regular relations were defined to exclude
forbidden inputs, it can be concluded that the (d-1)in/1 out rule weakly
applies to discrete functions.
Figure 6-3 explores the more extreme case – from our regular relation
perspective – of a single input variable. There exists one case out of six
where the single variable A, set equal to 1, can uniquely determine the
values of B and C. This is a case of 1 in/(d-1)out for nodes, which is
surprising if we are used to thinking only of regular relations.
In summary, it can be seen that the computational flow across discrete
relations depends far more on the precise nature of the relation than on
metamodel rules such as (d-1)in/1 out.
Next, let us examine the applicability of constraint potential, (N-K), to
determine the consistency of Boolean relations with K variables and N
relations.
As noted above, a Boolean relation of K variables can exist only at the
corners of a hypercube formed by the coordinates 0 and 1 along each of the
K dimensions of the relation. Therefore, each of the N relations can be
represented by a “binary number” containing 2K states of A’s for points
contained by the relation and empty spaces which do not contain the relation.
Figure 6-4 provides the general format of a Boolean model. Note that,
despite the fact that it is a rectangular array with the rows referring to
variables and columns referring to relations, this is not a constraint matrix as
142 Constraint theory
defined in chapter 2. The total number of rows equals the number of possible
states of any given Boolean function, and equals 2K, which is the power set
of the K states. The total number of columns equals all possible Boolean
functions which is the power set of the 2K possible states and equals
2^(2^K). In other words, the number of columns is the power set of the
number of states which is the power set of the number of variables.
Proof: Recall that, from Chapter 2, the model will be inconsistent if the
total allowability set is the null set. Also recall that the intersection of a null
set with any other set results in a null set. By examining Figure 5-4 (b), we
note that exactly half of the columns can be paired with the other half such
that their intersections are null sets. (Specifically, look at columns 1 and 16,
2 and 15, …etc.) Also note that no other columns have null set intersections.
Therefore the maximum number of relations which will not result in a null
set is exactly half the total number of columns. In short, the maximum N
which assures consistency is half the number of columns, or: Nmax = ( ½ )(
K K K
22 ) = ( 2-1 )( 22 ) = 22 1
. QED.
144 Constraint theory
Figure 6-5. The Constraint Potential for Boolean relations is weaker than for regular relations.
Thus, the constraint potential (N-K) for regular relations can be replaced
2K 1
by (N- 2 ) for Boolean relations, and as is shown in Figure 6-5, it is a
weaker indication of the model’s inconsistency. Once again, we see that the
specific properties of the relations must be examined and the metamodels, so
useful in Chapters 4 and 5, are less useful for discrete relations.
Additional properties and methods of discrete relations are therefore
discussed in the following.
Figure 6-9. Comparisons between pure inequality equations, equals or greater (less) than
relations, and boundary sets.
It was previously noted that, for regular relations, the BNS was required
to provide intrinsic point constraint, and for discrete relations, intrinsic point
constraint was always provided. We are now prepared to summarize the
point constraint situation for all classes of relations:
Theorem 35: In discrete relations, intrinsic point constraint always
occurs; in regular relations, intrinsic point constraint occurs in the presence
of a BNS; in inequality relations, intrinsic point constraint never occurs.
Proof: The first portion is true by T31; the second is true by T11; the last
is true because inequality relations do not include their boundary sets and
thus all intersections either include an infinite number of points or the null
set. QED
It is noteworthy that in a broad range of applications of math models, it is
the boundary set which is of the greatest interest. An example of this is the
general optimization techniques of linear programming where the
optimization of a linear criterion occurs only at the intersection of the
6. Discrete and Interval Relations 151
Figure 6-10. The number of inequality constraints can increase to infinity without driving the
model’s relation set to null (inconsistency).
6.7 SUMMARY
A Compact Summary
7.1 OVERVIEW
7.3 DEFINITIONS
The forty definitions employed in Chapter Two through Six are listed in
Table 7-2.
Most of these definitions either contribute to other, more complex
definitions, or to theorems. In seven cases – marked by an asterisk – the
definitions are a result of a set of theorems and represent a computational
procedure.
7.4 THEOREMS
The thirty-three theorems derived in Chapters two through five are listed
in Table 7-3.
The proofs of all these theorems – except T12, “Hall’s Theorem” – are
provided as they are first introduced in the chapters. The proof of Hall’s
theorem can be found in the reference.
7. The Logical Structure of Contstraint Theory 157
7.6 COMPLETENESS
The author makes no claim that these summaries and Constraint Theory
are complete.
Indeed, it his hope that this work will stimulate further research and study
into the increasingly important objective of managing complexity.
7. The Logical Structure of Contstraint Theory 159
The examples provided in this Chapter are drawn from applications in the
aerospace industry that the author has experienced. Rather than apologize for
this apparently undemocratic representation, it should be noted that it was
the aerospace industry which gave the first major fertile ground for the
discipline of systems engineering to manage the ever increasing complexity
of integrating new technologies, equipment and missions.
The first example applies constraint theory to the operations analysis of
new systems; it is an elaboration of the example presented in Chapter 1. The
second example deals with the kinematics of free-fall weapons, which,
despite the advent of “smart bombs” still has utility in many operational
scenarios. The third example is a dynamic analysis of the control of the
trajectory of an asteroid employing mass drivers.
Figure 8-1. US Air Force Research Laboratories Briefing chart on Cost as an Independent
Variable.
At this broad level of detail, the math model employed is similar to the
example of Chapter 1. Mathematically, this type of problem is ideally suited
to the viewpoint and tools of Constraint Theory. Simply stated, a model is
built, and rather than follow the original computational flow from
performance specifications to cost, the flow is reversed; the independent and
dependent variables are switched. The new problem can be stated as:
“Given budgetary constraints on newly developed systems, what is the
optimum system design which provides the most acceptable performance
without exceeding these budgetary constraints?”
8. Examples of Constraint Theory 165
Let us flesh out in more detail the AFRL chart to attain a model which
can perform the desired analysis. Table 8-1 presents a list of variables; note
that K=14 across three levels of detail. Table 8-2 lists the relations, with
N=16. Noting that N>K, we should be immediately concerned with model
consistency.
However, the 9 variables at the bottom of the graph are all inter-related,
as is shown in Figure 8-3 where 6 additional nodes are added to indicate
these relations. For example, node 7 describes how the cost of development,
CD, influences the radar signature, and node 9 describes how CD influences
the accuracy (CEP) of weapon delivery. Although we now have accumulated
several circuits, the total constraint potential is still less than zero, and there
are no local domains where the constraint potential exceeds zero – thus,
there are no BNSs. This can be verified by examining Figure 8-5 which
displays the constraint matrices for the various levels of detail in the model.
The constraint situation gets far more serious when the final 5 nodes at
the bottom of Table 8-2 are included in the model. Note that each of these
last five nodes are “policy demands” specifying additional requirements on
reliability, weight, radar signature, etc., without directly relating them to a
higher level system criterion. The addition of these is devastating to the
consistency of the model. As is shown in Figure 8-4, the application of nodes
168 Constraint theory
For the sake of the example, only the kinematics within the bomb trajectory
plane are considered. All the variables listed previously except X are
measurable by at least one type of airborne instrument. We may now state
the problem more specifically.
. . .
.
Which subsets of (X, Z, Z, r, r, S, S, V, d) form complete descriptions of
the aircraft in-plane kinematics? Call these describing sets.
. .
From physical reasoning, the immeasurable set (X, X, Z, Z) completely
defines the kinematics
. . where (X, Z) define a two-dimensional positional
vector, and (X, Z) define a two-dimensional velocity vector. Assuming for
the moment that every describing set has four variables, then
9!
C49 = = 126
4!(9 4)!
1: r = (X. 2 + Z2.)1/2
.
2: r = (XX + ZZ)(X2 + Z2)-1/2
3: X = r cos S
4: Z = r sin S
-1
5: . (Z/X)
S. = tan . 2
6: S = (ZX.2 – XZ)(X
. + Z2)-1
2 1/2
7: V = (X. + Z )
8: Z = -X tan d.
Each relation is correct. More correct relations can be written, but the
preceding list seems to be sufficient, whatever that means. Normally, the
analyst must rely on his intuition or judgment regarding the point at which
he should cease adding relations to his model. The bipartite graph and
constraint matrix meta models are presented in Figure 8-7.
The fundamental criterion for math model consistency is that the total
model relation set is not the null set. This simply means that the set of points
in the total ten-dimensional space which satisfied every one of the eight
relations must include at least one point, otherwise some part of the model is
intrinsically incompatible with another part. From the constraint point of
view, model inconsistency is due to overlapping domains of intrinsic
constraint.
Each relation in the example above is an n-1 dimensional regular surface
where n is the dimension of the relation. Thus they are all regular relations,
and the source of intrinsic constraint in this type of model is always a basic
nodal square (BNS), a special subgraph of the total bipartite graph which is
recognized by an equal number of vertices of each type. A logical first step
in the consistency testing is to search the meta models for BNSs, thereby
locating constraint sources.
Since for even this small model, a systematic search through all possible
submodels involves examination of 256 cases, it is clear that more powerful
means must be used to search for the BNS.
Every BNS is the union of simple nodal circuits. We know that every
BNS must be a circuit cluster and can contain no treelike appendages. . .
Examination of either Figures 8-7a or 8-7b reveals that the variables r, s, v,
and d form trees attached to the main circuit cluster; therefore, the nodes 2,
7, 8, and 6 cannot be part of a BNS. Temporarily stripping these nodes
(“trimming the trees”) from the model, we get the circuit cluster depicted in
Figure 8-8.
174 Constraint theory
Since Figure 8-8 has the same number of each type of vertex (constraint
potential equal zero), it is a nodal square. Since every node has a degree of
three, there cannot exist smaller nodal squares of dimension 2x2 within
8. Examples of Constraint Theory 175
Since the early 1990’s there has been an increasing awareness of the risk
to earth by a strike from a large comet or asteroid. It is estimated that if such
a near-earth object (NEO) of 1 km or larger struck the earth it would disrupt
our global ecology and cause a billion human casualties. Much has been
8. Examples of Constraint Theory 177
However, much less attention has been paid to the deflection or intercept
of these threats to human civilization. The methods which provide the
highest energy density delivered to the threatening object are nuclear. But
there is great resistance to testing and putting nuclear devices into space, the
bomb may split the NEO in unpredictable ways, it may pollute the
atmosphere, and the coupling of the energy to the NEO is critically
dependent on the NEO’s structure and composition.
The most robust and lowest technical risk approach is to employ
traditional chemical propulsion to the NEO by attaching a conventional
rocket to the object and transfer orbit-changing momentum. But the cost of
this approach is enormous; not only must the chemical energy be delivered
to the NEO, the reaction mass must be transported there also. Imagine a
Saturn V rocket engine firing continuously for a year or more!
A concept which has received far less attention is to employ a mass
driver – essentially a linear motor which converts electrical energy into
kinetic energy. Once a mass driver is transported and installed on the NEO,
the required energy can be obtained with a solar collector and the required
reaction mass can be obtained from the NEO itself. The following analysis
178 Constraint theory
employs constraint theory to perform trade studies of mass driver designs for
NEO deflection.
Refer to Figure 8-10 for a schematic of the mass driver installation on the
NEO. The NEO, or asteroid, has mass M and diameter da. A solar collector
of diameter ds collects energy for the mass driver. The mass driver is a long
coil which propels mass from the NEO in buckets with a velocity, v. The
8. Examples of Constraint Theory 179
action of the mass driver ejecta causes a reaction by the asteroid, moving it
in the opposite direction with velocity, V. In order to miss the earth within a
time-to-go, tgo, the NEO must move a distance S – normally considered to be
the earth’s radius. (Generally, the optimum application of force for orbit
changing is parallel to the asteroid’s velocity vector. However this analysis
stresses the case of final approaches with short time-to-go and small angles
between the orbits of earth and asteroid.)
The kinematic relations can be captured by the bipartite graph in Figure
8-11. The variables are:
s=f1(a,t)=(1/2)at2 Vf=f2(a,t)=at
Figure 8-12. NEO Kinematics, NEO energy required and solar energy available.
It would be tempting to assume that all the energy collected by the solar
array is available to provide the asteroid with the required energy, that is, the
relation designated as “*” in Figure 8-12 is simply: Er=Es. Unfortunately,
this is far from correct, since both energy and momentum relations must be
satisfied:
8. Examples of Constraint Theory 181
Energy: Es=(1/2)MVf2+(1/2)mv2
Momentum: MVf=mv
Define a new variable, k=v/V=M/v – that is, the ratio of asteroid mass to
ejecta mass, or equivalently, the ratio of ejecta velocity to asteroid velocity.
Substituting k into the energy equation yields:
In other words, only 1/(1+k) of the available solar energy couples to the
asteroid, the remainder accelerates the ejecta.
Paradoxically, the situation gets worse as the ejecta velocity increases
and the time-to-go increases. These results run counter to the “general
wisdom” that high ejecta velocity is good – after all, much hard work has
been expended in driving up the exhaust velocity of rocket engines and mass
drivers – and that the more time we have for deflection, the better off we are.
The general wisdom is still correct in their original domains and problem
statements, but does not precisely apply to the issues of asteroid deflection.
All these results can be summarized on a single chart, Figure 8-13.
Following the standard conventions employed to compare NEO deflection
concepts, the two fundamental independent variables are time-to-go and
asteroid diameter. The dependent variable is the diameter of the solar
collector, ds. At the center of the chart it can be seen that a solar collector of
only 100 meter diameter is required to deflect a 1 kilometer asteroid if one
year of time-to-go is available. The chart also presents an additional relation:
the locus of ds=da. On this locus, the solar collector is the same size as the
asteroid and the time-to-go becomes the dependent variable.
In summary, although the number of variables of this model was not
exceptionally high, the constraint theory methodology provided useful
kinematic and dynamic insight, as well as managing the computational flow
through a confusing set of equations.
182 Constraint theory
“These are not very brief ‘gists,’” complained the manager. “Haven’t you
heard the saying: ‘That which is good and short is doubly good’?”
“Yes, I have,” agreed the analyst, “but brevity depends on a shared
experience – which we didn’t have originally – and a deep understanding of
the language we’re using. Normally, the richer the language, the briefer the
meaningful messages can become. If nothing else, this book has introduced
you to new directions of mathematics, which is a special type of language.”
“Speaking of language, it can reasonably be argued that constraint theory
sits on a bridge between mathematics and qualitative language. All
languages have built into their grammatical structure the primitive concepts
of “none,” “one” and “many.” – I have no horse, I have one horse, I have
many horses. The rhythm and pivotal concepts of constraint theory follow in
analogous ways:
“But it’s regrettable, isn’t it, that the bipartite graph couldn’t be as useful
for discrete and interval relations as it is for regular relations?”
“I do wish it would be more broadly applicable,” agreed the analyst, but
can you name me any branch of mathematics that is equally useful over all
applications? Math tends to be very specialized into specific application
domains. Consider the vast body of math that analyzes linear systems,
despite the fact that most real-world systems are nonlinear. On the bright
side, regular relations presently represent the most common type of relation
used in math modeling, so if it’s limited to a domain, at least it’s the most
significant domain.”
“As I promised in Chapter One,” continued the analyst, “Constraint Theory
really occupies just a simple portion of all mathematics. Consider John
Barrow’s ‘Structure of Modern Mathematics’ [16] shown in Figure 9-1,
where Constraint Theory is represented only in the simplest lower right-hand
corner. We made substantial use of bipartite graphs which are with graph
theory and closely related to hypergraphs since the constraint matrix of a
9. Manager and Analyst Meet Again 185
“Do you think it a coincidence that the list of technologies are generally
relevant to the list of programs?” asked the analyst.
“Not if you realize that we’re really the same guy,” both observed. “And
the dialogues we used were merely an expository mechanism to describe a
new field and illuminate unfamiliar concepts. We were attempting – without
daring to be compared to them – to use dialogue in the manner of Plato,
Galileo and even Mark Twain in his philosophical work, ‘What is Man?’
Struct- K N E C B # # %
ure Examined Allowable Allowable
A 6 4 12 3 0 50 13 26%
B 6 4 11 2 0 50 12 24
C 6 4 10 1 0 50 9 18
D 8 6 18 5 1 154 6 4
E 8 6 18 5 0 154 23 15
F 8 6 17 4 1 154 6 4
G 8 6 16 3 1 154 6 4
H 8 6 15 2 1 154 6 4
J 8 6 14 1 1 154 3 2
K 8 6 13 0 0 154 6 4
TOTALS 1228 90 7.3%
LEGEND:
(Gross & Yellen, 2006, pp. 687 & 690; Anton, 1977, p. 127)
¾ a • ( b + c ) = ( a • b + a • c ) and ( a + b ) • c = ( a • c ) + ( b • c ).
Definition D-8: A vector space over a field (of scalars) F is a set V (of
vectors) together with an operation + on V and a mapping, called scalar
multiplication from the Cartesian product F × V to V ( (a, v) Æ av ), such
that the following conditions are satisfied for all scalars a, b F and all
vectors v, w V:
¾ ( V, + ) is an abelian group, where the notation “+” is being used to
denote both addition of scalars in field F and addition of vectors in
set V.
¾ ( a • b ) v = a ( bv )
¾ ( a + b ) v = av + bv
¾ a ( v + w ) v = av + aw
¾ ev = v, where e is the multiplicative identity of field F.
BINARY SET OPERATIONS (Gross & Yellen, 2006, pp. 197-198)
Appendix D 205
Definition D-9: Let s1, s2, . . . , sn be any sequence of objects, and let A
be a subset of S = { s1, s2, . . . , sn }. The characteristic vector of subset A,
denoted charvec ( A ), is the n-tuple whose jth component is 1 if sj A, and 0
otherwise.
In the model graph G of Figure D-1 below, EG = { e1, e2, e3, e4, e5, e6, e7 }.
Cycle A has as its edge set EA = { e1, e2, e3, e4 }, and cycle B has as its
edge set EB = { e3, e5, e6, e7 }, both of which are subsets of EG. By Definition
D-9 above, charvec ( EA ) = ( 1, 1, 1, 1, 0, 0, 0 ), and charvec ( EB ) = ( 0, 0,
1, 0, 1, 1, 1 ).
Characteristic Edges
Cycle Vector of
Edge Set e1 e2 e3 e4 e5 e6 e7
A charvec (EA) 1 1 1 1 0 0 0
B charvec (EB) 0 0 1 0 1 1 1
Definition D-10: For graphs A and B, the union of their edge sets EA and
EB is the set of all edges which are either in EA or in EB, or both.
Symbolically, e EA U EB if e EA or e EB.
In Figure D-2 below, edge set EA = { e1, e2, e3, e4 } and edge set EB = {
e3, e5, e6, e7 }. Thus, their union set EA U EB = { e1, e2, e3, e4, e5, e6, e7 }. Note
that charvec(EA U EB) is formed by combining the respective components of
charvec(EA) and charvec(EB) with the bitwise inclusive OR operator. In
C/C++, this operator has the syntactical symbol of “ | ”, e.g. ( 0 | 0 ) = 0; and
( 0 | 1 ) = ( 1 | 0 ) = ( 1 | 1 ) = 1. In this case, charvec ( EA U EB ) = ( 1, 1, 1,
1, 1, 1, 1 ).
Characteristic Edges
Cycle
Vectors e1 e2 e3 e4 e5 e6 e7
A charvec (EA) 1 1 1 1 0 0 0
B charvec (EB) 0 0 1 0 1 1 1
charvec ( EA U EB ) 1 1 1 1 1 1 1
Definition D-11: For graphs A and B, the intersection of their edge sets
EA and EB is the set of all edges which are in both EA and EB. Symbolically:
e EA EB if: e EA and e EB .
In Figure D-3 below, edge set EA = { e1, e2, e3, e4 } and edge set EB = {
e3, e5, e6, e7 }. Thus, their intersection set EA EB = { e3 }. Note that
charvec ( EA EB ) is formed by combining the respective components of
charvec ( EA ) and charvec ( EB ) with the logical operator AND where 1
means true, and 0 means false. Symbolically, ( 0 AND 0 ) = ( 0 AND 1 ) = ( 1
AND 0 ) = 0; ( 1 AND 1 ) = 1. In this case, charvec ( EA EB ) = ( 0, 0, 1, 0,
0, 0, 0).
Characteristic Edges
Cycle
Vectors e1 e2 e3 e4 e5 e6 e7
A charvec (EA) 1 1 1 1 0 0 0
B charvec (EB) 0 0 1 0 1 1 1
charvec ( EA ∩ EB ) 0 0 1 0 0 0 0
As demonstrated in Figure D-4 below, edge set EA = { e1, e2, e3, e4 } and
edge set EB = { e3, e5, e6, e7 }, therefore the difference set EA EB = { e1, e2,
e4 }. However, the difference set EB EA = { e5, e6, e7 }.
Definition D-13: For a graph G, let WE(G) denote the set of all subsets of
EG, i.e. the power set of all edges in G. The ring sum of two elements E1 , E2
WE(G) is defined as:
E1 E2 = ( E1 E2 ) U ( E2 E1 )
Figure D-5 below illustrates graphically an application of ring sum in
which two cycles A and B are combined to form a new cycle. Note that
charvec ( EA EB ) is formed by combining the respective components of
charvec ( EA ) and charvec ( EB ) with the bitwise exclusive OR operator. In
C/C++, this operator has the syntactical symbol of “ ^ ”, e.g. ( 0 ^ 0 ) = ( 1 ^
1 ) = 0; and ( 0 ^ 1 ) = ( 1 ^ 0 ) = 1.
13. K. Devlin, The Math Gene, Basic Books, New York, NY, 2000
19. Ivars Peterson, Math Trek, Science News Online, April 17, 1999
23. Tarjan, Robert Endre (1972). Depth-first search and linear graph
algorithms. SIAM Journal on Computing, Vol. 2, pp. 146-160.
References 213
26. Gross, Jonathan L. and Jay Yellen (2006). Graph Theory and Its
Applications. Boca Raton, FL: Chapman & Hall/CRC.
27. Tarjan, Robert Endre (1974, March). A note on finding the bridges
of a graph. Information Processing Letters, Vol. 2, No. 6, pp. 160-
161.
29. Anton, Howard (1977). Elementary Linear Algebra. New York, NY:
John Wiley & Sons, Inc.
Index
Abbot, 40 Boolean
abelian group, 203 algebra, 137
acceleration of asteroid, 179 equations, 138
accuracy, 167 relations, 138
adjacent, 72 Boolean equations, 141, 147
adjacent circuits, 89 Boolean relations, 141
aerospace industry, 163 Bourbaki, 31
Air Force Research Laboratories (AFRL), Brahe, 26
164 brute force, 124
airborne radar, 185 brute force procedure, 89, 90
aircraft, 171 budgetary constraints, 164
aircraft kinematics, 171 Burger, 40
aircraft velocity, 171 cardinals, 137
allowability, 147 Cartesian coordinates, 30
space, 152 charvec (characteristic vector)
subset, 152 definition, 110
allowable, 7, 40, 124, 156 chemical energy, 177
space, 138 chemical propulsion, 177
ambiguity, 141 circuit cluster, 56, 72, 79, 107, 131, 147,
analysis, 137 184
angle rate, 171 circuit rank, 75, 76, 78, 138, 184
applied mathematics, 185 Circuit Vector Addition, 75
articulation point, 97 circuit vector space, 75
artificial intelligence, 185 circuits, 105
asteroid, xiii, 163, 176, 179, 180 cognitive limitations, 124, 191
avionic equipment, 171 comet, 176
Barrow, 184 complex systems, xi
Basic Nodal Square, 65 computability, 46
biconnected component, 97 computational
binary operation *, 203 allowability, xii, 108
bipartite graph, xii, 8, 39, 46, 88, 103, flow, 141, 181
105, 108, 131, 157, 166, 168, 173, paths, 141
183, 184, 185, 197 request, xii, 66, 103, 105, 124, 156,
BNS, 65, 68, 79, 83, 105, 157, 173 179
detection of overlapping, 109 request disappointments, xiv, 189
kernels, 79 computational allowability, 103
location theorem, 157 Computer Assisted Engineering, 1
matrix, 111 computer program, xii
bombing problem, 171 computer science family, xi
bombing solution, 171 connected components, 54, 72, 184