Complex Systems Science
Complex Systems Science
Espoo 2004
Report 145
TEKNILLINEN KORKEAKOULU
TEKNISKA HGSKOLAN
HELSINKI UNIVERSITY OF TECHNOLOGY
TECHNISCHE UNIVERSITT HELSINKI
UNIVERSITE DE TECHNOLOGIE DHELSINKI
Report 145
Keywords: Chaos theory, complexity; agents, networks, cellular automata; fractals, power law
distributions; hierarchies, decentralization; emergence.
Distribution:
Helsinki University of Technology
Control Engineering Laboratory
P.O. Box 5500
FIN-02015 HUT, Finland
Tel. +358-9-451 5201
Fax. +358-9-451 5208
E-mail: control.engineering@hut.fi
http://www.control.hut.fi/
ISBN 951-22-7507-4
ISSN 0356-0872
Picaset Oy
Helsinki 2005
Preface
In Spring 2003, a postgraduate course on complex systems was organized at
HUT Control Engineering Laboratory.
Complex systems is a hot topic today: Complexity in technical systems and
in nature around us is overwhelming. The traditional modeling methods do
not seem to be powerful enough to capture this diversity, and it seems that
a paradigm shift is needed. It is complexity theory that promises to give us
tools for this. But what is it all about what is it now, and what are the
future possibilities? These issues are discussed in this report.
In addition to the printed version, this report is available also in PDF format
through the HTML page at the Internet address
http://www.control.hut.fi/hyotyniemi/publications
/04_report145/.
I am grateful to everybody who has contributed in this eort. All authors
are responsible for their own texts.
Contents
1 Introduction
1.1
. . . . . . . . . . . . . . .
1.2
1.3
Philosophical view . . . . . . . . . . . . . . . . . . . . 12
1.3.2
Top-down view . . . . . . . . . . . . . . . . . . . . . . 12
1.3.3
Bottom-up view . . . . . . . . . . . . . . . . . . . . . . 13
1.3.4
Systems view . . . . . . . . . . . . . . . . . . . . . . . 15
2 Complex Systems
17
2.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.2
2.3
2.2.1
2.2.2
Basic ideas
2.2.3
Cellular Automaton . . . . . . . . . . . . . . . . . . . . 18
2.2.4
Wolframs conclusions . . . . . . . . . . . . . . . . . . 19
2.2.5
Criticism . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.2.6
Provocation? . . . . . . . . . . . . . . . . . . . . . . . 20
. . . . . . . . . . . . . . . . . . . . . . . . 18
Networked systems . . . . . . . . . . . . . . . . . . . . . . . . 21
2.3.1
2.3.2
2.3.3
Networked structures . . . . . . . . . . . . . . . . . . . 22
1
CONTENTS
2.3.4
. . . . . . . . . . . . . . . 22
25
35
4.1.2
4.1.3
Hierarchic systems . . . . . . . . . . . . . . . . . . . . 37
4.2.2
4.2.3
4.2.4
4.3 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5 Towards Decentralization
43
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5.2 Intelligent agents that interact . . . . . . . . . . . . . . . . . . 45
5.3 Rationales for multiagent systems . . . . . . . . . . . . . . . . 46
5.4 Multiagent systems . . . . . . . . . . . . . . . . . . . . . . . . 47
5.4.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . 47
5.4.2
Motivations . . . . . . . . . . . . . . . . . . . . . . . . 48
CONTENTS
5.5.1
5.5.2
5.5.3
5.5.4
5.6
Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5.7
Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
6 Networks of Agents
6.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
6.1.1
6.2
57
Reductionism . . . . . . . . . . . . . . . . . . . . . . . 59
Knigsberg bridges . . . . . . . . . . . . . . . . . . . . 59
6.2.2
Random networks . . . . . . . . . . . . . . . . . . . . . 61
6.3
Degrees of separation . . . . . . . . . . . . . . . . . . . . . . . 62
6.4
6.5
6.5.2
6.5.3
6.6
6.7
6.8
Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7 Cellular Automata
73
7.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
7.2
7.2.2
7.2.3
7.2.4
CONTENTS
7.2.5
7.2.6
89
91
9.2.2
SOC . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
9.2.3
9.3.2
9.3.3
9.3.4
109
CONTENTS
123
. . . . . . . . . . . . . . . . . . . . . 124
. . . . . . . . . . . . . . . . . . . . . . . . . . . 129
139
159
CONTENTS
13.3.1 Rules of the beer distribution game . . . . . . . . . . . 163
13.3.2 Management ight simulators . . . . . . . . . . . . . . 166
13.4 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
13.5 Basic concepts of system dynamics . . . . . . . . . . . . . . . 169
13.5.1 Stocks and ows . . . . . . . . . . . . . . . . . . . . . 169
13.5.2 Causal loops . . . . . . . . . . . . . . . . . . . . . . . . 169
13.5.3 Equations behind the model . . . . . . . . . . . . . . . 172
13.5.4 Flight simulator . . . . . . . . . . . . . . . . . . . . . . 173
13.6 Literature and research institutes . . . . . . . . . . . . . . . . 175
13.6.1 System dynamics literature . . . . . . . . . . . . . . . 175
13.6.2 Research institutes . . . . . . . . . . . . . . . . . . . . 176
13.7 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176
14 Systems Theory
179
. . . . . . . 186
Session 1
Introduction
Heikki Hytyniemi
Helsinki University of Technology
Control Engineering Laboratory
heikki.hyotyniemi@hut.fi
1.1
The rst steps when constructing model for a system, complex or not, involve
outlining the system, determining its boundaries and dependency structures;
after that, data is collected. Model construction involves abstraction: The
7
Session 1. Introduction
available data is contracted so that only the most relevant information remains. Despite approved modeling principles a great deal of ingenuity and
intuition is necessary when pieces are put back together.
We see various examples of complex systems around us, intuitively understanding that there is something fundamentally similar beneath the surface.
However, the complex systems we see are so multifaceted that it is dicult to
see what this underlying similarity is. Typically the low-level actors, whatever is their physical appearance, are essentially identical, but after some
kind of dynamic processes involving positive feedback loops take place, some
kind of self-organization emerges where some kind of fractal and self-similar
patterns can be detected. For example, take the following examples:
In a genetic (cybernetic) system, the emergent phenomenon is specialization of tissues and organs, resulting from interaction between genes
and enzymes.
In a cognitive system, the emergent phenomenon is intelligence or consciousness, resulting of interaction between individual neurons.
In a social system, the emergent phenomenon is the birth of organizations, nations, and business enterprises, and the rivalry among them,
resulting from actions of individual humans.
In a memetic system, the emergent phenomenon is the formation of
ideologies and religions, and also scientic paradigms, resulting from
an interplay between ideas.
What can be said about all complex systems in general? From the modeling
point of view, one has to recognize that there exist much too little data to
draw denite conclusions. All examples of complex systems are too diverse,
there is too little similarity to uniquely determine the underlying similarities.
Will the General Theory of (All) Complex Systems remain just a dream
forever?
It has been claimed that complexity theory will revolutionize old ways of
thinking. Is this just empty jargon, or is there some relevance beneath the
bluster? Indeed, it can be claimed that ways of doing science will not be the
same any more. The natural sciences doing analysis on natural phonomena
will change drastically because the nature being studied will change. What
on earth does this claim mean?
As compared to traditional modeling problems, where, too, phenomena in
the physical environment are being explained, a deeper transition is taking
place. Let us study the power of new thinking in concrete terms:
The proponents of traditional science say that the intuitive nature and shaky denition of complex systems is a disadvantage
a pessimist would say that one cannot study complexity because one cannot dene it in the rst place. But one recognizes a
complex system when seeing one. As optimists we can think that
luckily enough we do not have to dene what a complex system
exactly is! A complex system is anything that looks like a complex system. The essense is there, no matter what is the origin
of the system, whether that creature is dwelling in biosphere or
in infosphere. This makes it possible to articially generate fresh
data examples of complex behavior for analysis purposes.
Complexity as a research topic, or searching for the general rules underlying
all complex systems, is perhaps a better manageable problem than studying
only one complex system at a time.
For example, in articial life research one is no more bound to the limitations of real, existing carbon-bound life forms: The science of the principles
governing the universal properties of life may transcend from life as we know
it to life as it could be. It is easier to see the big ideas and basic principles
when there exists a larger body of material to study this kind of wealth
can be reached when articial life forms are constructed. The computer is the
tool for creating strange worlds where creatures dened in terms of simple
mathematical formulae interact. Whether or not complex behaviors emerge
the boundaries of life can thus be mapped. The philosophers will also
have new material to ponder about: How to focus the age-old denitions of
what life is?
To have a glimpse of what kind of manifestations complex systems can have,
it is necessary to look at dierent kinds of examples. Thus, showing examples
of complex systems from dierent points of view is the goal in this report.
Whatever is the nal destiny of complexity research, one contribution that
will remain is this shift from analytic way of doing science towards synthetic
science. It is simulation that plays a central role in future science. However,
simulation must not be an end in itself.
1.2
One reason for the current interest in complex systems research is due to
Stephen Wolframs new book A New Kind of Science. However, complex
10
Session 1. Introduction
systems research has a much longer, colorful history, its mathematics dating
back to Poincar Orbits, continuing to Lorenz Attractors, and Mandelbrot
Sets, etc. No denite conclusions have ever been drawn, and the same challenges are invented once again after the previous generation of researchers has
left the eld. Research seems to be cyclic at least the branches of research
that are driven by fashionable buzzwords. Indeed, the society of complexity
researchers itself constitutes a complex system: The eld is a mosaic where
dierent ideas exist every now and then particularly good ideas emerge
from chaos, staying alive and ourishing, being seeds of further great ideas.
The goal here is to try and understand the eld of complex systems research.
What is the dynamics there? Can we estimate the future state, and perhaps
even personally contribute in reaching this new understanding? How to make
the research cycles into spirals, so that when the same old ideas pop up later,
they can be seen on some higher level of understanding?
When trying to understand complex systems research, history cannot be
forgotten. Typically in sciences, it is the results that are of importance,
and one can forget about the background with its detours and dead ends.
When seen in retrospect, the Kuhnian paradigm shifts only remain visible.
Now, on the other hand, one is facing a science during its making, and
everything is in turmoil the new stasis has not yet been reached. It is not
yet clear what is the hard core of the new paradigm. Modeling of dynamic
transients is much more challenging than static modeling of the steady states.
Without enough understanding of the Hegelian Spirit of Complexity, its
natural dynamics and inertias, there seems not to be much sense in the
bumping of this research eld. This Spirit determines what is interesting at
some time; the Spirit is an emergent phenomenon, not bound to individual
researchers but to the dynamics of the whole research community. If the
Spirit is not mature enough, an idea, however good, is not understood and
widely discussed, and it suers a memetic death, not becoming a landmark
along the evolution of the research branch. To understand the whims of this
free Spirit, one rst has to understand what is interesting. One has to know
what kind of understanding there already exists, how the current state has
been reached, and what are the painstaking problems at the moment.
It turns out that the results from complexity theory can be applied for analysis of this interestingness issue. As shown later, the interesting areas typically
reside between order and chaos. How to remain there in research work, not
to sway over the borderline to hopeless disorder and chaos this is a key
question. Indeed, it is easy to do ironic science in the chaotic domain,
introducing new and fancy ideas with no solid grounding; to remain on the
11
boundary between order and chaos in complexity research, one explicitly has
to connect new ideas to relevant mathematics. Mathematics is an emblem
of order; it is just a way to express things in a logical, consistent way, and
non-mathematics is not much more than hand-waving.
Stephen Wolfram speaks of New Science where simulation is the exclusive way
to do science, and mathematics has to be abandoned altogether. However,
after seeing the ood of endless images and patterns, one has to face the fact
that intuition has to be fetched from more concrete analyses. The hypothesis
here is that the traditional, mathematically oriented analytic science still
remains, and this way the ongoing further complexity research can continue
along the edge of chaos.
Sticking to the central role of mathematics may sound like a harsh limitation
but, truly, this restriction only opens up new horizons. Actually, one
could be more ambitious here than what Stephen Wolfram is: Even if there
will not exist New Science, it is the whole world of observations that will be
changed. Within the computer, there exist an innite number of New Worlds
to be conquered!
To summarize, simulation is a new way to have new data, but it is no substitute for theoretical analysis. However, traditional mathematical tools are
not well suited for the new challenges, and new ones need to be developed
for analysis of complex systems. Seeing the multitude of complex systems
and approaches to attacking them may help to see the problems one is facing
when trying to develop such mathematics.
1.3
The role of this report is to try to give an overall view over the eld of complex
systems research. The view is necessarily incomplete and biased, mainly
reecting the editors personal view. But if the view were not incomplete,
then (following the Gdelian intuition!) it would be certain that it would
be inconsistent. Dierent perspectives to the same thing may open up new
ways to understanding, and in what follows, the ideas of the selected four
projections are briey explained. All these four projections span a continuum
where nothing is black-and-white neither is it grey, but extremely colorful!
12
Session 1. Introduction
1.3.1
Philosophical view
1.3.2
Top-down view
There are dierent kinds of ideas of what kinds of structural constructs the
complex systems are qualitatively composed of. Even though the dierences
between paradigms are clear, it is interesting to see the mental inertia:
One often wants to see the world in terms of his/her own mental constructs,
being reluctant against fresh approaches; this holds true what comes to the
whole research community at certain times dierent ways of seeing the
world dominate. Looking complex systems research in a perspective, one can
recognize the shift from centrally organized to completely distributed control.
1. From strict hierarchies ...
Since the Aristotelian taxonomies, the traditional way of structuring
13
1.3.3
Bottom-up view
14
Session 1. Introduction
1.3.4
15
Systems view
It should not be a surprise that complex systems are complex one just
has to somehow tackle with such systems in practice. In the nal chapters,
examples of practical approaches, old and new, are presented. The objective of systems engineering is to manage and understand systems, however
complicated they might be.
1. From hard mathematics ...
One of the last theoretical success stories of modern system theory
was the framework that was intended for the hierarchical control of
large-scale systems. In the 1960s and 1970s they still thought that
sophisticated system models could be constructed, or understood. The
practitioners voted against such illusions, and selected postmodern
approaches (fuzzy and neural) instead (Chapter 12).
2. ... towards applying intuitions ...
When studying truly complex systems, no explicit models exist. The
proponents of system dynamics assume that qualitative approaches
suce, forgetting about exact details and numbers, and still utilizing
system theoretical tools. Simulation of more or less heuristically determined qualitative causality models is by no means mathematically
well justied, but without such assumptions not very much useful is
left that could be applied in practice (Chapter 13).
3. ... to reach a systemic view?
In general system theory the abstractions are brought to extremum
having no more concrete numerical models available one just has
to manipulate symbols, the methods being too abstract to be of any
real use, and one easily ends up doing mere philosophy. However, the
system theoretic approaches may still oer holistic ideas that are not
bound to todays practices. And, perhaps, from the system theoretic
considerations one can complete the spiral back to hard mathematics.
Can there exist a systemic view of complexity? How to see complex processes in a perspective? How the emergent patterns could be controlled,
and how to make something interesting emerge out from the simulations? What kind of holistic mathematics might be available when
analyzing complex systems? Specically, how to restrict the power
of tools so that they remain on the boundary between analyzability
and non-analyzability? Indeed, such questions can be studied. For
example, it seems that the idea of one single complexity theory has
16
Session 1. Introduction
to be abandoned perhaps the theory itself has to be fractal: To
make something interesting emerge from the mindless simulations, the
domain-specic semantics has to be coupled in the system structure.
The structures also dier in dierent application domains. This kind of
observations only make the questions more interesting and challenging:
The area of complexity research cannot be exhausted by the pioneers
alone. For example, most probably there will exist separate research
areas like complex cognitive system theory and complex automation
system theory, where the ways how emergent phenomena pop up dier
from each other. Correspondingly, the appropriate mathematical tools
are dierent, and the new intuitions and tools are dierent. The last
chapter hopefully gives some insight on what this all really means.
Session 2
Complex Systems A New
World View
Reino Virrankoski
Helsinki University of Technology,
Control Engineering Laboratory
reino.virrankoski@hut.fi
Because of the rapid increase of computational power, many numerical methods has been widely used in the complex system research. In addition to that
the development of information processing and networked systems like the
Internet has also eect to the eld of complex systems. This chapter gives
introduction to the ongoing change and possible promises.
2.1
Introduction
18
2.2
2.2.1
2.2.2
Basic ideas
2.2.3
Cellular Automaton
19
In the 1950s John von Neumann formulated cellular automaton. There are
cells located somewhere in space, and in each step of time the state of a single
cell is determined based on the states of neighbouring cells one step earlier.
One of the most well-known cellular automata was so-called Game of Life, a
two-dimensional cellular automaton presented by John Conway in the 1960s.
Stephen Wolfram start to work with cellular automata in the 1980s, and
found soon out that even cellular automata generated by very simple rules
have interesting properties. By using computer programs including Mathematica he and his company developed, Wolfram made numerous dierent
types of simulations with dierent types of cellular automata. He found out
interesting variations of chaotic and regular patterns in cellular automata,
nested structures, evolutionary structures and also interactions between different types of structures. Wolfram used cellular automata to investigate
the system sensitivity to the dierent initial values, and he used cellular
automata-like substitution systems to imitate some natural processes, like
uid vortexes, formation of snowakes and growth of plants and animals. He
built dierent classications to cellular automata types based on his ndings,
and found also analogies between cellular automata and traditional mathematics [1].
2.2.4
Wolframs conclusions
Based on his investigations, Wolfram claims that most of the complex natural
processes are based on the same type of simple rules that he implemented in
cellular automata. He goes even further, and postulates that even the laws
of fundamental physics could be based on similar types of simple rules, and
underneath all the complex phenomena we see in physics there could be some
simple program, which, if run long enough, would reproduce our universe
in every detail. As a comment for these postulates Wolfram writes: The
discovery of such subprogram would certainly be an exiting event as well
as dramatic endorsement for the new kind of science that I have developed on
this book. Furthermore, he suggests that by combining this with a certain
amount of mathematical and logical deduction, it will be possible at least as
far as reproducing the known laws of physics [1].
After huge amount of examples and comments Wolfram returns to the main
idea of his New Kind of Science, what he calls the Principle of Computational Equivalence: All processes, weather they are produced by human
eort or occur spontaneously in nature, can be viewed as computations [1].
This raises the question whether there is need for traditional mathematical
20
2.2.5
Criticism
2.2.6
Provocation?
21
When talking about big basic things in science, the scientic community
is today so smug that many radical new ideas are immediately shot down
without any further discussion, if presented conventionally. Hence, more
provocative style is needed for generating active debate.
2.3
2.3.1
Networked systems
New findings and inventions
The basic principles of the neural system in biological organism have been
known relatively long time, and during the latest decades a lot of new knowledge has been gathered in the area of neurosciences. In addition to that,
the development of computers and the rapid growth of computing recourses
have enabled the creation of dierent networked systems. Probably two most
well-known network categories are the Internet and articial neural network
algorithms developed for modelling and simulation.
The success of those two network types and the rapid increase in the number
of users has also generated new viewpoints to the complex systems, and,
generally, to the universe around us.
2.3.2
The traditional way to see many complex systems has been as a hierarchical
structure. Even in modern quantum physics there exist dierent electron
energy levels in the basic structure, and furthermore each of those stages is
split to hierarchical structure in lower level. In macroscale, space consists of
solar systems, galaxies and so on. There are of course a lot of hypotheses
and dierent types of theories concerning those models, but the traditional
way how the things have been seen is some kind of hierarchical structure.
Because of the traditional hierarchical way of thinking and relatively simple
structure of hierarchical models, those types of models have been widely used
in control engineering when trying to model and control complex systems.
In many cases this approach has been useful or at least good enough when
controlling industrial processes, for example.
22
2.3.3
Networked structures
In his book Emergence [3], Steven Johnson presents the idea that everything is linked rather than hierarchically structured. Furthermore, Johnson
weights the importance of local low-level interactions in the network. He
found analogies for example in the behaviour of ant colonies, economical systems and in the growth of cities. The structure of all of those systems is
network-type, and there seems to be a lot of local interactions that are not
centralized controlled. It could be possible that some general trends in the
behaviour are initially or centrally set, like rules in the markets and city and
instincts in the ant behaviour. Anyway, local interactions and decisions that
cannot be centrally controlled have still crucial eect to the development of
the whole system.
Thus, Johnson sees many systems as being constructed from bottom to up.
He also presents some discussion about the development of local structures
and local decision making clusters [3]. In his book Linked [4], AlbertLazlo Barabasi presents similar types of thoughts about network-structured
systems, and he formulates Agents as a functional objects in the network.
2.3.4
2.4
23
Bibliography
1. Wolfram, S., A New Kind of Science, Wolfram Media, Inc., 2002.
2. http://www.wolframresearch.com
3. Johnson, S., Emergence: The Connected Lives of Ants, Brains, Cities,
and Software, Touchstone, 2002.
4. Barabasi, A-L., Linked: The New Science of Networks, Perseus Publishing, 2002.
24
Session 3
What Kind of Science is This?
Heikki Pisto
Helsinki University of Technology
Control Engineering Laboratory
heikki.pisto@hut.fi
3.1
26
purely because of itself, not only for economical or other benets. This idea,
so called basic research, has proven to be as crucial in scientic developments
than the other half, applying. For example, celestial mechanics may seem to
be useless endeavour at rst, but it has always been one of the basic objects
of curiosity of mankind. After one learns that earth is round rather than at,
vast possibilities immediately open.
The Greeks had much interest in philosophy, and they considered study of
nature one branch of it, natural philosophy. They emphasised rationality,
and tried to form laws about nature works. On the other hand, they did not
consider testing of their theories with experiments too important, and some
of their ideas are somewhat confusing, even silly, to modern person. The
important idea, which emerged little by little during otherwise long and dark
Middle Ages, was empiristic way of acquiring information about nature. In
13th century Grosseteste worked with optics, and concerned to verify theories by experiments. His student Bacon used mathematics to describe optical
systems, and conducted systematic experiments with them. Later, the methods of Galilei, da Vinci and their successors of looking at the world through
experiments became very powerful and popular. From those days on natural
sciences have been increasingly related to mathematics. Observations needed
interpretations, theories, to be useful, and mathematics was practically the
only way to express them precisely. This symbiosis of mathematics and science is still in good health, but lately there have been some suspicions about
this old paradigm of doing science.
3.2
Every now and then scientists have felt that there are no more things to
discover. For a long time Newtons laws were good enough for everybody,
mainly because they explained almost all of the common phenomena that
occurred in everyday life. Electricity changed everything. This mysterious
force raised new questions but also made possible to arrange new experiments, which revealed new phenomena, which in turn needed new theories.
In early 20th century there were suddenly again plenty to do and explain, and
it took some 50 years for particle physicists to clean up the mess. Towards
the end of century there were again separate theories that were capable to
explain basically all observations. The only problem is that these theories
are not consistent with each other. Or is it a problem?
Today most of the questions concerning space, time and matter can be considered solved, at least for current practical purposes. The picture of physics
27
looks good, but it is The Reality? Hardly. And even if it was, the wholeness
is way too broad to be mastered by any individual. Still there are people,
who think it is worth trying. Some of the scientists have taken the quantum leap beyond experimentability, and continued making science in further
dimensions. Although this means quite a change in the traditional virtue
of science, it may not be wise to blame them for doing that. The ghost of
Democritus is still lurking among physicists.
The other way to do progress seems to be to construct something from these
already acquired building blocks of the Standard Model. Approaches are
numerous. Economically the most lucrative branch of science is probably
material sciences, where increase of knowledge has over the last 50 years
caused another industrial revolution. The so called Moores law, which says
that device complexity on a chip doubles every 18 months, has held true
for almost 40 years now, and probably will for at least 10 years. Stream of
new applications will continue to distant future, and yet new markets will
emerge. Unfortunately, as smaller details and scales are attained, the costs
of investments in industry have risen as fast as the markets. The so-called
Moores second law states, that also cost of facilities increase on a semi-log
scale. Modern semiconductor factory may easily cost billions and billions of
euros in the future, which eectively keeps prots from rising exponentially.
Biotechnology has also given big promises, but has not fully met them yet,
despite huge amounts of investments. It is the same phenomenon that concerns basic research: For ever increasing use of resources there is inevitably
diminishing amount of returns. It seems, that the progress is not slowing
down because of lack of adept scientists. The potential scientist material is
larger than ever. It is only that gigantic telescopes, particle colliders and fusion reactors seem to be the only way to go ever further to make any progress.
Could it be, that scientists are running out of good ideas?
3.3
Science journalist John Horgan has taken to himself the burden of revealing
the emperors new clothes. In his 1996 published book The End of Science
he argues, that although there are vast amount of scientists, more than ever
before, bustling around science, all they have left to do is to ll some gaps
and gure out applications. He bases his argument on dozens of interviews
he has made with some prominent scientists of late 20th century. They
include philosophers, biologists, physicists, social scientists, neurologists, and
people that he amusingly calls chaoplexologists. By chaoplexity he refers
28
to complex systems and chaos research, which has over last decades changed
names according to concurrent fashions. Through the interviews he creates
quite ambiguous construction about modern science. Some of the scientists
are looking for The Answer; somebody seems to already have found it. Some
believe in the Big Bang, some swear by evolution. And all this is avoured
with Horgans sceptical and rened comments.
Horgan certainly hit a sore point. To make science is a profession, and pure
science, basic research, is mostly publicly funded. In private sector people
working with technology and knowledge like to call it product development,
as contrary to science in public sector, even if they would be doing the same
thing. Engineers in the private companies want to please the capitalists,
scientists for one the taxpayers, the great public that is following their work.
To this public it may sound quite confusing, if a notable science journalist
declares, it is all over, everybody go to your homes, there is nothing more
to see. It may be a simple mans shot to make a handprint in the world, but
what if ...?
The most common objection to Horgans argument has been Thats what
they thought hundred years ago. Even though there are some famous quotes
to support this, it is probably not true. Horgans answer is simply No they
didnt. For most of the sciences and scientists, this is the fact. In the
end of 19th century science was by no means in a dead end, at least for a
scientically oriented mind. It is another thing, that for some spectators and
appliers of technology, like Horgan hundred years later, it might have seemed
that way. By Horgan this 100-years-argument means, that people want to
say, because science has advanced so rapidly over the past century or so, it
can and will continue to do so, possibly forever. This rhetoric turns out to
be a philosophical hypothesis, which sounds unconvincing, thus supporting
Horgans point. Yet that is not too convincing either in its strictest form, as
in the title of his book. If the truth is somewhere in between, what is the
catch?
Some, if not most, of Horgans critic is pointed at something he calls ironic
science. His most important observation about modern science is, that either
in lack of potential or because of pure curiosity about abstract, it is heading
more and more to speculative mode, separate from veriable physical experiments. Physical world is becoming engineering and is left for engineers,
because for potential future Einsteins, the prospects are bad. For example
there is superstring theory, which Horgan calls nave ironic science. Superstrings are mathematical constructions, that may make sense at certain
level of comprehension, but is it really enough to form a world-view? String
29
theory can be described as speculative; it is not possible to do any experiments in its bonus dimensions, with energies unimaginable. Also some of its
practitioners have almost religious relationship about their subject. When
they see the truth, it is the truth, and should be treated as such, despite of
the fact, that when none of its results are veriable. They can in no way
aect our lives.
It is known, that scientic revolutions result because of change of paradigm,
way of seeing and doing science. This speculative mode may well be a new
paradigm, at least if there really is nothing else to do. In that case, also
the experimental method of science, which has hundreds of years been so
successful, needs some new explaining. So far purpose of the theories has been
to make explanations, and if possible, predictions about nature. Experiments
have been necessary to evaluate usefulness of theories. If importance of
experiments is diminishing, or it the experiments are changing too expensive
to conduct, there is clearly a transformation going on between paradigms.
To Horgan it means, that science is going into decline. Obviously situation
in not such, that there is nothing to discover any more. But even though
Answers raise new questions, many of the questions are such by nature,
that they cannot be critically answered with objective truth. Lets consider
cosmology. Every layman can formulate questions, which are as impossible
to answer as those of religion. Science has its limits because of its (present)
method; there is nothing that can be done about it. And when we pass
the method of science, it is the end of science? In Horgans opinion ironic
science has its function, though. By handling unanswerable or unimportant
questions, it reminds human beings, how little they know, and maybe sets
their existence in perspective, if is doesnt happen to be there.
3.4
One common paradigm in doing science has been, that the truth in itself is
somehow simple and beautiful. One of the more famous formulations of this
is the so called the Occams Razor, after William of Occam (1300-1349),
which basically recommends the simplest of all rational explanations to an
observed phenomenon. This approach can be deceiving, for example in Horgans book it is gloomily quoted without reference, that in biology Occams
Razor cuts your throat. Indeed, lots of interest has been spent since end of
19th century on dynamically complex behaviour of systems of several simple
entities as opposed to studying properties of individual components. Many
times the n-body problem has been considered as starting point for com-
30
plex systems research. After that there has been numerous important and
fascinating examples of other such systems, for example strange attractors,
fractals and cellular automata. All of them have been in turn very fashionable subjects, only to be replaced in a few years with something else. All
these have the common treat that the beauty is not in the simplicity, but
in their complexity. More specically, what interests human mind is something between simple and complex, something that comes out of chaos and
appears to be somehow more than the original ingredients. Philosophically
speaking, for example in the case of cellular automata, all behaviour is essentially equal in system point of view. For external observer, in this case
human being, there is in some cases more to see. With certain rules, some
patterns expand innitely, some survive indenitely, and some repeat cycles,
some die out. This is fascinating. So are fractals, too. But, what kind of
practical results can be derived out of them? Stephen Wolfram says, that
every kinds of results. In his book New Kind of Science he asserts, that basically the whole reality is working like cellular automata, and that knowledge
about it can be acquired through simulations of other cellular automata.
His hypothesis of computational equivalence suggests, that everything can
be reduced to series of computations, which in turn can be carried out by
very simple logic. This would mean, that everything could be simulated,
and not just approximately, but precisely. This is another major deviation
from current paradigms of science. Like superstring theory, it is a quantum
leap into obscurity, and it seems quite dicult to work out the way back
to normal, concrete world. And, before the connection is again formed with
experimental reality, it is nothing more than metaphysics of ancient Greeks.
So what are we researching, when we study these so-called complex systems?
Many scientists have been very optimistic about what can be forged out of
them. Maladies, natural catastrophies, all kinds of messes can be cleaned
up in the future which will make human life better. Yet all the progress
so far has been discovering new intriguing dynamics in dierent kinds of
systems, but little more. Characteristic for the research has been, that for
every research subject there is real background somewhere in reality; complex
systems researchers seem to be dierent from mathematicians, who do not
care about nature as such. Only logic and consistency is relevant, abstract
is a virtue. In set theory, it is not very professional to use apples and horses
as members of sets. Then in complex side, people are more than eager to
develop great stories about what is going on in their complex world, and
relate new results straightforwardly with some other things in reality, even
if there is no real connection. Then on the bottom line, horses are happy
to eat apples, and it is much more probable for the research to get publicity
31
and funding. In the past it was work of the science journalists to make
quantum mechanics sound interesting, now the researchers themselves do it.
Is science becoming popular culture, and researchers becoming superstars?
Of course Einstein and Hawking are superstars, although few have any idea,
what is the content of their work. If there is any Philosophers Stone to be
found in chaoplexity eld, it will give to its nder the most juicy stories of
science history, and more fame that anyone before that. After all, some of
the contemporary complex researchers have already had a taste of this, but
everybody is not convinced yet.
3.5
Science has traditionally been prone to personality cults, even though great
majority of all-important discoveries seem later quite inevitable. Plancks
constant could well be somebody-elses-constant, Paulis exclusion rule be
Schulzennagels rule. All this encourages generations after another to continue making science. It might be dicult for a young scientist to content
oneself with only developing applications and giving lectures about existing
knowledge. So many ambitious scientists couldnt have resisted the idea of
inventing science rather than discovering it. If a theory great enough could
be constructed so, that it would not contradict with any of the existing theories but instead would expand and incorporate them, then one could truly
call that achievement regardless of whether it would be useful, or even experimentally provable at the moment. This approach has worked on several
occasions before, most notably for special relativity. Basically this is also
what Horgan calls ironic science. The rst deliberate ironic comment was
physicist Alan Sokals nonsense article in a journal of cultural and political
analysis Social Text in 1996, just to test, if there really existed any more
intellectual criticism among post-modern philosophers. The text passed the
editors, and was not revealed as a hoax, until Sokal himself did it. At least
in some peoples opinion more serious occasion was in 2002, when the infamous Bogdanov brothers got caught for at least four gibberish papers, for
which they also got their Ph.D:s in University of Bourgogne. The brothers
had their own TV-show in France, and they were local celebrities. The alarm
bells rang, and almost all of the people, who considered themselves scientists,
wondered, what was happening to this their beloved mission for The Answer
called science. Who can be so vain to do things like that? From outside
point of view, though, there is no dierence, if paper is written with hoax in
mind or by studying something of little importance in itself.
32
All this might cause, that talented young people hesitate to choose scientic
careers, and maybe prefer going for business or government, even after acquiring a scientic degree from university. Problem is not, Horgan argues,
that there are no more questions to be answered. After all one can always
wonder the meaning of life. If anything, there is shortage of good questions,
which have always been essential. It is hard to imagine theories that would
have even remotely same impact as evolution theory, general relativity or
quantum mechanics. There are still many unanswered questions in biology
and cosmology, for example. The nature of these questions is nonetheless
such that every hypothesis is necessarily very dicult to verify, if not impossible. The beginning of life may remain unanswered forever, unless some life
form resembling existing life really can be created articially and repeatably
from the elements. Also, the idea about the rst moments of universe is so
abstract, that one has take it with a grain of salt. It might be mathematically
consistent in the scientists head, but if it is true, that is a whole dierent
thing. For an individual scientist, it might bring fame and welfare, even if
the theory were false, as history can show. But from practical point of view,
it is quite safe to remain dubious about such things. Scepticism has always
been virtue in thinking, when it comes to matters, which have little tangible
eects on life itself. Whatever that is.
3.6
Where to go?
Its like the jazz musician, who was asked, where jazz is going,
and he said, If I knew, we would be there right now
Chaos research and complexity research is basically same thing under different names. Chaos refers to disorder, complexity to something elaborate
and sophisticated, but still somehow ordered to human mind. One could
say that weather is chaotic, you cannot predict that, let us say, for a year
forward, although you can with big condence assume that in Finland it is
chilly. Some cellular automaton may be complex, but it is not really chaotic,
because it is so discrete and computable. Both terms are quite ambiguous
and mean dierent things to dierent persons.
Ways of attacking complexity are numerous. Purely theoretical, mathematical ones, like fractals and cellular automata, have produced little more than
entertaining patterns. Some more practical approaches, like emulating nature
by networks and agents, have produced some real applications, like Internet,
where wholeness is so much more complicated than what the constructors
33
were planning in the beginning. It basically has a life of its own, and cannot
be controlled by any single quarter. From the net point of view the users
are not using the net as a traditional tool, but rather living in part the life
of the net. If there were no voluntary diverse action in the net, it would be
little more than an advanced telegraph. Internet is not designed to full different sexual desires, but because of the human nature, the services emerged
there. This for one encourages also other use of the net, also improving its
usefulness. How could have this been possible to predict?
This something-from-nothing-aspect has characterised much of the research
in the eld. It is always good thing, when something interesting emerges.
But, to make something really desired and useful eects, like future stock
prizes or a wristwatch, to emerge from something trivially available resource,
like computing capability or a ton or ore, appears to be daydreaming. But
also it is impossible to tell, what is possible and what is not. Life itself
should not be possible, because from human point of view, it is too complex
to work. The complexity of complexity has encouraged certain researchers to
make predictions about glorious future, starting from curing diseases, ending
with world peace. Human being would move on from being part of the
system to be the master of the system, knowing all relevant information,
taking into account all side eects of its action on the system. Also many
practical problems of an average person are complex in their nature, like
guessing lottery numbers. Others are usually computationally equivalent of
computing, if one has enough money for additional bottle of beer, or if one
can aord to sleep ten minutes more in the morning. Such considerations
will remain always the same, but in the best case in distant future, complex
systems research could make the lottery meaningless, or maybe at least the
stock exchange.
Up to the present the research of complexity and chaos has been more or
less alchemy with computers. The hype has caused some people, eager to
leave a handprint in the history, desperately seek for the Philosophers Stone,
which would convert complexity into simplicity. It should be evident, that
this is not going to happen. But, after all, research of chemistry was based
on alchemy. Alchemy was Isaac Newtons hobby, but not all alchemists were
Isaac Newtons. And indeed something has emerged, if nothing else, then
plenty of material for scientic Monty Python sketches.
Bibliography
1. Horgan, J.: The End of Science Broadway Books, 1997.
34
Session 4
Architecture of Complex Systems
Eugene M. Burmakin
Helsinki University of Technology
Industrial IT Laboratory
eugene@cs.hut.fi
4.1
Conceptions of Complexity
Last century has seen strong interest in complexity and complex systems.
There were three main waives of interest in complexity. The post-WWI
interest in the topic was focusing on the claim that the whole is more than the
sum of their parts, and was strongly anti-reductionistic in avor. The postWWII interest was rather focusing on the idea of feedback control and selfstabilization of complex systems. The current interest in complexity focuses
mainly on mechanisms that create and sustain complexity, and analytical
tools for describing and analyzing it.
4.1.1
Holism is a modern name for very old idea. In the words of its author, the
South African statesman and philosopher, J.C. Smuts:
Holism regards natural objects as wholes . . . It looks upon
nature as consisting of discrete, concrete bodies and things . . .
35
36
4.1.2
4.1.3
The current wave of interest in complexity has lots of common ideas with the
second one. But other new ideas are considered, such as catastrophe theory,
chaos theory, genetic algorithms, and cellular automata. The motivation for
the current interest is the need to tackle with complexity in global and largescale systems such as environment, society, organisms etc. Also, tools that
were developed for system complexity management in the second wave are
not appropriate for current systems under investigation due to their increased
complexity.
4.2
37
It is good to dene the term of complex systems, before going to the architecture of complex system. There are many denitions of a system. But
roughly, a complex system is made up of a large number of parts that have
many interactions. In such systems the whole is more than the sum of the
parts. Given the properties of parts and the laws of their interactions, it is
not a trivial thing to infer the properties of the whole.
4.2.1
Hierarchic systems
4.2.2
Let us quote here the watchmakers metaphor that was presented in [4.3]:
There once were two watchmakers, named Hora and Tempus, who
manufactured very ne watches. They were making their watches
in own premises, a phone was calling frequently new customers
were calling. Hora prospered and Tempus became poorer. What
was the reason?
The watches consisted of 1000 parts each. Tempus was assembling them part by part, and if his phone was ringing (the work
process was interrupted) he had to put the not nished watch
down, and it immediately fell to pieces and had to be reassembled from the elements.
The watches that Hora made were not less complex. But he
had designed them so that he could put together subassemblies
38
If we take into account some ideas from biological evolution and other elds,
a number of objections will arise against this metaphor.
The complex forms can arise from the simple ones by purely random
process. Direction is provided to the scheme by the stability of the
complex forms, once these come into existence.
Not all systems appear hierarchical. For example, most polymers are
linear chains of large number of identical components.
Multi-cellular organisms have evolved through multiplication and specialization of the cells of a single system, rather than through the merging of previously independent subsystems.
Hence, there are reasons to dismiss the metaphor. However, the systems that
evolve by means of specialization have the same kind of boxes-within-boxes
structures as well as systems that evolve by assembly of simple systems.
There are application were the metaphor works. Consider, for example, the
theorems and their proofs. The process starts with axioms and previously
proven theorems. Various transformations allow obtaining new expressions;
this process goes on until the theorem are proven. Another good example is
the development process of an empire:
Philip assembled his Macedonian empire and gave it to his son, to
be later combined with the Persian subassembly and others into
Alexanders greater system. On Alexanders death his empire
did not crumble into dust but fragmented into some of the major
subsystems that had composed it.
4.2.3
39
may be of dierent orders of magnitude. For example, in a formal organizations there will be more interactions between employees from the same
department than from dierent departments. The system can be considered
as decomposable if there are no interactions among the subsystems at all. In
practice, these interactions can be weak but not negligible. Hence, one may
move to the theory of nearly decomposable systems.
There are two main theoretical propositions found concerning nearly decomposable systems:
In a nearly decomposable system the short-run behavior of each of the
component subsystems is approximately independent of the short-run
behavior of the other components.
In the long run the long-run the behavior of any one of the components depends in only an aggregate way on the behavior of the other
components.
These ideas can be described using an example. Consider a building whose
outside walls (boundary of our system) provide perfect thermal isolation
from the environment. The building is divided into a large number of rooms;
the walls between them (subsystems boundaries) are not perfect isolators.
Each room is divided in partitions the cubicles being poorly isolated. A
thermometer hangs in each cubicle. There is wide variation in temperature
during the rst observation. But a few hours later the temperature variation
has become very small among rooms inside the building.
One can describe the process of reaching the equilibrium formally by setting
up the usual equation of heat ow. The equations can be represented by
the matrix of their coecients: r ij is the rate at which heat ows from the
ith cubicle to jth one. If cubicle i and j do not have a common wall then
r ij will be zero. If cubicles i and j are in the same room then r ij will be
large. If cubicles i and j are separated by the rooms wall then r ij will
be small but not zero. Hence, one can group together all these coecients,
getting a matrix where all of its large element will be located inside a string
of square submatrices along the main diagonal. We shall call a matrix with
this properties a nearly decomposable matrix (see Fig. 4.1).
Now it has been shown that a dynamic system that can be described in terms
of a nearly decomposable matrix has the properties of nearly decomposable
systems, stated earlier in this section. Hence, we have seen that hierarchies
have the property of near decomposability. Intra-component linkages are
generally stronger then inter-component linkages. This fact has the eect of
40
4.2.4
4.3. Conclusions
41
view, if there are important systems in the world that are complex without
being hierarchic, they may to a considerable extent escape our observation
and understanding. Analysis of their behavior would involve such detailed
knowledge and calculation of the interactions of their elementary parts that
it would be beyond our capacities of memory or computation.
There are two main types of description that seem to be available to us when
seeking for understanding of complex systems:
State description
A circle is the locus of all points equidistant from a given point.
Pictures, chemical structural formulas, blueprints, etc.
Process description
To construct a circle, rotate a compass with one arm xed the
other arm has returned to its starting point.
Recipes, dierential equations, etc.
4.3
Conclusions
Empirically a large proportion of the complex systems one can see have
hierarchical structure. Or one perceives them as hierarchies in order to solve
some complex problems. A complex system can be presented as a hierarchy in
order to make proper control of the systems. This approach of presentation is
very desirable in many cases due to the fact that it has many useful properties
such as near decomposability that is simplifying description and analysis of
the system.
Reference
Herbert A. Simon: The Sciences of the Articial. MIT press, Cambridge, Massachusetts, 1996 (third edition).
42
Session 5
Towards Decentralization
Matti Saastamoinen
matti.saastamoinen@honeywell.com
5.1
Introduction
44
5.2
45
"Agents" are autonomous, computational entities that can viewed as perceiving their environment through sensors and acting upon their environment
through eectors. Agents pursue goals and carry out tasks in order to meet
their design objectives, and in general these goals can be supplementary as
well as conict [3].
"Intelligent" indicates that the agents pursue their goals and execute their
tasks such that they optimize some given performance measures. To say that
agents are intelligent does not mean that they are omniscient or omnipotent,
nor does it mean that they never fail. Rather, it means that they operate
exibly and rationally in a variety of environmental circumstances, given the
information they have and their perceptual and eectual capabilities.
"Interacting" indicates that agents may be aected by other agents or perhaps by humans in pursuing their goals and executing their tasks. Interaction can take place indirectly through the environment in which they are
embedded or directly through a sheared language. To coordinate their goals
and tasks, agents have to explicitly take dependencies among their activities
into consideration. Two basic, contrasting patterns of coordination are cooperation and competition. In the case of cooperation, several agents work
together and draw on the board collection of their knowledge and capabilities
to achieve a common goal. Against that, in the case of competition, several
agents work against each other because their goals are conicting. Cooperating agents try to accomplish as a team what the individuals cannot, and
so fail or succeed together. Competitive agents try to maximize their own
benet at the expense of others, and so the success of one implies the failure of others. The following major characteristics of multiagent systems are
identied:
each agent has just incomplete information and is restricted in its capabilities;
system control is distributed;
data is decentralized; and
computation is asynchronous.
Broadly construed, both DAI and the traditional AI deal with computational
aspects of intelligence, but they do so from dierent points of view and under
46
dierent assumptions. Where traditional AI concentrates on agents as "intelligent stand-alone systems" and on intelligence as a property of systems that
act in isolation, DAI concentrates on agents as "intelligent connected systems" and on intelligence as a property of systems that interact. In this way,
DAI is not so much a specialization of traditional AI, but a generalization of
it.
5.3
The two major reasons that cause to study multiagent systems are:
Technological and Application Needs Multiagent systems oer a
promising and innovative way to understand, manage, and use distributed, large-scale, dynamic, open, and heterogeneous computing and
information systems. The Internet is the most prominent example of
such systems; other examples are multi-database systems and in-house
information systems. These systems are too complex to be completely
characterized and precisely described. DAI does not only aim at providing know-how for building sophisticated interactive systems from
scratch, but also for interconnecting existing legacy systems such hat
they coherently act as whole. Moreover, like no other discipline, DAI
aims at providing solutions to inherently distributed and inherently
complex applications [4].
Natural View of Intelligent Systems Multiagent systems oer a natural way to view and characterize intelligent systems. Intelligence and
interaction are deeply and inevitably coupled, and multiagent systems
reect this insight. Natural intelligent systems, like humans, do not
function in isolation. Instead, they are at the very least a part of the
environment in which they and other intelligent systems operate. Humans interact in various ways and at various levels, and most of what
humans have achieved is a result of interaction.
In addition, multiagent systems, as distributed systems, have the capacity to
oer several desirable properties:
Speed-up and Eciency Agents can operate asynchronously and in
parallel, and this can result in an increased speed.
47
5.4
5.4.1
Multiagent systems
Introduction
Agents operate and exist in some environment, which typically is both computational and physical. The environment might be open or closed, and
it might or might not contain other agents. Although there are situations
where an agent can operate usefully by itself, the increasing interconnection
and networking of computers is making such situations rare, and in the usual
state of aairs the agent interacts with other agents [5].
Communication protocols enable agents to exchange and understand messages. Interaction protocols enable agents to have conversations, which for
our purposes are structured exchanges of messages. As a concrete example
of these, a communication protocol might specify that the following types of
messages can be exchanged between two agents:
Propose a course of action
48
5.4.2
Motivations
5.5
49
Degree of decentralization
5.5.1
This is generally the easiest and most obvious organization for any sort of
agent that serves multiple users.
Advantages:
Easy to coordinate; no work has to be done by the agent itself to do
so.
Easy for users to know where to contact.
Lends itself to crossbar algorithms and similar cases in which the entire
knowledge base must be examined for each query or action.
If the server is used by very widely spread users, timezones may spread
out some of the load.
Disadvantages:
Doesnt scale: generally, the workload goes up as the square of the
number of users.
Not fault tolerant: the server is a single point of failure for both performance and security (it is a single obvious point to compromise).
The majority of users will nd themselves a good part of an Internet
diameter away from the server; this can be serious if low latency is
required of the server.
5.5.2
This describes a class of server where the basic algorithm is run in parallel on
a number of machines (typically, very-loosely-coupled parallelism, e.g., separate workstations, rather than a single MIMD or SIMD parallel architecture).
Such architectures in general can be divided into:
50
The choice of tight or loose consistency is generally a function of the operations being supported by the servers.
Advantages:
These architectures are handy when it is relatively simple to maintain
database consistency between servers (for example, if user requests or
actions taken on their behalf do not side-eect the database, then its
consistency is easier to maintain).
Load-balancing is fairly simple, and extra hosts can be added incrementally to accommodate increases in load.
The servers may be geographically distributed to improve either network load-balancing, time zone load-balancing, or fault-tolerance.
Disadvantages:
If the algorithm requires tight consistency, the requisite interserver
communications costs can eventually come to dominate the computation.
Even loosely-consistent servers will probably still suer from roughly
quadratic growth in load with the number of users. This implies that, to
keep up with even linear user growth, a quadratically-increasing number of servers must be put online; keeping up with typical exponential
growth, of course, is much harder.
5.5.3
51
5.5.4
52
Disadvantages:
Coordination between the peers becomes much more dicult.
Algorithms that require global consistency are probably impossible to
achieve with acceptable performance.
It may be dicult to keep all agents at similar software revision levels.
5.6
Applications
Many existing and potential industrial and commercial applications for DAI
and multiagent systems are described in the literature [7]. Basically following
examples of such applications are:
Electronic commerce and electronic markets, where "buyer" and "seller"
agents purchase and sell goods on behalf of their users.
Real-time monitoring and management of telecommunication networks,
where agents are responsible, e.g., for call forwarding and signal switching and transmission.
Modelling and optimization of in-house, in-town, national- or worldwide transportation systems, where agents represent, e.g., the transportation vehicles or the goods or customers to be transported.
Information handling in information environments like the Internat,
where multiple agents are responsible, e.g., for information ltering
and gathering.
Improving the ow of urban or air trac, where agents are responsible
for appropriately interpreting data arising at dierent sensor stations.
Automated meeting scheduling, where agents act on behalf of their
users to x meeting details like location, time and agenda.
Optimization of industrial manufacturing and production processes like
shopoor scheduling or supply chain management, where agents represent, e.g., dierent workcells or whole enterprises.
5.7. Conclusion
53
5.7
Conclusion
54
namically changing goals; and the problem solver cannot cooperatively solve
problems with other problem solvers [10].
In the future, systems will increasingly be designed, built, and operated in
a distributed manner. A large number of systems will be used by multiple
real-world parties. The problem of coordinating these parties and avoiding
manipulation cannot be tackled by technological or economic methods alone.
Instead, the successful solutions are likely to emerge from a deep understanding and careful hybridization of both [11].
Centralized systems are failing for two simple reasons: They cannot scale,
and they dont reect the real world of people [12].
Decentralized approaches often seem impractical, but they work in practice.
The Internet itself is a prime exampleit works because the content, the domain name system and the routers are radically distributed. But its the
human element that is really driving the pressure for decentralized solutions.
This shouldnt be too surprising. Biological phenomena like the human body
and the global biosphere have had billions of years to evolve, and they are
the most complex decentralized systems we encounter. Decentralization is
neither automatic nor absolute. The most decentralized system doesnt always win. The challenge is to nd the equilibrium pointsthe optimum group
sizes, the viable models and the appropriate social compromises.
Bibliography
1. Gerhard Weiss: Multiagent Systems, A Modern Approach to Distributed
Articial Intelligence The MIT Press, 1999, pp. 1.
2. http://webopedia.internet.com/Term/a/articialintelligence.html
3. Gerhard Weiss: Multiagent Systems, A Modern Approach to Distributed
Articial Intelligence The MIT Press, 1999, pp. 25.
4. Gerhard Weiss: Multiagent Systems, A Modern Approach to Distributed
Articial Intelligence The MIT Press, 1999, pp. 89.
5. Gerhard Weiss: Multiagent Systems, A Modern Approach to Distributed
Articial Intelligence The MIT Press, 1999, pp. 7981.
6. http://foner.www.media.mit.edu/people/foner/Essays/Agent-Coordination/degree-of-decentralization.html
5.7. Conclusion
55
56
Session 6
Networks of Agents
Jani Kaartinen
Jani.Kaartinen@hut.fi
It seems that in todays world everything is linked or networked together.
When studied closer, networks varying from internet to a coctail party or
ancient bacteria resemble each other and the same rules apply. However,
there are no solid tools for studying networks. That is why we have to learn
how to interpret these nets and to try to see if network thinking could enable
us to see something new or at least see things in a dierent light.
6.1
Introduction
In February 7th 2000 at 10:20 Pacic Standard Time, Yahoo one of the
most popular search engines at that time started to receive billions of
service requests. This would have been a celebrated event but there was a
problem, however. The requests were not produced by ordinary people trying
to nd information. They were produced by computer program running in
a large number of computers. The request was not very fruitful either, the
messages sent to hundreds of computers in Yahoos Santa Clara, California,
headquarters contained only the message Yes, I heard you! [24]. Yahoos
computers were under a Denial Of Service attack while millions of legitimate
customers, who wanted a movie title or an airline ticket, waited. Next day
the same attack was targeted at Amazon.com, eBay, CNN.com, ETrade and
Excite.
The damage caused to these giants of the Web was huge and thus a
57
58
very high-prole search was launched by the Federal Bureau of Investigation (FBI). The common opinion was that the attack must be doings of a
group of sophisticated crackers who had hijacked hundreds of computers in
schools, research labs and businesses and turned them into zombies screaming
Yes, I heard you! to Yahoo and the others.
Finally the FBI solved the crime. They did not nd the much-anticipated
cyberterrorist organization, however. Instead, they tracked down a single
fteen year old teenager living in the Canadian suburbs. Funny thing was
that the FBI would have never found the boy hadnt he been bragging with
his doings in a chat room where he called himself MaaBoy.
The MaaBoy successfully managed to halt the operation of billion-dollar
companies with access to best computer security experts in the world. Although, it was noticed that MaaBoy was not even very skilled among computers he still managed to organize this attack from his suburban home using
his modest desktop computer. And what is more interesting is that these big
companies were completely powerless against this type of rudimentary attack.
What made it possible for this fteenyearold to cause this kind of damage?
The answer lies in the structure of the complex network that he was using to
accomplish his task. Although its clear that he could not have done anything
to these computers directly, he utilized the properties and weaknesses of the
underlying network the Internet. If a mere youth can cause this kind of
harm, what could a small group of trained and skilled professionals achieve?
How vulnerable are we to such attacks?
In the light of the example above, it seems clear that the dierent properties
and the weaknesses of the Internet have to be studied carefully in order
to assure reliable operation of legitimate companies and other everyday
activities that we use it for. However, there have been ndings suggesting
that the same kind of structures and behaviors can be found in a number of
dierent kinds of networks varying from a social network in a cocktail party
to metabolic network within a cell.
Thus, a growing number of researchers believe that there must be some universal laws describing these nets and once we nd them for a certain network
we can possibly utilize the laws in some other nets and explain the behavior
more precisely.
6.1.1
59
Reductionism
Reductionism has been very popular method among scientists in the last
decade. The idea behind reductionism is that to comprehend nature we have
to study its smallest parts. Once we understand the parts, it will be easy
to grasp the whole. Divide and conquer; the devil is in the details. To
understand the universe we have to study atoms and superstrings, molecules
to comprehend life, individual genes to understand complex human behavior
and so on.
However, even if we know almost everything there is to know about the pieces
we are still as far as we have ever been from understanding the nature as a
whole. The main problem with reductionism is that once we start to put our
little pieces of the puzzle together, we run into the hard wall of complexity.
There are billions of ways to do this assembly, if you will, and we would need
something more than just the knowledge of the pieces. One possible remedy
to this problem is networks, they are everywhere. All we need is an eye
for them. Once the general laws governing the properties and formation of
dierent networks are studied and understood, we may have a very powerful
roadmap or blueprint describing a variety of our complex problems.
6.2
Graph theory can be considered the basis for todays thinking about networks. It was born in 1736 when Leonhard Euler, a Swiss born mathematician who spent most of his career in Berlin and St. Petersburg, oered a
proof to a problem that had been troubling the people of Knigsberg [8].
6.2.1
Knigsberg bridges
60
Once Euler heard about this he was intrigued about the problem and decided
to search for a solution. This he quickly found and was able to proof that such
a path does not exist. He even wrote a short paper describing his solution
[5]. The fact that Euler could solve the problem is not that interesting to
us, but rather the intermediate step that he took in the process: Euler had
the foresight to simplify his problem by considering the problem as a graph
consisting of links and nodes. The nodes in this case are the four pieces of
land and the links are the seven bridges connecting the nodes. Once this
is done it is easy to proof generally that if a node has an odd number of
links then it must be either a starting point or an end point of the journey.
This means that there cannot be more than two nodes with odd number
of links and in the case of the bridges there were four i.e. all of the nodes
had odd number of links. Thus the path cannot exist even if we spent our
life searching for it. This is important to notice: It does not depend on our
ability to nd such a path; rather it is a property of the graph. Finally the
people of Knigsberg agreed with Euler and later on they built yet another
bridge, solving the problem.
There were many consequences to this intermediate step Euler took in his
proof. First of all it launched an avalanche of studies and contributions
made by mathematical giants such as Cauchy, Hamilton, Cayley, Kircho
and Plya. On the other hand it started the thinking of networks. The fact
that there are many types of networks (or graphs at that time) and that they
can have properties hidden in their construction that enable or restrict the
ability to do dierent things with them was acknowledged. Also, by making
even very small changes to these properties we can drastically change the
behavior of the network and perhaps open hidden doors that enable new
possibilities to emerge as was the case in Knigsberg; just by adding a single
link to the graph the whole problem was solved.
61
After Eulers work the main goal of graph theory was to discover and catalogue various graphs and their properties. For example labyrinth problems,
rst solved in 1873, were very famous at that time along with problems like
nding a sequence of moves with a knight on a chess board such that each
square is visited only once and that the knight returns to its starting point.
Also things like the lattice formed by atoms in a crystal or the hexagonal
lattice made by bees in a beehive were studied.
6.2.2
Random networks
It took two centuries after the advent of graph theory before scientists started
moving from studying various properties of dierent graphs to asking how
networks form. What are the laws governing their appearance, structure and
changes in them? The rst answers to these questions were given by two
Hungarian mathematicians Paul Erdos and Alfrd Rnyi when in 1959 they
published their paper about random networks [9].
The idea behind random networks was that when creating a network the
links between nodes were added completely randomly. Once they started
adding random links one by one they also noticed that when the average
amount of links per node approached one, something special happened; a
single network emerged. Physicists would perhaps call this phenomenon a
phase transition, similar to the moment in which water freezes. Sociologists
would say that your subjects had just formed a community, but all would
agree that something special happened: After placing critical amount of links,
the network drastically changed. The properties of random networks were
further studied and one of Erdoss students, Bla Bollobs, proofed in 1982
that the histogram of links per node in a random network follows a Poisson
distribution [6].
Poisson distribution tells us that the majority of the nodes have the same
number of links as the average node has (peak value of the distribution)
and that it is extremely rare that there exists a node in the graph that
would deviate signicantly from this prototype node. When translated for
example to society, we end up with very democratic world in which everyone
has on the average the same number of friends or other social links. It tells
that most companies trade with roughly the same number of companies, most
neurons connect to roughly the same number of neurons, most Websites are
visited by roughly the same number of people and so on.
But is this the way that nature truly behaves? Of course it is not. Although
Erdos and Rnyi understood that there is very diverse spectrum of dierent
62
6.3
Degrees of separation
One interesting aspect in networks that has also intrigued scientists over the
decades is their smallness; even if the network at hand is very large, still often
it seems to be so interconnected that the average distance between nodes is
surprisingly small. One of the most important concepts in this area is the so
called six degrees of separation [13]. The concept was born from the studies
of society made by Frigyes Karinthy. In his book Lncszemek (or Chains
in english) he proposes that everybody is at most ve handshakes away from
each other. He presents few interesting examples in which he can connect
himself to any Nobelist or to any worker in Henry Fords car factory.
Although the six degrees of separation did not draw too much attention in
Karinthys time it was rediscovered almost three decades later, in 1967, by
Stanley Milgram, a Harvard professor who turned the concept into famous
study on our interconnectivity [21]. He modied Karinthys idea so that he
could estimate the distance between any two people in United States. In
his experiment he sent the following letter along with other information to
160 randomly chosen people living in Wichita and Omaha.
HOW TO TAKE PART IN THIS STUDY
1. ADD YOUR NAME TO THE ROSTER AT THE BOTTOM OF THIS
SHEET, so that the next person who receives this letter will know who
it came from.
2. DETACH ONE POSTCARD. FILL IT OUT AND RETURN TO HARVARD UNIVERSITY. No stamp is needed. The postcard is very important. It allows us to keep track of the progress of the folder as it
moves toward the target person.
63
64
billion plus documents that are connected to each other through hyperlinks
in the World Wide Web (WWW) [2].
The problem can also be approached completely analytically by using the
theory of random networks. It can be shown that generally the degree of
N
separation d for random networks is the following: d = log
, where N = size
log k
of the net and k = average number of links per node. As can be seen from
the logarithmic behavior of d, the degree of separation grows very slowly
even if the size of the network becomes large. This tells us in clear terms of
mathematics that when it comes to interconnected networks; we live in a
small world.
6.4
A sta writer at the New Yorker magazine called Malcolm Gladwell wrote a
book named The Tipping Point where he introduces a simple test: The task
is to go through a list of 248 surnames that he had compiled from Manhattan
phone book and count the number of people from your own circle of friends
and acquaintances that have their surname appearing on the list, counting
also multiple occurrences [11]. The purpose of the test was to measure the
social connectedness of dierent people. Gladwell tested about four hundred
people including several dierent types and also very homogenous groups
of people (similar age, education and income). The average was changing
according to education, age and social status of dierent groups and was
somewhere between 20 and 40. This was not so important, however, and
could be easily explained but the interesting thing in these experiments was
that the variation in the results was very large and independent of the group
or the homogeneity of the people in the group. The lowest score was typically
between 2 and 16 and the highest 95 and 118. Gladwell found that in every
test there are a small number of people that have the special ability to make
friends and acquaintances more easily than others. He called these people
connectors but in network terms they can be called hubs. Same kind of phenomena has been studied in many dierent networks (e.g. WWW, Internet,
social nets) [16], [10].
Yet, the random universe discussed above does not support connectors. If the
networks were random then the existence of the hubs would be practically
forbidden due to their increasingly small probability. If the WWW were a
random network, the probability of there being a page with ve hundred
incoming links would be 1099 . Yet the latest Web survey, covering roughly
20% of the full Web, found four hundred such documents. The document with
65
most incoming links had over two million links. The chance of nding such
a document from a random network is smaller than locating an individual
atom in the universe. Thus, new laws for describing networks had to be
found.
6.5
When researchers were studying the WWW it was found that the number
of links between HTML Hyper Text Markup Language) pages did not follow
a uniform distribution. From real data collected by robots crawling through
the WWW it was found that when the data was plotted on a log-log scale it
was a straight line. This was a strong indication that the distribution follows
a power law [7]. This implies that the number of WebPages with exactly k
links, denoted by N(k), follows N(k) k , where the parameter is the
degree exponent. For WWW the was recorded to be 2.1 for incoming links
and 2.5 for outgoing links [2].
Same kind of behavior was found in many other networks also. For example by utilizing the vast database of movies (The Internet Movie Database,
www.imdb.com) it was found that the number of actors that had links to
exactly k other actors decays following a power law [10]. Another example
is the cell where the number of molecules interacting with exactly k other
molecules follows also a power law [26].
6.5.1
One interesting implication of the existence of the power laws in certain systems has been around for a long time from the beginning of the 20th
century and is known as the 80/20 rule. The rule was invented by an
inuential Italian economist called Vilfredo Pareto. Pareto was an avid gardener and he happened to notice that 80 percent of his peas were produced
by only 20 percent of the peapods. He made also other similar observations,
for example he noticed that 80 percent of Italys land was owned by only 20
percent of the population and so on. More recently Paretos Law or Principle
got the name 80/20 rule and turned into the Murphys Law of management:
80 percent of prots are produced by only 20 percent of employees, 80 percent of customer service problems are produced by 20 percent of consumers,
80 percent of decisions are made during 20 percent of the meeting time, 80
percent of crime is committed by 20 percent of criminals, and so on [19].
66
The 80/20 rule is a clear implication of the existence of the power laws in
certain systems: It says that there are a smaller number of strong inuencers
(the hubs) in the system and the others are not so important.
6.5.2
Of course not all networks follow a power law distribution. There are many
dierent networks in the universe that have nothing to do with power laws.
But there exists also many real networks that do and they are called scalefree networks. The name comes from the fact that in networks that follow a
power law there is no intrinsic scale in them. There is no single node that we
could pick up and say that this is the prototype-node of this network, like
we can in a network that follows a Bell curve.
A good way to visualize this is to consider the dierences between a random
network and one described by a power-law degree distribution. Lets compare
a U.S. roadmap with an airline routing map. In the roadmap cities are the
nodes and the highways connecting them are the links. As one can imagine
the roadmap is fairly uniform network: Each major city has at least one link
to the highway system and on the other hand there are no cities that would
have hundreds of links. Another graph is the airline routing map where the
situation is completely dierent. In this map airports (nodes) are connected
to other airports by direct ights between them (links). In this case it is clear
that there exists a few major airports in the country that have a high number
of ights and a large number of smaller airports that have only few ights.
When the two networks are compared it is clear that they have very dierent
properties and this is why they also look very dierent. The roadmap follows
a Bell curve telling that vast majority of the nodes has the same number of
links and that it is extremely rare to nd a node that would dier greatly from
the average, which is located at the peak value of the degree distribution. On
the other hand the airline map follows a power law distribution and there is
no single node that could be selected as a prototype of the network. Rather
there exist few major inuencers (hubs) that take care of most of the trac
and shape the network to be scale-free. As can be seen, scale-free networks
are one special case of networks and they dene the topology of the network
to such that there is no prototype-node but the hubs in the net dene the
behavior [3].
6.5.3
67
Let us consider the two types of networks presented above, and think them
as communication networks where each node is a communication point and
the messages can be sent via links between nodes. Which would be more
reliable?
As usual, the answer is not unambiguous: If we consider the robustness
against random errors (equipment failures for example) then it has been
studied that the scale-free networks are extremely error tolerant. It has been
found that typically nearly 80% of the nodes can be randomly removed until
the network is crippled. On the other hand random network is fairly easily
halted. This dierence in behavior is because of the connecting force of
the hubs in the scale-free network. Since the probability of error occurring
in a node is exactly the same for every node in the network, the scale-free
network can aord to lose a large amount of the less important nodes
without signicant loss in performance.
There is a tradeo however, as the increasing number of hubs in the network
increases its stability against random errors at the same time its tolerance
against targeted attacks is decreased. This is clear because of the important
role of the hubs in the network and is exactly how the MaaBoy accomplished
in his attack (as described in the beginning). The attack was targeted at the
most important hubs and thus the damage was enormous. This tradeo
is very important to notice and should be considered carefully in network
design. Obviously the selection of the network topology should be based on
the boundary conditions (intended use, security aspects and so on) [1].
6.6
68
fads and viruses and the spread of others, social scientists and epidemiologists
have developed a tool called the threshold model. The model is based on the
fact that all people dier in their willingness to adopt a new idea. But in
general, with sucient positive evidence everyone can be convinced. Only the
level of positive evidence diers. Thus, the model gives every individual (or
generally every node) an individual threshold value indicating the likelihood
that he or she will adopt a given innovation. Another parameter called critical
threshold can be calculated utilizing the properties of the network and thus
a spreading rate for a given innovation can be estimated [12].
69
6.7
Today we are learning the language in which God created life, said President
Bill Clinton on June 25, 2000, at the White House ceremony announcing the
decoding of the 3 billion chemical letters of human genome. Was he right?
Now that we have the full DNA of a human being gured out, dont we have
all the information needed to cure dierent diseases and estimate medical
facts? The answer is: No! And the reason should be clear by now. As
explained earlier when talking about reductionism, we run again into the
hard wall of complexity. We only have the pieces but not the blueprint. To
cure most illnesses, we need to understand living systems in their integrity.
We need to understand how and when dierent genes work together, how
messages travel within the cell, which reactions are taking place or not in
this complex cellular network.
However, studies have shown that there is scalefreenetworkbehavior present
also in metabolic networks: For example, a study conducted on 43 dierent
organisms on how many reactions each molecule participates in showed that
all the nets were scalefree with three degrees of separation [15]. There are
also many other examples showing that networks are present in living organisms [25]. So the question is; if we improve our network-thinking abilities
can we also improve our understanding of living systems? Many researchers
seem to think that we can.
6.8
Conclusions
Networks are present almost everywhere, all we need is an eye for them.
Perhaps if we shift our way of thinking towards network-world we can open
up new doors for science. One interesting area of research would be the
dynamical models of the changes taking place in dierent networks. After
all, most of the real networks are under continuous change and the networks
describing them are only the skeletons of complexity. Identifying the underlying structures and their properties gives us a map for our journey through
complex worlds!
70
Bibliography
1. Albert R., Jeong H., Barabsi A-L.: Attack and Error Tolerance of
Complex Networks.
Nature, vol. 406, 2000, p. 378.
2. Albert R., Jeong H., Barabsi A-L.: Diameter of the World Wide Web.
Nature, vol. 401, 1999, pp. 130131.
3. Amaral L. A. N., Scala A., Barthlmy M., Stanley H. E.: Classes
of SmallWorld Networks. Proceedings of the National Academy of
Sciences, vol. 97, 2000, pp. 1114911152.
4. Barabsi A-L.: LINKED, The New Science of Networks. Perseus Publishing, Cambridge, Massachusettts, 2002.
5. Biggs N. L., Lloyd E. K., Wilson R. J.: Graph Theory: 17361936.
Clanderon Press, Oxford, England, 1976.
6. Bollobs B.: Degree Sequences of Random Graphs.
Discrete Mathematics, vol. 33, 1981, p. 1.
7. Buchanan M.: Ubiquity: The Science of History . . . Or Why the
World Is Simpler Than We Think. Crown Publishers, New York, 2001.
8. Durham W.: Euler: The Master of Us All. Mathematical Association
of America, Washington D.C., 1999.
9. Erds P., Rnyi A.: On Random Graphs I.
Math. Debrecen, vol. 6, 1959, pp. 290297.
10. Fass G., Ginelli M., Turtle B.: Six Degrees of Kevin Bacon. Plume,
New York, 1996.
11. Gladwell M.: The Tipping Point. Little, Brown, New York, 2000.
12. Granovetter M.: Threshold Models of Collective Behavior.
American Journal of Sociology, vol. 83, no. 6, 1978, pp. 14201443.
13. Guare John: Six Degrees of Separation. Random House, New York,
USA, 1990.
14. Jeong H., Mason S., Barabsi A-L., Oltvai Z. N.: Centrality and Lethality of Protein Networks.
Nature, vol. 411, 2001, pp. 4142.
6.8. Conclusions
71
15. Jeong H., Tombor B., Albert R., Oltvai Z. N., Barabsi A-L.: The
Large-Scale Organization of Metabolic Networks.
Nature, vol. 407, 2000, pp. 651654.
16. Kleinberg J.: Authoritative Sources in a Hyperlinked Environment.
Proceedings of the 9th Association for Computing Machinery Society for Industrial and Applied Mathematics. Symposium on Discrete
Algorithms, 1998.
17. Kleinfeld J.: Six Degrees of Separation: An Urban Myth.
Psychology Today, 2002.
18. Kleinfeld J.: The Small World Problem.
Society, vol. 39, 2002, pp. 6166.
19. Koch R.: The 80/20 Principle The Secret to Success by Achieving
More with Less. Currency, New York, USA, 1998.
20. Mandelbaum R.: Only in America.
New York Times Magazine, November 26, 2000.
21. Milgram S.: The Small World Problem.
Physiology Today, vol. 2, 1967, pp. 6067.
22. Montoya J. M., Sol R. V.: Small World Patterns in Food Webs.
http://www.santafe.edu/s/publications/Abstracts/00-10-059abs.html.
23. Newman M. E. J.: The Structure of Scientic Collaboration Networks.
Proceedings of the National Academy of Sciences of the United States
of America, vol. 98, January 16, 2001, p. 404409.
24. Taylor C.: Behind the Hack Attack.
Time Magazine, February 21, 2000.
25. Venter J. C. et al: The Sequence of the Human Genome.
Science, vol. 291, 2001, pp. 13041351.
26. Wagner A., Fell D.: The Small World Inside Large Metabolic Networks.
Proceedings of the Royal Society of London, Series BBiological Sciences, vol. 268, September 7, 2001, pp. 18031810.
27. Watts D. J., Strogatz S. H.: Collective Dynamics of SmallWorld
Networks.
Nature, vol. 393, 1998, pp. 440442.
72
28. Williams R. J., Martinez N. D., Berlow E. L., Dunne J. A., Barabsi
A-L.: Two Degrees of Separation in Complex Food Webs.
http://www.santafe.edu/s/publications/Abstracts/01-07-036abs.html.
Session 7
Cellular Automata
Juan Li
Communications Laboratory, HUT
First beginning with introduction to the history of cellular automata, this
paper pesents several important ideas in this eld. Then a brief analysis of
1-D cellular automata is given. Finally, some more advanced applications of
cellular automata are demonstrated.
7.1
Introduction
Automata is the plural of automaton. While the word automaton conjures up the
image of a mechanical toy or a soulless organism, in computer science it has a very precise
meaning. It refers to all machines whose output behaviour is not a direct consequence
of the current input, but of some past history of its inputs. They are characterised as
having an internal state which is a repository of this past experience. The inner state of
an automaton is private to the automaton, and is not available to an external observer.
73
74
make pretty pictures. For another, they are related to exciting new ideas
such as articial life and the edge of chaos. For a fairly simple example see
[1].
The cellular automaton provides a way of viewing whole populations of
interacting cells, each of which is itself a computer (automaton). By building appropriate rules into a cellular automaton, we can simulate many kinds
of complex behaviours, ranging from the motion of uids governed by the
Navier-Stokes equations to outbreaks of starsh on a coral reef.
7.2
7.2.1
7.2.2
Mathematician Stanislaw M. Ulam liked to invent pattern games for the computer at Los Alamos. Given certain xed rules, the computer would print
out ever-changing patterns. Many patterns grew almost as if they were alive.
A simple square would evolve into a delicate, coral-like growth. Two patterns
would ght over territory, sometimes leading to mutual annihilation. He developed 3-D games too, constructing thickets of coloured cubes as prescribed
by computer. He called the patterns recursively dened geometric objects.
Ulams games were cellular games. Each pattern was composed of square (or
triangular or hexagonal) cells. In eect, the games were played on limitless
75
chessboards. All growth and change of patterns took place in discrete jumps.
From moment to moment, the fate of a given cell depended only on the states
of its neighbouring cells. The advantage of the cellular structure is that it
allows a much simpler physics. Without the cellular structure, there would
be innitely many possible connections between components.
Ulam suggested that John von Neumann construct an abstract universe
for his analysis of machine reproduction. It would be an imaginary world
with self-consistent rules, as in Ulams computer games. It would be a world
complex enough to embrace all the essentials of machine operation, but otherwise as simple as possible. Von Neumann adopted an innite chessboard
as his universe. Each square cell could be in any of a number of states corresponding roughly to machine components. A machine was a pattern of
such cells. The rules governing the world would be a simplied physics. A
proof of machine reproduction would be easier to devise in such an imaginary
world, as all the nonessential points of engineering would be stripped away
[1]. Ironically, the name von Neumann is now strongly associated with the
old-fashioned, single-CPU computer architecture. He was also the major pioneer in parallel computing via his research on arrays of computer or cellular
automata.
In 1944, von Neumann was introduced to electronic computing via a description of the ENIAC by Goldstine. Shortly, he formed a group of scientists
headed by himself, to work on problems in computers, communications, control, time-series analysis, nd the communications and control aspects of the
nervous system. The last topic was included due to his great interest in the
work on neural networks of McCulloch and Pitts. In 1946, von Neumann proceeded to design the EDVAC (Electronic Discrete Variable Computer) which
was the rst design of the ideas on automata developed by Post (1936) and
Turing (1936), he had commenced studies on the complexity required for a
device or system to be self-reproductive. These studies also included work
on the problem of organizing a system from basically unreliable parts (a eld
of study which we now know as fault tolerant computing). At rst, von
Neumann investigated a continuous model of a self-reproducing automaton
based on a system of non-linear partial dierential equations, essentially
of the diusion type. He also pursued the idea of a kinematic automaton
which could, using a description of itself, proceed to mechanically assemble
a duplicate from available piece parts.
When von Neumann found it dicult to provide the rigorous and explicit
rules and instructions needed to realize such an automaton and when it became evident that the value of such an automata would be moot, he redirected
76
7.2.3
77
7.2.4
78
79
Petri dish as a whole weighs the same before and after. However, no such
restriction applies in Life, the amount of matter in the Life universe can
uctuate arbitrarily.
Simple rules can have complex consequences. The simple set of rules in Life
has such a wealth of implications that it is worth examining in detail. Life is
forward-deterministic: This means that a given pattern leads into one, and
only one sequel pattern. The Game of Life is not backward-deterministic; a
pattern usually has many patterns that may have preceded it. In short, a
conguration has only one future, but usually many possible pasts. This fact
is responsible for one of the occasional frustrations of playing Life. Sometimes
you will see something interesting happen, stop the program, and be unable
to backtrack and repeat it. There is no simple way you can program a
computer to go backward from a Life state there are too many possibilities.
Conway had tried to design Lifes rules so that unlimited growth patterns
were possible. It was not immediately obvious that he had succeeded, though.
Conway conceived of two ways of demonstrating that unlimited growth is
possible, if indeed it is. Both ways involved nding a pattern that does grow
forever, yet in a disciplined, predictable manner.
In 1970, in response to Conways challenge in the Scientic American that a
nite initial conguration of Life would not be able to generate innite populations, R.W. Gosper used a DEC PDP-6 computer to run Life simulations
far quicker than could be done by hand. Eventually, they found the glider
gun, which generated gliders continually, and a puer train that moved
on the grid, leaving behind a constant trail of live cells. They even found
puer trains that emitted gliders that collided to make glider guns which
then emitted gliders, but in a quadratically increasing number ....
From this breakthrough, Conway could prove that Life could indeed support
universal computation. Using glider streams to represent bits, he was able
to produce the equivalent of and-gates, or-gates and not-gates, as well as an
analogue of a computers internal storage.
7.2.5
80
7.2.6
81
Norman Packard was a physicist working with Stephen Wolfram at the Institute for Advanced Study, and chose to apply information theory to cellular
automata to emulate snowakes. He reasoned that rstly, it was not the
case that every snowake was identical. However, if every snowake had a
random structure, the information from each snowake would be meaningless. Therefore, there is a syntax within snowakes, several main types of
structures, which are capable of containing individual variations.
Packard discovered that dierent weather conditions result in snowakes taking on dierent general aspects. One set of conditions yields congurations
that look like plates, another determines snowakes shaped like collections
of rods, and another yields dendritic stars. He wrote a cellular automaton
simulation in which the o cells, those with a value of 0, represented water vapour, and the on cells, those assigned a value of 1, represented ice,
and appeared on the screen in colour. The snowake would grow outwards
from its boundary. A typical set of rules initiated a scan of a cells neighbourhood, summed the values of the surrounding cells, and lled the new
cells with either ice or water vapour, depending on whether the toal was odd
or even.
The resulting articial snowakes lacked the complexity of real snowakes,
particularly those with structures based on patterns of needle-like shapes
but they did have plates and dendrites growing from the corners of the plates,
from which more dendrites grew; they were easily idenable as snowakes.
7.3
82
n
C
n
n
n
n
n n
C n
n n
n
n
n
n n n
C n n
n n n
Its PROGRAM is the set of rules that denes how its state changes in
response to its current state, and that of its neighbours.
Example Fabric patterns
Imagine a CA consisting of a line of cells as follows:
States: 0 or 1
Neighbourhood: The two adjacent cells (or n C n)
Rules: The following list shows the new state of a cell for each possible
local conguration, i.e. arrangement of states for the cell and its two
neighbours. Because there are two possible states (0 or 1) for each of
3 cells, there are 23 = 8 rules required:
0 0 0
0 1 1
1 1 0
0
0
0
0
1
1
0 1
0 0
1 1
1
1
0
0 1
1 0
0 1
1 1
Suppose that we start with just a single cell in state 1. Then here is how the
array will change with time (see Fig. 7.2; here . denotes 0 for clarity).
When we plot the successive states as shown above, some properties emerge:
Self-organization: Notice that when we plot the successive states as shown
in the above example a pattern emerges. Even if the line of cells starts
with a random arrangement of states, the rules force patterns to emerge;
the pattern depends on the set of rules, for examples see Fig. 7.3.
83
84
7.4
ai
(t)
(t)
= ai1 + ai+1
mod 2,
(7.1)
N
1
i=0
ai xi
(7.2)
7.5. Applications
85
where x is a dummy variable, and the coecient of xi is the value of the site
at position i. In terms of characteristic polynomials, the cellular automaton
rule (7.1) takes on the particularly simple form
A(t+1) (x) = T (x)A(t) (x)
mod
(xN 1)
(7.3)
where
T (x) = (x + x1 ),
(7.4)
(7.5)
7.5
Applications
Cellular automata applications are diverse and numerous. The laws in our
Universe are only partly known and appear to be highly complex, whereas in
a cellular automaton laws are simple and completely known. One can then
test and analyse the global behaviour of a simplied universe, for example:
86
In dierent elds, cellular automata can be used as an alternative to dierential equations. Cellular automata can also be used as graphic generators2 .
The several images in Fig. 7.5 show some graphic eects.
According to Rucker et Walker op. cit., In five years you wont be able to watch
television for an hour without seeing some kind of cellular automata.
7.5. Applications
87
Bibliography
1. http://math.hws.edu/xJava/CA/
2. Kendall Preston, Jr. and Michael J.B.Du, Modern Cellular Automata:
Theory and Applications. Plenum Press, New York, 1984.
3. http://life.csu.edu.au/complex/tutorials/tutorial1.html
4. http://shakti.trincoll.edu/ pkenned3/alife.html
5. http://math.hws.edu/xJava/CA/CA.html
6. http://www.tu-bs.de/institute/WiR/weimar/ZAscriptnew/intro.html
7. http://www.stephenwolfram.com/publications/articles/ca/85-two/2/text.html
8. http://www.brunel.ac.uk/depts/AI/alife/al-ca.htm
9. http://www.rennard.org/alife/english/acintrogb01.html
10. http://www.ce.unipr.it/pardis/CNN/cnn.html
11. http://cafaq.com/soft/index.shtml
12. N. Howard, R. Taylor, and N. Allinson.The design and implementation
of a massively-parallel fuzzy architecture. Proc. IEEE, pages 545-552,
March 1992
88
Session 8
From Chaos ...
Complexity theory cannot be understood without taking a look at chaos theory. Chaos theory ourished in the 1980s; it had been observed that simple
rules can result in very complicated structures. Only after that, in complexity theory, the emphasis was turned in the opposite direction, searcing for
the underlying rules beneath the observed complexity. Thus, in complexity research there are many underlying assumptions, etc., that are directly
inherited from chaos research.
Unfortunately, this chapter was never delivered in the correct format to be
included in this report.
89
90
Session 9
. . . Towards New Order
Lasse Eriksson
Helsinki University of Technology
Control Engineering Laboratory
Lasse.Eriksson@hut.fi
In this chapter we are moving from chaos towards new order i.e. complexity. The complexity lies at the edge of chaos (EOC), which is the border
area between static and chaotic behavior. Other subjects in this chapter such
as self-organized criticality (SOC), highly optimized tolerance (HOT) and
measures of complexity are closely related to EOC.
9.1
Introduction to complexity
It is obvious that chaos theory existed long before anyone spoke about complexity. But there was a eld called Catastrophe theory even before chaos.
Catastrophe theory studies and classies phenomena characterized by sudden
shifts in behavior arising from small changes in circumstances. Originated
by the French mathematician Rene Thom in the 1960s, catastrophe theory
is a special branch of dynamical systems theory [1]. Anyway, as James P.
Sethna from Cornell University, Ithaca says: . . . the big fashion was topological defects. Everybody was . . . nding exotic systems to write papers about.
It was, in the end, a reasonable thing to do. The next fashion, catastrophe
theory, never became important for anything [2].
91
92
The successor of catastrophe theory, chaos theory has been described in the
previous chapter. In this chapter the emphasis is on the latest research area
of this eld, i.e. complexity and complex systems. There has been a lot of
discussion whether complexity is any dierent from chaos. Is there really
something new or are we talking about the same subject with a new name?
Langtons (father of EOC) famous egg diagram in Fig. 9.1 gives some insight
to this matter. The behavior of systems has been divided into four dierent
classes in the egg. These are Fixed, Periodic, Complex and Chaotic.
It is justied to say that complex behavior is dierent from chaotic behavior.
Thus complexity is not chaos and there is something new in this eld. But
yet the things studied in the eld of complexity are somehow familiar from
the past.
93
on the language that is used to model the system. The original Latin word
complexus signies entwined or twisted together. This may be interpreted
in the following way: in order to have a complex you need two or more components, which are joined in such a way that it is dicult to separate them
[5].
Another important term in the eld of complexity is emergence, which is [4]
1. What parts of a system do together that they would not do by themselves (collective behavior). How behavior at a larger scale of the system arises from the detailed structure, behavior and relationships on a
ner scale.
2. What a system does by virtue of its relationship to its environment
that it would not do by itself.
3. Act or process of becoming an emergent system.
Emergence refers to all the properties that we assign to a system that are
really properties of the relationship between a system and its environment.
9.2
Self-organized criticality
9.2.1
94
9.2.2
SOC
Per Bak, Chao Tang, and Kurt Wiesenfeld introduced the concept of selforganized criticality in 1987. SOC refers to tendency of large dissipative
systems to drive themselves to a critical state with a wide range of length
and time scales. As an example, consider a damped spring as in Fig. 9.2.
A mass m is attached to the spring, which is fastened into the wall on the
left. The spring constant is k and the damping coecient is B. The distance
from the equilibrium point is x(t).
m
x
B
Figure 9.2: A damped spring.
From the rst principles it is easy to model the spring system traditionally.
m
x(t) + B x(t)
+ kx(t) = 0
x(t) =
1
[B x(t)
+ kx(t)]
m
(9.1)
95
k = 0.1 N
m
B = 0.08 kg
s
m
=
1kg
x(0)
= 1 ms
(9.2)
1
s
Integrator
Integrator1
To Workspace
1/m
k
Gain2
x 10
3
2.5
3
2
ZOOM
0.5
Amplitude (m)
Amplitude (m)
1.5
0.5
1.5
3
2
165
0
20
40
60
80
100
Time (s)
120
140
160
180
200
170
175
180
185
190
195
Time (s)
The scale is very dierent in the two plots. As we can see from the simulation
results, the oscillatory behavior of the spring with a decreasing amplitude
theoretically continues forever. In real world, the motion would stop because
of the imperfections such as dust. Once the amplitude gets small enough, the
emotion suddenly stops. This generally occurs when the velocity is smallest
i.e. at the top or at the bottom of an oscillation. This is not the state
96
9.2.3
Adding sand slowly to a at sand pile will result only in some local rearrangement of particles. The individual grains, or degrees of freedom, do not
interact over large distances. Continuing the process will result in the slope
increasing to a critical value where an additional grain of sand gives rise to
avalanches of any size, from a single grain falling up to the full size of the
sand pile. The pile can no longer be described in terms of local degrees of
freedom, but only a holistic description in terms of one sand pile will do. The
distribution of avalanches follows a power law.
If the slope of the pile were too steep, one would obtain a large avalanche
97
and a collapse to a atter and more stable conguration. If the slope were
too shallow, the new sand would just accumulate to make the pile steeper.
If the process is modied, for instance by using wet sand instead of dry
sand, the pile will modify its slope during a transient period and return to
a new critical state. Consider for example snow screens. If they were built
to prevent avalanches, the snow pile would again respond by locally building
up to steeper states, and large avalanches would resume.
The sand pile model can be simulated and there are some Java-applets available on the Internet (e.g. [8]). Typically, the simulation model is such that
there is a 2-D lattice (cellular automaton) with N sites. Integer values zi
represent the local sand pile height at each site i. If the height exceeds some
critical height zcr (lets say three), then one grain is transferred from the
unstable site to each of the four neighboring site (this is called a toppling).
A toppling may initiate a chain reaction, where the total number of topplings
is a measure of the size of an avalanche. To explore the SOC of the sand
pile model, one can randomly add sand onto the pile and have the system
relax. The result is unpredictable and one can only simulate the resulting
avalanche to see the outcome. After adding a large amount of sand, the conguration seems random, but some subtle correlations exist (e.g. never do
two black cells lie adjacent to each other, nor does any site have four black
neighbors). Avalanche is triggered if a small amount of sand is added to a
site near the center. The size of an avalanche is calculated and the simulation
is repeated. After the simulations, it is possible to analyze the distribution
of the avalanche sizes. The distribution of avalanche size follows a power law
P (s) = s1 ,
= 2.1
(9.3)
98
9.3
In this section two approaches to the edge of chaos phenomenon are discussed
as well as the measures of complexity. Phase transitions and their relation to
Wolframs four classes (discussed in the earlier chapters) are also presented.
9.3.1
The edge of chaos is the interesting area, where complex rather than chaotic
or static behavior arises. Backed by Kaumans work on co-evolution, Wolframs cellular automata studies, and Baks investigations of self-organized
criticality, Langton has proposed (in 1990) the general thesis that complex
systems emerge and maintain on the edge of chaos, the narrow domain between frozen constancy and chaotic turbulence. The edge of chaos idea is
another step towards an elusive general denition of complexity [9].
In general, some cellular automata are boring since all cells either die after
few generations or they quickly settle into simple repeating patterns. You
could call these highly ordered cellular automata because their behavior
is predictable. Other cellular automata are boring because their behavior
seems to be totally random. You could call these chaotic cellular automata
because their behavior is totally unpredictable. On the other hand, some
cellular automata show interesting (complex, lifelike) behavior, which arises
near the border of between chaos and order. If these cellular automata were
more ordered, they would be predictable, and if they were less ordered, they
would be chaotic. This boundary is called the edge of chaos.
Christopher Langton introduced a 1-D cellular automaton, whose cells have
two states. The cells are either alive or dead. If a cell and its neighbors are
dead, they will remain dead in the next generation. Langton dened a simple
number that can be used to help predict whether a given cellular automaton
will fall in the ordered realm, in the chaotic realm, or near the boundary.
The number lambda (0 < < 1) can be computed from the rule space of the
cellular automaton. The value is simply the fraction of rules in which the
new state of the cell is living. The number of rules R of a cellular automaton
is determined by R = K N , where K is the number of possible states and N
is the number of neighbors. If = 0, the cells will die immediately and if
= 1, any cell with a living neighbor will live forever. Values of close to
zero give cellular automata in the ordered realm, and values near one give
cellular automata in the chaotic realm. The edge of chaos is somewhere in
between.
99
Value of does not simply represent the edge of chaos. It is more complicated. You could start with = 0 (death) and add randomly rules that lead
to life instead of death ( > 0). You would get a sequence of cellular
automata with values of increasing from zero to one. In the beginning,
the cellular automata would be highly ordered and in the end they would
be chaotic. Somewhere in between, at some critical value of , there would
be a transition from order to chaos. It is near this transition that the most
interesting cellular automata tend to be found, the ones that have the most
complex behavior. The critical value of is not a universal constant [10].
9.3.2
We can approach the edge of chaos from another point of view. This view
is based on measuring so called perturbation strength. Let us consider some
examples. The rst one is domino blocks in a row. The blocks are rst in
a stable state, standing still. The initial state is in a way minimally stable,
because a small perturbation can avalanche through the whole system. Once
the rst block is nudged, an avalanche is started. The system will become
stable again once all blocks are lying down. The nudge is called perturbation
and the duration of the avalanche is called transient. The strength of the
perturbation can be measured in terms of the eect it had i.e. the length
of time the disturbance lasted (or the transient length) plus the permanent
change that resulted (none in the domino case).
Other examples of perturbation strength are buildings in earthquakes, and
air molecules. For buildings, we require short transient length and return to
the initial state (buildings are almost static). Air molecules on the other hand
collide with each other continually, never settling down and never returning
to exactly the same state (molecules are chaotic). For air molecules the
transient length is innite, whereas for our best buildings it would be zero.
How about in the middle?
Consider yet another example, a room full of people. A sentence spoken may
be ignored (zero transient), may start a chain of responses which die out and
are forgotten by everyone (a short transient) or may be so interesting that
the participants will repeat it later to friends who will pass it on to other
people until it changes the world completely (an almost innite transient
e.g. old religions).
Systems with zero transient length are static and systems with innite transient length are chaotic. The instability with order as described in the examples is called the edge of chaos, a system midway between stable and
100
9.3.3
Complexity measures
9.3.4
101
Phase transitions
Phase transition studies came about from the work begun by John von Neumann and carried on by Steven Wolfram in their research of cellular automata. Consider what happens if we heat and cool systems: at high temperatures systems are in gaseous state (chaotic) and at low temperatures
systems are in solid state (static). At some point between high and low temperatures the system changes its state between the two i.e. it makes a phase
transition [11].
There are two kinds of phase transitions: rst order and second order. First
order we are familiar with when ice melts to water. Molecules are forced by
a rise in temperature to choose between order and chaos right at 0 C. This
is a deterministic choice. Second order phase transitions combine chaos and
order. There is a balance of ordered structures that ll up the phase space.
The liquid state is where complex behavior can arise [12].
Consider once again the egg diagram in Fig. 9.1. It shows the schematic
drawing of cellular automaton rule space indicating relative location of periodic, chaotic, and complex transition regimes. Crossing over the lines (in
the egg) produces a discrete jump between behaviors (rst order phase transitions). It is also possible that the transition regime acts as a smooth transition between periodic and chaotic activity (like EOC experiments with ).
This smooth change in dynamical behavior (smooth transition) is primarily
second-order, also called a critical transition [3].
In his research, Wolfram has divided cellular automata into four classes based
on their behavior. Dierent cellular automata seem to settle down to classes
which are: constant eld (Class I), isolated periodic structures (Class II), uniformly chaotic elds (Class III) and isolated structures showing complicated
internal behavior (Class IV). There is a relationship between Wolframs four
classes and the underlying phase transition structure in cellular automatas
rule space [13]. Figure 9.6 represents a schematic drawing of the relationship.
Phase transition feature allows us to control complexity by external forces.
Heating or perturbing the system drives it towards chaotic behavior, and
cooling or isolating the system drives it towards static behavior. This is
seen clearly in relation to brain temperature. Low temperature means static
behavior (hypothermia), medium temperature normal, organized behavior
and high temperatures chaotic behavior (fever).
102
9.4
103
According to Fig. 9.8 the sparks do not matter if the density of the forest
is less than the critical density (0.5927). If the forest density is higher, the
whole forest might get burned because of one spark hit anywhere in the forest.
If HOT is compared with SOC or EOC, we clearly see the dierence. SOC
104
and EOC assume that the interesting phenomena are at criticality. In HOT
state systems actually work over the critical point, where their performance is
optimized. In the forest model example this means that forest can be planted
denser than the critical density. In that way also the yield of the forest can
be increased, and the system is said to run in a HOT state. The idea is
to optimize the yield based on the knowledge of distribution of the sparks.
By HOT mechanism optimization, almost any distribution of sparks gives
a power law distribution of events. There exist both numeric and analytic
results for that.
One HOT mechanism is based on increasing the design degrees of freedom
(DDOF). The goal is to optimize the yield, or in other words, push it towards
the upper bound. The upper bound is the case where there would be no
sparks and the yield would be the same as the density of the forest. The
DDOFs are in this case the densities of dierent areas on the lattice. The
lattice can be divided into smaller lattices. Each small lattice represents a
design degree of freedom.
The HOT states specically optimize yield in the presence of a constraint.
A HOT state corresponds to forest, which is densely planted to maximize
the timber yield, with rebreaks arranged to minimize the spread damage.
The rebreaks could be of forest planted with density of the critical density.
Depending on the distribution of the sparks and the optimizing method, it
might be possible to plant the forest with density one everywhere else but in
the small rebreaks with a critical density. In practice this means that we
could have a lattice of total density 0.85 and with a yield 0.77, for example.
There are dierent HOT mechanisms that can be used to optimize the yield.
Figures 9.10, 9.11 and 9.12 represent three dierent densities of forest that
result, if three dierent methods are used in yield optimization. The assumed
distribution of sparks is presented in Fig. 9.9. In the upper left corner of
Fig. 9.9 the probability of sparks is almost 0.2, but in the other corners of
the lattice it is lower than 1010 . In the middle the probability decreases
from left upper corner towards right lower corner as can be seen.
The three design mechanisms in Figs. 9.10 - 9.12 are called grid, evolved
and DDOF. In grid design, the rebreaks are straight lines and the density
of the forest between the rebreaks is constant (in the small quadrangles).
The evolved mechanism produces such forest density that almost everywhere
the density is one. Only in the upper left corner of the lattice lower densities
are used. The DDOF mechanism produces as well a very dense forest. Even
the rebreaks are planted with a critical density and everywhere else the
density is one. The design problem in the DDOF is to optimize the areas
105
106
the system was not designed to handle. The real work with HOT is in new
Internet protocol design (optimizing the throughput of a network by operating in HOT state), forest re suppression, ecosystem management, analysis
of biological regulatory networks and convergent networking protocols [15].
9.5
Conclusions
In this chapter the emphasis has been on complexity. It has been approached
from dierent directions and some essential features and terminology have
been discussed. Catastrophe-, chaos- and complexity theory all have common
features, but dierent approaches. Complexity adds dimensions to chaos. All
these branches have a common problem, lack of useful applications. With
complexity there is still hope. . .
Self-organized criticality was discussed and it refers to tendency of large
dissipative systems to drive themselves into a critical state. Coupled systems
may collapse because of an avalanche that results from a perturbation to
the system. As an example of SOC a sandpile model (although not a very
realistic one) was discussed.
Edge of chaos is the border area where the interesting phenomena (i.e. complexity) arise. The edge lies between periodic and chaotic behavior. When
cellular automata are concerned the fraction of rules that lead to alive state
determine so called value. With some values of static behavior can be
detected and with some other values chaotic behavior is seen. Somewhere
in between with some critical value of complexity arises. That value of
represents the edge of chaos.
Highly optimized tolerance (HOT) was discussed in the last section. The idea
is to optimize the prot, yield or throughput of a complex system. By design
we can reduce the risk of catastrophes, but the resulting systems operating
in HOT states with considerable performance are yet fragile.
9.5. Conclusions
107
Bibliography
1. http://www.exploratorium.edu/complexity/CompLexicon/catastrophe.html
2. http://www.lassp.cornell.edu/sethna/OrderParameters/TopologicalDefects.html
3. http://www.theory.org/complexity/cdpt/html/node5.html
4. http://necsi.org/guide/concepts/
5. http://pespmc1.vub.ac.be/COMPLEXI.html
6. Bunde, A.;Havlin, S. (Eds.): Fractals in Science. Springer Verlag, 1994.
7. http://cmth.phy.bnl.gov/maslov/soc.htm
8. http://cmth.phy.bnl.gov/maslov/Sandpile.htm
9. http://pespmc1.vub.ac.be/CAS.html
10. http://math.hws.edu/xJava/CA/EdgeOfChaos.html
11. http://www.calresco.org/perturb.htm
12. http://www.wfu.edu/petrejh4/PhaseTransition.htm
13. http://delta.cs.cinvestav.mx/mcintosh/newweb/what/node8.html
14. Carlson, J.M.; Doyle, J.: Highly Optimized Tolerance: Robustness and
Design in Complex Systems. 1999
15. http://www.cds.caltech.edu/doyle/CmplxNets/HOT_intro.ppt
108
Session 10
Self-Similarity and Power Laws
Tiina Komulainen
Helsinki University of Technology
Laboratory of Process Control and Automation
tiina.komulainen@hut.fi
The aim of this chapter is to present the relate the common features of
complex systems, self-similarity and self-organization to power (scaling) laws
and to fractal dimension and to show how all these are intertwined together.
First, the basic concepts of self-similarity, self-organization, power laws and
fractals are represented. Then, the power laws and fractals are discussed
more detailed. Finally, examples of the applications of power laws and fractal
dimension are demonstrated.
10.1
Introduction
The power laws and fractal dimensions are just two sides of a coin and
they have a tight relationship joining them the together, as shown in Fig.
10.1. The relationships can be claried with a mathematical discussion.
The general equation for power law is shown in (10.1). It is a mathematical
pattern in which the frequency of an occurrence of a given size is inversely
proportionational to some power n of its size:
y(x) = xn .
(10.1)
109
110
(10.2)
It turns out that the power law can be expressed in linear form using
logarithms:
log(y(x)) = n log(x),
(10.3)
(10.4)
where is a constant factor and s is the scaling exponent, which is independent of x. Looking at (10.2), it is clear that the power law obeys the scaling
relationship.
The data emerging from the combination of self-similarity and self-organization
can not be described by either Normal or exponential distribution.The reason is, that emergence of order in complex systems is fundamentally based
on correlations between dierent levels of scale. The organization of phenomena that belong at each level in the hierarchy rules out a preferred scale
or dimension. The relationships in this type of systems are best described
by power laws and fractal dimension [3].
10.2. Self-Similarity
10.2
111
Self-Similarity
112
etc.). The process of human cognition facilitates scales and similarity. Human mind groups similar objects of the same size into a single level of scale.
This process has been compared with digital image compression, because
it reduces the amount of presented information by a complex structure [3].
Self-similarity in music comes from the coherent nature of the sounds. The
coherencies are agreeing with each other in every scale and dimension which
they are perceived. [5]. The expansion of the Universe from the big bang
and the collapse of a star to a singularity might both tend to self-similarity
in some circumstances [6].
A self-similar program is a program that can mutate itself into a new, more
complex program, that is also self-similar. For example, a self-similar language can be extended with a new language feature, which will be used to
parse or compile the rest of the program. Many languages can be made
self-similar. When language is simple but powerful (Lisp, SmallTalk), selfsimilarity is easier to achieve than when it is complex but powerful (Java,
C++) [7].
10.2.1
Self-organization
Two types of stable systems can be found in the physical universe: the
death state of perfect equilibrium and the innitely fertile condition of selforganized non-equilibrium. Self-organization provides useful models for many
complex features of the natural world, which are characterized by fractal geometries, cooperative behavior, self-similarity of structures, and power law
distributions. Openness to the environment and coherent behavior are necessary conditions for self-organization and the growth of complexity [8].
Because of a common conceptual framework or microphysics, self-organizing
systems are characterized by self-similarity and fractal geometries, in which
similar patterns are repeated with dierent sizes or time scales without changing their essential meaning. Similar geometric patterns are repeated at different sizes and are expressive of a fundamental unity of the system such as
braiding patterns ranging from streambeds to root systems and the human
lung [8].
Systems as diverse as metabolic networks or the world wide web are best
described as networks with complex topology. A common property of many
large networks is that the vertex connectivities follow a scale-free power-law
distribution. This feature is a consequence of two generic mechanisms shared
by many networks: Networks expand continuously by the addition of new vertices, and new vertices attach preferentially to already well connected sites.
10.2. Self-Similarity
113
10.2.2
Power laws
114
The Gutenberg-Richter law, used in seismography, is an empirical observation that earth quakes occur with a power law distribution. The crust of the
Earth, buckling under the stress of plates sliding around, produces earthquakes. Every day at some parts of the earth occur small earth quakes, that
are too weak to detect without instruments. A little more bigger earthquakes,
that rattle dishes are less common and the big earth quakes, that cause mass
destructions happen only once in a while [13].
Power laws are applied to monitor the the acoustic emissions in materials
which are used for bridges etc. Internal defects in materials make popping
sounds, acoustic emissions, under stress. Engineering materials contain tiny
cracks that grow under stress until they grow large enough to cause materials to fail. This can cause the failure of buildings, bridges and other societal
structures. One way to determine the condition of structures is to monitor
the acoustic emissions. This monitoring method is simple and inexpensive
form of non-destructive testing. With the help of the method it is also possible to develop new techniques to make better materials and design structures
less vulnerable to cracking [13].
Matthew-eect on scientic communities is a discovery, that the relationship
between the amount of citations received by members of a scientic community and their publishing size follows power law with exponent 1,270,0 .
The exponent is shown to be constant with time and relatively independent
of the nationality and size of a science system. Also the publishing size and
the size rank of the publishing community in a science system have a powerlaw relationship. Katz has determined the exponent to be 0,440.01. The
exponent should be relatively independent of the nationality and size of the
science system although, according to Katz, the rank of a specic community
in a science system can be quite unpredictable [14].
Some examples of power laws in biology and sociology are the laws of Kleiber,
which relates metabolic rate to body mass in animals; Taylors power law of
population uctuations and Yodas thinning law, which relates density of
stems to plant size [15].
In the following subchapters concentration is placed on Zipfs and Benfords
laws. Typical for the phenomena is that large occasions are fairly rare,
whereas smaller ones are much more frequent, and in between are cascades
of dierent sizes and frequencies. With Zipfs law the structure of the phenomena can be explained and it is possible to do some sort of prediction for
the earth quakes coming in the future.
10.2. Self-Similarity
10.2.3
115
Zipfs law
Zipfs law is named after the Harvard professor George Kingsley Zipf (1902
1950). The law is one of the scaling laws and it denes that the frequency
of occurrence of some event (P ), as a function of the rank (i) is a power-law
function [16,17].
In Zipfs law the quantity under study is inversely proportional to the rank.
The rank is determined by the frequency of occurrence and the exponent a
is close to unity [17].
Zipfs law describes both common and rare events. If an event is number 1
because it is most popular, Zipfs plot describes the common events (e.g., the
use of English words). On the other hand, if an event is number 1 because
it is unusual (biggest, highest, largest, ...), then it describes the rare events
(e.g., city population) [17].
Benoit Mandelbrot has shown that a more general power law is obtained by
adding constant terms to the denominator and power. In general, denominator is the rank plus rst constant c and power is one plus a second constant
d. Zipfs law is then the special case in which the two constants, c and d are
zero [16]:
Pi 1/(i + c)(1+d) .
(10.5)
For example the law for squares can be expressed with d = 1, so the power
will be a = 2 [16]:
Pi 1/(i + c)2 .
(10.6)
116
(10.7)
The modied Zipfs law consists of population coecient 5 million and constants c = 2/5 and d = 1/4.
Pi = 5, 000, 000 1/(i 2/5)(3/4) .
(10.8)
The sizes of the populations given by unmodied Zipfs law dier from the
real sizes approximately 30 % and the sizes by modied Zipfs law only about
10 %. The obvious conclusion is that the modied version of Zipfs law is
more accurate than the original one. Although the predictions of the original
one are quite good too [16].
Other famous examples of Zipfs law are: The frequency of English words
and the income or revenue of a company. The income of a company as a
function of the rank should also be called the Paretos law because Pareto
observed this at the end of the last century [17,19].
The english word example is illustrated by counting the top 50 words in
423 TIME magazine articles (total 245,412 occurrences of words), with the
as the number one (appearing 15861 times), of as number two (appearing
7239 times), to as the number three (6331 times), etc. When the number of
occurrences is plotted as the function of the rank (1, 2, 3, etc.), the functional
form is a power law function with exponent close to 1 [17].
Shiode and Batty have proposed a hypothesis of the applicability of Zipfs law
for the growth of WWW domains. Most of the mature domains follow closely
Zipfs law but the new domains, mostly in developing countries, do not. The
writers argue that as the WWW develops, all domains will ultimately follow
the same power laws as these technologies mature and adoption becomes
more uniform. The basis for this hypothesis was the observation that the
structure in the cross-sectional data, which the writers had collected, was
consistent with a system that was rapidly changing and had not yet reached
its steady state [18].
10.2. Self-Similarity
117
Table 10.1: Zipfs Law applied to population of cities in the USA. Unmodied
Zipfs law uses population size of 10 million and constants 0. Modied Zipfs
law uses population size of 5 million and constants -2/5 and -1/4 [16].
Rank (n) City
1
7
13
19
25
31
37
49
61
73
85
97
NewYork
Detroit
Baltimore
Washington, D.C.
New Orleans
Kansas City, Mo.
Virginia Beach, Va
Toledo
Arlington Texas
Baton Rouge, La.
Hialeah, Fla.
Bakerseld, Calif.
10.2.4
Population
(1990)
7.322.564
1.027.974
736.014
606.900
496.938
434.829
393.089
332.943
261.721
219.531
188.008
174.820
Unmodied
Zipfs law
10.000.000
1.428.571
769.231
526.316
400.000
322.581
270.270
204.082
163.934
136.986
117.647
103.093
Modied
Zipfs Law
7334.265
1214.261
747.639
558.258
452.656
384.308
336.015
271.639
230.205
201.033
179.243
162.270
Benfords Law
Benfords law (known also as the rst digit law, rst digit phenomenon or
leading digit phenomenon) states that the distribution of the rst digit is
not uniform. This law applies, if the numbers under investigation are not
entirely random, but somehow socially or naturally related. Generally the
law applies to data that is not dimensionless, meaning that the probability
distribution of the data is invariant under a change of scale, and data, that
is selected out of a variety of dierent sources. Benfords law does not apply
to uniform distributions, like lottery numbers [11,20,21].
Benfords law results that the probability of a number to be the rst digit
in tables of statistics , listings, etc., is biggest for one and smallest for nine
[11,20].
Many physical laws and human made systems require cutos to number
series, for example the street addresses begin from 1 and end up to some
cuto value. The cutos impose these systems to obey Benfords law, which
can be presented for a digit D (1, 2,. . . ,9) by the logarithmic distribution:
PD = log10 (1 + 1/D)
(10.9)
118
The base number of logarithm can also be other than 10 [21]. According to
this equation the probabilities for digits 1,. . . ,9 lie between 0.301,. . . ,0.045,
as shown in Fig. 10.2. The probability for 1 is over 6 times greater than for
9 [20].
The examples of Benfords law are numerous, including addresses, the area
and basin of rivers, population, constants, pressure, molecular and atomic
weight, the half-lives of radioactive atoms, cost data, powers and square root
of whole numbers, death rate, budget data of corporations, income tax and
even computer bugs [11,20,21].
Benfords law is a powerful and relatively simple tool for pointing suspicion
at frauds, embezzlers, tax evaders and sloppy accountants. The income tax
agencies of several nations and several states, including California, are using
detection software based on Benfords law, as are a score of large companies
and accounting businesses [21].
The social and natural impact aects all the listed examples. Budget data
is aected by the corporation size, particular industry a company belongs
to, the quality of the management, the state of the market, etc. The size
of a river basin depends on a depth and breadth of the river. Most of the
dependencies can be approximated by simple formulas: Linear, power or
exponential, oscillating, leading to saturation.
10.2. Self-Similarity
10.2.5
119
Fractal dimension
Fractals are characterised by three consepts: Self-similarity, response of measure to scale, and the recursive subdivision of space. Fractal dimension can
be measured by many dierent types of methods. Similar to all these method
is, that they all rely heavily on the power law when plotted to logarithmic
scale, which is the property relating fractal dimension to power laws [22,23].
One denition of fractal dimension D is the following equation:
D = log10 N/ log10 (1/R),
(10.10)
(10.11)
so that
N = RD .
(10.12)
(10.13)
(10.14)
120
1
4
1/3
1.33333
0.60206
0.47712
1.26185
2
16
1/9
1.77777
1.20412
0.95424
1.26185
3
64
1/27
2.37037
1.80618
1.43136
1.26185
4
256
1/81
3.16049
2.40824
1.90848
1.26185
5
1024
1/243
4.21399
3.01030
2.38561
1.26185
The formulas above indicate that N and R are related through a power law.
In general, a power law is a nonlinear relationship, which can be written in
the form N = a(1/R)D , where D is normally a non-integer constant and a is
a numerical constant which in the case of the Koch curve is 1.26.
Another way of dening the fractal dimension is box counting. In box counting the fractal is put on a grid, which is made of identical squares having size
of side h. Then the amount of non-empty squares, k, is counted. The magnication of the method equals to 1/h and the fractal dimension is dened
by equation: [4]
D = log10 (k)/ log10 (1/h).
(10.15)
10.3. References
10.3
121
References
1. Focardi, S.M.: Fat tails, scaling and stable laws: A critical look at
modeling extremal events in economic and nancial phenomena.
http://www.theintertekgroup.com/scaling.pdf, [18.3.2003].
2. J. McKim Malville: Power Law Distributions.
http://www.colorado.edu/Conferences/pilgrimage/papers
/Malville-Pap/kimpap1.html, [2.3.2003].
3. Salingaros, N.A. and West, B.J.: A universal rule for the distribution
of sizes.
http://www.math.utsa.edu/sphere/salingar/Universal.html, [15.3.2003].
4. Judd, C.: Fractals Self-Similarity.
http://www.bath.ac.uk/ma0cmj/FractalContents.html, [16.3.2003].
5. Yadegari, S.: Self-similarity.
http://www-crca.ucsd.edu/syadegar/MasterThesis/node25.html,
[16.3.2003].
6. Carr, B.J. and Coley, A.A.: Self-similarity in general relativity.
http://users.math.uni-potsdam.de/oeitner/QUELLEN
/ZUMCHAOS/selfsim1.htm, [16.3.2003].
7. Lebedev, A.: What is Self-Similar.
http://www.self-similar.com/what_is_self_similar.html, [16.3.2003].
8. J. McKim Malville: Power Law Distributions.
http://www.colorado.edu/Conferences/pilgrimage/papers
/Malville-Pap/kimpap1.html, [2.3.2003].
9. Barabasi, L.: Emergence of Scaling in Complex Networks.
http://discuss.santafe.edu/dynamics/stories/storyReader$10,
[18.3.2003].
10. Lucas, C.: Perturbation and Transients The Edge of Chaos.
http://www.calresco.org/perturb.htm, [17.3.2003].
11. Bogomolny, A.: Benfords Law and Zipf s Law.
http://www.cut-the-knot.org/do_you_know/zipfLaw.shtml, [15.3.2003].
12. Goldstein, J.: Edgeware Glossary.
http://www.plexusinstitute.com/edgeware/archive/think
/main_gloss.html, [15.3.2003].
122
Session 11
Turings Lure, Gdels Curse
Abouda Abdulla
Helsinki University of Technology
Communication Engineering Laboratory
abouda@cc.hut.fi
11.1
Computability theory
124
11.2
Gdelian undecidability
All frame works that are powerful enough are either incomplete or inconsistent. This is what is known as Gdelian undecidability principle. This
principle has two interpretations, in mathematics and in physics. In mathematics it says that there does not exist any reasonable nite formal system
from which all mathematical truth is derivable. And in physics it says that
with respect to predication, there are some systems that cannot be completely predicted like computation in Turing machine or predictbility of
Life pattern in the game of Life. In this article we consider the physical
interpretation [2].
It has been shown that any formalism that is powerful enough suers from
this Gdelian undecidability problem. A prototype of such a powerful system
is Turing machine.
11.3
Turing Machine
A Turing machine is a very simple machine that has all the power that any
digital computer has. A Turing machine is a particularly simple kind of computer, one whose operations are limited to reading and writing symbols on a
tape, or moving along the tape to the left or right. This tape is divided into
squares, any square of which may contain a symbol from a nite alphabet,
with the restriction that there can be only nitely many non-blank squares
on the tape. At any time, the Turing machine has a read/write head positioned at some square on the tape. Figure 11.1 shows how the tape looks like
with a series of As and Bs written on the tape and with read/write head
located on the rightmost of these symbols.
..
..
.
125
Furthermore, at any time, the Turing machine is in any one of a nite number
of internal states. The Turing machine is specied by a set of instructions of
the following form:
(current state, current symbol, new state, new symbol, left/right)
The instruction means that if the Turing machine is now in current state,
and the symbol under the read/write head is current symbol, then the Turing
machine changes its internal state to a new state, replaces the symbol on the
tape at its current position by a new symbol, and moves the read/write head
one square in the given direction left or right. It may happen that the Turing
machine is in a situation for which it has no instruction and that makes the
Turing machine to halt. Dening instructions of Turing machine can be
though of as programming the Turing machine.
There are several conventions commonly used. The convention that numbers
are represented in unary notation is adopted here. It means that the nonnegative integer number n is represented by using of successive 1s. Furthermore, if a function f (n1 ,n2 ,...,nk ) to be computed, we assume that initially
the tape consists of n1 ,n2 ,...,nk , with each sequence of 1s separated from
the previous one by a single blank, and with tape head initially located at
the right most bit of the rst input argument, and the state of the Turing
machine in some initial specied value. We say that the Turing machine has
computed m = f (n1 ,n2 ,...,nk ) if, when the machine halts, the tape consists
of the nal result and the read/write head at the right most bit of the result.
..
.
126
..
11.4
Cellular automata and other systems with simple underlying rules can produce complicated behaviors. To analyze the behavior of these systems, our
experience from traditional science might suggest that standard mathematical analysis should provide the appropriate basis for any framework. This
kind of analysis tends to be useful only when the overall behavior of the system is fairly simple. So what can one do when the over all behavior is more
complex? The main purpose of Wolframs book A New Kind of Science
is to develop a new kind of science that allows progress to be made in such
cases. The single most important idea that underlines this new science is the
notation of computation. So far we have been thinking of cellular automata
and other systems as simple computer programs. Now we will think of these
systems in terms of the computations they can perform [5].
In a typical case, the initial conditions for a system like cellular automaton or
Turing machine can be viewed as corresponding to the input to a computation, while the state of the system after some number of steps corresponding
to the output. The key idea is to think in purely abstract terms about the
computation that is performed. This abstract is useful due to two reasons,
the rst, because it allows us to discuss in a unied way systems that have
completely dierent underlying rules, and the second reason, because it becomes possible to imagine formulating principles that apply to a very wide
variety of dierent systems [5].
11.4.1
A Visual Turing Machine simulation program is available from the web page
http://www.cheran.ro/vturig
127
128
11.5
129
11.6
Game of Life
One form of cellular automata is Game of Life, which was invented by Cambridge mathematician John Conway. Game of Life is no-player game. Life
is played on an innite squared board. The cells in this game represent the
population. At any time some of the cells will be live and others will be
dead. The initial structure of live cells can be set as one wants but when
time starts to go on the cells birth or death is controlled by rules [6]. The
rules of this game are as follows2 :
Birth: A cell that is dead at time t becomes live at time t + 1 only if
exactly three of its eight neighbors were live at time t.
Death by overcrowding: A cell that is live at time t and has four or
more of its eight neighbors live at time t will be dead by time t + 1.
2
The Game of Life simulation program is available, for example, at the web page
http://www.bitstorm.org
130
131
tips of Blinkers die of exposure but the births on each side reform the line
in a perpendicular direction. The conguration will therefore oscillate with
period two forever. The nal pair of conguration known as Trac Lights.
Still Life
There are some common forms of still lives; some of these are shown in
Fig. 11.9, with their traditional names. The simple cases are usually loops in
which each live cell has two or three neighbors according to local curvature,
but the exact shape of the loop is important for eective birth control.
132
Life Cycle
There are some congurations whose Life history repeats itself after some
time. This period of time known as Life Cycle. The Blinker is the simplest
example of these congurations, where the Blinker repeats itself with period
larger than 1. Some congurations with their own Life cycles are shown in
Figs. 11.10 and 11.11.
Figure 11.11: Life cycle with period three: (a) Two eaters Gnash at
each other. (b) The Cambridge Pulsar
133
Figure 11.12: The Glider moves one square diagonally each four generations
134
11.6.1
Many computers have been programmed to play the game of Life. In this
section we will see how to dene Life patterns that can imitate computers. Computers are made from pieces of wire along pulses of electricity go.
To mimic these by certain lines in the plane along which Glider travel (see
Fig. 11.14.
135
The last crashing way has special interest, because the vanishing reactions
turn out to be surprisingly useful.
The Eater
Other phenomena that happens when two Gliders meet, an Eater will be
constructed. This is shown in Fig. 11.17.
136
Figure 11.16: A Glider Gun and vanishing reaction make a NOT gate
11.7. Conclusion
137
Figure 11.18: In (b), Glider dives into Pond and comes up with Ship.
In (c), Glider crashes into Ship and makes part of Glider Gun
11.7
Conclusion
The conclusion that can be drawn is that complex systems like cellular automata or Turing machine with very simple underlying rules are universal
systems which can emulate other complicated systems. When we come to
analyze such systems, we can see clearly that we do not have a framework
to deal with these systems. New kind of science to analyze these powerful
systems is needed.
Bibliography
1. http://www.cheran.ro/vturig
2. http://tph.tuwien.ac.at/ svozil/publ/1996-casti.htm
3. http://www.turing.org.uk/turing/
4. http://www.bitstorm.org
138
Figure 11.20: (a) AND gate, (b) OR gate, and (c) NOT gate
5. Wolfram, S.: A New kind of science.Wolfram Media, Champaign, Illinois, 2002.
6. Elwyn Berlekamp, Conway, J., and Guy, R.: Winning Ways, Vol. 2,
Chapter 25.
7. http://www.ams.org/new-in-math/cover/turing.html
8. http://www-csli.stanford.edu/hp/turing1.html
9. http://ddi.cs.uni-potsdam.de
10. Hytyniemi, H.: On universality and undecidability in dynamic systems. Helsinki University of Technology, Report 133, December 2002.
Session 12
Hierarchical Systems Theory
Jari Htnen
Systems Engineering Laboratory, University of Oulu
P.O. BOX 4300, FIN-90014 University of Oulu, Finland
12.1
Introduction
The design and analysis of large and complex system requires frequently that
the system is divided into smaller units, called sub-systems. The structure of
overall system resulting from the interconnections of sub-systems can be very
complex. One of the most common structures, is the hierarchical structure,
i.e. the lay-out of the structure is vertical. Consequently in this report only
hierarchical systems are considered, and a special emphasis is put on twolevel hierarchical systems. A diagram of a standard two-level hierarchical
system is shown in Fig. 12.1, where, as expected, two levels can be found,
namely the lower level and the upper level. The lower level consists of the
process level, where the process level has been divided into N sub-systems.
The sub-systems are connected to each other because there is either material
or information ows between these sub-systems. Each sub-system has its own
decision unit, which tries to control the behaviour of the sub-system so that
the objectives of this particular sub-system would be met. The decision unit
can also use feedback information from the sub-system to improve its control
policy. However, quite often the objectives of the sub-systems are conicting,
resulting in a poor overall performance. Hence an upper-level decision unit
139
140
Upper level
Coordinator
Lower level
Lower-level
decision making
Process
level
141
(i.e. only the faulty module has to replaced and its connections reestablished). The coordinator, however, is the weak point, because if
it stops working, the overall system cannot function anymore.
3) Even evolution seems to favour two-level hierachical systems. For example in a human body the brain can be considered as being the coordinator, whereas the rest of the body forms the sub-system level.
4) In the evolution of organisations two-level hierachical systems play a
major role. Even pre-historic tribes had a tribe leader, whom was
responsible for coordinating the actions of individual tribe members in
order to improve the overall well-being of the tribe.
In the following material it is shown how two-level hierarchical systems can
be analysed mathematically. Furthermore, it is shown how optimisation techniques can be used to coordinate the running of a two-level hierarchical systems. The material is a rather mathematical, but in order to understand it,
only a fairly modest background is needed in constrained optimisation theory. This report is based on [1], which is a nice introduction into the theory
of hierarchical systems.
12.2
12.2.1
Process level
Sub-system models
The process level consists of N inter connected sub-systems. For each subsystem i there exists a set of admissible inputs Ii , the set of admissible
outputs Oi and a sub-system model fi : Ii Oi . Because the sub-systems
are interconnected to each other, the set of inputs Ii is divided into the set
of free inputs Mi and the set of interconnected inputs Xi and consequently
Ii Mi Xi
(12.1)
In a similar fashion the the outputs are divided into the set of free ouput
variables Yi and set of constrained output variables Zi , i.e. the outputs
zi Zi are fed as inputs into other sub-systems, see Fig. 12.2. With this
division the set of admissible outputs Oi can be written as
O i Y i Zi
(12.2)
142
(12.3)
N
Cij zj ,
i = 1, 2, . . . , N
(12.4)
i=1
where xi Rni , zj Rmj and Cij is a real-valued ni nj -matrix. Furthermore, the element ckl
ij of Cij is either zero or one depending on whether or
not the lth component of zj is connected to the kth element of input vector
xi . Dene now
X := X1 X2 . . . XN M := M1 M2 . . . MN
Y := Y1 Y2 . . . YN
Z := Z1 Z2 . . . ZN
(12.5)
143
and
x = (x1 , x2 , . . . , xN ) m = (m1 , m2 , . . . , mN )
y = (y1 , y2, . . . , yN ) z = (z1 , z2 , . . . , zN )
(12.6)
(12.7)
where
(yi , zi ) = fi (mi , xi ),
i = 1, 2, . . . , N
(12.8)
(12.9)
and
xi =
N
Cij zj
(12.10)
i=1
(y, z) = f (m, x)
x = C(z)
(12.11)
(12.12)
(12.13)
144
y
z
x
(12.14)
12.2.2
As was explained earlier, for each sub-system i there exists a decision unit,
and the objective of the decision unit is to control the sub-system according
to its objectives by manipulating the free input variables mi . The objectives
of the lower-level decision units can vary a lot, but in this report each subsystem i is associated with a cost function gi , which is a function of the
145
(12.15)
D(Gi ) = D(fi )
(12.16)
where
(12.17)
with the notation fi (mi , xi ) = (fi1 (mi , xi ), fi2 (mi , xi )). In a similar fashion
an overall cost function g is dened where
g :M X Y Z R
(12.18)
The overall process model can then used to eliminate the output variables
resulting in equivalent representation of (12.18)
G : M D(G) R,
D(G) = D(F )
G(m) = g (m, K(m), P (m), S(m))
(12.19)
From now on it is assumed that the individual cost functions Gi and the
overall cost funtion G are related to each other with the equation
G(m) =
N
Gi (mi , Ki (m))
(12.20)
i=1
12.3
12.3.1
Upper level
Upper-level decision making
146
N
mD(F )
Gi (mi , xi )
(12.21)
i=1
(12.22)
D(fi = D(fi )
(12.23)
(12.24)
12.3.2
The sub-system decision unit i has optimise with a given the cost
i (mi , xi , ), i.e. the decision unit has to nd (mi , xi ) Mi Xi so
function G
i (mi , xi , ).
that resulting pair (mi , xi ) minimises the cost function G
Remark 3 The important point here is that now the optimisation problem
is now an unconstrained optimisation problem, and the fact that xi is determined by the behaviour of other sub-systems is not taken into account.
147
The optimal solution of the problem (if it exists in the rst place) is called
as the -optimal solution (m(), x()), where
m() = (m1 (), m2 (), . . . , mN ())
(12.25)
(12.26)
and
mD(F )
(12.27)
The question whether or not there exits a so that (12.27) holds depends
i and the set of coordistrongly on how the modied system variables fi , G
nation variables are chosen. Consider now the case where the modication
is done so that for an arbitrary , m D(F ) it holds that
fi (mi , Ki (m), ) = fi (mi , Ki (m))
i (mi , Ki (m), ) = Gi (mi , Ki (m))
G
(12.28)
(12.29)
G (m( )) = N
i=1 Gi (mi , Ki (m( ))
N
i (mi ( ), Ki (mi ( ), )
= i=1 G
N
i (mi , xi , )
= i=1 minmi ,xi D(fi ) G
N
i (mi , Ki (m), )
i=1 minmD(F ) G
N
minmD(F ) i=1 Gi (mi , Ki (m), )
N
i=1 Gi (mi , Ki (m ), )
N
= i Gi (mi , Ki (m ) = G(m )
(12.30)
148
12.4
Coordination
mD(F )
(12.31)
ii) There exists so that there exists a -optimal pair (m( ), x( )),
i.e.
i (mi ( ), xi ( ), )
i (mi , xi , ) = G
min(mi ,xi)D(fi ) G
i = 1, 2, . . . , N
(12.32)
iii) The objectives of the sub-system decision units and upper-level decision
unit are in harmony, i.e. m( ) D(F ) and
G(m ) = G (m( ))
(12.33)
12.4. Coordination
149
(12.34)
The coordination algorithm related to the coordination strategy can be described in the following way:
1) Select a suitable initial guess for
2) The lower-level decision units solve their own optimisation problems
resulting in the -optimal pair (m(), x())
3) If is not optimal (whether or not this can be tested in practise depends highly on the modication technique, see next section for further
details), the coordinator selects a new coordination variable using
the coordination strategy
(, m(), x())
(12.35)
150
12.5
(12.36)
N
i (xi , zi , ) = 0
(12.37)
N
Cij zj ,
i = 1, 2, . . . , N
(12.38)
i=1
is met. Hence in the balance principle only the cost functions are modied
whereas the sub-system model is equal to the original sub-system model.
Note that modication that satises (12.37) and (12.38) is called a zero-sum
modication. The dening property of a zero-sum modication is that the
eect of the modication disappears from the overall cost function N
i=1 Gi
when the system is in balance, i.e. the interconnection equations hold. In
this case the overall cost can be calculated as the sum of the individual values
of the cost functions in the following way:
G(m) =
N
i=1
i (mi , xi , ) =
G
N
i=1
Gi (mi , xi )
(12.39)
151
(12.40)
(mi ,xi )
min
i (mi , xi , )
G
(12.41)
where
D(i) = { |i () exists}
(12.42)
2
i , xi ), fi (mi , xi )
N
i=1
(12.43)
If the original overall optimisation problem has a solution m and there exists
a D(i ) so that the lower-level -optimal solution (m( ), x( ))
satises
x( ) =
N
i = 1, 2, . . . , N
(12.44)
i=1
(12.45)
( ) = N
i ( ) = N
i=1
i=1 Gi (mi ( ), xi ( ), )
(12.46)
() = min(m,x)D(f ) N
i=1 Gi (mi , xi , )
N
i (mi , xi , )
min(m,x)D(f ),x=K(m) i=1 G
N
= min(m,x)D(f ),x=K(m) i=1 Gi (mi , xi ) = G(m )
G (m( ))
(12.47)
152
and was shown previously in Section 12.3, G(m ) = G (m( )). Consequently in tha balancing principle the coordinator variable is taken to be
the solution of the maximisation problem
max ()
(12.48)
D()
In practise it can be dicult (impractical) to solve analytically this maximisation problem, and numerical methods have to be used instead. Typically
the gradient is available (see next section), and for example a steepestdescent algorithm can be used to solve iteratively the maximisation problem.
12.6
N
Gi (mi , xi ) +
i=1
N
j=1
i , xi
N
(12.49)
i=1
where < , > is the inner product in Xi (for simplicity it is assumed here
that Xi for i = 1, 2, . . . , N is always the same space and that it is reexive,
i.e. the dual space of Xi is Xi . Reexive spaces are for example the Euclidian
space RN (discrete-time case with nite time-axis) and L2 [0, T ] (continuoustime case with nite time-axis), which are one of the most commonly used
spaces in control theory. By changing the summation order it can be shown
that the Langrange function can be written equivalently as
L(m, x, ) = N
i=1 Li (m, x, )
2
Li (m, x, ) = Gi (m, xi ) + i , xi N
j=1 j , Cij fi (mi , xi )
(12.50)
In summary L(m, x, ) has been divided into a sum where each term Li (mi , xi , )
depends only on the variables related to sub-system i and the Langrange multiplier . This modication is clearly a zero-sum modication. As was shown
in the previous section, in the balancing technique the sub-system decision
problem is to solve with a given (which is now the Langrange-multpier)
the optimisation problem
i () :=
min
mi ,xi D(fi )
(12.51)
153
D()
(12.52)
N
(12.53)
i=1
As was mentioned previously, the upper level decision problem does not necessarily have a nice closed-form solution. Consequently in order to use numerical optimisation methods, the gradient i is needed. However, in the
Langrange technique this is just (can you show this?)
(o ) = [1 , 1 , . . . , N ]T :=
(12.54)
where
i = xi (o ) Cij fj2 (m(o ), xj (o))
(12.55)
(12.56)
(12.57)
and go to step 2.
Remark 4 If sub-system decision problems are linear programmes (i.e. they
are of the form cT x then it can be shown that the Langrange multipier can be
understood to be the price that a sub-system has to pay to the other sub-system
for the transfer of material (information). In other words coordination is
equivalent to an optimal pricing policy between departments, see [3]!
154
The toy example in the following section will illustrate the Langrange approach. Note that even this example considers only static sub-system models, this technique can be applied on dynamical systems models without any
major complications, see [2] and [1] for further details.
12.7
A toy example
y1 = 2m1 + u1 = P1 (m1 , u1 )
y2 = 2m2 + u2 = P2 (m2 , u2 )
(12.58)
(12.59)
(12.61)
10 4
4 2
m1
u1
4 + 2
2 1
10 4
4 2
m2
u2
8 + 1
4 2
(12.62)
() =
u1 m2
u2 m1
(12.63)
and the update law for the coordination variable (Langrange multplier) becomes
1 (k + 1)
2 (k + 1)
1 (k)
2 (k)
+k
u1 m2
u2 m1
(12.64)
12.8. Conclusions
155
Convergence of m1 and m2
25
l2norm of mmref
20
15
10
0
0
20
40
60
Iteration round k
80
100
12.8
Conclusions
156
J(k)
0.62
0.6
0.58
0.56
0.54
0.52
0.5
0
20
40
60
80
100
Iteration round k
Figure 12.5: The value of the modied cost function
the selshness of the sub-system decision making, an upper-level decision
making system has to be introduced. The objective of the upper-level decision making unit is to coordinate the decision making of each sub-system, so
that an overall harmony is achieved. A special emphasis was put on the so
called balancing technique, because it can be implemented numerically with
the well-known Langrange technique. Note that in this report no rigorous
convergence theory was presented for the iterative balancing technique. Convergence, however, can be proved in some special cases, see [2] for further
details.
The required theoretical background for understanding the theory of twolevel hierarchical systems in the most abstract setting is, unfortunately, reasonably sophisticated. Consequently the theory can be quite hard for an
engineer to digest, and therefore it has not found its way to the mainstream
material taught at graduate level in control engineering courses. On the
other hand it oers an interesting option for more mathematically oriented
engineers to do high-quality control for complex systems without resorting
to ad-hoc methods (i.e. fuzzy control, agents etc.).
Bibliography
[1] L. Hakkala, Hierarkisten jrjestelmien optimointi: koordinointiperiaattet, Report C 23, Systems Theory Laboratory, Helsinki University of
Technology, 1976
[2] Mesarovic, Macko, Takahara, Theory of Hierarchical Multilevel Systems,
Academic Press, New York, 1970
[3] Lasdon, Optimisation Theory for Large Systems, The Macmillan Publishing Company, New York, 1970
157
158
BIBLIOGRAPHY
Session 13
Qualitative Approaches
Petteri Kangas
petteri.kangas@hut.fi
As a tool to model complex systems, system dynamics is an alternative. It
can be used for wide range of applications. Mostly qualitative results can be
achieved by using this method. The most valuable benet is to understand the
inner functionality of the observed system. System dynamics has been used
for real applications during 50 years.
13.1
Introduction
In the previous chapters, complex systems have been introduced. Their structures have been presented, their behaviour has been introduced and dierent
ways to model complex systems have been shown. The aim of this chapter
is to show the basic ideas of system dynamics and system thinking. These
are tools for modelling and understanding dierent kind of complex systems,
from technology to biology and from economy to human behaviour. System
dynamics and thinking can be seen as a general tool for modelling dierent
kind of complex systems.
First, this chapter will give short historical perspective of system dynamics
during last 50 years. It tries to connect system dynamics to other elds of
research. A well-known example, the Beer distribution game is presented.
Few other applications are shown. At the end, the key concepts of system
dynamics are presented with a small example model. Finally, important
publications, software and research institutes are listed.
159
160
As an example of a dynamic system, Naills natural gas model [1] can be seen
in the gure 13.1.
13.2
The father of System Dynamics is Jay Forrester. He graduated from Engineering Collage at the University of Nebraska. After that, he got a job from
Massachusetts Institute of Technology as research assistant under supervision
of Gordon Brown, a pioneer in feedback control systems. At Servomechanism Laboratory, they built an experimental control for radar, which was
used during World War II [1], [2].
After the war, Jay Forrester continued to work at MIT. He started to build a
ight simulator for U.S. Navy. Soon they noticed that analog computers were
not able to handle problems this large. Development of digital computers
started. In 1947, the Digital Computer Laboratory of MIT was founded.
It was placed under the direction of Jay Forrester [1]. They designed the
Whirlwind digital computer, which later on was developed to SAGE (Semiautomatic Ground Environment) air defence system for North America [2].
At the middle of 50s Jay Forrester felt that pioneering days of digital computers were over. He was also running projects worth of several million dollars
and noticed that social systems are much harder to understand and control than are physical systems [1]. A business school was found in MIT and
Jay Forrester joined that school in 1956. He started to think management
problems with engineering background [2].
The Jay Forresters rst dynamic system was a simple pencil and paper
simulation of employment instability at General Electrics Kentucky plants.
Forrester was able to show that the instability of employment was due to the
internal structure of the rm, not external forces such as the business cycle
[2].
Next step was a compiler SIMPLE (Simulation of Industrial Management
with Lots of Equations) developed by Richard Bennett in 1958. Later, 1959,
SIMPLE and other system dynamics compiler was developed to DYNAMO
compiler by Jack Pugh [2]. The system dynamics became an industrial standard for over 30 years. Forrester published his classic book, Industrial Dynamics, in 1961 [1].
Urban dynamics by Jay Forrester was the book, the brought system dynamics
to even wider knowledge. It was the very rst non-corporate application of
161
162
system dynamics [1]. This model was developed with John Collins, the former
mayor of Boston and his experts in 1968. The results of this study were not
generally accepted, at rst. They showed that many policies widely used
for urban problems were in fact ineective or even caused worse problems.
Solutions, which seemed to be wrong at rst glance, were, at the other hand,
very eective [2].
The next turning point of system dynamics was year 1970. Jay Forrester was
invited by the Club of Rome, to a meeting. He was asked if system dynamics could be used to describe the "predicament of mankind". Jay Forrester
said YES and developed the rst model of worlds socioeconomic system,
called WORLD1. Later that model was rened and the WORLD2 model
was published in the Forresters book titled World Dynamics [2]. This model
and book showed that worlds socioeconomic system would collapse sometime during 21st century if no actions were taken to lessen the demands for
earths carrying capacity. Forresters formed Ph.D. student Dennis Meadows
developed the WORLD2 model further and the Club of Rome published a
book titled the Limits of Growth on based on the WORLD3 model [1]. Although, the Limits of Growth is widely criticised, it is an excellent example
of usability of system dynamics.
Last years Jay Forrester has been developing National model for U.S. In
addition, he has introduced system dynamics to education from kindergartens
to high schools [1].
At 1990 Peter Senge, a director of the Systems Thinking and Organizational
Learning Programs at MITs Sloan School of Management published a best
seller book the Fifth Discipline The Art & Practice of The Learning Organization. It re-introduced the system thinking to managers worldwide. The
relationship of system thinking and system dynamics is following: They look
the same kind of system from the same kind of perspective. They construct
the same causal loop diagrams but system thinking rarely takes additional
steps for constructing or testing computational simulations [3]. In Finland
Markku Perkola is using ideas of system thinking [4].
At the beginning of the new millennium, 2000, Professor John Sterman for
MITs Sloan School of Management, published a book Business Dynamics
System Thinking and Modelling for a Complex World. The roots of this
book relates back to the publications of Jay Forrester and the Fifth Discipline
by Peter Senge. The ideas of system dynamics and system thinking are
combined into this book. There are wide range of examples for using system
dynamics and thinking for dierent kind of applications, mostly economical.
In addition, the wide range of graphically attractive computer programs for
163
system dynamics modelling has brought the ideas of Jay Forrester back again.
System dynamics is seen as an option for modelling the complex world.
System dynamics has almost 50 years history behind it. It is a real tool used
for solving complex system problems, not a piece of ironic new science.
13.3
13.3.1
164
soon realise that even the rules of this game are very simple, the whole system
is complex. Overcompensations and uctuations appear soon.
The supply chain of the beer distribution game is shown in the gure 13.2
[7].
165
Figure 13.3: Simple, single-stage model, in the Beer distribution game [7].
right cause and eect [2]. Causal laws of distributors inventory can be seen
in the gure 13.4.
166
13.3.2
13.4. Applications
167
Figure 13.6: Multi-stage response to changes in retailer level after implementing Vendor Managed Inventory (VMI). Picture from the Beer distribution
game [7].
and system vision, and have capacity and commitment to continually learn.
They just do not execute a master plan put forth by the grand strategist
at top of the hierarchy [6]. Peter Senges ideas are somehow contradictory
to the hierarchical structure. They are similar to the ideas of networked
structure. Hierarchical and network structures are presented in the previous
chapters as an examples of complex systems.
13.4
Applications
The supply chain game, the Beer distribution game, is not the only applications of system dynamics and system thinking. Few other applications were
already presented in the beginning of this chapter. As earlier mentioned the
system dynamics is a real tool used for solve real life problem. Here are some
additional applications presented.
World models showed that the nite nature resources would be overused.
The result is overshoot and collapse of worlds socioeconomic system. Models
WORLD2 and WORLD3 were developed for the Club of Rome [1].
The life cycle theory of geologist M. King Hubbert is based on the idea that
there is a limited amount of certain resource, e.g. oil, available. If consumption is exponential, the limited resources will lead to peak in consumption
after growth and then long decline in consumption. Later Naill conrmed
168
this theory in his Masters thesis. The result was a three-year study of U.S.
energy transition problem. This problem is currently solved by importing
energy resources, which is cheaper than developing new domestic energy
supplies [1].
Naill also developed a model of natural gas resources. This model was aggregate. However, the model was rened and further developed to COAL1,
COAL2 and FOSSIL1. Name was FOSSIL1, since it looked at the transition
of an economy that is powered by fossil fuels to one that is powered by alternative energy sources. FOSSIL1 was later developed to FOSSIL2 and it was
used by the U.S. Department of Energy to analyse following things from the
late 1970s to the early 1990s [1]:
The net eect of supply side initiatives (including price deregulation)
on US oil imports.
The US vulnerability to oil supply disruptions due to political unrest in
the Middle East or the doubling of oil prices.
Policies aimed at stimulating US synfuel production.
The eects of tax initiatives (carbon, BTU, gasoline, oil import fees)
on the US energy system.
The eects of the Cooper-Synar CO2 Osets Bill on the US energy
system.
These days the FOSSIL2 model is rened to IDEAS model and it is still
actively used.
John Sterman, as a PhD student, noticed that the models above, COALFOSSIL-IDEAS-models, are isolated models from the rest of the economy.
Sterman showed that there are interactions between the energy sector and
the rest of economy. He built signicant energy-economy interactions in his
dissertation [1].
Fiddemanns presented Economy-climate interactions in his dissertation 1997.
The developed FREE (Feedback-Rich Energy Economy) model was the rst
one dynamic model of energy-economy interactions [1].
Applications presented above are shown in the gure 13.7.
These were few examples of the real life problems, which has been modelled by using system dynamics. As earlier mentioned it is a good tool for
understanding large complex systems.
169
13.5
The basic concepts of system dynamics are presented. The stocks and ows
are discussed. Causal loops are presented. In addition, a simple model of
Finnish car population is build. A ight simulator is built as well. The model
and ight simulator are built using a demo version of Powersim simulator.
This model is by no means complete or veried. It is just a humble example.
13.5.1
At the rst stock and ows are presented. We build an aging chain, which
describes the aging of Finnish cars /[5]/. Finnish cars are divided into three
dierent groups, 0-5 years old cars, 5-10 years old cars and over 10 years
old cars. Inows to model are brand new cars and old cars from Germany.
Outows are very old cars sold to Russia and wrecked cars. Stock and ows
are shown in the gure 13.8.
Correspondences between stock and ows and tanks and valves can be easily
seen.
13.5.2
Causal loops
170
171
old cars. The only inow is newly purchased cars. After 5 years these cars
are moved to the next storage, 5-10 years old cars. To this storage, there is
another inow, old cars from Germany. The originally new cars are moved
from the old car storage to the very old car storage after 5 years. The old
cars from Germany are staying only 2 years in the old car storage (Old cars
from Germany are 8 years old).
The nal storage, very old cars, has all the cars, which are over 10 years old.
Outow from the very old cars is wrecked cars. Very old cars are not sold to
Russia in this model.
Delays are added using causal loops to the gure 13.8 and the result can be
seen in the gure 13.9.
172
13.5.3
Even the system dynamics is usually considered a qualitative tool; the numbers have to be entered to the model if simulations are wanted to be run.
Next, we have to dene the ratios, which control the inows, outows and
ows between the dierent age groups.
New_cars0 = 250000
N ew _cars
= P urchace_ratio Aging_1
t
(13.1)
Old_cars0 = 250000
Old_cars
=
t
Old_cars_f rom_Germany Aging_2 + Aging_1
(13.2)
V ery_old_cars0 = 500000
V ery _old_cars
(13.3)
=
t
W recking_ratio V ery_old_cars_to_Russia + Aging_2
Aging_10,1,2,3,4 = 50000
Aging_1t = P urchace_ratiot5
(13.4)
Aging_20,1,2 = 50000 + 0
Aging_23,4,5 = 50000 + Old_cars_f rom_Germanyt2
Aging_2t = Aging_1t5 + Old_cars_f rom_Germanyt2
(13.5)
W recking_ratio =
(New_cars + Old_cars + V ery_old_cars)
T otal_amount_of _cars
W recking_ratio 0
(13.6)
Old_cars_f rom_Germany0 = 0
(13.7)
P urchace_ratio0 = 50000
(13.8)
V ery_old_cars_to_Russia = 0
(13.9)
Total amount of cars is dened as a curve, seen in the gure 13.10. The
amount of cars is predicted to increase. We can adjust the curve if forecasts
173
13.5.4
Flight simulator
174
tion. The most important thing is to watch the total amount of cars as well
as the amounts of dierent age groups. In Finland, we have the oldest cars
in the Europe. Those cars are unsafe and the pollution is huge. The goal of
politicians should be to lower the average age of Finnish cars. Politicians can
do this by lowering the taxes of imported cars from Germany or the taxes of
new cars in Finland.
With this ight simulator, it is possible to test how the average age of cars
behaves, if a great amount of cars is imported from Germany or if a great
amount of new cars is bought.
After testing dierent combinations, it can be seen that the only way to lower
the average age of Finnish cars is to lower the taxes of new cars. Later on the
model can be further developed and taxes and prices of cars can be added to
the model.
13.6
175
At the end of this chapter, current literature of the system dynamics and
system thinking is presented. The related software is discussed. The research
institutes studying system dynamics and thinking are listed.
13.6.1
Jay Forrester invented system dynamics in the late 50s. His work has been
referred earlier when the history of system dynamics was presented. At 90s
after the book The Fifth Disciple by Peter Senge, the system dynamic and
thinking seemed to gain popularity. Most likely visually attractive software
has brought system dynamics to wider knowledge. In addition, the ideas
of complex systems at the beginning of new millennium are new ground for
system dynamics.
The following books are published between 1990-2000. These are related to
Stermans Business Dynamics. The list is form Amazon.com [8]:
Jamshid Gharajedaghi, Systems Thinking: Managing Chaos and Complexity: A Platform for Designing Business Architecture, ButterworthHeinemann (Trd); (May 1999)
Barry Richmond, The "Thinking" in Systems Thinking: Seven Essential Skills, Pegasus Communications; (January 5, 2000)
Stephen G. Haines, The Managers Pocket Guide to Systems Thinking,
Human Resource Development Pr; Pocket edition (February 1999)
Joseph OConnor and Ian McDermott, The Art of Systems Thinking: Essential Skills for Creativity and Problem Solving, Thorsons Pub;
(April 1997)
Peter Checkland, Systems Thinking, Systems Practice: Includes a 30Year Retrospective, John Wiley & Sons; (September 16, 1999)
Stephen G. Haines, The Systems Thinking Approach to Strategic Planning and Management, Saint Lucie Press; (June 13, 2000)
Gerald M. Weinberg, An Introduction to General Systems Thinking,
Dorset House; (April 15, 2001)
176
13.6.2
Research institutes
13.7
Conclusions
Bibliography
[1] Radzicki, M., U.S. Department of Energys Introduction to System Dynamics A System Approach to Understanding Complex Policy Issues,
1997. Available at http://www.systemdynamics.org/DL-IntroSysDyn
[5.4.2003].
[2] Forrester, J., The Beginning of System Dynamics,
Banquet Talk at the international meeting of System Dynamics Society, Stuttgart, Germany, July 13, 1989. Available at
http://sysdyn.clexchange.org/sdep/papers/D-4165-1.pdf [30.3.2003].
[3] Home page of System Thinking Europe Oy.
Available at http://www.steoy.com/index.html [5.4.2003].
[4] Home page of System Dynamics Society.
Available at http://www.systemdynamics.org/ [5.4.2003].
[5] Sterman, J., Business Dynamics System Thinking and Modelling for
Complex World, McGraw-Hill, 2000.
[6] Senge, P., Fifth Discipline The Art & Practice of The learning Organization, Doubleday, New York, 1990.
[7] Forrester, M. & Forrester, N., Simple Beer Distribution Game Simulator.
Available at http://web.mit.edu/jsterman/www/SDG/MFS/simplebeer.html
[30.3.2003].
[8] List of books at Amazon.com.
Available at http://www.amazon.com/exec/obidos/ASIN/0750671637/
ref=pd_sxp_elt_l1/002-1327772-7212832 [30.3.2003].
[9] Sastry, A., & Sterman, J., Desert Island Dynamics: An Annotated
Survey of the Essential System Dynamics Literature. Available at
http://web.mit.edu/jsterman/www/DID.html [30.3.2003].
177
178
BIBLIOGRAPHY
Session 14
Towards a Systemic View of
Complexity?
Yuba Raj Adhikari
Laboratory of Process Control and Automation
Helsinki University of Technology, Finland
General System Theory concept is close to theory of control systems. In this
chapter, we try to nd a systematic view of complexity from the viewpoint of
system concepts. The rst part of the chapter covers about General System
Theory from Bertalanys book General System Theory [1]. In this book he
shows that reductionism is completely wrong and modelling the system from
the holistic viewpoint is the correct approach. One approach to looking at
complex systems in a systemic perspective and a way to control the emerging
patterns as well is demonstrated.
14.1
14.1.1
System Theory represents a novel paradigm in scientic thinking [1]. General system theory is similar to theory of evolution, which comprises about
everything between fossils digging, anatomy and the mathematical theory
of selection, or behavior theory extending from bird watching to sophisticated neurophysiological theories. Broadly speaking, three major aspects of
the theory are: System science, system technology and system philosophy.
179
180
System science deals with scientic exploration and theory of systems in the
various sciences, physics, biology, psychology, etc., and general system theory
as doctrine of principles applying to all systems. Technology has been led
to think not in terms of single machine but in those of systems. There is
system philosophy, i.e. the reorientation of thought and worldview ensuring
from the introduction of system as a new scientic paradigm. It has been
learned that for an understanding not only the elements but their interrelations as well are required. This is the domain of general system theory.
System thinking plays a dominant role in a wide range of elds from industrial enterprise and armaments to esoteric topics of pure science. Professions
and jobs have appeared going under names such as system design, system
analysis, system engineering and others. They are very nucleus of a new
technology and technocracy.
14.1.2
History
The idea of general system theory was rst introduced by the author of the
book General System Theory by Ludwig von Bertalany prior to cybernetics,
systems engineering and the emergence of related elds. There had been a
few preliminary works in eld of general system theory. Khlers physical
gestalten (1924) pointed in this direction but did not deal with the problem
in full generality, restricting its treatment to physics. Lotka (1925) dealt
with a general concept of system but being himself a statistician his interests
more lie in population problem than biological problem [1].
In early 20s, Bertalany advocated an organismic conception in biology that
emphasizes consideration of the organism as a whole or system, and sees the
main objective of biological sciences in the discovery of the principles of organization at its various levels. His rst statement goes back to Whiteheads
philosophy of organic mechanism, published in 1925. Canons work on
homeostasis appeared in 1929 and 1932.
In the rst year of the Center for Advanced Study in the Behavioral Sciences
(Palo Alto), the biomathematician A. Rapoport, the physiologist Ralph Gerard and Bertalany found themselves together. The project of a Society for
General System Theory was realized at the Annual Meeting of the American
Association for the Advancement of Science (AAAS) in 1954 and its name is
later changed to Society for General System Research, aliated to AAAS.
Local groups of the Society were established at various centers in United
States and subsequently in Europe.
Meanwhile another development had taken place. Norbert Wieners Cy-
181
14.1.3
Miniskirt and long hair are called teenage revolution; any new styling of automobiles or drug introduced by the pharmaceutical industry can also termed
so and in strictly technical sense one can speak of scientic revolutions. A
scientic revolution is dened by the appearance of new conceptual schemes
or paradigms. These bring to the fore aspects, which previously were not
seen or perceived, or even suppressed in normal science.
The system problem is essentially the problem of the limitations of analytical procedures in science. This used to be expressed by half-metaphysical
statements such as emergent evolution or the whole is more than a sum of
its parts but has a clear operational meaning. Analytical procedure means
that an entity investigated be resolved into, and hence can be constituted
or reconstituted from, the parts put together, these procedures being understood both in their material and conceptual sense. This is the basic principle
of classical science, which can be circumscribed in dierent ways: resolution
into isolable causal trains, seeking for atomic units in the various elds of
science etc.
Application of analytical procedure depends on two conditions. The rst
is that interactions between parts be non-existent or weak enough to be
neglected for certain research purposes. Only under this condition, the parts
can actually be worked out, logically, and mathematically, and then be put
together. The second condition is that the relations describing the behavior
of parts be linear; only then is the condition of summativity given e.g. partial
processes can be superimposed to obtain the total process etc.
These conditions are not fullled in entities called systems, i.e., consisting
of parts in interaction. The prototype of their description is a set of simultaneous dierential equations which are nonlinear in the general case.
A system or organised complexity may be circumscribed by the existence
of strong interactions or interactions which are nontrivial, i.e., nonlinear.
The methodological problem of system theory, therefore, is to provide for
problems which, compared with the analytical-summative ones of classical
science, are of a more general nature.
There are various approaches to deal with such problems. The more important approaches are as follows [1]:
182
183
14.1.4
There exist models, principles, and laws that apply to generalized systems
or their subclasses, irrespective of their particular kind, the nature of their
component elements, and the relations or forces between them. It seems
legitimate to ask for a theory, not of systems of a more or less special kind,
184
14.1.5
185
In any closed system, the nal state is unequivocally determined by the initial
conditions, e.g., the motion in a planetary system where the position of the
planet at a time is equivocally determined by its position at time T0 . In
equilibrium, the nal concentrations of the reactants depend on the initial
concentration. If either the initial conditions or the process is altered, the
nal state will also be changed. This is not so in an open system. Hence
the same nal state may be reached from dierent initial conditions and in
dierent ways.
An open system is dened as a system exchanging matter with its environment, presenting import and export, building up and breaking down of its
material components. Living systems are basically open system.
14.1.6
Three dierent kind of distinctions may be made when dealing with complexities of elements: Distinctions according to number, species and relations
of elements. In rst and second types, complexity can be understood as the
sum of elements in isolation (summative and constitutive, respectively). In
summative case the characteristics of elements within and outside the system
are the same. In the third type, not only the elements should be known, but
also the relations between them. Constitutive characteristics are those which
are dependent on the specic relations within the complex. The example for
the rst type is weight or molecular weight. An example in second type
is chemical characteristics (isomorphism). The meaning of the somewhat
mystical expression the whole is more than the sum of the parts is simply
that constitutive characteristics are not explainable from the characteristics
of isolated parts. The characteristics of the complex, therefore compared to
those of the elements, appear as new or emergent.
A system can be dened as a set of elements standing in interrelations. In
mathematics a system can be dened in various ways. We can choose a system of simultaneous dierential equations for dierent applications, e.g, law
of mass action, demographic problems, kinetics of cellular processes and the
theory of competition within an organism, etc. The dierential equations can
be used in describing the growth of the system. A solution of the equation,
the exponential law is useful in various dierent elds. As we talk about
the system or whole, every whole is based on the competition of its elements,
and presupposes the struggle between parts. The systems of dierential
186
14.2
There are still plenty of challenges we have to face in the eld of automation
and system engineering. Because of the new devices and technologies, there
is an information explosion what comes to available process data. The
process models are also becoming more and sophisticated. We need new
approaches to cope with the wealth of data and models. Theory of complex
systems promises wonderful solutions to such problems [2].
The properties of a complex system can also be pointed as below:
It consists of a large assemblage of interconnected nonlinearly interacting parts.
Parts evolve over time, adapting to the environment.
187
188
189
and might also say a bad modelling technique because everything is reduced
to elementary units as it is always allowed to avoid metaphysical questions
away and to take only the fruitful issues in developing the model [2]. System
thinking approach may help in dealing complexities. System thinking is a
better way to deal with our most dicult problems [11].
14.2.1
Thomas Kuhn put it forward that there are paradigm shifts within a science
[1] things are seen from another point of view, and new conceptual tools
are introduced. Systematic thinking may help in understanding complex
systems, and may visualize new challenges. A new, data-centered world
issuing to general system theory is trying to be demonstrated [2].
Theory of complex system may give new tools when searching for new tools
for mastering complicated automation systems. At dierent level of abstraction, the appropriate way of looking at the whole system changes altogether.
A good example can be modelling of gases [2]:
Elementary particles of gas behave stochastically (quantum theory to
be applied)
Atoms behave deterministically (Newtonian ideal gas model)
Atom groups behave stochastically (statistical mechanics)
Large volumes behave deterministically (states described by pressure
and temperature)
Still larger volumes behave stochastically (turbulence and uctuations
becoming acute)
Perfectly stirred volume behaves deterministically (ideal mixer model).
There still remain the same underlying laws, but the lower level tools are
not the most economical ones in describing the complexities. The key point
is to look at the phenomena from a higher abstraction level, ignoring the
details of the physical components. Rather than concentrating on the actual realizations of the dynamic signal, one looks at the statistical and static
relationships between the qualiers or the process parameters and corresponding qualities, or process behaviors. It can be claimed that a higherlevel statistical model emerges from the lower level deterministic behaviors.
190
14.3. Conclusion
191
Continuous nonlinearity: Smooth nonlinear functions can be approximated by locally linearising the function and using the nearest submodel to represent each sample.
Clustered data: Assuming there exit dierent operating modes and
fault conditions.
Independent components: They are capable of nicely capturing physically relevant phenomena.
Linear representations are too weak to capture the real-life complexity. If
the linear structures are summed or combined the resultant structure is still
linear. But still the linear basic underlying structure (ane) is assumed and
it is boosted with nonlinearity to make complex representations possible and
to facilitate emergence of sparsity.
14.2.2
14.3
Conclusion
192
of the same principles in dierent elds isolated from each other. Bertanlay
suggests the theories and models are useful in each level of the hierarchy of
the system structure. Holistic view of understanding the system is that not
only the parts but also the interactions in between them must be considered
once the system is modelled.
The most natural way to connect low-level models to high-level tools is simulation. New data-centered world might be a new approach to be issued in
General System Theory. Mastering a complex large-scale system is to understand what is happening, what is relevant and what is not in the wealth
of data.
Bibliography
1. Bertalany L.:General System Theory. (revised edition). Foundations,
Development, Applications. George Braziller, New York, 1969.
2. Hytyniemi, H.: Emergence and Complex Systems Towards a New
Science of Industrial Automation? Proceedings of the 4th International
Conference on Intelligent Processing and Manufacturing of Materials
(IPMM03), May 1823, 2003, Sendai, Japan.
3. Hytyniemi, H.: Towards New Languages for Systems Modelling . Proceedings of the 42nd Scandinavian Simulation Conference SIMS02,
September 2627, 2002, Oulu, Finland.
4. Hytyniemi, H.: Emergent Phenomena and Optimization of Parameters. SIMS02 Conference, Oulu, Finland,September 2627, 2002.
5. Hytyniemi, H.: Studies on Emergence and Cognition Part 1: LowLevel Functions. Finnish Articial Intelligence Conference (STeP02),
December 1617, 2002, Oulu, Finland.
6. Hytyniemi, H.: Studies on Emergence and Cognition Part 2:
Higher-Level Functionalities. Finnish Articial Intelligence Conference
(STeP02), December 16-17, 2002, Oulu, Finland.
7. http://www.new-paradigm.co.uk/complex-od.htm
8. Lloyd S.: Complexity the State of art . ESD Internal Conference
Extended Abstracts, MIT, March 25, 2002.
http://esd.mit.edu/wps/esd-wp-2002-05.pdf .
14.3. Conclusion
193
9. http://geog.tamu.edu/laura/classes/complexity.html
(10th , April, 2003).
10. Wolfram S.: A New Kind of Science.
Illinois, 2002.
11. http://www.thinking.net/Systems_Thinking/Intro_to_ST/intro_to_st.html,
(15th May, 2003)
12. Ilachinski, A.: Analytical Implication of Complexity. Systems and Tactics Team (Operating Forces Division) Center for Naval Analyzes.
Report 132
Gadoura, I. A.
Design of Robust Controllers for Telecom Power Supplies. September 2002.
Report 133
Hytyniemi, H.
On the Universality and Undecidability in Dynamic Systems. December 2002.
Report 134
Report 135
Blomqvist, E.
Security in Sensor Networks. February 2003.
Report 136
Zenger, K.
Modelling, Analysis and Controller Design of Time-Variable Flow Processes. March 2003.
Report 137
Hasu, V.
Adaptive Beamforming and Power Control in Wireless Communication Systems. August 2003.
Report 138
Report 139
Report 140
Viitamki, P.
Hybrid Modeling of Paper Machine Grade Changes. May 2004.
Report 141
Pyhnen, S.
Support Vector Machine Based Classification in Condition Monitoring of Induction Motors. June 2004.
Report 142
Elmusrati, M. S.
Radio Resource Scheduling and Smart Antennas in Cellular CDMA Communication Systems. August
2004.
Report 143
Tenno, A.
Modelling and Evaluation of Valve-Regulated Lead-Acid Batteries. September 2004.
Report 144
Hytyniemi, H.
Hebbian Neuron Grids: System Theoretic Approach. September 2004.
Report 145
Hytyniemi, H. (ed.)
Complex Systems: Science at the Edge of Chaos - Collected papers of the Spring 2003 postgraduate
seminar. October 2004.
ISBN 951-22-7507-4
ISSN 0356-0872
Picaset Oy, Helsinki 2005