SITutorial
SITutorial
1
Dynamic Social Impact Theory:
Axelrod’s Culture Model
Summary
z Populations of individuals are pictured as strings of
z Individuals influence one another, and in doing so symbols, or “features”
become more similar
z Probability of interaction between two individuals is a
z Patterns of belief held by individuals tend to correlate function of their similarity
within regions of a population
z Individuals become more similar as a result of
z This model is consistent with findings in the fields of interactions
social psychology, sociology, economics, and
z The observed dynamic is polarization , homogeneous
anthropology.
subpopulations that differ from one another
z No effect of similarity on probability of interaction z Individuals searching for solutions learn from the
z The effect of similarity is negative, in that it is experiences of others (individuals learn from their
dissimilarity
dissimilarity that creates boundaries between cultural neighbors)
regions z An observer of the population perceives phenomena of
z Interaction occurs if fitnesses are different which the individuals are the parts (individuals that
interact frequently become similar)
z Culture affects the performance of individuals that
comprise it (individuals gain benefit by imitating their
neighbors)
2
Features of Evolutionary Computation
Evolutionary Computation Algorithms
(EC) Paradigms
z EC paradigms utilize a population of points (potential 1. Initialize the population
solutions) in their search 2. Calculate the fitness of each individual in the
z EC paradigms use direct “fitness” information instead of population
function derivatives or other related knowledge
3. Reproduce selected individuals to form a new
z EC paradigms use probabilistic, rather than population
deterministic, transition rules
4. Perform evolutionary operations such as crossover and
mutation on the population
5. Loop to step 2 until some condition is met
Particle swarms
(physical position not a factor)
3
Introduction to Particle Swarm Introduction to Particle Swarm
Optimization Optimization (PSO), Continued
z A concept for optimizing nonlinear functions
z Has roots in artificial life and evolutionary computation
z A “swarm” is an apparently disorganized collection
(population) of moving individuals that tend to cluster z Developed by Kennedy and Eberhart (1995)
together while each individual seems to be moving in z Simple in concept
a random direction z Easy to implement
z We also use “swarm” to describe a certain family of z Computationally efficient
social processes z Effective on a variety of problems
4
PSO Adherence to Swarm Intelligence
Further Details of PSO
Principles
z Performance of each particle measured according to a z Proximity: n-dimensional space calculations carried out
predefined fitness function. over series of time steps
z Inertia weight influences tradeoff between global and z Quality: population responds to quality factors pbest
local exploration. and gbest (or lbest )
z Good approach is to reduce inertia weight during run z Stability: population changes state only when gbest (or
(i.e., from 0.9 to 0.4 over 1000 generations) lbest ) changes
z Usually set c1 and c2 to 2 z Adaptability: population does change state when gbest
z Usually set maximum velocity to dynamic range of (or lbest ) changes
variable
z Introduction
z Definitions and review of previous work
“Implementation of Evolutionary Fuzzy Systems”
z Advantages and disadvantages of previous
Authors: Shi, Eberhart, Chen approaches
IEEE Transactions on Fuzzy Systems z Using particle swarm optimization (PSO)
April 1999 z An example application
z Conclusions
5
Evolving Neural Networks with Particle
Introduction
Swarm Optimization
z Neural networks are very good at some problems, z Evolve neural network capable of being universal
such as mapping input vectors to outputs approximator, such as backpropagation or radial basis
z Evolutionary algorithms are very good at other function network.
problems, such as optimization z In backpropagation, most common PE transfer function
z Hybrid tools are possible that are better than either is sigmoidal function: output = 1/(1 + e - input )
approach by itself z Eberhart, Dobbins, and Simpson (1996) first used PSO
z Review articles on evolving neural networks: to evolve network weights (replaced backpropagation
Schaffer, Whitley, and Eshelman (1992); Yao (1995); learning algorithm)
and Fogel (1998) z PSO can also be used to indirectly evolve the structure
z Evolutionary algorithms usually used to evolve of a network. An added benefit is that the preprocessing
network weights, but sometimes used to evolve of input data is made unnecessary.
structures and/or learning algorithms
Evolving Neural Networks with Particle Evolving the Network Structure with
Swarm Optimization, Continued PSO
z Evolve both the network weights and the slopes of z If evolved slope is sufficiently small, sigmoidal output can
sigmoidal transfer functions of hidden and output PEs.
PEs. be clamped to 0.5, and hidden PE can be removed.
z If transfer function now is: output = 1/(1 + e -k*input ) Weights from bias PE to each PE in next layer are
increased by one-
one-half the value of the weight from the
then we are evolving k in addition to evolving the
PE being removed to the next-
next-layer PE. PEs are thus
weights.
pruned, reducing network complexity.
z The method is general, and can be applied to other
z If evolved slope is sufficiently high, sigmoid transfer
topologies and other transfer functions.
function can be replaced by step transfer function. This
z Flexibility is gained by allowing slopes to be positive or works with large negative or positive slopes. Network
negative. A change in sign for the slope is equivalent to computational complexity is thus reduced.
a change in signs of all input weights.
6
Outline Original Version with Inertia Weight
z
z
Brief review of particle swarm optimization
Types of dynamic systems
(
vid = wi vid + c1rand ()( pid − xid ) + c2 Rand () p gd − xid )
z Practical application requirements xid = xid + vid
z Previous work Where d is the dimension, c1 and c2 are positive constants,
z Experimental design rand and Rand are random functions, and w is the inertia
z Results weight. For neighborhood version, change pgd to pld.
z Conclusions and future effort
z Few practical problems are static; most are dynamic z Testing Parabolic Function
z Most time is spent re-
re-optimizing (re-
(re-scheduling, etc.) N
z Many systems involve machines and people error = ∑ ( xi − offset ) 2
z These systems have inertia
i =1
z 10-
10-100 seconds often available for re-
re-optimization
z Offset = offset + severity
z Eberhart’s Law of Sufficiency applies: If the solution is
good enough, fast enough, and cheap enough, then it is z Severity 0.01, .1, .5
sufficient z 2000 evaluations per change
z 3 dimensions, dynamic range –50 to +50
7
Previous Work: References Experimental Design
z Angeline,
Angeline, P.J. (1997) Tracking extrema in dynamic z Two possibilities with swarm
environments. Proc. Evol.
Evol. Programming VI, VI, z Continue on from where we were
Indianapolis, IN, Berlin: Springer-
Springer-Verlag,
Verlag, pp. 335-
335-345 z Re-
Re-initialize the swarm
z Bäck,
ck, T. (1998). On the behavior of evolutionary z Inertia weight of [0.5+(Rnd/2.0)] used
algorithms in dynamic environments. Proc. Int. Conf. on
Evol. z 20 particles; update interval of 100 generations
Evol. Computation,
Computation, Anchorage, AK. Piscataway, NJ:
IEEE Press, pp. 446-
446-451 z When change occurred:
z Retained the position of each particle
z Reset values of pbest (also of gbest)
gbest)
PSO average best over all runs PSO average best over all runs
Severity = 0.5 Severity = 0.1
Three dimensions Three dimensions
10000
10000
1000
1000
100
100
10
Average best value over all runs
10
1
1 0.1
0.1 0.01
0.01 0.001
0.001 0.0001
1E- 05
0.0001
1E- 06
1E-05
1E- 07
1E-06
1E- 08
1E-07
1E- 09
1E-08
1E- 10
1E-09
1E- 11
1E-10
PSO average best over all runs PSO average best over all runs
Severity = 0.1 Severity = 0.5
10 dimensions 10 dimensions
10000
10000
1000
1000
100
Average best value over all runs
100
Average best value over all runs
10
1 10
0.1
1
0.01
0.1
0.001
0.01
0.0001
0.00001 0.001
0.000001
0.0001
8
PSO average best over all runs
Comparison of Results:
Severity = 1.0
10 dimensions Error Values Obtained in 2000 Evaluations
Severity 0.1 Severity 0.5
10000
1000
100
Angeline 5x10-4 – 10-3 0.01-
0.01-0.10
Average best value over all runs
10
1
Bäck 2x10-5 10-3
0.1
0.01
Eberhart & 10-10 - 10-9 10-9 – 10-8
Shi
0.001
0.0001
Example Application:
Conclusions and Future Efforts
Reactive Power and Voltage Control
z Our results, including those in 10 dimensions and z Japanese electric utility
with severity = 1, are promising z PSO used to determine control strategy
z We are applying approach to other benchmark z Continuous and discrete control variables
functions, and to practical logistics applications
z Hybrid binary/real-
binary/real-valued version of PSO developed
z System voltage stability achieved using a
continuation power flow technique
evolutionary
programming
Machine Operations
9
Container Planning Sequences More Examples of Recent Applications
New Book
10