Matt I. - ShinyVerse
Download
Report
Transcript Matt I. - ShinyVerse
Evolving Neural Networks
Learning and Evolution: Their secret
conspiracy to take over the world.
Adaptation
There are two forms of adaptation
Learning
Training on a set of examples.
Fitting your behavior to training data.
Minimizing error.
Evolution
Population based search.
Random Mutation
Reproduction
Fitness Selection
Combining The Two
Combining the strategies of evolutionary
algorithms with learning produces
“evolutionary artificial neural networks”
or EANNs.
This increases the adaptability of both in
a way that neither system could achieve
on their own.
This also can give rise to extremely
complex relationships between the two.
What Can an EANN Do?
Adjust the network weights.
Learning rules.
Evolution
Build an architecture to fit the problem at
hand
Does it need hidden layers?
Is the propagation delay too large?
Is the environment dynamic?
Perhaps the m ost lofty goal is evolving
a learning rule for the network.
Evolving The Weights
Why evolve the weights? What’s wrong
with back propagation?
Backpropagation is a gradient accent
algorithm. These algorithms can get stuck
on a local maximum or local minimum
solutions.
Optimal solution
Local Max
Initial weights
Overcoming Local Min/Max
Since evolutionary algorithms are population
based they have no need for gradient
information.
Subsequently they are better in noisy or
complex environments
Though getting stuck on local maxes can be
avoided there is no guarantee that any maxima
will be found.
Population Samples
The Permutation Problem
A gigantic problem that reflects the noisy real
world solution is the permutation or competing
convention problem.
Caused by a many-to-one mapping from genotype to
phenotype within neural nets.
This problem kills the efficiency and effectiveness of
crossover because effective parents can produce
slacker offspring…..sorry mom and dad.
7
3
3
10
4
2
2
7
4
10
Does Evolution Beat
Backpropagation?
Absolutely!
D.L. Prados claims that GA-based training algorithms
completed the tests in 3 hours 40 minutes, while
networks using the generalized delta rule finished in
23 hours 40 minutes.
No way man!
H. Kitano claims that when testing a Genetic Algorithm
backpropagation combination, it was at best as
efficient as other backpropagation variants in small
networks, but rather crappy in large networks.
It just goes to show you that the best algorithm
is always problem dependent.
Evolving Network
Architectures
The architecture is very important to the
information processing capabilities of the
neural network.
Small networks without a hidden layer can’t solve
problems such as XOR, that are not linearly
separable.
Large networks can easily overfit a problem to
match the training data, constricting their ability
to generalize a problem set.
Constructing the Network
Constructive algorithms take a minimal
network and build up new layers nodes and
connections during training.
Destructive algorithms take a maximal
network and prunes unnecessary layers nodes
and connections during training.
The network is evaluated based on specific
performance criteria, for example, lowest
training error or lowest network complexity.
The Difficulties Involved
The architectural search space is infinitely large
Changes in nodes and connection can have a
discontinuous effect on network performance.
Mapping from network architecture to behavior
is indirect, dependent on evaluation and
epistatic…don’t ask.
Similar architectures may have highly divergent
performance.
Similar performance may be attained by diverse
architectures.
Evolving The Learning Rule
The optimal learning rules is highly dependent
on the architecture of the network
Designing an optimal learning rule is very hard
when little is known about the architecture,
which is generally the case.
Different rules apply to different problems.
Certain rules make it easier to learn patterns and in
this regard are more efficient.
Less efficient learning rules can learn exceptions to
the patterns.
Which one is better? Depends on who you ask.
Facing Reality
One can see the advantage of having a network
that is capable of learning to learn. Combined
with the ability to adjust it architecture, this sort
of neural net would seem to be approaching
intelligence.
The reality is that the relationship between
learning and evolution is extremely complex.
Subsequently research into evolving learning
rules is really in it’s infant stages.
Lamarckian Evolution.
On a side note Belew McInerney &
Schraudolf state that their findings
suggest a reason why Lamarckian
inheritance cannot be possible.
Due to issues like the permutation
problem it is impossible to transcribe
network behavior into a genomic
encoding since it is possible that there
are infinitely many encodings that will
produce a phenotype.
Can It Be Implemented?
Yes it can.
That’s all I got for this slide. It really seems
like a waste, but at least its not paper..
Go-Playing Neural Nets
Alex Lubberts and Risto Miikkulainen
have created a Co-Evolving Go-playing
Neural Network.
No program play go at any significant
level of experience.
They decided that a parasite host
relationship will foster a competitive
environment for evolution.
The Fight to Survive
The host
Attempt to find an optimal solution for
winning on scaled down 5X5 Go board.
The parasites
Attempt to find and exploit weaknesses
within the host population.
Populations are evaluated one at a time
so each population takes turns being a
host or a parasite.
Tricks of the Trade
Competitive Fitness Sharing
Unusual or special individuals are rewarded.
If an individual whups an opponent that is very tough
even if that individual lost most of it’s other games
that host may still be rewarded.
Shared Sampling
To cut down on the number of games a sample set is
pitted against opponents.
Hall of Fame
Old timers that have competed well are put into a
steel cage match with the tyros to ensure that new
generations are improving.
Conclusions
They found that co-evolution did in fact
increase the playing ability of their networks.
After 40 generations using their tournament
style selection the co-evolved networks had
nearly tripled the number of wins against a
similar network that was evolved without a
host parasite relationship.
EANN Used to Classify
Galaxies
Erick Cantu-Paz & Chandrika Kamath
Attempting to bring automation in the
classification of galaxies using neural
networks.
The learning algorithm must be carefully
tuned to the data.
The relevant features of a classification
problem may not be known before building
the network or tuning the learning rule.
How It Worked
Six combinations of GA and NN where
compared.
The interesting part is the evolution of feature
selection
GAs where to select what features are important
enough that the NN needs to know them in order to
classify a galaxy.
GAs consistently selected half the features and
supposedly of the half they selected most of the
features where relevant to classification such as
symmetry measures and angles.
Two point crossover was used along with a
fitness bias towards networks that learn quickly.
Findings
Several of the evolved networks where
competitive with human designed
networks.
The best evolving feature selection.
Identifying Bent-Double Galaxies 92.99%
accuracy.
Identifying Non-Bent 83.65% accurac
y.
Alright! He’s About to Shut
Up
So in conclusion there is a lot of room for
improvement in EANNs and there is a lot
to explore.
So quit what your doing and build an
evolving learning program. Tell your
friends that you have already given it a
body and soon you will make it capable
of replicating itself. If they are gullible
enough you might get to watch them
squirm.
References, The lazy way.
Richard Belew, John McInerney, Nicol Schraudolph: Evolving
Networks
http://www-cse.ucsd.edu/users/rik/papers/alife91/evol-net.ps
Erick Cantu-Paz C. Kamath: Evolving Neural Networks for the
Classification of Galaxies
http://www.llnl.gov/CASC/sapphire/pubs/147020.pdf
Xin Yao Evolving Artificial Neural Networks
http://www.cs.bham.ac.uk/~xin/papers/published_iproc_sep99.pdf
Alex Lubberts, Risto Miikkulainen Co-Evolving a Go-Playing
Neural Network
http://nn.cs.utexas.edu/downloads/papers/lubberts.coevolutiongecco01.pdf