SlideShare a Scribd company logo
1 of 3
Download to read offline
Methods of Combining Neural Networks and Genetic Algorithms
                                                 Talib S. Hussain
                                               Queen’s University
                                            hussain@qucis.queensu.ca

1. Introduction                                              method, then he can program the network structure
                                                             explicitly. However, if the problem is very complex or
          In the past decade, two areas of research which    has no known solution, the developer may not know
have become very popular are the fields of neural            what structure to give the network. To this end, most
networks (NNs) and genetic algorithms (GAs). Both are        neural network models include a learning rule which can
computational abstractions of biological information         change the network’s structure over the course of
processing systems, and both have captured the               training to arrive at a good final solution. Back-
imaginations of researchers all over the world. In           propagation is the most popular learning rule.
general, NNs are used as learning systems and GAs as
optimisation systems, but as many researchers have           1.2 Genetic Algorithms
discovered, they may be combined in a number of                        A variety of computational models based on
different ways resulting in highly successful adaptive       evolutionary processes have been proposed, and the
systems. In this tutorial, a summary will be given of        most popular models are those known as genetic
these combination methods. This summary is not meant         algorithms.      A genetic algorithm has four main
to be exhaustive, but rather to be indicative of the type    elements: the genetic code, a concise representation for
of research being conducted. For a more detailed             an individual solution; the population, a number of
discussion, see Yao (1993) and Schaffer et al. (1992).       individual solutions; the fitness function, an evaluation
          The tutorial is broken into three sections. In     of the usefulness of an individual; and the propagation
the first section, a brief introduction to the foundations   techniques, a set of methods for generating new
of neural networks and genetic algorithms is given. It is    individuals. The genetic algorithm works as follows.
assumed that the participants have a basic understanding     First, a population of individuals is generated by
of both fields, and this introduction is designed as a       randomly selecting different genes. The fitness of each
short refresher. In the second section, a variety of         individual is then evaluated, and the propagation
approaches to integrating NNs and GAs are presented.         techniques are applied to highly fit individuals to
In the final section, some of the key research issues are    generate a new population - the next generation. The
discussed.                                                   cycle of evaluate and propagate continues until a
                                                             satisfactory solution, hopefully optimal, is found.
1.1 Neural Networks                                                    In a typical genetic algorithm, the genetic code
          To set up the terminology for the rest of the      is a fixed-length bit string and the population is always a
paper, let us review the basics of a neural network. A       fixed size. The three most common propagation
neural network is a computational model consisting of a      techniques are elitism, mutation and crossover. In
number of connected elements, known as neurons. A            elitism, the exact individual survives into the next
neuron is a processing unit that receives input from         generation. In mutation, a new individual is created
outside the network and/or from other neurons, applies a     from an old one by changing a small number of
local transformation to that input, and provides a single    randomly selected bits in its gene. In crossover, a new
output signal which is passed on to other neurons and/or     individual is created from two old ones by randomly
outside the network. Each of the inputs is modified by a     selecting a split point in their genes are creating a new
value associated with the connection. This value is          gene with the left part from one parent and the right part
referred to as the connection strength, or weight, and       from another. In any genetic algorithm, the two key
roughly speaking, represents how much importance the         aspects are the genetic representation and the fitness
neuron attaches to that input source. The local              function. Together, these determine the type of problem
transformation is referred to as the activation function     which is being solved and the possible solutions which
and is usually sigmoidal in nature.                          may be generated.
          A typical neural network is capable of             2. Combining NNs and GAs
representing many functions, as proved by
Komolgorov’s Theorem, but finding the best network           2.1 Supportive and Collaborative
needed to solve a specific problem is a very open-ended              Researchers have combined NNs and GAs in a
problem. If the developer knows the exact solution           number of different ways. Schaffer et al. have noted
that these combinations can be classified into one of two     better convergence. Better still, since GAs are good at
general types - supportive combinations in which the          global search but inefficient at local finely tuned search,
NN and GA are applied sequentially, and collaborative         a hybrid approach combining GAs and gradient descent
combinations in which they are applied simultaneously.        are attractive. (Yao)
          In a supportive approach, the GA and the NN
are applied to two different stages of the problem. The       2.3 Evolution of architectures
most common combination is to use a GA to pre-                           In the second approach, the GA is used to
process the data set that is used to train a NN. For          select general structural parameters and the neural
instance, the GA may be used to reduce the                    learning is used separately to trained the network and
dimensionality of the data space by eliminating               determine its fitness. This includes evolution of both
redundant or unnecessary features. Supportive                 the topology (i.e., connectivity pattern) and activation
combinations are not highly interesting since the GA          functions of each node, although most work has
and NN are used very independently and either can             concentrated on the former and little has been done on
easily be replaced by an alternative technique. Some          the latter.
other possible combinations include: using a NN to                       In architecture evolution, the genetic code can
select the starting population for the GA; using a GA to      be either a direct or indirect encoding of the network’s
analyse the representations of a NN; and using a GA           topology. In a direct encoding, each connection is
and NN to solve the same problem and integrating their        explicitly represented (e.g., a matrix where 1 indicate
responses using a voting scheme. (Schaffer et al.)            the presence of a connection and 0 indicates no
          Alternatively, in a collaborative approach, the     connection).      In an indirect encoding, important
GA and NN are integrated into a single system in which        parameters of the network are represented and the
a population of neural networks is evolved. In other          details of the exact connectivity are left to
words, the goal of the system is to find the optimal          developmental rules (e.g., specify the number of hidden
neural network solution. Such collaborative approaches        nodes and assume full connectivity between layers).
are possible since neural network learning and genetic                   In both cases, the exact neural network is not
algorithms are both form of search. A neural network          specified since the weights are determined by the
learning rule performs a highly constrained search to         initialisation routine and the network’s learning
optimise the network’s structure, while a genetic             algorithm. Thus, the evaluation of a gene is noisy since
algorithm performs a very general population-based            it is dependent upon the evaluation of the trained
search to find an optimally fit gene. Both are examples       network, and the GA finds the best set of architectural
of biased search techniques, and “any algorithm that          parameters rather than the best neural network.
employs a bias to guide its future samples can be
mislead in a search space with the right structure. There     2.4 Evolution of learning rules
is always an Achilles heal.” (Schaffer et al, p. 4) The
                                                                        In the final approach, the GA is used similarly
primary reason researchers have looked at integrating
                                                              to the evolution of architecture, but a parametric
NNs and GAs is the belief that they may compensate for
                                                              representation of the network’s learning rule is also
each other’s search weaknesses.
                                                              encoded in the gene. The genetic coding of topology in
                                                              this case is generally indirect.
2.2 Evolution of Connection Weights
                                                                        Evolving learning rules does not refer simply
          A genetic algorithm can be applied to               to adapting learning algorithm parameters (e.g., learning
optimising a neural network in a variety of ways. Yao         rate, momentum, etc.) but to adapting the learning
has indicated three main approaches.- the evolution of        functions themselves. This is an area of research which
weights, the evolution of topology, and the evolution of      has received little attention. “The biggest problem here
learning rules. In each case, the GA’s genetic code           is how to encode the dynamic behaviour of a learning
varies highly.                                                rule into static genotypes. Trying to develop a universal
          In the first, the GA is used as the learning rule   representation scheme which can specify any kind of
of the NN. The genetic code is a direct encoding of the       dynamic behaviours is clearly impractical let alone the
neural network, with each weight being represented            prohibitive long computation time required to search
explicitly. The population of the GA are all NNs with         such a learning rule space.” (Yao, p. 214)
the same basic topology, but with different weight
values. Mutation and crossover thus affect only the
                                                              3. Issues
weights of the individuals. A key question in such
system is whether to use binary weights or real-valued
ones - the latter increases the search space greatly.                  Collaborative combinations of NNs and GAs
          Using GAs instead of gradient descent               have sparked the interest of a great number of
algorithms to train the weights can result in faster and      researchers because of their obvious analogy to natural
systems.    A wide variety of systems have been               grammatical encoding has recently received some
developed and a number of research issues have been           attention. (Gruau, 1994) Grammar encoding is quite
considered.                                                   powerful since it is compact but can represent a great
                                                              range of networks.
3.1 The Baldwin Effect
          In general, one may wonder whether it really is     4. Conclusions
of any use to have both neural learning and genetic                    Neural networks and genetic algorithms are
search operating in the same system. Perhaps using just       two highly popular areas of research, and integrating
genetic search would work given enough time, or               both techniques can often lead to highly successful
perhaps a very general neural learning technique would        learning systems. The participants of this tutorial are
be sufficiently powerful. This is quite possibly true, but    encouraged to try applying evolutionary neural network
an observation from natural systems known as the              solutions, or even developing new combinations of their
Baldwin Effect provide a clearer answer.                      own.
          The Baldwin Effect states that in an
evolutionary system, successful genes can propagate           References
faster, and in some cases only, if the individuals are
capable of learning. This principle has been clearly          French, R. & Messinger, A. (1994). “Genes, phenes and
demonstrated in an artificial evolutionary system by                   the Baldwin Effect: Learning and evolution in
French & Messinger (1994). Thus, an evolutionary                       a simulated population,” Artificial Life IV,
system with simple individuals which can learn is                      277-282.
generally more successful than one with non-learning          Gruau, F. (1994) “Automatic definition of modular
individuals and probably also better than a single highly              neural networks,” Adaptive Behaviour, 3, 151-
complex learning individual.                                           184.
                                                              Schaffer, D., Whitley, D. & Eshelman, L. (1992)
3.2 Generalisation                                                     “Combinations of Genetic Algorithms and
          In evolving a neural network, attention must be
                                                                       Neural Networks: A survey of the state of the
paid to the trade-off between evolutionary fitness and
                                                                       art,” Proceedings of the International
generalisation ability. In many tasks, the final network
                                                                       Workshop on Combinations of Genetic
is trained on a small set of data and applied to a much
                                                                       Algorithms and Neural Networks. D. Whitley
larger set of data. The goal of the learning is actually to
                                                                       and D. Schaffer (Eds.,) Los Alamitos, CA:
develop a neural network with the best performance on
                                                                       IEEE Computer Society Press, 1-37.
the entire problem and not just the training data.
                                                              Yao, X. (1993) “Evolutionary artificial neural networks”
However, this can easily be overlooked during the
                                                                       International Journal of Neural Systems, 4,
development process.
                                                                       203-222.
          Thus, one must be careful when evolving
neural networks not to select for highly specialised,
poorly generalising networks. This is especially true in
problem areas which are highly dynamic.



3.3 Encoding Methods
          The two main properties of an encoding of a
neural network in a GA are its compactness and
representation capability. A compact encoding is useful
since the GA can then be efficiently applied to problems
requiring large NN solutions. An encoding should be
powerful enough to represent a large class of NNs or
else the GA may not generate very good solutions. For
instance, direct encoding is generally quite powerful in
representation, but not compact, while parameterised
encoding is compact, yet often represents a highly
restrictive set of structures.
          The discussion so far has focused on direct
encoding and parametric encoding of neural network
structure. Other possibilities also exist. In particular,

More Related Content

What's hot

Simulated annealing.ppt
Simulated annealing.pptSimulated annealing.ppt
Simulated annealing.ppt
Kaal Nath
 
SVM Tutorial
SVM TutorialSVM Tutorial
SVM Tutorial
butest
 
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Simplilearn
 
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
Simplilearn
 

What's hot (20)

Simulated annealing.ppt
Simulated annealing.pptSimulated annealing.ppt
Simulated annealing.ppt
 
Activation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkActivation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural network
 
Ensemble methods in machine learning
Ensemble methods in machine learningEnsemble methods in machine learning
Ensemble methods in machine learning
 
SVM Tutorial
SVM TutorialSVM Tutorial
SVM Tutorial
 
SVM & KNN Presentation.pptx
SVM & KNN Presentation.pptxSVM & KNN Presentation.pptx
SVM & KNN Presentation.pptx
 
Simulated annealing
Simulated annealingSimulated annealing
Simulated annealing
 
Artificial Intelligence and Mathematics
Artificial Intelligence and MathematicsArtificial Intelligence and Mathematics
Artificial Intelligence and Mathematics
 
Divide&Conquer & Dynamic Programming
Divide&Conquer & Dynamic ProgrammingDivide&Conquer & Dynamic Programming
Divide&Conquer & Dynamic Programming
 
Introduction to soft computing
Introduction to soft computingIntroduction to soft computing
Introduction to soft computing
 
Class imbalance problem1
Class imbalance problem1Class imbalance problem1
Class imbalance problem1
 
Low-cost convolutional neural network for tomato plant diseases classification
Low-cost convolutional neural network for tomato plant diseases classificationLow-cost convolutional neural network for tomato plant diseases classification
Low-cost convolutional neural network for tomato plant diseases classification
 
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
 
Machine Learning
Machine Learning Machine Learning
Machine Learning
 
Support vector machine
Support vector machineSupport vector machine
Support vector machine
 
Regularization
RegularizationRegularization
Regularization
 
Backpropagation algo
Backpropagation  algoBackpropagation  algo
Backpropagation algo
 
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
Random Forest Algorithm - Random Forest Explained | Random Forest In Machine ...
 
ID3 ALGORITHM
ID3 ALGORITHMID3 ALGORITHM
ID3 ALGORITHM
 
Data mining applications
Data mining applicationsData mining applications
Data mining applications
 
House price prediction
House price predictionHouse price prediction
House price prediction
 

Similar to Methods of Combining Neural Networks and Genetic Algorithms

Whale optimization mirjalili
Whale optimization mirjaliliWhale optimization mirjalili
Whale optimization mirjalili
Prashant Kumar
 
A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...
eSAT Journals
 
Capital market applications of neural networks etc
Capital market applications of neural networks etcCapital market applications of neural networks etc
Capital market applications of neural networks etc
23tino
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
butest
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
butest
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
butest
 
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
ijasuc
 

Similar to Methods of Combining Neural Networks and Genetic Algorithms (20)

Genome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniquesGenome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniques
 
Whale optimization mirjalili
Whale optimization mirjaliliWhale optimization mirjalili
Whale optimization mirjalili
 
A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...
 
A clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction ofA clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction of
 
G44083642
G44083642G44083642
G44083642
 
Multilevel techniques for the clustering problem
Multilevel techniques for the clustering problemMultilevel techniques for the clustering problem
Multilevel techniques for the clustering problem
 
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
 
Analysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String RecognitionAnalysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String Recognition
 
Cornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 NetsCornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 Nets
 
Capital market applications of neural networks etc
Capital market applications of neural networks etcCapital market applications of neural networks etc
Capital market applications of neural networks etc
 
Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...
 
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERINGA SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
 
International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...
 
Literature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesLiterature Survey On Clustering Techniques
Literature Survey On Clustering Techniques
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
Ga 1 conference
Ga 1 conferenceGa 1 conference
Ga 1 conference
 
H43014046
H43014046H43014046
H43014046
 
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
 

More from ESCOM

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo Som
ESCOM
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales Som
ESCOM
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som Slides
ESCOM
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som Net
ESCOM
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networks
ESCOM
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales Kohonen
ESCOM
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia Adaptativa
ESCOM
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1
ESCOM
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3
ESCOM
 
Art2
Art2Art2
Art2
ESCOM
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo Art
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima Cognitron
ESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
ESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
ESCOM
 
Counterpropagation
CounterpropagationCounterpropagation
Counterpropagation
ESCOM
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAP
ESCOM
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1
ESCOM
 

More from ESCOM (20)

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo Som
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales Som
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som Slides
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som Net
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networks
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales Kohonen
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia Adaptativa
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3
 
Art2
Art2Art2
Art2
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo Art
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima Cognitron
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation
CounterpropagationCounterpropagation
Counterpropagation
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAP
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1
 

Recently uploaded

會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
中 央社
 

Recently uploaded (20)

Word Stress rules esl .pptx
Word Stress rules esl               .pptxWord Stress rules esl               .pptx
Word Stress rules esl .pptx
 
An overview of the various scriptures in Hinduism
An overview of the various scriptures in HinduismAn overview of the various scriptures in Hinduism
An overview of the various scriptures in Hinduism
 
IPL Online Quiz by Pragya; Question Set.
IPL Online Quiz by Pragya; Question Set.IPL Online Quiz by Pragya; Question Set.
IPL Online Quiz by Pragya; Question Set.
 
How To Create Editable Tree View in Odoo 17
How To Create Editable Tree View in Odoo 17How To Create Editable Tree View in Odoo 17
How To Create Editable Tree View in Odoo 17
 
When Quality Assurance Meets Innovation in Higher Education - Report launch w...
When Quality Assurance Meets Innovation in Higher Education - Report launch w...When Quality Assurance Meets Innovation in Higher Education - Report launch w...
When Quality Assurance Meets Innovation in Higher Education - Report launch w...
 
UChicago CMSC 23320 - The Best Commit Messages of 2024
UChicago CMSC 23320 - The Best Commit Messages of 2024UChicago CMSC 23320 - The Best Commit Messages of 2024
UChicago CMSC 23320 - The Best Commit Messages of 2024
 
How to Manage Closest Location in Odoo 17 Inventory
How to Manage Closest Location in Odoo 17 InventoryHow to Manage Closest Location in Odoo 17 Inventory
How to Manage Closest Location in Odoo 17 Inventory
 
MOOD STABLIZERS DRUGS.pptx
MOOD     STABLIZERS           DRUGS.pptxMOOD     STABLIZERS           DRUGS.pptx
MOOD STABLIZERS DRUGS.pptx
 
Graduate Outcomes Presentation Slides - English (v3).pptx
Graduate Outcomes Presentation Slides - English (v3).pptxGraduate Outcomes Presentation Slides - English (v3).pptx
Graduate Outcomes Presentation Slides - English (v3).pptx
 
Removal Strategy _ FEFO _ Working with Perishable Products in Odoo 17
Removal Strategy _ FEFO _ Working with Perishable Products in Odoo 17Removal Strategy _ FEFO _ Working with Perishable Products in Odoo 17
Removal Strategy _ FEFO _ Working with Perishable Products in Odoo 17
 
Features of Video Calls in the Discuss Module in Odoo 17
Features of Video Calls in the Discuss Module in Odoo 17Features of Video Calls in the Discuss Module in Odoo 17
Features of Video Calls in the Discuss Module in Odoo 17
 
PSYPACT- Practicing Over State Lines May 2024.pptx
PSYPACT- Practicing Over State Lines May 2024.pptxPSYPACT- Practicing Over State Lines May 2024.pptx
PSYPACT- Practicing Over State Lines May 2024.pptx
 
24 ĐỀ THAM KHẢO KÌ THI TUYỂN SINH VÀO LỚP 10 MÔN TIẾNG ANH SỞ GIÁO DỤC HẢI DƯ...
24 ĐỀ THAM KHẢO KÌ THI TUYỂN SINH VÀO LỚP 10 MÔN TIẾNG ANH SỞ GIÁO DỤC HẢI DƯ...24 ĐỀ THAM KHẢO KÌ THI TUYỂN SINH VÀO LỚP 10 MÔN TIẾNG ANH SỞ GIÁO DỤC HẢI DƯ...
24 ĐỀ THAM KHẢO KÌ THI TUYỂN SINH VÀO LỚP 10 MÔN TIẾNG ANH SỞ GIÁO DỤC HẢI DƯ...
 
Andreas Schleicher presents at the launch of What does child empowerment mean...
Andreas Schleicher presents at the launch of What does child empowerment mean...Andreas Schleicher presents at the launch of What does child empowerment mean...
Andreas Schleicher presents at the launch of What does child empowerment mean...
 
Spring gala 2024 photo slideshow - Celebrating School-Community Partnerships
Spring gala 2024 photo slideshow - Celebrating School-Community PartnershipsSpring gala 2024 photo slideshow - Celebrating School-Community Partnerships
Spring gala 2024 photo slideshow - Celebrating School-Community Partnerships
 
Basic Civil Engineering notes on Transportation Engineering, Modes of Transpo...
Basic Civil Engineering notes on Transportation Engineering, Modes of Transpo...Basic Civil Engineering notes on Transportation Engineering, Modes of Transpo...
Basic Civil Engineering notes on Transportation Engineering, Modes of Transpo...
 
II BIOSENSOR PRINCIPLE APPLICATIONS AND WORKING II
II BIOSENSOR PRINCIPLE APPLICATIONS AND WORKING IIII BIOSENSOR PRINCIPLE APPLICATIONS AND WORKING II
II BIOSENSOR PRINCIPLE APPLICATIONS AND WORKING II
 
Envelope of Discrepancy in Orthodontics: Enhancing Precision in Treatment
 Envelope of Discrepancy in Orthodontics: Enhancing Precision in Treatment Envelope of Discrepancy in Orthodontics: Enhancing Precision in Treatment
Envelope of Discrepancy in Orthodontics: Enhancing Precision in Treatment
 
An Overview of the Odoo 17 Knowledge App
An Overview of the Odoo 17 Knowledge AppAn Overview of the Odoo 17 Knowledge App
An Overview of the Odoo 17 Knowledge App
 
會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文會考英文
 

Methods of Combining Neural Networks and Genetic Algorithms

  • 1. Methods of Combining Neural Networks and Genetic Algorithms Talib S. Hussain Queen’s University hussain@qucis.queensu.ca 1. Introduction method, then he can program the network structure explicitly. However, if the problem is very complex or In the past decade, two areas of research which has no known solution, the developer may not know have become very popular are the fields of neural what structure to give the network. To this end, most networks (NNs) and genetic algorithms (GAs). Both are neural network models include a learning rule which can computational abstractions of biological information change the network’s structure over the course of processing systems, and both have captured the training to arrive at a good final solution. Back- imaginations of researchers all over the world. In propagation is the most popular learning rule. general, NNs are used as learning systems and GAs as optimisation systems, but as many researchers have 1.2 Genetic Algorithms discovered, they may be combined in a number of A variety of computational models based on different ways resulting in highly successful adaptive evolutionary processes have been proposed, and the systems. In this tutorial, a summary will be given of most popular models are those known as genetic these combination methods. This summary is not meant algorithms. A genetic algorithm has four main to be exhaustive, but rather to be indicative of the type elements: the genetic code, a concise representation for of research being conducted. For a more detailed an individual solution; the population, a number of discussion, see Yao (1993) and Schaffer et al. (1992). individual solutions; the fitness function, an evaluation The tutorial is broken into three sections. In of the usefulness of an individual; and the propagation the first section, a brief introduction to the foundations techniques, a set of methods for generating new of neural networks and genetic algorithms is given. It is individuals. The genetic algorithm works as follows. assumed that the participants have a basic understanding First, a population of individuals is generated by of both fields, and this introduction is designed as a randomly selecting different genes. The fitness of each short refresher. In the second section, a variety of individual is then evaluated, and the propagation approaches to integrating NNs and GAs are presented. techniques are applied to highly fit individuals to In the final section, some of the key research issues are generate a new population - the next generation. The discussed. cycle of evaluate and propagate continues until a satisfactory solution, hopefully optimal, is found. 1.1 Neural Networks In a typical genetic algorithm, the genetic code To set up the terminology for the rest of the is a fixed-length bit string and the population is always a paper, let us review the basics of a neural network. A fixed size. The three most common propagation neural network is a computational model consisting of a techniques are elitism, mutation and crossover. In number of connected elements, known as neurons. A elitism, the exact individual survives into the next neuron is a processing unit that receives input from generation. In mutation, a new individual is created outside the network and/or from other neurons, applies a from an old one by changing a small number of local transformation to that input, and provides a single randomly selected bits in its gene. In crossover, a new output signal which is passed on to other neurons and/or individual is created from two old ones by randomly outside the network. Each of the inputs is modified by a selecting a split point in their genes are creating a new value associated with the connection. This value is gene with the left part from one parent and the right part referred to as the connection strength, or weight, and from another. In any genetic algorithm, the two key roughly speaking, represents how much importance the aspects are the genetic representation and the fitness neuron attaches to that input source. The local function. Together, these determine the type of problem transformation is referred to as the activation function which is being solved and the possible solutions which and is usually sigmoidal in nature. may be generated. A typical neural network is capable of 2. Combining NNs and GAs representing many functions, as proved by Komolgorov’s Theorem, but finding the best network 2.1 Supportive and Collaborative needed to solve a specific problem is a very open-ended Researchers have combined NNs and GAs in a problem. If the developer knows the exact solution number of different ways. Schaffer et al. have noted
  • 2. that these combinations can be classified into one of two better convergence. Better still, since GAs are good at general types - supportive combinations in which the global search but inefficient at local finely tuned search, NN and GA are applied sequentially, and collaborative a hybrid approach combining GAs and gradient descent combinations in which they are applied simultaneously. are attractive. (Yao) In a supportive approach, the GA and the NN are applied to two different stages of the problem. The 2.3 Evolution of architectures most common combination is to use a GA to pre- In the second approach, the GA is used to process the data set that is used to train a NN. For select general structural parameters and the neural instance, the GA may be used to reduce the learning is used separately to trained the network and dimensionality of the data space by eliminating determine its fitness. This includes evolution of both redundant or unnecessary features. Supportive the topology (i.e., connectivity pattern) and activation combinations are not highly interesting since the GA functions of each node, although most work has and NN are used very independently and either can concentrated on the former and little has been done on easily be replaced by an alternative technique. Some the latter. other possible combinations include: using a NN to In architecture evolution, the genetic code can select the starting population for the GA; using a GA to be either a direct or indirect encoding of the network’s analyse the representations of a NN; and using a GA topology. In a direct encoding, each connection is and NN to solve the same problem and integrating their explicitly represented (e.g., a matrix where 1 indicate responses using a voting scheme. (Schaffer et al.) the presence of a connection and 0 indicates no Alternatively, in a collaborative approach, the connection). In an indirect encoding, important GA and NN are integrated into a single system in which parameters of the network are represented and the a population of neural networks is evolved. In other details of the exact connectivity are left to words, the goal of the system is to find the optimal developmental rules (e.g., specify the number of hidden neural network solution. Such collaborative approaches nodes and assume full connectivity between layers). are possible since neural network learning and genetic In both cases, the exact neural network is not algorithms are both form of search. A neural network specified since the weights are determined by the learning rule performs a highly constrained search to initialisation routine and the network’s learning optimise the network’s structure, while a genetic algorithm. Thus, the evaluation of a gene is noisy since algorithm performs a very general population-based it is dependent upon the evaluation of the trained search to find an optimally fit gene. Both are examples network, and the GA finds the best set of architectural of biased search techniques, and “any algorithm that parameters rather than the best neural network. employs a bias to guide its future samples can be mislead in a search space with the right structure. There 2.4 Evolution of learning rules is always an Achilles heal.” (Schaffer et al, p. 4) The In the final approach, the GA is used similarly primary reason researchers have looked at integrating to the evolution of architecture, but a parametric NNs and GAs is the belief that they may compensate for representation of the network’s learning rule is also each other’s search weaknesses. encoded in the gene. The genetic coding of topology in this case is generally indirect. 2.2 Evolution of Connection Weights Evolving learning rules does not refer simply A genetic algorithm can be applied to to adapting learning algorithm parameters (e.g., learning optimising a neural network in a variety of ways. Yao rate, momentum, etc.) but to adapting the learning has indicated three main approaches.- the evolution of functions themselves. This is an area of research which weights, the evolution of topology, and the evolution of has received little attention. “The biggest problem here learning rules. In each case, the GA’s genetic code is how to encode the dynamic behaviour of a learning varies highly. rule into static genotypes. Trying to develop a universal In the first, the GA is used as the learning rule representation scheme which can specify any kind of of the NN. The genetic code is a direct encoding of the dynamic behaviours is clearly impractical let alone the neural network, with each weight being represented prohibitive long computation time required to search explicitly. The population of the GA are all NNs with such a learning rule space.” (Yao, p. 214) the same basic topology, but with different weight values. Mutation and crossover thus affect only the 3. Issues weights of the individuals. A key question in such system is whether to use binary weights or real-valued ones - the latter increases the search space greatly. Collaborative combinations of NNs and GAs Using GAs instead of gradient descent have sparked the interest of a great number of algorithms to train the weights can result in faster and researchers because of their obvious analogy to natural
  • 3. systems. A wide variety of systems have been grammatical encoding has recently received some developed and a number of research issues have been attention. (Gruau, 1994) Grammar encoding is quite considered. powerful since it is compact but can represent a great range of networks. 3.1 The Baldwin Effect In general, one may wonder whether it really is 4. Conclusions of any use to have both neural learning and genetic Neural networks and genetic algorithms are search operating in the same system. Perhaps using just two highly popular areas of research, and integrating genetic search would work given enough time, or both techniques can often lead to highly successful perhaps a very general neural learning technique would learning systems. The participants of this tutorial are be sufficiently powerful. This is quite possibly true, but encouraged to try applying evolutionary neural network an observation from natural systems known as the solutions, or even developing new combinations of their Baldwin Effect provide a clearer answer. own. The Baldwin Effect states that in an evolutionary system, successful genes can propagate References faster, and in some cases only, if the individuals are capable of learning. This principle has been clearly French, R. & Messinger, A. (1994). “Genes, phenes and demonstrated in an artificial evolutionary system by the Baldwin Effect: Learning and evolution in French & Messinger (1994). Thus, an evolutionary a simulated population,” Artificial Life IV, system with simple individuals which can learn is 277-282. generally more successful than one with non-learning Gruau, F. (1994) “Automatic definition of modular individuals and probably also better than a single highly neural networks,” Adaptive Behaviour, 3, 151- complex learning individual. 184. Schaffer, D., Whitley, D. & Eshelman, L. (1992) 3.2 Generalisation “Combinations of Genetic Algorithms and In evolving a neural network, attention must be Neural Networks: A survey of the state of the paid to the trade-off between evolutionary fitness and art,” Proceedings of the International generalisation ability. In many tasks, the final network Workshop on Combinations of Genetic is trained on a small set of data and applied to a much Algorithms and Neural Networks. D. Whitley larger set of data. The goal of the learning is actually to and D. Schaffer (Eds.,) Los Alamitos, CA: develop a neural network with the best performance on IEEE Computer Society Press, 1-37. the entire problem and not just the training data. Yao, X. (1993) “Evolutionary artificial neural networks” However, this can easily be overlooked during the International Journal of Neural Systems, 4, development process. 203-222. Thus, one must be careful when evolving neural networks not to select for highly specialised, poorly generalising networks. This is especially true in problem areas which are highly dynamic. 3.3 Encoding Methods The two main properties of an encoding of a neural network in a GA are its compactness and representation capability. A compact encoding is useful since the GA can then be efficiently applied to problems requiring large NN solutions. An encoding should be powerful enough to represent a large class of NNs or else the GA may not generate very good solutions. For instance, direct encoding is generally quite powerful in representation, but not compact, while parameterised encoding is compact, yet often represents a highly restrictive set of structures. The discussion so far has focused on direct encoding and parametric encoding of neural network structure. Other possibilities also exist. In particular,