SlideShare une entreprise Scribd logo
1  sur  12
International Journal of Engineering Research and Development
e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com
Volume 7, Issue 3 (May 2013), PP. 44-55
44
A Comparison of Sofm With K-Mean Clustering And Ann
Pattern Reorganization:A Review
1,
Karishmatyagi,Viet Dadri 2,
Gauravsachan ,Viet Dadri
Abstract:- This paper mainly focus on how the input vectors organize themselves into different patterns,
that is, self-organizing maps, based on various algorithms like incremental batch algorithm, k-mean
clustering methods and linear vector quantization when applied in SOFM algorithm itself. It also throws
light on how SOFM algorithm works and defines various topologies. In this paper we have tried to show
how the map looks like after the number of iterations. Moreover we have compared the three known
strategies of self-organizing feature map and discussed the advantage of one technique over other.
Keywords:- feature maps, linear vector quantization, neural network, clusters, classification, neurons,
competitive layer, neuron-neighbours, network design
1. INTRODUCTION
The self-organizing map methods are used for classifying the unknown samples into some categories
based on their features without having the prior knowledge about the input data.It provides us with a
resultant topological network that resembles the input onto the plane exactly as it resembles in high
dimensional space ,that is, the points or neurons that are near to each other in space are also near each other
onto the plane in topological design. The SOM can thus serve as a cluster analysing tool for very large
datasets. These feature maps can also understand and classify the inputs that they have never met before.
1.1 Multiple Layers of Neurons
An artificial neural network consists of several layers. The first layer is an input layer, the last
layer is known as output layer and all the intermediate or middle layers are referred as hidden layer. Each of
these layers has a weight vectorW, a biasor threshold valueb, and an output matrix or vector a. To show
difference between the weight, threshold andoutput vectors in each layer, the number of thelayer is added as
a superscript to every variable respectively.
The network shown below has R1 number of inputs with S1 neurons in the first layer, S2 neurons in the
second layer and so on.Each layer hasdifferent numbers of neurons. A constant value of 1 is given to
each neuron as a threshold value. The central layers are all hidden layers and the output of one first
hidden layer becomes the input for next hidden layer. Thus layer
2 can be analysed as a one-layer network with S1 inputs, S2 neurons, and an S2 × S1 weight matrix
W2.Second layer has input a1 but output is as a2.Since we know all the inputs and outputs of this layer, it
alone can be treated as single-layer network. Multiple-layer networks are very strong and powerful. If we
have a network of two layers where the first layer is sigmoid and
Fig1: Multiple layer neural network
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
45
the second layer is linear, then with the help of multilayer networks and backpropogation algorithms ,this can be
combination of any input that can be transformed into any function although with a finite number of
discontinuities arbitrarily.
1.2 Self-organizing in neural networks
Self-organizing neural networks are designed to detect regularities and correlationsin the input they receive
and adapt their future responses to that input accordingly. The strategy followed by self- organizing maps to
recognize the clusters of similar type of inputvectors is that the way neurons physically near each other in the
neuronlayer respond to similar input vectors. Self-organizing maps do not havetarget vectors. They simply
divide the input vectors into clustersof similar vectors. There is no desired output for these types of networks.
1.3Important Self-Organizing functions
We can create competitive layers and self- organizing maps with competlayer and selforgmap, respectively
in Matlab.
1.4 Architecture of competitive layer(a base for SOM layer)
The n dist box takes the input vector p and weight vectorIW1,1, and calculates the negative of the
distances between them and produces the result as a vector having S1 elements.
Fig2: Competitive layer
The net input n1 is calculated by adding the threshold value to the calculated S1 value. If the
value of p is equals to the neuron’s weight vector andall biases are zero, then a neuron will have
maximum net input 0.The competitive transfer C gives the output as 0 for all the neurons accept
for the neurons that have maximum positive value, that is n1.For this the neuron has output as 1
which is declared as winner.If all the threshold values are 0 ,but the weight vector is not equal to the
input vector p then the neuron whose weight vector and the input vector has the least negative
distance are declared as winner with output 1.
2. SELF-ORGANIZING FEATURE MAPS
Self-organizing feature maps (SOFM) classify the input vectors on the basis of their grouping in the input space
and their neighbouring sections. Thus, a self- organizing maps consists of both the information ,that is, about the
distribution of input vectors and their topology.The neurons in the layer of an SOFM are arranged according to
the following topological functions defined below in detail as gridtop, hextop and randtop that organize the
neurons in a grid, hexagonal, or random pattern. Distances between neurons are calculated from their positions
with a distancefunction. Self-organizing feature map network uses the same procedure as of competitive layer to
find the winning neuron i*but together with updating the winning neuron .It also updates the neighbourhood Ni*
(d) of the winning neuron using the Kohonenrule as: iw(q) = iw(q −1) +α(p(q) − iw(q −1)) Here the
neighbourhood Ni* (d) contains the indices for all of the neurons thatlie within a radius d of the winning neuron
i*. Ni (d) = {j,dij≤ d} So,the weights of the neuron that is winner and its nearby neighbours move towards the p
when p is presented.
2.1 Incremental Batch algorithm
In this algorithm, before making any updating ,the whole data set is presented to the network. The algorithm
then determines a winning neuron for each input vector. Then Each weight vector moves to the average position
of all of the input vectors for which it is a winner, or for which it is in the neighbourhood of a winner. In the
figure on right hand side, thefirst diagram shows a two-dimensional neighbourhood of radius d = 1 around
neuron 13. The second diagram shows a neighbourhood of radius d = 2.
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
46
Fig3: Neurons with their respective radius and neighbourhood
These neighbourhoods are:N13 (1) = {8,12, 13, 14, and 18} andN13 (2) = {3, 7, 8, 9, 11, 12, 13, 14, 15, 17, 18,
19, and 23}. The neurons in an SOFM can be arranged into one, two or more dimensionalpatterns.The
performance of the network is not sensitive to the exact shape of the neighbourhoods.
2.1.1 Topologies used in feature map
In Matlab ,the following functions are used to specify the locations of neurons in the network:
1. gridtop
pos = gridtop(2,3)
pos =
0 1 0 1 0 1
0 0 1 1 2 2
It gives u an array of neurons of 2 X 3 size.Here neuron 1 has the position (0,0), neuron 2 has the
position (1,0), andneuron
3 has the position (0,1), etc. Let us take an example:
If we write the following code on matlab window.It will create a gridtop topology of
8X10 size of neurons. pos = gridtop(8,10); plotsom(pos)
The graph will look as:
plotsom(pos) that gives the following graph.
Fig 4:Agridtop topology of 8 X10 size
2. HEXTOP
Similarly a graph of 8X10 set of neurons in a hextop topology is created withthe following code:
pos = hextop(8,10);
plotsom(pos)
that gives the following graph.
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
47
Fig 5: A hextop topology of 8 X10 size
3. RANDTOP
This function creates neurons in an
N-dimensional random pattern. The following code generates a random pattern of neurons.
pos = randtop(2,3)
pos =
00.7620 0.62601.4218 0.0663 0.7862
0.0925 00.49840.6007 1.1222 1.4228
The 8X10 set of neurons in a randtop topology has the following code:
pos = randtop(8,10);
Fig 6: A randtop topology of 8 X10 size
2.1.2 Architecture
Fig 7: The architecture for SOFM
This architecture works on the basic principle of competitive network, except the difference that no threshold
value is used in calculation here.The competitive transfer function produces 1 for output element a1i
corresponding to i*, the winning neuron. All other output elements in a1 are 0.Neurons close to the winning
neuronare updated along with the winning neuron. We can choose anytopology and distance expressions of
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
48
neurons. 2.1.3 Creating a Self-Organizing Map Neural Network The feature maps arecan be created with the
function selforgmap in Matlab window. Thisfunction defines variables used in two phases of learning: •
Ordering-phase learning rate • Ordering-phase steps • Tuning-phase learning rate • Tuning-phase neighbourhood
distance Training: In this phase, the network identifies the winning neuron for all inputs. The weight vectors
then move towards the location of inputs for which they are winner or closest to the winner. The distance or size
of the neighbourhood is changed or updated during trainingthrough two phases. Ordering Phase:The
neighbourhood distancestarts at a given initial distance, and decreases to the tuning neighbourhood distance
(1.0).As we know that distance among the neighbouring neurons decreases during this phase so the neurons
rearrange themselves in the input space in the same way as are arranged physically in the network. Tuning
Phase: This phase lasts for the rest of training or adaption. The neighbourhood sizehas decreased below 1 so
only the winning neuron learns for each sample. One-Dimensional Self-Organizing Map Let us take an example
of 1D map. If there are 100 two-element unit input vectors spread evenly between 0° and90°. angles =
0:0.5*pi/99:0.5*pi; so the data will be plotted with the following code : P = [sin(angles); cos(angles)];
Fig 8: middle stage of 1-D SOM
As shown below, initially all neurons are at the centre of the figure.
Fig 9: starting stage of 1-D SOM
Initially all the weight vectors are concentrated at the centre of the inputvector space.During the training phase
all the weight vectors simultaneously move towards the input vectors.They also become ordered as the
neighbourhood size decreases. Finally thelayer adjusts its weights in such a way that each neuron responds
effectively to a regionof the input space occupied by input vectors. The placement of neighbouring neuron
weight vectors also represents the network design of the input vectors.
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
49
Fig10: last stage of 1-D SOM
After 120 cycles, the self-organizing feature map starts to rearrange even further more. After 500 rounds or
cycles the map shows how the neurons are more cleanly n evenly distributed in input space. Two-Dimensional
Self-Organizing Map First of all, some random input data is created with the following code: P = rands(2,1000);
Fig 11: a plot of these 1000 input vectors.
A 5-by-6 two-dimensional map of 30 neurons is used to classify these inputvectors.
After 120 cycles the SOFM look like as
Fig12: the map after 120 cycles
Fig13:a plot after 5000 cycles
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
50
So after having near about 5000 cycles, the map shows further smooth distribution of neurons in input space.
So we can say a two-dimensional feature map finds the topology of itsinputs’ space even much better.
Although we know that these feature maps doesnot take very long time to arrange themselves so that
neighbouring neurons can also classify themselves, but the map itself takes long time to arrange according
to the distribution of input neurons vector.
2.1.4 Training with the Batch Algorithm
The batch training algorithm is much faster than the simple incremental algorithm, and it is the default
algorithm for SOFM training. The following code creates 6X6
2-D feature map of 36 neurons:
x = simplecluster_dataset net = selforgmap([6 6]); net = train(net,x);
Fig15: After 200 iterations of the batch algorithm, the map is well distributed through the input space.
If we click the SOM sample hit option in the command window, we will get the following
window shown below as output, which shows the data points of each neuron and also shows how
these data points are distributed in input space.
Fig14: Batch trainingalgorithm command window
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
51
The above figure shows the various options that can be performed using batch training
algorithm which is used to create more efficient feature maps. By clicking the various options different
operations can be performed.
Fig 16: data points with their respective neurons.
In a batch training method the process updates the map at every iteration with the help of the
following equation shown below. Some advantages of batch method over the sequential version are:
a) it produces a map faster and b) it does not need a learning rate to converge.
This algorithm runs in the following 4 steps:
1. Retrieval of historical knowledge from a map
2. Starting of the new map.
3. Searching the winning neurons.
4. Updating of weight vectors
All the above steps are performed for each input vector of neurons.
An incremental method to form maps employing the batch training under non- stationary environment
does not need the historical data chunks to form an updated map. Instead, it only
needs some knowledge from the historical maps and the new data in order to be performed. An
experiment using a data set representing different data distribution has been set in earlier research which
shows the topology of the map during training. The map is able to not forget the topology previously
learned from the past data chunks. Therefore, a batch training proposal for non-
stationary environments might be considered.
2.2 K-means algorithm for generating feature maps
It belongs to the category of partitioning algorithms. It minimizesthe reconstruction error by
assigning a nearest neighbour in that compresses the data vectors onto a smaller set of reference vectors.
In this algorithm, the total number of clusters is defined priorly. The new clusters are made by splitting on
the variable with the largest distance. Here the data set is splitted into the selected number of
clusters by maximizing between relative to within - cluster variation. This algorithm is an
iterative procedure that assigns the different data vectors to the predefined number of clusters
that do not overlap each other. In a first trial k-means was used with the specification to produce two
clusters. The data set is splitted into two subsets as:
This figure shows the resulting subsets of clusters.
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
52
Fig 17: k-means parameterized for two clusters
k-means splits the dataset in the same way as a hyperplane cutting through the rings would do. It seems
that k-means is unable to take the nonlinear structure of the data set into account. Since the number of
clusters is a parameter to the k-means algorithm, it is a common practice for exploratory data
analysis to run k-means with several different numbers of clusters. K-means with four pre specified
clusters produces the result given in the below figure :
Fig 18: k-means parameterized to 4 clusters
While two of the resulting clusters contain only data of one ring, the other two, marked with a dot
receptively a cross in
figure 18, contain a mixture of the two rings. In summary it can be said, the k- means algorithm is
unable to classify this dataset correctly.
2.3Artificial Neural-Network Pattern Recognition Technique for creating feature
maps
In ANN architecture the classification procedure is implemented in three phases. All the phases
are described in detail below
Learning phase1-Self-organizing feature map (SOFM): This phase provides the “approximate”
quantization of the input space by adapting the weight vectors of the neurons in the feature map.
A problem with SOFM is that if the weights are initialized at small random values it gives different
results at different runs which is not desired when we are trying to optimize the performance of the
algorithm even more. Let us take the case of physician. He wants to review the classification
result of neuromuscular disorder of the patient caught in the form of EMG signals by
motor unit action potentials. For this case, the motor unit is to be identified first that has EMG signals and
then it has to be classified using self- organizing map with linear vector quantization technique.
In order to avoid the problem in this particular case as an example, the weights of the output nodes
should not initialized at small random values but at 0.0001 using vector quantization method. The stepsof
algorithm are shown in the following figure below:
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
53
Fig 19: algorithm steps for SOFM Learning phase 2-Learning vector
quantization (LVQ):
It adapt the weights vectors slightly to optimize the classification performance. For this the vector
quantization demands the knowledge of correctly classified inputs. The adaptation carried out during
the first learning phase should be correct and the segmented inputs should be correctly classified.
Updatingthe weight and winning neuron selection is done same as in phase 1. The implementation steps are
given in the following figure below:
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
54
Fig 20: Implementation steps for SOFM applied with LVQ for the particular case.
correctly with different input patterns. But the neural networktechnique for developing feature
map supported by linear vector quantization method is fast and efficient mechanism, although
we have tried to explain with a particular case but it is so..The use of theLVQ algorithm with the SOFM
algorithm optimizes the classification boundaries.It makes the algorithm fast and suitable for real-time
applications.
Phase 3: Classification:
This is the last phase where all the input vectors are defined to one of the output nodes.
The use of the LVQ algorithm with the SOFM algorithm optimizes the classification
boundaries through slight adaptation of the weights vectors in ANN pattern reorganization method.
III.CONCLUSION:
Through this paper we have tried to compare the three mainself- organizing feature
mapsdeveloping techniques.We have studied the batch training method in detail together with the
topological functions itsupportsand shown how the vectors organize themselves into different patterns at
each iteration. The k-mean algorithm we studied also develops the feature maps but it is not as efficient as
incremental batch algorithm as it is unable to classify the dataset
REFERENCES:
[1]. International Journal of Engineering and Advanced Technology (IJEAT) ISSN: 2249 –8958,
Volume-1.
[2]. Ieee transactions on neural networks, vol. 11, no. 3, may 2000
[3]. <html><meta http-equiv="content- type”
[4]. content="text/html;charset=utf-8">
[5]. <CITE>genome.tugraz.at/MedicalI nformatics2/<B>SOM</B>.pdf</CI TE>
[6]. Tamkang Journal of Science and Engineering, Vol. 5, No. 1, pp. 35-48 (2002)
[7]. [Char92] Charalambous, C.,“Conjugate gradient algorithm forefficienttraining of
artificial neural networks,” IEEE Proceedings, Vol. 139, No. 3,1992, pp. 301–310.
[8]. [Koho97] Kohonen, T., Self- Organizing Maps, Second Edition, Berlin:Springer- Verlag, 1997.
[9]. [Lipp87] Lippman, R.P., “An Introduction to computing with
AComparison Of SOFM With K-Mean Clustering And ANN Pattern …
55
[10]. neural nets,”IEEE ASSP Magazine, 1987, pp. 4–22.
[11]. Neural network toolbox using Matlab
[12]. Internet explorer- google search.

Contenu connexe

Tendances

Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyayabhishek upadhyay
 
Section5 Rbf
Section5 RbfSection5 Rbf
Section5 Rbfkylin
 
Paper id 26201482
Paper id 26201482Paper id 26201482
Paper id 26201482IJRAT
 
15 Machine Learning Multilayer Perceptron
15 Machine Learning Multilayer Perceptron15 Machine Learning Multilayer Perceptron
15 Machine Learning Multilayer PerceptronAndres Mendez-Vazquez
 
Introduction to Applied Machine Learning
Introduction to Applied Machine LearningIntroduction to Applied Machine Learning
Introduction to Applied Machine LearningSheilaJimenezMorejon
 
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagationKrish_ver2
 
Fixed-Point Code Synthesis for Neural Networks
Fixed-Point Code Synthesis for Neural NetworksFixed-Point Code Synthesis for Neural Networks
Fixed-Point Code Synthesis for Neural Networksgerogepatton
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksFrancesco Collova'
 
Adaptive resonance theory (art)
Adaptive resonance theory (art)Adaptive resonance theory (art)
Adaptive resonance theory (art)Ashutosh Tyagi
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashrisheetal katkar
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...bihira aggrey
 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyayabhishek upadhyay
 
Artificial neural network - Architectures
Artificial neural network - ArchitecturesArtificial neural network - Architectures
Artificial neural network - ArchitecturesErin Brunston
 

Tendances (20)

Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyay
 
MNN
MNNMNN
MNN
 
Neural networks
Neural networksNeural networks
Neural networks
 
Sefl Organizing Map
Sefl Organizing MapSefl Organizing Map
Sefl Organizing Map
 
Multi Layer Network
Multi Layer NetworkMulti Layer Network
Multi Layer Network
 
Section5 Rbf
Section5 RbfSection5 Rbf
Section5 Rbf
 
Paper id 26201482
Paper id 26201482Paper id 26201482
Paper id 26201482
 
15 Machine Learning Multilayer Perceptron
15 Machine Learning Multilayer Perceptron15 Machine Learning Multilayer Perceptron
15 Machine Learning Multilayer Perceptron
 
Introduction to Applied Machine Learning
Introduction to Applied Machine LearningIntroduction to Applied Machine Learning
Introduction to Applied Machine Learning
 
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagation
 
Associative memory network
Associative memory networkAssociative memory network
Associative memory network
 
Fixed-Point Code Synthesis for Neural Networks
Fixed-Point Code Synthesis for Neural NetworksFixed-Point Code Synthesis for Neural Networks
Fixed-Point Code Synthesis for Neural Networks
 
Max net
Max netMax net
Max net
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
 
Adaptive resonance theory (art)
Adaptive resonance theory (art)Adaptive resonance theory (art)
Adaptive resonance theory (art)
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashri
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
 
Artificial neural network - Architectures
Artificial neural network - ArchitecturesArtificial neural network - Architectures
Artificial neural network - Architectures
 

En vedette

Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...
Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...
Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...ArchiLab 7
 
Rauber, a. 1999: label_som_on the labeling of self-organising maps
Rauber, a. 1999:  label_som_on the labeling of self-organising mapsRauber, a. 1999:  label_som_on the labeling of self-organising maps
Rauber, a. 1999: label_som_on the labeling of self-organising mapsArchiLab 7
 
Information broucher 2014
Information broucher   2014Information broucher   2014
Information broucher 2014Siva Kiran
 
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentations
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentationsPaths of Wellbeing on Self-Organizing Maps + excerpts from other presentations
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentationsTimo Honkela
 
0c960528cf0be4df8f000000
0c960528cf0be4df8f0000000c960528cf0be4df8f000000
0c960528cf0be4df8f000000ELIMENG
 
Intrusion Detection System using Self Organizing Map: A Survey
Intrusion Detection System using Self Organizing Map: A SurveyIntrusion Detection System using Self Organizing Map: A Survey
Intrusion Detection System using Self Organizing Map: A SurveyIJERA Editor
 
Graph mining with kernel self-organizing map
Graph mining with kernel self-organizing mapGraph mining with kernel self-organizing map
Graph mining with kernel self-organizing maptuxette
 
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...ArchiLab 7
 
Van hulle springer:som
Van hulle springer:somVan hulle springer:som
Van hulle springer:somArchiLab 7
 
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage Problems
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage ProblemsEvolving a Team of Self-organizing UAVs to Address Spatial Coverage Problems
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage ProblemsWilfried Elmenreich
 
Food Recommendation System Using Clustering Analysis for Diabetic patients
Food Recommendation System Using Clustering Analysis for Diabetic patientsFood Recommendation System Using Clustering Analysis for Diabetic patients
Food Recommendation System Using Clustering Analysis for Diabetic patientsMaiyaporn Phanich
 
Multiple kernel Self-Organizing Maps
Multiple kernel Self-Organizing MapsMultiple kernel Self-Organizing Maps
Multiple kernel Self-Organizing Mapstuxette
 
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...Antonio Mora
 
Heterogeneous data fusion with multiple kernel growing self organizing maps
Heterogeneous data fusion with multiple kernel growing self organizing mapsHeterogeneous data fusion with multiple kernel growing self organizing maps
Heterogeneous data fusion with multiple kernel growing self organizing mapsPruthuvi Maheshakya Wijewardena
 
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...Alberto Pérez
 
Sociocracy 3.0 - Patterns for Self-organizing Teams
Sociocracy 3.0 - Patterns for Self-organizing TeamsSociocracy 3.0 - Patterns for Self-organizing Teams
Sociocracy 3.0 - Patterns for Self-organizing TeamsBernhard Bockelbrink
 
Bio inspired computational techniques applied to the analysis and visualizati...
Bio inspired computational techniques applied to the analysis and visualizati...Bio inspired computational techniques applied to the analysis and visualizati...
Bio inspired computational techniques applied to the analysis and visualizati...askroll
 

En vedette (20)

Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...
Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...
Honkela.t leinonen.t lonka.k_raike.a_2000: self-organizing maps and construct...
 
Rauber, a. 1999: label_som_on the labeling of self-organising maps
Rauber, a. 1999:  label_som_on the labeling of self-organising mapsRauber, a. 1999:  label_som_on the labeling of self-organising maps
Rauber, a. 1999: label_som_on the labeling of self-organising maps
 
Information broucher 2014
Information broucher   2014Information broucher   2014
Information broucher 2014
 
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentations
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentationsPaths of Wellbeing on Self-Organizing Maps + excerpts from other presentations
Paths of Wellbeing on Self-Organizing Maps + excerpts from other presentations
 
0c960528cf0be4df8f000000
0c960528cf0be4df8f0000000c960528cf0be4df8f000000
0c960528cf0be4df8f000000
 
Intrusion Detection System using Self Organizing Map: A Survey
Intrusion Detection System using Self Organizing Map: A SurveyIntrusion Detection System using Self Organizing Map: A Survey
Intrusion Detection System using Self Organizing Map: A Survey
 
Graph mining with kernel self-organizing map
Graph mining with kernel self-organizing mapGraph mining with kernel self-organizing map
Graph mining with kernel self-organizing map
 
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...
Fessant aknin oukhellou_midenet_2001:comparison_of_supervised_self_organizing...
 
Van hulle springer:som
Van hulle springer:somVan hulle springer:som
Van hulle springer:som
 
Ultra Fast SOM using CUDA
Ultra Fast SOM using CUDAUltra Fast SOM using CUDA
Ultra Fast SOM using CUDA
 
Know4BusinessMT_01
Know4BusinessMT_01Know4BusinessMT_01
Know4BusinessMT_01
 
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage Problems
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage ProblemsEvolving a Team of Self-organizing UAVs to Address Spatial Coverage Problems
Evolving a Team of Self-organizing UAVs to Address Spatial Coverage Problems
 
Food Recommendation System Using Clustering Analysis for Diabetic patients
Food Recommendation System Using Clustering Analysis for Diabetic patientsFood Recommendation System Using Clustering Analysis for Diabetic patients
Food Recommendation System Using Clustering Analysis for Diabetic patients
 
Multiple kernel Self-Organizing Maps
Multiple kernel Self-Organizing MapsMultiple kernel Self-Organizing Maps
Multiple kernel Self-Organizing Maps
 
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...
KohonAnts: A Self-Organizing Ant Algorithm for Clustering and Pattern Classif...
 
Brain computer interface
Brain computer interfaceBrain computer interface
Brain computer interface
 
Heterogeneous data fusion with multiple kernel growing self organizing maps
Heterogeneous data fusion with multiple kernel growing self organizing mapsHeterogeneous data fusion with multiple kernel growing self organizing maps
Heterogeneous data fusion with multiple kernel growing self organizing maps
 
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...
Web Page Clustering Using a Fuzzy Logic Based Representation and Self-Organiz...
 
Sociocracy 3.0 - Patterns for Self-organizing Teams
Sociocracy 3.0 - Patterns for Self-organizing TeamsSociocracy 3.0 - Patterns for Self-organizing Teams
Sociocracy 3.0 - Patterns for Self-organizing Teams
 
Bio inspired computational techniques applied to the analysis and visualizati...
Bio inspired computational techniques applied to the analysis and visualizati...Bio inspired computational techniques applied to the analysis and visualizati...
Bio inspired computational techniques applied to the analysis and visualizati...
 

Similaire à International Journal of Engineering Research and Development (IJERD)

Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...
Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...
Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...ijtsrd
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfSowmyaJyothi3
 
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...iosrjce
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORijac123
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsaciijournal
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)EdutechLearners
 
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATIONA STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATIONADEIJ Journal
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithmsaciijournal
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkNaren Chandra Kattla
 
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...IJERA Editor
 
Best data science course in pune. converted
Best data science course in pune. convertedBest data science course in pune. converted
Best data science course in pune. convertedsripadojwarumavilas
 
Echo state networks and locomotion patterns
Echo state networks and locomotion patternsEcho state networks and locomotion patterns
Echo state networks and locomotion patternsVito Strano
 
Soft computing based cryptographic technique using kohonen's selforganizing m...
Soft computing based cryptographic technique using kohonen's selforganizing m...Soft computing based cryptographic technique using kohonen's selforganizing m...
Soft computing based cryptographic technique using kohonen's selforganizing m...ijfcstjournal
 
Machine learning certification in gurgaon
Machine learning certification in gurgaon Machine learning certification in gurgaon
Machine learning certification in gurgaon TejaspathiLV
 
Data science training ang placements
Data science training ang placementsData science training ang placements
Data science training ang placementsbhuvan8999
 
data science course
data science coursedata science course
data science coursemarketer1234
 
machine learning training in bangalore
machine learning training in bangalore machine learning training in bangalore
machine learning training in bangalore kalojambhu
 
data science course in pune
data science course in punedata science course in pune
data science course in punemarketer1234
 

Similaire à International Journal of Engineering Research and Development (IJERD) (20)

Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...
Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...
Mobile Network Coverage Determination at 900MHz for Abuja Rural Areas using A...
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
 
H017376369
H017376369H017376369
H017376369
 
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
 
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATIONA STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
A STUDY OF METHODS FOR TRAINING WITH DIFFERENT DATASETS IN IMAGE CLASSIFICATION
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural Network
 
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
 
Best data science course in pune. converted
Best data science course in pune. convertedBest data science course in pune. converted
Best data science course in pune. converted
 
Echo state networks and locomotion patterns
Echo state networks and locomotion patternsEcho state networks and locomotion patterns
Echo state networks and locomotion patterns
 
ANN - UNIT 5.pptx
ANN - UNIT 5.pptxANN - UNIT 5.pptx
ANN - UNIT 5.pptx
 
Soft computing based cryptographic technique using kohonen's selforganizing m...
Soft computing based cryptographic technique using kohonen's selforganizing m...Soft computing based cryptographic technique using kohonen's selforganizing m...
Soft computing based cryptographic technique using kohonen's selforganizing m...
 
Machine learning certification in gurgaon
Machine learning certification in gurgaon Machine learning certification in gurgaon
Machine learning certification in gurgaon
 
Data science training ang placements
Data science training ang placementsData science training ang placements
Data science training ang placements
 
data science course
data science coursedata science course
data science course
 
machine learning training in bangalore
machine learning training in bangalore machine learning training in bangalore
machine learning training in bangalore
 
data science course in pune
data science course in punedata science course in pune
data science course in pune
 

Plus de IJERD Editor

A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksA Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksIJERD Editor
 
MEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEMEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEIJERD Editor
 
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...IJERD Editor
 
Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’IJERD Editor
 
Reducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignReducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignIJERD Editor
 
Router 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationRouter 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationIJERD Editor
 
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...IJERD Editor
 
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRMitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRIJERD Editor
 
Study on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingStudy on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingIJERD Editor
 
Spyware triggering system by particular string value
Spyware triggering system by particular string valueSpyware triggering system by particular string value
Spyware triggering system by particular string valueIJERD Editor
 
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...IJERD Editor
 
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeSecure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeIJERD Editor
 
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...IJERD Editor
 
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraGesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraIJERD Editor
 
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...IJERD Editor
 
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...IJERD Editor
 
Moon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingMoon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingIJERD Editor
 
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...IJERD Editor
 
Importance of Measurements in Smart Grid
Importance of Measurements in Smart GridImportance of Measurements in Smart Grid
Importance of Measurements in Smart GridIJERD Editor
 
Study of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderStudy of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderIJERD Editor
 

Plus de IJERD Editor (20)

A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksA Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
 
MEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEMEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACE
 
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
 
Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’
 
Reducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignReducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding Design
 
Router 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationRouter 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and Verification
 
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
 
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRMitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
 
Study on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingStudy on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive Manufacturing
 
Spyware triggering system by particular string value
Spyware triggering system by particular string valueSpyware triggering system by particular string value
Spyware triggering system by particular string value
 
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
 
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeSecure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
 
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
 
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraGesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
 
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
 
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
 
Moon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingMoon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF Dxing
 
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
 
Importance of Measurements in Smart Grid
Importance of Measurements in Smart GridImportance of Measurements in Smart Grid
Importance of Measurements in Smart Grid
 
Study of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderStudy of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powder
 

Dernier

Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEarley Information Science
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)wesley chun
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processorsdebabhi2
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?Antenna Manufacturer Coco
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 

Dernier (20)

Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptxEIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
EIS-Webinar-Prompt-Knowledge-Eng-2024-04-08.pptx
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 

International Journal of Engineering Research and Development (IJERD)

  • 1. International Journal of Engineering Research and Development e-ISSN: 2278-067X, p-ISSN: 2278-800X, www.ijerd.com Volume 7, Issue 3 (May 2013), PP. 44-55 44 A Comparison of Sofm With K-Mean Clustering And Ann Pattern Reorganization:A Review 1, Karishmatyagi,Viet Dadri 2, Gauravsachan ,Viet Dadri Abstract:- This paper mainly focus on how the input vectors organize themselves into different patterns, that is, self-organizing maps, based on various algorithms like incremental batch algorithm, k-mean clustering methods and linear vector quantization when applied in SOFM algorithm itself. It also throws light on how SOFM algorithm works and defines various topologies. In this paper we have tried to show how the map looks like after the number of iterations. Moreover we have compared the three known strategies of self-organizing feature map and discussed the advantage of one technique over other. Keywords:- feature maps, linear vector quantization, neural network, clusters, classification, neurons, competitive layer, neuron-neighbours, network design 1. INTRODUCTION The self-organizing map methods are used for classifying the unknown samples into some categories based on their features without having the prior knowledge about the input data.It provides us with a resultant topological network that resembles the input onto the plane exactly as it resembles in high dimensional space ,that is, the points or neurons that are near to each other in space are also near each other onto the plane in topological design. The SOM can thus serve as a cluster analysing tool for very large datasets. These feature maps can also understand and classify the inputs that they have never met before. 1.1 Multiple Layers of Neurons An artificial neural network consists of several layers. The first layer is an input layer, the last layer is known as output layer and all the intermediate or middle layers are referred as hidden layer. Each of these layers has a weight vectorW, a biasor threshold valueb, and an output matrix or vector a. To show difference between the weight, threshold andoutput vectors in each layer, the number of thelayer is added as a superscript to every variable respectively. The network shown below has R1 number of inputs with S1 neurons in the first layer, S2 neurons in the second layer and so on.Each layer hasdifferent numbers of neurons. A constant value of 1 is given to each neuron as a threshold value. The central layers are all hidden layers and the output of one first hidden layer becomes the input for next hidden layer. Thus layer 2 can be analysed as a one-layer network with S1 inputs, S2 neurons, and an S2 × S1 weight matrix W2.Second layer has input a1 but output is as a2.Since we know all the inputs and outputs of this layer, it alone can be treated as single-layer network. Multiple-layer networks are very strong and powerful. If we have a network of two layers where the first layer is sigmoid and Fig1: Multiple layer neural network
  • 2. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 45 the second layer is linear, then with the help of multilayer networks and backpropogation algorithms ,this can be combination of any input that can be transformed into any function although with a finite number of discontinuities arbitrarily. 1.2 Self-organizing in neural networks Self-organizing neural networks are designed to detect regularities and correlationsin the input they receive and adapt their future responses to that input accordingly. The strategy followed by self- organizing maps to recognize the clusters of similar type of inputvectors is that the way neurons physically near each other in the neuronlayer respond to similar input vectors. Self-organizing maps do not havetarget vectors. They simply divide the input vectors into clustersof similar vectors. There is no desired output for these types of networks. 1.3Important Self-Organizing functions We can create competitive layers and self- organizing maps with competlayer and selforgmap, respectively in Matlab. 1.4 Architecture of competitive layer(a base for SOM layer) The n dist box takes the input vector p and weight vectorIW1,1, and calculates the negative of the distances between them and produces the result as a vector having S1 elements. Fig2: Competitive layer The net input n1 is calculated by adding the threshold value to the calculated S1 value. If the value of p is equals to the neuron’s weight vector andall biases are zero, then a neuron will have maximum net input 0.The competitive transfer C gives the output as 0 for all the neurons accept for the neurons that have maximum positive value, that is n1.For this the neuron has output as 1 which is declared as winner.If all the threshold values are 0 ,but the weight vector is not equal to the input vector p then the neuron whose weight vector and the input vector has the least negative distance are declared as winner with output 1. 2. SELF-ORGANIZING FEATURE MAPS Self-organizing feature maps (SOFM) classify the input vectors on the basis of their grouping in the input space and their neighbouring sections. Thus, a self- organizing maps consists of both the information ,that is, about the distribution of input vectors and their topology.The neurons in the layer of an SOFM are arranged according to the following topological functions defined below in detail as gridtop, hextop and randtop that organize the neurons in a grid, hexagonal, or random pattern. Distances between neurons are calculated from their positions with a distancefunction. Self-organizing feature map network uses the same procedure as of competitive layer to find the winning neuron i*but together with updating the winning neuron .It also updates the neighbourhood Ni* (d) of the winning neuron using the Kohonenrule as: iw(q) = iw(q −1) +α(p(q) − iw(q −1)) Here the neighbourhood Ni* (d) contains the indices for all of the neurons thatlie within a radius d of the winning neuron i*. Ni (d) = {j,dij≤ d} So,the weights of the neuron that is winner and its nearby neighbours move towards the p when p is presented. 2.1 Incremental Batch algorithm In this algorithm, before making any updating ,the whole data set is presented to the network. The algorithm then determines a winning neuron for each input vector. Then Each weight vector moves to the average position of all of the input vectors for which it is a winner, or for which it is in the neighbourhood of a winner. In the figure on right hand side, thefirst diagram shows a two-dimensional neighbourhood of radius d = 1 around neuron 13. The second diagram shows a neighbourhood of radius d = 2.
  • 3. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 46 Fig3: Neurons with their respective radius and neighbourhood These neighbourhoods are:N13 (1) = {8,12, 13, 14, and 18} andN13 (2) = {3, 7, 8, 9, 11, 12, 13, 14, 15, 17, 18, 19, and 23}. The neurons in an SOFM can be arranged into one, two or more dimensionalpatterns.The performance of the network is not sensitive to the exact shape of the neighbourhoods. 2.1.1 Topologies used in feature map In Matlab ,the following functions are used to specify the locations of neurons in the network: 1. gridtop pos = gridtop(2,3) pos = 0 1 0 1 0 1 0 0 1 1 2 2 It gives u an array of neurons of 2 X 3 size.Here neuron 1 has the position (0,0), neuron 2 has the position (1,0), andneuron 3 has the position (0,1), etc. Let us take an example: If we write the following code on matlab window.It will create a gridtop topology of 8X10 size of neurons. pos = gridtop(8,10); plotsom(pos) The graph will look as: plotsom(pos) that gives the following graph. Fig 4:Agridtop topology of 8 X10 size 2. HEXTOP Similarly a graph of 8X10 set of neurons in a hextop topology is created withthe following code: pos = hextop(8,10); plotsom(pos) that gives the following graph.
  • 4. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 47 Fig 5: A hextop topology of 8 X10 size 3. RANDTOP This function creates neurons in an N-dimensional random pattern. The following code generates a random pattern of neurons. pos = randtop(2,3) pos = 00.7620 0.62601.4218 0.0663 0.7862 0.0925 00.49840.6007 1.1222 1.4228 The 8X10 set of neurons in a randtop topology has the following code: pos = randtop(8,10); Fig 6: A randtop topology of 8 X10 size 2.1.2 Architecture Fig 7: The architecture for SOFM This architecture works on the basic principle of competitive network, except the difference that no threshold value is used in calculation here.The competitive transfer function produces 1 for output element a1i corresponding to i*, the winning neuron. All other output elements in a1 are 0.Neurons close to the winning neuronare updated along with the winning neuron. We can choose anytopology and distance expressions of
  • 5. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 48 neurons. 2.1.3 Creating a Self-Organizing Map Neural Network The feature maps arecan be created with the function selforgmap in Matlab window. Thisfunction defines variables used in two phases of learning: • Ordering-phase learning rate • Ordering-phase steps • Tuning-phase learning rate • Tuning-phase neighbourhood distance Training: In this phase, the network identifies the winning neuron for all inputs. The weight vectors then move towards the location of inputs for which they are winner or closest to the winner. The distance or size of the neighbourhood is changed or updated during trainingthrough two phases. Ordering Phase:The neighbourhood distancestarts at a given initial distance, and decreases to the tuning neighbourhood distance (1.0).As we know that distance among the neighbouring neurons decreases during this phase so the neurons rearrange themselves in the input space in the same way as are arranged physically in the network. Tuning Phase: This phase lasts for the rest of training or adaption. The neighbourhood sizehas decreased below 1 so only the winning neuron learns for each sample. One-Dimensional Self-Organizing Map Let us take an example of 1D map. If there are 100 two-element unit input vectors spread evenly between 0° and90°. angles = 0:0.5*pi/99:0.5*pi; so the data will be plotted with the following code : P = [sin(angles); cos(angles)]; Fig 8: middle stage of 1-D SOM As shown below, initially all neurons are at the centre of the figure. Fig 9: starting stage of 1-D SOM Initially all the weight vectors are concentrated at the centre of the inputvector space.During the training phase all the weight vectors simultaneously move towards the input vectors.They also become ordered as the neighbourhood size decreases. Finally thelayer adjusts its weights in such a way that each neuron responds effectively to a regionof the input space occupied by input vectors. The placement of neighbouring neuron weight vectors also represents the network design of the input vectors.
  • 6. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 49 Fig10: last stage of 1-D SOM After 120 cycles, the self-organizing feature map starts to rearrange even further more. After 500 rounds or cycles the map shows how the neurons are more cleanly n evenly distributed in input space. Two-Dimensional Self-Organizing Map First of all, some random input data is created with the following code: P = rands(2,1000); Fig 11: a plot of these 1000 input vectors. A 5-by-6 two-dimensional map of 30 neurons is used to classify these inputvectors. After 120 cycles the SOFM look like as Fig12: the map after 120 cycles Fig13:a plot after 5000 cycles
  • 7. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 50 So after having near about 5000 cycles, the map shows further smooth distribution of neurons in input space. So we can say a two-dimensional feature map finds the topology of itsinputs’ space even much better. Although we know that these feature maps doesnot take very long time to arrange themselves so that neighbouring neurons can also classify themselves, but the map itself takes long time to arrange according to the distribution of input neurons vector. 2.1.4 Training with the Batch Algorithm The batch training algorithm is much faster than the simple incremental algorithm, and it is the default algorithm for SOFM training. The following code creates 6X6 2-D feature map of 36 neurons: x = simplecluster_dataset net = selforgmap([6 6]); net = train(net,x); Fig15: After 200 iterations of the batch algorithm, the map is well distributed through the input space. If we click the SOM sample hit option in the command window, we will get the following window shown below as output, which shows the data points of each neuron and also shows how these data points are distributed in input space. Fig14: Batch trainingalgorithm command window
  • 8. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 51 The above figure shows the various options that can be performed using batch training algorithm which is used to create more efficient feature maps. By clicking the various options different operations can be performed. Fig 16: data points with their respective neurons. In a batch training method the process updates the map at every iteration with the help of the following equation shown below. Some advantages of batch method over the sequential version are: a) it produces a map faster and b) it does not need a learning rate to converge. This algorithm runs in the following 4 steps: 1. Retrieval of historical knowledge from a map 2. Starting of the new map. 3. Searching the winning neurons. 4. Updating of weight vectors All the above steps are performed for each input vector of neurons. An incremental method to form maps employing the batch training under non- stationary environment does not need the historical data chunks to form an updated map. Instead, it only needs some knowledge from the historical maps and the new data in order to be performed. An experiment using a data set representing different data distribution has been set in earlier research which shows the topology of the map during training. The map is able to not forget the topology previously learned from the past data chunks. Therefore, a batch training proposal for non- stationary environments might be considered. 2.2 K-means algorithm for generating feature maps It belongs to the category of partitioning algorithms. It minimizesthe reconstruction error by assigning a nearest neighbour in that compresses the data vectors onto a smaller set of reference vectors. In this algorithm, the total number of clusters is defined priorly. The new clusters are made by splitting on the variable with the largest distance. Here the data set is splitted into the selected number of clusters by maximizing between relative to within - cluster variation. This algorithm is an iterative procedure that assigns the different data vectors to the predefined number of clusters that do not overlap each other. In a first trial k-means was used with the specification to produce two clusters. The data set is splitted into two subsets as: This figure shows the resulting subsets of clusters.
  • 9. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 52 Fig 17: k-means parameterized for two clusters k-means splits the dataset in the same way as a hyperplane cutting through the rings would do. It seems that k-means is unable to take the nonlinear structure of the data set into account. Since the number of clusters is a parameter to the k-means algorithm, it is a common practice for exploratory data analysis to run k-means with several different numbers of clusters. K-means with four pre specified clusters produces the result given in the below figure : Fig 18: k-means parameterized to 4 clusters While two of the resulting clusters contain only data of one ring, the other two, marked with a dot receptively a cross in figure 18, contain a mixture of the two rings. In summary it can be said, the k- means algorithm is unable to classify this dataset correctly. 2.3Artificial Neural-Network Pattern Recognition Technique for creating feature maps In ANN architecture the classification procedure is implemented in three phases. All the phases are described in detail below Learning phase1-Self-organizing feature map (SOFM): This phase provides the “approximate” quantization of the input space by adapting the weight vectors of the neurons in the feature map. A problem with SOFM is that if the weights are initialized at small random values it gives different results at different runs which is not desired when we are trying to optimize the performance of the algorithm even more. Let us take the case of physician. He wants to review the classification result of neuromuscular disorder of the patient caught in the form of EMG signals by motor unit action potentials. For this case, the motor unit is to be identified first that has EMG signals and then it has to be classified using self- organizing map with linear vector quantization technique. In order to avoid the problem in this particular case as an example, the weights of the output nodes should not initialized at small random values but at 0.0001 using vector quantization method. The stepsof algorithm are shown in the following figure below:
  • 10. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 53 Fig 19: algorithm steps for SOFM Learning phase 2-Learning vector quantization (LVQ): It adapt the weights vectors slightly to optimize the classification performance. For this the vector quantization demands the knowledge of correctly classified inputs. The adaptation carried out during the first learning phase should be correct and the segmented inputs should be correctly classified. Updatingthe weight and winning neuron selection is done same as in phase 1. The implementation steps are given in the following figure below:
  • 11. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 54 Fig 20: Implementation steps for SOFM applied with LVQ for the particular case. correctly with different input patterns. But the neural networktechnique for developing feature map supported by linear vector quantization method is fast and efficient mechanism, although we have tried to explain with a particular case but it is so..The use of theLVQ algorithm with the SOFM algorithm optimizes the classification boundaries.It makes the algorithm fast and suitable for real-time applications. Phase 3: Classification: This is the last phase where all the input vectors are defined to one of the output nodes. The use of the LVQ algorithm with the SOFM algorithm optimizes the classification boundaries through slight adaptation of the weights vectors in ANN pattern reorganization method. III.CONCLUSION: Through this paper we have tried to compare the three mainself- organizing feature mapsdeveloping techniques.We have studied the batch training method in detail together with the topological functions itsupportsand shown how the vectors organize themselves into different patterns at each iteration. The k-mean algorithm we studied also develops the feature maps but it is not as efficient as incremental batch algorithm as it is unable to classify the dataset REFERENCES: [1]. International Journal of Engineering and Advanced Technology (IJEAT) ISSN: 2249 –8958, Volume-1. [2]. Ieee transactions on neural networks, vol. 11, no. 3, may 2000 [3]. <html><meta http-equiv="content- type” [4]. content="text/html;charset=utf-8"> [5]. <CITE>genome.tugraz.at/MedicalI nformatics2/<B>SOM</B>.pdf</CI TE> [6]. Tamkang Journal of Science and Engineering, Vol. 5, No. 1, pp. 35-48 (2002) [7]. [Char92] Charalambous, C.,“Conjugate gradient algorithm forefficienttraining of artificial neural networks,” IEEE Proceedings, Vol. 139, No. 3,1992, pp. 301–310. [8]. [Koho97] Kohonen, T., Self- Organizing Maps, Second Edition, Berlin:Springer- Verlag, 1997. [9]. [Lipp87] Lippman, R.P., “An Introduction to computing with
  • 12. AComparison Of SOFM With K-Mean Clustering And ANN Pattern … 55 [10]. neural nets,”IEEE ASSP Magazine, 1987, pp. 4–22. [11]. Neural network toolbox using Matlab [12]. Internet explorer- google search.